[ 150.887733] ====================================================== [ 150.893903] WARNING: possible circular locking dependency detected [ 150.905917] ------------------------------------------------------ [ 150.912129] kfdtest/4081 is trying to acquire lock: [ 150.917002] ffff8f7f3762e118 (&mm->mmap_sem#2){++++}, at: __might_fault+0x3e/0x90 [ 150.924490] but task is already holding lock: [ 150.930320] ffff8f7f49d229e8 (&dqm->lock_hidden){+.+.}, at: destroy_queue_cpsch+0x29/0x210 [amdgpu] [ 150.939432] which lock already depends on the new lock. [ 150.947603] the existing dependency chain (in reverse order) is: [ 150.955074] -> #3 (&dqm->lock_hidden){+.+.}: [ 150.960822] __mutex_lock+0xa1/0x9f0 [ 150.964996] evict_process_queues_cpsch+0x22/0x120 [amdgpu] [ 150.971155] kfd_process_evict_queues+0x3b/0xc0 [amdgpu] [ 150.977054] kgd2kfd_quiesce_mm+0x25/0x60 [amdgpu] [ 150.982442] amdgpu_amdkfd_evict_userptr+0x35/0x70 [amdgpu] [ 150.988615] amdgpu_mn_invalidate_hsa+0x41/0x60 [amdgpu] [ 150.994448] __mmu_notifier_invalidate_range_start+0xa4/0x240 [ 151.000714] copy_page_range+0xd70/0xd80 [ 151.005159] dup_mm+0x3ca/0x550 [ 151.008816] copy_process+0x1bdc/0x1c70 [ 151.013183] _do_fork+0x76/0x6c0 [ 151.016929] __x64_sys_clone+0x8c/0xb0 [ 151.021201] do_syscall_64+0x4a/0x1d0 [ 151.025404] entry_SYSCALL_64_after_hwframe+0x49/0xbe [ 151.030977] -> #2 (&adev->notifier_lock){+.+.}: [ 151.036993] __mutex_lock+0xa1/0x9f0 [ 151.041168] amdgpu_mn_invalidate_hsa+0x30/0x60 [amdgpu] [ 151.047019] __mmu_notifier_invalidate_range_start+0xa4/0x240 [ 151.053277] copy_page_range+0xd70/0xd80 [ 151.057722] dup_mm+0x3ca/0x550 [ 151.061388] copy_process+0x1bdc/0x1c70 [ 151.065748] _do_fork+0x76/0x6c0 [ 151.069499] __x64_sys_clone+0x8c/0xb0 [ 151.073765] do_syscall_64+0x4a/0x1d0 [ 151.077952] entry_SYSCALL_64_after_hwframe+0x49/0xbe [ 151.083523] -> #1 (mmu_notifier_invalidate_range_start){+.+.}: [ 151.090833] change_protection+0x802/0xab0 [ 151.095448] mprotect_fixup+0x187/0x2d0 [ 151.099801] setup_arg_pages+0x124/0x250 [ 151.104251] load_elf_binary+0x3a4/0x1464 [ 151.108781] search_binary_handler+0x6c/0x210 [ 151.113656] __do_execve_file.isra.40+0x7f7/0xa50 [ 151.118875] do_execve+0x21/0x30 [ 151.122632] call_usermodehelper_exec_async+0x17e/0x190 [ 151.128393] ret_from_fork+0x24/0x30 [ 151.132489] -> #0 (&mm->mmap_sem#2){++++}: [ 151.138064] __lock_acquire+0x11a1/0x1490 [ 151.142597] lock_acquire+0x90/0x180 [ 151.146694] __might_fault+0x68/0x90 [ 151.150879] read_sdma_queue_counter+0x5f/0xb0 [amdgpu] [ 151.156693] update_sdma_queue_past_activity_stats+0x3b/0x90 [amdgpu] [ 151.163725] destroy_queue_cpsch+0x1ae/0x210 [amdgpu] [ 151.169373] pqm_destroy_queue+0xf0/0x250 [amdgpu] [ 151.174762] kfd_ioctl_destroy_queue+0x32/0x70 [amdgpu] [ 151.180577] kfd_ioctl+0x223/0x400 [amdgpu] [ 151.185284] ksys_ioctl+0x8f/0xb0 [ 151.189118] __x64_sys_ioctl+0x16/0x20 [ 151.193389] do_syscall_64+0x4a/0x1d0 [ 151.197569] entry_SYSCALL_64_after_hwframe+0x49/0xbe [ 151.203141] other info that might help us debug this: [ 151.211140] Chain exists of: &mm->mmap_sem#2 --> &adev->notifier_lock --> &dqm->lock_hidden [ 151.222535] Possible unsafe locking scenario: [ 151.228447] CPU0 CPU1 [ 151.232971] ---- ---- [ 151.237502] lock(&dqm->lock_hidden); [ 151.241254] lock(&adev->notifier_lock); [ 151.247774] lock(&dqm->lock_hidden); [ 151.254038] lock(&mm->mmap_sem#2); This commit fixes the warning by ensuring get_user() is not called while reading SDMA stats with dqm_lock held as get_user() could cause a page fault which leads to the circular locking scenario. Signed-off-by: Mukul Joshi <mukul.joshi@amd.com> Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
255 lines
8.7 KiB
C
255 lines
8.7 KiB
C
/*
|
|
* Copyright 2014 Advanced Micro Devices, Inc.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*
|
|
*/
|
|
|
|
#ifndef KFD_DEVICE_QUEUE_MANAGER_H_
|
|
#define KFD_DEVICE_QUEUE_MANAGER_H_
|
|
|
|
#include <linux/rwsem.h>
|
|
#include <linux/list.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/sched/mm.h>
|
|
#include "kfd_priv.h"
|
|
#include "kfd_mqd_manager.h"
|
|
|
|
|
|
#define VMID_NUM 16
|
|
|
|
struct device_process_node {
|
|
struct qcm_process_device *qpd;
|
|
struct list_head list;
|
|
};
|
|
|
|
/**
|
|
* struct device_queue_manager_ops
|
|
*
|
|
* @create_queue: Queue creation routine.
|
|
*
|
|
* @destroy_queue: Queue destruction routine.
|
|
*
|
|
* @update_queue: Queue update routine.
|
|
*
|
|
* @exeute_queues: Dispatches the queues list to the H/W.
|
|
*
|
|
* @register_process: This routine associates a specific process with device.
|
|
*
|
|
* @unregister_process: destroys the associations between process to device.
|
|
*
|
|
* @initialize: Initializes the pipelines and memory module for that device.
|
|
*
|
|
* @start: Initializes the resources/modules the the device needs for queues
|
|
* execution. This function is called on device initialization and after the
|
|
* system woke up after suspension.
|
|
*
|
|
* @stop: This routine stops execution of all the active queue running on the
|
|
* H/W and basically this function called on system suspend.
|
|
*
|
|
* @uninitialize: Destroys all the device queue manager resources allocated in
|
|
* initialize routine.
|
|
*
|
|
* @create_kernel_queue: Creates kernel queue. Used for debug queue.
|
|
*
|
|
* @destroy_kernel_queue: Destroys kernel queue. Used for debug queue.
|
|
*
|
|
* @set_cache_memory_policy: Sets memory policy (cached/ non cached) for the
|
|
* memory apertures.
|
|
*
|
|
* @process_termination: Clears all process queues belongs to that device.
|
|
*
|
|
* @evict_process_queues: Evict all active queues of a process
|
|
*
|
|
* @restore_process_queues: Restore all evicted queues queues of a process
|
|
*
|
|
* @get_wave_state: Retrieves context save state and optionally copies the
|
|
* control stack, if kept in the MQD, to the given userspace address.
|
|
*/
|
|
|
|
struct device_queue_manager_ops {
|
|
int (*create_queue)(struct device_queue_manager *dqm,
|
|
struct queue *q,
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*destroy_queue)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd,
|
|
struct queue *q);
|
|
|
|
int (*update_queue)(struct device_queue_manager *dqm,
|
|
struct queue *q);
|
|
|
|
int (*register_process)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*unregister_process)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*initialize)(struct device_queue_manager *dqm);
|
|
int (*start)(struct device_queue_manager *dqm);
|
|
int (*stop)(struct device_queue_manager *dqm);
|
|
void (*pre_reset)(struct device_queue_manager *dqm);
|
|
void (*uninitialize)(struct device_queue_manager *dqm);
|
|
int (*create_kernel_queue)(struct device_queue_manager *dqm,
|
|
struct kernel_queue *kq,
|
|
struct qcm_process_device *qpd);
|
|
|
|
void (*destroy_kernel_queue)(struct device_queue_manager *dqm,
|
|
struct kernel_queue *kq,
|
|
struct qcm_process_device *qpd);
|
|
|
|
bool (*set_cache_memory_policy)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd,
|
|
enum cache_policy default_policy,
|
|
enum cache_policy alternate_policy,
|
|
void __user *alternate_aperture_base,
|
|
uint64_t alternate_aperture_size);
|
|
|
|
int (*set_trap_handler)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd,
|
|
uint64_t tba_addr,
|
|
uint64_t tma_addr);
|
|
|
|
int (*process_termination)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*evict_process_queues)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
int (*restore_process_queues)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*get_wave_state)(struct device_queue_manager *dqm,
|
|
struct queue *q,
|
|
void __user *ctl_stack,
|
|
u32 *ctl_stack_used_size,
|
|
u32 *save_area_used_size);
|
|
};
|
|
|
|
struct device_queue_manager_asic_ops {
|
|
int (*update_qpd)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
bool (*set_cache_memory_policy)(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd,
|
|
enum cache_policy default_policy,
|
|
enum cache_policy alternate_policy,
|
|
void __user *alternate_aperture_base,
|
|
uint64_t alternate_aperture_size);
|
|
void (*init_sdma_vm)(struct device_queue_manager *dqm,
|
|
struct queue *q,
|
|
struct qcm_process_device *qpd);
|
|
struct mqd_manager * (*mqd_manager_init)(enum KFD_MQD_TYPE type,
|
|
struct kfd_dev *dev);
|
|
};
|
|
|
|
/**
|
|
* struct device_queue_manager
|
|
*
|
|
* This struct is a base class for the kfd queues scheduler in the
|
|
* device level. The device base class should expose the basic operations
|
|
* for queue creation and queue destruction. This base class hides the
|
|
* scheduling mode of the driver and the specific implementation of the
|
|
* concrete device. This class is the only class in the queues scheduler
|
|
* that configures the H/W.
|
|
*
|
|
*/
|
|
|
|
struct device_queue_manager {
|
|
struct device_queue_manager_ops ops;
|
|
struct device_queue_manager_asic_ops asic_ops;
|
|
|
|
struct mqd_manager *mqd_mgrs[KFD_MQD_TYPE_MAX];
|
|
struct packet_manager packets;
|
|
struct kfd_dev *dev;
|
|
struct mutex lock_hidden; /* use dqm_lock/unlock(dqm) */
|
|
struct list_head queues;
|
|
unsigned int saved_flags;
|
|
unsigned int processes_count;
|
|
unsigned int active_queue_count;
|
|
unsigned int active_cp_queue_count;
|
|
unsigned int gws_queue_count;
|
|
unsigned int total_queue_count;
|
|
unsigned int next_pipe_to_allocate;
|
|
unsigned int *allocated_queues;
|
|
uint64_t sdma_bitmap;
|
|
uint64_t xgmi_sdma_bitmap;
|
|
/* the pasid mapping for each kfd vmid */
|
|
uint16_t vmid_pasid[VMID_NUM];
|
|
uint64_t pipelines_addr;
|
|
uint64_t fence_gpu_addr;
|
|
unsigned int *fence_addr;
|
|
struct kfd_mem_obj *fence_mem;
|
|
bool active_runlist;
|
|
int sched_policy;
|
|
|
|
/* hw exception */
|
|
bool is_hws_hang;
|
|
bool is_resetting;
|
|
struct work_struct hw_exception_work;
|
|
struct kfd_mem_obj hiq_sdma_mqd;
|
|
bool sched_running;
|
|
};
|
|
|
|
void device_queue_manager_init_cik(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void device_queue_manager_init_cik_hawaii(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void device_queue_manager_init_vi(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void device_queue_manager_init_vi_tonga(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void device_queue_manager_init_v9(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void device_queue_manager_init_v10_navi10(
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
void program_sh_mem_settings(struct device_queue_manager *dqm,
|
|
struct qcm_process_device *qpd);
|
|
unsigned int get_cp_queues_num(struct device_queue_manager *dqm);
|
|
unsigned int get_queues_per_pipe(struct device_queue_manager *dqm);
|
|
unsigned int get_pipes_per_mec(struct device_queue_manager *dqm);
|
|
unsigned int get_num_sdma_queues(struct device_queue_manager *dqm);
|
|
unsigned int get_num_xgmi_sdma_queues(struct device_queue_manager *dqm);
|
|
|
|
static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd)
|
|
{
|
|
return (pdd->lds_base >> 16) & 0xFF;
|
|
}
|
|
|
|
static inline unsigned int
|
|
get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd)
|
|
{
|
|
return (pdd->lds_base >> 60) & 0x0E;
|
|
}
|
|
|
|
/* The DQM lock can be taken in MMU notifiers. Make sure no reclaim-FS
|
|
* happens while holding this lock anywhere to prevent deadlocks when
|
|
* an MMU notifier runs in reclaim-FS context.
|
|
*/
|
|
static inline void dqm_lock(struct device_queue_manager *dqm)
|
|
{
|
|
mutex_lock(&dqm->lock_hidden);
|
|
dqm->saved_flags = memalloc_nofs_save();
|
|
}
|
|
static inline void dqm_unlock(struct device_queue_manager *dqm)
|
|
{
|
|
memalloc_nofs_restore(dqm->saved_flags);
|
|
mutex_unlock(&dqm->lock_hidden);
|
|
}
|
|
|
|
int read_sdma_queue_counter(uint64_t q_rptr, uint64_t *val);
|
|
#endif /* KFD_DEVICE_QUEUE_MANAGER_H_ */
|