drm/amdgpu: Move kfd_mem_attach outside reservation
This is needed to avoid deadlocks with DMA buf import in the next patch. Also move PT/PD validation out of kfd_mem_attach, that way the caller can bo this unconditionally. Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> Acked-by: Oak Zeng <Oak.Zeng@amd.com> Acked-by: Ramesh Errabolu <Ramesh.Errabolu@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
b72ed8a2de
commit
9e5d275319
1 changed files with 44 additions and 31 deletions
|
@ -582,6 +582,34 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
|
||||||
|
struct amdgpu_bo **bo)
|
||||||
|
{
|
||||||
|
unsigned long bo_size = mem->bo->tbo.base.size;
|
||||||
|
struct drm_gem_object *gobj;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
ret = amdgpu_bo_reserve(mem->bo, false);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
ret = amdgpu_gem_object_create(adev, bo_size, 1,
|
||||||
|
AMDGPU_GEM_DOMAIN_CPU,
|
||||||
|
0, ttm_bo_type_sg,
|
||||||
|
mem->bo->tbo.base.resv,
|
||||||
|
&gobj);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
amdgpu_bo_unreserve(mem->bo);
|
||||||
|
|
||||||
|
*bo = gem_to_amdgpu_bo(gobj);
|
||||||
|
(*bo)->parent = amdgpu_bo_ref(mem->bo);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
/* kfd_mem_attach - Add a BO to a VM
|
/* kfd_mem_attach - Add a BO to a VM
|
||||||
*
|
*
|
||||||
* Everything that needs to bo done only once when a BO is first added
|
* Everything that needs to bo done only once when a BO is first added
|
||||||
|
@ -603,7 +631,6 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
|
||||||
uint64_t va = mem->va;
|
uint64_t va = mem->va;
|
||||||
struct kfd_mem_attachment *attachment[2] = {NULL, NULL};
|
struct kfd_mem_attachment *attachment[2] = {NULL, NULL};
|
||||||
struct amdgpu_bo *bo[2] = {NULL, NULL};
|
struct amdgpu_bo *bo[2] = {NULL, NULL};
|
||||||
struct drm_gem_object *gobj;
|
|
||||||
int i, ret;
|
int i, ret;
|
||||||
|
|
||||||
if (!va) {
|
if (!va) {
|
||||||
|
@ -637,15 +664,9 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
|
||||||
} else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) {
|
} else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) {
|
||||||
/* Create an SG BO to DMA-map userptrs on other GPUs */
|
/* Create an SG BO to DMA-map userptrs on other GPUs */
|
||||||
attachment[i]->type = KFD_MEM_ATT_USERPTR;
|
attachment[i]->type = KFD_MEM_ATT_USERPTR;
|
||||||
ret = amdgpu_gem_object_create(adev, bo_size, 1,
|
ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
|
||||||
AMDGPU_GEM_DOMAIN_CPU,
|
|
||||||
0, ttm_bo_type_sg,
|
|
||||||
mem->bo->tbo.base.resv,
|
|
||||||
&gobj);
|
|
||||||
if (ret)
|
if (ret)
|
||||||
goto unwind;
|
goto unwind;
|
||||||
bo[i] = gem_to_amdgpu_bo(gobj);
|
|
||||||
bo[i]->parent = amdgpu_bo_ref(mem->bo);
|
|
||||||
} else {
|
} else {
|
||||||
/* FIXME: Need to DMA-map other BO types */
|
/* FIXME: Need to DMA-map other BO types */
|
||||||
attachment[i]->type = KFD_MEM_ATT_SHARED;
|
attachment[i]->type = KFD_MEM_ATT_SHARED;
|
||||||
|
@ -670,13 +691,6 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
|
||||||
va += bo_size;
|
va += bo_size;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Allocate validate page tables if needed */
|
|
||||||
ret = vm_validate_pt_pd_bos(vm);
|
|
||||||
if (unlikely(ret)) {
|
|
||||||
pr_err("validate_pt_pd_bos() failed\n");
|
|
||||||
goto unwind;
|
|
||||||
}
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
unwind:
|
unwind:
|
||||||
|
@ -1483,12 +1497,12 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
|
||||||
pr_debug("Release VA 0x%llx - 0x%llx\n", mem->va,
|
pr_debug("Release VA 0x%llx - 0x%llx\n", mem->va,
|
||||||
mem->va + bo_size * (1 + mem->aql_queue));
|
mem->va + bo_size * (1 + mem->aql_queue));
|
||||||
|
|
||||||
|
ret = unreserve_bo_and_vms(&ctx, false, false);
|
||||||
|
|
||||||
/* Remove from VM internal data structures */
|
/* Remove from VM internal data structures */
|
||||||
list_for_each_entry_safe(entry, tmp, &mem->attachments, list)
|
list_for_each_entry_safe(entry, tmp, &mem->attachments, list)
|
||||||
kfd_mem_detach(entry);
|
kfd_mem_detach(entry);
|
||||||
|
|
||||||
ret = unreserve_bo_and_vms(&ctx, false, false);
|
|
||||||
|
|
||||||
/* Free the sync object */
|
/* Free the sync object */
|
||||||
amdgpu_sync_free(&mem->sync);
|
amdgpu_sync_free(&mem->sync);
|
||||||
|
|
||||||
|
@ -1565,6 +1579,12 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
||||||
mem->va + bo_size * (1 + mem->aql_queue),
|
mem->va + bo_size * (1 + mem->aql_queue),
|
||||||
avm, domain_string(domain));
|
avm, domain_string(domain));
|
||||||
|
|
||||||
|
if (!kfd_mem_is_attached(avm, mem)) {
|
||||||
|
ret = kfd_mem_attach(adev, mem, avm, mem->aql_queue);
|
||||||
|
if (ret)
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
ret = reserve_bo_and_vm(mem, avm, &ctx);
|
ret = reserve_bo_and_vm(mem, avm, &ctx);
|
||||||
if (unlikely(ret))
|
if (unlikely(ret))
|
||||||
goto out;
|
goto out;
|
||||||
|
@ -1578,15 +1598,9 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
||||||
bo->tbo.mem.mem_type == TTM_PL_SYSTEM)
|
bo->tbo.mem.mem_type == TTM_PL_SYSTEM)
|
||||||
is_invalid_userptr = true;
|
is_invalid_userptr = true;
|
||||||
|
|
||||||
if (!kfd_mem_is_attached(avm, mem)) {
|
|
||||||
ret = kfd_mem_attach(adev, mem, avm, mem->aql_queue);
|
|
||||||
if (ret)
|
|
||||||
goto attach_failed;
|
|
||||||
} else {
|
|
||||||
ret = vm_validate_pt_pd_bos(avm);
|
ret = vm_validate_pt_pd_bos(avm);
|
||||||
if (unlikely(ret))
|
if (unlikely(ret))
|
||||||
goto attach_failed;
|
goto out_unreserve;
|
||||||
}
|
|
||||||
|
|
||||||
if (mem->mapped_to_gpu_memory == 0 &&
|
if (mem->mapped_to_gpu_memory == 0 &&
|
||||||
!amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
|
!amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
|
||||||
|
@ -1597,7 +1611,7 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
||||||
ret = amdgpu_amdkfd_bo_validate(bo, domain, true);
|
ret = amdgpu_amdkfd_bo_validate(bo, domain, true);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_debug("Validate failed\n");
|
pr_debug("Validate failed\n");
|
||||||
goto map_bo_to_gpuvm_failed;
|
goto out_unreserve;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1612,13 +1626,13 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
||||||
is_invalid_userptr);
|
is_invalid_userptr);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_err("Failed to map bo to gpuvm\n");
|
pr_err("Failed to map bo to gpuvm\n");
|
||||||
goto map_bo_to_gpuvm_failed;
|
goto out_unreserve;
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = vm_update_pds(avm, ctx.sync);
|
ret = vm_update_pds(avm, ctx.sync);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_err("Failed to update page directories\n");
|
pr_err("Failed to update page directories\n");
|
||||||
goto map_bo_to_gpuvm_failed;
|
goto out_unreserve;
|
||||||
}
|
}
|
||||||
|
|
||||||
entry->is_mapped = true;
|
entry->is_mapped = true;
|
||||||
|
@ -1635,8 +1649,7 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
||||||
|
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
map_bo_to_gpuvm_failed:
|
out_unreserve:
|
||||||
attach_failed:
|
|
||||||
unreserve_bo_and_vms(&ctx, false, false);
|
unreserve_bo_and_vms(&ctx, false, false);
|
||||||
out:
|
out:
|
||||||
mutex_unlock(&mem->process_info->lock);
|
mutex_unlock(&mem->process_info->lock);
|
||||||
|
|
Loading…
Add table
Reference in a new issue