The problem is that when (for example) 4k pages are replaced with a single 2M page we need to wait for change to be flushed out by invalidating the TLB before the PT can be freed. Solve this by moving the TLB flush into a DMA-fence object which can be used to delay the freeing of the PT BOs until it is signaled. V2: (Shashank) - rebase - set dma_fence_error only in case of error - add tlb_flush fence only when PT/PD BO is locked (Felix) - use vm->pasid when f is NULL (Mukul) V4: - add a wait for (f->dependency) in tlb_fence_work (Christian) - move the misplaced fence_create call to the end (Philip) V5: - free the f->dependency properly V6: (Shashank) - light code movement, moved all the clean-up in previous patch - introduce params.needs_flush and its usage in this patch - rebase without TLB HW sequence patch V7: - Keep the vm->last_update_fence and tlb_cb code until we can fix the HW sequencing (Christian) - Move all the tlb_fence related code in a separate function so that its easier to read and review V9: Addressed review comments from Christian - start PT update only when we have callback memory allocated V10: - handle device unlock in OOM case (Christian, Mukul) - added Christian's R-B Cc: Christian Koenig <christian.koenig@amd.com> Cc: Felix Kuehling <Felix.Kuehling@amd.com> Cc: Rajneesh Bhardwaj <rajneesh.bhardwaj@amd.com> Cc: Alex Deucher <alexander.deucher@amd.com> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com> Acked-by: Rajneesh Bhardwaj <rajneesh.bhardwaj@amd.com> Tested-by: Rajneesh Bhardwaj <rajneesh.bhardwaj@amd.com> Reviewed-by: Shashank Sharma <shashank.sharma@amd.com> Reviewed-by: Christian Koenig <christian.koenig@amd.com> Signed-off-by: Christian Koenig <christian.koenig@amd.com> Signed-off-by: Shashank Sharma <shashank.sharma@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
312 lines
8.2 KiB
C
312 lines
8.2 KiB
C
/*
|
|
* Copyright 2019 Advanced Micro Devices, Inc.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include "amdgpu_vm.h"
|
|
#include "amdgpu_job.h"
|
|
#include "amdgpu_object.h"
|
|
#include "amdgpu_trace.h"
|
|
|
|
#define AMDGPU_VM_SDMA_MIN_NUM_DW 256u
|
|
#define AMDGPU_VM_SDMA_MAX_NUM_DW (16u * 1024u)
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_map_table - make sure new PDs/PTs are GTT mapped
|
|
*
|
|
* @table: newly allocated or validated PD/PT
|
|
*/
|
|
static int amdgpu_vm_sdma_map_table(struct amdgpu_bo_vm *table)
|
|
{
|
|
int r;
|
|
|
|
r = amdgpu_ttm_alloc_gart(&table->bo.tbo);
|
|
if (r)
|
|
return r;
|
|
|
|
if (table->shadow)
|
|
r = amdgpu_ttm_alloc_gart(&table->shadow->tbo);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Allocate a new job for @count PTE updates */
|
|
static int amdgpu_vm_sdma_alloc_job(struct amdgpu_vm_update_params *p,
|
|
unsigned int count)
|
|
{
|
|
enum amdgpu_ib_pool_type pool = p->immediate ? AMDGPU_IB_POOL_IMMEDIATE
|
|
: AMDGPU_IB_POOL_DELAYED;
|
|
struct drm_sched_entity *entity = p->immediate ? &p->vm->immediate
|
|
: &p->vm->delayed;
|
|
unsigned int ndw;
|
|
int r;
|
|
|
|
/* estimate how many dw we need */
|
|
ndw = AMDGPU_VM_SDMA_MIN_NUM_DW;
|
|
if (p->pages_addr)
|
|
ndw += count * 2;
|
|
ndw = min(ndw, AMDGPU_VM_SDMA_MAX_NUM_DW);
|
|
|
|
r = amdgpu_job_alloc_with_ib(p->adev, entity, AMDGPU_FENCE_OWNER_VM,
|
|
ndw * 4, pool, &p->job);
|
|
if (r)
|
|
return r;
|
|
|
|
p->num_dw_left = ndw;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_prepare - prepare SDMA command submission
|
|
*
|
|
* @p: see amdgpu_vm_update_params definition
|
|
* @resv: reservation object with embedded fence
|
|
* @sync_mode: synchronization mode
|
|
*
|
|
* Returns:
|
|
* Negativ errno, 0 for success.
|
|
*/
|
|
static int amdgpu_vm_sdma_prepare(struct amdgpu_vm_update_params *p,
|
|
struct dma_resv *resv,
|
|
enum amdgpu_sync_mode sync_mode)
|
|
{
|
|
struct amdgpu_sync sync;
|
|
int r;
|
|
|
|
r = amdgpu_vm_sdma_alloc_job(p, 0);
|
|
if (r)
|
|
return r;
|
|
|
|
if (!resv)
|
|
return 0;
|
|
|
|
amdgpu_sync_create(&sync);
|
|
r = amdgpu_sync_resv(p->adev, &sync, resv, sync_mode, p->vm);
|
|
if (!r)
|
|
r = amdgpu_sync_push_to_job(&sync, p->job);
|
|
amdgpu_sync_free(&sync);
|
|
return r;
|
|
}
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_commit - commit SDMA command submission
|
|
*
|
|
* @p: see amdgpu_vm_update_params definition
|
|
* @fence: resulting fence
|
|
*
|
|
* Returns:
|
|
* Negativ errno, 0 for success.
|
|
*/
|
|
static int amdgpu_vm_sdma_commit(struct amdgpu_vm_update_params *p,
|
|
struct dma_fence **fence)
|
|
{
|
|
struct amdgpu_ib *ib = p->job->ibs;
|
|
struct amdgpu_ring *ring;
|
|
struct dma_fence *f;
|
|
|
|
ring = container_of(p->vm->delayed.rq->sched, struct amdgpu_ring,
|
|
sched);
|
|
|
|
WARN_ON(ib->length_dw == 0);
|
|
amdgpu_ring_pad_ib(ring, ib);
|
|
|
|
if (p->needs_flush)
|
|
atomic64_inc(&p->vm->tlb_seq);
|
|
|
|
WARN_ON(ib->length_dw > p->num_dw_left);
|
|
f = amdgpu_job_submit(p->job);
|
|
|
|
if (p->unlocked) {
|
|
struct dma_fence *tmp = dma_fence_get(f);
|
|
|
|
swap(p->vm->last_unlocked, tmp);
|
|
dma_fence_put(tmp);
|
|
} else {
|
|
dma_resv_add_fence(p->vm->root.bo->tbo.base.resv, f,
|
|
DMA_RESV_USAGE_BOOKKEEP);
|
|
}
|
|
|
|
if (fence && !p->immediate) {
|
|
/*
|
|
* Most hw generations now have a separate queue for page table
|
|
* updates, but when the queue is shared with userspace we need
|
|
* the extra CPU round trip to correctly flush the TLB.
|
|
*/
|
|
set_bit(DRM_SCHED_FENCE_DONT_PIPELINE, &f->flags);
|
|
swap(*fence, f);
|
|
}
|
|
dma_fence_put(f);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_copy_ptes - copy the PTEs from mapping
|
|
*
|
|
* @p: see amdgpu_vm_update_params definition
|
|
* @bo: PD/PT to update
|
|
* @pe: addr of the page entry
|
|
* @count: number of page entries to copy
|
|
*
|
|
* Traces the parameters and calls the DMA function to copy the PTEs.
|
|
*/
|
|
static void amdgpu_vm_sdma_copy_ptes(struct amdgpu_vm_update_params *p,
|
|
struct amdgpu_bo *bo, uint64_t pe,
|
|
unsigned count)
|
|
{
|
|
struct amdgpu_ib *ib = p->job->ibs;
|
|
uint64_t src = ib->gpu_addr;
|
|
|
|
src += p->num_dw_left * 4;
|
|
|
|
pe += amdgpu_bo_gpu_offset_no_check(bo);
|
|
trace_amdgpu_vm_copy_ptes(pe, src, count, p->immediate);
|
|
|
|
amdgpu_vm_copy_pte(p->adev, ib, pe, src, count);
|
|
}
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_set_ptes - helper to call the right asic function
|
|
*
|
|
* @p: see amdgpu_vm_update_params definition
|
|
* @bo: PD/PT to update
|
|
* @pe: byte offset of the PDE/PTE, relative to start of PDB/PTB
|
|
* @addr: dst addr to write into pe
|
|
* @count: number of page entries to update
|
|
* @incr: increase next addr by incr bytes
|
|
* @flags: hw access flags
|
|
*
|
|
* Traces the parameters and calls the right asic functions
|
|
* to setup the page table using the DMA.
|
|
*/
|
|
static void amdgpu_vm_sdma_set_ptes(struct amdgpu_vm_update_params *p,
|
|
struct amdgpu_bo *bo, uint64_t pe,
|
|
uint64_t addr, unsigned count,
|
|
uint32_t incr, uint64_t flags)
|
|
{
|
|
struct amdgpu_ib *ib = p->job->ibs;
|
|
|
|
pe += amdgpu_bo_gpu_offset_no_check(bo);
|
|
trace_amdgpu_vm_set_ptes(pe, addr, count, incr, flags, p->immediate);
|
|
if (count < 3) {
|
|
amdgpu_vm_write_pte(p->adev, ib, pe, addr | flags,
|
|
count, incr);
|
|
} else {
|
|
amdgpu_vm_set_pte_pde(p->adev, ib, pe, addr,
|
|
count, incr, flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* amdgpu_vm_sdma_update - execute VM update
|
|
*
|
|
* @p: see amdgpu_vm_update_params definition
|
|
* @vmbo: PD/PT to update
|
|
* @pe: byte offset of the PDE/PTE, relative to start of PDB/PTB
|
|
* @addr: dst addr to write into pe
|
|
* @count: number of page entries to update
|
|
* @incr: increase next addr by incr bytes
|
|
* @flags: hw access flags
|
|
*
|
|
* Reserve space in the IB, setup mapping buffer on demand and write commands to
|
|
* the IB.
|
|
*/
|
|
static int amdgpu_vm_sdma_update(struct amdgpu_vm_update_params *p,
|
|
struct amdgpu_bo_vm *vmbo, uint64_t pe,
|
|
uint64_t addr, unsigned count, uint32_t incr,
|
|
uint64_t flags)
|
|
{
|
|
struct amdgpu_bo *bo = &vmbo->bo;
|
|
struct dma_resv_iter cursor;
|
|
unsigned int i, ndw, nptes;
|
|
struct dma_fence *fence;
|
|
uint64_t *pte;
|
|
int r;
|
|
|
|
/* Wait for PD/PT moves to be completed */
|
|
dma_resv_iter_begin(&cursor, bo->tbo.base.resv, DMA_RESV_USAGE_KERNEL);
|
|
dma_resv_for_each_fence_unlocked(&cursor, fence) {
|
|
dma_fence_get(fence);
|
|
r = drm_sched_job_add_dependency(&p->job->base, fence);
|
|
if (r) {
|
|
dma_fence_put(fence);
|
|
dma_resv_iter_end(&cursor);
|
|
return r;
|
|
}
|
|
}
|
|
dma_resv_iter_end(&cursor);
|
|
|
|
do {
|
|
ndw = p->num_dw_left;
|
|
ndw -= p->job->ibs->length_dw;
|
|
|
|
if (ndw < 32) {
|
|
r = amdgpu_vm_sdma_commit(p, NULL);
|
|
if (r)
|
|
return r;
|
|
|
|
r = amdgpu_vm_sdma_alloc_job(p, count);
|
|
if (r)
|
|
return r;
|
|
}
|
|
|
|
if (!p->pages_addr) {
|
|
/* set page commands needed */
|
|
if (vmbo->shadow)
|
|
amdgpu_vm_sdma_set_ptes(p, vmbo->shadow, pe, addr,
|
|
count, incr, flags);
|
|
amdgpu_vm_sdma_set_ptes(p, bo, pe, addr, count,
|
|
incr, flags);
|
|
return 0;
|
|
}
|
|
|
|
/* copy commands needed */
|
|
ndw -= p->adev->vm_manager.vm_pte_funcs->copy_pte_num_dw *
|
|
(vmbo->shadow ? 2 : 1);
|
|
|
|
/* for padding */
|
|
ndw -= 7;
|
|
|
|
nptes = min(count, ndw / 2);
|
|
|
|
/* Put the PTEs at the end of the IB. */
|
|
p->num_dw_left -= nptes * 2;
|
|
pte = (uint64_t *)&(p->job->ibs->ptr[p->num_dw_left]);
|
|
for (i = 0; i < nptes; ++i, addr += incr) {
|
|
pte[i] = amdgpu_vm_map_gart(p->pages_addr, addr);
|
|
pte[i] |= flags;
|
|
}
|
|
|
|
if (vmbo->shadow)
|
|
amdgpu_vm_sdma_copy_ptes(p, vmbo->shadow, pe, nptes);
|
|
amdgpu_vm_sdma_copy_ptes(p, bo, pe, nptes);
|
|
|
|
pe += nptes * 8;
|
|
count -= nptes;
|
|
} while (count);
|
|
|
|
return 0;
|
|
}
|
|
|
|
const struct amdgpu_vm_update_funcs amdgpu_vm_sdma_funcs = {
|
|
.map_table = amdgpu_vm_sdma_map_table,
|
|
.prepare = amdgpu_vm_sdma_prepare,
|
|
.update = amdgpu_vm_sdma_update,
|
|
.commit = amdgpu_vm_sdma_commit
|
|
};
|