iommu/amd: Remove special mapping code for dma_ops path
Use the iommu-api map/unmap functions instead. This will be required anyway when IOVA code is used for address allocation. Signed-off-by: Joerg Roedel <jroedel@suse.de>
This commit is contained in:
parent
b911b89b6d
commit
518d9b4503
1 changed files with 14 additions and 93 deletions
|
@ -2559,94 +2559,6 @@ static void update_domain(struct protection_domain *domain)
|
||||||
domain->updated = false;
|
domain->updated = false;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* This function fetches the PTE for a given address in the aperture
|
|
||||||
*/
|
|
||||||
static u64* dma_ops_get_pte(struct dma_ops_domain *dom,
|
|
||||||
unsigned long address)
|
|
||||||
{
|
|
||||||
struct aperture_range *aperture;
|
|
||||||
u64 *pte, *pte_page;
|
|
||||||
|
|
||||||
aperture = dom->aperture[APERTURE_RANGE_INDEX(address)];
|
|
||||||
if (!aperture)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
pte = aperture->pte_pages[APERTURE_PAGE_INDEX(address)];
|
|
||||||
if (!pte) {
|
|
||||||
pte = alloc_pte(&dom->domain, address, PAGE_SIZE, &pte_page,
|
|
||||||
GFP_ATOMIC);
|
|
||||||
aperture->pte_pages[APERTURE_PAGE_INDEX(address)] = pte_page;
|
|
||||||
} else
|
|
||||||
pte += PM_LEVEL_INDEX(0, address);
|
|
||||||
|
|
||||||
update_domain(&dom->domain);
|
|
||||||
|
|
||||||
return pte;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* This is the generic map function. It maps one 4kb page at paddr to
|
|
||||||
* the given address in the DMA address space for the domain.
|
|
||||||
*/
|
|
||||||
static dma_addr_t dma_ops_domain_map(struct dma_ops_domain *dom,
|
|
||||||
unsigned long address,
|
|
||||||
phys_addr_t paddr,
|
|
||||||
int direction)
|
|
||||||
{
|
|
||||||
u64 *pte, __pte;
|
|
||||||
|
|
||||||
WARN_ON(address > dom->aperture_size);
|
|
||||||
|
|
||||||
paddr &= PAGE_MASK;
|
|
||||||
|
|
||||||
pte = dma_ops_get_pte(dom, address);
|
|
||||||
if (!pte)
|
|
||||||
return DMA_ERROR_CODE;
|
|
||||||
|
|
||||||
__pte = paddr | IOMMU_PTE_P | IOMMU_PTE_FC;
|
|
||||||
|
|
||||||
if (direction == DMA_TO_DEVICE)
|
|
||||||
__pte |= IOMMU_PTE_IR;
|
|
||||||
else if (direction == DMA_FROM_DEVICE)
|
|
||||||
__pte |= IOMMU_PTE_IW;
|
|
||||||
else if (direction == DMA_BIDIRECTIONAL)
|
|
||||||
__pte |= IOMMU_PTE_IR | IOMMU_PTE_IW;
|
|
||||||
|
|
||||||
WARN_ON_ONCE(*pte);
|
|
||||||
|
|
||||||
*pte = __pte;
|
|
||||||
|
|
||||||
return (dma_addr_t)address;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* The generic unmapping function for on page in the DMA address space.
|
|
||||||
*/
|
|
||||||
static void dma_ops_domain_unmap(struct dma_ops_domain *dom,
|
|
||||||
unsigned long address)
|
|
||||||
{
|
|
||||||
struct aperture_range *aperture;
|
|
||||||
u64 *pte;
|
|
||||||
|
|
||||||
if (address >= dom->aperture_size)
|
|
||||||
return;
|
|
||||||
|
|
||||||
aperture = dom->aperture[APERTURE_RANGE_INDEX(address)];
|
|
||||||
if (!aperture)
|
|
||||||
return;
|
|
||||||
|
|
||||||
pte = aperture->pte_pages[APERTURE_PAGE_INDEX(address)];
|
|
||||||
if (!pte)
|
|
||||||
return;
|
|
||||||
|
|
||||||
pte += PM_LEVEL_INDEX(0, address);
|
|
||||||
|
|
||||||
WARN_ON_ONCE(!*pte);
|
|
||||||
|
|
||||||
*pte = 0ULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This function contains common code for mapping of a physically
|
* This function contains common code for mapping of a physically
|
||||||
* contiguous memory region into DMA address space. It is used by all
|
* contiguous memory region into DMA address space. It is used by all
|
||||||
|
@ -2657,7 +2569,7 @@ static dma_addr_t __map_single(struct device *dev,
|
||||||
struct dma_ops_domain *dma_dom,
|
struct dma_ops_domain *dma_dom,
|
||||||
phys_addr_t paddr,
|
phys_addr_t paddr,
|
||||||
size_t size,
|
size_t size,
|
||||||
int dir,
|
int direction,
|
||||||
bool align,
|
bool align,
|
||||||
u64 dma_mask)
|
u64 dma_mask)
|
||||||
{
|
{
|
||||||
|
@ -2665,6 +2577,7 @@ static dma_addr_t __map_single(struct device *dev,
|
||||||
dma_addr_t address, start, ret;
|
dma_addr_t address, start, ret;
|
||||||
unsigned int pages;
|
unsigned int pages;
|
||||||
unsigned long align_mask = 0;
|
unsigned long align_mask = 0;
|
||||||
|
int prot = 0;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
pages = iommu_num_pages(paddr, size, PAGE_SIZE);
|
pages = iommu_num_pages(paddr, size, PAGE_SIZE);
|
||||||
|
@ -2679,10 +2592,18 @@ static dma_addr_t __map_single(struct device *dev,
|
||||||
if (address == DMA_ERROR_CODE)
|
if (address == DMA_ERROR_CODE)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
|
if (direction == DMA_TO_DEVICE)
|
||||||
|
prot = IOMMU_PROT_IR;
|
||||||
|
else if (direction == DMA_FROM_DEVICE)
|
||||||
|
prot = IOMMU_PROT_IW;
|
||||||
|
else if (direction == DMA_BIDIRECTIONAL)
|
||||||
|
prot = IOMMU_PROT_IW | IOMMU_PROT_IR;
|
||||||
|
|
||||||
start = address;
|
start = address;
|
||||||
for (i = 0; i < pages; ++i) {
|
for (i = 0; i < pages; ++i) {
|
||||||
ret = dma_ops_domain_map(dma_dom, start, paddr, dir);
|
ret = iommu_map_page(&dma_dom->domain, start, paddr,
|
||||||
if (ret == DMA_ERROR_CODE)
|
PAGE_SIZE, prot, GFP_ATOMIC);
|
||||||
|
if (ret)
|
||||||
goto out_unmap;
|
goto out_unmap;
|
||||||
|
|
||||||
paddr += PAGE_SIZE;
|
paddr += PAGE_SIZE;
|
||||||
|
@ -2702,7 +2623,7 @@ out_unmap:
|
||||||
|
|
||||||
for (--i; i >= 0; --i) {
|
for (--i; i >= 0; --i) {
|
||||||
start -= PAGE_SIZE;
|
start -= PAGE_SIZE;
|
||||||
dma_ops_domain_unmap(dma_dom, start);
|
iommu_unmap_page(&dma_dom->domain, start, PAGE_SIZE);
|
||||||
}
|
}
|
||||||
|
|
||||||
dma_ops_free_addresses(dma_dom, address, pages);
|
dma_ops_free_addresses(dma_dom, address, pages);
|
||||||
|
@ -2733,7 +2654,7 @@ static void __unmap_single(struct dma_ops_domain *dma_dom,
|
||||||
start = dma_addr;
|
start = dma_addr;
|
||||||
|
|
||||||
for (i = 0; i < pages; ++i) {
|
for (i = 0; i < pages; ++i) {
|
||||||
dma_ops_domain_unmap(dma_dom, start);
|
iommu_unmap_page(&dma_dom->domain, start, PAGE_SIZE);
|
||||||
start += PAGE_SIZE;
|
start += PAGE_SIZE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue