iommu/vt-d: Add support for static identity domain
Software determines VT-d hardware support for passthrough translation by inspecting the capability register. If passthrough translation is not supported, the device is instructed to use DMA domain for its default domain. Add a global static identity domain with guaranteed attach semantics for IOMMUs that support passthrough translation mode. The global static identity domain is a dummy domain without corresponding dmar_domain structure. Consequently, the device's info->domain will be NULL with the identity domain is attached. Refactor the code accordingly. Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> Reviewed-by: Kevin Tian <kevin.tian@intel.com> Link: https://lore.kernel.org/r/20240809055431.36513-7-baolu.lu@linux.intel.com Signed-off-by: Joerg Roedel <jroedel@suse.de>
This commit is contained in:
parent
c7191984e5
commit
2031c469f8
2 changed files with 111 additions and 5 deletions
|
@ -3691,11 +3691,9 @@ int prepare_domain_attach_device(struct iommu_domain *domain,
|
||||||
static int intel_iommu_attach_device(struct iommu_domain *domain,
|
static int intel_iommu_attach_device(struct iommu_domain *domain,
|
||||||
struct device *dev)
|
struct device *dev)
|
||||||
{
|
{
|
||||||
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
if (info->domain)
|
device_block_translation(dev);
|
||||||
device_block_translation(dev);
|
|
||||||
|
|
||||||
ret = prepare_domain_attach_device(domain, dev);
|
ret = prepare_domain_attach_device(domain, dev);
|
||||||
if (ret)
|
if (ret)
|
||||||
|
@ -4301,11 +4299,17 @@ static void intel_iommu_remove_dev_pasid(struct device *dev, ioasid_t pasid,
|
||||||
struct iommu_domain *domain)
|
struct iommu_domain *domain)
|
||||||
{
|
{
|
||||||
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
||||||
struct dmar_domain *dmar_domain = to_dmar_domain(domain);
|
|
||||||
struct dev_pasid_info *curr, *dev_pasid = NULL;
|
struct dev_pasid_info *curr, *dev_pasid = NULL;
|
||||||
struct intel_iommu *iommu = info->iommu;
|
struct intel_iommu *iommu = info->iommu;
|
||||||
|
struct dmar_domain *dmar_domain;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
|
if (domain->type == IOMMU_DOMAIN_IDENTITY) {
|
||||||
|
intel_pasid_tear_down_entry(iommu, dev, pasid, false);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
dmar_domain = to_dmar_domain(domain);
|
||||||
spin_lock_irqsave(&dmar_domain->lock, flags);
|
spin_lock_irqsave(&dmar_domain->lock, flags);
|
||||||
list_for_each_entry(curr, &dmar_domain->dev_pasids, link_domain) {
|
list_for_each_entry(curr, &dmar_domain->dev_pasids, link_domain) {
|
||||||
if (curr->dev == dev && curr->pasid == pasid) {
|
if (curr->dev == dev && curr->pasid == pasid) {
|
||||||
|
@ -4532,9 +4536,111 @@ static const struct iommu_dirty_ops intel_dirty_ops = {
|
||||||
.read_and_clear_dirty = intel_iommu_read_and_clear_dirty,
|
.read_and_clear_dirty = intel_iommu_read_and_clear_dirty,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static int context_setup_pass_through(struct device *dev, u8 bus, u8 devfn)
|
||||||
|
{
|
||||||
|
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
||||||
|
struct intel_iommu *iommu = info->iommu;
|
||||||
|
struct context_entry *context;
|
||||||
|
|
||||||
|
spin_lock(&iommu->lock);
|
||||||
|
context = iommu_context_addr(iommu, bus, devfn, 1);
|
||||||
|
if (!context) {
|
||||||
|
spin_unlock(&iommu->lock);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (context_present(context) && !context_copied(iommu, bus, devfn)) {
|
||||||
|
spin_unlock(&iommu->lock);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
copied_context_tear_down(iommu, context, bus, devfn);
|
||||||
|
context_clear_entry(context);
|
||||||
|
context_set_domain_id(context, FLPT_DEFAULT_DID);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* In pass through mode, AW must be programmed to indicate the largest
|
||||||
|
* AGAW value supported by hardware. And ASR is ignored by hardware.
|
||||||
|
*/
|
||||||
|
context_set_address_width(context, iommu->msagaw);
|
||||||
|
context_set_translation_type(context, CONTEXT_TT_PASS_THROUGH);
|
||||||
|
context_set_fault_enable(context);
|
||||||
|
context_set_present(context);
|
||||||
|
if (!ecap_coherent(iommu->ecap))
|
||||||
|
clflush_cache_range(context, sizeof(*context));
|
||||||
|
context_present_cache_flush(iommu, FLPT_DEFAULT_DID, bus, devfn);
|
||||||
|
spin_unlock(&iommu->lock);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int context_setup_pass_through_cb(struct pci_dev *pdev, u16 alias, void *data)
|
||||||
|
{
|
||||||
|
struct device *dev = data;
|
||||||
|
|
||||||
|
if (dev != &pdev->dev)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
return context_setup_pass_through(dev, PCI_BUS_NUM(alias), alias & 0xff);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int device_setup_pass_through(struct device *dev)
|
||||||
|
{
|
||||||
|
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
||||||
|
|
||||||
|
if (!dev_is_pci(dev))
|
||||||
|
return context_setup_pass_through(dev, info->bus, info->devfn);
|
||||||
|
|
||||||
|
return pci_for_each_dma_alias(to_pci_dev(dev),
|
||||||
|
context_setup_pass_through_cb, dev);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int identity_domain_attach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
|
{
|
||||||
|
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
||||||
|
struct intel_iommu *iommu = info->iommu;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
device_block_translation(dev);
|
||||||
|
|
||||||
|
if (dev_is_real_dma_subdevice(dev))
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (sm_supported(iommu)) {
|
||||||
|
ret = intel_pasid_setup_pass_through(iommu, dev, IOMMU_NO_PASID);
|
||||||
|
if (!ret)
|
||||||
|
iommu_enable_pci_caps(info);
|
||||||
|
} else {
|
||||||
|
ret = device_setup_pass_through(dev);
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int identity_domain_set_dev_pasid(struct iommu_domain *domain,
|
||||||
|
struct device *dev, ioasid_t pasid)
|
||||||
|
{
|
||||||
|
struct device_domain_info *info = dev_iommu_priv_get(dev);
|
||||||
|
struct intel_iommu *iommu = info->iommu;
|
||||||
|
|
||||||
|
if (!pasid_supported(iommu) || dev_is_real_dma_subdevice(dev))
|
||||||
|
return -EOPNOTSUPP;
|
||||||
|
|
||||||
|
return intel_pasid_setup_pass_through(iommu, dev, pasid);
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct iommu_domain identity_domain = {
|
||||||
|
.type = IOMMU_DOMAIN_IDENTITY,
|
||||||
|
.ops = &(const struct iommu_domain_ops) {
|
||||||
|
.attach_dev = identity_domain_attach_dev,
|
||||||
|
.set_dev_pasid = identity_domain_set_dev_pasid,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
const struct iommu_ops intel_iommu_ops = {
|
const struct iommu_ops intel_iommu_ops = {
|
||||||
.blocked_domain = &blocking_domain,
|
.blocked_domain = &blocking_domain,
|
||||||
.release_domain = &blocking_domain,
|
.release_domain = &blocking_domain,
|
||||||
|
.identity_domain = &identity_domain,
|
||||||
.capable = intel_iommu_capable,
|
.capable = intel_iommu_capable,
|
||||||
.hw_info = intel_iommu_hw_info,
|
.hw_info = intel_iommu_hw_info,
|
||||||
.domain_alloc = intel_iommu_domain_alloc,
|
.domain_alloc = intel_iommu_domain_alloc,
|
||||||
|
|
|
@ -311,7 +311,7 @@ void intel_drain_pasid_prq(struct device *dev, u32 pasid)
|
||||||
domain = info->domain;
|
domain = info->domain;
|
||||||
pdev = to_pci_dev(dev);
|
pdev = to_pci_dev(dev);
|
||||||
sid = PCI_DEVID(info->bus, info->devfn);
|
sid = PCI_DEVID(info->bus, info->devfn);
|
||||||
did = domain_id_iommu(domain, iommu);
|
did = domain ? domain_id_iommu(domain, iommu) : FLPT_DEFAULT_DID;
|
||||||
qdep = pci_ats_queue_depth(pdev);
|
qdep = pci_ats_queue_depth(pdev);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
Loading…
Add table
Reference in a new issue