iommu/arm-smmu: use mutex instead of spinlock for locking page tables
When creating IO mappings, we lazily allocate our page tables using the standard, non-atomic allocator functions. This presents us with a problem, since our page tables are protected with a spinlock. This patch reworks the smmu_domain lock to use a mutex instead of a spinlock. iova_to_phys is then reworked so that it only reads the page tables, and can run in a lockless fashion, leaving the mutex to guard against concurrent mapping threads. Cc: <stable@vger.kernel.org> Signed-off-by: Will Deacon <will.deacon@arm.com>
This commit is contained in:
parent
dc1ccc4815
commit
a44a9791e7
|
@ -392,7 +392,7 @@ struct arm_smmu_domain {
|
||||||
struct arm_smmu_cfg root_cfg;
|
struct arm_smmu_cfg root_cfg;
|
||||||
phys_addr_t output_mask;
|
phys_addr_t output_mask;
|
||||||
|
|
||||||
spinlock_t lock;
|
struct mutex lock;
|
||||||
};
|
};
|
||||||
|
|
||||||
static DEFINE_SPINLOCK(arm_smmu_devices_lock);
|
static DEFINE_SPINLOCK(arm_smmu_devices_lock);
|
||||||
|
@ -900,7 +900,7 @@ static int arm_smmu_domain_init(struct iommu_domain *domain)
|
||||||
goto out_free_domain;
|
goto out_free_domain;
|
||||||
smmu_domain->root_cfg.pgd = pgd;
|
smmu_domain->root_cfg.pgd = pgd;
|
||||||
|
|
||||||
spin_lock_init(&smmu_domain->lock);
|
mutex_init(&smmu_domain->lock);
|
||||||
domain->priv = smmu_domain;
|
domain->priv = smmu_domain;
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
@ -1137,7 +1137,7 @@ static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
* Sanity check the domain. We don't currently support domains
|
* Sanity check the domain. We don't currently support domains
|
||||||
* that cross between different SMMU chains.
|
* that cross between different SMMU chains.
|
||||||
*/
|
*/
|
||||||
spin_lock(&smmu_domain->lock);
|
mutex_lock(&smmu_domain->lock);
|
||||||
if (!smmu_domain->leaf_smmu) {
|
if (!smmu_domain->leaf_smmu) {
|
||||||
/* Now that we have a master, we can finalise the domain */
|
/* Now that we have a master, we can finalise the domain */
|
||||||
ret = arm_smmu_init_domain_context(domain, dev);
|
ret = arm_smmu_init_domain_context(domain, dev);
|
||||||
|
@ -1152,7 +1152,7 @@ static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
dev_name(device_smmu->dev));
|
dev_name(device_smmu->dev));
|
||||||
goto err_unlock;
|
goto err_unlock;
|
||||||
}
|
}
|
||||||
spin_unlock(&smmu_domain->lock);
|
mutex_unlock(&smmu_domain->lock);
|
||||||
|
|
||||||
/* Looks ok, so add the device to the domain */
|
/* Looks ok, so add the device to the domain */
|
||||||
master = find_smmu_master(smmu_domain->leaf_smmu, dev->of_node);
|
master = find_smmu_master(smmu_domain->leaf_smmu, dev->of_node);
|
||||||
|
@ -1162,7 +1162,7 @@ static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
return arm_smmu_domain_add_master(smmu_domain, master);
|
return arm_smmu_domain_add_master(smmu_domain, master);
|
||||||
|
|
||||||
err_unlock:
|
err_unlock:
|
||||||
spin_unlock(&smmu_domain->lock);
|
mutex_unlock(&smmu_domain->lock);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1394,7 +1394,7 @@ static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
||||||
if (paddr & ~output_mask)
|
if (paddr & ~output_mask)
|
||||||
return -ERANGE;
|
return -ERANGE;
|
||||||
|
|
||||||
spin_lock(&smmu_domain->lock);
|
mutex_lock(&smmu_domain->lock);
|
||||||
pgd += pgd_index(iova);
|
pgd += pgd_index(iova);
|
||||||
end = iova + size;
|
end = iova + size;
|
||||||
do {
|
do {
|
||||||
|
@ -1410,7 +1410,7 @@ static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
||||||
} while (pgd++, iova != end);
|
} while (pgd++, iova != end);
|
||||||
|
|
||||||
out_unlock:
|
out_unlock:
|
||||||
spin_unlock(&smmu_domain->lock);
|
mutex_unlock(&smmu_domain->lock);
|
||||||
|
|
||||||
/* Ensure new page tables are visible to the hardware walker */
|
/* Ensure new page tables are visible to the hardware walker */
|
||||||
if (smmu->features & ARM_SMMU_FEAT_COHERENT_WALK)
|
if (smmu->features & ARM_SMMU_FEAT_COHERENT_WALK)
|
||||||
|
@ -1449,44 +1449,34 @@ static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova,
|
||||||
static phys_addr_t arm_smmu_iova_to_phys(struct iommu_domain *domain,
|
static phys_addr_t arm_smmu_iova_to_phys(struct iommu_domain *domain,
|
||||||
dma_addr_t iova)
|
dma_addr_t iova)
|
||||||
{
|
{
|
||||||
pgd_t *pgd;
|
pgd_t *pgdp, pgd;
|
||||||
pud_t *pud;
|
pud_t pud;
|
||||||
pmd_t *pmd;
|
pmd_t pmd;
|
||||||
pte_t *pte;
|
pte_t pte;
|
||||||
struct arm_smmu_domain *smmu_domain = domain->priv;
|
struct arm_smmu_domain *smmu_domain = domain->priv;
|
||||||
struct arm_smmu_cfg *root_cfg = &smmu_domain->root_cfg;
|
struct arm_smmu_cfg *root_cfg = &smmu_domain->root_cfg;
|
||||||
struct arm_smmu_device *smmu = root_cfg->smmu;
|
|
||||||
|
|
||||||
spin_lock(&smmu_domain->lock);
|
pgdp = root_cfg->pgd;
|
||||||
pgd = root_cfg->pgd;
|
if (!pgdp)
|
||||||
if (!pgd)
|
return 0;
|
||||||
goto err_unlock;
|
|
||||||
|
|
||||||
pgd += pgd_index(iova);
|
pgd = *(pgdp + pgd_index(iova));
|
||||||
if (pgd_none_or_clear_bad(pgd))
|
if (pgd_none(pgd))
|
||||||
goto err_unlock;
|
return 0;
|
||||||
|
|
||||||
pud = pud_offset(pgd, iova);
|
pud = *pud_offset(&pgd, iova);
|
||||||
if (pud_none_or_clear_bad(pud))
|
if (pud_none(pud))
|
||||||
goto err_unlock;
|
return 0;
|
||||||
|
|
||||||
pmd = pmd_offset(pud, iova);
|
pmd = *pmd_offset(&pud, iova);
|
||||||
if (pmd_none_or_clear_bad(pmd))
|
if (pmd_none(pmd))
|
||||||
goto err_unlock;
|
return 0;
|
||||||
|
|
||||||
pte = pmd_page_vaddr(*pmd) + pte_index(iova);
|
pte = *(pmd_page_vaddr(pmd) + pte_index(iova));
|
||||||
if (pte_none(pte))
|
if (pte_none(pte))
|
||||||
goto err_unlock;
|
return 0;
|
||||||
|
|
||||||
spin_unlock(&smmu_domain->lock);
|
return __pfn_to_phys(pte_pfn(pte)) | (iova & ~PAGE_MASK);
|
||||||
return __pfn_to_phys(pte_pfn(*pte)) | (iova & ~PAGE_MASK);
|
|
||||||
|
|
||||||
err_unlock:
|
|
||||||
spin_unlock(&smmu_domain->lock);
|
|
||||||
dev_warn(smmu->dev,
|
|
||||||
"invalid (corrupt?) page tables detected for iova 0x%llx\n",
|
|
||||||
(unsigned long long)iova);
|
|
||||||
return -EINVAL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int arm_smmu_domain_has_cap(struct iommu_domain *domain,
|
static int arm_smmu_domain_has_cap(struct iommu_domain *domain,
|
||||||
|
|
Loading…
Reference in New Issue