| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * cache.c - Intel VT-d cache invalidation |
| * |
| * Copyright (C) 2024 Intel Corporation |
| * |
| * Author: Lu Baolu <baolu.lu@linux.intel.com> |
| */ |
| |
| #define pr_fmt(fmt) "DMAR: " fmt |
| |
| #include <linux/dmar.h> |
| #include <linux/iommu.h> |
| #include <linux/memory.h> |
| #include <linux/pci.h> |
| #include <linux/spinlock.h> |
| |
| #include "iommu.h" |
| #include "pasid.h" |
| #include "trace.h" |
| |
| /* Check if an existing cache tag can be reused for a new association. */ |
| static bool cache_tage_match(struct cache_tag *tag, u16 domain_id, |
| struct intel_iommu *iommu, struct device *dev, |
| ioasid_t pasid, enum cache_tag_type type) |
| { |
| if (tag->type != type) |
| return false; |
| |
| if (tag->domain_id != domain_id || tag->pasid != pasid) |
| return false; |
| |
| if (type == CACHE_TAG_IOTLB || type == CACHE_TAG_NESTING_IOTLB) |
| return tag->iommu == iommu; |
| |
| if (type == CACHE_TAG_DEVTLB || type == CACHE_TAG_NESTING_DEVTLB) |
| return tag->dev == dev; |
| |
| return false; |
| } |
| |
| /* Assign a cache tag with specified type to domain. */ |
| static int cache_tag_assign(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid, |
| enum cache_tag_type type) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| struct intel_iommu *iommu = info->iommu; |
| struct cache_tag *tag, *temp; |
| unsigned long flags; |
| |
| tag = kzalloc(sizeof(*tag), GFP_KERNEL); |
| if (!tag) |
| return -ENOMEM; |
| |
| tag->type = type; |
| tag->iommu = iommu; |
| tag->domain_id = did; |
| tag->pasid = pasid; |
| tag->users = 1; |
| |
| if (type == CACHE_TAG_DEVTLB || type == CACHE_TAG_NESTING_DEVTLB) |
| tag->dev = dev; |
| else |
| tag->dev = iommu->iommu.dev; |
| |
| spin_lock_irqsave(&domain->cache_lock, flags); |
| list_for_each_entry(temp, &domain->cache_tags, node) { |
| if (cache_tage_match(temp, did, iommu, dev, pasid, type)) { |
| temp->users++; |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| kfree(tag); |
| trace_cache_tag_assign(temp); |
| return 0; |
| } |
| } |
| list_add_tail(&tag->node, &domain->cache_tags); |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| trace_cache_tag_assign(tag); |
| |
| return 0; |
| } |
| |
| /* Unassign a cache tag with specified type from domain. */ |
| static void cache_tag_unassign(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid, |
| enum cache_tag_type type) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| struct intel_iommu *iommu = info->iommu; |
| struct cache_tag *tag; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&domain->cache_lock, flags); |
| list_for_each_entry(tag, &domain->cache_tags, node) { |
| if (cache_tage_match(tag, did, iommu, dev, pasid, type)) { |
| trace_cache_tag_unassign(tag); |
| if (--tag->users == 0) { |
| list_del(&tag->node); |
| kfree(tag); |
| } |
| break; |
| } |
| } |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| } |
| |
| static int __cache_tag_assign_domain(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| int ret; |
| |
| ret = cache_tag_assign(domain, did, dev, pasid, CACHE_TAG_IOTLB); |
| if (ret || !info->ats_enabled) |
| return ret; |
| |
| ret = cache_tag_assign(domain, did, dev, pasid, CACHE_TAG_DEVTLB); |
| if (ret) |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_IOTLB); |
| |
| return ret; |
| } |
| |
| static void __cache_tag_unassign_domain(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_IOTLB); |
| |
| if (info->ats_enabled) |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_DEVTLB); |
| } |
| |
| static int __cache_tag_assign_parent_domain(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| int ret; |
| |
| ret = cache_tag_assign(domain, did, dev, pasid, CACHE_TAG_NESTING_IOTLB); |
| if (ret || !info->ats_enabled) |
| return ret; |
| |
| ret = cache_tag_assign(domain, did, dev, pasid, CACHE_TAG_NESTING_DEVTLB); |
| if (ret) |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_NESTING_IOTLB); |
| |
| return ret; |
| } |
| |
| static void __cache_tag_unassign_parent_domain(struct dmar_domain *domain, u16 did, |
| struct device *dev, ioasid_t pasid) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_NESTING_IOTLB); |
| |
| if (info->ats_enabled) |
| cache_tag_unassign(domain, did, dev, pasid, CACHE_TAG_NESTING_DEVTLB); |
| } |
| |
| static u16 domain_get_id_for_dev(struct dmar_domain *domain, struct device *dev) |
| { |
| struct device_domain_info *info = dev_iommu_priv_get(dev); |
| struct intel_iommu *iommu = info->iommu; |
| |
| /* |
| * The driver assigns different domain IDs for all domains except |
| * the SVA type. |
| */ |
| if (domain->domain.type == IOMMU_DOMAIN_SVA) |
| return FLPT_DEFAULT_DID; |
| |
| return domain_id_iommu(domain, iommu); |
| } |
| |
| /* |
| * Assign cache tags to a domain when it's associated with a device's |
| * PASID using a specific domain ID. |
| * |
| * On success (return value of 0), cache tags are created and added to the |
| * domain's cache tag list. On failure (negative return value), an error |
| * code is returned indicating the reason for the failure. |
| */ |
| int cache_tag_assign_domain(struct dmar_domain *domain, |
| struct device *dev, ioasid_t pasid) |
| { |
| u16 did = domain_get_id_for_dev(domain, dev); |
| int ret; |
| |
| /* domain->qi_bach will be freed in iommu_free_domain() path. */ |
| if (!domain->qi_batch) { |
| domain->qi_batch = kzalloc(sizeof(*domain->qi_batch), GFP_KERNEL); |
| if (!domain->qi_batch) |
| return -ENOMEM; |
| } |
| |
| ret = __cache_tag_assign_domain(domain, did, dev, pasid); |
| if (ret || domain->domain.type != IOMMU_DOMAIN_NESTED) |
| return ret; |
| |
| ret = __cache_tag_assign_parent_domain(domain->s2_domain, did, dev, pasid); |
| if (ret) |
| __cache_tag_unassign_domain(domain, did, dev, pasid); |
| |
| return ret; |
| } |
| |
| /* |
| * Remove the cache tags associated with a device's PASID when the domain is |
| * detached from the device. |
| * |
| * The cache tags must be previously assigned to the domain by calling the |
| * assign interface. |
| */ |
| void cache_tag_unassign_domain(struct dmar_domain *domain, |
| struct device *dev, ioasid_t pasid) |
| { |
| u16 did = domain_get_id_for_dev(domain, dev); |
| |
| __cache_tag_unassign_domain(domain, did, dev, pasid); |
| if (domain->domain.type == IOMMU_DOMAIN_NESTED) |
| __cache_tag_unassign_parent_domain(domain->s2_domain, did, dev, pasid); |
| } |
| |
| static unsigned long calculate_psi_aligned_address(unsigned long start, |
| unsigned long end, |
| unsigned long *_pages, |
| unsigned long *_mask) |
| { |
| unsigned long pages = aligned_nrpages(start, end - start + 1); |
| unsigned long aligned_pages = __roundup_pow_of_two(pages); |
| unsigned long bitmask = aligned_pages - 1; |
| unsigned long mask = ilog2(aligned_pages); |
| unsigned long pfn = IOVA_PFN(start); |
| |
| /* |
| * PSI masks the low order bits of the base address. If the |
| * address isn't aligned to the mask, then compute a mask value |
| * needed to ensure the target range is flushed. |
| */ |
| if (unlikely(bitmask & pfn)) { |
| unsigned long end_pfn = pfn + pages - 1, shared_bits; |
| |
| /* |
| * Since end_pfn <= pfn + bitmask, the only way bits |
| * higher than bitmask can differ in pfn and end_pfn is |
| * by carrying. This means after masking out bitmask, |
| * high bits starting with the first set bit in |
| * shared_bits are all equal in both pfn and end_pfn. |
| */ |
| shared_bits = ~(pfn ^ end_pfn) & ~bitmask; |
| mask = shared_bits ? __ffs(shared_bits) : MAX_AGAW_PFN_WIDTH; |
| aligned_pages = 1UL << mask; |
| } |
| |
| *_pages = aligned_pages; |
| *_mask = mask; |
| |
| return ALIGN_DOWN(start, VTD_PAGE_SIZE << mask); |
| } |
| |
| static void qi_batch_flush_descs(struct intel_iommu *iommu, struct qi_batch *batch) |
| { |
| if (!iommu || !batch->index) |
| return; |
| |
| qi_submit_sync(iommu, batch->descs, batch->index, 0); |
| |
| /* Reset the index value and clean the whole batch buffer. */ |
| memset(batch, 0, sizeof(*batch)); |
| } |
| |
| static void qi_batch_increment_index(struct intel_iommu *iommu, struct qi_batch *batch) |
| { |
| if (++batch->index == QI_MAX_BATCHED_DESC_COUNT) |
| qi_batch_flush_descs(iommu, batch); |
| } |
| |
| static void qi_batch_add_iotlb(struct intel_iommu *iommu, u16 did, u64 addr, |
| unsigned int size_order, u64 type, |
| struct qi_batch *batch) |
| { |
| qi_desc_iotlb(iommu, did, addr, size_order, type, &batch->descs[batch->index]); |
| qi_batch_increment_index(iommu, batch); |
| } |
| |
| static void qi_batch_add_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 pfsid, |
| u16 qdep, u64 addr, unsigned int mask, |
| struct qi_batch *batch) |
| { |
| /* |
| * According to VT-d spec, software is recommended to not submit any Device-TLB |
| * invalidation requests while address remapping hardware is disabled. |
| */ |
| if (!(iommu->gcmd & DMA_GCMD_TE)) |
| return; |
| |
| qi_desc_dev_iotlb(sid, pfsid, qdep, addr, mask, &batch->descs[batch->index]); |
| qi_batch_increment_index(iommu, batch); |
| } |
| |
| static void qi_batch_add_piotlb(struct intel_iommu *iommu, u16 did, u32 pasid, |
| u64 addr, unsigned long npages, bool ih, |
| struct qi_batch *batch) |
| { |
| /* |
| * npages == -1 means a PASID-selective invalidation, otherwise, |
| * a positive value for Page-selective-within-PASID invalidation. |
| * 0 is not a valid input. |
| */ |
| if (!npages) |
| return; |
| |
| qi_desc_piotlb(did, pasid, addr, npages, ih, &batch->descs[batch->index]); |
| qi_batch_increment_index(iommu, batch); |
| } |
| |
| static void qi_batch_add_pasid_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 pfsid, |
| u32 pasid, u16 qdep, u64 addr, |
| unsigned int size_order, struct qi_batch *batch) |
| { |
| /* |
| * According to VT-d spec, software is recommended to not submit any |
| * Device-TLB invalidation requests while address remapping hardware |
| * is disabled. |
| */ |
| if (!(iommu->gcmd & DMA_GCMD_TE)) |
| return; |
| |
| qi_desc_dev_iotlb_pasid(sid, pfsid, pasid, qdep, addr, size_order, |
| &batch->descs[batch->index]); |
| qi_batch_increment_index(iommu, batch); |
| } |
| |
| static void cache_tag_flush_iotlb(struct dmar_domain *domain, struct cache_tag *tag, |
| unsigned long addr, unsigned long pages, |
| unsigned long mask, int ih) |
| { |
| struct intel_iommu *iommu = tag->iommu; |
| u64 type = DMA_TLB_PSI_FLUSH; |
| |
| if (domain->use_first_level) { |
| qi_batch_add_piotlb(iommu, tag->domain_id, tag->pasid, addr, |
| pages, ih, domain->qi_batch); |
| return; |
| } |
| |
| /* |
| * Fallback to domain selective flush if no PSI support or the size |
| * is too big. |
| */ |
| if (!cap_pgsel_inv(iommu->cap) || |
| mask > cap_max_amask_val(iommu->cap) || pages == -1) { |
| addr = 0; |
| mask = 0; |
| ih = 0; |
| type = DMA_TLB_DSI_FLUSH; |
| } |
| |
| if (ecap_qis(iommu->ecap)) |
| qi_batch_add_iotlb(iommu, tag->domain_id, addr | ih, mask, type, |
| domain->qi_batch); |
| else |
| __iommu_flush_iotlb(iommu, tag->domain_id, addr | ih, mask, type); |
| } |
| |
| static void cache_tag_flush_devtlb_psi(struct dmar_domain *domain, struct cache_tag *tag, |
| unsigned long addr, unsigned long mask) |
| { |
| struct intel_iommu *iommu = tag->iommu; |
| struct device_domain_info *info; |
| u16 sid; |
| |
| info = dev_iommu_priv_get(tag->dev); |
| sid = PCI_DEVID(info->bus, info->devfn); |
| |
| if (tag->pasid == IOMMU_NO_PASID) { |
| qi_batch_add_dev_iotlb(iommu, sid, info->pfsid, info->ats_qdep, |
| addr, mask, domain->qi_batch); |
| if (info->dtlb_extra_inval) |
| qi_batch_add_dev_iotlb(iommu, sid, info->pfsid, info->ats_qdep, |
| addr, mask, domain->qi_batch); |
| return; |
| } |
| |
| qi_batch_add_pasid_dev_iotlb(iommu, sid, info->pfsid, tag->pasid, |
| info->ats_qdep, addr, mask, domain->qi_batch); |
| if (info->dtlb_extra_inval) |
| qi_batch_add_pasid_dev_iotlb(iommu, sid, info->pfsid, tag->pasid, |
| info->ats_qdep, addr, mask, |
| domain->qi_batch); |
| } |
| |
| static void cache_tag_flush_devtlb_all(struct dmar_domain *domain, struct cache_tag *tag) |
| { |
| struct intel_iommu *iommu = tag->iommu; |
| struct device_domain_info *info; |
| u16 sid; |
| |
| info = dev_iommu_priv_get(tag->dev); |
| sid = PCI_DEVID(info->bus, info->devfn); |
| |
| qi_batch_add_dev_iotlb(iommu, sid, info->pfsid, info->ats_qdep, 0, |
| MAX_AGAW_PFN_WIDTH, domain->qi_batch); |
| if (info->dtlb_extra_inval) |
| qi_batch_add_dev_iotlb(iommu, sid, info->pfsid, info->ats_qdep, 0, |
| MAX_AGAW_PFN_WIDTH, domain->qi_batch); |
| } |
| |
| /* |
| * Invalidates a range of IOVA from @start (inclusive) to @end (inclusive) |
| * when the memory mappings in the target domain have been modified. |
| */ |
| void cache_tag_flush_range(struct dmar_domain *domain, unsigned long start, |
| unsigned long end, int ih) |
| { |
| struct intel_iommu *iommu = NULL; |
| unsigned long pages, mask, addr; |
| struct cache_tag *tag; |
| unsigned long flags; |
| |
| addr = calculate_psi_aligned_address(start, end, &pages, &mask); |
| |
| spin_lock_irqsave(&domain->cache_lock, flags); |
| list_for_each_entry(tag, &domain->cache_tags, node) { |
| if (iommu && iommu != tag->iommu) |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| iommu = tag->iommu; |
| |
| switch (tag->type) { |
| case CACHE_TAG_IOTLB: |
| case CACHE_TAG_NESTING_IOTLB: |
| cache_tag_flush_iotlb(domain, tag, addr, pages, mask, ih); |
| break; |
| case CACHE_TAG_NESTING_DEVTLB: |
| /* |
| * Address translation cache in device side caches the |
| * result of nested translation. There is no easy way |
| * to identify the exact set of nested translations |
| * affected by a change in S2. So just flush the entire |
| * device cache. |
| */ |
| addr = 0; |
| mask = MAX_AGAW_PFN_WIDTH; |
| fallthrough; |
| case CACHE_TAG_DEVTLB: |
| cache_tag_flush_devtlb_psi(domain, tag, addr, mask); |
| break; |
| } |
| |
| trace_cache_tag_flush_range(tag, start, end, addr, pages, mask); |
| } |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| } |
| |
| /* |
| * Invalidates all ranges of IOVA when the memory mappings in the target |
| * domain have been modified. |
| */ |
| void cache_tag_flush_all(struct dmar_domain *domain) |
| { |
| struct intel_iommu *iommu = NULL; |
| struct cache_tag *tag; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&domain->cache_lock, flags); |
| list_for_each_entry(tag, &domain->cache_tags, node) { |
| if (iommu && iommu != tag->iommu) |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| iommu = tag->iommu; |
| |
| switch (tag->type) { |
| case CACHE_TAG_IOTLB: |
| case CACHE_TAG_NESTING_IOTLB: |
| cache_tag_flush_iotlb(domain, tag, 0, -1, 0, 0); |
| break; |
| case CACHE_TAG_DEVTLB: |
| case CACHE_TAG_NESTING_DEVTLB: |
| cache_tag_flush_devtlb_all(domain, tag); |
| break; |
| } |
| |
| trace_cache_tag_flush_all(tag); |
| } |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| } |
| |
| /* |
| * Invalidate a range of IOVA when new mappings are created in the target |
| * domain. |
| * |
| * - VT-d spec, Section 6.1 Caching Mode: When the CM field is reported as |
| * Set, any software updates to remapping structures other than first- |
| * stage mapping requires explicit invalidation of the caches. |
| * - VT-d spec, Section 6.8 Write Buffer Flushing: For hardware that requires |
| * write buffer flushing, software must explicitly perform write-buffer |
| * flushing, if cache invalidation is not required. |
| */ |
| void cache_tag_flush_range_np(struct dmar_domain *domain, unsigned long start, |
| unsigned long end) |
| { |
| struct intel_iommu *iommu = NULL; |
| unsigned long pages, mask, addr; |
| struct cache_tag *tag; |
| unsigned long flags; |
| |
| addr = calculate_psi_aligned_address(start, end, &pages, &mask); |
| |
| spin_lock_irqsave(&domain->cache_lock, flags); |
| list_for_each_entry(tag, &domain->cache_tags, node) { |
| if (iommu && iommu != tag->iommu) |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| iommu = tag->iommu; |
| |
| if (!cap_caching_mode(iommu->cap) || domain->use_first_level) { |
| iommu_flush_write_buffer(iommu); |
| continue; |
| } |
| |
| if (tag->type == CACHE_TAG_IOTLB || |
| tag->type == CACHE_TAG_NESTING_IOTLB) |
| cache_tag_flush_iotlb(domain, tag, addr, pages, mask, 0); |
| |
| trace_cache_tag_flush_range_np(tag, start, end, addr, pages, mask); |
| } |
| qi_batch_flush_descs(iommu, domain->qi_batch); |
| spin_unlock_irqrestore(&domain->cache_lock, flags); |
| } |