Home
last modified time | relevance | path

Searched refs:VTD_PAGE_SHIFT (Results 1 - 12 of 12) sorted by relevance

/kernel/linux/linux-6.6/drivers/iommu/intel/
H A Dsvm.c198 unsigned long align = (1ULL << (VTD_PAGE_SHIFT + shift)); in intel_flush_svm_range_dev()
200 unsigned long end = ALIGN(address + (pages << VTD_PAGE_SHIFT), align); in intel_flush_svm_range_dev()
203 __flush_svm_range_dev(svm, sdev, start, align >> VTD_PAGE_SHIFT, ih); in intel_flush_svm_range_dev()
232 0, 64 - VTD_PAGE_SHIFT); in intel_flush_svm_all()
233 quirk_extra_dev_tlb_flush(info, 0, 64 - VTD_PAGE_SHIFT, in intel_flush_svm_all()
253 (end - start + PAGE_SIZE - 1) >> VTD_PAGE_SHIFT, 0); in intel_arch_invalidate_secondary_tlbs()
580 event.fault.prm.addr = (u64)desc->addr << VTD_PAGE_SHIFT; in intel_svm_prq_report()
671 address = (u64)req->addr << VTD_PAGE_SHIFT; in prq_event_thread()
H A Diommu.c50 #define MAX_AGAW_PFN_WIDTH (MAX_AGAW_WIDTH - VTD_PAGE_SHIFT)
52 #define __DOMAIN_MAX_PFN(gaw) ((((uint64_t)1) << ((gaw) - VTD_PAGE_SHIFT)) - 1)
59 #define DOMAIN_MAX_ADDR(gaw) (((uint64_t)__DOMAIN_MAX_PFN(gaw)) << VTD_PAGE_SHIFT)
119 return mm_pfn << (PAGE_SHIFT - VTD_PAGE_SHIFT); in mm_to_dma_pfn_start()
123 return ((mm_pfn + 1) << (PAGE_SHIFT - VTD_PAGE_SHIFT)) - 1; in mm_to_dma_pfn_end()
394 int addr_width = agaw_to_width(domain->agaw) - VTD_PAGE_SHIFT; in domain_pfn_supported()
911 pgtable_walk(iommu, addr >> VTD_PAGE_SHIFT, bus, devfn, pgtable, level); in dmar_fault_dump_ptes()
948 pteval = ((uint64_t)virt_to_dma_pfn(tmp_page) << VTD_PAGE_SHIFT) | DMA_PTE_READ | DMA_PTE_WRITE; in pfn_to_dma_pte()
1427 !pci_enable_ats(pdev, VTD_PAGE_SHIFT)) { in iommu_enable_pci_caps()
1522 uint64_t addr = (uint64_t)pfn << VTD_PAGE_SHIFT; in iommu_flush_iotlb_psi()
[all...]
H A Diommu.h32 #define VTD_PAGE_SHIFT (12) macro
33 #define VTD_PAGE_SIZE (1UL << VTD_PAGE_SHIFT)
34 #define VTD_PAGE_MASK (((u64)-1) << VTD_PAGE_SHIFT)
H A Ddebugfs.c308 return BIT_ULL(VTD_PAGE_SHIFT + VTD_STRIDE_SHIFT * (level - 1)); in level_to_directory_size()
315 iova >> VTD_PAGE_SHIFT, path[5], path[4], in dump_page_info()
H A Dpasid.c445 qi_flush_dev_iotlb(iommu, sid, pfsid, qdep, 0, 64 - VTD_PAGE_SHIFT); in devtlb_invalidation_with_pasid()
447 qi_flush_dev_iotlb_pasid(iommu, sid, pfsid, pasid, qdep, 0, 64 - VTD_PAGE_SHIFT); in devtlb_invalidation_with_pasid()
H A Ddmar.c1535 addr |= (1ULL << (VTD_PAGE_SHIFT + mask - 1)) - 1; in qi_flush_dev_iotlb()
1575 unsigned long align = (1ULL << (VTD_PAGE_SHIFT + mask)); in qi_flush_piotlb()
1596 unsigned long mask = 1UL << (VTD_PAGE_SHIFT + size_order - 1); in qi_flush_dev_iotlb_pasid()
1634 desc.qw1 |= GENMASK_ULL(size_order + VTD_PAGE_SHIFT - 1, in qi_flush_dev_iotlb_pasid()
1635 VTD_PAGE_SHIFT); in qi_flush_dev_iotlb_pasid()
/kernel/linux/linux-5.10/drivers/iommu/intel/
H A Diommu.c68 #define MAX_AGAW_PFN_WIDTH (MAX_AGAW_WIDTH - VTD_PAGE_SHIFT)
70 #define __DOMAIN_MAX_PFN(gaw) ((((uint64_t)1) << ((gaw) - VTD_PAGE_SHIFT)) - 1)
77 #define DOMAIN_MAX_ADDR(gaw) (((uint64_t)__DOMAIN_MAX_PFN(gaw)) << VTD_PAGE_SHIFT)
155 return dma_pfn >> (PAGE_SHIFT - VTD_PAGE_SHIFT); in dma_to_mm_pfn()
160 return mm_pfn << (PAGE_SHIFT - VTD_PAGE_SHIFT); in mm_to_dma_pfn()
558 int addr_width = agaw_to_width(domain->agaw) - VTD_PAGE_SHIFT; in domain_pfn_supported()
1052 pteval = ((uint64_t)virt_to_dma_pfn(tmp_page) << VTD_PAGE_SHIFT) | DMA_PTE_READ | DMA_PTE_WRITE; in pfn_to_dma_pte()
1574 !pci_enable_ats(pdev, VTD_PAGE_SHIFT)) { in iommu_enable_dev_iotlb()
1653 uint64_t addr = (uint64_t)pfn << VTD_PAGE_SHIFT; in iommu_flush_iotlb_psi()
2360 return PAGE_ALIGN(host_addr + size) >> VTD_PAGE_SHIFT; in aligned_nrpages()
[all...]
H A Dsvm.c144 unsigned long align = (1ULL << (VTD_PAGE_SHIFT + shift)); in intel_flush_svm_range_dev()
146 unsigned long end = ALIGN(address + (pages << VTD_PAGE_SHIFT), align); in intel_flush_svm_range_dev()
149 __flush_svm_range_dev(svm, sdev, start, align >> VTD_PAGE_SHIFT, ih); in intel_flush_svm_range_dev()
173 (end - start + PAGE_SIZE - 1) >> VTD_PAGE_SHIFT, 0); in intel_invalidate_range()
865 event.fault.prm.addr = (u64)desc->addr << VTD_PAGE_SHIFT; in intel_svm_prq_report()
918 address = (u64)req->addr << VTD_PAGE_SHIFT; in prq_event_thread()
H A Ddebugfs.c306 return BIT_ULL(VTD_PAGE_SHIFT + VTD_STRIDE_SHIFT * (level - 1)); in level_to_directory_size()
313 iova >> VTD_PAGE_SHIFT, path[5], path[4], in dump_page_info()
H A Dpasid.c503 qi_flush_dev_iotlb(iommu, sid, pfsid, qdep, 0, 64 - VTD_PAGE_SHIFT); in devtlb_invalidation_with_pasid()
505 qi_flush_dev_iotlb_pasid(iommu, sid, pfsid, pasid, qdep, 0, 64 - VTD_PAGE_SHIFT); in devtlb_invalidation_with_pasid()
H A Ddmar.c1425 addr |= (1ULL << (VTD_PAGE_SHIFT + mask - 1)) - 1; in qi_flush_dev_iotlb()
1465 unsigned long align = (1ULL << (VTD_PAGE_SHIFT + mask)); in qi_flush_piotlb()
1486 unsigned long mask = 1UL << (VTD_PAGE_SHIFT + size_order - 1); in qi_flush_dev_iotlb_pasid()
1515 desc.qw1 |= GENMASK_ULL(size_order + VTD_PAGE_SHIFT - 1, in qi_flush_dev_iotlb_pasid()
1516 VTD_PAGE_SHIFT); in qi_flush_dev_iotlb_pasid()
/kernel/linux/linux-5.10/include/linux/
H A Dintel-iommu.h30 #define VTD_PAGE_SHIFT (12) macro
31 #define VTD_PAGE_SIZE (1UL << VTD_PAGE_SHIFT)
32 #define VTD_PAGE_MASK (((u64)-1) << VTD_PAGE_SHIFT)

Completed in 25 milliseconds