/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom.c | 136 flags = IOMMU_READ | IOMMU_CACHE; in usnic_uiom_get_pages() 257 int flags = IOMMU_READ | IOMMU_CACHE; in usnic_uiom_map_sorted_intervals()
|
/kernel/linux/linux-5.10/drivers/iommu/ |
H A D | io-pgtable-arm-v7s.c | 339 else if (prot & IOMMU_CACHE) in arm_v7s_prot_to_pte() 361 prot |= IOMMU_CACHE; in arm_v7s_pte_to_prot() 929 IOMMU_CACHE, GFP_KERNEL)) in arm_v7s_do_selftests()
|
H A D | io-pgtable-arm.c | 408 else if (prot & IOMMU_CACHE) in arm_lpae_prot_to_pte() 416 else if (prot & IOMMU_CACHE) in arm_lpae_prot_to_pte() 427 if (prot & IOMMU_CACHE && data->iop.fmt != ARM_MALI_LPAE) in arm_lpae_prot_to_pte() 1160 IOMMU_CACHE, GFP_KERNEL)) in arm_lpae_run_tests()
|
H A D | dma-iommu.c | 387 int prot = coherent ? IOMMU_CACHE : 0; in dma_info_to_prot() 628 if (!(ioprot & IOMMU_CACHE)) { in iommu_dma_alloc_remap()
|
/kernel/linux/linux-6.6/drivers/iommu/ |
H A D | io-pgtable-arm-v7s.c | 356 else if (prot & IOMMU_CACHE) in arm_v7s_prot_to_pte() 378 prot |= IOMMU_CACHE; in arm_v7s_pte_to_prot() 975 IOMMU_NOEXEC | IOMMU_CACHE, in arm_v7s_do_selftests()
|
H A D | of_iommu.c | 264 prot |= IOMMU_CACHE; in of_iommu_get_resv_regions()
|
H A D | io-pgtable-arm.c | 428 else if (prot & IOMMU_CACHE) in arm_lpae_prot_to_pte() 436 else if (prot & IOMMU_CACHE) in arm_lpae_prot_to_pte() 447 if (prot & IOMMU_CACHE && data->iop.fmt != ARM_MALI_LPAE) in arm_lpae_prot_to_pte() 1199 IOMMU_NOEXEC | IOMMU_CACHE, in arm_lpae_run_tests()
|
H A D | io-pgtable-dart.c | 229 if (!(prot & IOMMU_CACHE)) in dart_prot_to_pte()
|
H A D | dma-iommu.c | 642 int prot = coherent ? IOMMU_CACHE : 0; in dma_info_to_prot() 883 if (!(ioprot & IOMMU_CACHE)) { in __iommu_dma_alloc_noncontiguous()
|
/kernel/linux/linux-6.6/drivers/iommu/iommufd/ |
H A D | vfio_compat.c | 165 int iommu_prot = IOMMU_CACHE; in iommufd_vfio_map_dma()
|
H A D | ioas.c | 188 * operation. Support for IOMMU_CACHE is enforced by the in conv_iommu_prot() 191 int iommu_prot = IOMMU_CACHE; in conv_iommu_prot()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom.c | 252 int flags = IOMMU_READ | IOMMU_CACHE; in usnic_uiom_map_sorted_intervals()
|
/kernel/linux/linux-6.6/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 1438 npage << PAGE_SHIFT, prot | IOMMU_CACHE, in vfio_iommu_map() 1753 dma->prot | IOMMU_CACHE, GFP_KERNEL); in vfio_iommu_replay() 1848 IOMMU_READ | IOMMU_WRITE | IOMMU_CACHE, GFP_KERNEL); in vfio_test_domain_fgsp()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | iommu.h | 21 #define IOMMU_CACHE (1 << 2) /* DMA cache coherency */ macro
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | iommu.h | 20 #define IOMMU_CACHE (1 << 2) /* DMA cache coherency */ macro 122 IOMMU_CAP_CACHE_COHERENCY, /* IOMMU_CACHE is supported */ 325 * @enforce_cache_coherency: Prevent any kind of DMA from bypassing IOMMU_CACHE,
|
/kernel/linux/linux-5.10/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 84 int prot; /* IOMMU_CACHE */ 2223 domain->prot |= IOMMU_CACHE; in vfio_iommu_type1_attach_group() 2572 if (!(domain->prot & IOMMU_CACHE)) { in vfio_domains_have_iommu_cache()
|
/kernel/linux/linux-5.10/drivers/vhost/ |
H A D | vdpa.c | 563 return flags | IOMMU_CACHE; in perm_to_iommu_flags()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/msm/ |
H A D | msm_gem.c | 488 prot |= IOMMU_CACHE; in msm_gem_pin_vma_locked()
|
/kernel/linux/linux-6.6/drivers/acpi/arm64/ |
H A D | iort.c | 920 prot |= IOMMU_CACHE; in iort_get_rmrs()
|
/kernel/linux/linux-6.6/drivers/vhost/ |
H A D | vdpa.c | 877 return flags | IOMMU_CACHE; in perm_to_iommu_flags()
|
/kernel/linux/linux-5.10/drivers/iommu/intel/ |
H A D | iommu.c | 5673 if ((iommu_prot & IOMMU_CACHE) && dmar_domain->iommu_snooping) in intel_iommu_map()
|