/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | tlb.c | 206 u16 *new_asid, bool *need_flush) in choose_new_asid() 212 *need_flush = true; in choose_new_asid() 225 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid() 239 *need_flush = true; in choose_new_asid() 268 static void load_new_mm_cr3(pgd_t *pgdir, u16 new_asid, bool need_flush) in load_new_mm_cr3() argument 272 if (need_flush) { in load_new_mm_cr3() 430 bool need_flush; in switch_mm_irqs_off() local 524 need_flush = true; in switch_mm_irqs_off() 551 choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush); in switch_mm_irqs_off() 558 if (need_flush) { in switch_mm_irqs_off() 205 choose_new_asid(struct mm_struct *next, u64 next_tlb_gen, u16 *new_asid, bool *need_flush) choose_new_asid() argument [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | tlb.c | 220 u16 *new_asid, bool *need_flush) in choose_new_asid() 226 *need_flush = true; in choose_new_asid() 239 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid() 253 *need_flush = true; in choose_new_asid() 283 bool need_flush) in load_new_mm_cr3() 287 if (need_flush) { in load_new_mm_cr3() 504 bool need_flush; in switch_mm_irqs_off() local 606 need_flush = true; in switch_mm_irqs_off() 632 choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush); in switch_mm_irqs_off() 640 if (need_flush) { in switch_mm_irqs_off() 219 choose_new_asid(struct mm_struct *next, u64 next_tlb_gen, u16 *new_asid, bool *need_flush) choose_new_asid() argument 282 load_new_mm_cr3(pgd_t *pgdir, u16 new_asid, unsigned long lam, bool need_flush) load_new_mm_cr3() argument [all...] |
/kernel/linux/linux-5.10/arch/arm/mm/ |
H A D | pmsa-v7.c | 368 unsigned int subregions, bool need_flush) in mpu_setup_region() 387 if (need_flush) in mpu_setup_region() 441 bool need_flush = region == PMSAv7_RAM_REGION; in pmsav7_setup() local 448 xip[i].subreg, need_flush); in pmsav7_setup() 366 mpu_setup_region(unsigned int number, phys_addr_t start, unsigned int size_order, unsigned int properties, unsigned int subregions, bool need_flush) mpu_setup_region() argument
|
/kernel/linux/linux-6.6/arch/arm/mm/ |
H A D | pmsa-v7.c | 368 unsigned int subregions, bool need_flush) in mpu_setup_region() 387 if (need_flush) in mpu_setup_region() 441 bool need_flush = region == PMSAv7_RAM_REGION; in pmsav7_setup() local 448 xip[i].subreg, need_flush); in pmsav7_setup() 366 mpu_setup_region(unsigned int number, phys_addr_t start, unsigned int size_order, unsigned int properties, unsigned int subregions, bool need_flush) mpu_setup_region() argument
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | amd_gart_64.c | 89 static bool need_flush; /* global flush state. set for each gart wrap */ variable 106 need_flush = true; in alloc_iommu() 115 need_flush = true; in alloc_iommu() 119 need_flush = true; in alloc_iommu() 144 if (need_flush) { in flush_gart() 146 need_flush = false; in flush_gart()
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | amd_gart_64.c | 87 static bool need_flush; /* global flush state. set for each gart wrap */ variable 104 need_flush = true; in alloc_iommu() 113 need_flush = true; in alloc_iommu() 117 need_flush = true; in alloc_iommu() 142 if (need_flush) { in flush_gart() 144 need_flush = false; in flush_gart()
|
/kernel/linux/linux-5.10/mm/ |
H A D | highmem.c | 166 int need_flush = 0; in flush_all_zero_pkmaps() local 197 need_flush = 1; in flush_all_zero_pkmaps() 199 if (need_flush) in flush_all_zero_pkmaps()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_buffer.c | 350 bool need_flush = switch_mmu_context || gpu->flush_seq != new_flush_seq; in etnaviv_buffer_queue() local 368 if (need_flush || switch_context) { in etnaviv_buffer_queue() 375 if (need_flush) { in etnaviv_buffer_queue() 404 if (need_flush) { in etnaviv_buffer_queue()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_buffer.c | 351 bool need_flush = switch_mmu_context || gpu->flush_seq != new_flush_seq; in etnaviv_buffer_queue() local 369 if (need_flush || switch_context) { in etnaviv_buffer_queue() 376 if (need_flush) { in etnaviv_buffer_queue() 405 if (need_flush) { in etnaviv_buffer_queue()
|
/kernel/linux/linux-5.10/arch/sparc/kernel/ |
H A D | iommu-common.c | 19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() function 206 (n < pool->hint || need_flush(iommu))) { in iommu_tbl_range_alloc()
|
/kernel/linux/linux-6.6/arch/sparc/kernel/ |
H A D | iommu-common.c | 19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() function 206 (n < pool->hint || need_flush(iommu))) { in iommu_tbl_range_alloc()
|
/kernel/linux/linux-6.6/mm/ |
H A D | highmem.c | 188 int need_flush = 0; in flush_all_zero_pkmaps() local 221 need_flush = 1; in flush_all_zero_pkmaps() 223 if (need_flush) in flush_all_zero_pkmaps()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | cm.c | 528 int need_flush = 0; in mlx4_ib_cm_paravirt_clean() local 536 need_flush |= !cancel_delayed_work(&map->timeout); in mlx4_ib_cm_paravirt_clean() 542 if (need_flush) in mlx4_ib_cm_paravirt_clean()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | cm.c | 529 int need_flush = 0; in mlx4_ib_cm_paravirt_clean() local 537 need_flush |= !cancel_delayed_work(&map->timeout); in mlx4_ib_cm_paravirt_clean() 543 if (need_flush) in mlx4_ib_cm_paravirt_clean()
|
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
H A D | tdp_mmu.c | 791 int need_flush = 0; in set_tdp_spte() local 815 need_flush = 1; in set_tdp_spte() 818 if (need_flush) in set_tdp_spte()
|
/kernel/linux/linux-5.10/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 664 bool need_flush = false; in stage2_unmap_walker() local 675 need_flush = true; in stage2_unmap_walker() 687 if (need_flush) { in stage2_unmap_walker()
|
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 1097 bool need_flush = false; in stage2_unmap_walker() local 1113 need_flush = !stage2_has_fwb(pgt); in stage2_unmap_walker() 1123 if (need_flush && mm_ops->dcache_clean_inval_poc) in stage2_unmap_walker()
|
/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_log.c | 1693 bool need_flush) in xlog_write_iclog() 1733 if (need_flush) in xlog_write_iclog() 1851 bool need_flush = true, split = false; in xlog_sync() local 1911 need_flush = false; in xlog_sync() 1915 xlog_write_iclog(log, iclog, bno, count, need_flush); in xlog_sync() 1688 xlog_write_iclog( struct xlog *log, struct xlog_in_core *iclog, uint64_t bno, unsigned int count, bool need_flush) xlog_write_iclog() argument
|
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm-writecache.c | 975 bool need_flush = false; in writecache_resume() local 1043 need_flush = true; in writecache_resume() 1064 need_flush = true; in writecache_resume() 1071 if (need_flush) { in writecache_resume()
|
/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-writecache.c | 1000 bool need_flush = false; in writecache_resume() local 1071 need_flush = true; in writecache_resume() 1092 need_flush = true; in writecache_resume() 1099 if (need_flush) { in writecache_resume()
|
/kernel/linux/linux-5.10/fs/ceph/ |
H A D | snap.c | 535 capsnap->need_flush = true; in ceph_queue_cap_snap() 547 capsnap->need_flush ? "" : "no_flush"); in ceph_queue_cap_snap()
|
H A D | super.h | 222 bool need_flush; member
|
/kernel/linux/linux-6.6/fs/ceph/ |
H A D | snap.c | 574 capsnap->need_flush = true; in ceph_queue_cap_snap() 586 ceph_cap_string(dirty), capsnap->need_flush ? "" : "no_flush"); in ceph_queue_cap_snap()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_context.c | 461 unsigned int n, m, need_flush; in cpu_fill() local 465 err = i915_gem_object_prepare_write(obj, &need_flush); in cpu_fill()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_context.c | 480 unsigned int need_flush; in cpu_fill() local 485 err = i915_gem_object_prepare_write(obj, &need_flush); in cpu_fill()
|