/kernel/linux/linux-5.10/arch/csky/include/asm/ |
H A D | tlb.h | 11 if (!(tlb)->fullmm) \ 17 if (!(tlb)->fullmm) \
|
/kernel/linux/linux-5.10/include/asm-generic/ |
H A D | tlb.h | 92 * - mmu_gather::fullmm 267 unsigned int fullmm : 1; member 319 if (tlb->fullmm) { in __tlb_reset_range() 378 if (tlb->fullmm || tlb->need_flush_all) { in tlb_flush() 459 if (!tlb->fullmm && !tlb->need_flush_all) in tlb_change_page_size() 494 if (tlb->fullmm) in tlb_start_vma() 505 if (tlb->fullmm) in tlb_end_vma()
|
/kernel/linux/linux-6.6/include/asm-generic/ |
H A D | tlb.h | 94 * - mmu_gather::fullmm 307 unsigned int fullmm : 1; member 365 if (tlb->fullmm) { in __tlb_reset_range() 413 if (tlb->fullmm || tlb->need_flush_all) { in tlb_flush() 499 if (!tlb->fullmm && !tlb->need_flush_all) in tlb_change_page_size() 533 if (tlb->fullmm) in tlb_start_vma() 544 if (tlb->fullmm) in tlb_end_vma()
|
/kernel/linux/linux-6.6/mm/ |
H A D | mmu_gather.c | 304 bool fullmm) in __tlb_gather_mmu() 307 tlb->fullmm = fullmm; in __tlb_gather_mmu() 380 * The aarch64 yields better performance with fullmm by in tlb_finish_mmu() 384 * On x86 non-fullmm doesn't yield significant difference in tlb_finish_mmu() 385 * against fullmm. in tlb_finish_mmu() 387 tlb->fullmm = 1; in tlb_finish_mmu() 303 __tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, bool fullmm) __tlb_gather_mmu() argument
|
H A D | madvise.c | 501 tlb->fullmm); in madvise_cold_or_pageout_pte_range() 666 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range() 669 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range() 739 tlb->fullmm); in madvise_free_pte_range()
|
H A D | huge_memory.c | 1677 tlb->fullmm); in zap_huge_pmd() 1979 pudp_huge_get_and_clear_full(vma, addr, pud, tlb->fullmm); in zap_huge_pud()
|
H A D | memory.c | 1438 tlb->fullmm); in zap_pte_range() 1522 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range()
|
/kernel/linux/linux-5.10/arch/loongarch/include/asm/ |
H A D | tlb.h | 141 if (!(tlb)->fullmm) \ 158 if (tlb->fullmm) { in tlb_flush()
|
/kernel/linux/linux-5.10/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 897 pte_t *ptep, pte_t orig, int fullmm, 901 pte_t *ptep, pte_t orig, int fullmm, in maybe_tlb_batch_add() 911 tlb_batch_add(mm, vaddr, ptep, orig, fullmm, hugepage_shift); in maybe_tlb_batch_add() 925 pte_t *ptep, pte_t pte, int fullmm) in __set_pte_at() 930 maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT); in __set_pte_at() 940 #define pte_clear_not_present_full(mm,addr,ptep,fullmm) \ 941 __set_pte_at((mm), (addr), (ptep), __pte(0UL), (fullmm)) 900 maybe_tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) maybe_tlb_batch_add() argument 924 __set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, int fullmm) __set_pte_at() argument
|
/kernel/linux/linux-6.6/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 895 pte_t *ptep, pte_t orig, int fullmm, 899 pte_t *ptep, pte_t orig, int fullmm, in maybe_tlb_batch_add() 909 tlb_batch_add(mm, vaddr, ptep, orig, fullmm, hugepage_shift); in maybe_tlb_batch_add() 923 pte_t *ptep, pte_t pte, int fullmm) in __set_pte_at() 928 maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT); in __set_pte_at() 951 #define pte_clear_not_present_full(mm,addr,ptep,fullmm) \ 952 __set_pte_at((mm), (addr), (ptep), __pte(0UL), (fullmm)) 898 maybe_tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) maybe_tlb_batch_add() argument 922 __set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, int fullmm) __set_pte_at() argument
|
/kernel/linux/linux-6.6/arch/riscv/include/asm/ |
H A D | tlb.h | 19 if (tlb->fullmm || tlb->need_flush_all || tlb->freed_tables) in tlb_flush()
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | tlb.h | 19 if (!tlb->fullmm && !tlb->need_flush_all) { in tlb_flush()
|
/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | tlb.h | 15 if (!tlb->fullmm && !tlb->need_flush_all) { in tlb_flush()
|
/kernel/linux/linux-5.10/mm/ |
H A D | mmu_gather.c | 268 tlb->fullmm = !(start | (end+1)); in tlb_gather_mmu() 314 * The aarch64 yields better performance with fullmm by in tlb_finish_mmu() 318 * On x86 non-fullmm doesn't yield significant difference in tlb_finish_mmu() 319 * against fullmm. in tlb_finish_mmu() 321 tlb->fullmm = 1; in tlb_finish_mmu()
|
H A D | madvise.c | 454 tlb->fullmm); in madvise_cold_or_pageout_pte_range() 620 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range() 687 tlb->fullmm); in madvise_free_pte_range()
|
H A D | huge_memory.c | 1668 tlb->fullmm); in zap_huge_pmd() 1942 pudp_huge_get_and_clear_full(tlb->mm, addr, pud, tlb->fullmm); in zap_huge_pud()
|
H A D | memory.c | 1276 tlb->fullmm); in zap_pte_range() 1318 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range() 1341 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range()
|
/kernel/linux/linux-5.10/arch/arm64/include/asm/ |
H A D | tlb.h | 65 if (tlb->fullmm) { in tlb_flush()
|
/kernel/linux/linux-6.6/arch/arm64/include/asm/ |
H A D | tlb.h | 65 if (tlb->fullmm) { in tlb_flush()
|
/kernel/linux/linux-5.10/arch/sparc/mm/ |
H A D | tlb.c | 112 pte_t *ptep, pte_t orig, int fullmm, in tlb_batch_add() 139 if (!fullmm) in tlb_batch_add() 111 tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) tlb_batch_add() argument
|
/kernel/linux/linux-6.6/arch/sparc/mm/ |
H A D | tlb.c | 113 pte_t *ptep, pte_t orig, int fullmm, in tlb_batch_add() 142 if (!fullmm) in tlb_batch_add() 112 tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) tlb_batch_add() argument
|
/kernel/linux/linux-6.6/arch/powerpc/mm/book3s64/ |
H A D | radix_tlb.c | 609 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument 765 static enum tlb_flush_type flush_type_needed(struct mm_struct *mm, bool fullmm) in flush_type_needed() argument 792 * In the fullmm case there's no point doing the exit_flush_lazy_tlbs in flush_type_needed() 796 if (fullmm) in flush_type_needed() 875 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument 886 type = flush_type_needed(mm, fullmm); in __flush_all_mm() 1195 * A "fullmm" flush must always do a flush_all_mm (RIC=2) flush in radix__tlb_flush() 1199 if (tlb->fullmm) { in radix__tlb_flush()
|
/kernel/linux/linux-6.6/arch/loongarch/include/asm/ |
H A D | tlb.h | 148 if (tlb->fullmm) { in tlb_flush()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/book3s64/ |
H A D | radix_tlb.c | 600 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument 738 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument 750 if (!fullmm) { in __flush_all_mm() 1057 * A "fullmm" flush must always do a flush_all_mm (RIC=2) flush in radix__tlb_flush() 1061 if (tlb->fullmm || tlb->need_flush_all) { in radix__tlb_flush()
|