Home
last modified time | relevance | path

Searched refs:fullmm (Results 1 - 24 of 24) sorted by relevance

/kernel/linux/linux-5.10/arch/csky/include/asm/
H A Dtlb.h11 if (!(tlb)->fullmm) \
17 if (!(tlb)->fullmm) \
/kernel/linux/linux-5.10/include/asm-generic/
H A Dtlb.h92 * - mmu_gather::fullmm
267 unsigned int fullmm : 1; member
319 if (tlb->fullmm) { in __tlb_reset_range()
378 if (tlb->fullmm || tlb->need_flush_all) { in tlb_flush()
459 if (!tlb->fullmm && !tlb->need_flush_all) in tlb_change_page_size()
494 if (tlb->fullmm) in tlb_start_vma()
505 if (tlb->fullmm) in tlb_end_vma()
/kernel/linux/linux-6.6/include/asm-generic/
H A Dtlb.h94 * - mmu_gather::fullmm
307 unsigned int fullmm : 1; member
365 if (tlb->fullmm) { in __tlb_reset_range()
413 if (tlb->fullmm || tlb->need_flush_all) { in tlb_flush()
499 if (!tlb->fullmm && !tlb->need_flush_all) in tlb_change_page_size()
533 if (tlb->fullmm) in tlb_start_vma()
544 if (tlb->fullmm) in tlb_end_vma()
/kernel/linux/linux-6.6/mm/
H A Dmmu_gather.c304 bool fullmm) in __tlb_gather_mmu()
307 tlb->fullmm = fullmm; in __tlb_gather_mmu()
380 * The aarch64 yields better performance with fullmm by in tlb_finish_mmu()
384 * On x86 non-fullmm doesn't yield significant difference in tlb_finish_mmu()
385 * against fullmm. in tlb_finish_mmu()
387 tlb->fullmm = 1; in tlb_finish_mmu()
303 __tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, bool fullmm) __tlb_gather_mmu() argument
H A Dmadvise.c501 tlb->fullmm); in madvise_cold_or_pageout_pte_range()
666 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range()
669 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range()
739 tlb->fullmm); in madvise_free_pte_range()
H A Dhuge_memory.c1677 tlb->fullmm); in zap_huge_pmd()
1979 pudp_huge_get_and_clear_full(vma, addr, pud, tlb->fullmm); in zap_huge_pud()
H A Dmemory.c1438 tlb->fullmm); in zap_pte_range()
1522 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range()
/kernel/linux/linux-5.10/arch/loongarch/include/asm/
H A Dtlb.h141 if (!(tlb)->fullmm) \
158 if (tlb->fullmm) { in tlb_flush()
/kernel/linux/linux-5.10/arch/sparc/include/asm/
H A Dpgtable_64.h897 pte_t *ptep, pte_t orig, int fullmm,
901 pte_t *ptep, pte_t orig, int fullmm, in maybe_tlb_batch_add()
911 tlb_batch_add(mm, vaddr, ptep, orig, fullmm, hugepage_shift); in maybe_tlb_batch_add()
925 pte_t *ptep, pte_t pte, int fullmm) in __set_pte_at()
930 maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT); in __set_pte_at()
940 #define pte_clear_not_present_full(mm,addr,ptep,fullmm) \
941 __set_pte_at((mm), (addr), (ptep), __pte(0UL), (fullmm))
900 maybe_tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) maybe_tlb_batch_add() argument
924 __set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, int fullmm) __set_pte_at() argument
/kernel/linux/linux-6.6/arch/sparc/include/asm/
H A Dpgtable_64.h895 pte_t *ptep, pte_t orig, int fullmm,
899 pte_t *ptep, pte_t orig, int fullmm, in maybe_tlb_batch_add()
909 tlb_batch_add(mm, vaddr, ptep, orig, fullmm, hugepage_shift); in maybe_tlb_batch_add()
923 pte_t *ptep, pte_t pte, int fullmm) in __set_pte_at()
928 maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT); in __set_pte_at()
951 #define pte_clear_not_present_full(mm,addr,ptep,fullmm) \
952 __set_pte_at((mm), (addr), (ptep), __pte(0UL), (fullmm))
898 maybe_tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) maybe_tlb_batch_add() argument
922 __set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte, int fullmm) __set_pte_at() argument
/kernel/linux/linux-6.6/arch/riscv/include/asm/
H A Dtlb.h19 if (tlb->fullmm || tlb->need_flush_all || tlb->freed_tables) in tlb_flush()
/kernel/linux/linux-5.10/arch/x86/include/asm/
H A Dtlb.h19 if (!tlb->fullmm && !tlb->need_flush_all) { in tlb_flush()
/kernel/linux/linux-6.6/arch/x86/include/asm/
H A Dtlb.h15 if (!tlb->fullmm && !tlb->need_flush_all) { in tlb_flush()
/kernel/linux/linux-5.10/mm/
H A Dmmu_gather.c268 tlb->fullmm = !(start | (end+1)); in tlb_gather_mmu()
314 * The aarch64 yields better performance with fullmm by in tlb_finish_mmu()
318 * On x86 non-fullmm doesn't yield significant difference in tlb_finish_mmu()
319 * against fullmm. in tlb_finish_mmu()
321 tlb->fullmm = 1; in tlb_finish_mmu()
H A Dmadvise.c454 tlb->fullmm); in madvise_cold_or_pageout_pte_range()
620 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in madvise_free_pte_range()
687 tlb->fullmm); in madvise_free_pte_range()
H A Dhuge_memory.c1668 tlb->fullmm); in zap_huge_pmd()
1942 pudp_huge_get_and_clear_full(tlb->mm, addr, pud, tlb->fullmm); in zap_huge_pud()
H A Dmemory.c1276 tlb->fullmm); in zap_pte_range()
1318 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range()
1341 pte_clear_not_present_full(mm, addr, pte, tlb->fullmm); in zap_pte_range()
/kernel/linux/linux-5.10/arch/arm64/include/asm/
H A Dtlb.h65 if (tlb->fullmm) { in tlb_flush()
/kernel/linux/linux-6.6/arch/arm64/include/asm/
H A Dtlb.h65 if (tlb->fullmm) { in tlb_flush()
/kernel/linux/linux-5.10/arch/sparc/mm/
H A Dtlb.c112 pte_t *ptep, pte_t orig, int fullmm, in tlb_batch_add()
139 if (!fullmm) in tlb_batch_add()
111 tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) tlb_batch_add() argument
/kernel/linux/linux-6.6/arch/sparc/mm/
H A Dtlb.c113 pte_t *ptep, pte_t orig, int fullmm, in tlb_batch_add()
142 if (!fullmm) in tlb_batch_add()
112 tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig, int fullmm, unsigned int hugepage_shift) tlb_batch_add() argument
/kernel/linux/linux-6.6/arch/powerpc/mm/book3s64/
H A Dradix_tlb.c609 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument
765 static enum tlb_flush_type flush_type_needed(struct mm_struct *mm, bool fullmm) in flush_type_needed() argument
792 * In the fullmm case there's no point doing the exit_flush_lazy_tlbs in flush_type_needed()
796 if (fullmm) in flush_type_needed()
875 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument
886 type = flush_type_needed(mm, fullmm); in __flush_all_mm()
1195 * A "fullmm" flush must always do a flush_all_mm (RIC=2) flush in radix__tlb_flush()
1199 if (tlb->fullmm) { in radix__tlb_flush()
/kernel/linux/linux-6.6/arch/loongarch/include/asm/
H A Dtlb.h148 if (tlb->fullmm) { in tlb_flush()
/kernel/linux/linux-5.10/arch/powerpc/mm/book3s64/
H A Dradix_tlb.c600 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument
738 static void __flush_all_mm(struct mm_struct *mm, bool fullmm) in __flush_all_mm() argument
750 if (!fullmm) { in __flush_all_mm()
1057 * A "fullmm" flush must always do a flush_all_mm (RIC=2) flush in radix__tlb_flush()
1061 if (tlb->fullmm || tlb->need_flush_all) { in radix__tlb_flush()

Completed in 39 milliseconds