/kernel/linux/linux-5.10/mm/ |
H A D | huge_memory.c | 1250 void huge_pmd_set_accessed(struct vm_fault *vmf, pmd_t orig_pmd) in huge_pmd_set_accessed() argument 1257 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) in huge_pmd_set_accessed() 1260 entry = pmd_mkyoung(orig_pmd); in huge_pmd_set_accessed() 1271 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf, pmd_t orig_pmd) in do_huge_pmd_wp_page() argument 1280 if (is_huge_zero_pmd(orig_pmd)) in do_huge_pmd_wp_page() 1285 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page() 1290 page = pmd_page(orig_pmd); in do_huge_pmd_wp_page() 1299 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page() 1314 entry = pmd_mkyoung(orig_pmd); in do_huge_pmd_wp_page() 1575 pmd_t orig_pmd; in madvise_free_huge_pmd() local 1653 pmd_t orig_pmd; zap_huge_pmd() local [all...] |
H A D | madvise.c | 331 pmd_t orig_pmd; in madvise_cold_or_pageout_pte_range() local 339 orig_pmd = *pmd; in madvise_cold_or_pageout_pte_range() 340 if (is_huge_zero_pmd(orig_pmd)) in madvise_cold_or_pageout_pte_range() 343 if (unlikely(!pmd_present(orig_pmd))) { in madvise_cold_or_pageout_pte_range() 345 !is_pmd_migration_entry(orig_pmd)); in madvise_cold_or_pageout_pte_range() 349 page = pmd_page(orig_pmd); in madvise_cold_or_pageout_pte_range() 369 if (pmd_young(orig_pmd)) { in madvise_cold_or_pageout_pte_range() 371 orig_pmd = pmd_mkold(orig_pmd); in madvise_cold_or_pageout_pte_range() 373 set_pmd_at(mm, addr, pmd, orig_pmd); in madvise_cold_or_pageout_pte_range() [all...] |
H A D | memory.c | 4436 static inline vm_fault_t wp_huge_pmd(struct vm_fault *vmf, pmd_t orig_pmd) in wp_huge_pmd() argument 4439 if (userfaultfd_huge_pmd_wp(vmf->vma, orig_pmd)) in wp_huge_pmd() 4441 return do_huge_pmd_wp_page(vmf, orig_pmd); in wp_huge_pmd() 4658 pmd_t orig_pmd = *vmf.pmd; in __handle_mm_fault() local 4661 if (unlikely(is_swap_pmd(orig_pmd))) { in __handle_mm_fault() 4663 !is_pmd_migration_entry(orig_pmd)); in __handle_mm_fault() 4664 if (is_pmd_migration_entry(orig_pmd)) in __handle_mm_fault() 4668 if (pmd_trans_huge(orig_pmd) || pmd_devmap(orig_pmd)) { in __handle_mm_fault() 4669 if (pmd_protnone(orig_pmd) in __handle_mm_fault() [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | huge_memory.c | 1280 if (unlikely(!pmd_same(*vmf->pmd, vmf->orig_pmd))) in huge_pmd_set_accessed() 1296 pmd_t orig_pmd = vmf->orig_pmd; in do_huge_pmd_wp_page() local 1301 if (is_huge_zero_pmd(orig_pmd)) in do_huge_pmd_wp_page() 1306 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page() 1311 page = pmd_page(orig_pmd); in do_huge_pmd_wp_page() 1324 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page() 1359 entry = pmd_mkyoung(orig_pmd); in do_huge_pmd_wp_page() 1491 pmd_t oldpmd = vmf->orig_pmd; in do_huge_pmd_numa_page() 1584 pmd_t orig_pmd; in madvise_free_huge_pmd() local 1662 pmd_t orig_pmd; zap_huge_pmd() local [all...] |
H A D | madvise.c | 365 pmd_t orig_pmd; in madvise_cold_or_pageout_pte_range() local 373 orig_pmd = *pmd; in madvise_cold_or_pageout_pte_range() 374 if (is_huge_zero_pmd(orig_pmd)) in madvise_cold_or_pageout_pte_range() 377 if (unlikely(!pmd_present(orig_pmd))) { in madvise_cold_or_pageout_pte_range() 379 !is_pmd_migration_entry(orig_pmd)); in madvise_cold_or_pageout_pte_range() 383 folio = pfn_folio(pmd_pfn(orig_pmd)); in madvise_cold_or_pageout_pte_range() 406 if (pmd_young(orig_pmd)) { in madvise_cold_or_pageout_pte_range() 408 orig_pmd = pmd_mkold(orig_pmd); in madvise_cold_or_pageout_pte_range() 410 set_pmd_at(mm, addr, pmd, orig_pmd); in madvise_cold_or_pageout_pte_range() [all...] |
H A D | khugepaged.c | 737 pmd_t orig_pmd, in __collapse_huge_page_copy_failed() 750 pmd_populate(vma->vm_mm, pmd, pmd_pgtable(orig_pmd)); in __collapse_huge_page_copy_failed() 768 * @orig_pmd: the original raw pages' PMD 777 pmd_t orig_pmd, in __collapse_huge_page_copy() 810 __collapse_huge_page_copy_failed(pte, pmd, orig_pmd, vma, in __collapse_huge_page_copy() 735 __collapse_huge_page_copy_failed(pte_t *pte, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, struct list_head *compound_pagelist) __collapse_huge_page_copy_failed() argument 774 __collapse_huge_page_copy(pte_t *pte, struct page *page, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) __collapse_huge_page_copy() argument
|
H A D | memory.c | 4931 userfaultfd_huge_pmd_wp(vma, vmf->orig_pmd)) in wp_huge_pmd() 5153 vmf.orig_pmd = pmdp_get_lockless(vmf.pmd); in __handle_mm_fault() 5155 if (unlikely(is_swap_pmd(vmf.orig_pmd))) { in __handle_mm_fault() 5157 !is_pmd_migration_entry(vmf.orig_pmd)); in __handle_mm_fault() 5158 if (is_pmd_migration_entry(vmf.orig_pmd)) in __handle_mm_fault() 5162 if (pmd_trans_huge(vmf.orig_pmd) || pmd_devmap(vmf.orig_pmd)) { in __handle_mm_fault() 5163 if (pmd_protnone(vmf.orig_pmd) && vma_is_accessible(vma)) in __handle_mm_fault() 5167 !pmd_write(vmf.orig_pmd)) { in __handle_mm_fault()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | huge_mm.h | 14 void huge_pmd_set_accessed(struct vm_fault *vmf, pmd_t orig_pmd); 27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf, pmd_t orig_pmd); 294 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf, pmd_t orig_pmd); 448 pmd_t orig_pmd) in do_huge_pmd_numa_page() 447 do_huge_pmd_numa_page(struct vm_fault *vmf, pmd_t orig_pmd) do_huge_pmd_numa_page() argument
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | mm.h | 551 pmd_t orig_pmd; /* Value of PMD at the time of fault, member
|