Home
last modified time | relevance | path

Searched refs:orig_pmd (Results 1 - 9 of 9) sorted by relevance

/kernel/linux/linux-5.10/mm/
H A Dhuge_memory.c1250 void huge_pmd_set_accessed(struct vm_fault *vmf, pmd_t orig_pmd) in huge_pmd_set_accessed() argument
1257 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) in huge_pmd_set_accessed()
1260 entry = pmd_mkyoung(orig_pmd); in huge_pmd_set_accessed()
1271 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf, pmd_t orig_pmd) in do_huge_pmd_wp_page() argument
1280 if (is_huge_zero_pmd(orig_pmd)) in do_huge_pmd_wp_page()
1285 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page()
1290 page = pmd_page(orig_pmd); in do_huge_pmd_wp_page()
1299 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page()
1314 entry = pmd_mkyoung(orig_pmd); in do_huge_pmd_wp_page()
1575 pmd_t orig_pmd; in madvise_free_huge_pmd() local
1653 pmd_t orig_pmd; zap_huge_pmd() local
[all...]
H A Dmadvise.c331 pmd_t orig_pmd; in madvise_cold_or_pageout_pte_range() local
339 orig_pmd = *pmd; in madvise_cold_or_pageout_pte_range()
340 if (is_huge_zero_pmd(orig_pmd)) in madvise_cold_or_pageout_pte_range()
343 if (unlikely(!pmd_present(orig_pmd))) { in madvise_cold_or_pageout_pte_range()
345 !is_pmd_migration_entry(orig_pmd)); in madvise_cold_or_pageout_pte_range()
349 page = pmd_page(orig_pmd); in madvise_cold_or_pageout_pte_range()
369 if (pmd_young(orig_pmd)) { in madvise_cold_or_pageout_pte_range()
371 orig_pmd = pmd_mkold(orig_pmd); in madvise_cold_or_pageout_pte_range()
373 set_pmd_at(mm, addr, pmd, orig_pmd); in madvise_cold_or_pageout_pte_range()
[all...]
H A Dmemory.c4436 static inline vm_fault_t wp_huge_pmd(struct vm_fault *vmf, pmd_t orig_pmd) in wp_huge_pmd() argument
4439 if (userfaultfd_huge_pmd_wp(vmf->vma, orig_pmd)) in wp_huge_pmd()
4441 return do_huge_pmd_wp_page(vmf, orig_pmd); in wp_huge_pmd()
4658 pmd_t orig_pmd = *vmf.pmd; in __handle_mm_fault() local
4661 if (unlikely(is_swap_pmd(orig_pmd))) { in __handle_mm_fault()
4663 !is_pmd_migration_entry(orig_pmd)); in __handle_mm_fault()
4664 if (is_pmd_migration_entry(orig_pmd)) in __handle_mm_fault()
4668 if (pmd_trans_huge(orig_pmd) || pmd_devmap(orig_pmd)) { in __handle_mm_fault()
4669 if (pmd_protnone(orig_pmd) in __handle_mm_fault()
[all...]
/kernel/linux/linux-6.6/mm/
H A Dhuge_memory.c1280 if (unlikely(!pmd_same(*vmf->pmd, vmf->orig_pmd))) in huge_pmd_set_accessed()
1296 pmd_t orig_pmd = vmf->orig_pmd; in do_huge_pmd_wp_page() local
1301 if (is_huge_zero_pmd(orig_pmd)) in do_huge_pmd_wp_page()
1306 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page()
1311 page = pmd_page(orig_pmd); in do_huge_pmd_wp_page()
1324 if (unlikely(!pmd_same(*vmf->pmd, orig_pmd))) { in do_huge_pmd_wp_page()
1359 entry = pmd_mkyoung(orig_pmd); in do_huge_pmd_wp_page()
1491 pmd_t oldpmd = vmf->orig_pmd; in do_huge_pmd_numa_page()
1584 pmd_t orig_pmd; in madvise_free_huge_pmd() local
1662 pmd_t orig_pmd; zap_huge_pmd() local
[all...]
H A Dmadvise.c365 pmd_t orig_pmd; in madvise_cold_or_pageout_pte_range() local
373 orig_pmd = *pmd; in madvise_cold_or_pageout_pte_range()
374 if (is_huge_zero_pmd(orig_pmd)) in madvise_cold_or_pageout_pte_range()
377 if (unlikely(!pmd_present(orig_pmd))) { in madvise_cold_or_pageout_pte_range()
379 !is_pmd_migration_entry(orig_pmd)); in madvise_cold_or_pageout_pte_range()
383 folio = pfn_folio(pmd_pfn(orig_pmd)); in madvise_cold_or_pageout_pte_range()
406 if (pmd_young(orig_pmd)) { in madvise_cold_or_pageout_pte_range()
408 orig_pmd = pmd_mkold(orig_pmd); in madvise_cold_or_pageout_pte_range()
410 set_pmd_at(mm, addr, pmd, orig_pmd); in madvise_cold_or_pageout_pte_range()
[all...]
H A Dkhugepaged.c737 pmd_t orig_pmd, in __collapse_huge_page_copy_failed()
750 pmd_populate(vma->vm_mm, pmd, pmd_pgtable(orig_pmd)); in __collapse_huge_page_copy_failed()
768 * @orig_pmd: the original raw pages' PMD
777 pmd_t orig_pmd, in __collapse_huge_page_copy()
810 __collapse_huge_page_copy_failed(pte, pmd, orig_pmd, vma, in __collapse_huge_page_copy()
735 __collapse_huge_page_copy_failed(pte_t *pte, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, struct list_head *compound_pagelist) __collapse_huge_page_copy_failed() argument
774 __collapse_huge_page_copy(pte_t *pte, struct page *page, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) __collapse_huge_page_copy() argument
H A Dmemory.c4931 userfaultfd_huge_pmd_wp(vma, vmf->orig_pmd)) in wp_huge_pmd()
5153 vmf.orig_pmd = pmdp_get_lockless(vmf.pmd); in __handle_mm_fault()
5155 if (unlikely(is_swap_pmd(vmf.orig_pmd))) { in __handle_mm_fault()
5157 !is_pmd_migration_entry(vmf.orig_pmd)); in __handle_mm_fault()
5158 if (is_pmd_migration_entry(vmf.orig_pmd)) in __handle_mm_fault()
5162 if (pmd_trans_huge(vmf.orig_pmd) || pmd_devmap(vmf.orig_pmd)) { in __handle_mm_fault()
5163 if (pmd_protnone(vmf.orig_pmd) && vma_is_accessible(vma)) in __handle_mm_fault()
5167 !pmd_write(vmf.orig_pmd)) { in __handle_mm_fault()
/kernel/linux/linux-5.10/include/linux/
H A Dhuge_mm.h14 void huge_pmd_set_accessed(struct vm_fault *vmf, pmd_t orig_pmd);
27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf, pmd_t orig_pmd);
294 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf, pmd_t orig_pmd);
448 pmd_t orig_pmd) in do_huge_pmd_numa_page()
447 do_huge_pmd_numa_page(struct vm_fault *vmf, pmd_t orig_pmd) do_huge_pmd_numa_page() argument
/kernel/linux/linux-6.6/include/linux/
H A Dmm.h551 pmd_t orig_pmd; /* Value of PMD at the time of fault, member

Completed in 39 milliseconds