/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | mmu.h | 103 bool prefault); 106 u32 err, bool prefault) in kvm_mmu_do_page_fault() 110 return kvm_tdp_page_fault(vcpu, cr2_or_gpa, err, prefault); in kvm_mmu_do_page_fault() 112 return vcpu->arch.mmu->page_fault(vcpu, cr2_or_gpa, err, prefault); in kvm_mmu_do_page_fault() 105 kvm_mmu_do_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u32 err, bool prefault) kvm_mmu_do_page_fault() argument
|
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
H A D | tdp_mmu.c | 527 kvm_pfn_t pfn, bool prefault) in tdp_mmu_map_handle_target_level() 538 pfn, iter->old_spte, prefault, true, in tdp_mmu_map_handle_target_level() 563 if (!prefault) in tdp_mmu_map_handle_target_level() 575 bool prefault) in kvm_tdp_mmu_map() 648 pfn, prefault); in kvm_tdp_mmu_map() 524 tdp_mmu_map_handle_target_level(struct kvm_vcpu *vcpu, int write, int map_writable, struct tdp_iter *iter, kvm_pfn_t pfn, bool prefault) tdp_mmu_map_handle_target_level() argument 573 kvm_tdp_mmu_map(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, int map_writable, int max_level, kvm_pfn_t pfn, bool prefault) kvm_tdp_mmu_map() argument
|
H A D | tdp_mmu.h | 36 bool prefault);
|
H A D | paging_tmpl.h | 629 bool prefault) in fetch() 719 it.level, base_gfn, pfn, prefault, map_writable); in fetch() 786 bool prefault) in page_fault() 815 if (!prefault) in page_fault() 843 if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, in page_fault() 880 map_writable, prefault); in page_fault() 626 fetch(struct kvm_vcpu *vcpu, gpa_t addr, struct guest_walker *gw, u32 error_code, int max_level, kvm_pfn_t pfn, bool map_writable, bool prefault) fetch() argument 785 page_fault(struct kvm_vcpu *vcpu, gpa_t addr, u32 error_code, bool prefault) page_fault() argument
|
H A D | mmu.c | 2852 bool prefault, bool is_tdp) in __direct_map() 2897 write, level, base_gfn, pfn, prefault, in __direct_map() 3661 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, in try_async_pf() argument 3688 if (!prefault && kvm_can_do_async_pf(vcpu)) { in try_async_pf() 3703 bool prefault, int max_level, bool is_tdp) in direct_page_fault() 3729 if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) in direct_page_fault() 3745 pfn, prefault); in direct_page_fault() 3748 prefault, is_tdp); in direct_page_fault() 3757 u32 error_code, bool prefault) in nonpaging_page_fault() 3762 return direct_page_fault(vcpu, gpa & PAGE_MASK, error_code, prefault, in nonpaging_page_fault() 2850 __direct_map(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, int map_writable, int max_level, kvm_pfn_t pfn, bool prefault, bool is_tdp) __direct_map() argument 3702 direct_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, bool prefault, int max_level, bool is_tdp) direct_page_fault() argument 3756 nonpaging_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, bool prefault) nonpaging_page_fault() argument 3799 kvm_tdp_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u32 error_code, bool prefault) kvm_tdp_page_fault() argument [all...] |
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | kvm_host.h | 352 bool prefault);
|
/kernel/linux/linux-6.6/mm/ |
H A D | memory.c | 4392 bool prefault = in_range(vmf->address, addr, nr * PAGE_SIZE); in set_pte_range() local 4398 if (prefault && arch_wants_old_prefaulted_pte()) in set_pte_range()
|