/kernel/linux/linux-6.6/arch/mips/kvm/ |
H A D | mmu.c | 496 * @write_fault: Whether the fault was due to a write. 511 bool write_fault, in _kvm_mips_map_page_fast() 537 if (write_fault && !pte_dirty(*ptep)) { in _kvm_mips_map_page_fast() 566 * @write_fault: Whether the fault was due to a write. 587 bool write_fault, in kvm_mips_map_page() 602 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() 632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); in kvm_mips_map_page() 658 if (write_fault) { in kvm_mips_map_page() 685 bool write_fault) in kvm_mips_handle_vz_root_tlb_fault() 689 ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NUL in kvm_mips_handle_vz_root_tlb_fault() 510 _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, bool write_fault, pte_t *out_entry, pte_t *out_buddy) _kvm_mips_map_page_fast() argument 586 kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, bool write_fault, pte_t *out_entry, pte_t *out_buddy) kvm_mips_map_page() argument 683 kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr, struct kvm_vcpu *vcpu, bool write_fault) kvm_mips_handle_vz_root_tlb_fault() argument [all...] |
/kernel/linux/linux-5.10/arch/mips/kvm/ |
H A D | mmu.c | 571 * @write_fault: Whether the fault was due to a write. 586 bool write_fault, in _kvm_mips_map_page_fast() 612 if (write_fault && !pte_dirty(*ptep)) { in _kvm_mips_map_page_fast() 641 * @write_fault: Whether the fault was due to a write. 662 bool write_fault, in kvm_mips_map_page() 677 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() 707 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); in kvm_mips_map_page() 733 if (write_fault) { in kvm_mips_map_page() 963 bool write_fault) in kvm_mips_handle_vz_root_tlb_fault() 967 ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NUL in kvm_mips_handle_vz_root_tlb_fault() 585 _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, bool write_fault, pte_t *out_entry, pte_t *out_buddy) _kvm_mips_map_page_fast() argument 661 kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, bool write_fault, pte_t *out_entry, pte_t *out_buddy) kvm_mips_map_page() argument 961 kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr, struct kvm_vcpu *vcpu, bool write_fault) kvm_mips_handle_vz_root_tlb_fault() argument 977 kvm_mips_handle_kseg0_tlb_fault(unsigned long badvaddr, struct kvm_vcpu *vcpu, bool write_fault) kvm_mips_handle_kseg0_tlb_fault() argument 1014 kvm_mips_handle_mapped_seg_tlb_fault(struct kvm_vcpu *vcpu, struct kvm_mips_tlb *tlb, unsigned long gva, bool write_fault) kvm_mips_handle_mapped_seg_tlb_fault() argument [all...] |
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
H A D | paging_tmpl.h | 235 gpa_t addr, int write_fault) in update_accessed_dirty_bits() 256 if (level == walker->level && write_fault && in update_accessed_dirty_bits() 323 const int write_fault = access & PFERR_WRITE_MASK; in walk_addr_generic() local 444 if (!write_fault) in walk_addr_generic() 457 addr, write_fault); in walk_addr_generic() 470 errcode |= write_fault | user_fault; in walk_addr_generic() 494 if (write_fault) in walk_addr_generic() 632 bool write_fault = error_code & PFERR_WRITE_MASK; in fetch() local 718 ret = mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault, in fetch() 788 bool write_fault in page_fault() local 232 update_accessed_dirty_bits(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, struct guest_walker *walker, gpa_t addr, int write_fault) update_accessed_dirty_bits() argument [all...] |
/kernel/linux/linux-5.10/arch/arm64/kvm/ |
H A D | mmu.c | 746 bool write_fault, writable, force_pte = false; in user_mem_abort() local 763 write_fault = kvm_is_write_fault(vcpu); in user_mem_abort() 765 VM_BUG_ON(write_fault && exec_fault); in user_mem_abort() 767 if (fault_status == FSC_PERM && !write_fault && !exec_fault) { in user_mem_abort() 829 if (fault_status != FSC_PERM || (logging_active && write_fault)) { in user_mem_abort() 848 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writable); in user_mem_abort() 859 } else if (logging_active && !write_fault) { in user_mem_abort() 957 bool is_iabt, write_fault, writable; in kvm_handle_guest_abort() local 996 write_fault = kvm_is_write_fault(vcpu); in kvm_handle_guest_abort() 997 if (kvm_is_error_hva(hva) || (write_fault in kvm_handle_guest_abort() [all...] |
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_file.c | 1328 bool write_fault, in xfs_dax_fault() 1332 (write_fault && !vmf->cow_page) ? in xfs_dax_fault() 1341 bool write_fault, in xfs_dax_fault() 1363 bool write_fault) in __xfs_filemap_fault() 1370 trace_xfs_filemap_fault(ip, order, write_fault); in __xfs_filemap_fault() 1372 if (write_fault) { in __xfs_filemap_fault() 1377 if (IS_DAX(inode) || write_fault) in __xfs_filemap_fault() 1383 ret = xfs_dax_fault(vmf, order, write_fault, &pfn); in __xfs_filemap_fault() 1386 } else if (write_fault) { in __xfs_filemap_fault() 1395 if (write_fault) in __xfs_filemap_fault() 1325 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) xfs_dax_fault() argument 1338 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) xfs_dax_fault() argument 1360 __xfs_filemap_fault( struct vm_fault *vmf, unsigned int order, bool write_fault) __xfs_filemap_fault() argument [all...] |
/kernel/linux/linux-6.6/arch/x86/kvm/mmu/ |
H A D | paging_tmpl.h | 200 gpa_t addr, int write_fault) in update_accessed_dirty_bits() 221 if (level == walker->level && write_fault && in update_accessed_dirty_bits() 317 const int write_fault = access & PFERR_WRITE_MASK; in walk_addr_generic() local 454 if (!write_fault) in walk_addr_generic() 467 addr, write_fault); in walk_addr_generic() 477 errcode |= write_fault | user_fault; in walk_addr_generic() 502 if (write_fault) in walk_addr_generic() 197 update_accessed_dirty_bits(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, struct guest_walker *walker, gpa_t addr, int write_fault) update_accessed_dirty_bits() argument
|
/kernel/linux/linux-6.6/virt/kvm/ |
H A D | kvm_mm.h | 24 bool *async, bool write_fault, bool *writable);
|
H A D | kvm_main.c | 2515 static bool hva_to_pfn_fast(unsigned long addr, bool write_fault, in hva_to_pfn_fast() argument 2525 if (!(write_fault || writable)) in hva_to_pfn_fast() 2543 static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, in hva_to_pfn_slow() argument 2564 *writable = write_fault; in hva_to_pfn_slow() 2566 if (write_fault) in hva_to_pfn_slow() 2578 if (unlikely(!write_fault) && writable) { in hva_to_pfn_slow() 2591 static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) in vma_is_valid() argument 2596 if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) in vma_is_valid() 2613 unsigned long addr, bool write_fault, in hva_to_pfn_remapped() 2630 (write_fault in hva_to_pfn_remapped() 2612 hva_to_pfn_remapped(struct vm_area_struct *vma, unsigned long addr, bool write_fault, bool *writable, kvm_pfn_t *p_pfn) hva_to_pfn_remapped() argument 2695 hva_to_pfn(unsigned long addr, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable) hva_to_pfn() argument 2746 __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable, hva_t *hva) __gfn_to_pfn_memslot() argument 2778 gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, bool *writable) gfn_to_pfn_prot() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.h | 176 bool write_fault); 226 uint64_t addr, bool write_fault) in svm_range_restore_pages() 223 svm_range_restore_pages(struct amdgpu_device *adev, unsigned int pasid, uint32_t client_id, uint32_t node_id, uint64_t addr, bool write_fault) svm_range_restore_pages() argument
|
H A D | kfd_smi_events.h | 33 unsigned long address, bool write_fault,
|
H A D | kfd_smi_events.c | 254 unsigned long address, bool write_fault, in kfd_smi_event_page_fault_start() 259 address, node->id, write_fault ? 'W' : 'R'); in kfd_smi_event_page_fault_start() 253 kfd_smi_event_page_fault_start(struct kfd_node *node, pid_t pid, unsigned long address, bool write_fault, ktime_t ts) kfd_smi_event_page_fault_start() argument
|
H A D | kfd_svm.c | 2874 svm_fault_allowed(struct vm_area_struct *vma, bool write_fault) in svm_fault_allowed() argument 2878 if (write_fault) in svm_fault_allowed() 2889 uint64_t addr, bool write_fault) in svm_range_restore_pages() 3004 if (!svm_fault_allowed(vma, write_fault)) { in svm_range_restore_pages() 3006 write_fault ? "write" : "read"); in svm_range_restore_pages() 3024 write_fault, timestamp); in svm_range_restore_pages() 2887 svm_range_restore_pages(struct amdgpu_device *adev, unsigned int pasid, uint32_t vmid, uint32_t node_id, uint64_t addr, bool write_fault) svm_range_restore_pages() argument
|
/kernel/linux/linux-6.6/arch/arm64/kvm/ |
H A D | mmu.c | 1399 bool write_fault, writable, force_pte = false; in user_mem_abort() local 1416 write_fault = kvm_is_write_fault(vcpu); in user_mem_abort() 1418 VM_BUG_ON(write_fault && exec_fault); in user_mem_abort() 1420 if (fault_status == ESR_ELx_FSC_PERM && !write_fault && !exec_fault) { in user_mem_abort() 1432 (logging_active && write_fault)) { in user_mem_abort() 1508 write_fault, &writable, NULL); in user_mem_abort() 1528 } else if (logging_active && !write_fault) { in user_mem_abort() 1644 bool is_iabt, write_fault, writable; in kvm_handle_guest_abort() local 1703 write_fault = kvm_is_write_fault(vcpu); in kvm_handle_guest_abort() 1704 if (kvm_is_error_hva(hva) || (write_fault in kvm_handle_guest_abort() [all...] |
/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_file.c | 1251 bool write_fault) in __xfs_filemap_fault() 1257 trace_xfs_filemap_fault(ip, pe_size, write_fault); in __xfs_filemap_fault() 1259 if (write_fault) { in __xfs_filemap_fault() 1269 (write_fault && !vmf->cow_page) ? in __xfs_filemap_fault() 1275 if (write_fault) in __xfs_filemap_fault() 1283 if (write_fault) in __xfs_filemap_fault() 1248 __xfs_filemap_fault( struct vm_fault *vmf, enum page_entry_size pe_size, bool write_fault) __xfs_filemap_fault() argument
|
H A D | xfs_trace.h | 667 bool write_fault), 668 TP_ARGS(ip, pe_size, write_fault), 673 __field(bool, write_fault) 679 __entry->write_fault = write_fault; 681 TP_printk("dev %d:%d ino 0x%llx %s write_fault %d", 688 __entry->write_fault)
|
/kernel/linux/linux-5.10/virt/kvm/ |
H A D | kvm_main.c | 1895 static bool hva_to_pfn_fast(unsigned long addr, bool write_fault, in hva_to_pfn_fast() argument 1905 if (!(write_fault || writable)) in hva_to_pfn_fast() 1923 static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, in hva_to_pfn_slow() argument 1933 *writable = write_fault; in hva_to_pfn_slow() 1935 if (write_fault) in hva_to_pfn_slow() 1945 if (unlikely(!write_fault) && writable) { in hva_to_pfn_slow() 1958 static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) in vma_is_valid() argument 1963 if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) in vma_is_valid() 1978 bool write_fault, bool *writable, in hva_to_pfn_remapped() 1994 (write_fault in hva_to_pfn_remapped() 1976 hva_to_pfn_remapped(struct vm_area_struct *vma, unsigned long addr, bool *async, bool write_fault, bool *writable, kvm_pfn_t *p_pfn) hva_to_pfn_remapped() argument 2056 hva_to_pfn(unsigned long addr, bool atomic, bool *async, bool write_fault, bool *writable) hva_to_pfn() argument 2104 __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool *async, bool write_fault, bool *writable) __gfn_to_pfn_memslot() argument 2133 gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, bool *writable) gfn_to_pfn_prot() argument [all...] |
/kernel/linux/linux-5.10/arch/mips/include/asm/ |
H A D | kvm_host.h | 886 struct kvm_vcpu *vcpu, bool write_fault); 890 bool write_fault); 898 bool write_fault); 903 bool write_fault);
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | gmc_v10_0.c | 109 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v10_0_process_interrupt() local 137 if (amdgpu_vm_handle_fault(adev, entry->pasid, 0, 0, addr, write_fault)) in gmc_v10_0_process_interrupt()
|
H A D | amdgpu_vm.h | 473 bool write_fault);
|
H A D | gmc_v9_0.c | 551 bool write_fault = !!(entry->src_data[1] & 0x20); in gmc_v9_0_process_interrupt() local 597 addr, write_fault); in gmc_v9_0_process_interrupt() 620 addr, write_fault)) in gmc_v9_0_process_interrupt()
|
/kernel/linux/linux-6.6/tools/testing/selftests/user_events/ |
H A D | ftrace_test.c | 389 TEST_F(user, write_fault) { in TEST_F()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | kvm_host.h | 745 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, 750 bool atomic, bool *async, bool write_fault,
|
/kernel/linux/linux-5.10/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 465 bool write_fault) in follow_fault_pfn() 477 (write_fault ? FAULT_FLAG_WRITE : 0), in follow_fault_pfn() 490 if (write_fault && !pte_write(*ptep)) in follow_fault_pfn() 463 follow_fault_pfn(struct vm_area_struct *vma, struct mm_struct *mm, unsigned long vaddr, unsigned long *pfn, bool write_fault) follow_fault_pfn() argument
|
/kernel/linux/linux-6.6/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 514 bool write_fault) in follow_fault_pfn() 527 (write_fault ? FAULT_FLAG_WRITE : 0), in follow_fault_pfn() 542 if (write_fault && !pte_write(pte)) in follow_fault_pfn() 512 follow_fault_pfn(struct vm_area_struct *vma, struct mm_struct *mm, unsigned long vaddr, unsigned long *pfn, bool write_fault) follow_fault_pfn() argument
|
/kernel/linux/linux-6.6/arch/mips/include/asm/ |
H A D | kvm_host.h | 790 struct kvm_vcpu *vcpu, bool write_fault);
|