Lines Matching refs:fault_ipa
1394 static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
1488 fault_ipa &= ~(vma_pagesize - 1);
1490 gfn = fault_ipa >> PAGE_SHIFT;
1555 &fault_ipa);
1590 ret = kvm_pgtable_stage2_relax_perms(pgt, fault_ipa, prot);
1592 ret = kvm_pgtable_stage2_map(pgt, fault_ipa, vma_pagesize,
1611 static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa)
1616 trace_kvm_access_fault(fault_ipa);
1620 pte = kvm_pgtable_stage2_mkyoung(mmu->pgt, fault_ipa);
1641 phys_addr_t fault_ipa;
1650 fault_ipa = kvm_vcpu_get_fault_ipa(vcpu);
1655 if (fault_ipa >= BIT_ULL(get_kvm_ipa_limit())) {
1661 if (fault_ipa >= BIT_ULL(vcpu->arch.hw_mmu->pgt->ia_bits)) {
1662 fault_ipa |= kvm_vcpu_get_hfar(vcpu) & GENMASK(11, 0);
1665 kvm_inject_pabt(vcpu, fault_ipa);
1667 kvm_inject_dabt(vcpu, fault_ipa);
1678 if (kvm_handle_guest_sea(fault_ipa, kvm_vcpu_get_esr(vcpu)))
1685 kvm_vcpu_get_hfar(vcpu), fault_ipa);
1700 gfn = fault_ipa >> PAGE_SHIFT;
1744 fault_ipa |= kvm_vcpu_get_hfar(vcpu) & ((1 << 12) - 1);
1745 ret = io_mem_abort(vcpu, fault_ipa);
1750 VM_BUG_ON(fault_ipa >= kvm_phys_size(vcpu->kvm));
1753 handle_access_fault(vcpu, fault_ipa);
1758 ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status);