Lines Matching defs:address
659 vcpu->arch.apf.nested_apf_token = fault->address;
663 fault->address);
678 * Invalidate the TLB entry for the faulting address, if it exists,
683 kvm_mmu_invalidate_gva(vcpu, fault_mmu, fault->address,
1636 * non-canonical address is written on Intel but not on
1641 * value on AMD. Making the address canonical ensures that
6746 static void emulator_invlpg(struct x86_emulate_ctxt *ctxt, ulong address)
6748 kvm_mmu_invlpg(emul_to_vcpu(ctxt), address);
7370 * last retried eip and the last fault address, if we meet the eip
7371 * and the address again, we can break out of the potential infinite
7656 /* Save the faulting GPA (cr2) in the address field */
7657 ctxt->exception.address = cr2_or_gpa;
7665 /* Sanitize the address out of an abundance of paranoia. */
7666 ctxt->exception.address = 0;
8761 put_smstate(u32, buf, 0x7f74, dt.address);
8765 put_smstate(u32, buf, 0x7f58, dt.address);
8816 put_smstate(u64, buf, 0x7e88, dt.address);
8826 put_smstate(u64, buf, 0x7e68, dt.address);
8874 dt.address = dt.size = 0;
9036 * The physical address of apic access page is stored in the VMCS.
9306 * care about the messed up debug address registers. But if
9781 sregs->idt.base = dt.address;
9784 sregs->gdt.base = dt.address;
9930 dt.address = sregs->idt.base;
9933 dt.address = sregs->gdt.base;
10069 * Translate a guest virtual address to a guest physical address.
10851 * If the gfn and userspace address are not aligned wrt each
11304 fault.address = work->arch.token;
11506 fault.address = gva;