Lines Matching defs:addr

156 	u64 addr;
2362 u64 addr)
2364 iterator->addr = addr;
2381 = vcpu->arch.mmu->pae_root[(addr >> 30) & 3];
2390 struct kvm_vcpu *vcpu, u64 addr)
2393 addr);
2401 iterator->index = SPTE_INDEX(iterator->addr, iterator->level);
3227 for_each_shadow_entry(vcpu, fault->addr, it) {
3295 gva_t gva = fault->is_tdp ? 0 : fault->addr;
3441 sptep = kvm_tdp_mmu_fast_pf_get_last_sptep(vcpu, fault->addr, &spte);
3443 sptep = fast_pf_get_last_sptep(vcpu, fault->addr, &spte);
4056 static bool mmio_info_in_cache(struct kvm_vcpu *vcpu, u64 addr, bool direct)
4066 return vcpu_match_mmio_gpa(vcpu, addr);
4068 return vcpu_match_mmio_gva(vcpu, addr);
4077 static int get_walk(struct kvm_vcpu *vcpu, u64 addr, u64 *sptes, int *root_level)
4083 for (shadow_walk_init(&iterator, vcpu, addr),
4097 static bool get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep)
4107 leaf = kvm_tdp_mmu_get_walk(vcpu, addr, sptes, &root);
4109 leaf = get_walk(vcpu, addr, sptes, &root);
4135 pr_err("%s: reserved bits set on MMU-present spte, addr 0x%llx, hierarchy:\n",
4136 __func__, addr);
4146 static int handle_mmio_page_fault(struct kvm_vcpu *vcpu, u64 addr, bool direct)
4151 if (mmio_info_in_cache(vcpu, addr, direct))
4154 reserved = get_mmio_spte(vcpu, addr, &spte);
4166 addr = 0;
4168 trace_handle_mmio_page_fault(addr, gfn, access);
4169 vcpu_cache_mmio_info(vcpu, addr, gfn, access);
4199 static void shadow_page_table_clear_flood(struct kvm_vcpu *vcpu, gva_t addr)
4205 for_each_shadow_entry_lockless(vcpu, addr, iterator, spte)
4294 trace_kvm_try_async_get_page(fault->addr, fault->gfn);
4296 trace_kvm_async_pf_repeated_fault(fault->addr, fault->gfn);
4299 } else if (kvm_arch_setup_async_pf(vcpu, fault->addr, fault->gfn)) {
5769 u64 addr, hpa_t root_hpa)
5773 vcpu_clear_mmio_info(vcpu, addr);
5787 for_each_shadow_entry_using_root(vcpu, root_hpa, addr, iterator) {
5806 u64 addr, unsigned long roots)
5815 if (is_noncanonical_address(addr, vcpu))
5818 static_call(kvm_x86_flush_tlb_gva)(vcpu, addr);
5825 __kvm_mmu_invalidate_addr(vcpu, mmu, addr, mmu->root.hpa);
5829 __kvm_mmu_invalidate_addr(vcpu, mmu, addr, mmu->prev_roots[i].hpa);