Lines Matching defs:gfn
4918 * Take the srcu lock as memslots will be accessed to check the gfn
10419 unsigned long gfn = gpa_to_gfn(APIC_DEFAULT_PHYS_BASE);
10422 kvm_zap_gfn_range(kvm, gfn, gfn+1);
12603 * If the gfn and userspace address are not aligned wrt each
12982 static inline u32 kvm_async_pf_hash_fn(gfn_t gfn)
12986 return hash_32(gfn & 0xffffffff, order_base_2(ASYNC_PF_PER_VCPU));
12994 static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
12996 u32 key = kvm_async_pf_hash_fn(gfn);
13001 vcpu->arch.apf.gfns[key] = gfn;
13004 static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn)
13007 u32 key = kvm_async_pf_hash_fn(gfn);
13010 (vcpu->arch.apf.gfns[key] != gfn &&
13017 bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
13019 return vcpu->arch.apf.gfns[kvm_async_pf_gfn_slot(vcpu, gfn)] == gfn;
13022 static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
13026 i = j = kvm_async_pf_gfn_slot(vcpu, gfn);
13028 if (WARN_ON_ONCE(vcpu->arch.apf.gfns[i] != gfn))
13126 kvm_add_async_pf_gfn(vcpu, work->arch.gfn);
13163 kvm_del_async_pf_gfn(vcpu, work->arch.gfn);