Home
last modified time | relevance | path

Searched refs:pte_access (Results 1 - 10 of 10) sorted by relevance

/kernel/linux/linux-6.6/arch/x86/kvm/mmu/
H A Dpaging_tmpl.h90 unsigned int pte_access; member
310 u64 pt_access, pte_access; in walk_addr_generic() local
349 pte_access = ~0; in walk_addr_generic()
367 pt_access = pte_access; in walk_addr_generic()
415 pte_access = pt_access & (pte ^ walk_nx_mask); in walk_addr_generic()
432 accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0; in walk_addr_generic()
435 walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask); in walk_addr_generic()
436 errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); in walk_addr_generic()
455 FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pt in walk_addr_generic()
537 unsigned pte_access; prefetch_gpte() local
909 unsigned pte_access; sync_spte() local
[all...]
H A Dspte.c139 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, in make_spte()
147 WARN_ON_ONCE(!pte_access && !shadow_present_mask); in make_spte()
176 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte()
178 pte_access &= ~ACC_EXEC_MASK; in make_spte()
181 if (pte_access & ACC_EXEC_MASK) in make_spte()
186 if (pte_access & ACC_USER_MASK) in make_spte()
198 pte_access &= ~ACC_WRITE_MASK; in make_spte()
205 if (pte_access & ACC_WRITE_MASK) { in make_spte()
225 pte_access &= ~ACC_WRITE_MASK; in make_spte()
230 if (pte_access in make_spte()
137 make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, const struct kvm_memory_slot *slot, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool prefetch, bool can_unsync, bool host_writable, u64 *new_spte) make_spte() argument
[all...]
H A Dspte.h476 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
H A Dmmu.c2901 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
2919 mark_mmio_spte(vcpu, sptep, gfn, pte_access); in mmu_set_spte()
2942 wrprot = make_spte(vcpu, sp, slot, pte_access, gfn, pfn, *sptep, prefetch, in mmu_set_spte()
2962 rmap_add(vcpu, slot, sptep, gfn, pte_access); in mmu_set_spte()
2964 /* Already rmapped but the pte_access bits may have changed. */ in mmu_set_spte()
2965 kvm_mmu_page_set_access(sp, spte_index(sptep), pte_access); in mmu_set_spte()
2900 mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, u64 *sptep, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, struct kvm_page_fault *fault) mmu_set_spte() argument
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/
H A Dspte.c84 int make_spte(struct kvm_vcpu *vcpu, unsigned int pte_access, int level, in make_spte() argument
107 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte()
109 pte_access &= ~ACC_EXEC_MASK; in make_spte()
112 if (pte_access & ACC_EXEC_MASK) in make_spte()
117 if (pte_access & ACC_USER_MASK) in make_spte()
129 pte_access &= ~ACC_WRITE_MASK; in make_spte()
136 if (pte_access & ACC_WRITE_MASK) { in make_spte()
152 pte_access &= ~ACC_WRITE_MASK; in make_spte()
157 if (pte_access & ACC_WRITE_MASK) in make_spte()
H A Dpaging_tmpl.h95 unsigned int pte_access; member
316 u64 pt_access, pte_access; in walk_addr_generic() local
356 pte_access = ~0; in walk_addr_generic()
362 pt_access = pte_access; in walk_addr_generic()
407 pte_access = pt_access & (pte ^ walk_nx_mask); in walk_addr_generic()
424 accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0; in walk_addr_generic()
427 walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask); in walk_addr_generic()
428 errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); in walk_addr_generic()
445 FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pt in walk_addr_generic()
531 unsigned pte_access; prefetch_gpte() local
1032 unsigned pte_access; sync_page() local
[all...]
H A Dspte.h246 int make_spte(struct kvm_vcpu *vcpu, unsigned int pte_access, int level,
H A Dmmu.c2561 unsigned int pte_access, int level, in set_spte()
2569 if (set_mmio_spte(vcpu, sptep, gfn, pfn, pte_access)) in set_spte()
2574 ret = make_spte(vcpu, pte_access, level, gfn, pfn, *sptep, speculative, in set_spte()
2588 unsigned int pte_access, bool write_fault, int level, in mmu_set_spte()
2622 set_spte_ret = set_spte(vcpu, sptep, pte_access, level, gfn, pfn, in mmu_set_spte()
2560 set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned int pte_access, int level, gfn_t gfn, kvm_pfn_t pfn, bool speculative, bool can_unsync, bool host_writable) set_spte() argument
2587 mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned int pte_access, bool write_fault, int level, gfn_t gfn, kvm_pfn_t pfn, bool speculative, bool host_writable) mmu_set_spte() argument
/kernel/linux/linux-5.10/arch/x86/kvm/
H A Dmmu.h167 unsigned pte_access, unsigned pte_pkey, in permission_fault()
189 bool fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault()
206 ((pte_access & PT_USER_MASK) << (PFERR_RSVD_BIT - PT_USER_SHIFT)); in permission_fault()
166 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned pte_access, unsigned pte_pkey, unsigned pfec) permission_fault() argument
/kernel/linux/linux-6.6/arch/x86/kvm/
H A Dmmu.h187 unsigned pte_access, unsigned pte_pkey, in permission_fault()
214 fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault()
230 ((pte_access & PT_USER_MASK) << (PFERR_RSVD_BIT - PT_USER_SHIFT)); in permission_fault()
186 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned pte_access, unsigned pte_pkey, u64 access) permission_fault() argument

Completed in 25 milliseconds