/kernel/linux/linux-6.6/arch/x86/kvm/mmu/ |
H A D | paging_tmpl.h | 90 unsigned int pte_access; member 310 u64 pt_access, pte_access; in walk_addr_generic() local 349 pte_access = ~0; in walk_addr_generic() 367 pt_access = pte_access; in walk_addr_generic() 415 pte_access = pt_access & (pte ^ walk_nx_mask); in walk_addr_generic() 432 accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0; in walk_addr_generic() 435 walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask); in walk_addr_generic() 436 errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); in walk_addr_generic() 455 FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pt in walk_addr_generic() 537 unsigned pte_access; prefetch_gpte() local 909 unsigned pte_access; sync_spte() local [all...] |
H A D | spte.c | 139 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, in make_spte() 147 WARN_ON_ONCE(!pte_access && !shadow_present_mask); in make_spte() 176 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 178 pte_access &= ~ACC_EXEC_MASK; in make_spte() 181 if (pte_access & ACC_EXEC_MASK) in make_spte() 186 if (pte_access & ACC_USER_MASK) in make_spte() 198 pte_access &= ~ACC_WRITE_MASK; in make_spte() 205 if (pte_access & ACC_WRITE_MASK) { in make_spte() 225 pte_access &= ~ACC_WRITE_MASK; in make_spte() 230 if (pte_access in make_spte() 137 make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, const struct kvm_memory_slot *slot, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool prefetch, bool can_unsync, bool host_writable, u64 *new_spte) make_spte() argument [all...] |
H A D | spte.h | 476 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
|
H A D | mmu.c | 2901 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte() 2919 mark_mmio_spte(vcpu, sptep, gfn, pte_access); in mmu_set_spte() 2942 wrprot = make_spte(vcpu, sp, slot, pte_access, gfn, pfn, *sptep, prefetch, in mmu_set_spte() 2962 rmap_add(vcpu, slot, sptep, gfn, pte_access); in mmu_set_spte() 2964 /* Already rmapped but the pte_access bits may have changed. */ in mmu_set_spte() 2965 kvm_mmu_page_set_access(sp, spte_index(sptep), pte_access); in mmu_set_spte() 2900 mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, u64 *sptep, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, struct kvm_page_fault *fault) mmu_set_spte() argument
|
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
H A D | spte.c | 84 int make_spte(struct kvm_vcpu *vcpu, unsigned int pte_access, int level, in make_spte() argument 107 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 109 pte_access &= ~ACC_EXEC_MASK; in make_spte() 112 if (pte_access & ACC_EXEC_MASK) in make_spte() 117 if (pte_access & ACC_USER_MASK) in make_spte() 129 pte_access &= ~ACC_WRITE_MASK; in make_spte() 136 if (pte_access & ACC_WRITE_MASK) { in make_spte() 152 pte_access &= ~ACC_WRITE_MASK; in make_spte() 157 if (pte_access & ACC_WRITE_MASK) in make_spte()
|
H A D | paging_tmpl.h | 95 unsigned int pte_access; member 316 u64 pt_access, pte_access; in walk_addr_generic() local 356 pte_access = ~0; in walk_addr_generic() 362 pt_access = pte_access; in walk_addr_generic() 407 pte_access = pt_access & (pte ^ walk_nx_mask); in walk_addr_generic() 424 accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0; in walk_addr_generic() 427 walker->pte_access = FNAME(gpte_access)(pte_access ^ walk_nx_mask); in walk_addr_generic() 428 errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); in walk_addr_generic() 445 FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pt in walk_addr_generic() 531 unsigned pte_access; prefetch_gpte() local 1032 unsigned pte_access; sync_page() local [all...] |
H A D | spte.h | 246 int make_spte(struct kvm_vcpu *vcpu, unsigned int pte_access, int level,
|
H A D | mmu.c | 2561 unsigned int pte_access, int level, in set_spte() 2569 if (set_mmio_spte(vcpu, sptep, gfn, pfn, pte_access)) in set_spte() 2574 ret = make_spte(vcpu, pte_access, level, gfn, pfn, *sptep, speculative, in set_spte() 2588 unsigned int pte_access, bool write_fault, int level, in mmu_set_spte() 2622 set_spte_ret = set_spte(vcpu, sptep, pte_access, level, gfn, pfn, in mmu_set_spte() 2560 set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned int pte_access, int level, gfn_t gfn, kvm_pfn_t pfn, bool speculative, bool can_unsync, bool host_writable) set_spte() argument 2587 mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned int pte_access, bool write_fault, int level, gfn_t gfn, kvm_pfn_t pfn, bool speculative, bool host_writable) mmu_set_spte() argument
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | mmu.h | 167 unsigned pte_access, unsigned pte_pkey, in permission_fault() 189 bool fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault() 206 ((pte_access & PT_USER_MASK) << (PFERR_RSVD_BIT - PT_USER_SHIFT)); in permission_fault() 166 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned pte_access, unsigned pte_pkey, unsigned pfec) permission_fault() argument
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | mmu.h | 187 unsigned pte_access, unsigned pte_pkey, in permission_fault() 214 fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault() 230 ((pte_access & PT_USER_MASK) << (PFERR_RSVD_BIT - PT_USER_SHIFT)); in permission_fault() 186 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, unsigned pte_access, unsigned pte_pkey, u64 access) permission_fault() argument
|