Lines Matching refs:index
160 unsigned index;
285 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index);
721 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index)
727 return sp->shadowed_translation[index] >> PAGE_SHIFT;
729 return sp->gfn + (index << ((sp->role.level - 1) * SPTE_LEVEL_BITS));
738 static u32 kvm_mmu_page_get_access(struct kvm_mmu_page *sp, int index)
741 return sp->shadowed_translation[index] & ACC_ALL;
758 static void kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index,
762 sp->shadowed_translation[index] = (gfn << PAGE_SHIFT) | access;
766 WARN_ONCE(access != kvm_mmu_page_get_access(sp, index),
769 sp->gfn, kvm_mmu_page_get_access(sp, index), access);
771 WARN_ONCE(gfn != kvm_mmu_page_get_gfn(sp, index),
774 sp->gfn, kvm_mmu_page_get_gfn(sp, index), gfn);
777 static void kvm_mmu_page_set_access(struct kvm_mmu_page *sp, int index,
780 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index);
782 kvm_mmu_page_set_translation(sp, index, gfn, access);
1120 int pos; /* index of the sptep */
2337 * _least_ significant bit of the PDE index.
2401 iterator->index = SPTE_INDEX(iterator->addr, iterator->level);
2402 iterator->sptep = ((u64 *)__va(iterator->shadow_addr)) + iterator->index;
5034 * code (minus the P bit) and the page table's U bit form an index into the
5791 int ret = kvm_sync_spte(vcpu, sp, iterator.index);
6379 int index;
6383 for (index = 0; index < SPTE_ENT_PER_PAGE; index++) {
6384 sptep = &sp->spt[index];
6385 gfn = kvm_mmu_page_get_gfn(sp, index);
6406 spte = make_huge_page_split_spte(kvm, huge_spte, sp->role, index);