Lines Matching refs:spte
152 struct kvm_mmu_page *sp, u64 *spte,
169 drop_spte(vcpu->kvm, spte);
534 u64 *spte, pt_element_t gpte)
541 if (FNAME(prefetch_invalid_gpte)(vcpu, sp, spte, gpte))
556 mmu_set_spte(vcpu, slot, spte, pte_access, gfn, pfn, NULL);
589 u64 *spte;
608 spte = sp->spt + i;
610 for (i = 0; i < PTE_PREFETCH_NUM; i++, spte++) {
611 if (spte == sptep)
614 if (is_shadow_present_pte(*spte))
617 if (!FNAME(prefetch_gpte)(vcpu, sp, spte, gptep[i]))
821 * we will cache the incorrect access into mmio spte.
895 * - The spte has a reference to the struct page, so the pfn for a given gfn
899 * < 0: failed to sync spte
900 * 0: the spte is synced and no tlb flushing is required
901 * > 0: the spte is synced and tlb flushing is required
907 u64 *sptep, spte;
960 spte = *sptep;
961 host_writable = spte & shadow_host_writable_mask;
964 spte_to_pfn(spte), spte, true, false,
965 host_writable, &spte);
967 return mmu_spte_update(sptep, spte);