Lines Matching defs:efer
295 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer)
298 u64 old_efer = vcpu->arch.efer;
299 vcpu->arch.efer = efer;
303 efer |= EFER_NX;
305 if (!(efer & EFER_LMA))
306 efer &= ~EFER_LME;
309 if ((old_efer & EFER_SVME) != (efer & EFER_SVME)) {
310 if (!(efer & EFER_SVME)) {
329 vcpu->arch.efer = old_efer;
342 svm->vmcb->save.efer = efer | EFER_SVME;
591 uint64_t efer;
594 rdmsrl(MSR_EFER, efer);
595 if (efer & EFER_SVME) {
601 wrmsrl(MSR_EFER, efer & ~EFER_SVME);
627 uint64_t efer;
630 rdmsrl(MSR_EFER, efer);
631 if (efer & EFER_SVME)
640 wrmsrl(MSR_EFER, efer | EFER_SVME);
1876 if (vcpu->arch.efer & EFER_LME) {
1878 vcpu->arch.efer |= EFER_LMA;
1880 svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
1884 vcpu->arch.efer &= ~EFER_LMA;
1886 svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
2984 /* check for svm_disable while efer.svme is set */
2985 if (svm_dis && (vcpu->arch.efer & EFER_SVME))
3422 pr_err("vmpl: %d cpl: %d efer: %016llx\n",
3423 save->vmpl, save->cpl, save->efer);
4675 if (!(smram64->efer & EFER_SVME))