Lines Matching refs:vmcs01
283 if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01))
284 vmx_switch_vmcs(vcpu, &vmx->vmcs01);
298 vmcs_clear(vmx->vmcs01.shadow_vmcs);
299 free_vmcs(vmx->vmcs01.shadow_vmcs);
300 vmx->vmcs01.shadow_vmcs = NULL;
324 * vmcs01 of the vcpu before calling free_nested().
495 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap;
1557 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
1592 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
2195 * Conceptually we want to copy the PML index from vmcs01 here,
2196 * and then back to vmcs01 on nested vmexit. But since we flush
2235 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01,
2247 exec_control = __pin_controls_get(vmcs01);
2262 exec_control = __exec_controls_get(vmcs01); /* L0's desires */
2299 exec_control = __secondary_exec_controls_get(vmcs01);
2349 * Similarly, take vmcs01's PERF_GLOBAL_CTRL in the hope that if KVM is
2353 exec_control = __vm_entry_controls_get(vmcs01);
2372 exec_control = __vm_exit_controls_get(vmcs01);
2480 * "or"ing of the EB of vmcs01 and vmcs12, because when enable_ept,
2481 * vmcs01's EB.PF is 0 so the "or" will take vmcs12's value, and when
2519 * with L0's requirements for its guest (a.k.a. vmcs01), so we can run the L2
3371 * Overwrite vmcs01.GUEST_CR3 with L1's CR3 if EPT is disabled *and*
3377 * vmcs01.GUEST_CR3 results in the unwind naturally setting arch.cr3 to
3378 * the correct value. Smashing vmcs01.GUEST_CR3 is safe because nested
3379 * VM-Exits, and the unwind, reset KVM's MMU, i.e. vmcs01.GUEST_CR3 is
3381 * L1. Don't stuff vmcs01.GUEST_CR3 when using nested early checks as
3384 * path would need to manually save/restore vmcs01.GUEST_CR3.
3391 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12);
3395 vmx_switch_vmcs(vcpu, &vmx->vmcs01);
3400 vmx_switch_vmcs(vcpu, &vmx->vmcs01);
3489 vmx_switch_vmcs(vcpu, &vmx->vmcs01);
4073 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01);
4077 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01);
4081 vmx->loaded_vmcs = &vmx->vmcs01;
4222 * This function should be called when the active VMCS is L1's (vmcs01).
4247 * CR0_GUEST_HOST_MASK is already set in the original vmcs01
4384 * as vmcs01.GUEST_DR7 contains a userspace defined value
4412 * from vmcs01 (if necessary). The PDPTRs are not loaded on
4563 vmx_switch_vmcs(vcpu, &vmx->vmcs01);
4839 * We should allocate a shadow vmcs for vmcs01 only when L1
4842 * here when vmcs01 already have an allocated shadow vmcs.
4844 WARN_ON(loaded_vmcs == &vmx->vmcs01 && loaded_vmcs->shadow_vmcs);
5278 vmcs_load(vmx->vmcs01.shadow_vmcs);
5282 vmcs_clear(vmx->vmcs01.shadow_vmcs);
5298 __pa(vmx->vmcs01.shadow_vmcs));
6178 * in the shadow or enlightened vmcs linked to vmcs01, unless
6435 * in vmcs01 and prepare_vmcs02, when it bitwise-or's the control
6436 * fields of vmcs01 and vmcs02, will turn these bits off - and