Lines Matching refs:to_vmx

547 	evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs;
899 if (to_vmx(vcpu)->rmode.vm86_active)
1291 struct vcpu_vmx *vmx = to_vmx(vcpu);
1425 struct vcpu_vmx *vmx = to_vmx(vcpu);
1495 struct vcpu_vmx *vmx = to_vmx(vcpu);
1508 vmx_prepare_switch_to_host(to_vmx(vcpu));
1518 struct vcpu_vmx *vmx = to_vmx(vcpu);
1536 struct vcpu_vmx *vmx = to_vmx(vcpu);
1599 struct vcpu_vmx *vmx = to_vmx(vcpu);
1679 if (to_vmx(vcpu)->exit_reason.enclave_mode) {
1688 union vmx_exit_reason exit_reason = to_vmx(vcpu)->exit_reason;
1757 struct vcpu_vmx *vmx = to_vmx(vcpu);
1808 struct vcpu_vmx *vmx = to_vmx(vcpu);
1992 struct vcpu_vmx *vmx = to_vmx(vcpu);
2026 msr_info->data = to_vmx(vcpu)->spec_ctrl;
2058 msr_info->data = to_vmx(vcpu)->msr_ia32_sgxlepubkeyhash
2171 struct vcpu_vmx *vmx = to_vmx(vcpu);
2250 if (intel_pmu_lbr_is_enabled(vcpu) && !to_vmx(vcpu)->lbr_desc.event &&
2330 !(to_vmx(vcpu)->msr_ia32_feature_control &
2500 if (!(exec_controls_get(to_vmx(vcpu)) & CPU_BASED_CR3_LOAD_EXITING))
3005 struct vcpu_vmx *vmx = to_vmx(vcpu);
3075 struct vcpu_vmx *vmx = to_vmx(vcpu);
3122 struct vcpu_vmx *vmx = to_vmx(vcpu);
3149 vmx_segment_cache_clear(to_vmx(vcpu));
3171 struct vcpu_vmx *vmx = to_vmx(vcpu);
3196 return to_vmx(vcpu)->vpid;
3274 if (to_vmx(vcpu)->nested.vmxon)
3282 struct vcpu_vmx *vmx = to_vmx(vcpu);
3430 if (to_vmx(vcpu)->nested.vmxon && !nested_cr4_valid(vcpu, cr4))
3439 struct vcpu_vmx *vmx = to_vmx(vcpu);
3502 struct vcpu_vmx *vmx = to_vmx(vcpu);
3540 if (to_vmx(vcpu)->rmode.vm86_active) {
3544 return vmx_read_guest_seg_base(to_vmx(vcpu), seg);
3549 struct vcpu_vmx *vmx = to_vmx(vcpu);
3578 struct vcpu_vmx *vmx = to_vmx(vcpu);
3617 to_vmx(vcpu)->emulation_required = vmx_emulation_required(vcpu);
3622 u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS);
3968 struct vcpu_vmx *vmx = to_vmx(vcpu);
4012 struct vcpu_vmx *vmx = to_vmx(vcpu);
4051 struct vcpu_vmx *vmx = to_vmx(vcpu);
4104 struct vcpu_vmx *vmx = to_vmx(vcpu);
4120 struct vcpu_vmx *vmx = to_vmx(vcpu);
4140 struct vcpu_vmx *vmx = to_vmx(vcpu);
4213 struct vcpu_vmx *vmx = to_vmx(vcpu);
4251 struct vcpu_vmx *vmx = to_vmx(vcpu);
4444 struct vcpu_vmx *vmx = to_vmx(vcpu);
4833 struct vcpu_vmx *vmx = to_vmx(vcpu);
4860 struct vcpu_vmx *vmx = to_vmx(vcpu);
4919 exec_controls_setbit(to_vmx(vcpu), CPU_BASED_INTR_WINDOW_EXITING);
4930 exec_controls_setbit(to_vmx(vcpu), CPU_BASED_NMI_WINDOW_EXITING);
4935 struct vcpu_vmx *vmx = to_vmx(vcpu);
4963 struct vcpu_vmx *vmx = to_vmx(vcpu);
4994 struct vcpu_vmx *vmx = to_vmx(vcpu);
5008 struct vcpu_vmx *vmx = to_vmx(vcpu);
5031 if (!enable_vnmi && to_vmx(vcpu)->loaded_vmcs->soft_vnmi_blocked)
5041 if (to_vmx(vcpu)->nested.nested_run_pending)
5063 if (to_vmx(vcpu)->nested.nested_run_pending)
5110 to_vmx(vcpu)->vcpu.arch.event_exit_inst_len =
5186 struct vcpu_vmx *vmx = to_vmx(vcpu);
5566 exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_MOV_DR_EXITING);
5602 exec_controls_setbit(to_vmx(vcpu), CPU_BASED_MOV_DR_EXITING);
5624 exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_INTR_WINDOW_EXITING);
5691 struct vcpu_vmx *vmx = to_vmx(vcpu);
5760 if (!(to_vmx(vcpu)->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
5826 exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_NMI_WINDOW_EXITING);
5835 struct vcpu_vmx *vmx = to_vmx(vcpu);
5843 struct vcpu_vmx *vmx = to_vmx(vcpu);
5894 struct vcpu_vmx *vmx = to_vmx(vcpu);
5910 struct vcpu_vmx *vmx = to_vmx(vcpu);
5991 if (!(to_vmx(vcpu)->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
6006 struct vcpu_vmx *vmx = to_vmx(vcpu);
6053 to_vmx(vcpu)->exit_reason.bus_lock_detected = true;
6150 struct vcpu_vmx *vmx = to_vmx(vcpu);
6178 struct vcpu_vmx *vmx = to_vmx(vcpu);
6235 struct vcpu_vmx *vmx = to_vmx(vcpu);
6414 struct vcpu_vmx *vmx = to_vmx(vcpu);
6599 if (to_vmx(vcpu)->exit_reason.bus_lock_detected) {
6690 to_vmx(vcpu)->nested.l1_tpr_threshold = tpr_threshold;
6697 struct vcpu_vmx *vmx = to_vmx(vcpu);
6760 to_vmx(vcpu)->nested.reload_vmcs01_apic_access_page = true;
6764 if (!(secondary_exec_controls_get(to_vmx(vcpu)) &
6869 struct vcpu_vmx *vmx = to_vmx(vcpu);
6926 struct vcpu_vmx *vmx = to_vmx(vcpu);
6989 struct vcpu_vmx *vmx = to_vmx(vcpu);
7162 struct vcpu_vmx *vmx = to_vmx(vcpu);
7221 switch (to_vmx(vcpu)->exit_reason.basic) {
7234 struct vcpu_vmx *vmx = to_vmx(vcpu);
7287 struct vcpu_vmx *vmx = to_vmx(vcpu);
7443 struct vcpu_vmx *vmx = to_vmx(vcpu);
7459 vmx = to_vmx(vcpu);
7657 struct vcpu_vmx *vmx = to_vmx(vcpu);
7697 struct vcpu_vmx *vmx = to_vmx(vcpu);
7766 struct vcpu_vmx *vmx = to_vmx(vcpu);
7919 to_vmx(vcpu)->req_immediate_exit = true;
8046 vmx = to_vmx(vcpu);
8081 to_vmx(vcpu)->hv_deadline_tsc = -1;
8093 struct vcpu_vmx *vmx = to_vmx(vcpu);
8117 to_vmx(vcpu)->msr_ia32_feature_control_valid_bits |=
8120 to_vmx(vcpu)->msr_ia32_feature_control_valid_bits &=
8128 if (to_vmx(vcpu)->nested.nested_run_pending)
8135 struct vcpu_vmx *vmx = to_vmx(vcpu);
8156 struct vcpu_vmx *vmx = to_vmx(vcpu);
8183 return to_vmx(vcpu)->nested.vmxon && !is_guest_mode(vcpu);
8189 struct hrtimer *timer = &to_vmx(vcpu)->nested.preemption_timer;