Lines Matching refs:vcpu_vmx

365 static __always_inline void vmx_disable_fb_clear(struct vcpu_vmx *vmx)
379 static __always_inline void vmx_enable_fb_clear(struct vcpu_vmx *vmx)
388 static void vmx_update_fb_clear_dis(struct kvm_vcpu *vcpu, struct vcpu_vmx *vmx)
519 static inline void vmx_segment_cache_clear(struct vcpu_vmx *vmx)
720 struct vmx_uret_msr *vmx_find_uret_msr(struct vcpu_vmx *vmx, u32 msr)
730 static int vmx_set_guest_uret_msr(struct vcpu_vmx *vmx,
830 static bool vmx_segment_cache_test_set(struct vcpu_vmx *vmx, unsigned seg,
845 static u16 vmx_read_guest_seg_selector(struct vcpu_vmx *vmx, unsigned seg)
854 static ulong vmx_read_guest_seg_base(struct vcpu_vmx *vmx, unsigned seg)
863 static u32 vmx_read_guest_seg_limit(struct vcpu_vmx *vmx, unsigned seg)
872 static u32 vmx_read_guest_seg_ar(struct vcpu_vmx *vmx, unsigned seg)
943 static bool msr_write_intercepted(struct vcpu_vmx *vmx, u32 msr)
951 unsigned int __vmx_vcpu_run_flags(struct vcpu_vmx *vmx)
969 static __always_inline void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx,
987 static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr)
1027 static __always_inline void add_atomic_switch_msr_special(struct vcpu_vmx *vmx,
1039 static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr,
1105 static bool update_transition_efer(struct vcpu_vmx *vmx)
1187 static inline bool pt_can_write_msr(struct vcpu_vmx *vmx)
1227 static void pt_guest_enter(struct vcpu_vmx *vmx)
1244 static void pt_guest_exit(struct vcpu_vmx *vmx)
1291 struct vcpu_vmx *vmx = to_vmx(vcpu);
1363 static void vmx_prepare_switch_to_host(struct vcpu_vmx *vmx)
1403 static u64 vmx_read_guest_kernel_gs_base(struct vcpu_vmx *vmx)
1412 static void vmx_write_guest_kernel_gs_base(struct vcpu_vmx *vmx, u64 data)
1425 struct vcpu_vmx *vmx = to_vmx(vcpu);
1495 struct vcpu_vmx *vmx = to_vmx(vcpu);
1518 struct vcpu_vmx *vmx = to_vmx(vcpu);
1536 struct vcpu_vmx *vmx = to_vmx(vcpu);
1599 struct vcpu_vmx *vmx = to_vmx(vcpu);
1757 struct vcpu_vmx *vmx = to_vmx(vcpu);
1808 struct vcpu_vmx *vmx = to_vmx(vcpu);
1849 static void vmx_setup_uret_msr(struct vcpu_vmx *vmx, unsigned int msr,
1867 static void vmx_setup_uret_msrs(struct vcpu_vmx *vmx)
1949 static inline bool is_vmx_feature_control_msr_valid(struct vcpu_vmx *vmx,
1992 struct vcpu_vmx *vmx = to_vmx(vcpu);
2171 struct vcpu_vmx *vmx = to_vmx(vcpu);
3005 struct vcpu_vmx *vmx = to_vmx(vcpu);
3075 struct vcpu_vmx *vmx = to_vmx(vcpu);
3122 struct vcpu_vmx *vmx = to_vmx(vcpu);
3171 struct vcpu_vmx *vmx = to_vmx(vcpu);
3282 struct vcpu_vmx *vmx = to_vmx(vcpu);
3439 struct vcpu_vmx *vmx = to_vmx(vcpu);
3502 struct vcpu_vmx *vmx = to_vmx(vcpu);
3549 struct vcpu_vmx *vmx = to_vmx(vcpu);
3578 struct vcpu_vmx *vmx = to_vmx(vcpu);
3948 static void vmx_msr_bitmap_l01_changed(struct vcpu_vmx *vmx)
3968 struct vcpu_vmx *vmx = to_vmx(vcpu);
4012 struct vcpu_vmx *vmx = to_vmx(vcpu);
4051 struct vcpu_vmx *vmx = to_vmx(vcpu);
4104 struct vcpu_vmx *vmx = to_vmx(vcpu);
4120 struct vcpu_vmx *vmx = to_vmx(vcpu);
4140 struct vcpu_vmx *vmx = to_vmx(vcpu);
4213 struct vcpu_vmx *vmx = to_vmx(vcpu);
4251 struct vcpu_vmx *vmx = to_vmx(vcpu);
4300 void vmx_set_constant_host_state(struct vcpu_vmx *vmx)
4367 void set_cr4_guest_host_mask(struct vcpu_vmx *vmx)
4383 static u32 vmx_pin_based_exec_ctrl(struct vcpu_vmx *vmx)
4444 struct vcpu_vmx *vmx = to_vmx(vcpu);
4470 static u32 vmx_exec_control(struct vcpu_vmx *vmx)
4514 static u64 vmx_tertiary_exec_control(struct vcpu_vmx *vmx)
4534 vmx_adjust_secondary_exec_control(struct vcpu_vmx *vmx, u32 *exec_control,
4594 static u32 vmx_secondary_exec_control(struct vcpu_vmx *vmx)
4713 static void init_vmcs(struct vcpu_vmx *vmx)
4833 struct vcpu_vmx *vmx = to_vmx(vcpu);
4860 struct vcpu_vmx *vmx = to_vmx(vcpu);
4935 struct vcpu_vmx *vmx = to_vmx(vcpu);
4963 struct vcpu_vmx *vmx = to_vmx(vcpu);
4994 struct vcpu_vmx *vmx = to_vmx(vcpu);
5008 struct vcpu_vmx *vmx = to_vmx(vcpu);
5186 struct vcpu_vmx *vmx = to_vmx(vcpu);
5691 struct vcpu_vmx *vmx = to_vmx(vcpu);
5835 struct vcpu_vmx *vmx = to_vmx(vcpu);
5843 struct vcpu_vmx *vmx = to_vmx(vcpu);
5894 struct vcpu_vmx *vmx = to_vmx(vcpu);
5910 struct vcpu_vmx *vmx = to_vmx(vcpu);
6006 struct vcpu_vmx *vmx = to_vmx(vcpu);
6150 struct vcpu_vmx *vmx = to_vmx(vcpu);
6168 static void vmx_destroy_pml_buffer(struct vcpu_vmx *vmx)
6178 struct vcpu_vmx *vmx = to_vmx(vcpu);
6235 struct vcpu_vmx *vmx = to_vmx(vcpu);
6414 struct vcpu_vmx *vmx = to_vmx(vcpu);
6697 struct vcpu_vmx *vmx = to_vmx(vcpu);
6869 struct vcpu_vmx *vmx = to_vmx(vcpu);
6926 struct vcpu_vmx *vmx = to_vmx(vcpu);
6955 static void handle_exception_irqoff(struct vcpu_vmx *vmx)
6989 struct vcpu_vmx *vmx = to_vmx(vcpu);
7026 static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx)
7120 static void vmx_complete_interrupts(struct vcpu_vmx *vmx)
7137 static void atomic_switch_perf_msrs(struct vcpu_vmx *vmx)
7162 struct vcpu_vmx *vmx = to_vmx(vcpu);
7186 void noinstr vmx_update_host_rsp(struct vcpu_vmx *vmx, unsigned long host_rsp)
7194 void noinstr vmx_spec_ctrl_restore_host(struct vcpu_vmx *vmx,
7234 struct vcpu_vmx *vmx = to_vmx(vcpu);
7287 struct vcpu_vmx *vmx = to_vmx(vcpu);
7443 struct vcpu_vmx *vmx = to_vmx(vcpu);
7455 struct vcpu_vmx *vmx;
7458 BUILD_BUG_ON(offsetof(struct vcpu_vmx, vcpu) != 0);
7632 static void vmcs_set_secondary_exec_control(struct vcpu_vmx *vmx, u32 new_ctl)
7657 struct vcpu_vmx *vmx = to_vmx(vcpu);
7697 struct vcpu_vmx *vmx = to_vmx(vcpu);
7766 struct vcpu_vmx *vmx = to_vmx(vcpu);
8042 struct vcpu_vmx *vmx;
8093 struct vcpu_vmx *vmx = to_vmx(vcpu);
8135 struct vcpu_vmx *vmx = to_vmx(vcpu);
8156 struct vcpu_vmx *vmx = to_vmx(vcpu);
8711 r = kvm_init(sizeof(struct vcpu_vmx), __alignof__(struct vcpu_vmx),