Home
last modified time | relevance | path

Searched refs:vmcs12 (Results 1 - 25 of 26) sorted by relevance

12

/kernel/linux/linux-6.6/arch/x86/kvm/vmx/
H A Dnested.h6 #include "vmcs12.h"
39 static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu) in get_vmcs12()
44 static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu) in get_shadow_vmcs12()
85 * the value+mask loaded into vmcs02 may not match the vmcs12 fields.
87 static inline unsigned long nested_read_cr0(struct vmcs12 *fields) in nested_read_cr0()
92 static inline unsigned long nested_read_cr4(struct vmcs12 *fields) in nested_read_cr4()
131 static inline bool nested_cpu_has(struct vmcs12 *vmcs12, u32 bit) in nested_cpu_has() argument
133 return vmcs12->cpu_based_vm_exec_control & bit; in nested_cpu_has()
136 static inline bool nested_cpu_has2(struct vmcs12 *vmcs1 argument
143 nested_cpu_has_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_preemption_timer() argument
149 nested_cpu_has_nmi_exiting(struct vmcs12 *vmcs12) nested_cpu_has_nmi_exiting() argument
154 nested_cpu_has_virtual_nmis(struct vmcs12 *vmcs12) nested_cpu_has_virtual_nmis() argument
159 nested_cpu_has_mtf(struct vmcs12 *vmcs12) nested_cpu_has_mtf() argument
164 nested_cpu_has_ept(struct vmcs12 *vmcs12) nested_cpu_has_ept() argument
169 nested_cpu_has_xsaves(struct vmcs12 *vmcs12) nested_cpu_has_xsaves() argument
174 nested_cpu_has_pml(struct vmcs12 *vmcs12) nested_cpu_has_pml() argument
179 nested_cpu_has_virt_x2apic_mode(struct vmcs12 *vmcs12) nested_cpu_has_virt_x2apic_mode() argument
184 nested_cpu_has_vpid(struct vmcs12 *vmcs12) nested_cpu_has_vpid() argument
189 nested_cpu_has_apic_reg_virt(struct vmcs12 *vmcs12) nested_cpu_has_apic_reg_virt() argument
194 nested_cpu_has_vid(struct vmcs12 *vmcs12) nested_cpu_has_vid() argument
199 nested_cpu_has_posted_intr(struct vmcs12 *vmcs12) nested_cpu_has_posted_intr() argument
204 nested_cpu_has_vmfunc(struct vmcs12 *vmcs12) nested_cpu_has_vmfunc() argument
209 nested_cpu_has_eptp_switching(struct vmcs12 *vmcs12) nested_cpu_has_eptp_switching() argument
216 nested_cpu_has_shadow_vmcs(struct vmcs12 *vmcs12) nested_cpu_has_shadow_vmcs() argument
221 nested_cpu_has_save_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_save_preemption_timer() argument
242 nested_cpu_has_encls_exit(struct vmcs12 *vmcs12) nested_cpu_has_encls_exit() argument
260 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_guest_cr0_valid() local
[all...]
H A Dnested.c56 #define SHADOW_FIELD_RO(x, y) { x, offsetof(struct vmcs12, y) },
63 #define SHADOW_FIELD_RW(x, y) { x, offsetof(struct vmcs12, y) },
381 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in nested_ept_inject_page_fault() local
403 nested_ept_invalidate_addr(vcpu, vmcs12->ept_pointer, in nested_ept_inject_page_fault()
408 vmcs12->guest_physical_address = fault->address; in nested_ept_inject_page_fault()
441 static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12, in nested_vmx_is_page_fault_vmexit() argument
446 bit = (vmcs12->exception_bitmap & (1u << PF_VECTOR)) != 0; in nested_vmx_is_page_fault_vmexit()
448 (error_code & vmcs12 in nested_vmx_is_page_fault_vmexit()
456 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_is_exception_vmexit() local
471 nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_io_bitmap_controls() argument
484 nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_msr_bitmap_controls() argument
496 nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_tpr_shadow_controls() argument
568 nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_prepare_msr_bitmap() argument
670 nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_cache_shadow_vmcs12() argument
689 nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_flush_cached_shadow_vmcs12() argument
718 nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apic_access_controls() argument
728 nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apicv_controls() argument
786 nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_exit_msr_switch_controls() argument
800 nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_entry_msr_switch_controls() argument
811 nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_pml_controls() argument
824 nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_unrestricted_guest_controls() argument
833 nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_mode_based_ept_exec_controls() argument
842 nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_shadow_vmcs_controls() argument
1026 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_msr_store_list_has_msr() local
1130 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_has_guest_tlb_tag() local
1136 nested_vmx_transition_tlb_flush(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool is_vmenter) nested_vmx_transition_tlb_flush() argument
1521 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_shadow_to_vmcs12() local
1556 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_vmcs12_to_shadow() local
1581 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_enlightened_to_vmcs12() local
1825 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_vmcs12_to_enlightened() local
2066 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_handle_enlightened_vmptrld() local
2114 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_calc_preemption_timer_value() local
2152 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) nested_vmx_calc_efer() argument
2224 prepare_vmcs02_early_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_early_rare() argument
2239 prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, struct vmcs12 *vmcs12) prepare_vmcs02_early() argument
2404 prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_rare() argument
2533 prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool from_vmentry, enum vm_entry_failure_code *entry_failure_code) prepare_vmcs02() argument
2674 nested_vmx_check_nmi_controls(struct vmcs12 *vmcs12) nested_vmx_check_nmi_controls() argument
2735 nested_check_vm_execution_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_execution_controls() argument
2794 nested_check_vm_exit_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_exit_controls() argument
2811 nested_check_vm_entry_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_entry_controls() argument
2883 nested_vmx_check_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_controls() argument
2897 nested_vmx_check_address_space_size(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_address_space_size() argument
2908 nested_vmx_check_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_host_state() argument
2977 nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_vmcs_link_ptr() argument
3010 nested_check_guest_non_reg_state(struct vmcs12 *vmcs12) nested_check_guest_non_reg_state() argument
3020 nested_vmx_check_guest_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) nested_vmx_check_guest_state() argument
3194 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_get_vmcs12_pages() local
3307 struct vmcs12 *vmcs12; nested_vmx_write_pml_buffer() local
3389 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_enter_non_root_mode() local
3552 struct vmcs12 *vmcs12; nested_vmx_run() local
3708 vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr0() argument
3718 vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr4() argument
3727 vmcs12_save_pending_event(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info) vmcs12_save_pending_event() argument
3796 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_mark_vmcs12_pages_dirty() local
3861 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_inject_exception_vmexit() local
4254 sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12_rare() argument
4299 copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) copy_vmcs02_to_vmcs12_rare() argument
4328 sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12() argument
4409 prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info, unsigned long exit_qualification) prepare_vmcs12() argument
4465 load_vmcs12_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) load_vmcs12_host_state() argument
4615 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_restore_host_state() local
4727 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_vmexit() local
5345 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmread() local
5451 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmwrite() local
5837 nested_vmx_eptp_switching(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_eptp_switching() argument
5873 struct vmcs12 *vmcs12; handle_vmfunc() local
5928 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_io_bitmaps() local
5958 nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_io() argument
5982 nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, union vmx_exit_reason exit_reason) nested_vmx_exit_handled_msr() argument
6020 nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_cr() argument
6090 nested_vmx_exit_handled_encls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_encls() argument
6105 nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, gpa_t bitmap) nested_vmx_exit_handled_vmcs_access() argument
6129 nested_vmx_exit_handled_mtf(struct vmcs12 *vmcs12) nested_vmx_exit_handled_mtf() argument
6229 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_l1_wants_exit() local
6393 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_reflect_vmexit() local
6410 struct vmcs12 *vmcs12; vmx_get_nested_state() local
6542 struct vmcs12 *vmcs12; vmx_set_nested_state() local
[all...]
H A Dsgx.h18 void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12);
26 struct vmcs12 *vmcs12) in vmx_write_encls_bitmap()
25 vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmx_write_encls_bitmap() argument
H A Dhyperv.c574 int nested_evmcs_check_controls(struct vmcs12 *vmcs12) in nested_evmcs_check_controls() argument
577 vmcs12->pin_based_vm_exec_control))) in nested_evmcs_check_controls()
581 vmcs12->cpu_based_vm_exec_control))) in nested_evmcs_check_controls()
585 vmcs12->secondary_vm_exec_control))) in nested_evmcs_check_controls()
589 vmcs12->vm_exit_controls))) in nested_evmcs_check_controls()
593 vmcs12->vm_entry_controls))) in nested_evmcs_check_controls()
601 if (WARN_ON_ONCE(vmcs12->vm_function_control >> 32)) in nested_evmcs_check_controls()
605 vmcs12->vm_function_control))) in nested_evmcs_check_controls()
H A Dvmcs12.h10 * struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a
11 * single nested guest (L2), hence the name vmcs12. Any VMX implementation has
12 * a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is
27 struct __packed vmcs12 { struct
192 * layout of struct vmcs12 is changed. MSR_IA32_VMX_BASIC returns this id, and
202 * and any VMCS region. Although only sizeof(struct vmcs12) are used by the
209 * For save/restore compatibility, the vmcs12 field offsets must not change.
212 ASSERT_STRUCT_OFFSET(struct vmcs12, field, loc)
385 static inline u64 vmcs12_read_any(struct vmcs12 *vmcs12, unsigne argument
405 vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field, u16 offset, u64 field_value) vmcs12_write_any() argument
[all...]
H A Dsgx.c467 void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) in vmx_write_encls_bitmap() argument
503 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap()
504 vmcs12 = get_vmcs12(vcpu); in vmx_write_encls_bitmap()
505 if (vmcs12 && nested_cpu_has_encls_exit(vmcs12)) in vmx_write_encls_bitmap()
506 bitmap |= vmcs12->encls_exiting_bitmap; in vmx_write_encls_bitmap()
H A Dhyperv.h15 #include "vmcs12.h"
61 * is 'struct vmcs12' or 'struct hv_enlightened_vmcs' as it takes in evmcs_read_any()
196 int nested_evmcs_check_controls(struct vmcs12 *vmcs12);
H A Dvmx.c65 #include "vmcs12.h"
1756 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_update_emulated_instruction() local
1773 if (nested_cpu_has_mtf(vmcs12) && in vmx_update_emulated_instruction()
1906 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_get_l2_tsc_offset() local
1908 if (nested_cpu_has(vmcs12, CPU_BASED_USE_TSC_OFFSETTING)) in vmx_get_l2_tsc_offset()
1909 return vmcs12->tsc_offset; in vmx_get_l2_tsc_offset()
1916 struct vmcs12 *vmcs12 in vmx_get_l2_tsc_multiplier() local
5397 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr0() local
5423 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr4() local
6681 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_update_cr8_intercept() local
7925 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept_io() local
7961 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept() local
[all...]
H A Dvmx.h129 struct vmcs12 *cached_vmcs12;
135 struct vmcs12 *cached_shadow_vmcs12;
138 * GPA to HVA cache for accessing vmcs12->vmcs_link_pointer
149 * with the data held by struct vmcs12.
H A Dvmcs12.c4 #include "vmcs12.h"
6 #define VMCS12_OFFSET(x) offsetof(struct vmcs12, x)
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/
H A Dnested.h6 #include "vmcs12.h"
40 static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu) in get_vmcs12()
45 static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu) in get_shadow_vmcs12()
63 * have vmcs12 if it is true. in vmx_has_valid_vmcs12()
92 static inline unsigned long nested_read_cr0(struct vmcs12 *fields) in nested_read_cr0()
97 static inline unsigned long nested_read_cr4(struct vmcs12 *fields) in nested_read_cr4()
136 static inline bool nested_cpu_has(struct vmcs12 *vmcs12, u32 bit) in nested_cpu_has() argument
138 return vmcs12->cpu_based_vm_exec_control & bit; in nested_cpu_has()
141 static inline bool nested_cpu_has2(struct vmcs12 *vmcs1 argument
148 nested_cpu_has_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_preemption_timer() argument
154 nested_cpu_has_nmi_exiting(struct vmcs12 *vmcs12) nested_cpu_has_nmi_exiting() argument
159 nested_cpu_has_virtual_nmis(struct vmcs12 *vmcs12) nested_cpu_has_virtual_nmis() argument
164 nested_cpu_has_mtf(struct vmcs12 *vmcs12) nested_cpu_has_mtf() argument
169 nested_cpu_has_ept(struct vmcs12 *vmcs12) nested_cpu_has_ept() argument
174 nested_cpu_has_xsaves(struct vmcs12 *vmcs12) nested_cpu_has_xsaves() argument
179 nested_cpu_has_pml(struct vmcs12 *vmcs12) nested_cpu_has_pml() argument
184 nested_cpu_has_virt_x2apic_mode(struct vmcs12 *vmcs12) nested_cpu_has_virt_x2apic_mode() argument
189 nested_cpu_has_vpid(struct vmcs12 *vmcs12) nested_cpu_has_vpid() argument
194 nested_cpu_has_apic_reg_virt(struct vmcs12 *vmcs12) nested_cpu_has_apic_reg_virt() argument
199 nested_cpu_has_vid(struct vmcs12 *vmcs12) nested_cpu_has_vid() argument
204 nested_cpu_has_posted_intr(struct vmcs12 *vmcs12) nested_cpu_has_posted_intr() argument
209 nested_cpu_has_vmfunc(struct vmcs12 *vmcs12) nested_cpu_has_vmfunc() argument
214 nested_cpu_has_eptp_switching(struct vmcs12 *vmcs12) nested_cpu_has_eptp_switching() argument
221 nested_cpu_has_shadow_vmcs(struct vmcs12 *vmcs12) nested_cpu_has_shadow_vmcs() argument
226 nested_cpu_has_save_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_save_preemption_timer() argument
260 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_guest_cr0_valid() local
[all...]
H A Dnested.c59 #define SHADOW_FIELD_RO(x, y) { x, offsetof(struct vmcs12, y) },
66 #define SHADOW_FIELD_RW(x, y) { x, offsetof(struct vmcs12, y) },
336 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in nested_ept_inject_page_fault() local
351 vmcs12->guest_physical_address = fault->address; in nested_ept_inject_page_fault()
377 static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12, in nested_vmx_is_page_fault_vmexit() argument
382 bit = (vmcs12->exception_bitmap & (1u << PF_VECTOR)) != 0; in nested_vmx_is_page_fault_vmexit()
384 (error_code & vmcs12->page_fault_error_code_mask) != in nested_vmx_is_page_fault_vmexit()
385 vmcs12 in nested_vmx_is_page_fault_vmexit()
396 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_exception() local
431 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_inject_page_fault_nested() local
447 nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_io_bitmap_controls() argument
460 nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_msr_bitmap_controls() argument
472 nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_tpr_shadow_controls() argument
564 nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_prepare_msr_bitmap() argument
663 nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_cache_shadow_vmcs12() argument
682 nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_flush_cached_shadow_vmcs12() argument
705 nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apic_access_controls() argument
715 nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apicv_controls() argument
776 nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_exit_msr_switch_controls() argument
790 nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_entry_msr_switch_controls() argument
801 nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_pml_controls() argument
814 nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_unrestricted_guest_controls() argument
823 nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_mode_based_ept_exec_controls() argument
832 nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_shadow_vmcs_controls() argument
1016 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_msr_store_list_has_msr() local
1183 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_has_guest_tlb_tag() local
1189 nested_vmx_transition_tlb_flush(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool is_vmenter) nested_vmx_transition_tlb_flush() argument
1558 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_shadow_to_vmcs12() local
1593 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_vmcs12_to_shadow() local
1618 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_enlightened_to_vmcs12() local
1834 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_vmcs12_to_enlightened() local
2064 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_handle_enlightened_vmptrld() local
2113 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_calc_preemption_timer_value() local
2151 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) nested_vmx_calc_efer() argument
2220 prepare_vmcs02_early_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_early_rare() argument
2235 prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, struct vmcs12 *vmcs12) prepare_vmcs02_early() argument
2398 prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_rare() argument
2527 prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) prepare_vmcs02() argument
2649 nested_vmx_check_nmi_controls(struct vmcs12 *vmcs12) nested_vmx_check_nmi_controls() argument
2711 nested_check_vm_execution_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_execution_controls() argument
2770 nested_check_vm_exit_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_exit_controls() argument
2787 nested_check_vm_entry_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_entry_controls() argument
2859 nested_vmx_check_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_controls() argument
2873 nested_vmx_check_address_space_size(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_address_space_size() argument
2884 nested_vmx_check_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_host_state() argument
2959 nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_vmcs_link_ptr() argument
2988 nested_check_guest_non_reg_state(struct vmcs12 *vmcs12) nested_check_guest_non_reg_state() argument
2997 nested_vmx_check_guest_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) nested_vmx_check_guest_state() argument
3164 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_get_vmcs12_pages() local
3263 struct vmcs12 *vmcs12; nested_vmx_write_pml_buffer() local
3345 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_enter_non_root_mode() local
3507 struct vmcs12 *vmcs12; nested_vmx_run() local
3649 vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr0() argument
3659 vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr4() argument
3668 vmcs12_save_pending_event(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info) vmcs12_save_pending_event() argument
3737 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_mark_vmcs12_pages_dirty() local
3792 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_inject_exception_vmexit() local
4016 sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12_rare() argument
4063 copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) copy_vmcs02_to_vmcs12_rare() argument
4092 sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12() argument
4170 prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info, unsigned long exit_qualification) prepare_vmcs12() argument
4224 load_vmcs12_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) load_vmcs12_host_state() argument
4373 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_restore_host_state() local
4488 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_vmexit() local
5103 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmread() local
5187 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmwrite() local
5572 nested_vmx_eptp_switching(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_eptp_switching() argument
5608 struct vmcs12 *vmcs12; handle_vmfunc() local
5654 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_io_bitmaps() local
5684 nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_io() argument
5708 nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, union vmx_exit_reason exit_reason) nested_vmx_exit_handled_msr() argument
5746 nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_cr() argument
5816 nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, gpa_t bitmap) nested_vmx_exit_handled_vmcs_access() argument
5840 nested_vmx_exit_handled_mtf(struct vmcs12 *vmcs12) nested_vmx_exit_handled_mtf() argument
5929 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_l1_wants_exit() local
6088 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_reflect_vmexit() local
6105 struct vmcs12 *vmcs12; vmx_get_nested_state() local
6232 struct vmcs12 *vmcs12; vmx_set_nested_state() local
[all...]
H A Dvmcs12.h10 * struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a
11 * single nested guest (L2), hence the name vmcs12. Any VMX implementation has
12 * a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is
27 struct __packed vmcs12 { struct
190 * layout of struct vmcs12 is changed. MSR_IA32_VMX_BASIC returns this id, and
200 * and any VMCS region. Although only sizeof(struct vmcs12) are used by the
213 * For save/restore compatibility, the vmcs12 field offsets must not change.
216 BUILD_BUG_ON_MSG(offsetof(struct vmcs12, field) != (loc), \
217 "Offset of " #field " in struct vmcs12 has changed.")
392 static inline u64 vmcs12_read_any(struct vmcs12 *vmcs1 argument
412 vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field, u16 offset, u64 field_value) vmcs12_write_any() argument
[all...]
H A Devmcs.c376 int nested_evmcs_check_controls(struct vmcs12 *vmcs12) in nested_evmcs_check_controls() argument
381 unsupp_ctl = vmcs12->pin_based_vm_exec_control & in nested_evmcs_check_controls()
390 unsupp_ctl = vmcs12->secondary_vm_exec_control & in nested_evmcs_check_controls()
399 unsupp_ctl = vmcs12->vm_exit_controls & in nested_evmcs_check_controls()
408 unsupp_ctl = vmcs12->vm_entry_controls & in nested_evmcs_check_controls()
417 unsupp_ctl = vmcs12->vm_function_control & EVMCS1_UNSUPPORTED_VMFUNC; in nested_evmcs_check_controls()
H A Devmcs.h13 #include "vmcs12.h"
203 int nested_evmcs_check_controls(struct vmcs12 *vmcs12);
H A Dvmx.c63 #include "vmcs12.h"
1695 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_update_emulated_instruction() local
1708 if (nested_cpu_has_mtf(vmcs12) && in vmx_update_emulated_instruction()
1831 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_write_l1_tsc_offset() local
1841 (vmcs12->cpu_based_vm_exec_control & CPU_BASED_USE_TSC_OFFSETTING)) in vmx_write_l1_tsc_offset()
1842 g_tsc_offset = vmcs12->tsc_offset; in vmx_write_l1_tsc_offset()
5086 struct vmcs12 *vmcs12 in handle_set_cr0() local
5119 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr4() local
6304 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); update_cr8_intercept() local
7491 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept_io() local
7527 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept() local
[all...]
H A Dvmx.h114 struct vmcs12 *cached_vmcs12;
120 struct vmcs12 *cached_shadow_vmcs12;
124 * with the data held by struct vmcs12.
H A Dvmcs12.c3 #include "vmcs12.h"
6 #define VMCS12_OFFSET(x) offsetof(struct vmcs12, x)
/kernel/linux/linux-5.10/arch/x86/kvm/
H A DMakefile21 kvm-intel-y += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o \
/kernel/linux/linux-6.6/arch/x86/kvm/
H A DMakefile25 kvm-intel-y += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o \
/kernel/linux/linux-5.10/arch/x86/include/uapi/asm/
H A Dkvm.h426 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
/kernel/linux/linux-5.10/tools/arch/x86/include/uapi/asm/
H A Dkvm.h426 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/asm-x86/asm/
H A Dkvm.h351 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/asm-x86/asm/
H A Dkvm.h351 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
/kernel/linux/linux-6.6/arch/x86/include/uapi/asm/
H A Dkvm.h464 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member

Completed in 33 milliseconds

12