/kernel/linux/linux-6.6/arch/x86/kvm/vmx/ |
H A D | nested.h | 6 #include "vmcs12.h" 39 static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu) in get_vmcs12() 44 static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu) in get_shadow_vmcs12() 85 * the value+mask loaded into vmcs02 may not match the vmcs12 fields. 87 static inline unsigned long nested_read_cr0(struct vmcs12 *fields) in nested_read_cr0() 92 static inline unsigned long nested_read_cr4(struct vmcs12 *fields) in nested_read_cr4() 131 static inline bool nested_cpu_has(struct vmcs12 *vmcs12, u32 bit) in nested_cpu_has() argument 133 return vmcs12->cpu_based_vm_exec_control & bit; in nested_cpu_has() 136 static inline bool nested_cpu_has2(struct vmcs12 *vmcs1 argument 143 nested_cpu_has_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_preemption_timer() argument 149 nested_cpu_has_nmi_exiting(struct vmcs12 *vmcs12) nested_cpu_has_nmi_exiting() argument 154 nested_cpu_has_virtual_nmis(struct vmcs12 *vmcs12) nested_cpu_has_virtual_nmis() argument 159 nested_cpu_has_mtf(struct vmcs12 *vmcs12) nested_cpu_has_mtf() argument 164 nested_cpu_has_ept(struct vmcs12 *vmcs12) nested_cpu_has_ept() argument 169 nested_cpu_has_xsaves(struct vmcs12 *vmcs12) nested_cpu_has_xsaves() argument 174 nested_cpu_has_pml(struct vmcs12 *vmcs12) nested_cpu_has_pml() argument 179 nested_cpu_has_virt_x2apic_mode(struct vmcs12 *vmcs12) nested_cpu_has_virt_x2apic_mode() argument 184 nested_cpu_has_vpid(struct vmcs12 *vmcs12) nested_cpu_has_vpid() argument 189 nested_cpu_has_apic_reg_virt(struct vmcs12 *vmcs12) nested_cpu_has_apic_reg_virt() argument 194 nested_cpu_has_vid(struct vmcs12 *vmcs12) nested_cpu_has_vid() argument 199 nested_cpu_has_posted_intr(struct vmcs12 *vmcs12) nested_cpu_has_posted_intr() argument 204 nested_cpu_has_vmfunc(struct vmcs12 *vmcs12) nested_cpu_has_vmfunc() argument 209 nested_cpu_has_eptp_switching(struct vmcs12 *vmcs12) nested_cpu_has_eptp_switching() argument 216 nested_cpu_has_shadow_vmcs(struct vmcs12 *vmcs12) nested_cpu_has_shadow_vmcs() argument 221 nested_cpu_has_save_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_save_preemption_timer() argument 242 nested_cpu_has_encls_exit(struct vmcs12 *vmcs12) nested_cpu_has_encls_exit() argument 260 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_guest_cr0_valid() local [all...] |
H A D | nested.c | 56 #define SHADOW_FIELD_RO(x, y) { x, offsetof(struct vmcs12, y) }, 63 #define SHADOW_FIELD_RW(x, y) { x, offsetof(struct vmcs12, y) }, 381 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in nested_ept_inject_page_fault() local 403 nested_ept_invalidate_addr(vcpu, vmcs12->ept_pointer, in nested_ept_inject_page_fault() 408 vmcs12->guest_physical_address = fault->address; in nested_ept_inject_page_fault() 441 static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12, in nested_vmx_is_page_fault_vmexit() argument 446 bit = (vmcs12->exception_bitmap & (1u << PF_VECTOR)) != 0; in nested_vmx_is_page_fault_vmexit() 448 (error_code & vmcs12 in nested_vmx_is_page_fault_vmexit() 456 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_is_exception_vmexit() local 471 nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_io_bitmap_controls() argument 484 nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_msr_bitmap_controls() argument 496 nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_tpr_shadow_controls() argument 568 nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_prepare_msr_bitmap() argument 670 nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_cache_shadow_vmcs12() argument 689 nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_flush_cached_shadow_vmcs12() argument 718 nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apic_access_controls() argument 728 nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apicv_controls() argument 786 nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_exit_msr_switch_controls() argument 800 nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_entry_msr_switch_controls() argument 811 nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_pml_controls() argument 824 nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_unrestricted_guest_controls() argument 833 nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_mode_based_ept_exec_controls() argument 842 nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_shadow_vmcs_controls() argument 1026 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_msr_store_list_has_msr() local 1130 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_has_guest_tlb_tag() local 1136 nested_vmx_transition_tlb_flush(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool is_vmenter) nested_vmx_transition_tlb_flush() argument 1521 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_shadow_to_vmcs12() local 1556 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_vmcs12_to_shadow() local 1581 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_enlightened_to_vmcs12() local 1825 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_vmcs12_to_enlightened() local 2066 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_handle_enlightened_vmptrld() local 2114 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_calc_preemption_timer_value() local 2152 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) nested_vmx_calc_efer() argument 2224 prepare_vmcs02_early_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_early_rare() argument 2239 prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, struct vmcs12 *vmcs12) prepare_vmcs02_early() argument 2404 prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_rare() argument 2533 prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool from_vmentry, enum vm_entry_failure_code *entry_failure_code) prepare_vmcs02() argument 2674 nested_vmx_check_nmi_controls(struct vmcs12 *vmcs12) nested_vmx_check_nmi_controls() argument 2735 nested_check_vm_execution_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_execution_controls() argument 2794 nested_check_vm_exit_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_exit_controls() argument 2811 nested_check_vm_entry_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_entry_controls() argument 2883 nested_vmx_check_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_controls() argument 2897 nested_vmx_check_address_space_size(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_address_space_size() argument 2908 nested_vmx_check_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_host_state() argument 2977 nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_vmcs_link_ptr() argument 3010 nested_check_guest_non_reg_state(struct vmcs12 *vmcs12) nested_check_guest_non_reg_state() argument 3020 nested_vmx_check_guest_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) nested_vmx_check_guest_state() argument 3194 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_get_vmcs12_pages() local 3307 struct vmcs12 *vmcs12; nested_vmx_write_pml_buffer() local 3389 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_enter_non_root_mode() local 3552 struct vmcs12 *vmcs12; nested_vmx_run() local 3708 vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr0() argument 3718 vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr4() argument 3727 vmcs12_save_pending_event(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info) vmcs12_save_pending_event() argument 3796 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_mark_vmcs12_pages_dirty() local 3861 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_inject_exception_vmexit() local 4254 sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12_rare() argument 4299 copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) copy_vmcs02_to_vmcs12_rare() argument 4328 sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12() argument 4409 prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info, unsigned long exit_qualification) prepare_vmcs12() argument 4465 load_vmcs12_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) load_vmcs12_host_state() argument 4615 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_restore_host_state() local 4727 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_vmexit() local 5345 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmread() local 5451 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmwrite() local 5837 nested_vmx_eptp_switching(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_eptp_switching() argument 5873 struct vmcs12 *vmcs12; handle_vmfunc() local 5928 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_io_bitmaps() local 5958 nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_io() argument 5982 nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, union vmx_exit_reason exit_reason) nested_vmx_exit_handled_msr() argument 6020 nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_cr() argument 6090 nested_vmx_exit_handled_encls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_encls() argument 6105 nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, gpa_t bitmap) nested_vmx_exit_handled_vmcs_access() argument 6129 nested_vmx_exit_handled_mtf(struct vmcs12 *vmcs12) nested_vmx_exit_handled_mtf() argument 6229 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_l1_wants_exit() local 6393 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_reflect_vmexit() local 6410 struct vmcs12 *vmcs12; vmx_get_nested_state() local 6542 struct vmcs12 *vmcs12; vmx_set_nested_state() local [all...] |
H A D | sgx.h | 18 void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12); 26 struct vmcs12 *vmcs12) in vmx_write_encls_bitmap() 25 vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmx_write_encls_bitmap() argument
|
H A D | hyperv.c | 574 int nested_evmcs_check_controls(struct vmcs12 *vmcs12) in nested_evmcs_check_controls() argument 577 vmcs12->pin_based_vm_exec_control))) in nested_evmcs_check_controls() 581 vmcs12->cpu_based_vm_exec_control))) in nested_evmcs_check_controls() 585 vmcs12->secondary_vm_exec_control))) in nested_evmcs_check_controls() 589 vmcs12->vm_exit_controls))) in nested_evmcs_check_controls() 593 vmcs12->vm_entry_controls))) in nested_evmcs_check_controls() 601 if (WARN_ON_ONCE(vmcs12->vm_function_control >> 32)) in nested_evmcs_check_controls() 605 vmcs12->vm_function_control))) in nested_evmcs_check_controls()
|
H A D | vmcs12.h | 10 * struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a 11 * single nested guest (L2), hence the name vmcs12. Any VMX implementation has 12 * a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is 27 struct __packed vmcs12 { struct 192 * layout of struct vmcs12 is changed. MSR_IA32_VMX_BASIC returns this id, and 202 * and any VMCS region. Although only sizeof(struct vmcs12) are used by the 209 * For save/restore compatibility, the vmcs12 field offsets must not change. 212 ASSERT_STRUCT_OFFSET(struct vmcs12, field, loc) 385 static inline u64 vmcs12_read_any(struct vmcs12 *vmcs12, unsigne argument 405 vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field, u16 offset, u64 field_value) vmcs12_write_any() argument [all...] |
H A D | sgx.c | 467 void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) in vmx_write_encls_bitmap() argument 503 if (!vmcs12 && is_guest_mode(vcpu)) in vmx_write_encls_bitmap() 504 vmcs12 = get_vmcs12(vcpu); in vmx_write_encls_bitmap() 505 if (vmcs12 && nested_cpu_has_encls_exit(vmcs12)) in vmx_write_encls_bitmap() 506 bitmap |= vmcs12->encls_exiting_bitmap; in vmx_write_encls_bitmap()
|
H A D | hyperv.h | 15 #include "vmcs12.h" 61 * is 'struct vmcs12' or 'struct hv_enlightened_vmcs' as it takes in evmcs_read_any() 196 int nested_evmcs_check_controls(struct vmcs12 *vmcs12);
|
H A D | vmx.c | 65 #include "vmcs12.h" 1756 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_update_emulated_instruction() local 1773 if (nested_cpu_has_mtf(vmcs12) && in vmx_update_emulated_instruction() 1906 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_get_l2_tsc_offset() local 1908 if (nested_cpu_has(vmcs12, CPU_BASED_USE_TSC_OFFSETTING)) in vmx_get_l2_tsc_offset() 1909 return vmcs12->tsc_offset; in vmx_get_l2_tsc_offset() 1916 struct vmcs12 *vmcs12 in vmx_get_l2_tsc_multiplier() local 5397 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr0() local 5423 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr4() local 6681 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_update_cr8_intercept() local 7925 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept_io() local 7961 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept() local [all...] |
H A D | vmx.h | 129 struct vmcs12 *cached_vmcs12; 135 struct vmcs12 *cached_shadow_vmcs12; 138 * GPA to HVA cache for accessing vmcs12->vmcs_link_pointer 149 * with the data held by struct vmcs12.
|
H A D | vmcs12.c | 4 #include "vmcs12.h" 6 #define VMCS12_OFFSET(x) offsetof(struct vmcs12, x)
|
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/ |
H A D | nested.h | 6 #include "vmcs12.h" 40 static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu) in get_vmcs12() 45 static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu) in get_shadow_vmcs12() 63 * have vmcs12 if it is true. in vmx_has_valid_vmcs12() 92 static inline unsigned long nested_read_cr0(struct vmcs12 *fields) in nested_read_cr0() 97 static inline unsigned long nested_read_cr4(struct vmcs12 *fields) in nested_read_cr4() 136 static inline bool nested_cpu_has(struct vmcs12 *vmcs12, u32 bit) in nested_cpu_has() argument 138 return vmcs12->cpu_based_vm_exec_control & bit; in nested_cpu_has() 141 static inline bool nested_cpu_has2(struct vmcs12 *vmcs1 argument 148 nested_cpu_has_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_preemption_timer() argument 154 nested_cpu_has_nmi_exiting(struct vmcs12 *vmcs12) nested_cpu_has_nmi_exiting() argument 159 nested_cpu_has_virtual_nmis(struct vmcs12 *vmcs12) nested_cpu_has_virtual_nmis() argument 164 nested_cpu_has_mtf(struct vmcs12 *vmcs12) nested_cpu_has_mtf() argument 169 nested_cpu_has_ept(struct vmcs12 *vmcs12) nested_cpu_has_ept() argument 174 nested_cpu_has_xsaves(struct vmcs12 *vmcs12) nested_cpu_has_xsaves() argument 179 nested_cpu_has_pml(struct vmcs12 *vmcs12) nested_cpu_has_pml() argument 184 nested_cpu_has_virt_x2apic_mode(struct vmcs12 *vmcs12) nested_cpu_has_virt_x2apic_mode() argument 189 nested_cpu_has_vpid(struct vmcs12 *vmcs12) nested_cpu_has_vpid() argument 194 nested_cpu_has_apic_reg_virt(struct vmcs12 *vmcs12) nested_cpu_has_apic_reg_virt() argument 199 nested_cpu_has_vid(struct vmcs12 *vmcs12) nested_cpu_has_vid() argument 204 nested_cpu_has_posted_intr(struct vmcs12 *vmcs12) nested_cpu_has_posted_intr() argument 209 nested_cpu_has_vmfunc(struct vmcs12 *vmcs12) nested_cpu_has_vmfunc() argument 214 nested_cpu_has_eptp_switching(struct vmcs12 *vmcs12) nested_cpu_has_eptp_switching() argument 221 nested_cpu_has_shadow_vmcs(struct vmcs12 *vmcs12) nested_cpu_has_shadow_vmcs() argument 226 nested_cpu_has_save_preemption_timer(struct vmcs12 *vmcs12) nested_cpu_has_save_preemption_timer() argument 260 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_guest_cr0_valid() local [all...] |
H A D | nested.c | 59 #define SHADOW_FIELD_RO(x, y) { x, offsetof(struct vmcs12, y) }, 66 #define SHADOW_FIELD_RW(x, y) { x, offsetof(struct vmcs12, y) }, 336 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in nested_ept_inject_page_fault() local 351 vmcs12->guest_physical_address = fault->address; in nested_ept_inject_page_fault() 377 static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12, in nested_vmx_is_page_fault_vmexit() argument 382 bit = (vmcs12->exception_bitmap & (1u << PF_VECTOR)) != 0; in nested_vmx_is_page_fault_vmexit() 384 (error_code & vmcs12->page_fault_error_code_mask) != in nested_vmx_is_page_fault_vmexit() 385 vmcs12 in nested_vmx_is_page_fault_vmexit() 396 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_exception() local 431 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_inject_page_fault_nested() local 447 nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_io_bitmap_controls() argument 460 nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_msr_bitmap_controls() argument 472 nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_tpr_shadow_controls() argument 564 nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_prepare_msr_bitmap() argument 663 nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_cache_shadow_vmcs12() argument 682 nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_flush_cached_shadow_vmcs12() argument 705 nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apic_access_controls() argument 715 nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_apicv_controls() argument 776 nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_exit_msr_switch_controls() argument 790 nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_entry_msr_switch_controls() argument 801 nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_pml_controls() argument 814 nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_unrestricted_guest_controls() argument 823 nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_mode_based_ept_exec_controls() argument 832 nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_shadow_vmcs_controls() argument 1016 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_msr_store_list_has_msr() local 1183 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_has_guest_tlb_tag() local 1189 nested_vmx_transition_tlb_flush(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, bool is_vmenter) nested_vmx_transition_tlb_flush() argument 1558 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_shadow_to_vmcs12() local 1593 struct vmcs12 *vmcs12 = get_vmcs12(&vmx->vcpu); copy_vmcs12_to_shadow() local 1618 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_enlightened_to_vmcs12() local 1834 struct vmcs12 *vmcs12 = vmx->nested.cached_vmcs12; copy_vmcs12_to_enlightened() local 2064 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_handle_enlightened_vmptrld() local 2113 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_calc_preemption_timer_value() local 2151 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) nested_vmx_calc_efer() argument 2220 prepare_vmcs02_early_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_early_rare() argument 2235 prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, struct vmcs12 *vmcs12) prepare_vmcs02_early() argument 2398 prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) prepare_vmcs02_rare() argument 2527 prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) prepare_vmcs02() argument 2649 nested_vmx_check_nmi_controls(struct vmcs12 *vmcs12) nested_vmx_check_nmi_controls() argument 2711 nested_check_vm_execution_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_execution_controls() argument 2770 nested_check_vm_exit_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_exit_controls() argument 2787 nested_check_vm_entry_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_check_vm_entry_controls() argument 2859 nested_vmx_check_controls(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_controls() argument 2873 nested_vmx_check_address_space_size(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_address_space_size() argument 2884 nested_vmx_check_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_host_state() argument 2959 nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_check_vmcs_link_ptr() argument 2988 nested_check_guest_non_reg_state(struct vmcs12 *vmcs12) nested_check_guest_non_reg_state() argument 2997 nested_vmx_check_guest_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, enum vm_entry_failure_code *entry_failure_code) nested_vmx_check_guest_state() argument 3164 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_get_vmcs12_pages() local 3263 struct vmcs12 *vmcs12; nested_vmx_write_pml_buffer() local 3345 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_enter_non_root_mode() local 3507 struct vmcs12 *vmcs12; nested_vmx_run() local 3649 vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr0() argument 3659 vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) vmcs12_guest_cr4() argument 3668 vmcs12_save_pending_event(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info) vmcs12_save_pending_event() argument 3737 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_mark_vmcs12_pages_dirty() local 3792 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_inject_exception_vmexit() local 4016 sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12_rare() argument 4063 copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) copy_vmcs02_to_vmcs12_rare() argument 4092 sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) sync_vmcs02_to_vmcs12() argument 4170 prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, u32 vm_exit_reason, u32 exit_intr_info, unsigned long exit_qualification) prepare_vmcs12() argument 4224 load_vmcs12_host_state(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) load_vmcs12_host_state() argument 4373 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_restore_host_state() local 4488 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_vmexit() local 5103 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmread() local 5187 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) handle_vmwrite() local 5572 nested_vmx_eptp_switching(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_eptp_switching() argument 5608 struct vmcs12 *vmcs12; handle_vmfunc() local 5654 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_check_io_bitmaps() local 5684 nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_io() argument 5708 nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, union vmx_exit_reason exit_reason) nested_vmx_exit_handled_msr() argument 5746 nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12) nested_vmx_exit_handled_cr() argument 5816 nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12, gpa_t bitmap) nested_vmx_exit_handled_vmcs_access() argument 5840 nested_vmx_exit_handled_mtf(struct vmcs12 *vmcs12) nested_vmx_exit_handled_mtf() argument 5929 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_l1_wants_exit() local 6088 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); nested_vmx_reflect_vmexit() local 6105 struct vmcs12 *vmcs12; vmx_get_nested_state() local 6232 struct vmcs12 *vmcs12; vmx_set_nested_state() local [all...] |
H A D | vmcs12.h | 10 * struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a 11 * single nested guest (L2), hence the name vmcs12. Any VMX implementation has 12 * a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is 27 struct __packed vmcs12 { struct 190 * layout of struct vmcs12 is changed. MSR_IA32_VMX_BASIC returns this id, and 200 * and any VMCS region. Although only sizeof(struct vmcs12) are used by the 213 * For save/restore compatibility, the vmcs12 field offsets must not change. 216 BUILD_BUG_ON_MSG(offsetof(struct vmcs12, field) != (loc), \ 217 "Offset of " #field " in struct vmcs12 has changed.") 392 static inline u64 vmcs12_read_any(struct vmcs12 *vmcs1 argument 412 vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field, u16 offset, u64 field_value) vmcs12_write_any() argument [all...] |
H A D | evmcs.c | 376 int nested_evmcs_check_controls(struct vmcs12 *vmcs12) in nested_evmcs_check_controls() argument 381 unsupp_ctl = vmcs12->pin_based_vm_exec_control & in nested_evmcs_check_controls() 390 unsupp_ctl = vmcs12->secondary_vm_exec_control & in nested_evmcs_check_controls() 399 unsupp_ctl = vmcs12->vm_exit_controls & in nested_evmcs_check_controls() 408 unsupp_ctl = vmcs12->vm_entry_controls & in nested_evmcs_check_controls() 417 unsupp_ctl = vmcs12->vm_function_control & EVMCS1_UNSUPPORTED_VMFUNC; in nested_evmcs_check_controls()
|
H A D | evmcs.h | 13 #include "vmcs12.h" 203 int nested_evmcs_check_controls(struct vmcs12 *vmcs12);
|
H A D | vmx.c | 63 #include "vmcs12.h" 1695 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_update_emulated_instruction() local 1708 if (nested_cpu_has_mtf(vmcs12) && in vmx_update_emulated_instruction() 1831 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); in vmx_write_l1_tsc_offset() local 1841 (vmcs12->cpu_based_vm_exec_control & CPU_BASED_USE_TSC_OFFSETTING)) in vmx_write_l1_tsc_offset() 1842 g_tsc_offset = vmcs12->tsc_offset; in vmx_write_l1_tsc_offset() 5086 struct vmcs12 *vmcs12 in handle_set_cr0() local 5119 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); handle_set_cr4() local 6304 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); update_cr8_intercept() local 7491 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept_io() local 7527 struct vmcs12 *vmcs12 = get_vmcs12(vcpu); vmx_check_intercept() local [all...] |
H A D | vmx.h | 114 struct vmcs12 *cached_vmcs12; 120 struct vmcs12 *cached_shadow_vmcs12; 124 * with the data held by struct vmcs12.
|
H A D | vmcs12.c | 3 #include "vmcs12.h" 6 #define VMCS12_OFFSET(x) offsetof(struct vmcs12, x)
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | Makefile | 21 kvm-intel-y += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o \
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | Makefile | 25 kvm-intel-y += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o \
|
/kernel/linux/linux-5.10/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 426 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
|
/kernel/linux/linux-5.10/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 426 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/asm-x86/asm/ |
H A D | kvm.h | 351 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/asm-x86/asm/ |
H A D | kvm.h | 351 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
|
/kernel/linux/linux-6.6/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 464 __u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE]; member
|