/kernel/linux/linux-5.10/arch/loongarch/kvm/ |
H A D | hypcall.c | 95 kvm_make_request(KVM_REQ_EVENT, dst_vcpu); in kvm_pv_kick_cpu()
|
H A D | emulate.c | 41 if (kvm_check_request(KVM_REQ_EVENT, vcpu)) { in _kvm_emu_idle()
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | lapic.c | 745 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in apic_update_ppr() 1104 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1112 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1133 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1144 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1261 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in apic_set_eoi() 1276 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in kvm_apic_set_eoi_accelerated() 2546 * KVM_REQ_EVENT. in kvm_get_apic_interrupt() 2554 * triggered KVM_REQ_EVENT already. in kvm_get_apic_interrupt() 2637 kvm_make_request(KVM_REQ_EVENT, vcp in kvm_apic_set_state() [all...] |
H A D | x86.c | 549 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_multiple_exception() 4241 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_interrupt() 4256 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_interrupt() 4521 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_x86_set_vcpu_events() 7161 kvm_make_request(KVM_REQ_EVENT, vcpu); in toggle_interruptibility() 7410 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_smm_changed() 7728 * For STI, interrupts are shadowed; so KVM_REQ_EVENT will in x86_emulate_instruction() 7734 kvm_make_request(KVM_REQ_EVENT, vcpu); in x86_emulate_instruction() 8599 * will re-request KVM_REQ_EVENT. Sometimes however an event is pending in inject_pending_event() 8675 kvm_make_request(KVM_REQ_EVENT, vcp in process_nmi() [all...] |
H A D | i8259.c | 62 kvm_make_request(KVM_REQ_EVENT, vcpu);
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | lapic.c | 972 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in apic_update_ppr() 1335 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1357 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1368 kvm_make_request(KVM_REQ_EVENT, vcpu); in __apic_accept_irq() 1486 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in apic_set_eoi() 1501 kvm_make_request(KVM_REQ_EVENT, apic->vcpu); in kvm_apic_set_eoi_accelerated() 2915 * KVM_REQ_EVENT. in kvm_get_apic_interrupt() 2923 * triggered KVM_REQ_EVENT already. in kvm_get_apic_interrupt() 3020 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_apic_set_state()
|
H A D | smm.c | 122 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_smm_changed() 138 kvm_make_request(KVM_REQ_EVENT, vcpu); in process_smi()
|
H A D | x86.c | 653 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_multiple_exception() 4993 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_interrupt() 5008 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_interrupt() 5354 kvm_make_request(KVM_REQ_EVENT, vcpu); in kvm_vcpu_ioctl_x86_set_vcpu_events() 8389 kvm_make_request(KVM_REQ_EVENT, vcpu); in toggle_interruptibility() 9053 * For STI, interrupts are shadowed; so KVM_REQ_EVENT will in x86_emulate_instruction() 9059 kvm_make_request(KVM_REQ_EVENT, vcpu); in x86_emulate_instruction() 10190 * will re-request KVM_REQ_EVENT. Sometimes however an event is pending in kvm_check_and_inject_events() 10308 kvm_make_request(KVM_REQ_EVENT, vcpu); in process_nmi() 10353 * pending. At the same time, KVM_REQ_EVENT ma in __kvm_vcpu_update_apicv() [all...] |
H A D | i8259.c | 64 kvm_make_request(KVM_REQ_EVENT, vcpu);
|
/kernel/linux/linux-5.10/arch/loongarch/include/asm/ |
H A D | kvm_host.h | 43 #define KVM_REQ_EVENT KVM_ARCH_REQ(4) macro
|
/kernel/linux/linux-5.10/arch/x86/kvm/svm/ |
H A D | svm.c | 1845 kvm_make_request(KVM_REQ_EVENT, vcpu); in db_interception() 2113 * again while processing KVM_REQ_EVENT if needed. in svm_set_gif() 2124 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in svm_set_gif() 2281 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in iret_interception() 2807 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in interrupt_window_interception() 3425 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in svm_complete_interrupts() 3435 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in svm_complete_interrupts()
|
/kernel/linux/linux-6.6/arch/x86/kvm/svm/ |
H A D | svm.c | 2098 kvm_make_request(KVM_REQ_EVENT, vcpu); in db_interception() 2431 * again while processing KVM_REQ_EVENT if needed. in svm_set_gif() 2443 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in svm_set_gif() 2596 kvm_make_request(KVM_REQ_EVENT, vcpu); in iret_interception() 3197 kvm_make_request(KVM_REQ_EVENT, vcpu); in interrupt_window_interception() 3656 kvm_make_request(KVM_REQ_EVENT, vcpu); in svm_complete_interrupt_delivery() 4039 kvm_make_request(KVM_REQ_EVENT, vcpu); in svm_complete_interrupts() 4049 kvm_make_request(KVM_REQ_EVENT, vcpu); in svm_complete_interrupts()
|
H A D | nested.c | 657 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in nested_vmcb02_prepare_control() 1046 * virtual interrupt masking). Raise KVM_REQ_EVENT to ensure that in nested_svm_vmexit() 1060 kvm_make_request(KVM_REQ_EVENT, &svm->vcpu); in nested_svm_vmexit()
|
H A D | sev.c | 2903 kvm_make_request(KVM_REQ_EVENT, vcpu); in sev_handle_vmgexit()
|
/kernel/linux/linux-6.6/arch/x86/kvm/vmx/ |
H A D | vmx.c | 1779 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_update_emulated_instruction() 4222 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_deliver_nested_posted_interrupt() 4286 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_deliver_interrupt() 5626 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_interrupt_window() 5828 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_nmi_window() 5854 if (kvm_test_request(KVM_REQ_EVENT, vcpu)) in handle_invalid_guest_state() 6892 * delivery (RVI) or KVM_REQ_EVENT. Virtual interrupt delivery is in vmx_sync_pir_to_irr() 6896 * wants to exit on interrupts, KVM_REQ_EVENT is needed to synthesize a in vmx_sync_pir_to_irr() 6899 * interrupts into L2, and so KVM_REQ_EVENT is again needed. in vmx_sync_pir_to_irr() 6908 kvm_make_request(KVM_REQ_EVENT, vcp in vmx_sync_pir_to_irr() [all...] |
H A D | nested.c | 2105 kvm_make_request(KVM_REQ_EVENT, &vmx->vcpu); in vmx_preemption_timer_fn() 3502 kvm_make_request(KVM_REQ_EVENT, vcpu); in nested_vmx_enter_non_root_mode() 3636 kvm_make_request(KVM_REQ_EVENT, vcpu); in nested_vmx_run() 5270 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_vmxoff() 6713 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_set_nested_state()
|
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/ |
H A D | nested.c | 2104 kvm_make_request(KVM_REQ_EVENT, &vmx->vcpu); in vmx_preemption_timer_fn() 3454 * to perform pending event evaluation by requesting a KVM_REQ_EVENT. in nested_vmx_enter_non_root_mode() 3457 kvm_make_request(KVM_REQ_EVENT, vcpu); in nested_vmx_enter_non_root_mode() 3587 kvm_make_request(KVM_REQ_EVENT, vcpu); in nested_vmx_run() 5038 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_vmoff()
|
H A D | vmx.c | 4092 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_deliver_nested_posted_interrupt() 5308 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_interrupt_window() 5534 kvm_make_request(KVM_REQ_EVENT, vcpu); in handle_nmi_window() 5552 if (kvm_test_request(KVM_REQ_EVENT, vcpu)) in handle_invalid_guest_state() 6468 * processed via KVM_REQ_EVENT, not RVI, because we do not use in vmx_sync_pir_to_irr() 6472 kvm_make_request(KVM_REQ_EVENT, vcpu); in vmx_sync_pir_to_irr() 6629 kvm_make_request(KVM_REQ_EVENT, vcpu); in __vmx_complete_interrupts()
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | kvm_host.h | 62 #define KVM_REQ_EVENT KVM_ARCH_REQ(6) macro
|
/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | kvm_host.h | 79 #define KVM_REQ_EVENT KVM_ARCH_REQ(6) macro
|