/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | pmu.c | 218 config = kvm_x86_ops.pmu_ops->pmc_perf_hw_id(pmc); in reprogram_gp_counter() 270 kvm_x86_ops.pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter() 279 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter() 301 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event() 323 return kvm_x86_ops.pmu_ops->is_valid_rdpmc_ecx(vcpu, idx); in kvm_pmu_is_valid_rdpmc_ecx() 373 pmc = kvm_x86_ops.pmu_ops->rdpmc_ecx_to_pmc(vcpu, idx, &mask); in kvm_pmu_rdpmc() 378 (kvm_x86_ops.get_cpl(vcpu) != 0) && in kvm_pmu_rdpmc() 394 return kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr) || in kvm_pmu_is_valid_msr() 395 kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr() 401 struct kvm_pmc *pmc = kvm_x86_ops in kvm_pmu_mark_pmc_in_use() [all...] |
H A D | x86.c | 114 struct kvm_x86_ops kvm_x86_ops __read_mostly; 115 EXPORT_SYMBOL_GPL(kvm_x86_ops); variable 716 if (kvm_x86_ops.get_cpl(vcpu) <= required_cpl) in kvm_require_cpl() 858 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in kvm_set_cr0() 871 kvm_x86_ops.set_cr0(vcpu, cr0); in kvm_set_cr0() 981 if (kvm_x86_ops.get_cpl(vcpu) != 0 || in kvm_set_xcr() 998 if (!kvm_x86_ops.is_valid_cr4(vcpu, cr4)) in kvm_valid_cr4() 1035 kvm_x86_ops.set_cr4(vcpu, cr4); in kvm_set_cr4() 1123 kvm_x86_ops in kvm_update_dr7() [all...] |
H A D | kvm_cache_regs.h | 71 kvm_x86_ops.cache_reg(vcpu, reg); in kvm_register_read() 111 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_PDPTR); in kvm_pdptr_read() 121 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR0); in kvm_read_cr0_bits() 135 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR4); in kvm_read_cr4_bits() 142 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR3); in kvm_read_cr3()
|
H A D | mmu.h | 98 kvm_x86_ops.load_mmu_pgd(vcpu, root_hpa | kvm_get_active_pcid(vcpu), in kvm_mmu_load_pgd() 170 int cpl = kvm_x86_ops.get_cpl(vcpu); in permission_fault() 171 unsigned long rflags = kvm_x86_ops.get_rflags(vcpu); in permission_fault()
|
H A D | irq.c | 146 if (kvm_x86_ops.migrate_timers) in __kvm_migrate_timers() 147 kvm_x86_ops.migrate_timers(vcpu); in __kvm_migrate_timers()
|
H A D | lapic.c | 122 return kvm_x86_ops.set_hv_timer in kvm_can_use_hv_timer() 492 kvm_x86_ops.hwapic_irr_update(vcpu, in apic_clear_irr() 523 kvm_x86_ops.hwapic_isr_update(vcpu, vec); in apic_set_isr() 571 kvm_x86_ops.hwapic_isr_update(vcpu, in apic_clear_isr() 709 highest_irr = kvm_x86_ops.sync_pir_to_irr(apic->vcpu); in apic_has_interrupt_for_ppr() 1102 if (kvm_x86_ops.deliver_posted_interrupt(vcpu, vector)) { in __apic_accept_irq() 1838 kvm_x86_ops.cancel_hv_timer(apic->vcpu); in cancel_hv_timer() 1855 if (kvm_x86_ops.set_hv_timer(vcpu, ktimer->tscdeadline, &expired)) in start_hv_timer() 2287 kvm_x86_ops.set_virtual_apic_mode(vcpu); in kvm_lapic_set_base() 2364 kvm_x86_ops in kvm_lapic_reset() [all...] |
H A D | x86.h | 100 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in is_64_bit_mode() 131 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_vcpu_flush_tlb_current() 241 return is_smm(vcpu) || kvm_x86_ops.apic_init_signal_blocked(vcpu); in kvm_vcpu_latch_init()
|
H A D | pmu.h | 92 return kvm_x86_ops.pmu_ops->pmc_is_enabled(pmc); in pmc_is_enabled()
|
H A D | hyperv.c | 1165 kvm_x86_ops.patch_hypercall(vcpu, instructions); in kvm_hv_set_msr_pw() 1762 if (kvm_x86_ops.get_cpl(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall() 1986 if (kvm_x86_ops.nested_ops->get_evmcs_version) in kvm_vcpu_ioctl_get_hv_cpuid() 1987 evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu); in kvm_vcpu_ioctl_get_hv_cpuid()
|
H A D | trace.h | 259 kvm_x86_ops.get_exit_info(vcpu, &__entry->info1, \ 745 __entry->csbase = kvm_x86_ops.get_segment_base(vcpu, VCPU_SREG_CS);
|
H A D | cpuid.c | 184 kvm_x86_ops.vcpu_after_set_cpuid(vcpu); in kvm_vcpu_after_set_cpuid()
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | kvm_host.h | 1093 struct kvm_x86_ops { struct 1319 struct kvm_x86_ops *runtime_ops; 1331 extern struct kvm_x86_ops kvm_x86_ops; 1336 return __vmalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_arch_alloc_vm() 1343 if (kvm_x86_ops.tlb_remote_flush && in kvm_arch_flush_remote_tlb() 1344 !kvm_x86_ops.tlb_remote_flush(kvm)) in kvm_arch_flush_remote_tlb() 1431 * kvm_x86_ops.skip_emulated_instruction() implementations. 1731 if (kvm_x86_ops.vcpu_blocking) in kvm_arch_vcpu_blocking() 1732 kvm_x86_ops in kvm_arch_vcpu_blocking() [all...] |
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | kvm_onhyperv.c | 116 if (kvm_x86_ops.flush_remote_tlbs == hv_flush_remote_tlbs) { in hv_track_root_tdp()
|
H A D | x86.c | 136 struct kvm_x86_ops kvm_x86_ops __read_mostly; 140 *(((struct kvm_x86_ops *)0)->func)); 643 kvm_x86_ops.nested_ops->leave_nested(vcpu); in kvm_leave_nested() 663 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, nr, error_code)) { in kvm_multiple_exception() 4602 r = kvm_x86_ops.nested_ops->get_state ? in kvm_vm_ioctl_check_extension() 4603 kvm_x86_ops.nested_ops->get_state(NULL, NULL, 0) : 0; in kvm_vm_ioctl_check_extension() 4606 r = kvm_x86_ops.enable_l2_tlb_flush != NULL; in kvm_vm_ioctl_check_extension() 4609 r = kvm_x86_ops.nested_ops->enable_evmcs != NULL; in kvm_vm_ioctl_check_extension() 5619 if (!kvm_x86_ops in kvm_vcpu_ioctl_enable_cap() [all...] |
H A D | hyperv.c | 2344 kvm_x86_ops.nested_ops->hv_inject_synthetic_vmexit_post_tlb_flush(vcpu); in kvm_hv_hypercall_complete() 2728 if (kvm_x86_ops.nested_ops->get_evmcs_version) in kvm_get_hv_cpuid() 2729 evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu); in kvm_get_hv_cpuid()
|
H A D | lapic.c | 156 return kvm_x86_ops.set_hv_timer in kvm_can_use_hv_timer() 935 if (kvm_x86_ops.sync_pir_to_irr) in apic_has_interrupt_for_ppr()
|
/kernel/linux/linux-6.6/arch/x86/kvm/svm/ |
H A D | svm_onhyperv.h | 16 static struct kvm_x86_ops svm_x86_ops;
|
/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | kvm_host.h | 1551 struct kvm_x86_ops { struct 1781 struct kvm_x86_ops *runtime_ops; 1796 extern struct kvm_x86_ops kvm_x86_ops; 1799 DECLARE_STATIC_CALL(kvm_x86_##func, *(((struct kvm_x86_ops *)0)->func)); 1810 return __vmalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_arch_alloc_vm() 1819 if (kvm_x86_ops.flush_remote_tlbs && in kvm_arch_flush_remote_tlbs() 1894 * kvm_x86_ops.skip_emulated_instruction() implementations if
|
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
H A D | spte.c | 123 spte |= kvm_x86_ops.get_mt_mask(vcpu, gfn, in make_spte()
|
H A D | mmu.c | 186 return kvm_x86_ops.tlb_remote_flush_with_range; in kvm_available_flush_tlb_with_range() 194 if (range && kvm_x86_ops.tlb_remote_flush_with_range) in kvm_flush_remote_tlbs_with_range() 195 ret = kvm_x86_ops.tlb_remote_flush_with_range(kvm, range); in kvm_flush_remote_tlbs_with_range() 1186 * Similar to the !kvm_x86_ops.slot_disable_log_dirty case, in spte_set_dirty() 1281 if (kvm_x86_ops.enable_log_dirty_pt_masked) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1282 kvm_x86_ops.enable_log_dirty_pt_masked(kvm, slot, gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked() 4856 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_mmu_load() 5137 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invalidate_gva() 5196 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invpcid_gva()
|
H A D | paging_tmpl.h | 260 if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr)) in update_accessed_dirty_bits()
|
/kernel/linux/linux-6.6/arch/x86/kvm/mmu/ |
H A D | mmu_internal.h | 157 return kvm_x86_ops.cpu_dirty_log_size && sp->role.guest_mode; in kvm_mmu_page_ad_need_write_protect()
|
H A D | paging_tmpl.h | 225 if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr)) in update_accessed_dirty_bits()
|
H A D | mmu.c | 274 return kvm_x86_ops.flush_remote_tlbs_range; in kvm_available_flush_remote_tlbs_range() 279 if (!kvm_x86_ops.flush_remote_tlbs_range) in kvm_arch_flush_remote_tlbs_range() 1397 if (kvm_x86_ops.cpu_dirty_log_size) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1405 return kvm_x86_ops.cpu_dirty_log_size; in kvm_cpu_dirty_log_size() 1586 if (kvm_x86_ops.set_apic_access_page_addr && in kvm_unmap_gfn_range()
|
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/ |
H A D | vmx.c | 3189 if (kvm_x86_ops.tlb_remote_flush) { in vmx_load_mmu_pgd() 7696 if (kvm_x86_ops.set_hv_timer) in vmx_post_block() 7792 static struct kvm_x86_ops vmx_x86_ops __initdata = {
|