Home
last modified time | relevance | path

Searched refs:kvm_x86_ops (Results 1 - 25 of 29) sorted by relevance

12

/kernel/linux/linux-5.10/arch/x86/kvm/
H A Dpmu.c218 config = kvm_x86_ops.pmu_ops->pmc_perf_hw_id(pmc); in reprogram_gp_counter()
270 kvm_x86_ops.pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter()
279 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter()
301 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event()
323 return kvm_x86_ops.pmu_ops->is_valid_rdpmc_ecx(vcpu, idx); in kvm_pmu_is_valid_rdpmc_ecx()
373 pmc = kvm_x86_ops.pmu_ops->rdpmc_ecx_to_pmc(vcpu, idx, &mask); in kvm_pmu_rdpmc()
378 (kvm_x86_ops.get_cpl(vcpu) != 0) && in kvm_pmu_rdpmc()
394 return kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr) || in kvm_pmu_is_valid_msr()
395 kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr()
401 struct kvm_pmc *pmc = kvm_x86_ops in kvm_pmu_mark_pmc_in_use()
[all...]
H A Dx86.c114 struct kvm_x86_ops kvm_x86_ops __read_mostly;
115 EXPORT_SYMBOL_GPL(kvm_x86_ops); variable
716 if (kvm_x86_ops.get_cpl(vcpu) <= required_cpl) in kvm_require_cpl()
858 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in kvm_set_cr0()
871 kvm_x86_ops.set_cr0(vcpu, cr0); in kvm_set_cr0()
981 if (kvm_x86_ops.get_cpl(vcpu) != 0 || in kvm_set_xcr()
998 if (!kvm_x86_ops.is_valid_cr4(vcpu, cr4)) in kvm_valid_cr4()
1035 kvm_x86_ops.set_cr4(vcpu, cr4); in kvm_set_cr4()
1123 kvm_x86_ops in kvm_update_dr7()
[all...]
H A Dkvm_cache_regs.h71 kvm_x86_ops.cache_reg(vcpu, reg); in kvm_register_read()
111 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_PDPTR); in kvm_pdptr_read()
121 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR0); in kvm_read_cr0_bits()
135 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR4); in kvm_read_cr4_bits()
142 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR3); in kvm_read_cr3()
H A Dmmu.h98 kvm_x86_ops.load_mmu_pgd(vcpu, root_hpa | kvm_get_active_pcid(vcpu), in kvm_mmu_load_pgd()
170 int cpl = kvm_x86_ops.get_cpl(vcpu); in permission_fault()
171 unsigned long rflags = kvm_x86_ops.get_rflags(vcpu); in permission_fault()
H A Dirq.c146 if (kvm_x86_ops.migrate_timers) in __kvm_migrate_timers()
147 kvm_x86_ops.migrate_timers(vcpu); in __kvm_migrate_timers()
H A Dlapic.c122 return kvm_x86_ops.set_hv_timer in kvm_can_use_hv_timer()
492 kvm_x86_ops.hwapic_irr_update(vcpu, in apic_clear_irr()
523 kvm_x86_ops.hwapic_isr_update(vcpu, vec); in apic_set_isr()
571 kvm_x86_ops.hwapic_isr_update(vcpu, in apic_clear_isr()
709 highest_irr = kvm_x86_ops.sync_pir_to_irr(apic->vcpu); in apic_has_interrupt_for_ppr()
1102 if (kvm_x86_ops.deliver_posted_interrupt(vcpu, vector)) { in __apic_accept_irq()
1838 kvm_x86_ops.cancel_hv_timer(apic->vcpu); in cancel_hv_timer()
1855 if (kvm_x86_ops.set_hv_timer(vcpu, ktimer->tscdeadline, &expired)) in start_hv_timer()
2287 kvm_x86_ops.set_virtual_apic_mode(vcpu); in kvm_lapic_set_base()
2364 kvm_x86_ops in kvm_lapic_reset()
[all...]
H A Dx86.h100 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in is_64_bit_mode()
131 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_vcpu_flush_tlb_current()
241 return is_smm(vcpu) || kvm_x86_ops.apic_init_signal_blocked(vcpu); in kvm_vcpu_latch_init()
H A Dpmu.h92 return kvm_x86_ops.pmu_ops->pmc_is_enabled(pmc); in pmc_is_enabled()
H A Dhyperv.c1165 kvm_x86_ops.patch_hypercall(vcpu, instructions); in kvm_hv_set_msr_pw()
1762 if (kvm_x86_ops.get_cpl(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall()
1986 if (kvm_x86_ops.nested_ops->get_evmcs_version) in kvm_vcpu_ioctl_get_hv_cpuid()
1987 evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu); in kvm_vcpu_ioctl_get_hv_cpuid()
H A Dtrace.h259 kvm_x86_ops.get_exit_info(vcpu, &__entry->info1, \
745 __entry->csbase = kvm_x86_ops.get_segment_base(vcpu, VCPU_SREG_CS);
H A Dcpuid.c184 kvm_x86_ops.vcpu_after_set_cpuid(vcpu); in kvm_vcpu_after_set_cpuid()
/kernel/linux/linux-5.10/arch/x86/include/asm/
H A Dkvm_host.h1093 struct kvm_x86_ops { struct
1319 struct kvm_x86_ops *runtime_ops;
1331 extern struct kvm_x86_ops kvm_x86_ops;
1336 return __vmalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_arch_alloc_vm()
1343 if (kvm_x86_ops.tlb_remote_flush && in kvm_arch_flush_remote_tlb()
1344 !kvm_x86_ops.tlb_remote_flush(kvm)) in kvm_arch_flush_remote_tlb()
1431 * kvm_x86_ops.skip_emulated_instruction() implementations.
1731 if (kvm_x86_ops.vcpu_blocking) in kvm_arch_vcpu_blocking()
1732 kvm_x86_ops in kvm_arch_vcpu_blocking()
[all...]
/kernel/linux/linux-6.6/arch/x86/kvm/
H A Dkvm_onhyperv.c116 if (kvm_x86_ops.flush_remote_tlbs == hv_flush_remote_tlbs) { in hv_track_root_tdp()
H A Dx86.c136 struct kvm_x86_ops kvm_x86_ops __read_mostly;
140 *(((struct kvm_x86_ops *)0)->func));
643 kvm_x86_ops.nested_ops->leave_nested(vcpu); in kvm_leave_nested()
663 kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, nr, error_code)) { in kvm_multiple_exception()
4602 r = kvm_x86_ops.nested_ops->get_state ? in kvm_vm_ioctl_check_extension()
4603 kvm_x86_ops.nested_ops->get_state(NULL, NULL, 0) : 0; in kvm_vm_ioctl_check_extension()
4606 r = kvm_x86_ops.enable_l2_tlb_flush != NULL; in kvm_vm_ioctl_check_extension()
4609 r = kvm_x86_ops.nested_ops->enable_evmcs != NULL; in kvm_vm_ioctl_check_extension()
5619 if (!kvm_x86_ops in kvm_vcpu_ioctl_enable_cap()
[all...]
H A Dhyperv.c2344 kvm_x86_ops.nested_ops->hv_inject_synthetic_vmexit_post_tlb_flush(vcpu); in kvm_hv_hypercall_complete()
2728 if (kvm_x86_ops.nested_ops->get_evmcs_version) in kvm_get_hv_cpuid()
2729 evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu); in kvm_get_hv_cpuid()
H A Dlapic.c156 return kvm_x86_ops.set_hv_timer in kvm_can_use_hv_timer()
935 if (kvm_x86_ops.sync_pir_to_irr) in apic_has_interrupt_for_ppr()
/kernel/linux/linux-6.6/arch/x86/kvm/svm/
H A Dsvm_onhyperv.h16 static struct kvm_x86_ops svm_x86_ops;
/kernel/linux/linux-6.6/arch/x86/include/asm/
H A Dkvm_host.h1551 struct kvm_x86_ops { struct
1781 struct kvm_x86_ops *runtime_ops;
1796 extern struct kvm_x86_ops kvm_x86_ops;
1799 DECLARE_STATIC_CALL(kvm_x86_##func, *(((struct kvm_x86_ops *)0)->func));
1810 return __vmalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_arch_alloc_vm()
1819 if (kvm_x86_ops.flush_remote_tlbs && in kvm_arch_flush_remote_tlbs()
1894 * kvm_x86_ops.skip_emulated_instruction() implementations if
/kernel/linux/linux-5.10/arch/x86/kvm/mmu/
H A Dspte.c123 spte |= kvm_x86_ops.get_mt_mask(vcpu, gfn, in make_spte()
H A Dmmu.c186 return kvm_x86_ops.tlb_remote_flush_with_range; in kvm_available_flush_tlb_with_range()
194 if (range && kvm_x86_ops.tlb_remote_flush_with_range) in kvm_flush_remote_tlbs_with_range()
195 ret = kvm_x86_ops.tlb_remote_flush_with_range(kvm, range); in kvm_flush_remote_tlbs_with_range()
1186 * Similar to the !kvm_x86_ops.slot_disable_log_dirty case, in spte_set_dirty()
1281 if (kvm_x86_ops.enable_log_dirty_pt_masked) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1282 kvm_x86_ops.enable_log_dirty_pt_masked(kvm, slot, gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked()
4856 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_mmu_load()
5137 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invalidate_gva()
5196 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invpcid_gva()
H A Dpaging_tmpl.h260 if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr)) in update_accessed_dirty_bits()
/kernel/linux/linux-6.6/arch/x86/kvm/mmu/
H A Dmmu_internal.h157 return kvm_x86_ops.cpu_dirty_log_size && sp->role.guest_mode; in kvm_mmu_page_ad_need_write_protect()
H A Dpaging_tmpl.h225 if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr)) in update_accessed_dirty_bits()
H A Dmmu.c274 return kvm_x86_ops.flush_remote_tlbs_range; in kvm_available_flush_remote_tlbs_range()
279 if (!kvm_x86_ops.flush_remote_tlbs_range) in kvm_arch_flush_remote_tlbs_range()
1397 if (kvm_x86_ops.cpu_dirty_log_size) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1405 return kvm_x86_ops.cpu_dirty_log_size; in kvm_cpu_dirty_log_size()
1586 if (kvm_x86_ops.set_apic_access_page_addr && in kvm_unmap_gfn_range()
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/
H A Dvmx.c3189 if (kvm_x86_ops.tlb_remote_flush) { in vmx_load_mmu_pgd()
7696 if (kvm_x86_ops.set_hv_timer) in vmx_post_block()
7792 static struct kvm_x86_ops vmx_x86_ops __initdata = {

Completed in 74 milliseconds

12