Home
last modified time | relevance | path

Searched refs:vcpu_is_preempted (Results 1 - 25 of 37) sorted by relevance

12

/kernel/linux/linux-5.10/arch/arm64/include/asm/
H A Dspinlock.h21 #define vcpu_is_preempted vcpu_is_preempted macro
22 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
/kernel/linux/linux-6.6/arch/arm64/include/asm/
H A Dspinlock.h21 #define vcpu_is_preempted vcpu_is_preempted macro
22 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
/kernel/linux/linux-5.10/arch/loongarch/include/asm/
H A Dqspinlock.h34 #define vcpu_is_preempted vcpu_is_preempted macro
35 static inline bool vcpu_is_preempted(long cpu) in vcpu_is_preempted() function
H A Dparavirt.h51 bool (*vcpu_is_preempted)(long cpu); member
81 return pv_lock_ops.vcpu_is_preempted(cpu); in pv_vcpu_is_preempted()
/kernel/linux/linux-5.10/arch/powerpc/include/asm/
H A Dparavirt.h89 #define vcpu_is_preempted vcpu_is_preempted macro
90 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
/kernel/linux/linux-5.10/arch/x86/include/asm/
H A Dqspinlock.h59 #define vcpu_is_preempted vcpu_is_preempted macro
60 static inline bool vcpu_is_preempted(long cpu) in vcpu_is_preempted() function
/kernel/linux/linux-6.6/arch/powerpc/include/asm/
H A Dparavirt.h105 #define vcpu_is_preempted vcpu_is_preempted macro
106 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
129 * The result of vcpu_is_preempted() is used in a in vcpu_is_preempted()
/kernel/linux/linux-6.6/arch/x86/include/asm/
H A Dqspinlock.h60 #define vcpu_is_preempted vcpu_is_preempted macro
61 static inline bool vcpu_is_preempted(long cpu) in vcpu_is_preempted() function
/kernel/linux/linux-5.10/arch/x86/kernel/
H A Dparavirt_patch.c57 unsigned char vcpu_is_preempted[2]; member
61 .vcpu_is_preempted = { 0x31, 0xc0 }, // xor %eax, %eax
96 case PARAVIRT_PATCH(lock.vcpu_is_preempted): in native_patch()
98 return PATCH(lock, vcpu_is_preempted, insn_buff, len); in native_patch()
H A Dparavirt-spinlocks.c32 return pv_ops.lock.vcpu_is_preempted.func == in pv_is_native_vcpu_is_preempted()
H A Dkvm.c607 if (vcpu_is_preempted(cpu)) { in kvm_smp_send_call_func_ipi()
1006 pv_ops.lock.vcpu_is_preempted = in kvm_spinlock_init()
H A Dparavirt.c410 .lock.vcpu_is_preempted =
/kernel/linux/linux-5.10/arch/loongarch/kernel/
H A Dparavirt-spinlocks.c36 return pv_lock_ops.vcpu_is_preempted == __native_vcpu_is_preempted; in pv_is_native_vcpu_is_preempted()
45 .vcpu_is_preempted = __native_vcpu_is_preempted,
H A Dparavirt.c237 pv_lock_ops.vcpu_is_preempted = __kvm_vcpu_is_preempted; in kvm_spinlock_init()
/kernel/linux/linux-6.6/arch/x86/kernel/
H A Dparavirt-spinlocks.c32 return pv_ops.lock.vcpu_is_preempted.func == in pv_is_native_vcpu_is_preempted()
H A Dkvm.c640 if (!idle_cpu(cpu) && vcpu_is_preempted(cpu)) { in kvm_smp_send_call_func_ipi()
823 pv_ops.lock.vcpu_is_preempted = in kvm_guest_init()
/kernel/linux/linux-5.10/arch/x86/hyperv/
H A Dhv_spinlock.c80 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(hv_vcpu_is_preempted); in hv_init_spinlocks()
/kernel/linux/linux-6.6/arch/x86/hyperv/
H A Dhv_spinlock.c84 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(hv_vcpu_is_preempted); in hv_init_spinlocks()
/kernel/linux/linux-6.6/arch/s390/include/asm/
H A Dspinlock.h25 #define vcpu_is_preempted arch_vcpu_is_preempted macro
/kernel/linux/linux-5.10/arch/s390/include/asm/
H A Dspinlock.h25 #define vcpu_is_preempted arch_vcpu_is_preempted macro
/kernel/linux/linux-5.10/arch/x86/xen/
H A Dspinlock.c144 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(xen_vcpu_stolen); in xen_init_spinlocks()
/kernel/linux/linux-5.10/kernel/locking/
H A Dosq_lock.c140 * is implemented with a monitor-wait. vcpu_is_preempted() relies on in osq_lock()
144 vcpu_is_preempted(node_cpu(node->prev)))) in osq_lock()
/kernel/linux/linux-6.6/kernel/locking/
H A Dosq_lock.c140 * is implemented with a monitor-wait. vcpu_is_preempted() relies on in osq_lock()
144 vcpu_is_preempted(node_cpu(node->prev)))) in osq_lock()
/kernel/linux/linux-6.6/arch/x86/xen/
H A Dspinlock.c144 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(xen_vcpu_stolen); in xen_init_spinlocks()
/kernel/linux/linux-6.6/arch/powerpc/lib/
H A Dqspinlock.c374 if (vcpu_is_preempted(owner)) { in propagate_yield_cpu()
709 if (vcpu_is_preempted(next_cpu)) in queued_spin_lock_mcs_queue()

Completed in 10 milliseconds

12