/kernel/linux/linux-6.6/arch/x86/events/amd/ |
H A D | brs.c | 381 * On ctxswin, sched_in = true, called after the PMU has started 382 * On ctxswout, sched_in = false, called before the PMU is stopped 384 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 397 if (sched_in) in amd_pmu_brs_sched_task()
|
H A D | lbr.c | 376 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_lbr_sched_task() argument 385 if (cpuc->lbr_users && sched_in) in amd_pmu_lbr_sched_task()
|
/kernel/linux/linux-5.10/tools/perf/ |
H A D | builtin-sched.c | 1115 struct thread *sched_out, *sched_in; in latency_switch_event() local 1135 sched_in = machine__findnew_thread(machine, -1, next_pid); in latency_switch_event() 1136 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1152 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1154 if (thread_atoms_insert(sched, sched_in)) in latency_switch_event() 1156 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1172 thread__put(sched_in); in latency_switch_event() 1533 struct thread *sched_in; in map_switch_event() local 1570 sched_in = map__findnew_thread(sched, machine, -1, next_pid); in map_switch_event() 1571 if (sched_in in map_switch_event() [all...] |
/kernel/linux/linux-6.6/tools/perf/ |
H A D | builtin-sched.c | 1137 struct thread *sched_out, *sched_in; in latency_switch_event() local 1157 sched_in = machine__findnew_thread(machine, -1, next_pid); in latency_switch_event() 1158 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1174 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1176 if (thread_atoms_insert(sched, sched_in)) in latency_switch_event() 1178 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1194 thread__put(sched_in); in latency_switch_event() 1570 struct thread *sched_in; in map_switch_event() local 1610 sched_in = map__findnew_thread(sched, machine, -1, next_pid); in map_switch_event() 1611 if (sched_in in map_switch_event() [all...] |
/kernel/linux/linux-6.6/arch/x86/events/ |
H A D | perf_event.h | 819 bool sched_in); 1312 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1353 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1378 static inline void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 1540 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1551 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in);
|
H A D | core.c | 2623 static void x86_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in x86_pmu_sched_task() argument 2625 static_call_cond(x86_pmu_sched_task)(pmu_ctx, sched_in); in x86_pmu_sched_task()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | preempt.h | 281 * @sched_in: we're about to be rescheduled: 288 * Please note that sched_in and out are called under different 290 * while sched_in is called without rq lock and irq enabled. This 294 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member
|
H A D | perf_event.h | 430 bool sched_in);
|
/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | kvm-x86-ops.h | 105 KVM_X86_OP(sched_in)
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | preempt.h | 325 * @sched_in: we're about to be rescheduled: 332 * Please note that sched_in and out are called under different 334 * while sched_in is called without rq lock and irq enabled. This 338 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member
|
H A D | perf_event.h | 456 bool sched_in);
|
/kernel/linux/linux-6.6/arch/s390/kernel/ |
H A D | perf_pai_crypto.c | 381 static void paicrypt_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in paicrypt_sched_task() argument 386 if (!sched_in) in paicrypt_sched_task()
|
H A D | perf_pai_ext.c | 470 static void paiext_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in paiext_sched_task() argument 475 if (!sched_in) in paiext_sched_task()
|
/kernel/linux/linux-5.10/arch/x86/events/ |
H A D | perf_event.h | 711 bool sched_in); 1179 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in); 1190 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in);
|
H A D | core.c | 2426 static void x86_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in x86_pmu_sched_task() argument 2428 static_call_cond(x86_pmu_sched_task)(ctx, sched_in); in x86_pmu_sched_task()
|
/kernel/linux/linux-5.10/arch/x86/events/intel/ |
H A D | lbr.c | 622 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_lbr_sched_task() argument 637 if (sched_in) in intel_pmu_lbr_sched_task() 650 if (sched_in) in intel_pmu_lbr_sched_task()
|
H A D | ds.c | 917 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 921 if (!sched_in && pebs_needs_sched_cb(cpuc)) in intel_pmu_pebs_sched_task()
|
H A D | core.c | 4194 bool sched_in) in intel_pmu_sched_task() 4196 intel_pmu_pebs_sched_task(ctx, sched_in); in intel_pmu_sched_task() 4197 intel_pmu_lbr_sched_task(ctx, sched_in); in intel_pmu_sched_task() 4193 intel_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) intel_pmu_sched_task() argument
|
/kernel/linux/linux-6.6/arch/x86/events/intel/ |
H A D | lbr.c | 541 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in intel_pmu_lbr_sched_task() argument 556 if (sched_in) in intel_pmu_lbr_sched_task() 569 if (sched_in) in intel_pmu_lbr_sched_task()
|
H A D | ds.c | 1128 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 1132 if (!sched_in && pebs_needs_sched_cb(cpuc)) in intel_pmu_pebs_sched_task()
|
/kernel/linux/linux-5.10/arch/powerpc/perf/ |
H A D | core-book3s.c | 130 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) {} in power_pmu_sched_task() argument 401 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in power_pmu_sched_task() argument 406 if (sched_in) in power_pmu_sched_task()
|
/kernel/linux/linux-6.6/arch/powerpc/perf/ |
H A D | core-book3s.c | 135 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) {} in power_pmu_sched_task() argument 454 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in power_pmu_sched_task() argument 459 if (sched_in) in power_pmu_sched_task()
|
/kernel/linux/linux-6.6/kernel/events/ |
H A D | core.c | 3479 static void perf_ctx_sched_task_cb(struct perf_event_context *ctx, bool sched_in) in perf_ctx_sched_task_cb() argument 3488 pmu_ctx->pmu->sched_task(pmu_ctx, sched_in); in perf_ctx_sched_task_cb() 3618 static void __perf_pmu_sched_task(struct perf_cpu_pmu_context *cpc, bool sched_in) in __perf_pmu_sched_task() argument 3632 pmu->sched_task(cpc->task_epc, sched_in); in __perf_pmu_sched_task() 3640 bool sched_in) in perf_pmu_sched_task() 3650 __perf_pmu_sched_task(cpc, sched_in); in perf_pmu_sched_task() 3654 struct task_struct *next_prev, bool sched_in); 9052 struct task_struct *next_prev, bool sched_in) in perf_event_switch() 9064 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch() 9072 if (!sched_in in perf_event_switch() 3638 perf_pmu_sched_task(struct task_struct *prev, struct task_struct *next, bool sched_in) perf_pmu_sched_task() argument 9051 perf_event_switch(struct task_struct *task, struct task_struct *next_prev, bool sched_in) perf_event_switch() argument [all...] |
/kernel/linux/linux-5.10/kernel/events/ |
H A D | core.c | 3576 static void __perf_pmu_sched_task(struct perf_cpu_context *cpuctx, bool sched_in) in __perf_pmu_sched_task() argument 3588 pmu->sched_task(cpuctx->task_ctx, sched_in); in __perf_pmu_sched_task() 3596 bool sched_in) in perf_pmu_sched_task() 3608 __perf_pmu_sched_task(cpuctx, sched_in); in perf_pmu_sched_task() 3613 struct task_struct *next_prev, bool sched_in); 8597 struct task_struct *next_prev, bool sched_in) in perf_event_switch() 8609 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch() 8617 if (!sched_in && task->state == TASK_RUNNING) in perf_event_switch() 3594 perf_pmu_sched_task(struct task_struct *prev, struct task_struct *next, bool sched_in) perf_pmu_sched_task() argument 8596 perf_event_switch(struct task_struct *task, struct task_struct *next_prev, bool sched_in) perf_event_switch() argument
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | kvm_host.h | 1216 void (*sched_in)(struct kvm_vcpu *kvm, int cpu); member
|