/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_stats.h | 165 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \ 166 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \ 171 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \ 172 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \ 177 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \ 178 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \ 183 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \ 184 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \ 189 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \ 190 per_cpu_ptr(m [all...] |
H A D | xfs_stats.c | 15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val() 71 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format() 72 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format() 73 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format() 74 defer_relog += per_cpu_ptr(stats, i)->s.defer_relog; in xfs_stats_format() 100 vn_active = per_cpu_ptr(stats, c)->s.vn_active; in xfs_stats_clearall() 101 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall() 102 per_cpu_ptr(stats, c)->s.vn_active = vn_active; in xfs_stats_clearall()
|
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_stats.h | 165 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \ 166 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \ 171 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \ 172 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \ 177 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \ 178 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \ 183 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \ 184 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \ 189 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \ 190 per_cpu_ptr(m [all...] |
H A D | xfs_stats.c | 15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val() 71 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format() 72 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format() 73 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format() 74 defer_relog += per_cpu_ptr(stats, i)->s.defer_relog; in xfs_stats_format() 100 vn_active = per_cpu_ptr(stats, c)->s.vn_active; in xfs_stats_clearall() 101 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall() 102 per_cpu_ptr(stats, c)->s.vn_active = vn_active; in xfs_stats_clearall()
|
/kernel/linux/linux-5.10/arch/x86/events/amd/ |
H A D | uncore.c | 75 return *per_cpu_ptr(amd_uncore_nb, event->cpu); in event_to_amd_uncore() 77 return *per_cpu_ptr(amd_uncore_llc, event->cpu); in event_to_amd_uncore() 386 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; in amd_uncore_cpu_up_prepare() 400 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc; in amd_uncore_cpu_up_prepare() 407 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; in amd_uncore_cpu_up_prepare() 420 that = *per_cpu_ptr(uncores, cpu); in amd_uncore_find_online_sibling() 445 uncore = *per_cpu_ptr(amd_uncore_nb, cpu); in amd_uncore_cpu_starting() 450 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore; in amd_uncore_cpu_starting() 454 uncore = *per_cpu_ptr(amd_uncore_llc, cpu); in amd_uncore_cpu_starting() 458 *per_cpu_ptr(amd_uncore_ll in amd_uncore_cpu_starting() [all...] |
/kernel/linux/linux-6.6/arch/x86/events/amd/ |
H A D | uncore.c | 75 return *per_cpu_ptr(amd_uncore_nb, event->cpu); in event_to_amd_uncore() 77 return *per_cpu_ptr(amd_uncore_llc, event->cpu); in event_to_amd_uncore() 446 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; in amd_uncore_cpu_up_prepare() 460 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; in amd_uncore_cpu_up_prepare() 464 *per_cpu_ptr(amd_uncore_llc, cpu) = NULL; in amd_uncore_cpu_up_prepare() 478 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc; in amd_uncore_cpu_up_prepare() 505 that = *per_cpu_ptr(uncores, cpu); in amd_uncore_find_online_sibling() 530 uncore = *per_cpu_ptr(amd_uncore_nb, cpu); in amd_uncore_cpu_starting() 535 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore; in amd_uncore_cpu_starting() 539 uncore = *per_cpu_ptr(amd_uncore_ll in amd_uncore_cpu_starting() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | context_tracking_state.h | 66 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_cpu() 73 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_cpu_acquire() 85 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_nesting_cpu() 97 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_nmi_nesting_cpu()
|
H A D | part_stat.h | 29 (per_cpu_ptr((part)->bd_stats, (cpu))->field) 39 res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \ 48 memset(per_cpu_ptr(part->bd_stats, i), value, in part_stat_set_all()
|
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-clt-stats.c | 27 s = per_cpu_ptr(stats->pcpu_stats, con->cpu); in rtrs_clt_update_wc_stats() 54 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_cnt_to_str() 61 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_cnt_to_str() 88 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 118 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_rdma_stats() 134 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_cpu_migr_stats()
|
/kernel/linux/linux-6.6/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-clt-stats.c | 27 s = per_cpu_ptr(stats->pcpu_stats, con->cpu); in rtrs_clt_update_wc_stats() 47 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_from_cnt_to_str() 66 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_to_cnt_to_str() 90 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 119 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_rdma_stats() 135 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_cpu_migr_stats()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | smpboot.c | 172 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread() 197 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread() 230 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread() 249 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread() 272 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads() 277 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | cpuacct.c | 97 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_read() 98 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_read() 137 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_write() 138 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_write() 270 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_stats_show() 278 cputime.sum_exec_runtime += *per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_stats_show() 342 *per_cpu_ptr(ca->cpuusage, cpu) += cputime; in cpuacct_charge()
|
H A D | topology.c | 918 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask() 979 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group() 1039 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups() 1189 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group() 1197 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group() 1198 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group() 1521 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations() 1522 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations() 1524 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations() 1525 *per_cpu_ptr(sd in claim_allocations() [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
H A D | percpu.h | 28 #define per_cpu_ptr(ptr, cpu) \ macro 69 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), thread_cpu_id), \ 76 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), this_cpu_add_impl_cpu), \
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | part_stat.h | 28 (per_cpu_ptr((part)->dkstats, (cpu))->field) 38 res += per_cpu_ptr((part)->dkstats, _cpu)->field; \ 47 memset(per_cpu_ptr(part->dkstats, i), value, in part_stat_set_all()
|
/kernel/linux/linux-5.10/fs/squashfs/ |
H A D | decompressor_multi_percpu.c | 40 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 54 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 71 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy()
|
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | percpu_freelist.c | 15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init() 65 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_push_nmi() 112 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate() 131 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop() 165 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop_nmi()
|
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | percpu_freelist.c | 15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init() 66 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_push_nmi() 110 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate() 128 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop() 160 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop_nmi()
|
/kernel/linux/linux-5.10/drivers/powercap/ |
H A D | idle_inject.c | 92 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_wakeup() 137 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_fn() 254 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_stop() 284 per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_should_run()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | topology.c | 870 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask() 929 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group() 989 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups() 1139 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group() 1147 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group() 1148 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group() 1351 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations() 1352 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations() 1354 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations() 1355 *per_cpu_ptr(sd in claim_allocations() [all...] |
H A D | cpuacct.c | 98 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_read() 99 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_read() 137 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_write() 138 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_write() 269 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_stats_show() 337 *per_cpu_ptr(ca->cpuusage, cpu) += cputime; in cpuacct_charge()
|
/kernel/linux/linux-5.10/net/xfrm/ |
H A D | xfrm_ipcomp.c | 45 u8 *scratch = *per_cpu_ptr(ipcomp_scratches, cpu); in ipcomp_decompress() 46 struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); in ipcomp_decompress() 212 vfree(*per_cpu_ptr(scratches, i)); in ipcomp_free_scratches() 238 *per_cpu_ptr(scratches, i) = scratch; in ipcomp_alloc_scratches() 266 struct crypto_comp *tfm = *per_cpu_ptr(tfms, cpu); in ipcomp_free_tfms() 308 *per_cpu_ptr(tfms, cpu) = tfm; in ipcomp_alloc_tfms()
|
/kernel/linux/linux-5.10/kernel/ |
H A D | smpboot.c | 172 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread() 197 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread() 230 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread() 249 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread() 272 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads() 277 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
|
/kernel/linux/linux-5.10/kernel/irq/ |
H A D | matrix.c | 138 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu() 159 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed() 215 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed() 256 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed() 298 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed() 394 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc() 421 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free() 504 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()
|
/kernel/linux/linux-6.6/kernel/irq/ |
H A D | matrix.c | 138 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu() 159 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed() 215 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed() 256 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed() 299 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed() 394 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc() 421 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free() 504 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()
|