Home
last modified time | relevance | path

Searched refs:per_cpu_ptr (Results 1 - 25 of 646) sorted by relevance

12345678910>>...26

/kernel/linux/linux-5.10/fs/xfs/
H A Dxfs_stats.h165 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \
166 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
171 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \
172 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
177 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
178 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
183 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \
184 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
189 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \
190 per_cpu_ptr(m
[all...]
H A Dxfs_stats.c15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val()
71 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format()
72 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format()
73 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format()
74 defer_relog += per_cpu_ptr(stats, i)->s.defer_relog; in xfs_stats_format()
100 vn_active = per_cpu_ptr(stats, c)->s.vn_active; in xfs_stats_clearall()
101 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall()
102 per_cpu_ptr(stats, c)->s.vn_active = vn_active; in xfs_stats_clearall()
/kernel/linux/linux-6.6/fs/xfs/
H A Dxfs_stats.h165 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \
166 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
171 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \
172 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
177 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
178 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
183 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \
184 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
189 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \
190 per_cpu_ptr(m
[all...]
H A Dxfs_stats.c15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val()
71 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format()
72 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format()
73 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format()
74 defer_relog += per_cpu_ptr(stats, i)->s.defer_relog; in xfs_stats_format()
100 vn_active = per_cpu_ptr(stats, c)->s.vn_active; in xfs_stats_clearall()
101 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall()
102 per_cpu_ptr(stats, c)->s.vn_active = vn_active; in xfs_stats_clearall()
/kernel/linux/linux-5.10/arch/x86/events/amd/
H A Duncore.c75 return *per_cpu_ptr(amd_uncore_nb, event->cpu); in event_to_amd_uncore()
77 return *per_cpu_ptr(amd_uncore_llc, event->cpu); in event_to_amd_uncore()
386 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; in amd_uncore_cpu_up_prepare()
400 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc; in amd_uncore_cpu_up_prepare()
407 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; in amd_uncore_cpu_up_prepare()
420 that = *per_cpu_ptr(uncores, cpu); in amd_uncore_find_online_sibling()
445 uncore = *per_cpu_ptr(amd_uncore_nb, cpu); in amd_uncore_cpu_starting()
450 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore; in amd_uncore_cpu_starting()
454 uncore = *per_cpu_ptr(amd_uncore_llc, cpu); in amd_uncore_cpu_starting()
458 *per_cpu_ptr(amd_uncore_ll in amd_uncore_cpu_starting()
[all...]
/kernel/linux/linux-6.6/arch/x86/events/amd/
H A Duncore.c75 return *per_cpu_ptr(amd_uncore_nb, event->cpu); in event_to_amd_uncore()
77 return *per_cpu_ptr(amd_uncore_llc, event->cpu); in event_to_amd_uncore()
446 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; in amd_uncore_cpu_up_prepare()
460 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; in amd_uncore_cpu_up_prepare()
464 *per_cpu_ptr(amd_uncore_llc, cpu) = NULL; in amd_uncore_cpu_up_prepare()
478 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc; in amd_uncore_cpu_up_prepare()
505 that = *per_cpu_ptr(uncores, cpu); in amd_uncore_find_online_sibling()
530 uncore = *per_cpu_ptr(amd_uncore_nb, cpu); in amd_uncore_cpu_starting()
535 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore; in amd_uncore_cpu_starting()
539 uncore = *per_cpu_ptr(amd_uncore_ll in amd_uncore_cpu_starting()
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dcontext_tracking_state.h66 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_cpu()
73 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_cpu_acquire()
85 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_nesting_cpu()
97 struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); in ct_dynticks_nmi_nesting_cpu()
H A Dpart_stat.h29 (per_cpu_ptr((part)->bd_stats, (cpu))->field)
39 res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \
48 memset(per_cpu_ptr(part->bd_stats, i), value, in part_stat_set_all()
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt-stats.c27 s = per_cpu_ptr(stats->pcpu_stats, con->cpu); in rtrs_clt_update_wc_stats()
54 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_cnt_to_str()
61 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_cnt_to_str()
88 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str()
118 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_rdma_stats()
134 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_cpu_migr_stats()
/kernel/linux/linux-6.6/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt-stats.c27 s = per_cpu_ptr(stats->pcpu_stats, con->cpu); in rtrs_clt_update_wc_stats()
47 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_from_cnt_to_str()
66 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_stats_migration_to_cnt_to_str()
90 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str()
119 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_rdma_stats()
135 s = per_cpu_ptr(stats->pcpu_stats, cpu); in rtrs_clt_reset_cpu_migr_stats()
/kernel/linux/linux-6.6/kernel/
H A Dsmpboot.c172 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread()
197 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread()
230 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread()
249 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread()
272 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads()
277 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
/kernel/linux/linux-6.6/kernel/sched/
H A Dcpuacct.c97 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_read()
98 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_read()
137 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_write()
138 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_write()
270 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_stats_show()
278 cputime.sum_exec_runtime += *per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_stats_show()
342 *per_cpu_ptr(ca->cpuusage, cpu) += cputime; in cpuacct_charge()
H A Dtopology.c918 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask()
979 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group()
1039 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups()
1189 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1197 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
1198 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
1521 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations()
1522 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations()
1524 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations()
1525 *per_cpu_ptr(sd in claim_allocations()
[all...]
/kernel/linux/linux-5.10/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/
H A Dpercpu.h28 #define per_cpu_ptr(ptr, cpu) \ macro
69 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), thread_cpu_id), \
76 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), this_cpu_add_impl_cpu), \
/kernel/linux/linux-5.10/include/linux/
H A Dpart_stat.h28 (per_cpu_ptr((part)->dkstats, (cpu))->field)
38 res += per_cpu_ptr((part)->dkstats, _cpu)->field; \
47 memset(per_cpu_ptr(part->dkstats, i), value, in part_stat_set_all()
/kernel/linux/linux-5.10/fs/squashfs/
H A Ddecompressor_multi_percpu.c40 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create()
54 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create()
71 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy()
/kernel/linux/linux-5.10/kernel/bpf/
H A Dpercpu_freelist.c15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init()
65 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_push_nmi()
112 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate()
131 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop()
165 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop_nmi()
/kernel/linux/linux-6.6/kernel/bpf/
H A Dpercpu_freelist.c15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init()
66 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_push_nmi()
110 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate()
128 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop()
160 head = per_cpu_ptr(s->freelist, cpu); in ___pcpu_freelist_pop_nmi()
/kernel/linux/linux-5.10/drivers/powercap/
H A Didle_inject.c92 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_wakeup()
137 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_fn()
254 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_stop()
284 per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_should_run()
/kernel/linux/linux-5.10/kernel/sched/
H A Dtopology.c870 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask()
929 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group()
989 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups()
1139 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1147 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
1148 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
1351 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations()
1352 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations()
1354 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations()
1355 *per_cpu_ptr(sd in claim_allocations()
[all...]
H A Dcpuacct.c98 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_read()
99 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_read()
137 u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); in cpuacct_cpuusage_write()
138 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_cpuusage_write()
269 u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; in cpuacct_stats_show()
337 *per_cpu_ptr(ca->cpuusage, cpu) += cputime; in cpuacct_charge()
/kernel/linux/linux-5.10/net/xfrm/
H A Dxfrm_ipcomp.c45 u8 *scratch = *per_cpu_ptr(ipcomp_scratches, cpu); in ipcomp_decompress()
46 struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); in ipcomp_decompress()
212 vfree(*per_cpu_ptr(scratches, i)); in ipcomp_free_scratches()
238 *per_cpu_ptr(scratches, i) = scratch; in ipcomp_alloc_scratches()
266 struct crypto_comp *tfm = *per_cpu_ptr(tfms, cpu); in ipcomp_free_tfms()
308 *per_cpu_ptr(tfms, cpu) = tfm; in ipcomp_alloc_tfms()
/kernel/linux/linux-5.10/kernel/
H A Dsmpboot.c172 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread()
197 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread()
230 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread()
249 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread()
272 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads()
277 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
/kernel/linux/linux-5.10/kernel/irq/
H A Dmatrix.c138 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu()
159 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed()
215 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed()
256 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed()
298 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed()
394 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc()
421 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free()
504 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()
/kernel/linux/linux-6.6/kernel/irq/
H A Dmatrix.c138 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu()
159 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed()
215 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed()
256 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed()
299 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed()
394 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc()
421 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free()
504 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()

Completed in 18 milliseconds

12345678910>>...26