/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | idle.h | 14 *this_cpu_ptr(&idle_entry_purr_snap) = mfspr(SPRN_PURR); in snapshot_purr_idle_entry() 19 *this_cpu_ptr(&idle_entry_spurr_snap) = mfspr(SPRN_SPURR); in snapshot_spurr_idle_entry() 25 u64 in_purr = *this_cpu_ptr(&idle_entry_purr_snap); in update_idle_purr_accounting() 34 u64 *idle_spurr_cycles_ptr = this_cpu_ptr(&idle_spurr_cycles); in update_idle_spurr_accounting() 35 u64 in_spurr = *this_cpu_ptr(&idle_entry_spurr_snap); in update_idle_spurr_accounting() 89 return *this_cpu_ptr(&idle_spurr_cycles); in read_this_idle_spurr()
|
H A D | xics.h | 110 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_push_cppr() 123 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_pop_cppr() 133 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_set_base_cppr() 145 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_cppr_top()
|
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | idle.h | 14 *this_cpu_ptr(&idle_entry_purr_snap) = mfspr(SPRN_PURR); in snapshot_purr_idle_entry() 19 *this_cpu_ptr(&idle_entry_spurr_snap) = mfspr(SPRN_SPURR); in snapshot_spurr_idle_entry() 25 u64 in_purr = *this_cpu_ptr(&idle_entry_purr_snap); in update_idle_purr_accounting() 34 u64 *idle_spurr_cycles_ptr = this_cpu_ptr(&idle_spurr_cycles); in update_idle_spurr_accounting() 35 u64 in_spurr = *this_cpu_ptr(&idle_entry_spurr_snap); in update_idle_spurr_accounting() 89 return *this_cpu_ptr(&idle_spurr_cycles); in read_this_idle_spurr()
|
H A D | xics.h | 115 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_push_cppr() 128 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_pop_cppr() 138 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_set_base_cppr() 150 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_cppr_top()
|
/kernel/linux/linux-5.10/kernel/ |
H A D | irq_work.c | 57 if (llist_add(&work->llnode, this_cpu_ptr(&lazy_list)) && in __irq_work_queue_local() 61 if (llist_add(&work->llnode, this_cpu_ptr(&raised_list))) in __irq_work_queue_local() 120 raised = this_cpu_ptr(&raised_list); in irq_work_needs_cpu() 121 lazy = this_cpu_ptr(&lazy_list); in irq_work_needs_cpu() 179 irq_work_run_list(this_cpu_ptr(&raised_list)); in irq_work_run() 180 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_run() 186 struct llist_head *raised = this_cpu_ptr(&raised_list); in irq_work_tick() 190 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_tick()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | local_lock_internal.h | 68 local_lock_acquire(this_cpu_ptr(lock)); \ 74 local_lock_acquire(this_cpu_ptr(lock)); \ 80 local_lock_acquire(this_cpu_ptr(lock)); \ 85 local_lock_release(this_cpu_ptr(lock)); \ 91 local_lock_release(this_cpu_ptr(lock)); \ 97 local_lock_release(this_cpu_ptr(lock)); \ 119 spin_lock(this_cpu_ptr((__lock))); \ 133 spin_unlock(this_cpu_ptr((__lock))); \
|
/kernel/linux/linux-6.6/kernel/ |
H A D | irq_work.c | 35 if (!llist_empty(this_cpu_ptr(&lazy_list)) && tsk) in wake_irq_workd() 51 return !llist_empty(this_cpu_ptr(&lazy_list)); in irq_workd_should_run() 103 list = this_cpu_ptr(&lazy_list); in __irq_work_queue_local() 105 list = this_cpu_ptr(&raised_list); in __irq_work_queue_local() 188 raised = this_cpu_ptr(&raised_list); in irq_work_needs_cpu() 189 lazy = this_cpu_ptr(&lazy_list); in irq_work_needs_cpu() 261 irq_work_run_list(this_cpu_ptr(&raised_list)); in irq_work_run() 263 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_run() 271 struct llist_head *raised = this_cpu_ptr(&raised_list); in irq_work_tick() 277 irq_work_run_list(this_cpu_ptr( in irq_work_tick() [all...] |
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/nvhe/ |
H A D | stacktrace.c | 28 struct kvm_nvhe_stacktrace_info *stacktrace_info = this_cpu_ptr(&kvm_stacktrace_info); in hyp_prepare_backtrace() 29 struct kvm_nvhe_init_params *params = this_cpu_ptr(&kvm_init_params); in hyp_prepare_backtrace() 32 stacktrace_info->overflow_stack_base = (unsigned long)this_cpu_ptr(overflow_stack); in hyp_prepare_backtrace() 44 unsigned long low = (unsigned long)this_cpu_ptr(overflow_stack); in stackinfo_get_overflow() 55 struct kvm_nvhe_init_params *params = this_cpu_ptr(&kvm_init_params); in stackinfo_get_hyp() 96 unsigned long *stacktrace = this_cpu_ptr(pkvm_stacktrace); in pkvm_save_backtrace_entry()
|
H A D | psci-relay.c | 160 boot_args = this_cpu_ptr(&suspend_args); in psci_cpu_suspend() 161 init_params = this_cpu_ptr(&kvm_init_params); in psci_cpu_suspend() 187 boot_args = this_cpu_ptr(&suspend_args); in psci_system_suspend() 188 init_params = this_cpu_ptr(&kvm_init_params); in psci_system_suspend() 208 host_ctxt = &this_cpu_ptr(&kvm_host_data)->host_ctxt; in __kvm_host_psci_cpu_entry() 211 boot_args = this_cpu_ptr(&cpu_on_args); in __kvm_host_psci_cpu_entry() 213 boot_args = this_cpu_ptr(&suspend_args); in __kvm_host_psci_cpu_entry()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | local_lock_internal.h | 66 local_lock_acquire(this_cpu_ptr(lock)); \ 72 local_lock_acquire(this_cpu_ptr(lock)); \ 78 local_lock_acquire(this_cpu_ptr(lock)); \ 83 local_lock_release(this_cpu_ptr(lock)); \ 89 local_lock_release(this_cpu_ptr(lock)); \ 95 local_lock_release(this_cpu_ptr(lock)); \
|
/kernel/linux/linux-6.6/arch/x86/kernel/cpu/mce/ |
H A D | intel.c | 146 if (machine_check_poll(0, this_cpu_ptr(&mce_banks_owned))) in mce_intel_cmci_poll() 169 owned = this_cpu_ptr(mce_banks_owned); in cmci_toggle_interrupt_mode() 266 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 276 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_discover() 298 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 322 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 333 WARN_ON(!test_bit(i, this_cpu_ptr(mce_poll_banks))); in cmci_discover() 358 machine_check_poll(0, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 367 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 372 __clear_bit(bank, this_cpu_ptr(mce_banks_owne in __cmci_disable_bank() [all...] |
/kernel/linux/linux-5.10/arch/x86/kernel/cpu/mce/ |
H A D | intel.c | 139 if (machine_check_poll(0, this_cpu_ptr(&mce_banks_owned))) in mce_intel_cmci_poll() 162 owned = this_cpu_ptr(mce_banks_owned); in cmci_toggle_interrupt_mode() 259 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 269 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_discover() 291 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 315 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 326 WARN_ON(!test_bit(i, this_cpu_ptr(mce_poll_banks))); in cmci_discover() 351 machine_check_poll(0, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 360 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 365 __clear_bit(bank, this_cpu_ptr(mce_banks_owne in __cmci_disable_bank() [all...] |
/kernel/linux/linux-5.10/net/core/ |
H A D | dst_cache.c | 68 return dst_cache_per_cpu_get(dst_cache, this_cpu_ptr(dst_cache->cache)); in dst_cache_get() 80 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip4() 98 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip4() 113 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip6() 114 dst_cache_per_cpu_dst_set(this_cpu_ptr(dst_cache->cache), dst, in dst_cache_set_ip6() 129 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip6()
|
/kernel/linux/linux-6.6/net/core/ |
H A D | dst_cache.c | 68 return dst_cache_per_cpu_get(dst_cache, this_cpu_ptr(dst_cache->cache)); in dst_cache_get() 80 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip4() 98 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip4() 113 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip6() 114 dst_cache_per_cpu_dst_set(this_cpu_ptr(dst_cache->cache), dst, in dst_cache_set_ip6() 129 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip6()
|
/kernel/linux/linux-5.10/arch/xtensa/kernel/ |
H A D | hw_breakpoint.c | 177 i = alloc_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_install_hw_breakpoint() 184 i = alloc_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_install_hw_breakpoint() 215 i = free_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_uninstall_hw_breakpoint() 223 i = free_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_uninstall_hw_breakpoint() 268 struct perf_event *bp = this_cpu_ptr(wp_on_reg)[i]; in restore_dbreak() 280 struct perf_event **bp = this_cpu_ptr(bp_on_reg); in check_hw_breakpoint() 289 struct perf_event **bp = this_cpu_ptr(wp_on_reg); in check_hw_breakpoint()
|
/kernel/linux/linux-6.6/arch/x86/events/amd/ |
H A D | brs.c | 205 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_enable() 221 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_enable_all() 228 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_disable() 257 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_disable_all() 283 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_drain() 386 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_brs_sched_task() 407 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in perf_amd_brs_lopwr_cb()
|
H A D | lbr.c | 99 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_filter() 162 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_read() 324 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_reset() 343 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_add() 363 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_del() 378 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_sched_task() 391 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_enable_all() 412 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_disable_all()
|
/kernel/linux/linux-6.6/arch/xtensa/kernel/ |
H A D | hw_breakpoint.c | 178 i = alloc_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_install_hw_breakpoint() 185 i = alloc_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_install_hw_breakpoint() 216 i = free_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_uninstall_hw_breakpoint() 224 i = free_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_uninstall_hw_breakpoint() 269 struct perf_event *bp = this_cpu_ptr(wp_on_reg)[i]; in restore_dbreak() 281 struct perf_event **bp = this_cpu_ptr(bp_on_reg); in check_hw_breakpoint() 290 struct perf_event **bp = this_cpu_ptr(wp_on_reg); in check_hw_breakpoint()
|
/kernel/linux/linux-5.10/arch/x86/events/intel/ |
H A D | bts.c | 228 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_start() 256 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 257 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_start() 287 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_stop() 301 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 302 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_stop() 335 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_enable_local() 355 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_disable_local() 446 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 447 struct bts_ctx *bts = this_cpu_ptr( in intel_bts_interrupt() [all...] |
/kernel/linux/linux-6.6/arch/x86/events/intel/ |
H A D | bts.c | 234 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_start() 262 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 263 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_start() 293 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_stop() 307 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 308 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_stop() 341 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_enable_local() 361 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_disable_local() 452 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 453 struct bts_ctx *bts = this_cpu_ptr( in intel_bts_interrupt() [all...] |
/kernel/linux/linux-5.10/kernel/time/ |
H A D | tick-sched.c | 313 irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); in tick_nohz_full_kick() 461 ts = this_cpu_ptr(&tick_cpu_sched); in __tick_nohz_task_switch() 560 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_tick_stopped() 1052 __tick_nohz_idle_stop_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_stop_tick() 1057 tick_nohz_retain_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_retain_tick() 1078 ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_enter() 1098 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_exit() 1111 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_got_tick() 1141 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_sleep_length() 1191 struct tick_sched *ts = this_cpu_ptr( in tick_nohz_get_idle_calls() [all...] |
/kernel/linux/linux-6.6/kernel/time/ |
H A D | tick-sched.c | 344 irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); in tick_nohz_full_kick() 529 ts = this_cpu_ptr(&tick_cpu_sched); in __tick_nohz_task_switch() 626 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_tick_stopped() 1110 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_stop_tick() 1146 tick_nohz_retain_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_retain_tick() 1167 ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_enter() 1187 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_exit() 1200 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_got_tick() 1234 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_sleep_length() 1284 struct tick_sched *ts = this_cpu_ptr( in tick_nohz_get_idle_calls() [all...] |
/kernel/linux/linux-5.10/arch/s390/kernel/ |
H A D | perf_cpum_cf_diag.c | 109 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_trailer() 132 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_enable() 158 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_disable() 371 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_getctrset() 523 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_start() 524 struct cf_diag_csd *csd = this_cpu_ptr(&cf_diag_csd); in cf_diag_start() 546 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_stop() 547 struct cf_diag_csd *csd = this_cpu_ptr(&cf_diag_csd); in cf_diag_stop() 566 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_add() 590 struct cpu_cf_events *cpuhw = this_cpu_ptr( in cf_diag_del() [all...] |
/kernel/linux/linux-6.6/arch/arm64/kvm/ |
H A D | vmid.c | 135 atomic64_set(this_cpu_ptr(&active_vmids), VMID_ACTIVE_INVALID); in kvm_arm_vmid_clear_active() 155 old_active_vmid = atomic64_read(this_cpu_ptr(&active_vmids)); in kvm_arm_vmid_update() 157 0 != atomic64_cmpxchg_relaxed(this_cpu_ptr(&active_vmids), in kvm_arm_vmid_update() 168 atomic64_set(this_cpu_ptr(&active_vmids), vmid); in kvm_arm_vmid_update()
|
/kernel/linux/linux-6.6/kernel/irq/ |
H A D | ipi-mux.c | 31 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_mask() 38 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_unmask() 56 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_send_mask() 122 struct ipi_mux_cpu *icpu = this_cpu_ptr(ipi_mux_pcpu); in ipi_mux_process()
|