Home
last modified time | relevance | path

Searched refs:READ_ONCE (Results 1 - 25 of 2457) sorted by relevance

12345678910>>...99

/kernel/linux/linux-6.6/kernel/rcu/
H A Dtree_stall.h30 int cpu_stall_timeout = READ_ONCE(rcu_exp_cpu_stall_timeout); in rcu_exp_jiffies_till_stall_check()
59 int till_stall_check = READ_ONCE(rcu_cpu_stall_timeout); in rcu_jiffies_till_stall_check()
103 return !time_before(j, READ_ONCE(rcu_state.gp_start) + d); in rcu_gp_might_be_stalled()
181 rcu_state.n_force_qs_gpstart = READ_ONCE(rcu_state.n_force_qs); in record_gp_stall_check_time()
200 if (!READ_ONCE(rcu_kick_kthreads)) in rcu_stall_kick_kthreads()
202 j = READ_ONCE(rcu_state.jiffies_kick_kthreads); in rcu_stall_kick_kthreads()
204 (rcu_gp_in_progress() || READ_ONCE(rcu_state.gp_flags))) { in rcu_stall_kick_kthreads()
409 unsigned long j = jiffies - READ_ONCE(rcu_state.gp_activity); in rcu_is_gp_kthread_starving()
430 j = jiffies - READ_ONCE(rdp->rcuc_activity); in rcu_is_rcuc_kthread_starving()
538 data_race(READ_ONCE(rcu_stat in rcu_check_gp_kthread_starvation()
[all...]
H A Dsync.c78 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_func()
79 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_func()
155 wait_event(rsp->gp_wait, READ_ONCE(rsp->gp_state) >= GP_PASSED); in rcu_sync_enter()
170 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_exit()
171 WARN_ON_ONCE(READ_ONCE(rsp->gp_count) == 0); in rcu_sync_exit()
193 WARN_ON_ONCE(READ_ONCE(rsp->gp_count)); in rcu_sync_dtor()
194 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_dtor()
H A Dsrcutiny.c99 int newval = READ_ONCE(ssp->srcu_lock_nesting[idx]) - 1; in __srcu_read_unlock()
102 if (!newval && READ_ONCE(ssp->srcu_gp_waiting) && in_task()) in __srcu_read_unlock()
120 if (ssp->srcu_gp_running || ULONG_CMP_GE(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max))) in srcu_drive_gp()
133 swait_event_exclusive(ssp->srcu_wq, !READ_ONCE(ssp->srcu_lock_nesting[idx])); in srcu_drive_gp()
153 if (ULONG_CMP_LT(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max))) in srcu_drive_gp()
163 if (ULONG_CMP_GE(READ_ONCE(ssp->srcu_idx_max), cookie)) in srcu_gp_start_if_needed()
166 if (!READ_ONCE(ssp->srcu_gp_running)) { in srcu_gp_start_if_needed()
228 ret = (READ_ONCE(ssp->srcu_idx) + 3) & ~0x1; in get_state_synchronize_srcu()
255 unsigned long cur_s = READ_ONCE(ssp->srcu_idx); in poll_state_synchronize_srcu()
/kernel/linux/linux-6.6/net/ipv4/
H A Dtcp_plb.c30 if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled)) in tcp_plb_update_state()
34 if (cong_ratio < READ_ONCE(net->ipv4.sysctl_tcp_plb_cong_thresh)) in tcp_plb_update_state()
37 READ_ONCE(net->ipv4.sysctl_tcp_plb_rehash_rounds)) in tcp_plb_update_state()
52 if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled)) in tcp_plb_check_rehash()
56 READ_ONCE(net->ipv4.sysctl_tcp_plb_rehash_rounds); in tcp_plb_check_rehash()
58 idle_rehash = READ_ONCE(net->ipv4.sysctl_tcp_plb_idle_rehash_rounds) && in tcp_plb_check_rehash()
61 READ_ONCE(net->ipv4.sysctl_tcp_plb_idle_rehash_rounds); in tcp_plb_check_rehash()
71 max_suspend = 2 * READ_ONCE(net->ipv4.sysctl_tcp_plb_suspend_rto_sec) * HZ; in tcp_plb_check_rehash()
96 if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled)) in tcp_plb_update_state_upon_rto()
99 pause = READ_ONCE(ne in tcp_plb_update_state_upon_rto()
[all...]
H A Dtcp_timer.c33 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_rto_to_user_timeout()
50 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_probe0_to_user_timeout()
72 WRITE_ONCE(sk->sk_err, READ_ONCE(sk->sk_err_soft) ? : ETIMEDOUT); in tcp_write_err()
115 if (READ_ONCE(sk->sk_err_soft)) in tcp_out_of_resources()
148 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries()
151 if (READ_ONCE(sk->sk_err_soft) && !alive) in tcp_orphan_retries()
168 if (!READ_ONCE(net->ipv4.sysctl_tcp_mtu_probing)) in tcp_mtu_probing()
176 mss = min(READ_ONCE(net->ipv4.sysctl_tcp_base_mss), mss); in tcp_mtu_probing()
177 mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_mtu_probe_floor)); in tcp_mtu_probing()
178 mss = max(mss, READ_ONCE(ne in tcp_mtu_probing()
[all...]
/kernel/linux/linux-5.10/kernel/rcu/
H A Dtree_stall.h30 int till_stall_check = READ_ONCE(rcu_cpu_stall_timeout); in rcu_jiffies_till_stall_check()
74 return !time_before(j, READ_ONCE(rcu_state.gp_start) + d); in rcu_gp_might_be_stalled()
144 rcu_state.n_force_qs_gpstart = READ_ONCE(rcu_state.n_force_qs); in record_gp_stall_check_time()
163 if (!READ_ONCE(rcu_kick_kthreads)) in rcu_stall_kick_kthreads()
165 j = READ_ONCE(rcu_state.jiffies_kick_kthreads); in rcu_stall_kick_kthreads()
167 (rcu_gp_in_progress() || READ_ONCE(rcu_state.gp_flags))) { in rcu_stall_kick_kthreads()
388 unsigned long j = jiffies - READ_ONCE(rcu_state.gp_activity); in rcu_is_gp_kthread_starving()
531 if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall))) in print_other_cpu_stall()
578 if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall))) in print_cpu_stall()
607 if ((rcu_stall_is_suppressed() && !READ_ONCE(rcu_kick_kthread in check_cpu_stall()
[all...]
H A Dtasks.h143 int j = READ_ONCE(i); // Prevent the compiler from reading twice in tasks_gp_state_getname()
166 if (needwake && READ_ONCE(rtp->kthread_ptr)) in call_rcu_tasks_generic()
214 READ_ONCE(rtp->cbs_head)); in rcu_tasks_kthread()
356 rtst = READ_ONCE(rcu_task_stall_timeout); in rcu_tasks_wait_gp()
408 if (t != current && READ_ONCE(t->on_rq) && !is_idle_task(t)) { in rcu_tasks_pertask()
410 t->rcu_tasks_nvcsw = READ_ONCE(t->nvcsw); in rcu_tasks_pertask()
445 if (!READ_ONCE(t->rcu_tasks_holdout) || in check_holdout_task()
446 t->rcu_tasks_nvcsw != READ_ONCE(t->nvcsw) || in check_holdout_task()
447 !READ_ONCE(t->on_rq) || in check_holdout_task()
804 int nq = READ_ONCE( in rcu_read_unlock_trace_special()
[all...]
H A Dsync.c78 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_func()
79 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_func()
155 wait_event(rsp->gp_wait, READ_ONCE(rsp->gp_state) >= GP_PASSED); in rcu_sync_enter()
170 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_exit()
171 WARN_ON_ONCE(READ_ONCE(rsp->gp_count) == 0); in rcu_sync_exit()
193 WARN_ON_ONCE(READ_ONCE(rsp->gp_count)); in rcu_sync_dtor()
194 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_dtor()
H A Drcu_segcblist.h15 return READ_ONCE(rclp->len); in rcu_cblist_n_cbs()
40 return !READ_ONCE(rsclp->head); in rcu_segcblist_empty()
49 return READ_ONCE(rsclp->len); in rcu_segcblist_n_cbs()
75 return !READ_ONCE(*READ_ONCE(rsclp->tails[seg])); in rcu_segcblist_restempty()
/kernel/linux/linux-6.6/io_uring/
H A Dfs.c60 ren->old_dfd = READ_ONCE(sqe->fd); in io_renameat_prep()
61 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_renameat_prep()
62 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_renameat_prep()
63 ren->new_dfd = READ_ONCE(sqe->len); in io_renameat_prep()
64 ren->flags = READ_ONCE(sqe->rename_flags); in io_renameat_prep()
114 un->dfd = READ_ONCE(sqe->fd); in io_unlinkat_prep()
116 un->flags = READ_ONCE(sqe->unlink_flags); in io_unlinkat_prep()
120 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_unlinkat_prep()
164 mkd->dfd = READ_ONCE(sqe->fd); in io_mkdirat_prep()
165 mkd->mode = READ_ONCE(sq in io_mkdirat_prep()
[all...]
H A Dsync.c32 sync->off = READ_ONCE(sqe->off); in io_sfr_prep()
33 sync->len = READ_ONCE(sqe->len); in io_sfr_prep()
34 sync->flags = READ_ONCE(sqe->sync_range_flags); in io_sfr_prep()
60 sync->flags = READ_ONCE(sqe->fsync_flags); in io_fsync_prep()
64 sync->off = READ_ONCE(sqe->off); in io_fsync_prep()
65 sync->len = READ_ONCE(sqe->len); in io_fsync_prep()
92 sync->off = READ_ONCE(sqe->off); in io_fallocate_prep()
93 sync->len = READ_ONCE(sqe->addr); in io_fallocate_prep()
94 sync->mode = READ_ONCE(sqe->len); in io_fallocate_prep()
H A Dstatx.c33 sx->dfd = READ_ONCE(sqe->fd); in io_statx_prep()
34 sx->mask = READ_ONCE(sqe->len); in io_statx_prep()
35 path = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_statx_prep()
36 sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_statx_prep()
37 sx->flags = READ_ONCE(sqe->statx_flags); in io_statx_prep()
/kernel/linux/linux-6.6/include/net/
H A Dbusy_poll.h37 return READ_ONCE(sysctl_net_busy_poll); in net_busy_loop_on()
42 return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current); in sk_can_busy_loop()
77 unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll); in busy_loop_timeout()
93 unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec); in sk_busy_loop_timeout()
108 unsigned int napi_id = READ_ONCE(sk->sk_napi_id); in sk_busy_loop()
112 READ_ONCE(sk->sk_prefer_busy_poll), in sk_busy_loop()
113 READ_ONCE(sk->sk_busy_poll_budget) ?: BUSY_POLL_BUDGET); in sk_busy_loop()
134 if (unlikely(READ_ONCE(sk->sk_napi_id) != skb->napi_id)) in sk_mark_napi_id()
156 if (!READ_ONCE(sk->sk_napi_id)) in __sk_mark_napi_id_once()
/kernel/linux/linux-6.6/include/linux/
H A Dsrcutiny.h67 idx = ((READ_ONCE(ssp->srcu_idx) + 1) & 0x2) >> 1; in __srcu_read_lock()
68 WRITE_ONCE(ssp->srcu_lock_nesting[idx], READ_ONCE(ssp->srcu_lock_nesting[idx]) + 1); in __srcu_read_lock()
88 idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1; in srcu_torture_stats_print()
91 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])), in srcu_torture_stats_print()
92 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])), in srcu_torture_stats_print()
93 data_race(READ_ONCE(ssp->srcu_idx)), in srcu_torture_stats_print()
94 data_race(READ_ONCE(ssp->srcu_idx_max))); in srcu_torture_stats_print()
/kernel/linux/linux-5.10/arch/arm64/include/asm/
H A Dpreempt.h12 return READ_ONCE(current_thread_info()->preempt.count); in preempt_count()
46 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_add()
53 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_sub()
61 u64 pc = READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
70 * interrupt occurring between the non-atomic READ_ONCE/WRITE_ONCE in __preempt_count_dec_and_test()
73 return !pc || !READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
78 u64 pc = READ_ONCE(current_thread_info()->preempt_count); in should_resched()
/kernel/linux/linux-5.10/arch/arm64/mm/
H A Dmmu.c161 pte_t old_pte = READ_ONCE(*ptep); in init_pte()
170 READ_ONCE(pte_val(*ptep)))); in init_pte()
185 pmd_t pmd = READ_ONCE(*pmdp); in alloc_init_cont_pte()
193 pmd = READ_ONCE(*pmdp); in alloc_init_cont_pte()
222 pmd_t old_pmd = READ_ONCE(*pmdp); in init_pmd()
236 READ_ONCE(pmd_val(*pmdp)))); in init_pmd()
242 pmd_val(old_pmd) != READ_ONCE(pmd_val(*pmdp))); in init_pmd()
256 pud_t pud = READ_ONCE(*pudp); in alloc_init_cont_pmd()
267 pud = READ_ONCE(*pudp); in alloc_init_cont_pmd()
307 p4d_t p4d = READ_ONCE(*p4d in alloc_init_pud()
[all...]
/kernel/linux/linux-6.6/arch/arm64/include/asm/
H A Dpreempt.h13 return READ_ONCE(current_thread_info()->preempt.count); in preempt_count()
47 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_add()
54 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_sub()
62 u64 pc = READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
71 * interrupt occurring between the non-atomic READ_ONCE/WRITE_ONCE in __preempt_count_dec_and_test()
74 return !pc || !READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
79 u64 pc = READ_ONCE(current_thread_info()->preempt_count); in should_resched()
/kernel/linux/linux-5.10/include/net/
H A Dbusy_poll.h34 return READ_ONCE(sysctl_net_busy_poll); in net_busy_loop_on()
39 return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current); in sk_can_busy_loop()
74 unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll); in busy_loop_timeout()
90 unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec); in sk_busy_loop_timeout()
105 unsigned int napi_id = READ_ONCE(sk->sk_napi_id); in sk_busy_loop()
139 if (!READ_ONCE(sk->sk_napi_id)) in sk_mark_napi_id_once()
/kernel/linux/linux-5.10/net/mptcp/
H A Dmptcp_diag.c128 info->mptcpi_subflows = READ_ONCE(msk->pm.subflows); in mptcp_diag_get_info()
129 info->mptcpi_add_addr_signal = READ_ONCE(msk->pm.add_addr_signaled); in mptcp_diag_get_info()
130 info->mptcpi_add_addr_accepted = READ_ONCE(msk->pm.add_addr_accepted); in mptcp_diag_get_info()
131 info->mptcpi_subflows_max = READ_ONCE(msk->pm.subflows_max); in mptcp_diag_get_info()
132 val = READ_ONCE(msk->pm.add_addr_signal_max); in mptcp_diag_get_info()
134 val = READ_ONCE(msk->pm.add_addr_accept_max); in mptcp_diag_get_info()
138 if (READ_ONCE(msk->can_ack)) in mptcp_diag_get_info()
141 info->mptcpi_token = READ_ONCE(msk->token); in mptcp_diag_get_info()
142 info->mptcpi_write_seq = READ_ONCE(msk->write_seq); in mptcp_diag_get_info()
144 info->mptcpi_rcv_nxt = READ_ONCE(ms in mptcp_diag_get_info()
[all...]
H A Dpm.c63 pm->subflows_max, READ_ONCE(pm->accept_subflow)); in mptcp_pm_allow_new_subflow()
66 if (!READ_ONCE(pm->accept_subflow)) in mptcp_pm_allow_new_subflow()
70 if (READ_ONCE(pm->accept_subflow)) { in mptcp_pm_allow_new_subflow()
104 if (!READ_ONCE(pm->work_pending)) in mptcp_pm_fully_established()
109 if (READ_ONCE(pm->work_pending)) in mptcp_pm_fully_established()
127 if (!READ_ONCE(pm->work_pending)) in mptcp_pm_subflow_established()
132 if (READ_ONCE(pm->work_pending)) in mptcp_pm_subflow_established()
149 READ_ONCE(pm->accept_addr)); in mptcp_pm_add_addr_received()
153 if (!READ_ONCE(pm->accept_addr)) in mptcp_pm_add_addr_received()
186 *echo = READ_ONCE(ms in mptcp_pm_add_addr_signal()
[all...]
/kernel/linux/linux-5.10/arch/s390/kernel/
H A Didle.c61 idle_count = READ_ONCE(idle->idle_count); in show_idle_count()
62 if (READ_ONCE(idle->clock_idle_enter)) in show_idle_count()
78 idle_time = READ_ONCE(idle->idle_time); in show_idle_time()
79 idle_enter = READ_ONCE(idle->clock_idle_enter); in show_idle_time()
80 idle_exit = READ_ONCE(idle->clock_idle_exit); in show_idle_time()
104 idle_enter = READ_ONCE(idle->clock_idle_enter); in arch_cpu_idle_time()
105 idle_exit = READ_ONCE(idle->clock_idle_exit); in arch_cpu_idle_time()
/kernel/linux/linux-5.10/drivers/powercap/
H A Didle_inject.c114 duration_us = READ_ONCE(ii_dev->run_duration_us); in idle_inject_timer_fn()
115 duration_us += READ_ONCE(ii_dev->idle_duration_us); in idle_inject_timer_fn()
144 play_idle_precise(READ_ONCE(ii_dev->idle_duration_us) * NSEC_PER_USEC, in idle_inject_fn()
145 READ_ONCE(ii_dev->latency_us) * NSEC_PER_USEC); in idle_inject_fn()
172 *run_duration_us = READ_ONCE(ii_dev->run_duration_us); in idle_inject_get_duration()
173 *idle_duration_us = READ_ONCE(ii_dev->idle_duration_us); in idle_inject_get_duration()
198 unsigned int idle_duration_us = READ_ONCE(ii_dev->idle_duration_us); in idle_inject_start()
199 unsigned int run_duration_us = READ_ONCE(ii_dev->run_duration_us); in idle_inject_start()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/
H A Den_port.c165 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats()
166 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats()
176 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats()
177 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats()
254 sw_rx_dropped += READ_ONCE(ring->dropped); in mlx4_en_DUMP_ETH_STATS()
255 priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok); in mlx4_en_DUMP_ETH_STATS()
256 priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none); in mlx4_en_DUMP_ETH_STATS()
257 priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete); in mlx4_en_DUMP_ETH_STATS()
258 priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages); in mlx4_en_DUMP_ETH_STATS()
259 priv->xdp_stats.rx_xdp_drop += READ_ONCE(rin in mlx4_en_DUMP_ETH_STATS()
[all...]
/kernel/linux/linux-5.10/arch/s390/include/asm/
H A Dpreempt.h17 return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED; in preempt_count()
25 old = READ_ONCE(S390_lowcore.preempt_count); in preempt_count_set()
44 return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED); in test_preempt_need_resched()
74 return unlikely(READ_ONCE(S390_lowcore.preempt_count) == in should_resched()
84 return READ_ONCE(S390_lowcore.preempt_count); in preempt_count()
/kernel/linux/linux-6.6/arch/s390/include/asm/
H A Dpreempt.h17 return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED; in preempt_count()
25 old = READ_ONCE(S390_lowcore.preempt_count); in preempt_count_set()
44 return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED); in test_preempt_need_resched()
74 return unlikely(READ_ONCE(S390_lowcore.preempt_count) == in should_resched()
84 return READ_ONCE(S390_lowcore.preempt_count); in preempt_count()

Completed in 13 milliseconds

12345678910>>...99