/kernel/linux/linux-5.10/fs/fscache/ |
H A D | stats.c | 142 atomic_read(&fscache_n_cookie_index), in fscache_stats_show() 143 atomic_read(&fscache_n_cookie_data), in fscache_stats_show() 144 atomic_read(&fscache_n_cookie_special)); in fscache_stats_show() 147 atomic_read(&fscache_n_object_alloc), in fscache_stats_show() 148 atomic_read(&fscache_n_object_no_alloc), in fscache_stats_show() 149 atomic_read(&fscache_n_object_avail), in fscache_stats_show() 150 atomic_read(&fscache_n_object_dead)); in fscache_stats_show() 152 atomic_read(&fscache_n_checkaux_none), in fscache_stats_show() 153 atomic_read(&fscache_n_checkaux_okay), in fscache_stats_show() 154 atomic_read( in fscache_stats_show() [all...] |
H A D | histogram.c | 37 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); in fscache_histogram_show() 38 n[1] = atomic_read(&fscache_ops_histogram[index]); in fscache_histogram_show() 39 n[2] = atomic_read(&fscache_objs_histogram[index]); in fscache_histogram_show() 40 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); in fscache_histogram_show() 41 n[4] = atomic_read(&fscache_retrieval_histogram[index]); in fscache_histogram_show()
|
/kernel/linux/linux-6.6/fs/fscache/ |
H A D | stats.c | 59 atomic_read(&fscache_n_cookies), in fscache_stats_show() 60 atomic_read(&fscache_n_volumes), in fscache_stats_show() 61 atomic_read(&fscache_n_volumes_collision), in fscache_stats_show() 62 atomic_read(&fscache_n_volumes_nomem) in fscache_stats_show() 66 atomic_read(&fscache_n_acquires), in fscache_stats_show() 67 atomic_read(&fscache_n_acquires_ok), in fscache_stats_show() 68 atomic_read(&fscache_n_acquires_oom)); in fscache_stats_show() 71 atomic_read(&fscache_n_cookies_lru), in fscache_stats_show() 72 atomic_read(&fscache_n_cookies_lru_expired), in fscache_stats_show() 73 atomic_read( in fscache_stats_show() [all...] |
/kernel/linux/linux-6.6/fs/netfs/ |
H A D | stats.c | 34 atomic_read(&netfs_n_rh_readahead), in netfs_stats_show() 35 atomic_read(&netfs_n_rh_readpage), in netfs_stats_show() 36 atomic_read(&netfs_n_rh_write_begin), in netfs_stats_show() 37 atomic_read(&netfs_n_rh_write_zskip), in netfs_stats_show() 38 atomic_read(&netfs_n_rh_rreq), in netfs_stats_show() 39 atomic_read(&netfs_n_rh_sreq)); in netfs_stats_show() 41 atomic_read(&netfs_n_rh_zero), in netfs_stats_show() 42 atomic_read(&netfs_n_rh_short_read), in netfs_stats_show() 43 atomic_read(&netfs_n_rh_write_zskip)); in netfs_stats_show() 45 atomic_read( in netfs_stats_show() [all...] |
/kernel/linux/linux-6.6/net/rxrpc/ |
H A D | proc.c | 181 atomic_read(&conn->active), in rxrpc_connection_seq_show() 350 atomic_read(&local->active_users), in rxrpc_local_seq_show() 406 atomic_read(&rxnet->stat_tx_data_send), in rxrpc_stats_show() 407 atomic_read(&rxnet->stat_tx_data_send_frag), in rxrpc_stats_show() 408 atomic_read(&rxnet->stat_tx_data_send_fail)); in rxrpc_stats_show() 411 atomic_read(&rxnet->stat_tx_data), in rxrpc_stats_show() 412 atomic_read(&rxnet->stat_tx_data_retrans), in rxrpc_stats_show() 413 atomic_read(&rxnet->stat_tx_data_underflow), in rxrpc_stats_show() 414 atomic_read(&rxnet->stat_tx_data_cwnd_reset)); in rxrpc_stats_show() 417 atomic_read( in rxrpc_stats_show() [all...] |
/kernel/linux/linux-5.10/net/netfilter/ipvs/ |
H A D | ip_vs_nq.c | 45 return atomic_read(&dest->activeconns) + 1; in ip_vs_nq_dest_overhead() 77 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 83 if (atomic_read(&dest->activeconns) == 0) { in ip_vs_nq_schedule() 90 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 91 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 107 atomic_read(&least->activeconns), in ip_vs_nq_schedule() 109 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
H A D | ip_vs_lblcr.c | 173 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 189 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 190 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 202 atomic_read(&least->activeconns), in ip_vs_dest_set_min() 204 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 222 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 235 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 236 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 237 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() 247 atomic_read( in ip_vs_dest_set_max() [all...] |
H A D | ip_vs_sed.c | 49 return atomic_read(&dest->activeconns) + 1; in ip_vs_sed_dest_overhead() 80 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 97 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 98 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 108 atomic_read(&least->activeconns), in ip_vs_sed_schedule() 110 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
H A D | ip_vs_wlc.c | 52 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 69 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 70 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 80 atomic_read(&least->activeconns), in ip_vs_wlc_schedule() 82 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
/kernel/linux/linux-6.6/net/netfilter/ipvs/ |
H A D | ip_vs_nq.c | 45 return atomic_read(&dest->activeconns) + 1; in ip_vs_nq_dest_overhead() 77 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 83 if (atomic_read(&dest->activeconns) == 0) { in ip_vs_nq_schedule() 90 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 91 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 107 atomic_read(&least->activeconns), in ip_vs_nq_schedule() 109 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
H A D | ip_vs_lblcr.c | 173 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 189 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 190 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 202 atomic_read(&least->activeconns), in ip_vs_dest_set_min() 204 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 222 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 235 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 236 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 237 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() 247 atomic_read( in ip_vs_dest_set_max() [all...] |
H A D | ip_vs_sed.c | 49 return atomic_read(&dest->activeconns) + 1; in ip_vs_sed_dest_overhead() 80 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 97 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 98 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 108 atomic_read(&least->activeconns), in ip_vs_sed_schedule() 110 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
H A D | ip_vs_wlc.c | 52 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 69 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 70 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 80 atomic_read(&least->activeconns), in ip_vs_wlc_schedule() 82 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_ring.h | 71 const unsigned int idx = atomic_read(var); in pvrdma_idx() 80 __u32 idx = atomic_read(var) + 1; /* Increment. */ in pvrdma_idx_ring_inc() 89 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_space() 90 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_space() 103 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_data() 104 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_data()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_ring.h | 71 const unsigned int idx = atomic_read(var); in pvrdma_idx() 80 __u32 idx = atomic_read(var) + 1; /* Increment. */ in pvrdma_idx_ring_inc() 89 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_space() 90 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_space() 103 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_data() 104 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_data()
|
/kernel/linux/linux-5.10/drivers/lightnvm/ |
H A D | pblk-rl.c | 31 rb_space = atomic_read(&rl->rb_space); in pblk_rl_is_limit() 38 int rb_user_cnt = atomic_read(&rl->rb_user_cnt); in pblk_rl_user_may_insert() 39 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_user_may_insert() 52 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_inserted() 60 int rb_gc_cnt = atomic_read(&rl->rb_gc_cnt); in pblk_rl_gc_may_insert() 100 return atomic_read(&rl->free_blocks); in pblk_rl_nr_free_blks() 105 return atomic_read(&rl->free_user_blocks); in pblk_rl_nr_user_free_blks() 113 int werr_gc_needed = atomic_read(&rl->werr_lines); in __pblk_rl_update_rates() 161 int blk_in_line = atomic_read(&line->blk_in_line); in pblk_rl_free_lines_inc() 173 int blk_in_line = atomic_read( in pblk_rl_free_lines_dec() [all...] |
/kernel/linux/linux-5.10/sound/core/seq/ |
H A D | seq_lock.c | 16 if (atomic_read(lockp) < 0) { in snd_use_lock_sync_helper() 17 pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper() 20 while (atomic_read(lockp) > 0) { in snd_use_lock_sync_helper() 22 pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper()
|
/kernel/linux/linux-6.6/sound/core/seq/ |
H A D | seq_lock.c | 16 if (atomic_read(lockp) < 0) { in snd_use_lock_sync_helper() 17 pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper() 20 while (atomic_read(lockp) > 0) { in snd_use_lock_sync_helper() 22 pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | membarrier.c | 81 atomic_read(&mm->membarrier_state)); in ipi_sync_rq_state() 185 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 192 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 198 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 204 (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1)) in membarrier_private_expedited() 296 int membarrier_state = atomic_read(&mm->membarrier_state); in sync_runqueues_membarrier_state() 300 if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) { in sync_runqueues_membarrier_state() 359 if (atomic_read(&mm->membarrier_state) & in membarrier_register_global_expedited() 399 if ((atomic_read(&mm->membarrier_state) & ready_state) == ready_state) in membarrier_register_private_expedited()
|
/kernel/linux/linux-5.10/mm/ |
H A D | zswapd_control.c | 67 return atomic_read(&zram_wm_ratio); in get_zram_wm_ratio() 72 return atomic_read(&compress_ratio); in get_compress_ratio() 77 return atomic_read(&inactive_file_ratio); in get_inactive_file_ratio() 82 return atomic_read(&active_file_ratio); in get_active_file_ratio() 87 return atomic_read(&avail_buffers); in get_avail_buffers() 92 return atomic_read(&min_avail_buffers); in get_min_avail_buffers() 97 return atomic_read(&high_avail_buffers); in get_high_avail_buffers() 102 return atomic_read(&max_reclaim_size); in get_zswapd_max_reclaim_size() 163 if (atomic_read(&min_avail_buffers) == 0) in avail_buffers_params_write() 531 seq_printf(m, "avail_buffers: %u\n", atomic_read( in avail_buffers_params_show() [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | zswapd_control.c | 68 return atomic_read(&zram_wm_ratio); in get_zram_wm_ratio() 73 return atomic_read(&compress_ratio); in get_compress_ratio() 78 return atomic_read(&inactive_file_ratio); in get_inactive_file_ratio() 83 return atomic_read(&active_file_ratio); in get_active_file_ratio() 88 return atomic_read(&avail_buffers); in get_avail_buffers() 93 return atomic_read(&min_avail_buffers); in get_min_avail_buffers() 98 return atomic_read(&high_avail_buffers); in get_high_avail_buffers() 103 return atomic_read(&max_reclaim_size); in get_zswapd_max_reclaim_size() 164 if (atomic_read(&min_avail_buffers) == 0) in avail_buffers_params_write() 532 seq_printf(m, "avail_buffers: %u\n", atomic_read( in avail_buffers_params_show() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/bcm/ |
H A D | util.c | 374 atomic_read(&ipriv->session_count)); in spu_debugfs_read() 377 atomic_read(&ipriv->stream_count)); in spu_debugfs_read() 380 atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER])); in spu_debugfs_read() 383 atomic_read(&ipriv->op_counts[SPU_OP_CIPHER])); in spu_debugfs_read() 386 op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]); in spu_debugfs_read() 397 atomic_read(&ipriv->op_counts[SPU_OP_HASH])); in spu_debugfs_read() 399 op_cnt = atomic_read(&ipriv->hash_cnt[alg]); in spu_debugfs_read() 409 atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC])); in spu_debugfs_read() 412 atomic_read(&ipriv->op_counts[SPU_OP_HMAC])); in spu_debugfs_read() 414 op_cnt = atomic_read( in spu_debugfs_read() [all...] |
/kernel/linux/linux-6.6/drivers/crypto/bcm/ |
H A D | util.c | 376 atomic_read(&ipriv->session_count)); in spu_debugfs_read() 379 atomic_read(&ipriv->stream_count)); in spu_debugfs_read() 382 atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER])); in spu_debugfs_read() 385 atomic_read(&ipriv->op_counts[SPU_OP_CIPHER])); in spu_debugfs_read() 388 op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]); in spu_debugfs_read() 399 atomic_read(&ipriv->op_counts[SPU_OP_HASH])); in spu_debugfs_read() 401 op_cnt = atomic_read(&ipriv->hash_cnt[alg]); in spu_debugfs_read() 411 atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC])); in spu_debugfs_read() 414 atomic_read(&ipriv->op_counts[SPU_OP_HMAC])); in spu_debugfs_read() 416 op_cnt = atomic_read( in spu_debugfs_read() [all...] |
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | membarrier.c | 210 atomic_read(&mm->membarrier_state)); in ipi_sync_rq_state() 242 membarrier_state = atomic_read(&next_mm->membarrier_state); in membarrier_update_current_mm() 323 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 330 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 336 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 342 (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1)) in membarrier_private_expedited() 433 int membarrier_state = atomic_read(&mm->membarrier_state); in sync_runqueues_membarrier_state() 437 if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) { in sync_runqueues_membarrier_state() 495 if (atomic_read(&mm->membarrier_state) & in membarrier_register_global_expedited() 535 if ((atomic_read( in membarrier_register_private_expedited() [all...] |
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | locking.c | 145 WARN_ON(atomic_read(&eb->spinning_readers) == 0); in btrfs_assert_spinning_readers_put() 161 BUG_ON(!atomic_read(&eb->read_locks)); in btrfs_assert_tree_read_locked() 344 if (READ_ONCE(eb->blocking_writers) || atomic_read(&eb->blocking_readers)) in btrfs_try_tree_write_lock() 349 if (READ_ONCE(eb->blocking_writers) || atomic_read(&eb->blocking_readers)) { in btrfs_try_tree_write_lock() 406 WARN_ON(atomic_read(&eb->blocking_readers) == 0); in btrfs_tree_read_unlock_blocking() 429 wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0); 433 if (atomic_read(&eb->blocking_readers) || 620 if (atomic_read(&lock->readers)) in btrfs_drew_try_write_lock() 627 if (atomic_read(&lock->readers)) { in btrfs_drew_try_write_lock() 640 wait_event(lock->pending_writers, !atomic_read( in btrfs_drew_write_lock() [all...] |