/kernel/linux/linux-5.10/kernel/sched/ |
H A D | loadavg.c | 22 * nr_active = 0; 24 * nr_active += cpu_of(cpu)->nr_running + cpu_of(cpu)->nr_uninterruptible; 26 * avenrun[n] = avenrun[0] * exp_n + nr_active * (1 - exp_n) 32 * to calculating nr_active. 37 * So assuming nr_active := 0 when we start out -- true per definition, we 81 long nr_active, delta = 0; in calc_load_fold_active() local 83 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 84 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active() 86 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 87 delta = nr_active in calc_load_fold_active() [all...] |
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | loadavg.c | 21 * nr_active = 0; 23 * nr_active += cpu_of(cpu)->nr_running + cpu_of(cpu)->nr_uninterruptible; 25 * avenrun[n] = avenrun[0] * exp_n + nr_active * (1 - exp_n) 31 * to calculating nr_active. 36 * So assuming nr_active := 0 when we start out -- true per definition, we 80 long nr_active, delta = 0; in calc_load_fold_active() local 82 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 83 nr_active += (int)this_rq->nr_uninterruptible; in calc_load_fold_active() 85 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 86 delta = nr_active in calc_load_fold_active() [all...] |
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | vmscan.h | 423 unsigned long nr_active, unsigned long nr_deactivated, 426 TP_ARGS(nid, nr_taken, nr_active, nr_deactivated, nr_referenced, priority, file), 431 __field(unsigned long, nr_active) 441 __entry->nr_active = nr_active; 448 TP_printk("nid=%d nr_taken=%ld nr_active=%ld nr_deactivated=%ld nr_referenced=%ld priority=%d flags=%s", 451 __entry->nr_active, __entry->nr_deactivated, __entry->nr_referenced,
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | vmscan.h | 437 unsigned long nr_active, unsigned long nr_deactivated, 440 TP_ARGS(nid, nr_taken, nr_active, nr_deactivated, nr_referenced, priority, file), 445 __field(unsigned long, nr_active) 455 __entry->nr_active = nr_active; 462 TP_printk("nid=%d nr_taken=%ld nr_active=%ld nr_deactivated=%ld nr_referenced=%ld priority=%d flags=%s", 465 __entry->nr_active, __entry->nr_deactivated, __entry->nr_referenced,
|
/kernel/linux/linux-5.10/block/ |
H A D | blk-mq.h | 209 atomic_inc(&hctx->nr_active); in __blk_mq_inc_active_requests() 217 atomic_dec(&hctx->nr_active); in __blk_mq_dec_active_requests() 224 return atomic_read(&hctx->nr_active); in __blk_mq_active_requests()
|
H A D | blk-mq-debugfs.c | 619 seq_printf(m, "%d\n", atomic_read(&hctx->nr_active)); in hctx_active_show()
|
/kernel/linux/linux-6.6/mm/ |
H A D | dmapool.c | 53 size_t nr_active; member 83 pool->name, pool->nr_active, in pools_show() 187 pool->nr_active++; in pool_block_pop() 377 if (pool->nr_active) { in dma_pool_destroy() 458 pool->nr_active--; in dma_pool_free()
|
H A D | zswapd.c | 596 u64 nr_active, nr_inactive, nr_zram, nr_eswap, zram_ratio; in zswapd_shrink_anon() local 609 nr_active = lruvec_lru_size(lruvec, LRU_ACTIVE_ANON, MAX_NR_ZONES); in zswapd_shrink_anon() 615 (nr_inactive + nr_active + nr_zram + nr_eswap + 1)); in zswapd_shrink_anon() 621 nr[LRU_ACTIVE_ANON] = nr_active >> (unsigned int)sc->priority; in zswapd_shrink_anon()
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/cell/spufs/ |
H A D | sched.c | 686 cbe_spu_info[node].nr_active--; in find_victim() 715 cbe_spu_info[node].nr_active++; in __spu_schedule() 756 cbe_spu_info[node].nr_active--; in spu_unschedule() 957 int nr_active = 0, node; in count_active_contexts() local 960 nr_active += cbe_spu_info[node].nr_active; in count_active_contexts() 961 nr_active += spu_prio->nr_waiting; in count_active_contexts() 963 return nr_active; in count_active_contexts()
|
/kernel/linux/linux-6.6/arch/powerpc/platforms/cell/spufs/ |
H A D | sched.c | 680 cbe_spu_info[node].nr_active--; in find_victim() 709 cbe_spu_info[node].nr_active++; in __spu_schedule() 750 cbe_spu_info[node].nr_active--; in spu_unschedule() 951 int nr_active = 0, node; in count_active_contexts() local 954 nr_active += cbe_spu_info[node].nr_active; in count_active_contexts() 955 nr_active += spu_prio->nr_waiting; in count_active_contexts() 957 return nr_active; in count_active_contexts()
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/prog_tests/ |
H A D | setget_sockopt.c | 64 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp() 144 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_ktls()
|
/kernel/linux/linux-6.6/block/ |
H A D | blk-mq.h | 279 atomic_inc(&hctx->nr_active); in __blk_mq_inc_active_requests() 288 atomic_sub(val, &hctx->nr_active); in __blk_mq_sub_active_requests() 300 return atomic_read(&hctx->nr_active); in __blk_mq_active_requests()
|
/kernel/linux/linux-5.10/mm/ |
H A D | zswapd.c | 595 u64 nr_active, nr_inactive, nr_zram, nr_eswap, zram_ratio; in zswapd_shrink_anon() local 608 nr_active = lruvec_lru_size(lruvec, LRU_ACTIVE_ANON, MAX_NR_ZONES); in zswapd_shrink_anon() 614 (nr_inactive + nr_active + nr_zram + nr_eswap + 1)); in zswapd_shrink_anon() 620 nr[LRU_ACTIVE_ANON] = nr_active >> (unsigned int)sc->priority; in zswapd_shrink_anon()
|
/kernel/linux/linux-5.10/fs/ceph/ |
H A D | mdsmap.c | 407 int i, nr_active = 0; in ceph_mdsmap_is_cluster_available() local 416 nr_active++; in ceph_mdsmap_is_cluster_available() 418 return nr_active > 0; in ceph_mdsmap_is_cluster_available()
|
/kernel/linux/linux-6.6/fs/ceph/ |
H A D | mdsmap.c | 425 int i, nr_active = 0; in ceph_mdsmap_is_cluster_available() local 434 nr_active++; in ceph_mdsmap_is_cluster_available() 436 return nr_active > 0; in ceph_mdsmap_is_cluster_available()
|
/kernel/linux/linux-5.10/kernel/ |
H A D | workqueue.c | 208 int nr_active; /* L: nr of active works */ member 1160 pwq->nr_active++; in pwq_activate_inactive_work() 1184 /* uncolored work items don't participate in flushing or nr_active */ in pwq_dec_nr_in_flight() 1190 pwq->nr_active--; in pwq_dec_nr_in_flight() 1193 if (pwq->nr_active < pwq->max_active) in pwq_dec_nr_in_flight() 1298 * on the inactive_works list, will confuse pwq->nr_active in try_to_grab_pending() 1495 if (likely(pwq->nr_active < pwq->max_active)) { in __queue_work() 1497 pwq->nr_active++; in __queue_work() 2546 if (pwq->nr_active && need_to_create_worker(pool)) { in rescuer_thread() 2966 drained = !pwq->nr_active in drain_workqueue() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
H A D | setget_sockopt.c | 21 int nr_active; variable 383 nr_active += !(bpf_test_sockopt(skops, sk) || in skops_sockopt()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | blk-mq.h | 139 * @nr_active: Number of active requests. Only used when a tag set is 142 atomic_t nr_active; member
|
H A D | perf_event.h | 791 * nr_active, and the list: 810 int nr_active; member 816 * Set when nr_events != nr_active, except tolerant to events not
|
/kernel/linux/linux-6.6/kernel/ |
H A D | workqueue.c | 243 * nr_active management and WORK_STRUCT_INACTIVE: 245 * When pwq->nr_active >= max_active, new work item is queued to 250 * nr_active and all work items in pwq->inactive_works are marked with 255 * in nr_active. For non-barrier work item, it is marked with 258 int nr_active; /* L: nr of active works */ member 290 * As sharing a single nr_active across multiple sockets can be very expensive, 302 atomic_t nr; /* per-node nr_active */ 357 struct wq_node_nr_active *node_nr_active[]; /* I: per-node nr_active */ 1467 * - %NULL for per-cpu workqueues as they don't need to use shared nr_active. 1579 return !pwq->nr_active in pwq_is_empty() [all...] |
/kernel/linux/linux-5.10/net/mptcp/ |
H A D | protocol.c | 1092 int i, nr_active = 0; in mptcp_subflow_get_send() local 1131 nr_active += !subflow->backup; in mptcp_subflow_get_send() 1148 pr_debug("msk=%p nr_active=%d ssk=%p:%lld backup=%p:%lld", in mptcp_subflow_get_send() 1149 msk, nr_active, send_info[0].ssk, send_info[0].ratio, in mptcp_subflow_get_send() 1153 if (!nr_active) in mptcp_subflow_get_send()
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | spu.h | 183 int nr_active; member
|
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | spu.h | 183 int nr_active; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | blk-mq.h | 405 * @nr_active: Number of active requests. Only used when a tag set is 408 atomic_t nr_active; member
|
/kernel/linux/linux-6.6/net/mptcp/ |
H A D | protocol.c | 1423 int i, nr_active = 0; in mptcp_subflow_get_send() local 1441 nr_active += !subflow->backup; in mptcp_subflow_get_send() 1460 if (!nr_active) in mptcp_subflow_get_send()
|