/kernel/linux/linux-5.10/kernel/sched/ |
H A D | pelt.h | 49 unsigned int enqueued; in cfs_se_util_change() local 55 enqueued = avg->util_est.enqueued; in cfs_se_util_change() 56 if (!(enqueued & UTIL_AVG_UNCHANGED)) in cfs_se_util_change() 60 enqueued &= ~UTIL_AVG_UNCHANGED; in cfs_se_util_change() 61 WRITE_ONCE(avg->util_est.enqueued, enqueued); in cfs_se_util_change()
|
H A D | fair.c | 342 * enqueued) or force our parent to appear after us when it is in list_add_leaf_cfs_rq() 343 * enqueued. The fact that we always enqueue bottom-up in list_add_leaf_cfs_rq() 428 /* Do the two (enqueued) entities belong to the same group ? */ 1039 * Task is being enqueued - update stats: 2772 * without p->mm even though we still had it when we enqueued this in task_numa_work() 3953 return max(ue.ewma, (ue.enqueued & ~UTIL_AVG_UNCHANGED)); in _task_util_est() 3996 unsigned int enqueued; in util_est_enqueue() local 4002 enqueued = cfs_rq->avg.util_est.enqueued; in util_est_enqueue() 4003 enqueued in util_est_enqueue() 4012 unsigned int enqueued; util_est_dequeue() local [all...] |
H A D | debug.c | 606 cfs_rq->avg.util_est.enqueued); in print_cfs_rq() 1029 PM(se.avg.util_est.enqueued, ~UTIL_AVG_UNCHANGED); in proc_sched_show_task()
|
H A D | sched.h | 2784 READ_ONCE(rq->cfs.avg.util_est.enqueued)); in cpu_util_cfs()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | pelt.h | 49 unsigned int enqueued; in cfs_se_util_change() local 55 enqueued = avg->util_est.enqueued; in cfs_se_util_change() 56 if (!(enqueued & UTIL_AVG_UNCHANGED)) in cfs_se_util_change() 60 enqueued &= ~UTIL_AVG_UNCHANGED; in cfs_se_util_change() 61 WRITE_ONCE(avg->util_est.enqueued, enqueued); in cfs_se_util_change()
|
H A D | fair.c | 410 * enqueued) or force our parent to appear after us when it is in list_add_leaf_cfs_rq() 411 * enqueued. The fact that we always enqueue bottom-up in list_add_leaf_cfs_rq() 496 /* Do the two (enqueued) entities belong to the same group ? */ 1326 * Task is being enqueued - update stats: 3296 * without p->mm even though we still had it when we enqueued this in task_numa_work() 4799 return max(ue.ewma, (ue.enqueued & ~UTIL_AVG_UNCHANGED)); in _task_util_est() 4842 unsigned int enqueued; in util_est_enqueue() local 4848 enqueued = cfs_rq->avg.util_est.enqueued; in util_est_enqueue() 4849 enqueued in util_est_enqueue() 4858 unsigned int enqueued; util_est_dequeue() local [all...] |
H A D | debug.c | 684 cfs_rq->avg.util_est.enqueued); in print_cfs_rq() 1100 PM(se.avg.util_est.enqueued, ~UTIL_AVG_UNCHANGED); in proc_sched_show_task()
|
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | reada.c | 33 * pointers that lie in the given range will also be enqueued. The reads will 778 u64 enqueued; in reada_start_for_fsdevs() local 783 enqueued = 0; in reada_start_for_fsdevs() 787 enqueued += reada_start_machine_dev(device); in reada_start_for_fsdevs() 789 total += enqueued; in reada_start_for_fsdevs() 790 } while (enqueued && total < 10000); in reada_start_for_fsdevs() 799 u64 enqueued = 0; in __reada_start_machine() local 803 enqueued += reada_start_for_fsdevs(fs_devices); in __reada_start_machine() 805 enqueued += reada_start_for_fsdevs(seed_devs); in __reada_start_machine() 808 if (enqueued in __reada_start_machine() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-xsk.c | 397 int total_enqueued, enqueued; in dpaa2_xsk_tx() local 428 enqueued = 0; in dpaa2_xsk_tx() 432 batch - total_enqueued, &enqueued); in dpaa2_xsk_tx() 438 total_enqueued += enqueued; in dpaa2_xsk_tx()
|
H A D | dpaa2-eth.c | 338 int total_enqueued = 0, retries = 0, enqueued; in dpaa2_eth_xdp_flush() local 351 0, num_fds - total_enqueued, &enqueued); in dpaa2_eth_xdp_flush() 356 total_enqueued += enqueued; in dpaa2_eth_xdp_flush() 369 int enqueued, i; in dpaa2_eth_xdp_tx_flush() local 374 enqueued = dpaa2_eth_xdp_flush(priv, fq, &fq->xdp_tx_fds); in dpaa2_eth_xdp_tx_flush() 377 percpu_stats->tx_packets += enqueued; in dpaa2_eth_xdp_tx_flush() 379 for (i = 0; i < enqueued; i++) { in dpaa2_eth_xdp_tx_flush() 383 for (i = enqueued; i < fq->xdp_tx_fds.num; i++) { in dpaa2_eth_xdp_tx_flush() 1399 int total_enqueued = 0, retries = 0, enqueued; in __dpaa2_eth_tx() local 1493 prio, num_fds - total_enqueued, &enqueued); in __dpaa2_eth_tx() 2863 int enqueued, i, err; dpaa2_eth_xdp_xmit() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-eth.c | 257 int total_enqueued = 0, retries = 0, enqueued; in dpaa2_eth_xdp_flush() local 270 0, num_fds - total_enqueued, &enqueued); in dpaa2_eth_xdp_flush() 275 total_enqueued += enqueued; in dpaa2_eth_xdp_flush() 288 int enqueued, i; in dpaa2_eth_xdp_tx_flush() local 293 enqueued = dpaa2_eth_xdp_flush(priv, fq, &fq->xdp_tx_fds); in dpaa2_eth_xdp_tx_flush() 296 percpu_stats->tx_packets += enqueued; in dpaa2_eth_xdp_tx_flush() 298 for (i = 0; i < enqueued; i++) { in dpaa2_eth_xdp_tx_flush() 302 for (i = enqueued; i < fq->xdp_tx_fds.num; i++) { in dpaa2_eth_xdp_tx_flush() 1845 * - stop MAC Rx and wait for all Rx frames to be enqueued to software in dpaa2_eth_stop() 1848 * frames are enqueued bac in dpaa2_eth_stop() 2348 int enqueued, i, err; dpaa2_eth_xdp_xmit() local [all...] |
/kernel/linux/linux-5.10/drivers/firewire/ |
H A D | net.c | 201 u8 enqueued; member 843 free = (ptask->outstanding_pkts == 0 && ptask->enqueued); in fwnet_transmit_packet_done() 920 free = ptask->enqueued; in fwnet_transmit_packet_failed() 1021 free = (ptask->outstanding_pkts == 0 && !ptask->enqueued); in fwnet_send_packet() 1023 ptask->enqueued = true; in fwnet_send_packet() 1040 free = (ptask->outstanding_pkts == 0 && !ptask->enqueued); in fwnet_send_packet() 1042 ptask->enqueued = true; in fwnet_send_packet() 1346 ptask->enqueued = 0; in fwnet_tx()
|
/kernel/linux/linux-6.6/drivers/firewire/ |
H A D | net.c | 201 u8 enqueued; member 833 free = (ptask->outstanding_pkts == 0 && ptask->enqueued); in fwnet_transmit_packet_done() 910 free = ptask->enqueued; in fwnet_transmit_packet_failed() 1011 free = (ptask->outstanding_pkts == 0 && !ptask->enqueued); in fwnet_send_packet() 1013 ptask->enqueued = true; in fwnet_send_packet() 1030 free = (ptask->outstanding_pkts == 0 && !ptask->enqueued); in fwnet_send_packet() 1032 ptask->enqueued = true; in fwnet_send_packet() 1336 ptask->enqueued = 0; in fwnet_tx()
|
/kernel/linux/linux-5.10/arch/arc/kernel/ |
H A D | entry-compact.S | 210 ; off to "C" code where SIGBUS in enqueued
|
/kernel/linux/linux-6.6/arch/arc/kernel/ |
H A D | entry-compact.S | 210 ; off to "C" code where SIGBUS in enqueued
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | sched.h | 387 * @enqueued: instantaneous estimated utilization of a task/cpu 397 * The enqueued attribute has a slightly different meaning for tasks and cpus: 399 * - cfs_rq: the sum of util_est.enqueued for each RUNNABLE task on that CPU 400 * Thus, the util_est.enqueued of a task represents the contribution on the 401 * estimated utilization of the CPU where that task is currently enqueued. 410 * This information is mapped into the MSB bit of util_est.enqueued at dequeue 411 * time. Since max value of util_est.enqueued for a task is 1024 (PELT util_avg 415 unsigned int enqueued; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | sched.h | 464 * @enqueued: instantaneous estimated utilization of a task/cpu 474 * The enqueued attribute has a slightly different meaning for tasks and cpus: 476 * - cfs_rq: the sum of util_est.enqueued for each RUNNABLE task on that CPU 477 * Thus, the util_est.enqueued of a task represents the contribution on the 478 * estimated utilization of the CPU where that task is currently enqueued. 487 * This information is mapped into the MSB bit of util_est.enqueued at dequeue 488 * time. Since max value of util_est.enqueued for a task is 1024 (PELT util_avg 492 unsigned int enqueued; member
|
/kernel/linux/linux-5.10/drivers/block/ |
H A D | skd_main.c | 1459 int limit, int *enqueued) in skd_isr_completion_posted() 1458 skd_isr_completion_posted(struct skd_device *skdev, int limit, int *enqueued) skd_isr_completion_posted() argument
|