/kernel/linux/linux-5.10/include/trace/events/ |
H A D | page_pool.h | 17 s32 inflight, u32 hold, u32 release), 19 TP_ARGS(pool, inflight, hold, release), 23 __field(s32, inflight) 31 __entry->inflight = inflight; 37 TP_printk("page_pool=%p inflight=%d hold=%u release=%u cnt=%llu", 38 __entry->pool, __entry->inflight, __entry->hold,
|
H A D | wbt.h | 127 * @inflight: tracked writes inflight 132 int step, unsigned int inflight), 134 TP_ARGS(bdi, status, step, inflight), 140 __field(unsigned int, inflight) 148 __entry->inflight = inflight; 151 TP_printk("%s: status=%u, step=%d, inflight=%u", __entry->name, 152 __entry->status, __entry->step, __entry->inflight)
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | page_pool.h | 17 s32 inflight, u32 hold, u32 release), 19 TP_ARGS(pool, inflight, hold, release), 23 __field(s32, inflight) 31 __entry->inflight = inflight; 37 TP_printk("page_pool=%p inflight=%d hold=%u release=%u cnt=%llu", 38 __entry->pool, __entry->inflight, __entry->hold,
|
H A D | wbt.h | 127 * @inflight: tracked writes inflight 132 int step, unsigned int inflight), 134 TP_ARGS(bdi, status, step, inflight), 140 __field(unsigned int, inflight) 148 __entry->inflight = inflight; 151 TP_printk("%s: status=%u, step=%d, inflight=%u", __entry->name, 152 __entry->status, __entry->step, __entry->inflight)
|
/kernel/linux/linux-5.10/net/core/ |
H A D | page_pool.c | 271 s32 inflight; in page_pool_inflight() local 273 inflight = _distance(hold_cnt, release_cnt); in page_pool_inflight() 275 trace_page_pool_release(pool, inflight, hold_cnt, release_cnt); in page_pool_inflight() 276 WARN(inflight < 0, "Negative(%d) inflight packet-pages", inflight); in page_pool_inflight() 278 return inflight; in page_pool_inflight() 292 /* Always account for inflight pages, even if we didn't in page_pool_release_page() 471 int inflight; in page_pool_release() local 474 inflight in page_pool_release() 485 int inflight; page_pool_release_retry() local [all...] |
/kernel/linux/linux-6.6/block/ |
H A D | genhd.c | 123 unsigned int inflight = 0; in part_in_flight() local 127 inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) + in part_in_flight() 130 if ((int)inflight < 0) in part_in_flight() 131 inflight = 0; in part_in_flight() 133 return inflight; in part_in_flight() 137 unsigned int inflight[2]) in part_in_flight_rw() 141 inflight[0] = 0; in part_in_flight_rw() 142 inflight[1] = 0; in part_in_flight_rw() 144 inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu); in part_in_flight_rw() 145 inflight[ in part_in_flight_rw() 136 part_in_flight_rw(struct block_device *part, unsigned int inflight[2]) part_in_flight_rw() argument 952 unsigned int inflight; part_stat_show() local 1000 unsigned int inflight[2]; part_inflight_show() local 1231 unsigned int inflight; diskstats_show() local [all...] |
H A D | blk-wbt.c | 199 int inflight, limit; in wbt_rqw_done() local 201 inflight = atomic_dec_return(&rqw->inflight); in wbt_rqw_done() 218 if (inflight && inflight >= limit) in wbt_rqw_done() 222 int diff = limit - inflight; in wbt_rqw_done() 224 if (!inflight || diff >= rwb->wb_background / 2) in wbt_rqw_done() 292 ret += atomic_read(&rwb->rq_wait[i].inflight); in wbt_inflight() 426 unsigned int inflight = wbt_inflight(rwb); in wb_timer_fn() local 434 trace_wbt_timer(rwb->rqos.disk->bdi, status, rqd->scale_step, inflight); in wb_timer_fn() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 269 * Currently we are not tracking the rq->context being inflight in schedule_in() 270 * (ce->inflight = rq->engine). It is only used by the execlists in schedule_in() 272 * required if we generalise the inflight tracking. in schedule_in() 290 struct i915_request **first = execlists->inflight; in __guc_dequeue() 307 * We write directly into the execlists->inflight queue and don't use in __guc_dequeue() 323 port - execlists->inflight); in __guc_dequeue() 340 *port = schedule_in(last, port - execlists->inflight); in __guc_dequeue() 344 execlists->active = execlists->inflight; in __guc_dequeue() 356 for (port = execlists->inflight; (rq = *port); port++) { in guc_submission_tasklet() 362 if (port != execlists->inflight) { in guc_submission_tasklet() [all...] |
/kernel/linux/linux-6.6/drivers/firmware/arm_scmi/ |
H A D | smc.c | 48 * @inflight: Atomic flag to protect access to Tx/Rx shared memory area. 62 atomic_t inflight; member 91 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_init() 96 static bool smc_xfer_inflight(struct scmi_xfer *xfer, atomic_t *inflight) in smc_xfer_inflight() argument 100 ret = atomic_cmpxchg(inflight, INFLIGHT_NONE, xfer->hdr.seq); in smc_xfer_inflight() 110 spin_until_cond(smc_xfer_inflight(xfer, &scmi_info->inflight)); in smc_channel_lock_acquire() 118 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_release()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
H A D | intel_context_types.h | 60 struct intel_engine_cs *inflight; member 61 #define intel_context_inflight(ce) ptr_mask_bits(READ_ONCE((ce)->inflight), 2) 62 #define intel_context_inflight_count(ce) ptr_unmask_bits(READ_ONCE((ce)->inflight), 2)
|
/kernel/linux/linux-5.10/tools/io_uring/ |
H A D | io_uring-bench.c | 80 int inflight; member 272 s->inflight -= reaped; in reap_events() 292 if (!prepped && s->inflight < DEPTH) { in submitter_fn() 293 to_prep = min(DEPTH - s->inflight, BATCH_SUBMIT); in submitter_fn() 296 s->inflight += prepped; in submitter_fn() 300 if (to_submit && (s->inflight + to_submit <= DEPTH)) in submitter_fn() 303 to_wait = min(s->inflight + to_submit, BATCH_COMPLETE); in submitter_fn() 341 if (s->inflight) in submitter_fn() 580 printf("IOPS=%lu, IOS/call=%ld/%ld, inflight=%u (%s)\n", in main() 581 this_done - done, rpc, ipc, s->inflight, in main() [all...] |
/kernel/linux/linux-5.10/net/unix/ |
H A D | garbage.c | 169 atomic_long_dec(&usk->inflight); in dec_inflight() 174 atomic_long_inc(&usk->inflight); in inc_inflight() 179 atomic_long_inc(&u->inflight); in inc_inflight_move_tail() 193 /* If number of inflight sockets is insane, in wait_for_unix_gc() 254 inflight_refs = atomic_long_read(&u->inflight); in unix_gc() 290 if (atomic_long_read(&u->inflight) > 0) { in unix_gc() 299 * inflight counters for these as well, and remove the skbuffs in unix_gc() 307 * cycle. Restore these to the inflight list. in unix_gc() 336 * the inflight list in unix_gc()
|
H A D | scm.c | 57 if (atomic_long_inc_return(&u->inflight) == 1) { in unix_inflight() 79 BUG_ON(!atomic_long_read(&u->inflight)); in unix_notinflight() 82 if (atomic_long_dec_and_test(&u->inflight)) in unix_notinflight()
|
/kernel/linux/linux-6.6/net/unix/ |
H A D | garbage.c | 169 atomic_long_dec(&usk->inflight); in dec_inflight() 174 atomic_long_inc(&usk->inflight); in inc_inflight() 179 atomic_long_inc(&u->inflight); in inc_inflight_move_tail() 193 /* If number of inflight sockets is insane, in wait_for_unix_gc() 243 inflight_refs = atomic_long_read(&u->inflight); in unix_gc() 274 if (atomic_long_read(&u->inflight) > 0) { in unix_gc() 283 * inflight counters for these as well, and remove the skbuffs in unix_gc() 299 * cycle. Restore these to the inflight list. in unix_gc() 328 * the inflight list in unix_gc()
|
H A D | scm.c | 56 if (atomic_long_inc_return(&u->inflight) == 1) { in unix_inflight() 78 BUG_ON(!atomic_long_read(&u->inflight)); in unix_notinflight() 81 if (atomic_long_dec_and_test(&u->inflight)) in unix_notinflight()
|
/kernel/linux/linux-5.10/block/ |
H A D | blk-wbt.c | 131 int inflight, limit; in wbt_rqw_done() local 133 inflight = atomic_dec_return(&rqw->inflight); in wbt_rqw_done() 159 if (inflight && inflight >= limit) in wbt_rqw_done() 163 int diff = limit - inflight; in wbt_rqw_done() 165 if (!inflight || diff >= rwb->wb_background / 2) in wbt_rqw_done() 357 unsigned int inflight = wbt_inflight(rwb); in wb_timer_fn() local 363 inflight); in wb_timer_fn() 405 if (rqd->scale_step || inflight) in wb_timer_fn() [all...] |
H A D | genhd.c | 117 unsigned int inflight = 0; in part_in_flight() local 121 inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) + in part_in_flight() 124 if ((int)inflight < 0) in part_in_flight() 125 inflight = 0; in part_in_flight() 127 return inflight; in part_in_flight() 130 static void part_in_flight_rw(struct hd_struct *part, unsigned int inflight[2]) in part_in_flight_rw() argument 134 inflight[0] = 0; in part_in_flight_rw() 135 inflight[1] = 0; in part_in_flight_rw() 137 inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu); in part_in_flight_rw() 138 inflight[ in part_in_flight_rw() 1302 unsigned int inflight; part_stat_show() local 1345 unsigned int inflight[2]; part_inflight_show() local 1614 unsigned int inflight; diskstats_show() local [all...] |
/kernel/linux/linux-6.6/net/core/ |
H A D | page_pool.c | 499 s32 inflight; in page_pool_inflight() local 501 inflight = _distance(hold_cnt, release_cnt); in page_pool_inflight() 503 trace_page_pool_release(pool, inflight, hold_cnt, release_cnt); in page_pool_inflight() 504 WARN(inflight < 0, "Negative(%d) inflight packet-pages", inflight); in page_pool_inflight() 506 return inflight; in page_pool_inflight() 520 /* Always account for inflight pages, even if we didn't in page_pool_return_page() 851 int inflight; in page_pool_release() local 854 inflight in page_pool_release() 865 int inflight; page_pool_release_retry() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/chelsio/ |
H A D | chcr_core.c | 56 if (atomic_read(&dev->inflight)) { in detach_work_fn() 60 atomic_read(&dev->inflight)); in detach_work_fn() 65 atomic_read(&dev->inflight)); in detach_work_fn() 103 atomic_set(&dev->inflight, 0); in chcr_dev_add() 122 atomic_set(&dev->inflight, 0); in chcr_dev_init() 232 if (atomic_read(&dev->inflight) != 0) { in chcr_detach_device()
|
/kernel/linux/linux-6.6/drivers/crypto/chelsio/ |
H A D | chcr_core.c | 56 if (atomic_read(&dev->inflight)) { in detach_work_fn() 60 atomic_read(&dev->inflight)); in detach_work_fn() 65 atomic_read(&dev->inflight)); in detach_work_fn() 103 atomic_set(&dev->inflight, 0); in chcr_dev_add() 122 atomic_set(&dev->inflight, 0); in chcr_dev_init() 232 if (atomic_read(&dev->inflight) != 0) { in chcr_detach_device()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | i915_scheduler.c | 196 const struct i915_request *inflight; in kick_submission() local 208 inflight = execlists_active(&engine->execlists); in kick_submission() 209 if (!inflight) in kick_submission() 216 if (inflight->context == rq->context) in kick_submission() 220 "bumping queue-priority-hint:%d for rq:%llx:%lld, inflight:%llx:%lld prio %d\n", in kick_submission() 223 inflight->fence.context, inflight->fence.seqno, in kick_submission() 224 inflight->sched.attr.priority); in kick_submission() 227 if (need_preempt(prio, rq_prio(inflight))) in kick_submission()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/ |
H A D | intel_context_types.h | 90 struct intel_engine_cs *inflight; member 94 __intel_context_inflight(READ_ONCE((ce)->inflight)) 96 __intel_context_inflight_count(READ_ONCE((ce)->inflight))
|
/kernel/linux/linux-5.10/drivers/vhost/ |
H A D | scsi.c | 68 /* Refcount for the inflight reqs */ 112 /* Used to track inflight cmd */ 113 struct vhost_scsi_inflight *inflight; member 182 * Reference counting for inflight reqs, used for flush operation. At 188 * Indicate current inflight in use, protected by vq->mutex. 224 struct vhost_scsi_inflight *inflight; member 251 struct vhost_scsi_inflight *inflight; in vhost_scsi_done_inflight() local 253 inflight = container_of(kref, struct vhost_scsi_inflight, kref); in vhost_scsi_done_inflight() 254 complete(&inflight->comp); in vhost_scsi_done_inflight() 287 struct vhost_scsi_inflight *inflight; in vhost_scsi_get_inflight() local 297 vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight) vhost_scsi_put_inflight() argument 347 struct vhost_scsi_inflight *inflight = tv_cmd->inflight; vhost_scsi_release_cmd_res() local 366 struct vhost_scsi_inflight *inflight = tmf->inflight; vhost_scsi_release_tmf_res() local [all...] |
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_bbr.c | 42 * it briefly enters PROBE_RTT to cut inflight to a minimum value to re-probe 83 BBR_PROBE_RTT, /* cut inflight to min to probe min_rtt */ 402 /* Ensure gain cycling gets inflight above BDP even for small BDPs. */ in bbr_quantization_budget() 409 /* Find inflight based on min RTT and the estimated bottleneck bandwidth. */ 412 u32 inflight; in bbr_inflight() local 414 inflight = bbr_bdp(sk, bw, gain); in bbr_inflight() 415 inflight = bbr_quantization_budget(sk, inflight); in bbr_inflight() 417 return inflight; in bbr_inflight() 424 * inflight leve 559 u32 inflight, bw; bbr_is_next_cycle_phase() local [all...] |
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_bbr.c | 42 * it briefly enters PROBE_RTT to cut inflight to a minimum value to re-probe 85 BBR_PROBE_RTT, /* cut inflight to min to probe min_rtt */ 404 /* Ensure gain cycling gets inflight above BDP even for small BDPs. */ in bbr_quantization_budget() 411 /* Find inflight based on min RTT and the estimated bottleneck bandwidth. */ 414 u32 inflight; in bbr_inflight() local 416 inflight = bbr_bdp(sk, bw, gain); in bbr_inflight() 417 inflight = bbr_quantization_budget(sk, inflight); in bbr_inflight() 419 return inflight; in bbr_inflight() 426 * inflight leve 561 u32 inflight, bw; bbr_is_next_cycle_phase() local [all...] |