/kernel/linux/linux-6.6/block/ |
H A D | blk-wbt.c | 94 struct rq_depth rq_depth; member 307 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() 360 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() 370 } else if (rwb->rq_depth.max_depth <= 2) { in calc_wb_limits() 371 rwb->wb_normal = rwb->rq_depth.max_depth; in calc_wb_limits() 374 rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2; in calc_wb_limits() 375 rwb->wb_background = (rwb->rq_depth in calc_wb_limits() [all...] |
H A D | blk-rq-qos.h | 51 struct rq_depth { struct 99 bool rq_depth_scale_up(struct rq_depth *rqd); 100 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 101 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
H A D | blk-rq-qos.c | 110 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() 158 bool rq_depth_scale_up(struct rq_depth *rqd) in rq_depth_scale_up() 177 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle) in rq_depth_scale_down()
|
/kernel/linux/linux-5.10/block/ |
H A D | blk-wbt.c | 238 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() 291 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() 301 } else if (rwb->rq_depth.max_depth <= 2) { in calc_wb_limits() 302 rwb->wb_normal = rwb->rq_depth.max_depth; in calc_wb_limits() 305 rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2; in calc_wb_limits() 306 rwb->wb_background = (rwb->rq_depth.max_depth + 3) / 4; in calc_wb_limits() 312 if (!rq_depth_scale_up(&rwb->rq_depth)) in scale_up() 322 if (!rq_depth_scale_down(&rwb->rq_depth, hard_throttl in scale_down() [all...] |
H A D | blk-iolatency.c | 143 struct rq_depth rq_depth; member 282 return rq_wait_inc_below(rqw, iolat->rq_depth.max_depth); in iolat_acquire_inflight() 374 unsigned long old = iolat->rq_depth.max_depth; in scale_change() 386 iolat->rq_depth.max_depth = old; in scale_change() 391 iolat->rq_depth.max_depth = max(old, 1UL); in scale_change() 449 if (iolat->rq_depth.max_depth == 1 && direction < 0) { in check_scale_change() 457 iolat->rq_depth.max_depth = UINT_MAX; in check_scale_change() 512 if (unlikely(issue_as_root && iolat->rq_depth.max_depth != UINT_MAX)) { in iolatency_record_time() 920 if (iolat->rq_depth in iolatency_ssd_stat() [all...] |
H A D | blk-rq-qos.h | 51 struct rq_depth { struct 154 bool rq_depth_scale_up(struct rq_depth *rqd); 155 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 156 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
H A D | blk-wbt.h | 70 struct rq_depth rq_depth; member
|
H A D | blk-rq-qos.c | 116 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() 164 bool rq_depth_scale_up(struct rq_depth *rqd) in rq_depth_scale_up() 183 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle) in rq_depth_scale_down()
|
/kernel/linux/linux-5.10/net/9p/ |
H A D | trans_rdma.c | 61 * @rq_depth: The depth of the Receive Queue. 88 int rq_depth; member 120 * @rq_depth: The depth of the RQ. Should be greater than or equal to SQ depth 127 int rq_depth; member 159 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 160 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 184 opts->rq_depth = P9_RDMA_RQ_DEPTH; in parse_opts() 221 opts->rq_depth = option; in parse_opts() 234 opts->rq_depth = max(opts->rq_depth, opt in parse_opts() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_queue.c | 175 for (i = 0; i < funq->rq_depth; i++) { in fun_clean_rq() 192 for (i = 0; i < funq->rq_depth; i++) { in fun_fill_rq() 209 funq->rq_tail = funq->rq_depth - 1; in fun_fill_rq() 222 if (++funq->rq_buf_idx == funq->rq_depth) in fun_rq_update_pos() 344 funq->rq_depth; in fun_process_cq() 375 funq->rqes = fun_alloc_ring_mem(funq->fdev->dev, funq->rq_depth, in fun_alloc_rqes() 396 fun_free_ring_mem(dev, funq->rq_depth, sizeof(*funq->rqes), in fun_free_queue() 419 if (req->rq_depth) { in fun_alloc_queue() 458 if (req->rq_depth) { in fun_alloc_queue() 460 funq->rq_depth in fun_alloc_queue() [all...] |
H A D | fun_queue.h | 47 u32 rq_depth; member 124 u32 rq_depth; member
|
H A D | fun_dev.h | 94 u16 rq_depth; member
|
H A D | fun_dev.c | 232 .rq_depth = areq->rq_depth, in fun_enable_admin_queue() 280 if (areq->rq_depth) { in fun_enable_admin_queue() 582 if (cq_count < 2 || sq_count < 2 + !!fdev->admin_q->rq_depth) in fun_get_dev_limits()
|
/kernel/linux/linux-6.6/net/9p/ |
H A D | trans_rdma.c | 58 * @rq_depth: The depth of the Receive Queue. 85 int rq_depth; member 119 * @rq_depth: The depth of the RQ. Should be greater than or equal to SQ depth 126 int rq_depth; member 158 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 159 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 183 opts->rq_depth = P9_RDMA_RQ_DEPTH; in parse_opts() 220 opts->rq_depth = option; in parse_opts() 233 opts->rq_depth = max(opts->rq_depth, opt in parse_opts() [all...] |
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 373 unsigned int ctxts, rq_depth; in svc_rdma_accept() local 407 rq_depth = newxprt->sc_max_requests + newxprt->sc_max_bc_requests; in svc_rdma_accept() 408 if (rq_depth > dev->attrs.max_qp_wr) { in svc_rdma_accept() 411 rq_depth = dev->attrs.max_qp_wr; in svc_rdma_accept() 412 newxprt->sc_max_requests = rq_depth - 2; in svc_rdma_accept() 418 newxprt->sc_sq_depth = rq_depth + ctxts; in svc_rdma_accept() 436 ib_alloc_cq_any(dev, newxprt, rq_depth, IB_POLL_WORKQUEUE); in svc_rdma_accept() 446 qp_attr.cap.max_recv_wr = rq_depth; in svc_rdma_accept()
|
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 371 unsigned int ctxts, rq_depth; in svc_rdma_accept() local 406 rq_depth = newxprt->sc_max_requests + newxprt->sc_max_bc_requests + in svc_rdma_accept() 408 if (rq_depth > dev->attrs.max_qp_wr) { in svc_rdma_accept() 411 rq_depth = dev->attrs.max_qp_wr; in svc_rdma_accept() 413 newxprt->sc_max_requests = rq_depth - 2; in svc_rdma_accept() 419 newxprt->sc_sq_depth = rq_depth + ctxts; in svc_rdma_accept() 437 ib_alloc_cq_any(dev, newxprt, rq_depth, IB_POLL_WORKQUEUE); in svc_rdma_accept() 447 qp_attr.cap.max_recv_wr = rq_depth; in svc_rdma_accept()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth.h | 76 unsigned int rq_depth; member 116 unsigned int rq_depth; member
|
H A D | funeth_ethtool.c | 566 ring->rx_pending = fp->rq_depth; in fun_get_ringparam() 594 fp->rq_depth == ring->rx_pending) in fun_set_ringparam() 600 .rq_depth = ring->rx_pending, in fun_set_ringparam() 610 fp->rq_depth = ring->rx_pending; in fun_set_ringparam() 611 fp->cq_depth = 2 * fp->rq_depth; in fun_set_ringparam()
|
H A D | funeth_main.c | 509 qset->rq_depth, qset->rxq_start, qset->state); in fun_alloc_rings() 842 .rq_depth = fp->rq_depth, in funeth_open() 1643 .rq_depth = fp->rq_depth, in fun_change_num_queues() 1783 fp->rq_depth = min_t(unsigned int, RQ_DEPTH, fdev->q_depth); in fun_create_netdev() 2003 .rq_depth = ADMIN_RQ_DEPTH, in funeth_probe()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_dev.c | 265 * @rq_depth: rq depth 271 unsigned int rq_depth) in set_hw_ioctxt() 285 hw_ioctxt.rq_depth = ilog2(rq_depth); in set_hw_ioctxt() 434 * @rq_depth: the receive queue depth 438 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth) in hinic_hwdev_ifup() argument 462 func_to_io->rq_depth = rq_depth; in hinic_hwdev_ifup() 489 err = set_hw_ioctxt(hwdev, sq_depth, rq_depth); in hinic_hwdev_ifup() 270 set_hw_ioctxt(struct hinic_hwdev *hwdev, unsigned int sq_depth, unsigned int rq_depth) set_hw_ioctxt() argument
|
H A D | hinic_dev.h | 100 u16 rq_depth; member
|
H A D | hinic_hw_io.h | 74 u16 rq_depth; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_dev.c | 271 * @rq_depth: rq depth 277 unsigned int rq_depth) in set_hw_ioctxt() 291 hw_ioctxt.rq_depth = ilog2(rq_depth); in set_hw_ioctxt() 440 * @rq_depth: the receive queue depth 444 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth) in hinic_hwdev_ifup() argument 468 func_to_io->rq_depth = rq_depth; in hinic_hwdev_ifup() 495 err = set_hw_ioctxt(hwdev, sq_depth, rq_depth); in hinic_hwdev_ifup() 276 set_hw_ioctxt(struct hinic_hwdev *hwdev, unsigned int sq_depth, unsigned int rq_depth) set_hw_ioctxt() argument
|
H A D | hinic_dev.h | 96 u16 rq_depth; member
|
H A D | hinic_hw_io.h | 74 u16 rq_depth; member
|