Home
last modified time | relevance | path

Searched refs:queue_depth (Results 1 - 25 of 207) sorted by relevance

123456789

/kernel/linux/linux-5.10/drivers/target/
H A Dtarget_core_tpg.c157 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth()
159 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth()
161 if (!acl->queue_depth) { in target_set_nacl_queue_depth()
165 acl->queue_depth = 1; in target_set_nacl_queue_depth()
173 u32 queue_depth; in target_alloc_node_acl() local
190 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl()
192 queue_depth = 1; in target_alloc_node_acl()
193 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl()
217 acl->queue_depth, in target_add_node_acl()
156 target_set_nacl_queue_depth(struct se_portal_group *tpg, struct se_node_acl *acl, u32 queue_depth) target_set_nacl_queue_depth() argument
378 core_tpg_set_initiator_node_queue_depth( struct se_node_acl *acl, u32 queue_depth) core_tpg_set_initiator_node_queue_depth() argument
[all...]
/kernel/linux/linux-6.6/drivers/target/
H A Dtarget_core_tpg.c158 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth()
160 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth()
162 if (!acl->queue_depth) { in target_set_nacl_queue_depth()
166 acl->queue_depth = 1; in target_set_nacl_queue_depth()
174 u32 queue_depth; in target_alloc_node_acl() local
191 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl()
193 queue_depth = 1; in target_alloc_node_acl()
194 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl()
218 acl->queue_depth, in target_add_node_acl()
157 target_set_nacl_queue_depth(struct se_portal_group *tpg, struct se_node_acl *acl, u32 queue_depth) target_set_nacl_queue_depth() argument
379 core_tpg_set_initiator_node_queue_depth( struct se_node_acl *acl, u32 queue_depth) core_tpg_set_initiator_node_queue_depth() argument
[all...]
/kernel/linux/linux-5.10/drivers/ata/
H A Dlibata-sata.c1019 * @queue_depth: new queue depth
1026 int queue_depth) in __ata_change_queue_depth()
1031 if (queue_depth < 1 || queue_depth == sdev->queue_depth) in __ata_change_queue_depth()
1032 return sdev->queue_depth; in __ata_change_queue_depth()
1036 return sdev->queue_depth; in __ata_change_queue_depth()
1041 if (queue_depth == 1 || !ata_ncq_enabled(dev)) { in __ata_change_queue_depth()
1043 queue_depth = 1; in __ata_change_queue_depth()
1048 queue_depth in __ata_change_queue_depth()
1025 __ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, int queue_depth) __ata_change_queue_depth() argument
1074 ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) ata_scsi_change_queue_depth() argument
[all...]
/kernel/linux/linux-6.6/drivers/s390/block/
H A Ddasd_genhd.c28 static unsigned int queue_depth = 32; variable
31 module_param(queue_depth, uint, 0444);
32 MODULE_PARM_DESC(queue_depth, "Default queue depth for new DASD devices");
54 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc()
/kernel/linux/linux-6.6/arch/um/drivers/
H A Dvector_kern.c250 int queue_depth; in vector_advancehead() local
258 qi->queue_depth -= advance; in vector_advancehead()
264 if (qi->queue_depth == 0) { in vector_advancehead()
268 queue_depth = qi->queue_depth; in vector_advancehead()
270 return queue_depth; in vector_advancehead()
280 int queue_depth; in vector_advancetail() local
286 qi->queue_depth += advance; in vector_advancetail()
287 queue_depth = qi->queue_depth; in vector_advancetail()
336 int queue_depth; vector_enqueue() local
411 int result = 0, send_len, queue_depth = qi->max_depth; vector_send() local
1051 int queue_depth = 0; vector_net_start_xmit() local
[all...]
/kernel/linux/linux-5.10/arch/um/drivers/
H A Dvector_kern.c251 int queue_depth; in vector_advancehead() local
259 qi->queue_depth -= advance; in vector_advancehead()
265 if (qi->queue_depth == 0) { in vector_advancehead()
269 queue_depth = qi->queue_depth; in vector_advancehead()
271 return queue_depth; in vector_advancehead()
281 int queue_depth; in vector_advancetail() local
287 qi->queue_depth += advance; in vector_advancetail()
288 queue_depth = qi->queue_depth; in vector_advancetail()
337 int queue_depth; vector_enqueue() local
412 int result = 0, send_len, queue_depth = qi->max_depth; vector_send() local
1069 int queue_depth = 0; vector_net_start_xmit() local
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt.c63 size_t max_depth = clt->queue_depth; in __rtrs_get_permit()
474 flags = atomic_inc_return(&con->io_cnt) % sess->queue_depth ? in rtrs_post_send_rdma()
490 if (WARN_ON(msg_id >= sess->queue_depth)) in process_io_rsp()
543 if (WARN_ON(buf_id >= sess->queue_depth)) in rtrs_clt_rkey_rsp_done()
713 q_size = sess->queue_depth; in post_recv_sess()
973 flags = atomic_inc_return(&con->io_cnt) % sess->queue_depth ? in rtrs_post_rdma_write_sg()
1212 for (i = 0; i < sess->queue_depth; ++i) { in fail_all_outstanding_reqs()
1238 for (i = 0; i < sess->queue_depth; ++i) { in free_sess_reqs()
1255 sess->reqs = kcalloc(sess->queue_depth, sizeof(*sess->reqs), in alloc_sess_reqs()
1260 for (i = 0; i < sess->queue_depth; in alloc_sess_reqs()
1702 u16 version, queue_depth; rtrs_rdma_conn_established() local
[all...]
H A Drtrs-srv.c152 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids()
171 sess->ops_ids = kcalloc(srv->queue_depth, sizeof(*sess->ops_ids), in rtrs_srv_alloc_ops_ids()
176 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids()
302 flags = (atomic_inc_return(&id->con->wr_cnt) % srv->queue_depth) ? in rdma_write_sg()
429 flags = (atomic_inc_return(&con->wr_cnt) % srv->queue_depth) ? in send_io_resp_imm()
605 * Here we map queue_depth chunks to MR. Firstly we have to in map_cont_bufs()
613 mrs_num = srv->queue_depth; in map_cont_bufs()
617 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs()
618 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs()
637 srv->queue_depth in map_cont_bufs()
[all...]
H A Drtrs-clt.h140 size_t queue_depth; member
171 size_t queue_depth; member
/kernel/linux/linux-6.6/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt.c71 size_t max_depth = clt->queue_depth; in __rtrs_get_permit()
499 if (WARN_ON(msg_id >= clt_path->queue_depth)) in process_io_rsp()
553 if (WARN_ON(buf_id >= clt_path->queue_depth)) in rtrs_clt_rkey_rsp_done()
725 q_size = clt_path->queue_depth; in post_recv_path()
1323 for (i = 0; i < clt_path->queue_depth; ++i) { in fail_all_outstanding_reqs()
1349 for (i = 0; i < clt_path->queue_depth; ++i) { in free_path_reqs()
1365 clt_path->reqs = kcalloc(clt_path->queue_depth, in alloc_path_reqs()
1371 for (i = 0; i < clt_path->queue_depth; ++i) { in alloc_path_reqs()
1411 clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL); in alloc_permits()
1416 clt->permits = kcalloc(clt->queue_depth, permit_siz in alloc_permits()
1825 u16 version, queue_depth; rtrs_rdma_conn_established() local
[all...]
H A Drtrs-srv.c109 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids()
138 srv_path->ops_ids = kcalloc(srv->queue_depth, in rtrs_srv_alloc_ops_ids()
144 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids()
577 * Here we map queue_depth chunks to MR. Firstly we have to in map_cont_bufs()
585 mrs_num = srv->queue_depth; in map_cont_bufs()
589 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs()
590 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs()
607 srv->queue_depth - chunks); in map_cont_bufs()
655 chunk_bits = ilog2(srv->queue_depth - 1) + 1; in map_cont_bufs()
989 q_size = srv->queue_depth; in post_recv_path()
[all...]
/kernel/linux/linux-5.10/drivers/target/iscsi/
H A Discsi_target_device.c37 * struct se_node_acl->queue_depth. The value in struct se_node_acl->queue_depth in iscsit_determine_maxcmdsn()
41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn()
42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
/kernel/linux/linux-6.6/drivers/target/iscsi/
H A Discsi_target_device.c37 * struct se_node_acl->queue_depth. The value in struct se_node_acl->queue_depth in iscsit_determine_maxcmdsn()
41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn()
42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
/kernel/linux/linux-5.10/drivers/s390/crypto/
H A Dap_card.c47 return scnprintf(buf, PAGE_SIZE, "%d\n", ac->queue_depth); in depth_show()
209 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument
221 ac->queue_depth = queue_depth; in ap_card_create()
H A Dap_bus.h162 int queue_depth; /* AP queue depth.*/ member
287 struct ap_card *ap_card_create(int id, int queue_depth, int raw_device_type,
/kernel/linux/linux-6.6/drivers/s390/crypto/
H A Dap_card.c47 return sysfs_emit(buf, "%d\n", ac->queue_depth); in depth_show()
232 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument
244 ac->queue_depth = queue_depth; in ap_card_create()
/kernel/linux/linux-6.6/drivers/ata/
H A Dlibata-sata.c1049 * @queue_depth: new queue depth
1056 int queue_depth) in ata_change_queue_depth()
1065 if (!dev || queue_depth < 1 || queue_depth == sdev->queue_depth) { in ata_change_queue_depth()
1067 return sdev->queue_depth; in ata_change_queue_depth()
1076 if (queue_depth > max_queue_depth) { in ata_change_queue_depth()
1085 if (queue_depth == 1 || !ata_ncq_supported(dev)) { in ata_change_queue_depth()
1087 queue_depth = 1; in ata_change_queue_depth()
1094 if (queue_depth in ata_change_queue_depth()
1055 ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, int queue_depth) ata_change_queue_depth() argument
1116 ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) ata_scsi_change_queue_depth() argument
[all...]
/kernel/linux/linux-5.10/block/
H A Dblk-rq-qos.c128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
144 rqd->queue_depth); in rq_depth_calc_max_depth()
148 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
H A Dblk-mq.c2328 for (i = 0; i < set->queue_depth; i++) { in blk_mq_clear_rq_mapping()
2651 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping()
2662 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
2686 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
2863 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request()
2868 set->queue_depth); in __blk_mq_alloc_map_and_request()
3152 unsigned int queue_depth, in blk_mq_init_sq_queue()
3162 set->queue_depth = queue_depth; in blk_mq_init_sq_queue()
3331 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
2650 blk_mq_clear_flush_rq_mapping(struct blk_mq_tags *tags, unsigned int queue_depth, struct request *flush_rq) blk_mq_clear_flush_rq_mapping() argument
3150 blk_mq_init_sq_queue(struct blk_mq_tag_set *set, const struct blk_mq_ops *ops, unsigned int queue_depth, unsigned int set_flags) blk_mq_init_sq_queue() argument
[all...]
/kernel/linux/linux-5.10/drivers/block/null_blk/
H A Dnull_blk.h29 unsigned int queue_depth; member
86 unsigned int queue_depth; member
/kernel/linux/linux-6.6/include/net/mana/
H A Dhw_channel.h139 u16 queue_depth; member
153 u16 queue_depth; member
/kernel/linux/linux-6.6/drivers/block/null_blk/
H A Dnull_blk.h33 unsigned int queue_depth; member
133 unsigned int queue_depth; member
H A Dmain.c768 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag()
769 if (tag >= nq->queue_depth) in get_tag()
1780 nq->queue_depth = nullb->queue_depth; in null_init_queue()
1882 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands()
1886 nq->tag_map = bitmap_zalloc(nq->queue_depth, GFP_KERNEL); in setup_commands()
1892 for (i = 0; i < nq->queue_depth; i++) { in setup_commands()
1912 nullb->queue_depth = nullb->dev->hw_queue_depth; in setup_queues()
1965 unsigned int queue_depth; in null_init_tag_set() local
1971 queue_depth in null_init_tag_set()
[all...]
/kernel/linux/linux-6.6/block/
H A Dblk-rq-qos.c122 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
138 rqd->queue_depth); in rq_depth_calc_max_depth()
142 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
H A Dblk-mq.c3621 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping()
3632 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
3657 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
3848 set->queue_depth); in __blk_mq_alloc_map_and_rqs()
4310 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
4343 set->queue_depth); in __blk_mq_alloc_rq_maps()
4370 * may reduce the depth asked for, if memory is tight. set->queue_depth
4378 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs()
4384 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs()
4385 if (set->queue_depth < se in blk_mq_alloc_set_map_and_rqs()
3620 blk_mq_clear_flush_rq_mapping(struct blk_mq_tags *tags, unsigned int queue_depth, struct request *flush_rq) blk_mq_clear_flush_rq_mapping() argument
4581 blk_mq_alloc_sq_tag_set(struct blk_mq_tag_set *set, const struct blk_mq_ops *ops, unsigned int queue_depth, unsigned int set_flags) blk_mq_alloc_sq_tag_set() argument
[all...]

Completed in 34 milliseconds

123456789