/kernel/linux/linux-5.10/drivers/target/ |
H A D | target_core_tpg.c | 157 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() 159 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth() 161 if (!acl->queue_depth) { in target_set_nacl_queue_depth() 165 acl->queue_depth = 1; in target_set_nacl_queue_depth() 173 u32 queue_depth; in target_alloc_node_acl() local 190 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl() 192 queue_depth = 1; in target_alloc_node_acl() 193 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl() 217 acl->queue_depth, in target_add_node_acl() 156 target_set_nacl_queue_depth(struct se_portal_group *tpg, struct se_node_acl *acl, u32 queue_depth) target_set_nacl_queue_depth() argument 378 core_tpg_set_initiator_node_queue_depth( struct se_node_acl *acl, u32 queue_depth) core_tpg_set_initiator_node_queue_depth() argument [all...] |
/kernel/linux/linux-6.6/drivers/target/ |
H A D | target_core_tpg.c | 158 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() 160 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth() 162 if (!acl->queue_depth) { in target_set_nacl_queue_depth() 166 acl->queue_depth = 1; in target_set_nacl_queue_depth() 174 u32 queue_depth; in target_alloc_node_acl() local 191 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl() 193 queue_depth = 1; in target_alloc_node_acl() 194 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl() 218 acl->queue_depth, in target_add_node_acl() 157 target_set_nacl_queue_depth(struct se_portal_group *tpg, struct se_node_acl *acl, u32 queue_depth) target_set_nacl_queue_depth() argument 379 core_tpg_set_initiator_node_queue_depth( struct se_node_acl *acl, u32 queue_depth) core_tpg_set_initiator_node_queue_depth() argument [all...] |
/kernel/linux/linux-5.10/drivers/ata/ |
H A D | libata-sata.c | 1019 * @queue_depth: new queue depth 1026 int queue_depth) in __ata_change_queue_depth() 1031 if (queue_depth < 1 || queue_depth == sdev->queue_depth) in __ata_change_queue_depth() 1032 return sdev->queue_depth; in __ata_change_queue_depth() 1036 return sdev->queue_depth; in __ata_change_queue_depth() 1041 if (queue_depth == 1 || !ata_ncq_enabled(dev)) { in __ata_change_queue_depth() 1043 queue_depth = 1; in __ata_change_queue_depth() 1048 queue_depth in __ata_change_queue_depth() 1025 __ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, int queue_depth) __ata_change_queue_depth() argument 1074 ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) ata_scsi_change_queue_depth() argument [all...] |
/kernel/linux/linux-6.6/drivers/s390/block/ |
H A D | dasd_genhd.c | 28 static unsigned int queue_depth = 32; variable 31 module_param(queue_depth, uint, 0444); 32 MODULE_PARM_DESC(queue_depth, "Default queue depth for new DASD devices"); 54 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc()
|
/kernel/linux/linux-6.6/arch/um/drivers/ |
H A D | vector_kern.c | 250 int queue_depth; in vector_advancehead() local 258 qi->queue_depth -= advance; in vector_advancehead() 264 if (qi->queue_depth == 0) { in vector_advancehead() 268 queue_depth = qi->queue_depth; in vector_advancehead() 270 return queue_depth; in vector_advancehead() 280 int queue_depth; in vector_advancetail() local 286 qi->queue_depth += advance; in vector_advancetail() 287 queue_depth = qi->queue_depth; in vector_advancetail() 336 int queue_depth; vector_enqueue() local 411 int result = 0, send_len, queue_depth = qi->max_depth; vector_send() local 1051 int queue_depth = 0; vector_net_start_xmit() local [all...] |
/kernel/linux/linux-5.10/arch/um/drivers/ |
H A D | vector_kern.c | 251 int queue_depth; in vector_advancehead() local 259 qi->queue_depth -= advance; in vector_advancehead() 265 if (qi->queue_depth == 0) { in vector_advancehead() 269 queue_depth = qi->queue_depth; in vector_advancehead() 271 return queue_depth; in vector_advancehead() 281 int queue_depth; in vector_advancetail() local 287 qi->queue_depth += advance; in vector_advancetail() 288 queue_depth = qi->queue_depth; in vector_advancetail() 337 int queue_depth; vector_enqueue() local 412 int result = 0, send_len, queue_depth = qi->max_depth; vector_send() local 1069 int queue_depth = 0; vector_net_start_xmit() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-clt.c | 63 size_t max_depth = clt->queue_depth; in __rtrs_get_permit() 474 flags = atomic_inc_return(&con->io_cnt) % sess->queue_depth ? in rtrs_post_send_rdma() 490 if (WARN_ON(msg_id >= sess->queue_depth)) in process_io_rsp() 543 if (WARN_ON(buf_id >= sess->queue_depth)) in rtrs_clt_rkey_rsp_done() 713 q_size = sess->queue_depth; in post_recv_sess() 973 flags = atomic_inc_return(&con->io_cnt) % sess->queue_depth ? in rtrs_post_rdma_write_sg() 1212 for (i = 0; i < sess->queue_depth; ++i) { in fail_all_outstanding_reqs() 1238 for (i = 0; i < sess->queue_depth; ++i) { in free_sess_reqs() 1255 sess->reqs = kcalloc(sess->queue_depth, sizeof(*sess->reqs), in alloc_sess_reqs() 1260 for (i = 0; i < sess->queue_depth; in alloc_sess_reqs() 1702 u16 version, queue_depth; rtrs_rdma_conn_established() local [all...] |
H A D | rtrs-srv.c | 152 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids() 171 sess->ops_ids = kcalloc(srv->queue_depth, sizeof(*sess->ops_ids), in rtrs_srv_alloc_ops_ids() 176 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids() 302 flags = (atomic_inc_return(&id->con->wr_cnt) % srv->queue_depth) ? in rdma_write_sg() 429 flags = (atomic_inc_return(&con->wr_cnt) % srv->queue_depth) ? in send_io_resp_imm() 605 * Here we map queue_depth chunks to MR. Firstly we have to in map_cont_bufs() 613 mrs_num = srv->queue_depth; in map_cont_bufs() 617 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs() 618 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs() 637 srv->queue_depth in map_cont_bufs() [all...] |
H A D | rtrs-clt.h | 140 size_t queue_depth; member 171 size_t queue_depth; member
|
/kernel/linux/linux-6.6/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-clt.c | 71 size_t max_depth = clt->queue_depth; in __rtrs_get_permit() 499 if (WARN_ON(msg_id >= clt_path->queue_depth)) in process_io_rsp() 553 if (WARN_ON(buf_id >= clt_path->queue_depth)) in rtrs_clt_rkey_rsp_done() 725 q_size = clt_path->queue_depth; in post_recv_path() 1323 for (i = 0; i < clt_path->queue_depth; ++i) { in fail_all_outstanding_reqs() 1349 for (i = 0; i < clt_path->queue_depth; ++i) { in free_path_reqs() 1365 clt_path->reqs = kcalloc(clt_path->queue_depth, in alloc_path_reqs() 1371 for (i = 0; i < clt_path->queue_depth; ++i) { in alloc_path_reqs() 1411 clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL); in alloc_permits() 1416 clt->permits = kcalloc(clt->queue_depth, permit_siz in alloc_permits() 1825 u16 version, queue_depth; rtrs_rdma_conn_established() local [all...] |
H A D | rtrs-srv.c | 109 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids() 138 srv_path->ops_ids = kcalloc(srv->queue_depth, in rtrs_srv_alloc_ops_ids() 144 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids() 577 * Here we map queue_depth chunks to MR. Firstly we have to in map_cont_bufs() 585 mrs_num = srv->queue_depth; in map_cont_bufs() 589 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs() 590 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs() 607 srv->queue_depth - chunks); in map_cont_bufs() 655 chunk_bits = ilog2(srv->queue_depth - 1) + 1; in map_cont_bufs() 989 q_size = srv->queue_depth; in post_recv_path() [all...] |
/kernel/linux/linux-5.10/drivers/target/iscsi/ |
H A D | iscsi_target_device.c | 37 * struct se_node_acl->queue_depth. The value in struct se_node_acl->queue_depth in iscsit_determine_maxcmdsn() 41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn() 42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
|
/kernel/linux/linux-6.6/drivers/target/iscsi/ |
H A D | iscsi_target_device.c | 37 * struct se_node_acl->queue_depth. The value in struct se_node_acl->queue_depth in iscsit_determine_maxcmdsn() 41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn() 42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
|
/kernel/linux/linux-5.10/drivers/s390/crypto/ |
H A D | ap_card.c | 47 return scnprintf(buf, PAGE_SIZE, "%d\n", ac->queue_depth); in depth_show() 209 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument 221 ac->queue_depth = queue_depth; in ap_card_create()
|
H A D | ap_bus.h | 162 int queue_depth; /* AP queue depth.*/ member 287 struct ap_card *ap_card_create(int id, int queue_depth, int raw_device_type,
|
/kernel/linux/linux-6.6/drivers/s390/crypto/ |
H A D | ap_card.c | 47 return sysfs_emit(buf, "%d\n", ac->queue_depth); in depth_show() 232 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument 244 ac->queue_depth = queue_depth; in ap_card_create()
|
/kernel/linux/linux-6.6/drivers/ata/ |
H A D | libata-sata.c | 1049 * @queue_depth: new queue depth 1056 int queue_depth) in ata_change_queue_depth() 1065 if (!dev || queue_depth < 1 || queue_depth == sdev->queue_depth) { in ata_change_queue_depth() 1067 return sdev->queue_depth; in ata_change_queue_depth() 1076 if (queue_depth > max_queue_depth) { in ata_change_queue_depth() 1085 if (queue_depth == 1 || !ata_ncq_supported(dev)) { in ata_change_queue_depth() 1087 queue_depth = 1; in ata_change_queue_depth() 1094 if (queue_depth in ata_change_queue_depth() 1055 ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, int queue_depth) ata_change_queue_depth() argument 1116 ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) ata_scsi_change_queue_depth() argument [all...] |
/kernel/linux/linux-5.10/block/ |
H A D | blk-rq-qos.c | 128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 144 rqd->queue_depth); in rq_depth_calc_max_depth() 148 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
|
H A D | blk-mq.c | 2328 for (i = 0; i < set->queue_depth; i++) { in blk_mq_clear_rq_mapping() 2651 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() 2662 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping() 2686 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 2863 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request() 2868 set->queue_depth); in __blk_mq_alloc_map_and_request() 3152 unsigned int queue_depth, in blk_mq_init_sq_queue() 3162 set->queue_depth = queue_depth; in blk_mq_init_sq_queue() 3331 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 2650 blk_mq_clear_flush_rq_mapping(struct blk_mq_tags *tags, unsigned int queue_depth, struct request *flush_rq) blk_mq_clear_flush_rq_mapping() argument 3150 blk_mq_init_sq_queue(struct blk_mq_tag_set *set, const struct blk_mq_ops *ops, unsigned int queue_depth, unsigned int set_flags) blk_mq_init_sq_queue() argument [all...] |
/kernel/linux/linux-5.10/drivers/block/null_blk/ |
H A D | null_blk.h | 29 unsigned int queue_depth; member 86 unsigned int queue_depth; member
|
/kernel/linux/linux-6.6/include/net/mana/ |
H A D | hw_channel.h | 139 u16 queue_depth; member 153 u16 queue_depth; member
|
/kernel/linux/linux-6.6/drivers/block/null_blk/ |
H A D | null_blk.h | 33 unsigned int queue_depth; member 133 unsigned int queue_depth; member
|
H A D | main.c | 768 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag() 769 if (tag >= nq->queue_depth) in get_tag() 1780 nq->queue_depth = nullb->queue_depth; in null_init_queue() 1882 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands() 1886 nq->tag_map = bitmap_zalloc(nq->queue_depth, GFP_KERNEL); in setup_commands() 1892 for (i = 0; i < nq->queue_depth; i++) { in setup_commands() 1912 nullb->queue_depth = nullb->dev->hw_queue_depth; in setup_queues() 1965 unsigned int queue_depth; in null_init_tag_set() local 1971 queue_depth in null_init_tag_set() [all...] |
/kernel/linux/linux-6.6/block/ |
H A D | blk-rq-qos.c | 122 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 138 rqd->queue_depth); in rq_depth_calc_max_depth() 142 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
|
H A D | blk-mq.c | 3621 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() 3632 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping() 3657 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 3848 set->queue_depth); in __blk_mq_alloc_map_and_rqs() 4310 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 4343 set->queue_depth); in __blk_mq_alloc_rq_maps() 4370 * may reduce the depth asked for, if memory is tight. set->queue_depth 4378 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs() 4384 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs() 4385 if (set->queue_depth < se in blk_mq_alloc_set_map_and_rqs() 3620 blk_mq_clear_flush_rq_mapping(struct blk_mq_tags *tags, unsigned int queue_depth, struct request *flush_rq) blk_mq_clear_flush_rq_mapping() argument 4581 blk_mq_alloc_sq_tag_set(struct blk_mq_tag_set *set, const struct blk_mq_ops *ops, unsigned int queue_depth, unsigned int set_flags) blk_mq_alloc_sq_tag_set() argument [all...] |