/kernel/linux/linux-5.10/block/ |
H A D | blk-mq-sched.c | 130 max_dispatch = hctx->queue->nr_requests; in __blk_mq_do_dispatch_sched() 525 hctx->sched_tags = blk_mq_alloc_rq_map(set, hctx_idx, q->nr_requests, in blk_mq_sched_alloc_tags() 530 ret = blk_mq_alloc_rqs(set, hctx->sched_tags, hctx_idx, q->nr_requests); in blk_mq_sched_alloc_tags() 565 q->nr_requests = q->tag_set->queue_depth; in blk_mq_init_sched() 574 q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth, in blk_mq_init_sched()
|
H A D | blk-iolatency.c | 332 unsigned long qd = blkiolat->rqos.q->nr_requests; in scale_cookie_change() 372 unsigned long qd = iolat->blkiolat->rqos.q->nr_requests; in scale_change() 995 iolat->rq_depth.queue_depth = blkg->q->nr_requests; in iolatency_pd_init()
|
H A D | blk-sysfs.c | 63 return queue_var_show(q->nr_requests, (page)); in queue_requests_show() 564 QUEUE_RW_ENTRY(queue_requests, "nr_requests");
|
H A D | blk-mq.c | 3331 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 3616 if (q->nr_requests == nr) in blk_mq_update_nr_requests() 3646 q->nr_requests = nr; in blk_mq_update_nr_requests()
|
H A D | blk-core.c | 581 q->nr_requests = BLKDEV_MAX_RQ; in blk_alloc_queue()
|
/kernel/linux/linux-5.10/drivers/s390/block/ |
H A D | scm_blk.c | 30 static unsigned int nr_requests = 64; variable 33 module_param(nr_requests, uint, S_IRUGO); 34 MODULE_PARM_DESC(nr_requests, "Number of parallel requests."); 457 bdev->tag_set.nr_hw_queues = nr_requests; in scm_blk_dev_setup() 458 bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests; in scm_blk_dev_setup() 557 ret = scm_alloc_rqs(nr_requests); in scm_blk_init()
|
/kernel/linux/linux-6.6/drivers/s390/block/ |
H A D | scm_blk.c | 29 static unsigned int nr_requests = 64; variable 32 module_param(nr_requests, uint, S_IRUGO); 33 MODULE_PARM_DESC(nr_requests, "Number of parallel requests."); 456 bdev->tag_set.nr_hw_queues = nr_requests; in scm_blk_dev_setup() 457 bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests; in scm_blk_dev_setup() 552 ret = scm_alloc_rqs(nr_requests); in scm_blk_init()
|
/kernel/linux/linux-6.6/block/ |
H A D | blk-mq-sched.c | 100 max_dispatch = hctx->queue->nr_requests; in __blk_mq_do_dispatch_sched() 392 q->nr_requests); in blk_mq_sched_alloc_map_and_rqs() 429 * updating nr_requests. in blk_mq_init_sched_shared_tags() 456 q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth, in blk_mq_init_sched()
|
H A D | blk-iolatency.c | 333 unsigned long qd = blkiolat->rqos.disk->queue->nr_requests; in scale_cookie_change() 375 unsigned long qd = iolat->blkiolat->rqos.disk->queue->nr_requests; in scale_change()
|
H A D | blk-mq-tag.c | 663 q->nr_requests - q->tag_set->reserved_tags); in blk_mq_tag_update_sched_shared_tags()
|
H A D | elevator.c | 708 q->nr_requests = q->tag_set->queue_depth; in elevator_disable()
|
H A D | blk-sysfs.c | 52 return queue_var_show(q->nr_requests, page); in queue_requests_show() 489 QUEUE_RW_ENTRY(queue_requests, "nr_requests");
|
H A D | blk-core.c | 443 q->nr_requests = BLKDEV_DEFAULT_RQ; in blk_alloc_queue()
|
H A D | blk-mq.c | 4310 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 4632 if (q->nr_requests == nr) in blk_mq_update_nr_requests() 4659 q->nr_requests = nr; in blk_mq_update_nr_requests()
|
H A D | bfq-iosched.c | 695 unsigned limit = data->q->nr_requests; in bfq_limit_depth()
|
/kernel/linux/linux-5.10/drivers/dma/ |
H A D | stm32-mdma.c | 276 u32 nr_requests; member 1499 if (config.request >= dmadev->nr_requests) { in stm32_mdma_of_xlate() 1535 u32 nr_channels, nr_requests; in stm32_mdma_probe() local 1551 &nr_requests); in stm32_mdma_probe() 1553 nr_requests = STM32_MDMA_MAX_REQUESTS; in stm32_mdma_probe() 1555 nr_requests); in stm32_mdma_probe() 1568 dmadev->nr_requests = nr_requests; in stm32_mdma_probe()
|
H A D | owl-dma.c | 1093 int ret, i, nr_channels, nr_requests; in owl_dma_probe() local 1109 ret = of_property_read_u32(np, "dma-requests", &nr_requests); in owl_dma_probe() 1116 nr_channels, nr_requests); in owl_dma_probe() 1121 od->nr_vchans = nr_requests; in owl_dma_probe()
|
/kernel/linux/linux-6.6/drivers/dma/ |
H A D | stm32-mdma.c | 255 u32 nr_requests; member 1563 if (config.request >= dmadev->nr_requests) { in stm32_mdma_of_xlate() 1598 u32 nr_channels, nr_requests; in stm32_mdma_probe() local 1614 &nr_requests); in stm32_mdma_probe() 1616 nr_requests = STM32_MDMA_MAX_REQUESTS; in stm32_mdma_probe() 1618 nr_requests); in stm32_mdma_probe() 1632 dmadev->nr_requests = nr_requests; in stm32_mdma_probe()
|
H A D | owl-dma.c | 1095 int ret, i, nr_channels, nr_requests; in owl_dma_probe() local 1111 ret = of_property_read_u32(np, "dma-requests", &nr_requests); in owl_dma_probe() 1118 nr_channels, nr_requests); in owl_dma_probe() 1123 od->nr_vchans = nr_requests; in owl_dma_probe()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | blkdev.h | 444 unsigned long nr_requests; /* Max # of requests */ member 718 return q->nr_requests; in blk_queue_depth()
|
/kernel/linux/linux-5.10/drivers/target/ |
H A D | target_core_iblock.c | 110 dev->dev_attrib.hw_queue_depth = q->nr_requests; in iblock_configure_device()
|
/kernel/linux/linux-6.6/drivers/target/ |
H A D | target_core_iblock.c | 132 dev->dev_attrib.hw_queue_depth = q->nr_requests; in iblock_configure_device()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | blkdev.h | 478 unsigned long nr_requests; /* Max # of requests */ member 853 return q->nr_requests; in blk_queue_depth()
|
/kernel/linux/linux-5.10/drivers/block/ |
H A D | skd_main.c | 503 WARN_ONCE(tag >= skd_max_queue_depth, "%#x > %#x (nr_requests = %lu)\n", in skd_mq_queue_rq() 504 tag, skd_max_queue_depth, q->nr_requests); in skd_mq_queue_rq()
|