/kernel/linux/linux-5.10/block/ |
H A D | blk-mq-cpumap.c | 19 unsigned int nr_queues, const int q) in queue_index() 21 return qmap->queue_offset + (q % nr_queues); in queue_index() 38 unsigned int nr_queues = qmap->nr_queues; in blk_mq_map_queues() local 49 if (q >= nr_queues) in blk_mq_map_queues() 51 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 63 if (q < nr_queues) { in blk_mq_map_queues() 64 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 68 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 18 queue_index(struct blk_mq_queue_map *qmap, unsigned int nr_queues, const int q) queue_index() argument
|
H A D | blk-mq-pci.c | 32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues() 44 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues()
|
H A D | blk-mq-virtio.c | 33 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues()
|
H A D | blk-mq-rdma.c | 30 for (queue = 0; queue < map->nr_queues; queue++) { in blk_mq_rdma_map_queues()
|
/kernel/linux/linux-5.10/drivers/crypto/cavium/cpt/ |
H A D | cptvf_main.c | 42 if (cptvf->nr_queues) { in init_worker_threads() 44 cptvf->nr_queues); in init_worker_threads() 47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads() 69 if (cptvf->nr_queues) { in cleanup_worker_threads() 71 cptvf->nr_queues); in cleanup_worker_threads() 74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads() 100 pqinfo->nr_queues = 0; in free_pending_queues() 104 u32 nr_queues) in alloc_pending_queues() 111 pqinfo->nr_queues = nr_queues; in alloc_pending_queues() 103 alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen, u32 nr_queues) alloc_pending_queues() argument 139 init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) init_pending_queues() argument 315 cptvf_sw_init(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) cptvf_sw_init() argument [all...] |
H A D | cptvf.h | 85 u32 nr_queues; /* Number of queues supported */ member 91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \ 110 u32 nr_queues; member
|
/kernel/linux/linux-6.6/drivers/crypto/cavium/cpt/ |
H A D | cptvf_main.c | 42 if (cptvf->nr_queues) { in init_worker_threads() 44 cptvf->nr_queues); in init_worker_threads() 47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads() 69 if (cptvf->nr_queues) { in cleanup_worker_threads() 71 cptvf->nr_queues); in cleanup_worker_threads() 74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads() 100 pqinfo->nr_queues = 0; in free_pending_queues() 104 u32 nr_queues) in alloc_pending_queues() 110 pqinfo->nr_queues = nr_queues; in alloc_pending_queues() 103 alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen, u32 nr_queues) alloc_pending_queues() argument 136 init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) init_pending_queues() argument 312 cptvf_sw_init(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) cptvf_sw_init() argument [all...] |
H A D | cptvf.h | 85 u32 nr_queues; /* Number of queues supported */ member 91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \ 110 u32 nr_queues; member
|
/kernel/linux/linux-5.10/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_sriov.c | 58 int nr_queues = 0; in vf_mode_to_nr_queues() local 62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues() 65 nr_queues = 8; in vf_mode_to_nr_queues() 68 nr_queues = 4; in vf_mode_to_nr_queues() 71 nr_queues = 2; in vf_mode_to_nr_queues() 74 nr_queues = 1; in vf_mode_to_nr_queues() 78 return nr_queues; in vf_mode_to_nr_queues()
|
H A D | nitrox_lib.c | 91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues() 102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues() 142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues() 155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues() 161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
|
H A D | nitrox_dev.h | 152 * @nr_queues: number of queues enabled in VF 160 int nr_queues; member 224 * @nr_queues: Number of command queues 250 u16 nr_queues; member
|
H A D | nitrox_mbx.c | 66 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response() 74 vfdev->nr_queues = 0; in pf2vf_send_response()
|
/kernel/linux/linux-6.6/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_sriov.c | 58 int nr_queues = 0; in vf_mode_to_nr_queues() local 62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues() 65 nr_queues = 8; in vf_mode_to_nr_queues() 68 nr_queues = 4; in vf_mode_to_nr_queues() 71 nr_queues = 2; in vf_mode_to_nr_queues() 74 nr_queues = 1; in vf_mode_to_nr_queues() 78 return nr_queues; in vf_mode_to_nr_queues()
|
H A D | nitrox_lib.c | 91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues() 102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues() 142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues() 155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues() 161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
|
H A D | nitrox_dev.h | 152 * @nr_queues: number of queues enabled in VF 160 int nr_queues; member 224 * @nr_queues: Number of command queues 250 u16 nr_queues; member
|
H A D | nitrox_mbx.c | 68 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response() 76 vfdev->nr_queues = 0; in pf2vf_send_response()
|
/kernel/linux/linux-6.6/drivers/ufs/core/ |
H A D | ufs-mcq.c | 172 hba->nr_queues[HCTX_TYPE_DEFAULT] = rw_queues; in ufshcd_mcq_config_nr_queues() 173 rem -= hba->nr_queues[HCTX_TYPE_DEFAULT]; in ufshcd_mcq_config_nr_queues() 179 hba->nr_queues[HCTX_TYPE_POLL] = poll_queues; in ufshcd_mcq_config_nr_queues() 180 rem -= hba->nr_queues[HCTX_TYPE_POLL]; in ufshcd_mcq_config_nr_queues() 184 hba->nr_queues[HCTX_TYPE_READ] = read_queues; in ufshcd_mcq_config_nr_queues() 185 rem -= hba->nr_queues[HCTX_TYPE_READ]; in ufshcd_mcq_config_nr_queues() 188 if (!hba->nr_queues[HCTX_TYPE_DEFAULT]) in ufshcd_mcq_config_nr_queues() 189 hba->nr_queues[HCTX_TYPE_DEFAULT] = min3(rem, rw_queues, in ufshcd_mcq_config_nr_queues() 193 host->nr_hw_queues += hba->nr_queues[i]; in ufshcd_mcq_config_nr_queues() 377 if (i < hba->nr_hw_queues - hba->nr_queues[HCTX_TYPE_POL in ufshcd_mcq_make_queues_operational() [all...] |
/kernel/linux/linux-6.6/block/ |
H A D | blk-mq-pci.c | 31 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues() 43 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues()
|
H A D | blk-mq-cpumap.c | 23 masks = group_cpus_evenly(qmap->nr_queues); in blk_mq_map_queues() 30 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues()
|
H A D | blk-mq-virtio.c | 32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues()
|
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | auxtrace.c | 197 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument 203 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array() 206 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array() 210 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array() 220 queues->nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__init() 221 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init() 230 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local 234 if (!nr_queues) in auxtrace_queues__grow() 235 nr_queues in auxtrace_queues__grow() [all...] |
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | auxtrace.c | 200 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument 206 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array() 209 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array() 213 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array() 223 queues->nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__init() 224 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init() 233 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local 237 if (!nr_queues) in auxtrace_queues__grow() 238 nr_queues in auxtrace_queues__grow() [all...] |
/kernel/linux/linux-6.6/drivers/block/null_blk/ |
H A D | main.c | 1530 if (nullb->nr_queues != 1) in nullb_to_queue() 1531 index = raw_smp_processor_id() / ((nr_cpu_ids + nullb->nr_queues - 1) / nullb->nr_queues); in nullb_to_queue() 1626 map->nr_queues = submit_queues; in null_map_queues() 1629 map->nr_queues = 0; in null_map_queues() 1632 map->nr_queues = poll_queues; in null_map_queues() 1636 qoff += map->nr_queues; in null_map_queues() 1763 for (i = 0; i < nullb->nr_queues; i++) in cleanup_queues() 1774 nullb->nr_queues--; in null_exit_hctx() 1798 nullb->nr_queues in null_init_hctx() [all...] |
/kernel/linux/linux-6.6/drivers/block/rnbd/ |
H A D | rnbd-clt.c | 1171 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1173 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1180 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues() 1182 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues() 1186 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1187 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues() 1188 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues() 1192 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1193 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
|
/kernel/linux/linux-5.10/drivers/block/null_blk/ |
H A D | null_blk.h | 93 unsigned int nr_queues; member
|