/kernel/linux/linux-5.10/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_mtl.c | 61 static void sxgbe_mtl_set_txfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_txfifosize() argument 68 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 70 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 73 static void sxgbe_mtl_set_rxfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_rxfifosize() argument 80 reg_val = readl(ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 82 writel(reg_val, ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 85 static void sxgbe_mtl_enable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_enable_txqueue() argument 89 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 91 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 94 static void sxgbe_mtl_disable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_disable_txqueue() argument 103 sxgbe_mtl_fc_active(void __iomem *ioaddr, int queue_num, int threshold) sxgbe_mtl_fc_active() argument 115 sxgbe_mtl_fc_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fc_enable() argument 124 sxgbe_mtl_fc_deactive(void __iomem *ioaddr, int queue_num, int threshold) sxgbe_mtl_fc_deactive() argument 136 sxgbe_mtl_fep_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fep_enable() argument 146 sxgbe_mtl_fep_disable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fep_disable() argument 156 sxgbe_mtl_fup_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fup_enable() argument 166 sxgbe_mtl_fup_disable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fup_disable() argument 177 sxgbe_set_tx_mtl_mode(void __iomem *ioaddr, int queue_num, int tx_mode) sxgbe_set_tx_mtl_mode() argument 208 sxgbe_set_rx_mtl_mode(void __iomem *ioaddr, int queue_num, int rx_mode) sxgbe_set_rx_mtl_mode() argument [all...] |
H A D | sxgbe_mtl.h | 64 void (*mtl_set_txfifosize)(void __iomem *ioaddr, int queue_num, 67 void (*mtl_set_rxfifosize)(void __iomem *ioaddr, int queue_num, 70 void (*mtl_enable_txqueue)(void __iomem *ioaddr, int queue_num); 72 void (*mtl_disable_txqueue)(void __iomem *ioaddr, int queue_num); 74 void (*set_tx_mtl_mode)(void __iomem *ioaddr, int queue_num, 77 void (*set_rx_mtl_mode)(void __iomem *ioaddr, int queue_num, 82 void (*mtl_fc_active)(void __iomem *ioaddr, int queue_num, 85 void (*mtl_fc_deactive)(void __iomem *ioaddr, int queue_num, 88 void (*mtl_fc_enable)(void __iomem *ioaddr, int queue_num); 90 void (*mtl_fep_enable)(void __iomem *ioaddr, int queue_num); [all...] |
H A D | sxgbe_main.c | 563 int queue_num, ret; in init_dma_desc_rings() local 569 SXGBE_FOR_EACH_QUEUE(SXGBE_TX_QUEUES, queue_num) { in init_dma_desc_rings() 570 ret = init_tx_ring(priv->device, queue_num, in init_dma_desc_rings() 571 priv->txq[queue_num], tx_rsize); in init_dma_desc_rings() 580 priv->txq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 584 SXGBE_FOR_EACH_QUEUE(SXGBE_RX_QUEUES, queue_num) { in init_dma_desc_rings() 585 ret = init_rx_ring(netd, queue_num, in init_dma_desc_rings() 586 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings() 595 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 603 while (queue_num in init_dma_desc_rings() 637 int queue_num; dma_free_tx_skbufs() local 647 int queue_num; free_dma_desc_resources() local 667 int queue_num; txring_mem_alloc() local 681 int queue_num; rxring_mem_alloc() local 701 int queue_num; sxgbe_mtl_operation_mode() local 798 u8 queue_num; sxgbe_tx_all_clean() local 819 sxgbe_restart_tx_queue(struct sxgbe_priv_data *priv, int queue_num) sxgbe_restart_tx_queue() argument 855 int queue_num; sxgbe_reset_all_tx_queues() local 955 int queue_num; sxgbe_init_dma_engine() local 981 int queue_num; sxgbe_init_mtl_engine() local 998 int queue_num; sxgbe_disable_mtl_engine() local 1027 u8 queue_num; sxgbe_tx_init_coalesce() local 1041 u8 queue_num; sxgbe_tx_del_timer() local 1061 int ret, queue_num; sxgbe_open() local 2065 u8 queue_num; sxgbe_drv_probe() local 2206 u8 queue_num; sxgbe_drv_remove() local [all...] |
H A D | sxgbe_common.h | 194 #define SXGBE_FOR_EACH_QUEUE(max_queues, queue_num) \ 195 for (queue_num = 0; queue_num < max_queues; queue_num++) 358 void (*enable_rxqueue)(void __iomem *ioaddr, int queue_num); 359 void (*disable_rxqueue)(void __iomem *ioaddr, int queue_num);
|
H A D | sxgbe_core.c | 165 static void sxgbe_core_enable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_enable_rxqueue() argument 170 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_enable_rxqueue() 175 static void sxgbe_core_disable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_disable_rxqueue() argument 180 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_disable_rxqueue()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_mtl.c | 61 static void sxgbe_mtl_set_txfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_txfifosize() argument 68 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 70 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 73 static void sxgbe_mtl_set_rxfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_rxfifosize() argument 80 reg_val = readl(ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 82 writel(reg_val, ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 85 static void sxgbe_mtl_enable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_enable_txqueue() argument 89 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 91 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 94 static void sxgbe_mtl_disable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_disable_txqueue() argument 103 sxgbe_mtl_fc_active(void __iomem *ioaddr, int queue_num, int threshold) sxgbe_mtl_fc_active() argument 115 sxgbe_mtl_fc_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fc_enable() argument 124 sxgbe_mtl_fc_deactive(void __iomem *ioaddr, int queue_num, int threshold) sxgbe_mtl_fc_deactive() argument 136 sxgbe_mtl_fep_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fep_enable() argument 146 sxgbe_mtl_fep_disable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fep_disable() argument 156 sxgbe_mtl_fup_enable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fup_enable() argument 166 sxgbe_mtl_fup_disable(void __iomem *ioaddr, int queue_num) sxgbe_mtl_fup_disable() argument 177 sxgbe_set_tx_mtl_mode(void __iomem *ioaddr, int queue_num, int tx_mode) sxgbe_set_tx_mtl_mode() argument 208 sxgbe_set_rx_mtl_mode(void __iomem *ioaddr, int queue_num, int rx_mode) sxgbe_set_rx_mtl_mode() argument [all...] |
H A D | sxgbe_mtl.h | 64 void (*mtl_set_txfifosize)(void __iomem *ioaddr, int queue_num, 67 void (*mtl_set_rxfifosize)(void __iomem *ioaddr, int queue_num, 70 void (*mtl_enable_txqueue)(void __iomem *ioaddr, int queue_num); 72 void (*mtl_disable_txqueue)(void __iomem *ioaddr, int queue_num); 74 void (*set_tx_mtl_mode)(void __iomem *ioaddr, int queue_num, 77 void (*set_rx_mtl_mode)(void __iomem *ioaddr, int queue_num, 82 void (*mtl_fc_active)(void __iomem *ioaddr, int queue_num, 85 void (*mtl_fc_deactive)(void __iomem *ioaddr, int queue_num, 88 void (*mtl_fc_enable)(void __iomem *ioaddr, int queue_num); 90 void (*mtl_fep_enable)(void __iomem *ioaddr, int queue_num); [all...] |
H A D | sxgbe_main.c | 563 int queue_num, ret; in init_dma_desc_rings() local 569 SXGBE_FOR_EACH_QUEUE(SXGBE_TX_QUEUES, queue_num) { in init_dma_desc_rings() 570 ret = init_tx_ring(priv->device, queue_num, in init_dma_desc_rings() 571 priv->txq[queue_num], tx_rsize); in init_dma_desc_rings() 580 priv->txq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 584 SXGBE_FOR_EACH_QUEUE(SXGBE_RX_QUEUES, queue_num) { in init_dma_desc_rings() 585 ret = init_rx_ring(netd, queue_num, in init_dma_desc_rings() 586 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings() 595 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 603 while (queue_num in init_dma_desc_rings() 637 int queue_num; dma_free_tx_skbufs() local 647 int queue_num; free_dma_desc_resources() local 667 int queue_num; txring_mem_alloc() local 681 int queue_num; rxring_mem_alloc() local 701 int queue_num; sxgbe_mtl_operation_mode() local 798 u8 queue_num; sxgbe_tx_all_clean() local 819 sxgbe_restart_tx_queue(struct sxgbe_priv_data *priv, int queue_num) sxgbe_restart_tx_queue() argument 855 int queue_num; sxgbe_reset_all_tx_queues() local 958 int queue_num; sxgbe_init_dma_engine() local 984 int queue_num; sxgbe_init_mtl_engine() local 1001 int queue_num; sxgbe_disable_mtl_engine() local 1030 u8 queue_num; sxgbe_tx_init_coalesce() local 1044 u8 queue_num; sxgbe_tx_del_timer() local 1064 int ret, queue_num; sxgbe_open() local 2068 u8 queue_num; sxgbe_drv_probe() local 2209 u8 queue_num; sxgbe_drv_remove() local [all...] |
H A D | sxgbe_common.h | 194 #define SXGBE_FOR_EACH_QUEUE(max_queues, queue_num) \ 195 for (queue_num = 0; queue_num < max_queues; queue_num++) 358 void (*enable_rxqueue)(void __iomem *ioaddr, int queue_num); 359 void (*disable_rxqueue)(void __iomem *ioaddr, int queue_num);
|
H A D | sxgbe_core.c | 166 static void sxgbe_core_enable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_enable_rxqueue() argument 171 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_enable_rxqueue() 176 static void sxgbe_core_disable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_disable_rxqueue() argument 181 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_disable_rxqueue()
|
/kernel/linux/linux-5.10/tools/testing/selftests/netfilter/ |
H A D | nf-queue.c | 22 unsigned int queue_num; member 33 printf("Usage: %s [-c|-v [-vv] ] [-t timeout] [-q queue_num] [-Qdst_queue ] [ -d ms_delay ] [-G]\n", p); in help() 121 nfq_build_cfg_request(char *buf, uint8_t command, int queue_num) in nfq_build_cfg_request() argument 137 nfg->res_id = htons(queue_num); in nfq_build_cfg_request() 145 nfq_build_cfg_params(char *buf, uint8_t mode, int range, int queue_num) in nfq_build_cfg_params() argument 160 nfg->res_id = htons(queue_num); in nfq_build_cfg_params() 168 nfq_build_verdict(char *buf, int id, int queue_num, uint32_t verd) in nfq_build_verdict() argument 183 nfg->res_id = htons(queue_num); in nfq_build_verdict() 210 unsigned int queue_num; in open_queue() local 227 queue_num in open_queue() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/netfilter/ |
H A D | nf-queue.c | 22 unsigned int queue_num; member 33 printf("Usage: %s [-c|-v [-vv] ] [-t timeout] [-q queue_num] [-Qdst_queue ] [ -d ms_delay ] [-G]\n", p); in help() 121 nfq_build_cfg_request(char *buf, uint8_t command, int queue_num) in nfq_build_cfg_request() argument 137 nfg->res_id = htons(queue_num); in nfq_build_cfg_request() 145 nfq_build_cfg_params(char *buf, uint8_t mode, int range, int queue_num) in nfq_build_cfg_params() argument 160 nfg->res_id = htons(queue_num); in nfq_build_cfg_params() 168 nfq_build_verdict(char *buf, int id, int queue_num, uint32_t verd) in nfq_build_verdict() argument 183 nfg->res_id = htons(queue_num); in nfq_build_verdict() 210 unsigned int queue_num; in open_queue() local 227 queue_num in open_queue() [all...] |
/kernel/linux/linux-6.6/samples/pktgen/ |
H A D | pktgen_sample06_numa_awared_queue_irq_affinity.sh | 68 queue_num=$i 69 info "queue number is $queue_num" 70 pg_set $dev "queue_map_min $queue_num" 71 pg_set $dev "queue_map_max $queue_num"
|
/kernel/linux/linux-5.10/net/netfilter/ |
H A D | nfnetlink_queue.c | 68 u_int16_t queue_num; /* number of this queue */ member 96 static inline u_int8_t instance_hashfn(u_int16_t queue_num) in instance_hashfn() argument 98 return ((queue_num >> 8) ^ queue_num) % INSTANCE_BUCKETS; in instance_hashfn() 102 instance_lookup(struct nfnl_queue_net *q, u_int16_t queue_num) in instance_lookup() argument 107 head = &q->instance_table[instance_hashfn(queue_num)]; in instance_lookup() 109 if (inst->queue_num == queue_num) in instance_lookup() 116 instance_create(struct nfnl_queue_net *q, u_int16_t queue_num, u32 portid) in instance_create() argument 123 if (instance_lookup(q, queue_num)) { in instance_create() 1026 verdict_instance_lookup(struct nfnl_queue_net *q, u16 queue_num, u32 nlportid) verdict_instance_lookup() argument 1073 u16 queue_num = ntohs(nfmsg->res_id); nfqnl_recv_verdict_batch() local 1176 u_int16_t queue_num = ntohs(nfmsg->res_id); nfqnl_recv_verdict() local 1262 u_int16_t queue_num = ntohs(nfmsg->res_id); nfqnl_recv_config() local [all...] |
/kernel/linux/linux-6.6/net/netfilter/ |
H A D | nfnetlink_queue.c | 70 u_int16_t queue_num; /* number of this queue */ member 98 static inline u_int8_t instance_hashfn(u_int16_t queue_num) in instance_hashfn() argument 100 return ((queue_num >> 8) ^ queue_num) % INSTANCE_BUCKETS; in instance_hashfn() 104 instance_lookup(struct nfnl_queue_net *q, u_int16_t queue_num) in instance_lookup() argument 109 head = &q->instance_table[instance_hashfn(queue_num)]; in instance_lookup() 111 if (inst->queue_num == queue_num) in instance_lookup() 118 instance_create(struct nfnl_queue_net *q, u_int16_t queue_num, u32 portid) in instance_create() argument 125 if (instance_lookup(q, queue_num)) { in instance_create() 1066 verdict_instance_lookup(struct nfnl_queue_net *q, u16 queue_num, u32 nlportid) verdict_instance_lookup() argument 1106 u16 queue_num = ntohs(info->nfmsg->res_id); nfqnl_recv_verdict_batch() local 1217 u_int16_t queue_num = ntohs(info->nfmsg->res_id); nfqnl_recv_verdict() local 1302 u_int16_t queue_num = ntohs(info->nfmsg->res_id); nfqnl_recv_config() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/ |
H A D | hns3_debugfs.c | 22 u32 queue_num, queue_max; in hns3_dbg_queue_info() local 32 cnt = kstrtouint(&cmd_buf[11], 0, &queue_num); in hns3_dbg_queue_info() 34 queue_num = 0; in hns3_dbg_queue_info() 36 queue_max = queue_num + 1; in hns3_dbg_queue_info() 40 if (queue_num >= h->kinfo.num_tqps) { in hns3_dbg_queue_info() 42 "Queue number(%u) is out of range(0-%u)\n", queue_num, in hns3_dbg_queue_info() 47 for (i = queue_num; i < queue_max; i++) { in hns3_dbg_queue_info() 296 " [rtc] [ppp] [rcb] [tqp <queue_num>] [mac]]\n", in hns3_dbg_help()
|
/kernel/linux/linux-5.10/drivers/net/dsa/ |
H A D | bcm_sf2_cfp.c | 216 unsigned int queue_num, in bcm_sf2_cfp_act_pol_set() 228 CHANGE_TC | queue_num << NEW_TC_SHIFT; in bcm_sf2_cfp_act_pol_set() 348 unsigned int queue_num, in bcm_sf2_cfp_ipv4_rule_set() 461 queue_num, true); in bcm_sf2_cfp_ipv4_rule_set() 629 unsigned int queue_num, in bcm_sf2_cfp_ipv6_rule_set() 763 queue_num, false); in bcm_sf2_cfp_ipv6_rule_set() 821 queue_num, true); in bcm_sf2_cfp_ipv6_rule_set() 853 unsigned int queue_num, port_num; in bcm_sf2_cfp_rule_insert() local 906 queue_num = ring_cookie % SF2_NUM_EGRESS_QUEUES; in bcm_sf2_cfp_rule_insert() 914 queue_num, f in bcm_sf2_cfp_rule_insert() 212 bcm_sf2_cfp_act_pol_set(struct bcm_sf2_priv *priv, unsigned int rule_index, int src_port, unsigned int port_num, unsigned int queue_num, bool fwd_map_change) bcm_sf2_cfp_act_pol_set() argument 346 bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port, unsigned int port_num, unsigned int queue_num, struct ethtool_rx_flow_spec *fs) bcm_sf2_cfp_ipv4_rule_set() argument 627 bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port, unsigned int port_num, unsigned int queue_num, struct ethtool_rx_flow_spec *fs) bcm_sf2_cfp_ipv6_rule_set() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/dsa/ |
H A D | bcm_sf2_cfp.c | 216 unsigned int queue_num, in bcm_sf2_cfp_act_pol_set() 228 CHANGE_TC | queue_num << NEW_TC_SHIFT; in bcm_sf2_cfp_act_pol_set() 348 unsigned int queue_num, in bcm_sf2_cfp_ipv4_rule_set() 461 queue_num, true); in bcm_sf2_cfp_ipv4_rule_set() 629 unsigned int queue_num, in bcm_sf2_cfp_ipv6_rule_set() 763 queue_num, false); in bcm_sf2_cfp_ipv6_rule_set() 821 queue_num, true); in bcm_sf2_cfp_ipv6_rule_set() 853 unsigned int queue_num, port_num; in bcm_sf2_cfp_rule_insert() local 903 queue_num = ring_cookie % SF2_NUM_EGRESS_QUEUES; in bcm_sf2_cfp_rule_insert() 911 queue_num, f in bcm_sf2_cfp_rule_insert() 212 bcm_sf2_cfp_act_pol_set(struct bcm_sf2_priv *priv, unsigned int rule_index, int src_port, unsigned int port_num, unsigned int queue_num, bool fwd_map_change) bcm_sf2_cfp_act_pol_set() argument 346 bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port, unsigned int port_num, unsigned int queue_num, struct ethtool_rx_flow_spec *fs) bcm_sf2_cfp_ipv4_rule_set() argument 627 bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port, unsigned int port_num, unsigned int queue_num, struct ethtool_rx_flow_spec *fs) bcm_sf2_cfp_ipv6_rule_set() argument [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | blk-mq.h | 135 /** @queue_num: Index of this hardware queue. */ 136 unsigned int queue_num; member 591 return rq->tag | (hctx->queue_num << BLK_QC_T_SHIFT); in request_to_qc_t() 593 return rq->internal_tag | (hctx->queue_num << BLK_QC_T_SHIFT) | in request_to_qc_t()
|
/kernel/linux/linux-5.10/drivers/soc/ti/ |
H A D | knav_qmss_acc.c | 284 cmd->queue_num, cmd->timer_config); in knav_acc_write() 287 writel_relaxed(cmd->queue_num, &pdsp->acc_command->queue_num); in knav_acc_write() 324 cmd->queue_num = info->list_entries << 16; in knav_acc_setup_cmd() 325 cmd->queue_num |= queue_base; in knav_acc_setup_cmd()
|
/kernel/linux/linux-6.6/drivers/soc/ti/ |
H A D | knav_qmss_acc.c | 284 cmd->queue_num, cmd->timer_config); in knav_acc_write() 287 writel_relaxed(cmd->queue_num, &pdsp->acc_command->queue_num); in knav_acc_write() 324 cmd->queue_num = info->list_entries << 16; in knav_acc_setup_cmd() 325 cmd->queue_num |= queue_base; in knav_acc_setup_cmd()
|
/kernel/linux/linux-5.10/block/ |
H A D | blk-mq-tag.c | 596 new = blk_mq_alloc_rq_map(set, hctx->queue_num, tdepth, in blk_mq_tag_update_depth() 600 ret = blk_mq_alloc_rqs(set, new, hctx->queue_num, tdepth); in blk_mq_tag_update_depth() 606 blk_mq_free_rqs(set, *tagsptr, hctx->queue_num); in blk_mq_tag_update_depth() 640 return (rq->mq_hctx->queue_num << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
|
/kernel/linux/linux-5.10/drivers/net/wireless/broadcom/b43legacy/ |
H A D | main.c | 2452 int queue_num; in b43legacy_tx_work() local 2462 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_tx_work() 2463 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43legacy_tx_work() 2464 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43legacy_tx_work() 2470 wl->tx_queue_stopped[queue_num] = 1; in b43legacy_tx_work() 2471 ieee80211_stop_queue(wl->hw, queue_num); in b43legacy_tx_work() 2472 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43legacy_tx_work() 2481 wl->tx_queue_stopped[queue_num] in b43legacy_tx_work() 2917 int queue_num; b43legacy_wireless_core_stop() local 3790 int queue_num; b43legacy_wireless_init() local [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/broadcom/b43legacy/ |
H A D | main.c | 2452 int queue_num; in b43legacy_tx_work() local 2462 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_tx_work() 2463 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43legacy_tx_work() 2464 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43legacy_tx_work() 2470 wl->tx_queue_stopped[queue_num] = 1; in b43legacy_tx_work() 2471 ieee80211_stop_queue(wl->hw, queue_num); in b43legacy_tx_work() 2472 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43legacy_tx_work() 2481 wl->tx_queue_stopped[queue_num] in b43legacy_tx_work() 2918 int queue_num; b43legacy_wireless_core_stop() local 3791 int queue_num; b43legacy_wireless_init() local [all...] |
/kernel/linux/linux-5.10/drivers/dma/ |
H A D | fsl-qdma.c | 486 int queue_num, block_number; in fsl_qdma_alloc_queue_resources() local 490 queue_num = fsl_qdma->n_queues; in fsl_qdma_alloc_queue_resources() 493 if (queue_num > FSL_QDMA_QUEUE_MAX) in fsl_qdma_alloc_queue_resources() 494 queue_num = FSL_QDMA_QUEUE_MAX; in fsl_qdma_alloc_queue_resources() 495 len = sizeof(*queue_head) * queue_num * block_number; in fsl_qdma_alloc_queue_resources() 501 queue_size, queue_num); in fsl_qdma_alloc_queue_resources() 507 for (i = 0; i < queue_num; i++) { in fsl_qdma_alloc_queue_resources() 514 queue_temp = queue_head + i + (j * queue_num); in fsl_qdma_alloc_queue_resources()
|