/kernel/linux/linux-5.10/include/net/ |
H A D | red.h | 147 int qcount; /* Number of packets since last random member 168 v->qcount = -1; in red_set_vars() 287 v->qcount = -1; in red_restart() 386 max_P*(qavg - qth_min)/(qth_max-qth_min) < rnd/qcount in red_mark_probability() 390 return !(((qavg - p->qth_min) >> p->Wlog) * v->qcount < v->qR); in red_mark_probability() 421 v->qcount = -1; in red_action() 425 if (++v->qcount) { in red_action() 427 v->qcount = 0; in red_action() 437 v->qcount = -1; in red_action()
|
/kernel/linux/linux-6.6/include/net/ |
H A D | red.h | 146 int qcount; /* Number of packets since last random member 167 v->qcount = -1; in red_set_vars() 286 v->qcount = -1; in red_restart() 385 max_P*(qavg - qth_min)/(qth_max-qth_min) < rnd/qcount in red_mark_probability() 389 return !(((qavg - p->qth_min) >> p->Wlog) * v->qcount < v->qR); in red_mark_probability() 420 v->qcount = -1; in red_action() 424 if (++v->qcount) { in red_action() 426 v->qcount = 0; in red_action() 436 v->qcount = -1; in red_action()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_tc_mqprio.c | 14 u32 speed, qcount = 0, qoffset = 0; in cxgb4_mqprio_validate() local 49 qcount += mqprio->qopt.count[i]; in cxgb4_mqprio_validate() 81 if (qoffset >= adap->tids.neotids || qcount > adap->tids.neotids) in cxgb4_mqprio_validate() 431 u32 qoffset, qcount, tot_qcount, qid, hwqid; in cxgb4_mqprio_enable_offload() local 446 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 447 for (j = 0; j < qcount; j++) { in cxgb4_mqprio_enable_offload() 486 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 487 if (qcount) { in cxgb4_mqprio_enable_offload() 490 qcount = pi->nqsets; in cxgb4_mqprio_enable_offload() 494 ret = netdev_set_tc_queue(dev, i, qcount, qoffse in cxgb4_mqprio_enable_offload() 537 u32 qoffset, qcount; cxgb4_mqprio_disable_offload() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_tc_mqprio.c | 14 u32 speed, qcount = 0, qoffset = 0; in cxgb4_mqprio_validate() local 49 qcount += mqprio->qopt.count[i]; in cxgb4_mqprio_validate() 81 if (qoffset >= adap->tids.neotids || qcount > adap->tids.neotids) in cxgb4_mqprio_validate() 431 u32 qoffset, qcount, tot_qcount, qid, hwqid; in cxgb4_mqprio_enable_offload() local 446 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 447 for (j = 0; j < qcount; j++) { in cxgb4_mqprio_enable_offload() 486 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 487 if (qcount) { in cxgb4_mqprio_enable_offload() 490 qcount = pi->nqsets; in cxgb4_mqprio_enable_offload() 494 ret = netdev_set_tc_queue(dev, i, qcount, qoffse in cxgb4_mqprio_enable_offload() 537 u32 qoffset, qcount; cxgb4_mqprio_disable_offload() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_vf.c | 456 int err, qcount; in otx2vf_probe() local 478 qcount = num_online_cpus(); in otx2vf_probe() 479 netdev = alloc_etherdev_mqs(sizeof(*vf), qcount, qcount); in otx2vf_probe() 496 hw->rx_queues = qcount; in otx2vf_probe() 497 hw->tx_queues = qcount; in otx2vf_probe() 498 hw->max_queues = qcount; in otx2vf_probe() 547 err = otx2_set_real_num_queues(netdev, qcount, qcount); in otx2vf_probe()
|
H A D | otx2_pf.c | 2005 int err, qcount; in otx2_probe() local 2029 qcount = min_t(int, num_online_cpus(), OTX2_MAX_CQ_CNT); in otx2_probe() 2031 netdev = alloc_etherdev_mqs(sizeof(*pf), qcount, qcount); in otx2_probe() 2048 hw->rx_queues = qcount; in otx2_probe() 2049 hw->tx_queues = qcount; in otx2_probe() 2050 hw->max_queues = qcount; in otx2_probe()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_vf.c | 538 int err, qcount, qos_txqs; in otx2vf_probe() local 563 qcount = num_online_cpus(); in otx2vf_probe() 564 qos_txqs = min_t(int, qcount, OTX2_QOS_MAX_LEAF_NODES); in otx2vf_probe() 565 netdev = alloc_etherdev_mqs(sizeof(*vf), qcount + qos_txqs, qcount); in otx2vf_probe() 582 hw->rx_queues = qcount; in otx2vf_probe() 583 hw->tx_queues = qcount; in otx2vf_probe() 584 hw->max_queues = qcount; in otx2vf_probe() 585 hw->non_qos_queues = qcount; in otx2vf_probe() 638 err = otx2_set_real_num_queues(netdev, qcount, qcoun in otx2vf_probe() [all...] |
H A D | otx2_pf.c | 2857 int err, qcount, qos_txqs; in otx2_probe() local 2884 qcount = min_t(int, num_online_cpus(), OTX2_MAX_CQ_CNT); in otx2_probe() 2885 qos_txqs = min_t(int, qcount, OTX2_QOS_MAX_LEAF_NODES); in otx2_probe() 2887 netdev = alloc_etherdev_mqs(sizeof(*pf), qcount + qos_txqs, qcount); in otx2_probe() 2904 hw->rx_queues = qcount; in otx2_probe() 2905 hw->tx_queues = qcount; in otx2_probe() 2906 hw->non_qos_queues = qcount; in otx2_probe() 2907 hw->max_queues = qcount; in otx2_probe()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_dcb_lib.c | 219 u16 qoffset, qcount; in ice_vsi_cfg_dcb_rings() local 240 qcount = vsi->tc_cfg.tc_info[n].qcount_tx; in ice_vsi_cfg_dcb_rings() 241 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 244 qcount = vsi->tc_cfg.tc_info[n].qcount_rx; in ice_vsi_cfg_dcb_rings() 245 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 262 qcount = vsi->mqprio_qopt.qopt.count[n]; in ice_vsi_cfg_dcb_rings() 263 for (i = qoffset; i < (qoffset + qcount); i++) { in ice_vsi_cfg_dcb_rings()
|
H A D | ice_main.c | 7965 * This function validates MQPRIO params, such as qcount (power of 2 wherever 7966 * needed), and make sure user doesn't specify qcount and BW rate limit 7995 int qcount = mqprio_qopt->qopt.count[i]; in ice_validate_mqprio_qopt() local 7998 if (!qcount) in ice_validate_mqprio_qopt() 8001 if (is_power_of_2(qcount)) { in ice_validate_mqprio_qopt() 8003 qcount > non_power_of_2_qcount) { in ice_validate_mqprio_qopt() 8004 dev_err(dev, "qcount[%d] cannot be greater than non power of 2 qcount[%d]\n", in ice_validate_mqprio_qopt() 8005 qcount, non_power_of_2_qcount); in ice_validate_mqprio_qopt() 8008 if (qcount > max_rss_q_cn in ice_validate_mqprio_qopt() [all...] |
/kernel/linux/linux-5.10/net/sched/ |
H A D | sch_choke.c | 227 q->vars.qcount = -1; in choke_enqueue() 240 q->vars.qcount = -1; in choke_enqueue() 250 } else if (++q->vars.qcount) { in choke_enqueue() 252 q->vars.qcount = 0; in choke_enqueue()
|
/kernel/linux/linux-6.6/net/sched/ |
H A D | sch_choke.c | 226 q->vars.qcount = -1; in choke_enqueue() 239 q->vars.qcount = -1; in choke_enqueue() 249 } else if (++q->vars.qcount) { in choke_enqueue() 251 q->vars.qcount = 0; in choke_enqueue()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_dcb_lib.c | 198 u16 qoffset, qcount; in ice_vsi_cfg_dcb_rings() local 219 qcount = vsi->tc_cfg.tc_info[n].qcount_tx; in ice_vsi_cfg_dcb_rings() 220 for (i = qoffset; i < (qoffset + qcount); i++) { in ice_vsi_cfg_dcb_rings()
|
/kernel/linux/linux-5.10/fs/gfs2/ |
H A D | trace_gfs2.h | 278 __field( u64, qcount ) 295 __entry->qcount = gl->gl_stats.stats[GFS2_LKS_QCOUNT]; 310 (long long)__entry->qcount)
|
/kernel/linux/linux-6.6/fs/gfs2/ |
H A D | trace_gfs2.h | 275 __field( u64, qcount ) 292 __entry->qcount = gl->gl_stats.stats[GFS2_LKS_QCOUNT]; 307 (long long)__entry->qcount)
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 1807 u16 qcount = 0, max_qcount, qmap, sections = 0; in i40e_vsi_setup_queue_map_mqprio() local 1832 qcount = vsi->mqprio_qopt.qopt.count[i]; in i40e_vsi_setup_queue_map_mqprio() 1833 if (qcount > max_qcount) in i40e_vsi_setup_queue_map_mqprio() 1834 max_qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1836 vsi->tc_config.tc_info[i].qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1844 vsi->tc_config.tc_info[i].qcount = 1; in i40e_vsi_setup_queue_map_mqprio() 1850 vsi->num_queue_pairs = offset + qcount; in i40e_vsi_setup_queue_map_mqprio() 1901 u16 qcount; in i40e_vsi_setup_queue_map() local 1970 qcount in i40e_vsi_setup_queue_map() 3583 u16 qoffset, qcount; i40e_vsi_config_dcb_rings() local 6029 u16 qcount, qmap, sections = 0; i40e_channel_setup_queue_map() local 7124 u16 qcount; i40e_vsi_set_default_tc_config() local 11810 u16 qcount; i40e_pf_config_rss() local 11867 u16 qcount; i40e_reconfig_rss_queues() local [all...] |
H A D | i40e_txrx.c | 3549 u16 qcount; in i40e_lan_select_queue() local 3566 qcount = vsi->tc_config.tc_info[tclass].qcount; in i40e_lan_select_queue() 3567 hash = i40e_swdcb_skb_tx_hash(netdev, skb, qcount); in i40e_lan_select_queue()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 1928 u16 qcount = 0, max_qcount, qmap, sections = 0; in i40e_vsi_setup_queue_map_mqprio() local 1953 qcount = vsi->mqprio_qopt.qopt.count[i]; in i40e_vsi_setup_queue_map_mqprio() 1954 if (qcount > max_qcount) in i40e_vsi_setup_queue_map_mqprio() 1955 max_qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1957 vsi->tc_config.tc_info[i].qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1965 vsi->tc_config.tc_info[i].qcount = 1; in i40e_vsi_setup_queue_map_mqprio() 1971 vsi->num_queue_pairs = offset + qcount; in i40e_vsi_setup_queue_map_mqprio() 2022 u16 qcount; in i40e_vsi_setup_queue_map() local 2091 qcount in i40e_vsi_setup_queue_map() 3772 u16 qoffset, qcount; i40e_vsi_config_dcb_rings() local 6228 u16 qcount, qmap, sections = 0; i40e_channel_setup_queue_map() local 7653 u16 qcount; i40e_vsi_set_default_tc_config() local 12446 u16 qcount; i40e_pf_config_rss() local 12503 u16 qcount; i40e_reconfig_rss_queues() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_main.c | 2110 int err, qcount; in nicvf_probe() local 2147 qcount = netif_get_num_default_rss_queues(); in nicvf_probe() 2152 qcount = min_t(int, num_online_cpus(), in nicvf_probe() 2156 netdev = alloc_etherdev_mqs(sizeof(struct nicvf), qcount, qcount); in nicvf_probe() 2170 nic->max_queues = qcount; in nicvf_probe()
|
H A D | nicvf_queues.c | 383 int tail, qcount; in nicvf_refill_rbdr() local 400 qcount = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS0, rbdr_idx); in nicvf_refill_rbdr() 401 qcount &= 0x7FFFF; in nicvf_refill_rbdr() 403 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr() 406 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_main.c | 2099 int err, qcount; in nicvf_probe() local 2128 qcount = netif_get_num_default_rss_queues(); in nicvf_probe() 2133 qcount = min_t(int, num_online_cpus(), in nicvf_probe() 2137 netdev = alloc_etherdev_mqs(sizeof(struct nicvf), qcount, qcount); in nicvf_probe() 2151 nic->max_queues = qcount; in nicvf_probe()
|
H A D | nicvf_queues.c | 384 int tail, qcount; in nicvf_refill_rbdr() local 401 qcount = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS0, rbdr_idx); in nicvf_refill_rbdr() 402 qcount &= 0x7FFFF; in nicvf_refill_rbdr() 404 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr() 407 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 240 u8 qcount[MAX_TQM_ALLOC_REQ]; member
|
H A D | qplib_res.c | 408 if (!tqmctx->qcount[i]) in bnxt_qplib_alloc_tqm_rings() 410 hwq_attr.depth = ctx->qpc_count * tqmctx->qcount[i]; in bnxt_qplib_alloc_tqm_rings()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 255 u8 qcount[MAX_TQM_ALLOC_REQ]; member
|