/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_main.c | 441 static void fun_free_rings(struct net_device *netdev, struct fun_qset *qset) in fun_free_rings() argument 444 struct funeth_txq **xdpqs = qset->xdpqs; in fun_free_rings() 445 struct funeth_rxq **rxqs = qset->rxqs; in fun_free_rings() 447 /* qset may not specify any queues to operate on. In that case the in fun_free_rings() 453 qset->txqs = fp->txqs; in fun_free_rings() 454 qset->nrxqs = netdev->real_num_rx_queues; in fun_free_rings() 455 qset->ntxqs = netdev->real_num_tx_queues; in fun_free_rings() 456 qset->nxdpqs = fp->num_xdpqs; in fun_free_rings() 468 free_rxqs(rxqs, qset->nrxqs, qset in fun_free_rings() 479 fun_alloc_rings(struct net_device *netdev, struct fun_qset *qset) fun_alloc_rings() argument 530 fun_advance_ring_state(struct net_device *dev, struct fun_qset *qset) fun_advance_ring_state() argument 741 fun_down(struct net_device *dev, struct fun_qset *qset) fun_down() argument 768 fun_up(struct net_device *dev, struct fun_qset *qset) fun_up() argument 837 struct fun_qset qset = { funeth_open() local 863 struct fun_qset qset = { .state = FUN_QSTATE_DESTROYED }; funeth_close() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 23 if (!pfvf->qset.pool) in otx2_qos_aura_pool_free() 26 pool = &pfvf->qset.pool[pool_id]; in otx2_qos_aura_pool_free() 35 struct otx2_qset *qset = &pfvf->qset; in otx2_qos_sq_aura_pool_init() local 50 num_sqbs = (qset->sqe_cnt + num_sqbs) / num_sqbs; in otx2_qos_sq_aura_pool_init() 57 pool = &pfvf->qset.pool[pool_id]; in otx2_qos_sq_aura_pool_init() 76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init() 118 struct otx2_qset *qset = &pfvf->qset; in otx2_qos_sq_free_sqbs() local 124 sq = &qset in otx2_qos_sq_free_sqbs() 261 struct otx2_qset *qset = &pfvf->qset; otx2_qos_disable_sq() local [all...] |
H A D | otx2_common.c | 78 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats() 80 if (!pfvf->qset.rq) in otx2_update_rq_stats() 89 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats() 91 if (!pfvf->qset.sq) in otx2_update_sq_stats() 824 sq = &pfvf->qset.sq[qidx]; in otx2_sqb_flush() 855 struct otx2_qset *qset = &pfvf->qset; in otx2_rq_init() local 873 aq->rq.xqe_pass = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init() 874 aq->rq.xqe_drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init() 892 sq = &pfvf->qset in otx2_sq_aq_init() 925 struct otx2_qset *qset = &pfvf->qset; otx2_sq_init() local 987 struct otx2_qset *qset = &pfvf->qset; otx2_cq_init() local 1187 struct otx2_qset *qset = &pfvf->qset; otx2_sq_free_sqbs() local 1429 struct otx2_qset *qset = &pfvf->qset; otx2_sq_aura_pool_init() local 1550 struct otx2_qset *qset = &pfvf->qset; otx2_config_npa() local 1771 struct otx2_qset *qset = &pfvf->qset; otx2_free_cints() local [all...] |
H A D | otx2_pf.c | 1261 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1292 sq = &pf->qset.sq[qidx]; in otx2_q_intr_handler() 1391 struct otx2_qset *qset = &pf->qset; in otx2_disable_napi() local 1396 cq_poll = &qset->napi[qidx]; in otx2_disable_napi() 1405 struct otx2_qset *qset = &pf->qset; in otx2_free_cq_res() local 1411 for (qidx = 0; qidx < qset->cq_cnt; qidx++) { in otx2_free_cq_res() 1412 cq = &qset->cq[qidx]; in otx2_free_cq_res() 1419 struct otx2_qset *qset in otx2_free_sq_res() local 1594 struct otx2_qset *qset = &pf->qset; otx2_free_hw_resources() local 1764 struct otx2_qset *qset = &pf->qset; otx2_open() local 1973 struct otx2_qset *qset = &pf->qset; otx2_stop() local [all...] |
H A D | otx2_txrx.c | 463 sq = &pfvf->qset.sq[qidx]; in otx2_tx_napi_handler() 478 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[qidx], in otx2_tx_napi_handler() 536 struct otx2_qset *qset; in otx2_napi_handler() local 542 qset = &pfvf->qset; in otx2_napi_handler() 548 cq = &qset->cq[cq_idx]; in otx2_napi_handler() 1225 pool = &pfvf->qset.pool[pool_id]; in otx2_cleanup_rx_cqes() 1260 sq = &pfvf->qset.sq[qidx]; in otx2_cleanup_tx_cqes() 1325 sq = &pfvf->qset.sq[sq_idx]; in otx2_free_pending_sqe() 1373 sq = &pfvf->qset in otx2_xdp_sq_append_pkt() [all...] |
H A D | otx2_ethtool.c | 83 static void otx2_get_qset_strings(struct otx2_nic *pfvf, u8 **data, int qset) in otx2_get_qset_strings() argument 85 int start_qidx = qset * pfvf->hw.rx_queues; in otx2_get_qset_strings() 164 *((*data)++) = ((u64 *)&pfvf->qset.rq[qidx].stats) in otx2_get_qset_stats() 175 *((*data)++) = ((u64 *)&pfvf->qset.sq[qidx].stats) in otx2_get_qset_stats() 382 struct otx2_qset *qs = &pfvf->qset; in otx2_get_ringparam() 402 struct otx2_qset *qs = &pfvf->qset; in otx2_set_ringparam()
|
H A D | cn10k.c | 99 aq->sq.cq_limit = ((SEND_CQ_SKID * 256) / (pfvf->qset.sqe_cnt)); in cn10k_sq_aq_init()
|
H A D | otx2_common.h | 472 struct otx2_qset qset; member 599 pfvf->qset.rqe_cnt = Q_COUNT(Q_SIZE_1K); in otx2_setup_dev_hw_settings()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_idc.c | 58 * @qset: Resource to be allocated 60 int ice_add_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset) in ice_add_rdma_qset() argument 70 if (WARN_ON(!pf || !qset)) in ice_add_rdma_qset() 87 max_rdmaqs[qset->tc]++; in ice_add_rdma_qset() 88 qs_handle = qset->qs_handle; in ice_add_rdma_qset() 97 status = ice_ena_vsi_rdma_qset(vsi->port_info, vsi->idx, qset->tc, in ice_add_rdma_qset() 103 vsi->qset_handle[qset->tc] = qset->qs_handle; in ice_add_rdma_qset() 104 qset->teid = qset_teid; in ice_add_rdma_qset() 113 * @qset 115 ice_del_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset) ice_del_rdma_qset() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 65 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats() 67 if (!pfvf->qset.rq) in otx2_update_rq_stats() 76 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats() 78 if (!pfvf->qset.sq) in otx2_update_sq_stats() 706 struct otx2_qset *qset = &pfvf->qset; in otx2_rq_init() local 724 aq->rq.xqe_pass = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init() 725 aq->rq.xqe_drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init() 739 struct otx2_qset *qset = &pfvf->qset; in otx2_sq_init() local 816 struct otx2_qset *qset = &pfvf->qset; otx2_cq_init() local 1012 struct otx2_qset *qset = &pfvf->qset; otx2_sq_free_sqbs() local 1211 struct otx2_qset *qset = &pfvf->qset; otx2_sq_aura_pool_init() local 1329 struct otx2_qset *qset = &pfvf->qset; otx2_config_npa() local 1547 struct otx2_qset *qset = &pfvf->qset; otx2_free_cints() local [all...] |
H A D | otx2_pf.c | 1150 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1243 struct otx2_qset *qset = &pf->qset; in otx2_disable_napi() local 1248 cq_poll = &qset->napi[qidx]; in otx2_disable_napi() 1256 struct otx2_qset *qset = &pf->qset; in otx2_free_cq_res() local 1262 for (qidx = 0; qidx < qset->cq_cnt; qidx++) { in otx2_free_cq_res() 1263 cq = &qset->cq[qidx]; in otx2_free_cq_res() 1270 struct otx2_qset *qset = &pf->qset; in otx2_free_sq_res() local 1390 struct otx2_qset *qset = &pf->qset; otx2_free_hw_resources() local 1457 struct otx2_qset *qset = &pf->qset; otx2_open() local 1637 struct otx2_qset *qset = &pf->qset; otx2_stop() local [all...] |
H A D | otx2_ethtool.c | 78 static void otx2_get_qset_strings(struct otx2_nic *pfvf, u8 **data, int qset) in otx2_get_qset_strings() argument 80 int start_qidx = qset * pfvf->hw.rx_queues; in otx2_get_qset_strings() 147 *((*data)++) = ((u64 *)&pfvf->qset.rq[qidx].stats) in otx2_get_qset_stats() 158 *((*data)++) = ((u64 *)&pfvf->qset.sq[qidx].stats) in otx2_get_qset_stats() 237 pfvf->qset.cq_cnt = pfvf->hw.tx_queues + pfvf->hw.rx_queues; in otx2_set_channels() 301 struct otx2_qset *qs = &pfvf->qset; in otx2_get_ringparam() 314 struct otx2_qset *qs = &pfvf->qset; in otx2_set_ringparam()
|
H A D | otx2_txrx.c | 372 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[cq->cint_idx], in otx2_tx_napi_handler() 402 struct otx2_qset *qset; in otx2_napi_handler() local 407 qset = &pfvf->qset; in otx2_napi_handler() 413 cq = &qset->cq[cq_idx]; in otx2_napi_handler() 939 sq = &pfvf->qset.sq[cq->cint_idx]; in otx2_cleanup_tx_cqes()
|
H A D | otx2_common.h | 243 struct otx2_qset qset; member 315 pfvf->qset.rqe_cnt = Q_COUNT(Q_SIZE_1K); in otx2_setup_dev_hw_settings()
|
H A D | otx2_vf.c | 368 sq = &vf->qset.sq[qidx]; in otx2vf_xmit()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | main.c | 161 * irdma_lan_register_qset - Register qset with LAN driver 170 struct iidc_rdma_qset_params qset = {}; in irdma_lan_register_qset() local 173 qset.qs_handle = tc_node->qs_handle; in irdma_lan_register_qset() 174 qset.tc = tc_node->traffic_class; in irdma_lan_register_qset() 175 qset.vport_id = vsi->vsi_idx; in irdma_lan_register_qset() 176 ret = ice_add_rdma_qset(pf, &qset); in irdma_lan_register_qset() 178 ibdev_dbg(&iwdev->ibdev, "WS: LAN alloc_res for rdma qset failed.\n"); in irdma_lan_register_qset() 182 tc_node->l2_sched_node_id = qset.teid; in irdma_lan_register_qset() 183 vsi->qos[tc_node->user_pri].l2_sched_node_id = qset.teid; in irdma_lan_register_qset() 189 * irdma_lan_unregister_qset - Unregister qset wit 198 struct iidc_rdma_qset_params qset = {}; irdma_lan_unregister_qset() local [all...] |
H A D | i40iw_if.c | 113 u16 qset; in i40iw_open() local 136 qset = cdev_info->params.qos.prio_qos[i].qs_handle; in i40iw_open() 138 l2params.qs_handle_list[i] = qset; in i40iw_open() 140 last_qset = qset; in i40iw_open() 141 else if ((qset != last_qset) && (qset != IRDMA_NO_QSET)) in i40iw_open()
|
/kernel/linux/linux-6.6/include/linux/net/intel/ |
H A D | iidc.h | 76 int ice_add_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset); 77 int ice_del_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset);
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
H A D | nic_main.c | 478 u8 qset, rq_idx = 0; in nic_config_cpi() local 504 qset = cfg->vf_id; in nic_config_cpi() 508 (qset << 3) | rq_idx); in nic_config_cpi() 567 u8 qset, idx = 0; in nic_config_rss() local 579 qset = nic->vf_sqs[cfg->vf_id][svf - 1]; in nic_config_rss() 581 qset = cfg->vf_id; in nic_config_rss() 583 (qset << 3) | (cfg->ind_tbl[idx] & 0x7)); in nic_config_rss()
|
H A D | nicvf_ethtool.c | 211 static void nicvf_get_qset_strings(struct nicvf *nic, u8 **data, int qset) in nicvf_get_qset_strings() argument 214 int start_qidx = qset * MAX_RCV_QUEUES_PER_QS; in nicvf_get_qset_strings()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/thunder/ |
H A D | nic_main.c | 478 u8 qset, rq_idx = 0; in nic_config_cpi() local 504 qset = cfg->vf_id; in nic_config_cpi() 508 (qset << 3) | rq_idx); in nic_config_cpi() 567 u8 qset, idx = 0; in nic_config_rss() local 579 qset = nic->vf_sqs[cfg->vf_id][svf - 1]; in nic_config_rss() 581 qset = cfg->vf_id; in nic_config_rss() 583 (qset << 3) | (cfg->ind_tbl[idx] & 0x7)); in nic_config_rss()
|
H A D | nicvf_ethtool.c | 212 static void nicvf_get_qset_strings(struct nicvf *nic, u8 **data, int qset) in nicvf_get_qset_strings() argument 215 int start_qidx = qset * MAX_RCV_QUEUES_PER_QS; in nicvf_get_qset_strings()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_main.c | 1313 u16 qset; in i40iw_initialize_dev() local 1347 qset = ldev->params.qos.prio_qos[i].qs_handle; in i40iw_initialize_dev() 1348 l2params.qs_handle_list[i] = qset; in i40iw_initialize_dev() 1350 last_qset = qset; in i40iw_initialize_dev() 1351 else if ((qset != last_qset) && (qset != I40IW_NO_QSET)) in i40iw_initialize_dev()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | cxgb3_main.c | 671 &adap->params.sge.qset[qset_idx], ntxq, dev, in setup_sge_qsets() 1956 const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset]; in get_sge_param() 1988 q = &adapter->params.sge.qset[pi->first_qset]; in set_sge_param() 2012 qsp = &adapter->params.sge.qset[i]; in set_coalesce() 2025 struct qset_params *q = adapter->params.sge.qset; in get_coalesce() 2188 /* Allow setting of any available qset when offload enabled */ in cxgb_extension_ioctl() 2202 q = &adapter->params.sge.qset[t.qset_idx]; in cxgb_extension_ioctl() 2236 qset[i]; in cxgb_extension_ioctl() 2278 q = &adapter->params.sge.qset[q1 + t.qset_idx]; in cxgb_extension_ioctl() 2823 * Sleep a while before ringing the driver qset db in db_drop_task() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | cxgb3_main.c | 670 &adap->params.sge.qset[qset_idx], ntxq, dev, in setup_sge_qsets() 1957 const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset]; in get_sge_param() 1991 q = &adapter->params.sge.qset[pi->first_qset]; in set_sge_param() 2017 qsp = &adapter->params.sge.qset[i]; in set_coalesce() 2032 struct qset_params *q = adapter->params.sge.qset; in get_coalesce() 2188 /* Allow setting of any available qset when offload enabled */ in cxgb_siocdevprivate() 2202 q = &adapter->params.sge.qset[t.qset_idx]; in cxgb_siocdevprivate() 2236 qset[i]; in cxgb_siocdevprivate() 2278 q = &adapter->params.sge.qset[q1 + t.qset_idx]; in cxgb_siocdevprivate() 2821 * Sleep a while before ringing the driver qset db in db_drop_task() [all...] |