/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 31 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument 42 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg() 504 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() 524 qidx += ((nic->sqs_id + 1) * MAX_SND_QUEUES_PER_QS); in nicvf_init_snd_queue() 525 if (qidx < nic->pnicvf->xdp_tx_queues) { in nicvf_init_snd_queue() 627 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() 630 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue() 632 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue() 635 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue() 639 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() 503 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument 626 nicvf_reclaim_snd_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_snd_queue() argument 638 nicvf_reclaim_rcv_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_rcv_queue() argument 648 nicvf_reclaim_cmp_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_cmp_queue() argument 659 nicvf_reclaim_rbdr(struct nicvf *nic, struct rbdr *rbdr, int qidx) nicvf_reclaim_rbdr() argument 744 nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_rcv_queue_config() argument 817 nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_cmp_queue_config() argument 858 nicvf_snd_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_snd_queue_config() argument 916 nicvf_rbdr_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_rbdr_config() argument 990 int qidx; nicvf_free_resources() local 1008 int qidx; nicvf_alloc_resources() local 1068 int qidx; nicvf_config_data_transfer() local 1159 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument 1170 nicvf_sq_disable(struct nicvf *nic, int qidx) nicvf_sq_disable() argument 1179 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument [all...] |
H A D | nicvf_main.c | 75 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument 78 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx() 80 return qidx; in nicvf_netdev_qidx() 104 u64 qidx, u64 val) in nicvf_queue_reg_write() 108 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write() 111 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument 115 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read() 992 int qidx; in nicvf_handle_qs_err() local 998 for (qidx = 0; qidx < q in nicvf_handle_qs_err() 103 nicvf_queue_reg_write(struct nicvf *nic, u64 offset, u64 qidx, u64 val) nicvf_queue_reg_write() argument 1046 int qidx = cq_poll->cq_idx; nicvf_intr_handler() local 1065 u8 qidx; nicvf_rbdr_intr_handler() local 1316 int qidx; nicvf_free_cq_poll() local 1329 int irq, qidx; nicvf_stop() local 1454 int cpu, err, qidx; nicvf_open() local 1662 int qidx, cpu; nicvf_update_stats() local [all...] |
H A D | nicvf_ethtool.c | 213 int stats, qidx; in nicvf_get_qset_strings() local 216 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 218 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 224 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings() 226 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 300 int stat, qidx; in nicvf_get_qset_stats() local [all...] |
H A D | nicvf_queues.h | 336 int qidx, bool enable); 338 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx); 339 void nicvf_sq_disable(struct nicvf *nic, int qidx); 342 struct snd_queue *sq, int qidx); 365 u64 qidx, u64 val); 367 u64 offset, u64 qidx);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 32 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument 43 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg() 505 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() 525 qidx += ((nic->sqs_id + 1) * MAX_SND_QUEUES_PER_QS); in nicvf_init_snd_queue() 526 if (qidx < nic->pnicvf->xdp_tx_queues) { in nicvf_init_snd_queue() 628 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() 631 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue() 633 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue() 636 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue() 640 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() 504 nicvf_init_snd_queue(struct nicvf *nic, struct snd_queue *sq, int q_len, int qidx) nicvf_init_snd_queue() argument 627 nicvf_reclaim_snd_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_snd_queue() argument 639 nicvf_reclaim_rcv_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_rcv_queue() argument 649 nicvf_reclaim_cmp_queue(struct nicvf *nic, struct queue_set *qs, int qidx) nicvf_reclaim_cmp_queue() argument 660 nicvf_reclaim_rbdr(struct nicvf *nic, struct rbdr *rbdr, int qidx) nicvf_reclaim_rbdr() argument 745 nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_rcv_queue_config() argument 818 nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_cmp_queue_config() argument 859 nicvf_snd_queue_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_snd_queue_config() argument 917 nicvf_rbdr_config(struct nicvf *nic, struct queue_set *qs, int qidx, bool enable) nicvf_rbdr_config() argument 991 int qidx; nicvf_free_resources() local 1009 int qidx; nicvf_alloc_resources() local 1069 int qidx; nicvf_config_data_transfer() local 1160 nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx) nicvf_sq_enable() argument 1171 nicvf_sq_disable(struct nicvf *nic, int qidx) nicvf_sq_disable() argument 1180 nicvf_sq_free_used_descs(struct net_device *netdev, struct snd_queue *sq, int qidx) nicvf_sq_free_used_descs() argument [all...] |
H A D | nicvf_main.c | 75 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument 78 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx() 80 return qidx; in nicvf_netdev_qidx() 104 u64 qidx, u64 val) in nicvf_queue_reg_write() 108 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write() 111 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument 115 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read() 989 int qidx; in nicvf_handle_qs_err() local 995 for (qidx = 0; qidx < q in nicvf_handle_qs_err() 103 nicvf_queue_reg_write(struct nicvf *nic, u64 offset, u64 qidx, u64 val) nicvf_queue_reg_write() argument 1043 int qidx = cq_poll->cq_idx; nicvf_intr_handler() local 1062 u8 qidx; nicvf_rbdr_intr_handler() local 1313 int qidx; nicvf_free_cq_poll() local 1326 int irq, qidx; nicvf_stop() local 1451 int cpu, err, qidx; nicvf_open() local 1658 int qidx, cpu; nicvf_update_stats() local [all...] |
H A D | nicvf_ethtool.c | 214 int stats, qidx; in nicvf_get_qset_strings() local 217 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 219 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 225 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings() 227 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 301 int stat, qidx; in nicvf_get_qset_stats() local [all...] |
H A D | nicvf_queues.h | 336 int qidx, bool enable); 338 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx); 339 void nicvf_sq_disable(struct nicvf *nic, int qidx); 342 struct snd_queue *sq, int qidx); 365 u64 qidx, u64 val); 367 u64 offset, u64 qidx);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 33 static int otx2_qos_sq_aura_pool_init(struct otx2_nic *pfvf, int qidx) in otx2_qos_sq_aura_pool_init() argument 56 pool_id = otx2_get_pool_idx(pfvf, AURA_NIX_SQ, qidx); in otx2_qos_sq_aura_pool_init() 76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init() 116 static void otx2_qos_sq_free_sqbs(struct otx2_nic *pfvf, int qidx) in otx2_qos_sq_free_sqbs() argument 124 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 140 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 151 static void otx2_qos_sqb_flush(struct otx2_nic *pfvf, int qidx) in otx2_qos_sqb_flush() argument 157 incr = (u64)qidx << 32; in otx2_qos_sqb_flush() 165 static int otx2_qos_ctx_disable(struct otx2_nic *pfvf, u16 qidx, int aura_id) in otx2_qos_ctx_disable() argument 176 cn10k_sq_aq->qidx in otx2_qos_ctx_disable() 222 int qidx; otx2_qos_get_qid() local 230 otx2_qos_free_qid(struct otx2_nic *pfvf, int qidx) otx2_qos_free_qid() argument 235 otx2_qos_enable_sq(struct otx2_nic *pfvf, int qidx) otx2_qos_enable_sq() argument 259 otx2_qos_disable_sq(struct otx2_nic *pfvf, int qidx) otx2_qos_disable_sq() argument [all...] |
H A D | otx2_common.c | 20 struct otx2_nic *pfvf, int qidx) in otx2_nix_rq_op_stats() 22 u64 incr = (u64)qidx << 32; in otx2_nix_rq_op_stats() 33 struct otx2_nic *pfvf, int qidx) in otx2_nix_sq_op_stats() 35 u64 incr = (u64)qidx << 32; in otx2_nix_sq_op_stats() 76 int otx2_update_rq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_rq_stats() argument 78 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats() 83 otx2_nix_rq_op_stats(&rq->stats, pfvf, qidx); in otx2_update_rq_stats() 87 int otx2_update_sq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_sq_stats() argument 89 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats() 94 if (qidx > in otx2_update_sq_stats() 19 otx2_nix_rq_op_stats(struct queue_stats *stats, struct otx2_nic *pfvf, int qidx) otx2_nix_rq_op_stats() argument 32 otx2_nix_sq_op_stats(struct queue_stats *stats, struct otx2_nic *pfvf, int qidx) otx2_nix_sq_op_stats() argument 509 otx2_config_irq_coalescing(struct otx2_nic *pfvf, int qidx) otx2_config_irq_coalescing() argument 818 int qidx, sqe_tail, sqe_head; otx2_sqb_flush() local 853 otx2_rq_init(struct otx2_nic *pfvf, u16 qidx, u16 lpb_aura) otx2_rq_init() argument 886 otx2_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura) otx2_sq_aq_init() argument 923 otx2_sq_init(struct otx2_nic *pfvf, u16 qidx, u16 sqb_aura) otx2_sq_init() argument 985 otx2_cq_init(struct otx2_nic *pfvf, u16 qidx) otx2_cq_init() argument 1080 int qidx; otx2_pool_refill_task() local 1096 int qidx, err; otx2_config_nix_queues() local 1190 int sqb, qidx; otx2_sq_free_sqbs() local 1428 int qidx, pool_id, stack_pages, num_sqbs; otx2_sq_aura_pool_init() local 1773 int irq, qidx; otx2_free_cints() local [all...] |
H A D | cn10k.c | 75 int cn10k_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura) in cn10k_sq_aq_init() argument 85 aq->sq.cq = pfvf->hw.rx_queues + qidx; in cn10k_sq_aq_init() 89 aq->sq.smq = otx2_get_smq_idx(pfvf, qidx); in cn10k_sq_aq_init() 102 aq->qidx = qidx; in cn10k_sq_aq_init() 138 void cn10k_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx) in cn10k_sqe_flush() argument 321 aq->qidx = rq_idx; in cn10k_map_unmap_rq_policer() 345 int qidx, rc; in cn10k_free_matchall_ipolicer() local 350 for (qidx = 0; qidx < h in cn10k_free_matchall_ipolicer() 468 int qidx, rc; cn10k_set_matchall_ipolicer_rate() local [all...] |
H A D | otx2_pf.c | 1258 u64 qidx = 0; in otx2_q_intr_handler() local 1261 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1263 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler() 1265 otx2_write64(pf, NIX_LF_CQ_OP_INT, (qidx << 44) | in otx2_q_intr_handler() 1273 qidx, otx2_read64(pf, NIX_LF_ERR_INT)); in otx2_q_intr_handler() 1277 qidx); in otx2_q_intr_handler() 1281 qidx); in otx2_q_intr_handler() 1288 for (qidx in otx2_q_intr_handler() 1373 int qidx = cq_poll->cint_idx; otx2_cq_intr_handler() local 1393 int qidx; otx2_disable_napi() local 1407 int qidx; otx2_free_cq_res() local 1421 int qidx; otx2_free_sq_res() local 1601 int qidx; otx2_free_hw_resources() local 1765 int err = 0, qidx, vec; otx2_open() local 1975 int qidx, vec, wrk; otx2_stop() local 2043 int qidx = skb_get_queue_mapping(skb); otx2_xmit() local 2585 otx2_xdp_xmit_tx(struct otx2_nic *pf, struct xdp_frame *xdpf, int qidx) otx2_xdp_xmit_tx() argument 2612 int qidx = smp_processor_id(); otx2_xdp_xmit() local [all...] |
H A D | otx2_txrx.c | 196 int qidx) in otx2_skb_add_frag() 227 pfvf->hw_ops->aura_freeptr(pfvf, qidx, iova & ~0x07ULL); in otx2_skb_add_frag() 255 int qidx) in otx2_free_rcv_seg() 268 pfvf->hw_ops->aura_freeptr(pfvf, qidx, in otx2_free_rcv_seg() 275 struct nix_cqe_rx_s *cqe, int qidx) in otx2_check_rcv_errors() 283 qidx, parse->errlev, parse->errcode); in otx2_check_rcv_errors() 334 otx2_free_rcv_seg(pfvf, cqe, qidx); in otx2_check_rcv_errors() 450 int tx_pkts = 0, tx_bytes = 0, qidx; in otx2_tx_napi_handler() local 462 qidx = cq->cq_idx - pfvf->hw.rx_queues; in otx2_tx_napi_handler() 463 sq = &pfvf->qset.sq[qidx]; in otx2_tx_napi_handler() 194 otx2_skb_add_frag(struct otx2_nic *pfvf, struct sk_buff *skb, u64 iova, int len, struct nix_rx_parse_s *parse, int qidx) otx2_skb_add_frag() argument 254 otx2_free_rcv_seg(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, int qidx) otx2_free_rcv_seg() argument 274 otx2_check_rcv_errors(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, int qidx) otx2_check_rcv_errors() argument 595 otx2_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx) otx2_sqe_flush() argument 750 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) otx2_sqe_add_hdr() argument 880 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_tso() argument 1141 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_skb() argument 1210 otx2_cleanup_rx_cqes(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, int qidx) otx2_cleanup_rx_cqes() argument 1257 int qidx; otx2_cleanup_tx_cqes() local 1367 otx2_xdp_sq_append_pkt(struct otx2_nic *pfvf, u64 iova, int len, u16 qidx) otx2_xdp_sq_append_pkt() argument 1407 int qidx = cq->cq_idx; otx2_xdp_rcv_pkt_handler() local [all...] |
H A D | otx2_ethtool.c | 86 int qidx, stats; in otx2_get_qset_strings() local 88 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) { in otx2_get_qset_strings() 90 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in otx2_get_qset_strings() 96 for (qidx = 0; qidx < otx2_get_total_tx_queues(pfvf); qidx++) { in otx2_get_qset_strings() 98 if (qidx >= pfvf->hw.non_qos_queues) in otx2_get_qset_strings() 100 qidx in otx2_get_qset_strings() 153 int stat, qidx; otx2_get_qset_stats() local 492 int qidx; otx2_set_coalesce() local [all...] |
H A D | otx2_dcbnl.c | 159 cn10k_sq_aq->qidx = prio; in otx2_pfc_update_sq_smq_mapping() 174 sq_aq->qidx = prio; in otx2_pfc_update_sq_smq_mapping() 325 void otx2_update_bpid_in_rqctx(struct otx2_nic *pfvf, int vlan_prio, int qidx, in otx2_update_bpid_in_rqctx() argument 333 if (pfvf->queue_to_pfc_map[qidx] && pfc_enable) { in otx2_update_bpid_in_rqctx() 336 pfvf->queue_to_pfc_map[qidx], qidx); in otx2_update_bpid_in_rqctx() 345 pfvf->queue_to_pfc_map[qidx] = vlan_prio; in otx2_update_bpid_in_rqctx() 357 aq->qidx = qidx; in otx2_update_bpid_in_rqctx() 372 npa_aq->aura_id = qidx; in otx2_update_bpid_in_rqctx() [all...] |
H A D | qos.h | 27 void otx2_qos_free_qid(struct otx2_nic *pfvf, int qidx); 28 int otx2_qos_enable_sq(struct otx2_nic *pfvf, int qidx); 29 void otx2_qos_disable_sq(struct otx2_nic *pfvf, int qidx);
|
H A D | otx2_common.h | 371 int (*sq_aq_init)(void *dev, u16 qidx, u16 sqb_aura); 373 int size, int qidx); 913 static inline u16 otx2_get_smq_idx(struct otx2_nic *pfvf, u16 qidx) in otx2_get_smq_idx() argument 917 if (qidx < NIX_PF_PFC_PRIO_MAX && pfvf->pfc_alloc_status[qidx]) in otx2_get_smq_idx() 918 return pfvf->pfc_schq_list[NIX_TXSCH_LVL_SMQ][qidx]; in otx2_get_smq_idx() 920 /* check if qidx falls under QOS queues */ in otx2_get_smq_idx() 921 if (qidx >= pfvf->hw.non_qos_queues) in otx2_get_smq_idx() 922 smq = pfvf->qos.qid_to_sqmap[qidx - pfvf->hw.non_qos_queues]; in otx2_get_smq_idx() 961 void otx2_config_irq_coalescing(struct otx2_nic *pfvf, int qidx); [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_trace.h | 23 __field(u32, qidx) 31 __entry->qidx = txq->qidx; 39 __get_str(devname), __entry->qidx, __entry->sqe_idx, 53 __field(u32, qidx) 61 __entry->qidx = txq->qidx; 69 __get_str(devname), __entry->qidx, __entry->sqe_idx, 84 __field(u32, qidx) 94 __entry->qidx [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 20 struct otx2_nic *pfvf, int qidx) in otx2_nix_rq_op_stats() 22 u64 incr = (u64)qidx << 32; in otx2_nix_rq_op_stats() 33 struct otx2_nic *pfvf, int qidx) in otx2_nix_sq_op_stats() 35 u64 incr = (u64)qidx << 32; in otx2_nix_sq_op_stats() 63 int otx2_update_rq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_rq_stats() argument 65 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats() 70 otx2_nix_rq_op_stats(&rq->stats, pfvf, qidx); in otx2_update_rq_stats() 74 int otx2_update_sq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_sq_stats() argument 76 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats() 81 otx2_nix_sq_op_stats(&sq->stats, pfvf, qidx); in otx2_update_sq_stats() 19 otx2_nix_rq_op_stats(struct queue_stats *stats, struct otx2_nic *pfvf, int qidx) otx2_nix_rq_op_stats() argument 32 otx2_nix_sq_op_stats(struct queue_stats *stats, struct otx2_nic *pfvf, int qidx) otx2_nix_sq_op_stats() argument 457 otx2_config_irq_coalescing(struct otx2_nic *pfvf, int qidx) otx2_config_irq_coalescing() argument 666 int qidx, sqe_tail, sqe_head; otx2_sqb_flush() local 704 otx2_rq_init(struct otx2_nic *pfvf, u16 qidx, u16 lpb_aura) otx2_rq_init() argument 737 otx2_sq_init(struct otx2_nic *pfvf, u16 qidx, u16 sqb_aura) otx2_sq_init() argument 814 otx2_cq_init(struct otx2_nic *pfvf, u16 qidx) otx2_cq_init() argument 888 int qidx, free_ptrs = 0; otx2_pool_refill_task() local 924 int qidx, err; otx2_config_nix_queues() local 1015 int sqb, qidx; otx2_sq_free_sqbs() local 1210 int qidx, pool_id, stack_pages, num_sqbs; otx2_sq_aura_pool_init() local 1549 int irq, qidx; otx2_free_cints() local [all...] |
H A D | otx2_pf.c | 1147 u64 qidx = 0; in otx2_q_intr_handler() local 1150 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1152 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler() 1154 otx2_write64(pf, NIX_LF_CQ_OP_INT, (qidx << 44) | in otx2_q_intr_handler() 1161 qidx, otx2_read64(pf, NIX_LF_ERR_INT)); in otx2_q_intr_handler() 1165 qidx); in otx2_q_intr_handler() 1168 qidx); in otx2_q_intr_handler() 1175 for (qidx in otx2_q_intr_handler() 1226 int qidx = cq_poll->cint_idx; otx2_cq_intr_handler() local 1245 int qidx; otx2_disable_napi() local 1258 int qidx; otx2_free_cq_res() local 1272 int qidx; otx2_free_sq_res() local 1394 int qidx, err; otx2_free_hw_resources() local 1458 int err = 0, qidx, vec; otx2_open() local 1639 int qidx, vec, wrk; otx2_stop() local 1706 int qidx = skb_get_queue_mapping(skb); otx2_xmit() local [all...] |
H A D | otx2_ethtool.c | 81 int qidx, stats; in otx2_get_qset_strings() local 83 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) { in otx2_get_qset_strings() 85 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in otx2_get_qset_strings() 90 for (qidx = 0; qidx < pfvf->hw.tx_queues; qidx++) { in otx2_get_qset_strings() 92 sprintf(*data, "txq%d: %s", qidx + start_qidx, in otx2_get_qset_strings() 136 int stat, qidx; in otx2_get_qset_stats() local 372 int qidx; otx2_set_coalesce() local [all...] |
H A D | otx2_txrx.c | 189 int qidx) in otx2_free_rcv_seg() 202 otx2_aura_freeptr(pfvf, qidx, *seg_addr & ~0x07ULL); in otx2_free_rcv_seg() 208 struct nix_cqe_rx_s *cqe, int qidx) in otx2_check_rcv_errors() 216 qidx, parse->errlev, parse->errcode); in otx2_check_rcv_errors() 268 otx2_free_rcv_seg(pfvf, cqe, qidx); in otx2_check_rcv_errors() 577 struct sk_buff *skb, u16 qidx) in otx2_sqe_add_hdr() 590 sqe_hdr->sq = qidx; in otx2_sqe_add_hdr() 705 struct sk_buff *skb, u16 qidx) in otx2_sq_append_tso() 707 struct netdev_queue *txq = netdev_get_tx_queue(pfvf->netdev, qidx); in otx2_sq_append_tso() 736 otx2_sqe_add_hdr(pfvf, sq, sqe_hdr, skb, qidx); in otx2_sq_append_tso() 188 otx2_free_rcv_seg(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, int qidx) otx2_free_rcv_seg() argument 207 otx2_check_rcv_errors(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, int qidx) otx2_check_rcv_errors() argument 575 otx2_sqe_add_hdr(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct nix_sqe_hdr_s *sqe_hdr, struct sk_buff *skb, u16 qidx) otx2_sqe_add_hdr() argument 704 otx2_sq_append_tso(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_tso() argument 840 otx2_sq_append_skb(struct net_device *netdev, struct otx2_snd_queue *sq, struct sk_buff *skb, u16 qidx) otx2_sq_append_skb() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
H A D | chcr_ipsec.c | 420 u32 qidx; in copy_esn_pktxt() local 428 qidx = skb->queue_mapping; in copy_esn_pktxt() 429 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_esn_pktxt() 470 u32 ctrl0, qidx; in copy_cpltx_pktxt() local 476 qidx = skb->queue_mapping; in copy_cpltx_pktxt() 477 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_cpltx_pktxt() 515 unsigned int qidx; in copy_key_cpltx_pktxt() local 519 qidx = skb->queue_mapping; in copy_key_cpltx_pktxt() 520 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_key_cpltx_pktxt() 575 int qidx in ch_ipsec_crypto_wreq() local 713 int qidx, left, credits; ch_ipsec_xmit() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
H A D | chcr_ipsec.c | 424 u32 qidx; in copy_esn_pktxt() local 432 qidx = skb->queue_mapping; in copy_esn_pktxt() 433 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_esn_pktxt() 474 u32 ctrl0, qidx; in copy_cpltx_pktxt() local 480 qidx = skb->queue_mapping; in copy_cpltx_pktxt() 481 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_cpltx_pktxt() 519 unsigned int qidx; in copy_key_cpltx_pktxt() local 523 qidx = skb->queue_mapping; in copy_key_cpltx_pktxt() 524 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_key_cpltx_pktxt() 579 int qidx in ch_ipsec_crypto_wreq() local 717 int qidx, left, credits; ch_ipsec_xmit() local [all...] |
/kernel/linux/linux-6.6/drivers/dma/ptdma/ |
H A D | ptdma-dev.c | 72 u8 *q_desc = (u8 *)&cmd_q->qbase[cmd_q->qidx]; in pt_core_execute_cmd() 84 cmd_q->qidx = (cmd_q->qidx + 1) % CMD_Q_LEN; in pt_core_execute_cmd() 90 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_core_execute_cmd() 136 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_do_cmd_complete() 216 cmd_q->qidx = 0; in pt_core_init()
|