/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 124 * tx_map_skb - dma mapping for skb and return sges 127 * @sges: returned sges 132 struct hinic_sge *sges) in tx_map_skb() 148 hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb)); in tx_map_skb() 161 hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag)); in tx_map_skb() 168 dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]), in tx_map_skb() 169 sges[j + 1].len, DMA_TO_DEVICE); in tx_map_skb() 171 dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[ in tx_map_skb() 131 tx_map_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_map_skb() argument 182 tx_unmap_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_unmap_skb() argument 651 tx_free_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_free_skb() argument [all...] |
H A D | hinic_tx.h | 36 struct hinic_sge *sges; member
|
H A D | hinic_hw_qp.c | 594 * @sges: sges for use by the wqe for send for buf addresses 595 * @nr_sges: number of sges 598 struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, in hinic_sq_prepare_wqe() 608 sq_wqe->buf_descs[i].sge = sges[i]; in hinic_sq_prepare_wqe() 776 * hinic_sq_get_sges - get sges from the wqe 777 * @sq_wqe: wqe to get the sges from its buffer addresses 778 * @sges: returned sges 779 * @nr_sges: number sges t 597 hinic_sq_prepare_wqe(struct hinic_sq *sq, u16 prod_idx, struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_prepare_wqe() argument 781 hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_get_sges() argument [all...] |
H A D | hinic_hw_qp.h | 179 struct hinic_sq_wqe *wqe, struct hinic_sge *sges, 204 void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 125 * tx_map_skb - dma mapping for skb and return sges 128 * @sges: returned sges 133 struct hinic_sge *sges) in tx_map_skb() 149 hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb)); in tx_map_skb() 162 hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag)); in tx_map_skb() 169 dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]), in tx_map_skb() 170 sges[j + 1].len, DMA_TO_DEVICE); in tx_map_skb() 172 dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[ in tx_map_skb() 132 tx_map_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_map_skb() argument 183 tx_unmap_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_unmap_skb() argument 653 tx_free_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_free_skb() argument [all...] |
H A D | hinic_tx.h | 36 struct hinic_sge *sges; member
|
H A D | hinic_hw_qp.c | 591 * @sges: sges for use by the wqe for send for buf addresses 592 * @nr_sges: number of sges 595 struct hinic_sge *sges, int nr_sges) in hinic_sq_prepare_wqe() 604 sq_wqe->buf_descs[i].sge = sges[i]; in hinic_sq_prepare_wqe() 772 * hinic_sq_get_sges - get sges from the wqe 773 * @sq_wqe: wqe to get the sges from its buffer addresses 774 * @sges: returned sges 775 * @nr_sges: number sges t 594 hinic_sq_prepare_wqe(struct hinic_sq *sq, struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_prepare_wqe() argument 777 hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_get_sges() argument [all...] |
H A D | hinic_hw_qp.h | 179 struct hinic_sge *sges, int nr_sges); 203 void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
|
/kernel/linux/linux-5.10/include/rdma/ |
H A D | rw.h | 30 struct ib_sge *sges; member
|
/kernel/linux/linux-6.6/include/rdma/ |
H A D | rw.h | 30 struct ib_sge *sges; member
|
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | rw.c | 205 ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); in rdma_rw_init_map_wrs() 206 if (!ctx->map.sges) in rdma_rw_init_map_wrs() 245 kfree(ctx->map.sges); in rdma_rw_init_map_wrs() 597 kfree(ctx->map.sges); in rdma_rw_ctx_destroy()
|
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | rw.c | 206 ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); in rdma_rw_init_map_wrs() 207 if (!ctx->map.sges) in rdma_rw_init_map_wrs() 246 kfree(ctx->map.sges); in rdma_rw_init_map_wrs() 588 kfree(ctx->map.sges); in rdma_rw_ctx_destroy()
|
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_rw.c | 53 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument 73 if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges, in svc_rdma_get_rw_ctxt() 82 trace_svcrdma_no_rwctx_err(rdma, sges); in svc_rdma_get_rw_ctxt()
|
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_rw.c | 54 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument 74 if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges, in svc_rdma_get_rw_ctxt() 83 trace_svcrdma_no_rwctx_err(rdma, sges); in svc_rdma_get_rw_ctxt()
|
/kernel/linux/linux-5.10/drivers/scsi/bfa/ |
H A D | bfi_ms.h | 677 struct bfi_sge_s sges[BFI_SGE_INLINE_MAX]; member
|
H A D | bfa_svc.h | 331 struct bfa_sge_s sges[BFI_SGE_INLINE_MAX]; member
|
H A D | bfi.h | 143 struct bfi_sge_s sges[BFI_SGPG_SGES_MAX]; member
|
H A D | bfa_fcpim.c | 2476 sge = &m->sges[0]; in bfa_ioim_send_ioreq() 2494 sgpge = sgpg->sgpg->sges; in bfa_ioim_send_ioreq()
|
/kernel/linux/linux-6.6/drivers/scsi/bfa/ |
H A D | bfi_ms.h | 677 struct bfi_sge_s sges[BFI_SGE_INLINE_MAX]; member
|
H A D | bfa_svc.h | 331 struct bfa_sge_s sges[BFI_SGE_INLINE_MAX]; member
|
H A D | bfi.h | 143 struct bfi_sge_s sges[BFI_SGPG_SGES_MAX]; member
|
H A D | bfa_fcpim.c | 2476 sge = &m->sges[0]; in bfa_ioim_send_ioreq() 2494 sgpge = sgpg->sgpg->sges; in bfa_ioim_send_ioreq()
|