Home
last modified time | relevance | path

Searched refs:sges (Results 1 - 22 of 22) sorted by relevance

/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c124 * tx_map_skb - dma mapping for skb and return sges
127 * @sges: returned sges
132 struct hinic_sge *sges) in tx_map_skb()
148 hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb)); in tx_map_skb()
161 hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag)); in tx_map_skb()
168 dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]), in tx_map_skb()
169 sges[j + 1].len, DMA_TO_DEVICE); in tx_map_skb()
171 dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[ in tx_map_skb()
131 tx_map_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_map_skb() argument
182 tx_unmap_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_unmap_skb() argument
651 tx_free_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_free_skb() argument
[all...]
H A Dhinic_tx.h36 struct hinic_sge *sges; member
H A Dhinic_hw_qp.c594 * @sges: sges for use by the wqe for send for buf addresses
595 * @nr_sges: number of sges
598 struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, in hinic_sq_prepare_wqe()
608 sq_wqe->buf_descs[i].sge = sges[i]; in hinic_sq_prepare_wqe()
776 * hinic_sq_get_sges - get sges from the wqe
777 * @sq_wqe: wqe to get the sges from its buffer addresses
778 * @sges: returned sges
779 * @nr_sges: number sges t
597 hinic_sq_prepare_wqe(struct hinic_sq *sq, u16 prod_idx, struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_prepare_wqe() argument
781 hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_get_sges() argument
[all...]
H A Dhinic_hw_qp.h179 struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
204 void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c125 * tx_map_skb - dma mapping for skb and return sges
128 * @sges: returned sges
133 struct hinic_sge *sges) in tx_map_skb()
149 hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb)); in tx_map_skb()
162 hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag)); in tx_map_skb()
169 dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]), in tx_map_skb()
170 sges[j + 1].len, DMA_TO_DEVICE); in tx_map_skb()
172 dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[ in tx_map_skb()
132 tx_map_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_map_skb() argument
183 tx_unmap_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_unmap_skb() argument
653 tx_free_skb(struct hinic_dev *nic_dev, struct sk_buff *skb, struct hinic_sge *sges) tx_free_skb() argument
[all...]
H A Dhinic_tx.h36 struct hinic_sge *sges; member
H A Dhinic_hw_qp.c591 * @sges: sges for use by the wqe for send for buf addresses
592 * @nr_sges: number of sges
595 struct hinic_sge *sges, int nr_sges) in hinic_sq_prepare_wqe()
604 sq_wqe->buf_descs[i].sge = sges[i]; in hinic_sq_prepare_wqe()
772 * hinic_sq_get_sges - get sges from the wqe
773 * @sq_wqe: wqe to get the sges from its buffer addresses
774 * @sges: returned sges
775 * @nr_sges: number sges t
594 hinic_sq_prepare_wqe(struct hinic_sq *sq, struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_prepare_wqe() argument
777 hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, int nr_sges) hinic_sq_get_sges() argument
[all...]
H A Dhinic_hw_qp.h179 struct hinic_sge *sges, int nr_sges);
203 void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
/kernel/linux/linux-5.10/include/rdma/
H A Drw.h30 struct ib_sge *sges; member
/kernel/linux/linux-6.6/include/rdma/
H A Drw.h30 struct ib_sge *sges; member
/kernel/linux/linux-5.10/drivers/infiniband/core/
H A Drw.c205 ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); in rdma_rw_init_map_wrs()
206 if (!ctx->map.sges) in rdma_rw_init_map_wrs()
245 kfree(ctx->map.sges); in rdma_rw_init_map_wrs()
597 kfree(ctx->map.sges); in rdma_rw_ctx_destroy()
/kernel/linux/linux-6.6/drivers/infiniband/core/
H A Drw.c206 ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); in rdma_rw_init_map_wrs()
207 if (!ctx->map.sges) in rdma_rw_init_map_wrs()
246 kfree(ctx->map.sges); in rdma_rw_init_map_wrs()
588 kfree(ctx->map.sges); in rdma_rw_ctx_destroy()
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/
H A Dsvc_rdma_rw.c53 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument
73 if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges, in svc_rdma_get_rw_ctxt()
82 trace_svcrdma_no_rwctx_err(rdma, sges); in svc_rdma_get_rw_ctxt()
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/
H A Dsvc_rdma_rw.c54 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument
74 if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges, in svc_rdma_get_rw_ctxt()
83 trace_svcrdma_no_rwctx_err(rdma, sges); in svc_rdma_get_rw_ctxt()
/kernel/linux/linux-5.10/drivers/scsi/bfa/
H A Dbfi_ms.h677 struct bfi_sge_s sges[BFI_SGE_INLINE_MAX]; member
H A Dbfa_svc.h331 struct bfa_sge_s sges[BFI_SGE_INLINE_MAX]; member
H A Dbfi.h143 struct bfi_sge_s sges[BFI_SGPG_SGES_MAX]; member
H A Dbfa_fcpim.c2476 sge = &m->sges[0]; in bfa_ioim_send_ioreq()
2494 sgpge = sgpg->sgpg->sges; in bfa_ioim_send_ioreq()
/kernel/linux/linux-6.6/drivers/scsi/bfa/
H A Dbfi_ms.h677 struct bfi_sge_s sges[BFI_SGE_INLINE_MAX]; member
H A Dbfa_svc.h331 struct bfa_sge_s sges[BFI_SGE_INLINE_MAX]; member
H A Dbfi.h143 struct bfi_sge_s sges[BFI_SGPG_SGES_MAX]; member
H A Dbfa_fcpim.c2476 sge = &m->sges[0]; in bfa_ioim_send_ioreq()
2494 sgpge = sgpg->sgpg->sges; in bfa_ioim_send_ioreq()

Completed in 31 milliseconds