Lines Matching defs:sge
43 struct ib_sge sge[NVMET_RDMA_MAX_INLINE_SGE + 1];
260 struct ib_sge *sge;
267 sge = &c->sge[1];
269 for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
270 if (sge->length)
271 ib_dma_unmap_page(ndev->device, sge->addr,
272 sge->length, DMA_FROM_DEVICE);
282 struct ib_sge *sge;
292 sge = &c->sge[1];
295 for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
300 sge->addr = ib_dma_map_page(ndev->device,
302 if (ib_dma_mapping_error(ndev->device, sge->addr))
304 sge->length = min_t(int, len, PAGE_SIZE);
305 sge->lkey = ndev->pd->local_dma_lkey;
306 len -= sge->length;
311 for (; i >= 0; i--, sg--, sge--) {
312 if (sge->length)
313 ib_dma_unmap_page(ndev->device, sge->addr,
314 sge->length, DMA_FROM_DEVICE);
329 c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd,
331 if (ib_dma_mapping_error(ndev->device, c->sge[0].addr))
334 c->sge[0].length = sizeof(*c->nvme_cmd);
335 c->sge[0].lkey = ndev->pd->local_dma_lkey;
343 c->wr.sg_list = c->sge;
349 ib_dma_unmap_single(ndev->device, c->sge[0].addr,
363 ib_dma_unmap_single(ndev->device, c->sge[0].addr,
507 cmd->sge[0].addr, cmd->sge[0].length,
982 cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length,