/kernel/linux/linux-6.6/drivers/ufs/core/ |
H A D | ufs_bsg.c | 57 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer() 73 struct scatterlist *sg_list = NULL; in ufs_bsg_exec_advanced_rpmb_req() local 111 sg_cnt = dma_map_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() 114 sg_list = payload->sg_list; in ufs_bsg_exec_advanced_rpmb_req() 120 &rpmb_reply->ehs_rsp, sg_cnt, sg_list, dir); in ufs_bsg_exec_advanced_rpmb_req() 123 dma_unmap_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() 166 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
|
/kernel/linux/linux-5.10/drivers/virt/ |
H A D | fsl_hypervisor.c | 153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local 236 * sg_list is the list of fh_sg_list objects that we pass to the in ioctl_memcpy() 246 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy() 263 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy() 266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy() 267 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 269 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy() 271 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy() 272 count = param.count - sg_list[ in ioctl_memcpy() [all...] |
/kernel/linux/linux-6.6/drivers/virt/ |
H A D | fsl_hypervisor.c | 153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local 236 * sg_list is the list of fh_sg_list objects that we pass to the in ioctl_memcpy() 246 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy() 263 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy() 266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy() 267 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 269 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy() 271 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy() 272 count = param.count - sg_list[ in ioctl_memcpy() [all...] |
/kernel/linux/linux-5.10/drivers/scsi/qla2xxx/ |
H A D | qla_bsg.c | 47 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 52 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 55 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 58 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 193 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 220 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 351 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 354 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 360 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() 363 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/ |
H A D | erdma_qp.c | 221 bytes += send_wr->sg_list[i].length; in fill_inline_data() 225 remain_size = send_wr->sg_list[i].length; in fill_inline_data() 232 (void *)(uintptr_t)send_wr->sg_list[i].addr + in fill_inline_data() 273 bytes += send_wr->sg_list[i].length; in fill_sgl() 274 memcpy(sgl + sgl_offset, &send_wr->sg_list[i], in fill_sgl() 358 read_sqe->length = cpu_to_le32(send_wr->sg_list[0].length); in erdma_push_one_sqe() 359 read_sqe->sink_stag = cpu_to_le32(send_wr->sg_list[0].lkey); in erdma_push_one_sqe() 361 cpu_to_le32(lower_32_bits(send_wr->sg_list[0].addr)); in erdma_push_one_sqe() 363 cpu_to_le32(upper_32_bits(send_wr->sg_list[0].addr)); in erdma_push_one_sqe() 369 sge->length = cpu_to_le32(send_wr->sg_list[ in erdma_push_one_sqe() [all...] |
/kernel/linux/linux-6.6/drivers/scsi/qla2xxx/ |
H A D | qla_bsg.c | 53 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 58 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 68 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 71 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 207 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 234 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 370 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 373 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 379 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() 382 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() [all...] |
/kernel/linux/linux-5.10/drivers/tee/amdtee/ |
H A D | call.c | 315 cmd->sg_list.count = count; in handle_map_shmem() 320 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem() 321 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem() 322 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem() 323 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem() 326 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem() 328 cmd->sg_list.buf[i].low_addr); in handle_map_shmem() 329 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem() 330 pr_debug("list size = 0x%x\n", cmd->sg_list in handle_map_shmem() [all...] |
/kernel/linux/linux-6.6/drivers/tee/amdtee/ |
H A D | call.c | 315 cmd->sg_list.count = count; in handle_map_shmem() 320 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem() 321 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem() 322 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem() 323 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem() 326 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem() 328 cmd->sg_list.buf[i].low_addr); in handle_map_shmem() 329 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem() 330 pr_debug("list size = 0x%x\n", cmd->sg_list in handle_map_shmem() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
H A D | rc.c | 206 ss->sge = wqe->sg_list[0]; in rvt_restart_sge() 207 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
|
H A D | qp.c | 666 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey() 1116 sz = struct_size(swq, sg_list, init_attr->cap.max_send_sge); in rvt_create_qp() 1881 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv() 1882 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv() 1883 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv() 1930 wr->sg_list[0].length < sizeof(u64) || in rvt_qp_valid_operation() 1931 wr->sg_list[ in rvt_qp_valid_operation() [all...] |
/kernel/linux/linux-5.10/include/rdma/ |
H A D | rdmavt_mr.h | 77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member 98 ss->sge = *ss->sg_list++; in rvt_put_ss() 126 *sge = *ss->sg_list++; in rvt_update_sge()
|
/kernel/linux/linux-6.6/include/rdma/ |
H A D | rdmavt_mr.h | 77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member 98 ss->sge = *ss->sg_list++; in rvt_put_ss() 126 *sge = *ss->sg_list++; in rvt_update_sge()
|
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/ |
H A D | rc.c | 164 ss->sge = wqe->sg_list[0]; in rvt_restart_sge() 165 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
|
H A D | qp.c | 626 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey() 1078 sz = struct_size(swq, sg_list, init_attr->cap.max_send_sge); in rvt_create_qp() 1829 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv() 1830 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv() 1831 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv() 1878 wr->sg_list[0].length < sizeof(u64) || in rvt_qp_valid_operation() 1879 wr->sg_list[ in rvt_qp_valid_operation() [all...] |
/kernel/linux/linux-5.10/block/ |
H A D | bsg-lib.c | 155 kfree(job->request_payload.sg_list); in bsg_teardown_job() 156 kfree(job->reply_payload.sg_list); in bsg_teardown_job() 210 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer() 211 if (!buf->sg_list) in bsg_map_buffer() 213 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer() 214 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer() 248 kfree(job->request_payload.sg_list); in bsg_prepare_job()
|
/kernel/linux/linux-6.6/block/ |
H A D | bsg-lib.c | 163 kfree(job->request_payload.sg_list); in bsg_teardown_job() 164 kfree(job->reply_payload.sg_list); in bsg_teardown_job() 218 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer() 219 if (!buf->sg_list) in bsg_map_buffer() 221 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer() 222 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer() 256 kfree(job->request_payload.sg_list); in bsg_prepare_job()
|
/kernel/linux/linux-5.10/drivers/dma/ |
H A D | imx-dma.c | 159 struct scatterlist *sg_list; member 810 kfree(imxdmac->sg_list); in imxdma_free_chan_resources() 811 imxdmac->sg_list = NULL; in imxdma_free_chan_resources() 887 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic() 889 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic() 891 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic() 894 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic() 897 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic() 898 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic() 899 imxdmac->sg_list[ in imxdma_prep_dma_cyclic() [all...] |
/kernel/linux/linux-6.6/drivers/dma/ |
H A D | imx-dma.c | 159 struct scatterlist *sg_list; member 790 kfree(imxdmac->sg_list); in imxdma_free_chan_resources() 791 imxdmac->sg_list = NULL; in imxdma_free_chan_resources() 867 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic() 869 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic() 871 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic() 874 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic() 877 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic() 878 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic() 879 imxdmac->sg_list[ in imxdma_prep_dma_cyclic() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_ctrl_net.c | 47 msg->sg_list[0].msg = buf; in init_send_req() 48 msg->sg_list[0].sz = msg->hdr.s.sz; in init_send_req() 287 msg->sg_list[0].msg, in process_mbox_resp() 303 req = (struct octep_ctrl_net_f2h_req *)msg->sg_list[0].msg; in process_mbox_notify() 340 msg.sg_list[0].sz = msg_sz; in octep_ctrl_net_recv_fw_messages() 341 msg.sg_list[0].msg = &data; in octep_ctrl_net_recv_fw_messages()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback() 172 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 184 *sge = *ssge.sg_list++; in qib_ud_loopback() 317 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req() 318 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback() 172 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 184 *sge = *ssge.sg_list++; in qib_ud_loopback() 318 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req() 319 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/qedr/ |
H A D | qedr_roce_cm.c | 110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet() 409 send_size += swr->sg_list[i].length; in qedr_gsi_build_header() 534 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet() 535 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet() 642 wr->sg_list[0].addr, in qedr_gsi_post_recv() 643 wr->sg_list[0].length, in qedr_gsi_post_recv() 655 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv() 693 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/qedr/ |
H A D | qedr_roce_cm.c | 110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet() 407 send_size += swr->sg_list[i].length; in qedr_gsi_build_header() 532 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet() 533 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet() 640 wr->sg_list[0].addr, in qedr_gsi_post_recv() 641 wr->sg_list[0].length, in qedr_gsi_post_recv() 653 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv() 691 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
|
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | uverbs_std_types_mr.c | 53 struct ib_sge *sg_list; in UVERBS_METHOD_ADVISE_MR() local 76 sg_list = uverbs_attr_get_alloced_ptr(attrs, in UVERBS_METHOD_ADVISE_MR() 78 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_METHOD_ADVISE_MR()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.h | 72 struct sg_list { struct 93 struct sg_list *sg;
|