/kernel/linux/linux-5.10/drivers/nvme/target/ |
H A D | passthru.c | 83 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_passthru_override_id_ctrl() 85 id->sgls |= cpu_to_le32(1 << 2); in nvmet_passthru_override_id_ctrl() 87 id->sgls |= cpu_to_le32(1 << 20); in nvmet_passthru_override_id_ctrl()
|
H A D | discovery.c | 281 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_execute_disc_identify() 283 id->sgls |= cpu_to_le32(1 << 2); in nvmet_execute_disc_identify() 285 id->sgls |= cpu_to_le32(1 << 20); in nvmet_execute_disc_identify()
|
H A D | admin-cmd.c | 425 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_execute_identify_ctrl() 427 id->sgls |= cpu_to_le32(1 << 2); in nvmet_execute_identify_ctrl() 429 id->sgls |= cpu_to_le32(1 << 20); in nvmet_execute_identify_ctrl()
|
/kernel/linux/linux-6.6/drivers/nvme/target/ |
H A D | discovery.c | 287 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_execute_disc_identify() 289 id->sgls |= cpu_to_le32(1 << 2); in nvmet_execute_disc_identify() 291 id->sgls |= cpu_to_le32(1 << 20); in nvmet_execute_disc_identify()
|
H A D | passthru.c | 140 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_passthru_override_id_ctrl() 142 id->sgls |= cpu_to_le32(1 << 2); in nvmet_passthru_override_id_ctrl() 144 id->sgls |= cpu_to_le32(1 << 20); in nvmet_passthru_override_id_ctrl()
|
H A D | admin-cmd.c | 448 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_execute_identify_ctrl() 450 id->sgls |= cpu_to_le32(1 << 2); in nvmet_execute_identify_ctrl() 452 id->sgls |= cpu_to_le32(1 << 20); in nvmet_execute_identify_ctrl()
|
/kernel/linux/linux-5.10/drivers/misc/genwqe/ |
H A D | card_dev.c | 846 if (req->sgls[i].sgl != NULL) in ddcb_cmd_cleanup() 847 genwqe_free_sync_sgl(cd, &req->sgls[i]); in ddcb_cmd_cleanup() 958 rc = genwqe_alloc_sync_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 964 genwqe_setup_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 968 cpu_to_be64(req->sgls[i].sgl_dma_addr); in ddcb_cmd_fixups()
|
H A D | card_base.h | 387 struct genwqe_sgl sgls[DDCB_FIXUPS]; member
|
/kernel/linux/linux-6.6/drivers/misc/genwqe/ |
H A D | card_dev.c | 846 if (req->sgls[i].sgl != NULL) in ddcb_cmd_cleanup() 847 genwqe_free_sync_sgl(cd, &req->sgls[i]); in ddcb_cmd_cleanup() 958 rc = genwqe_alloc_sync_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 964 genwqe_setup_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 968 cpu_to_be64(req->sgls[i].sgl_dma_addr); in ddcb_cmd_fixups()
|
H A D | card_base.h | 387 struct genwqe_sgl sgls[DDCB_FIXUPS]; member
|
/kernel/linux/linux-6.6/drivers/scsi/elx/efct/ |
H A D | efct_hw.c | 610 struct efc_dma **sgls = NULL; in efct_hw_init_prereg_io() local 614 sgls = kmalloc_array(sgls_per_request, sizeof(*sgls), GFP_KERNEL); in efct_hw_init_prereg_io() 615 if (!sgls) in efct_hw_init_prereg_io() 623 kfree(sgls); in efct_hw_init_prereg_io() 628 /* Copy address of SGL's into local sgls[] array, break in efct_hw_init_prereg_io() 641 sgls[n] = hw->io[idx + n]->sgl; in efct_hw_init_prereg_io() 645 hw->io[idx]->indicator, n, sgls, NULL, &req)) { in efct_hw_init_prereg_io() 667 kfree(sgls); in efct_hw_init_prereg_io()
|
/kernel/linux/linux-6.6/drivers/nvme/host/ |
H A D | nvme.h | 320 u32 sgls; member 1059 return ctrl->sgls & ((1 << 0) | (1 << 1)); in nvme_ctrl_sgl_supported()
|
H A D | rdma.c | 1020 if (!(ctrl->ctrl.sgls & (1 << 2))) { in nvme_rdma_setup_ctrl() 1023 "Mandatory keyed sgls are not supported!\n"); in nvme_rdma_setup_ctrl() 1047 if (ctrl->ctrl.sgls & (1 << 20)) in nvme_rdma_setup_ctrl()
|
H A D | core.c | 3090 ctrl->sgls = le32_to_cpu(id->sgls); in nvme_init_identify()
|
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | uverbs_cmd.c | 2000 const struct ib_sge __user *sgls; in ib_uverbs_post_send() local 2010 sgls = uverbs_request_next_ptr( in ib_uverbs_post_send() 2012 if (IS_ERR(sgls)) in ib_uverbs_post_send() 2013 return PTR_ERR(sgls); in ib_uverbs_post_send() 2140 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_post_send() 2191 const struct ib_sge __user *sgls; in ib_uverbs_unmarshall_recv() local 2200 sgls = uverbs_request_next_ptr( in ib_uverbs_unmarshall_recv() 2202 if (IS_ERR(sgls)) in ib_uverbs_unmarshall_recv() 2203 return ERR_CAST(sgls); in ib_uverbs_unmarshall_recv() 2254 if (copy_from_user(next->sg_list, sgls in ib_uverbs_unmarshall_recv() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | uverbs_cmd.c | 2005 const struct ib_sge __user *sgls; in ib_uverbs_post_send() local 2015 sgls = uverbs_request_next_ptr( in ib_uverbs_post_send() 2017 if (IS_ERR(sgls)) in ib_uverbs_post_send() 2018 return PTR_ERR(sgls); in ib_uverbs_post_send() 2145 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_post_send() 2196 const struct ib_sge __user *sgls; in ib_uverbs_unmarshall_recv() local 2205 sgls = uverbs_request_next_ptr( in ib_uverbs_unmarshall_recv() 2207 if (IS_ERR(sgls)) in ib_uverbs_unmarshall_recv() 2208 return ERR_CAST(sgls); in ib_uverbs_unmarshall_recv() 2259 if (copy_from_user(next->sg_list, sgls in ib_uverbs_unmarshall_recv() [all...] |
/kernel/linux/linux-5.10/drivers/nvme/host/ |
H A D | nvme.h | 297 u32 sgls; member
|
H A D | rdma.c | 1123 if (!(ctrl->ctrl.sgls & (1 << 2))) { in nvme_rdma_setup_ctrl() 1126 "Mandatory keyed sgls are not supported!\n"); in nvme_rdma_setup_ctrl() 1143 if (ctrl->ctrl.sgls & (1 << 20)) in nvme_rdma_setup_ctrl()
|
H A D | pci.c | 531 if (!(dev->ctrl.sgls & ((1 << 0) | (1 << 1)))) in nvme_pci_use_sgls() 852 dev->ctrl.sgls & ((1 << 0) | (1 << 1))) in nvme_map_data()
|
H A D | core.c | 3132 ctrl->sgls = le32_to_cpu(id->sgls); in nvme_init_identify()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | nvme.h | 300 __le32 sgls; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | nvme.h | 341 __le32 sgls; member
|