Lines Matching refs:iod
76 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
78 sg_free_table_chained(&iod->sg_table, NVME_INLINE_SG_CNT);
125 struct nvme_loop_iod *iod =
128 iod->req.execute(&iod->req);
137 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
149 iod->cmd.common.flags |= NVME_CMD_SGL_METABUF;
150 iod->req.port = queue->ctrl->port;
151 if (!nvmet_req_init(&iod->req, &queue->nvme_cq,
156 iod->sg_table.sgl = iod->first_sgl;
157 if (sg_alloc_table_chained(&iod->sg_table,
159 iod->sg_table.sgl, NVME_INLINE_SG_CNT)) {
164 iod->req.sg = iod->sg_table.sgl;
165 iod->req.sg_cnt = blk_rq_map_sg(req->q, req, iod->sg_table.sgl);
166 iod->req.transfer_len = blk_rq_payload_bytes(req);
169 queue_work(nvmet_wq, &iod->work);
177 struct nvme_loop_iod *iod = &ctrl->async_event_iod;
179 memset(&iod->cmd, 0, sizeof(iod->cmd));
180 iod->cmd.common.opcode = nvme_admin_async_event;
181 iod->cmd.common.command_id = NVME_AQ_BLK_MQ_DEPTH;
182 iod->cmd.common.flags |= NVME_CMD_SGL_METABUF;
184 if (!nvmet_req_init(&iod->req, &queue->nvme_cq, &queue->nvme_sq,
190 queue_work(nvmet_wq, &iod->work);
194 struct nvme_loop_iod *iod, unsigned int queue_idx)
196 iod->req.cmd = &iod->cmd;
197 iod->req.cqe = &iod->cqe;
198 iod->queue = &ctrl->queues[queue_idx];
199 INIT_WORK(&iod->work, nvme_loop_execute_work);
208 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
211 nvme_req(req)->cmd = &iod->cmd;