Lines Matching refs:req
22 struct nvmet_req req;
74 static void nvme_loop_complete_rq(struct request *req)
76 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
79 nvme_complete_rq(req);
91 static void nvme_loop_queue_response(struct nvmet_req *req)
94 container_of(req->sq, struct nvme_loop_queue, nvme_sq);
95 struct nvme_completion *cqe = req->cqe;
128 iod->req.execute(&iod->req);
136 struct request *req = bd->rq;
137 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req);
141 if (!nvmf_check_ready(&queue->ctrl->ctrl, req, queue_ready))
142 return nvmf_fail_nonready_command(&queue->ctrl->ctrl, req);
144 ret = nvme_setup_cmd(ns, req, &iod->cmd);
148 blk_mq_start_request(req);
150 iod->req.port = queue->ctrl->port;
151 if (!nvmet_req_init(&iod->req, &queue->nvme_cq,
155 if (blk_rq_nr_phys_segments(req)) {
158 blk_rq_nr_phys_segments(req),
160 nvme_cleanup_cmd(req);
164 iod->req.sg = iod->sg_table.sgl;
165 iod->req.sg_cnt = blk_rq_map_sg(req->q, req, iod->sg_table.sgl);
166 iod->req.transfer_len = blk_rq_payload_bytes(req);
184 if (!nvmet_req_init(&iod->req, &queue->nvme_cq, &queue->nvme_sq,
196 iod->req.cmd = &iod->cmd;
197 iod->req.cqe = &iod->cqe;
204 struct request *req, unsigned int hctx_idx,
209 nvme_req(req)->ctrl = &ctrl->ctrl;
210 return nvme_loop_init_iod(ctrl, blk_mq_rq_to_pdu(req),