Home
last modified time | relevance | path

Searched refs:nvme_req (Results 1 - 25 of 29) sorted by relevance

12

/kernel/linux/linux-6.6/drivers/nvme/host/
H A Dtrace.h63 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
99 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
101 __entry->cid = nvme_req(req)->cmd->common.command_id;
102 __entry->result = le64_to_cpu(nvme_req(req)->result.u64);
103 __entry->retries = nvme_req(req)->retries;
104 __entry->flags = nvme_req(req)->flags;
105 __entry->status = nvme_req(req)->status;
152 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
H A Dfault_inject.c71 fault_inject = &nvme_req(req)->ctrl->fault_inject; in nvme_should_fail()
79 nvme_req(req)->status = status; in nvme_should_fail()
H A Dioctl.c162 nvme_req(req)->flags |= NVME_REQ_USERCMD; in nvme_alloc_user_request()
247 ctrl = nvme_req(req)->ctrl; in nvme_submit_user_cmd()
252 *result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_submit_user_cmd()
479 if (nvme_req(req)->flags & NVME_REQ_CANCELLED) in nvme_uring_task_meta_cb()
482 status = nvme_req(req)->status; in nvme_uring_task_meta_cb()
484 result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_uring_task_meta_cb()
514 if (nvme_req(req)->flags & NVME_REQ_CANCELLED) { in nvme_uring_cmd_end_io()
517 pdu->nvme_status = nvme_req(req)->status; in nvme_uring_cmd_end_io()
521 pdu->u.result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_uring_cmd_end_io()
H A Dnvme.h194 static inline struct nvme_request *nvme_req(struct request *req) in nvme_req() function
563 return nvme_cid_install_genctr(nvme_req(rq)->genctr) | rq->tag; in nvme_cid()
579 if (unlikely(nvme_genctr_mask(nvme_req(rq)->genctr) != genctr)) { in nvme_find_rq()
580 dev_err(nvme_req(rq)->ctrl->device, in nvme_find_rq()
582 tag, genctr, nvme_genctr_mask(nvme_req(rq)->genctr)); in nvme_find_rq()
707 struct nvme_request *rq = nvme_req(req); in nvme_try_complete_req()
H A Dcore.c300 crd = (nvme_req(req)->status & NVME_SC_CRD) >> 11; in nvme_retry_req()
302 delay = nvme_req(req)->ctrl->crdt[crd - 1] * 100; in nvme_retry_req()
304 nvme_req(req)->retries++; in nvme_retry_req()
312 struct nvme_request *nr = nvme_req(req); in nvme_log_error()
349 if (likely(nvme_req(req)->status == 0)) in nvme_decide_disposition()
352 if ((nvme_req(req)->status & 0x7ff) == NVME_SC_AUTH_REQUIRED) in nvme_decide_disposition()
356 (nvme_req(req)->status & NVME_SC_DNR) || in nvme_decide_disposition()
357 nvme_req(req)->retries >= nvme_max_retries) in nvme_decide_disposition()
361 if (nvme_is_path_error(nvme_req(req)->status) || in nvme_decide_disposition()
377 le64_to_cpu(nvme_req(re in nvme_end_req_zoned()
[all...]
H A Dmultipath.c86 u16 status = nvme_req(req)->status & 0x7ff; in nvme_failover_req()
133 nvme_req(rq)->flags |= NVME_MPATH_IO_STATS; in nvme_mpath_start_request()
134 nvme_req(rq)->start_time = bdev_start_io_acct(disk->part0, req_op(rq), in nvme_mpath_start_request()
143 if (!(nvme_req(rq)->flags & NVME_MPATH_IO_STATS)) in nvme_mpath_end_request()
147 nvme_req(rq)->start_time); in nvme_mpath_end_request()
H A Dfabrics.h194 nvme_req(rq)->status = NVME_SC_HOST_ABORTED_CMD; in nvmf_complete_timed_out_request()
H A Drdma.c302 nvme_req(rq)->ctrl = &ctrl->ctrl; in nvme_rdma_init_request()
309 req->metadata_sgl = (void *)nvme_req(rq) + in nvme_rdma_init_request()
314 nvme_req(rq)->cmd = req->sqe.data; in nvme_rdma_init_request()
1986 struct nvme_command *c = nvme_req(rq)->cmd; in nvme_rdma_queue_rq()
2075 nvme_req(rq)->status = NVME_SC_INVALID_PI; in nvme_rdma_check_pi_status()
2082 nvme_req(rq)->status = NVME_SC_GUARD_CHECK; in nvme_rdma_check_pi_status()
2085 nvme_req(rq)->status = NVME_SC_REFTAG_CHECK; in nvme_rdma_check_pi_status()
2088 nvme_req(rq)->status = NVME_SC_APPTAG_CHECK; in nvme_rdma_check_pi_status()
H A Dapple.c602 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in apple_nvme_handle_cqe()
789 struct nvme_request *nreq = nvme_req(req); in apple_nvme_init_request()
904 nvme_req(req)->status = NVME_SC_HOST_ABORTED_CMD; in apple_nvme_timeout()
905 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in apple_nvme_timeout()
930 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in apple_nvme_timeout()
H A Dpci.c425 nvme_req(req)->ctrl = set->driver_data; in nvme_pci_init_request()
426 nvme_req(req)->cmd = &iod->cmd; in nvme_pci_init_request()
1041 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in nvme_handle_cqe()
1223 "Abort status: 0x%x", nvme_req(req)->status); in abort_endio()
1333 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout()
1351 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout()
1368 nvme_get_opcode_str(nvme_req(req)->cmd->common.opcode), in nvme_timeout()
/kernel/linux/linux-5.10/drivers/nvme/host/
H A Dtrace.h63 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
99 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
102 __entry->result = le64_to_cpu(nvme_req(req)->result.u64);
103 __entry->retries = nvme_req(req)->retries;
104 __entry->flags = nvme_req(req)->flags;
105 __entry->status = nvme_req(req)->status;
152 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
H A Dfault_inject.c71 fault_inject = &nvme_req(req)->ctrl->fault_inject; in nvme_should_fail()
79 nvme_req(req)->status = status; in nvme_should_fail()
H A Dnvme.h184 static inline struct nvme_request *nvme_req(struct request *req) in nvme_req() function
508 return nvme_cid_install_genctr(nvme_req(rq)->genctr) | rq->tag; in nvme_cid()
524 if (unlikely(nvme_genctr_mask(nvme_req(rq)->genctr) != genctr)) { in nvme_find_rq()
525 dev_err(nvme_req(rq)->ctrl->device, in nvme_find_rq()
527 tag, genctr, nvme_genctr_mask(nvme_req(rq)->genctr)); in nvme_find_rq()
625 struct nvme_request *rq = nvme_req(req); in nvme_try_complete_req()
H A Dcore.c246 crd = (nvme_req(req)->status & NVME_SC_CRD) >> 11; in nvme_retry_req()
250 nvme_req(req)->retries++; in nvme_retry_req()
263 if (likely(nvme_req(req)->status == 0)) in nvme_decide_disposition()
267 (nvme_req(req)->status & NVME_SC_DNR) || in nvme_decide_disposition()
268 nvme_req(req)->retries >= nvme_max_retries) in nvme_decide_disposition()
272 if (nvme_is_path_error(nvme_req(req)->status) || in nvme_decide_disposition()
285 blk_status_t status = nvme_error_status(nvme_req(req)->status); in nvme_end_req()
290 le64_to_cpu(nvme_req(req)->result.u64)); in nvme_end_req()
301 if (nvme_req(req)->ctrl->kas) in nvme_complete_rq()
302 nvme_req(re in nvme_complete_rq()
[all...]
H A Dlightnvm.c639 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io()
640 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io()
643 kfree(nvme_req(rq)->cmd); in nvme_nvm_end_io()
825 if (nvme_req(rq)->flags & NVME_REQ_CANCELLED) in nvme_nvm_submit_user_cmd()
827 else if (nvme_req(rq)->status & 0x7ff) in nvme_nvm_submit_user_cmd()
830 *result = nvme_req(rq)->status & 0x7ff; in nvme_nvm_submit_user_cmd()
832 *status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_submit_user_cmd()
H A Drdma.c305 nvme_req(rq)->ctrl = &ctrl->ctrl; in nvme_rdma_init_request()
312 req->metadata_sgl = (void *)nvme_req(rq) + in nvme_rdma_init_request()
2009 nvme_req(rq)->status = NVME_SC_HOST_ABORTED_CMD; in nvme_rdma_complete_timed_out()
2145 nvme_req(rq)->status = NVME_SC_INVALID_PI; in nvme_rdma_check_pi_status()
2152 nvme_req(rq)->status = NVME_SC_GUARD_CHECK; in nvme_rdma_check_pi_status()
2155 nvme_req(rq)->status = NVME_SC_REFTAG_CHECK; in nvme_rdma_check_pi_status()
2158 nvme_req(rq)->status = NVME_SC_APPTAG_CHECK; in nvme_rdma_check_pi_status()
H A Dfabrics.c560 nvme_req(rq)->status = NVME_SC_HOST_PATH_ERROR; in nvmf_fail_nonready_command()
570 struct nvme_request *req = nvme_req(rq); in __nvmf_check_ready()
H A Dpci.c425 nvme_req(req)->ctrl = &dev->ctrl; in nvme_init_request()
1207 "Abort status: 0x%x", nvme_req(req)->status); in abort_endio()
1310 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout()
1328 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout()
H A Dmultipath.c71 u16 status = nvme_req(req)->status & 0x7ff; in nvme_failover_req()
/kernel/linux/linux-5.10/drivers/nvme/target/
H A Dpassthru.c160 status = nvme_req(rq)->status; in nvmet_passthru_execute_cmd_work()
173 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_execute_cmd_work()
183 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_req_done()
184 nvmet_req_complete(req, nvme_req(rq)->status); in nvmet_passthru_req_done()
H A Dloop.c19 struct nvme_request nvme_req; member
64 static void nvme_loop_queue_response(struct nvmet_req *nvme_req);
209 nvme_req(req)->ctrl = &ctrl->ctrl; in nvme_loop_init_request()
H A Dfc.c226 nvmet_req_to_fod(struct nvmet_req *nvme_req) in nvmet_req_to_fod() argument
228 return container_of(nvme_req, struct nvmet_fc_fcp_iod, req); in nvmet_req_to_fod()
1903 static void nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req);
2484 nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req) in nvmet_fc_fcp_nvme_cmd_done() argument
2486 struct nvmet_fc_fcp_iod *fod = nvmet_req_to_fod(nvme_req); in nvmet_fc_fcp_nvme_cmd_done()
/kernel/linux/linux-6.6/drivers/nvme/target/
H A Dpassthru.c218 struct nvme_ctrl *ctrl = nvme_req(rq)->ctrl; in nvmet_passthru_execute_cmd_work()
241 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_execute_cmd_work()
254 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_req_done()
255 nvmet_req_complete(req, nvme_req(rq)->status); in nvmet_passthru_req_done()
H A Dloop.c19 struct nvme_request nvme_req; member
64 static void nvme_loop_queue_response(struct nvmet_req *nvme_req);
210 nvme_req(req)->ctrl = &ctrl->ctrl; in nvme_loop_init_request()
211 nvme_req(req)->cmd = &iod->cmd; in nvme_loop_init_request()
H A Dfc.c230 nvmet_req_to_fod(struct nvmet_req *nvme_req) in nvmet_req_to_fod() argument
232 return container_of(nvme_req, struct nvmet_fc_fcp_iod, req); in nvmet_req_to_fod()
1925 static void nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req);
2507 nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req) in nvmet_fc_fcp_nvme_cmd_done() argument
2509 struct nvmet_fc_fcp_iod *fod = nvmet_req_to_fod(nvme_req); in nvmet_fc_fcp_nvme_cmd_done()

Completed in 58 milliseconds

12