/kernel/linux/linux-5.10/block/ |
H A D | blk-exec.c | 38 * @at_head: insert request at head or tail of queue 49 struct request *rq, int at_head, in blk_execute_rq_nowait() 64 blk_mq_sched_insert_request(rq, at_head, true, false); in blk_execute_rq_nowait() 73 * @at_head: insert request at head or tail of queue 81 struct request *rq, int at_head) in blk_execute_rq() 87 blk_execute_rq_nowait(q, bd_disk, rq, at_head, blk_end_sync_rq); in blk_execute_rq() 48 blk_execute_rq_nowait(struct request_queue *q, struct gendisk *bd_disk, struct request *rq, int at_head, rq_end_io_fn *done) blk_execute_rq_nowait() argument 80 blk_execute_rq(struct request_queue *q, struct gendisk *bd_disk, struct request *rq, int at_head) blk_execute_rq() argument
|
H A D | blk-mq-sched.c | 427 void blk_mq_sched_insert_request(struct request *rq, bool at_head, in blk_mq_sched_insert_request() argument 459 at_head = (rq->rq_flags & RQF_FLUSH_SEQ) ? true : at_head; in blk_mq_sched_insert_request() 460 blk_mq_request_bypass_insert(rq, at_head, false); in blk_mq_sched_insert_request() 468 e->type->ops.insert_requests(hctx, &list, at_head); in blk_mq_sched_insert_request() 471 __blk_mq_insert_request(hctx, rq, at_head); in blk_mq_sched_insert_request()
|
H A D | blk-mq.h | 45 void blk_mq_add_to_requeue_list(struct request *rq, bool at_head, 70 bool at_head); 71 void blk_mq_request_bypass_insert(struct request *rq, bool at_head,
|
H A D | mq-deadline.c | 487 bool at_head) in dd_insert_request() 507 if (at_head || blk_rq_is_passthrough(rq)) { in dd_insert_request() 508 if (at_head) in dd_insert_request() 530 struct list_head *list, bool at_head) in dd_insert_requests() 541 dd_insert_request(hctx, rq, at_head); in dd_insert_requests() 486 dd_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, bool at_head) dd_insert_request() argument 529 dd_insert_requests(struct blk_mq_hw_ctx *hctx, struct list_head *list, bool at_head) dd_insert_requests() argument
|
H A D | blk-mq-sched.h | 20 void blk_mq_sched_insert_request(struct request *rq, bool at_head,
|
H A D | scsi_ioctl.c | 288 int at_head = 0; in sg_io() local 311 at_head = 1; in sg_io() 360 blk_execute_rq(q, bd_disk, rq, at_head); in sg_io()
|
H A D | blk-mq.c | 838 void blk_mq_add_to_requeue_list(struct request *rq, bool at_head, in blk_mq_add_to_requeue_list() argument 851 if (at_head) { in blk_mq_add_to_requeue_list() 1837 bool at_head) in __blk_mq_insert_req_list() 1846 if (at_head) in __blk_mq_insert_req_list() 1853 bool at_head) in __blk_mq_insert_request() 1859 __blk_mq_insert_req_list(hctx, rq, at_head); in __blk_mq_insert_request() 1866 * @at_head: true if the request should be inserted at the head of the list. 1872 void blk_mq_request_bypass_insert(struct request *rq, bool at_head, in blk_mq_request_bypass_insert() argument 1878 if (at_head) in blk_mq_request_bypass_insert() 1835 __blk_mq_insert_req_list(struct blk_mq_hw_ctx *hctx, struct request *rq, bool at_head) __blk_mq_insert_req_list() argument 1852 __blk_mq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, bool at_head) __blk_mq_insert_request() argument
|
H A D | kyber-iosched.c | 585 struct list_head *rq_list, bool at_head) in kyber_insert_requests() 596 if (at_head) in kyber_insert_requests() 584 kyber_insert_requests(struct blk_mq_hw_ctx *hctx, struct list_head *rq_list, bool at_head) kyber_insert_requests() argument
|
H A D | bfq-iosched.c | 5531 bool at_head) in bfq_insert_request() 5554 if (!bfqq || at_head || blk_rq_is_passthrough(rq)) { in bfq_insert_request() 5555 if (at_head) in bfq_insert_request() 5589 struct list_head *list, bool at_head) in bfq_insert_requests() 5596 bfq_insert_request(hctx, rq, at_head); in bfq_insert_requests() 5980 * re-preparation, can happen only for pass_through or at_head in bfq_finish_requeue_request() 5530 bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, bool at_head) bfq_insert_request() argument 5588 bfq_insert_requests(struct blk_mq_hw_ctx *hctx, struct list_head *list, bool at_head) bfq_insert_requests() argument
|
/kernel/linux/linux-6.6/drivers/scsi/ |
H A D | scsi_ioctl.c | 416 int at_head = 0; in sg_io() local 439 at_head = 1; in sg_io() 467 blk_execute_rq(rq, at_head); in sg_io()
|
H A D | sg.c | 782 int k, at_head; in sg_common_write() local 825 at_head = 0; in sg_common_write() 827 at_head = 1; in sg_common_write() 832 blk_execute_rq_nowait(srp->rq, at_head); in sg_common_write()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | blk-mq.h | 998 void blk_execute_rq_nowait(struct request *rq, bool at_head); 999 blk_status_t blk_execute_rq(struct request *rq, bool at_head);
|
/kernel/linux/linux-6.6/drivers/nvme/host/ |
H A D | nvme.h | 839 int qid, int at_head, 1097 int nvme_execute_rq(struct request *rq, bool at_head);
|
H A D | core.c | 1027 int nvme_execute_rq(struct request *rq, bool at_head) in nvme_execute_rq() argument 1031 status = blk_execute_rq(rq, at_head); in nvme_execute_rq() 1046 int qid, int at_head, blk_mq_req_flags_t flags) in __nvme_submit_sync_cmd() 1067 ret = nvme_execute_rq(req, at_head); in __nvme_submit_sync_cmd() 1044 __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd, union nvme_result *result, void *buffer, unsigned bufflen, int qid, int at_head, blk_mq_req_flags_t flags) __nvme_submit_sync_cmd() argument
|
/kernel/linux/linux-5.10/drivers/scsi/ |
H A D | sg.c | 779 int k, at_head; in sg_common_write() local 823 at_head = 0; in sg_common_write() 825 at_head = 1; in sg_common_write() 830 srp->rq, at_head, sg_rq_end_io); in sg_common_write()
|
/kernel/linux/linux-6.6/block/ |
H A D | blk-mq.c | 1319 * @at_head: insert request at head or tail of queue 1328 void blk_execute_rq_nowait(struct request *rq, bool at_head) in blk_execute_rq_nowait() argument 1342 if (current->plug && !at_head) { in blk_execute_rq_nowait() 1347 blk_mq_insert_request(rq, at_head ? BLK_MQ_INSERT_AT_HEAD : 0); in blk_execute_rq_nowait() 1387 * @at_head: insert request at head or tail of queue 1394 blk_status_t blk_execute_rq(struct request *rq, bool at_head) in blk_execute_rq() argument 1408 blk_mq_insert_request(rq, at_head ? BLK_MQ_INSERT_AT_HEAD : 0); in blk_execute_rq()
|
/kernel/linux/linux-5.10/drivers/nvme/host/ |
H A D | nvme.h | 700 unsigned timeout, int qid, int at_head,
|
H A D | core.c | 907 struct gendisk *bd_disk, struct request *rq, int at_head) in nvme_execute_rq_polled() 915 blk_execute_rq_nowait(q, bd_disk, rq, at_head, nvme_end_sync_rq); in nvme_execute_rq_polled() 929 unsigned timeout, int qid, int at_head, in __nvme_submit_sync_cmd() 952 nvme_execute_rq_polled(req->q, NULL, req, at_head); in __nvme_submit_sync_cmd() 954 blk_execute_rq(req->q, NULL, req, at_head); in __nvme_submit_sync_cmd() 906 nvme_execute_rq_polled(struct request_queue *q, struct gendisk *bd_disk, struct request *rq, int at_head) nvme_execute_rq_polled() argument 927 __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd, union nvme_result *result, void *buffer, unsigned bufflen, unsigned timeout, int qid, int at_head, blk_mq_req_flags_t flags, bool poll) __nvme_submit_sync_cmd() argument
|
/kernel/linux/linux-5.10/drivers/scsi/smartpqi/ |
H A D | smartpqi_init.c | 5076 struct pqi_io_request *io_request, bool at_head) in pqi_add_to_raid_bypass_retry_list() 5081 if (at_head) in pqi_add_to_raid_bypass_retry_list() 5074 pqi_add_to_raid_bypass_retry_list( struct pqi_ctrl_info *ctrl_info, struct pqi_io_request *io_request, bool at_head) pqi_add_to_raid_bypass_retry_list() argument
|