at_head          5489 block/bfq-iosched.c 			       bool at_head)
at_head          5509 block/bfq-iosched.c 	if (!bfqq || at_head || blk_rq_is_passthrough(rq)) {
at_head          5510 block/bfq-iosched.c 		if (at_head)
at_head          5544 block/bfq-iosched.c 				struct list_head *list, bool at_head)
at_head          5551 block/bfq-iosched.c 		bfq_insert_request(hctx, rq, at_head);
at_head            49 block/blk-exec.c 			   struct request *rq, int at_head,
at_head            62 block/blk-exec.c 	blk_mq_sched_insert_request(rq, at_head, true, false);
at_head            78 block/blk-exec.c 		   struct request *rq, int at_head)
at_head            84 block/blk-exec.c 	blk_execute_rq_nowait(q, bd_disk, rq, at_head, blk_end_sync_rq);
at_head           384 block/blk-mq-sched.c void blk_mq_sched_insert_request(struct request *rq, bool at_head,
at_head           422 block/blk-mq-sched.c 		at_head = (rq->rq_flags & RQF_FLUSH_SEQ) ? true : at_head;
at_head           423 block/blk-mq-sched.c 		blk_mq_request_bypass_insert(rq, at_head, false);
at_head           431 block/blk-mq-sched.c 		e->type->ops.insert_requests(hctx, &list, at_head);
at_head           434 block/blk-mq-sched.c 		__blk_mq_insert_request(hctx, rq, at_head);
at_head            22 block/blk-mq-sched.h void blk_mq_sched_insert_request(struct request *rq, bool at_head,
at_head           778 block/blk-mq.c void blk_mq_add_to_requeue_list(struct request *rq, bool at_head,
at_head           791 block/blk-mq.c 	if (at_head) {
at_head          1643 block/blk-mq.c 					    bool at_head)
at_head          1652 block/blk-mq.c 	if (at_head)
at_head          1659 block/blk-mq.c 			     bool at_head)
at_head          1665 block/blk-mq.c 	__blk_mq_insert_req_list(hctx, rq, at_head);
at_head          1673 block/blk-mq.c void blk_mq_request_bypass_insert(struct request *rq, bool at_head,
at_head          1679 block/blk-mq.c 	if (at_head)
at_head            44 block/blk-mq.h void blk_mq_add_to_requeue_list(struct request *rq, bool at_head,
at_head            68 block/blk-mq.h 				bool at_head);
at_head            69 block/blk-mq.h void blk_mq_request_bypass_insert(struct request *rq, bool at_head,
at_head           588 block/kyber-iosched.c 				  struct list_head *rq_list, bool at_head)
at_head           599 block/kyber-iosched.c 		if (at_head)
at_head           484 block/mq-deadline.c 			      bool at_head)
at_head           501 block/mq-deadline.c 	if (at_head || blk_rq_is_passthrough(rq)) {
at_head           502 block/mq-deadline.c 		if (at_head)
at_head           524 block/mq-deadline.c 			       struct list_head *list, bool at_head)
at_head           535 block/mq-deadline.c 		dd_insert_request(hctx, rq, at_head);
at_head           284 block/scsi_ioctl.c 	int at_head = 0;
at_head           307 block/scsi_ioctl.c 		at_head = 1;
at_head           357 block/scsi_ioctl.c 	blk_execute_rq(q, bd_disk, rq, at_head);
at_head           798 drivers/nvme/host/core.c 		struct gendisk *bd_disk, struct request *rq, int at_head)
at_head           806 drivers/nvme/host/core.c 	blk_execute_rq_nowait(q, bd_disk, rq, at_head, nvme_end_sync_rq);
at_head           820 drivers/nvme/host/core.c 		unsigned timeout, int qid, int at_head,
at_head           839 drivers/nvme/host/core.c 		nvme_execute_rq_polled(req->q, NULL, req, at_head);
at_head           841 drivers/nvme/host/core.c 		blk_execute_rq(req->q, NULL, req, at_head);
at_head           492 drivers/nvme/host/nvme.h 		unsigned timeout, int qid, int at_head,
at_head           792 drivers/scsi/sg.c 	int k, at_head;
at_head           836 drivers/scsi/sg.c 		at_head = 0;
at_head           838 drivers/scsi/sg.c 		at_head = 1;
at_head           843 drivers/scsi/sg.c 			      srp->rq, at_head, sg_rq_end_io);
at_head          5063 drivers/scsi/smartpqi/smartpqi_init.c 	struct pqi_io_request *io_request, bool at_head)
at_head          5068 drivers/scsi/smartpqi/smartpqi_init.c 	if (at_head)