at_head 5489 block/bfq-iosched.c bool at_head) at_head 5509 block/bfq-iosched.c if (!bfqq || at_head || blk_rq_is_passthrough(rq)) { at_head 5510 block/bfq-iosched.c if (at_head) at_head 5544 block/bfq-iosched.c struct list_head *list, bool at_head) at_head 5551 block/bfq-iosched.c bfq_insert_request(hctx, rq, at_head); at_head 49 block/blk-exec.c struct request *rq, int at_head, at_head 62 block/blk-exec.c blk_mq_sched_insert_request(rq, at_head, true, false); at_head 78 block/blk-exec.c struct request *rq, int at_head) at_head 84 block/blk-exec.c blk_execute_rq_nowait(q, bd_disk, rq, at_head, blk_end_sync_rq); at_head 384 block/blk-mq-sched.c void blk_mq_sched_insert_request(struct request *rq, bool at_head, at_head 422 block/blk-mq-sched.c at_head = (rq->rq_flags & RQF_FLUSH_SEQ) ? true : at_head; at_head 423 block/blk-mq-sched.c blk_mq_request_bypass_insert(rq, at_head, false); at_head 431 block/blk-mq-sched.c e->type->ops.insert_requests(hctx, &list, at_head); at_head 434 block/blk-mq-sched.c __blk_mq_insert_request(hctx, rq, at_head); at_head 22 block/blk-mq-sched.h void blk_mq_sched_insert_request(struct request *rq, bool at_head, at_head 778 block/blk-mq.c void blk_mq_add_to_requeue_list(struct request *rq, bool at_head, at_head 791 block/blk-mq.c if (at_head) { at_head 1643 block/blk-mq.c bool at_head) at_head 1652 block/blk-mq.c if (at_head) at_head 1659 block/blk-mq.c bool at_head) at_head 1665 block/blk-mq.c __blk_mq_insert_req_list(hctx, rq, at_head); at_head 1673 block/blk-mq.c void blk_mq_request_bypass_insert(struct request *rq, bool at_head, at_head 1679 block/blk-mq.c if (at_head) at_head 44 block/blk-mq.h void blk_mq_add_to_requeue_list(struct request *rq, bool at_head, at_head 68 block/blk-mq.h bool at_head); at_head 69 block/blk-mq.h void blk_mq_request_bypass_insert(struct request *rq, bool at_head, at_head 588 block/kyber-iosched.c struct list_head *rq_list, bool at_head) at_head 599 block/kyber-iosched.c if (at_head) at_head 484 block/mq-deadline.c bool at_head) at_head 501 block/mq-deadline.c if (at_head || blk_rq_is_passthrough(rq)) { at_head 502 block/mq-deadline.c if (at_head) at_head 524 block/mq-deadline.c struct list_head *list, bool at_head) at_head 535 block/mq-deadline.c dd_insert_request(hctx, rq, at_head); at_head 284 block/scsi_ioctl.c int at_head = 0; at_head 307 block/scsi_ioctl.c at_head = 1; at_head 357 block/scsi_ioctl.c blk_execute_rq(q, bd_disk, rq, at_head); at_head 798 drivers/nvme/host/core.c struct gendisk *bd_disk, struct request *rq, int at_head) at_head 806 drivers/nvme/host/core.c blk_execute_rq_nowait(q, bd_disk, rq, at_head, nvme_end_sync_rq); at_head 820 drivers/nvme/host/core.c unsigned timeout, int qid, int at_head, at_head 839 drivers/nvme/host/core.c nvme_execute_rq_polled(req->q, NULL, req, at_head); at_head 841 drivers/nvme/host/core.c blk_execute_rq(req->q, NULL, req, at_head); at_head 492 drivers/nvme/host/nvme.h unsigned timeout, int qid, int at_head, at_head 792 drivers/scsi/sg.c int k, at_head; at_head 836 drivers/scsi/sg.c at_head = 0; at_head 838 drivers/scsi/sg.c at_head = 1; at_head 843 drivers/scsi/sg.c srp->rq, at_head, sg_rq_end_io); at_head 5063 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_io_request *io_request, bool at_head) at_head 5068 drivers/scsi/smartpqi/smartpqi_init.c if (at_head)