next_rq           570 block/bfq-iosched.c 		if (sector > blk_rq_pos(bfqq->next_rq))
next_rq           572 block/bfq-iosched.c 		else if (sector < blk_rq_pos(bfqq->next_rq))
next_rq           631 block/bfq-iosched.c 	if (!bfqq->next_rq)
next_rq           636 block/bfq-iosched.c 			blk_rq_pos(bfqq->next_rq), &parent, &p);
next_rq           943 block/bfq-iosched.c 	struct request *next_rq = bfqq->next_rq;
next_rq           946 block/bfq-iosched.c 	if (!next_rq)
next_rq           958 block/bfq-iosched.c 				 bfq_serv_to_charge(next_rq, bfqq)),
next_rq          1500 block/bfq-iosched.c 			       bfq_serv_to_charge(bfqq->next_rq, bfqq));
next_rq          1868 block/bfq-iosched.c 	struct request *next_rq, *prev;
next_rq          2055 block/bfq-iosched.c 	prev = bfqq->next_rq;
next_rq          2056 block/bfq-iosched.c 	next_rq = bfq_choose_req(bfqd, bfqq->next_rq, rq, bfqd->last_position);
next_rq          2057 block/bfq-iosched.c 	bfqq->next_rq = next_rq;
next_rq          2063 block/bfq-iosched.c 	if (unlikely(!bfqd->nonrot_with_queueing && prev != bfqq->next_rq))
next_rq          2080 block/bfq-iosched.c 		if (prev != bfqq->next_rq)
next_rq          2159 block/bfq-iosched.c 	if (bfqq->next_rq == rq) {
next_rq          2160 block/bfq-iosched.c 		bfqq->next_rq = bfq_find_next_rq(bfqd, bfqq, rq);
next_rq          2175 block/bfq-iosched.c 		bfqq->next_rq = NULL;
next_rq          2273 block/bfq-iosched.c 		struct request *prev, *next_rq;
next_rq          2285 block/bfq-iosched.c 		prev = bfqq->next_rq;
next_rq          2286 block/bfq-iosched.c 		next_rq = bfq_choose_req(bfqd, bfqq->next_rq, req,
next_rq          2288 block/bfq-iosched.c 		bfqq->next_rq = next_rq;
next_rq          2294 block/bfq-iosched.c 		if (prev != bfqq->next_rq) {
next_rq          2346 block/bfq-iosched.c 	if (bfqq->next_rq == next)
next_rq          2347 block/bfq-iosched.c 		bfqq->next_rq = rq;
next_rq          2435 block/bfq-iosched.c 	if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
next_rq          2438 block/bfq-iosched.c 	if (blk_rq_pos(__bfqq->next_rq) < sector)
next_rq          2446 block/bfq-iosched.c 	if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
next_rq          3533 block/bfq-iosched.c 	struct request *next_rq;
next_rq          3680 block/bfq-iosched.c 	next_rq = bfqq->next_rq;
next_rq          3681 block/bfq-iosched.c 	if (next_rq)
next_rq          3683 block/bfq-iosched.c 					    bfq_serv_to_charge(next_rq, bfqq));
next_rq          3686 block/bfq-iosched.c 			next_rq ? blk_rq_sectors(next_rq) : 0,
next_rq          4295 block/bfq-iosched.c 		    bfq_serv_to_charge(bfqq->next_rq, bfqq) <=
next_rq          4315 block/bfq-iosched.c 			    blk_rq_sectors(bfqq->next_rq) >=
next_rq          4337 block/bfq-iosched.c 	struct request *next_rq;
next_rq          4364 block/bfq-iosched.c 	next_rq = bfqq->next_rq;
next_rq          4369 block/bfq-iosched.c 	if (next_rq) {
next_rq          4370 block/bfq-iosched.c 		if (bfq_serv_to_charge(next_rq, bfqq) >
next_rq          4420 block/bfq-iosched.c 			bfqq->bic->bfqq[0]->next_rq ?
next_rq          4496 block/bfq-iosched.c 		    icq_to_bic(async_bfqq->next_rq->elv.icq) == bfqq->bic &&
next_rq          4497 block/bfq-iosched.c 		    bfq_serv_to_charge(async_bfqq->next_rq, async_bfqq) <=
next_rq          4502 block/bfq-iosched.c 			   bfqq->next_rq &&
next_rq          4503 block/bfq-iosched.c 			   bfq_serv_to_charge(bfqq->waker_bfqq->next_rq,
next_rq          4595 block/bfq-iosched.c 	struct request *rq = bfqq->next_rq;
next_rq           259 block/bfq-iosched.h 	struct request *next_rq;
next_rq            50 block/mq-deadline.c 	struct request *next_rq[2];
next_rq           100 block/mq-deadline.c 	if (dd->next_rq[data_dir] == rq)
next_rq           101 block/mq-deadline.c 		dd->next_rq[data_dir] = deadline_latter_request(rq);
next_rq           169 block/mq-deadline.c 	dd->next_rq[READ] = NULL;
next_rq           170 block/mq-deadline.c 	dd->next_rq[WRITE] = NULL;
next_rq           171 block/mq-deadline.c 	dd->next_rq[data_dir] = deadline_latter_request(rq);
next_rq           245 block/mq-deadline.c 	rq = dd->next_rq[data_dir];
next_rq           273 block/mq-deadline.c 	struct request *rq, *next_rq;
next_rq           335 block/mq-deadline.c 	next_rq = deadline_next_request(dd, data_dir);
next_rq           336 block/mq-deadline.c 	if (deadline_check_fifo(dd, data_dir) || !next_rq) {
next_rq           348 block/mq-deadline.c 		rq = next_rq;
next_rq           698 block/mq-deadline.c 	struct request *rq = dd->next_rq[ddir];				\
next_rq           790 drivers/lightnvm/pblk-core.c next_rq:
next_rq           848 drivers/lightnvm/pblk-core.c 		goto next_rq;
next_rq           219 drivers/lightnvm/pblk-gc.c next_rq:
next_rq           269 drivers/lightnvm/pblk-gc.c 		goto next_rq;
next_rq           385 drivers/lightnvm/pblk-recovery.c next_rq:
next_rq           467 drivers/lightnvm/pblk-recovery.c 		goto next_rq;