Lines Matching refs:rq
59 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
61 int blk_rq_append_bio(struct request_queue *q, struct request *rq,
65 void blk_dequeue_request(struct request *rq);
67 bool __blk_end_bidi_request(struct request *rq, int error,
99 static inline int blk_mark_rq_complete(struct request *rq) in blk_mark_rq_complete() argument
101 return test_and_set_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags); in blk_mark_rq_complete()
104 static inline void blk_clear_rq_complete(struct request *rq) in blk_clear_rq_complete() argument
106 clear_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags); in blk_clear_rq_complete()
112 #define ELV_ON_HASH(rq) ((rq)->cmd_flags & REQ_HASHED) argument
114 void blk_insert_flush(struct request *rq);
118 struct request *rq; in __elv_next_request() local
123 rq = list_entry_rq(q->queue_head.next); in __elv_next_request()
124 return rq; in __elv_next_request()
153 static inline void elv_activate_rq(struct request_queue *q, struct request *rq) in elv_activate_rq() argument
158 e->type->ops.elevator_activate_req_fn(q, rq); in elv_activate_rq()
161 static inline void elv_deactivate_rq(struct request_queue *q, struct request *rq) in elv_deactivate_rq() argument
166 e->type->ops.elevator_deactivate_req_fn(q, rq); in elv_deactivate_rq()
185 int attempt_back_merge(struct request_queue *q, struct request *rq);
186 int attempt_front_merge(struct request_queue *q, struct request *rq);
187 int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
189 void blk_recalc_rq_segments(struct request *rq);
190 void blk_rq_set_mixed_merge(struct request *rq);
191 bool blk_rq_merge_ok(struct request *rq, struct bio *bio);
192 int blk_try_merge(struct request *rq, struct bio *bio);
228 static inline int blk_do_io_stat(struct request *rq) in blk_do_io_stat() argument
230 return rq->rq_disk && in blk_do_io_stat()
231 (rq->cmd_flags & REQ_IO_STAT) && in blk_do_io_stat()
232 (rq->cmd_type == REQ_TYPE_FS); in blk_do_io_stat()