Lines Matching refs:rq
65 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
67 int blk_rq_append_bio(struct request_queue *q, struct request *rq,
71 void blk_dequeue_request(struct request *rq);
73 bool __blk_end_bidi_request(struct request *rq, int error,
127 static inline int blk_mark_rq_complete(struct request *rq) in blk_mark_rq_complete() argument
129 return test_and_set_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags); in blk_mark_rq_complete()
132 static inline void blk_clear_rq_complete(struct request *rq) in blk_clear_rq_complete() argument
134 clear_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags); in blk_clear_rq_complete()
140 #define ELV_ON_HASH(rq) ((rq)->cmd_flags & REQ_HASHED) argument
142 void blk_insert_flush(struct request *rq);
146 struct request *rq; in __elv_next_request() local
151 rq = list_entry_rq(q->queue_head.next); in __elv_next_request()
152 return rq; in __elv_next_request()
181 static inline void elv_activate_rq(struct request_queue *q, struct request *rq) in elv_activate_rq() argument
186 e->type->ops.elevator_activate_req_fn(q, rq); in elv_activate_rq()
189 static inline void elv_deactivate_rq(struct request_queue *q, struct request *rq) in elv_deactivate_rq() argument
194 e->type->ops.elevator_deactivate_req_fn(q, rq); in elv_deactivate_rq()
213 int attempt_back_merge(struct request_queue *q, struct request *rq);
214 int attempt_front_merge(struct request_queue *q, struct request *rq);
215 int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
217 void blk_recalc_rq_segments(struct request *rq);
218 void blk_rq_set_mixed_merge(struct request *rq);
219 bool blk_rq_merge_ok(struct request *rq, struct bio *bio);
220 int blk_try_merge(struct request *rq, struct bio *bio);
254 static inline int blk_do_io_stat(struct request *rq) in blk_do_io_stat() argument
256 return rq->rq_disk && in blk_do_io_stat()
257 (rq->cmd_flags & REQ_IO_STAT) && in blk_do_io_stat()
258 (rq->cmd_type == REQ_TYPE_FS); in blk_do_io_stat()