Lines Matching refs:queue

52 	struct request_queue *q = mq->queue;  in mmc_queue_thread()
201 mq->queue = blk_init_queue(mmc_request_fn, lock); in mmc_init_queue()
202 if (!mq->queue) in mmc_init_queue()
207 mq->queue->queuedata = mq; in mmc_init_queue()
209 blk_queue_prep_rq(mq->queue, mmc_prep_request); in mmc_init_queue()
210 queue_flag_set_unlocked(QUEUE_FLAG_NONROT, mq->queue); in mmc_init_queue()
211 queue_flag_clear_unlocked(QUEUE_FLAG_ADD_RANDOM, mq->queue); in mmc_init_queue()
213 mmc_queue_setup_discard(mq->queue, card); in mmc_init_queue()
246 blk_queue_bounce_limit(mq->queue, BLK_BOUNCE_ANY); in mmc_init_queue()
247 blk_queue_max_hw_sectors(mq->queue, bouncesz / 512); in mmc_init_queue()
248 blk_queue_max_segments(mq->queue, bouncesz / 512); in mmc_init_queue()
249 blk_queue_max_segment_size(mq->queue, bouncesz); in mmc_init_queue()
273 blk_queue_bounce_limit(mq->queue, limit); in mmc_init_queue()
274 blk_queue_max_hw_sectors(mq->queue, in mmc_init_queue()
276 blk_queue_max_segments(mq->queue, host->max_segs); in mmc_init_queue()
277 blk_queue_max_segment_size(mq->queue, host->max_seg_size); in mmc_init_queue()
317 blk_cleanup_queue(mq->queue); in mmc_init_queue()
323 struct request_queue *q = mq->queue; in mmc_cleanup_queue()
415 struct request_queue *q = mq->queue; in mmc_queue_suspend()
435 struct request_queue *q = mq->queue; in mmc_queue_resume()
460 unsigned int max_seg_sz = queue_max_segment_size(mq->queue); in mmc_queue_packed_map_sg()
476 sg_len += blk_rq_map_sg(mq->queue, req, __sg); in mmc_queue_packed_map_sg()
502 return blk_rq_map_sg(mq->queue, mqrq->req, mqrq->sg); in mmc_queue_map_sg()
511 sg_len = blk_rq_map_sg(mq->queue, mqrq->req, mqrq->bounce_sg); in mmc_queue_map_sg()