Lines Matching refs:rq
120 struct request_queue *rq; member
610 static void do_blkif_request(struct request_queue *rq) in do_blkif_request() argument
620 while ((req = blk_peek_request(rq)) != NULL) { in do_blkif_request()
640 blk_requeue_request(rq, req); in do_blkif_request()
643 blk_stop_queue(rq); in do_blkif_request()
658 struct request_queue *rq; in xlvbd_init_blk_queue() local
661 rq = blk_init_queue(do_blkif_request, &info->io_lock); in xlvbd_init_blk_queue()
662 if (rq == NULL) in xlvbd_init_blk_queue()
665 queue_flag_set_unlocked(QUEUE_FLAG_VIRT, rq); in xlvbd_init_blk_queue()
668 queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, rq); in xlvbd_init_blk_queue()
669 blk_queue_max_discard_sectors(rq, get_capacity(gd)); in xlvbd_init_blk_queue()
670 rq->limits.discard_granularity = info->discard_granularity; in xlvbd_init_blk_queue()
671 rq->limits.discard_alignment = info->discard_alignment; in xlvbd_init_blk_queue()
673 queue_flag_set_unlocked(QUEUE_FLAG_SECDISCARD, rq); in xlvbd_init_blk_queue()
677 blk_queue_logical_block_size(rq, sector_size); in xlvbd_init_blk_queue()
678 blk_queue_physical_block_size(rq, physical_sector_size); in xlvbd_init_blk_queue()
679 blk_queue_max_hw_sectors(rq, (segments * PAGE_SIZE) / 512); in xlvbd_init_blk_queue()
682 blk_queue_segment_boundary(rq, PAGE_SIZE - 1); in xlvbd_init_blk_queue()
683 blk_queue_max_segment_size(rq, PAGE_SIZE); in xlvbd_init_blk_queue()
686 blk_queue_max_segments(rq, segments); in xlvbd_init_blk_queue()
689 blk_queue_dma_alignment(rq, 511); in xlvbd_init_blk_queue()
692 blk_queue_bounce_limit(rq, BLK_BOUNCE_ANY); in xlvbd_init_blk_queue()
694 gd->queue = rq; in xlvbd_init_blk_queue()
713 blk_queue_flush(info->rq, info->feature_flush); in xlvbd_flush()
803 BUG_ON(info->rq != NULL); in xlvbd_alloc_gendisk()
866 info->rq = gd->queue; in xlvbd_alloc_gendisk()
893 if (info->rq == NULL) in xlvbd_release_gendisk()
899 blk_stop_queue(info->rq); in xlvbd_release_gendisk()
914 blk_cleanup_queue(info->rq); in xlvbd_release_gendisk()
915 info->rq = NULL; in xlvbd_release_gendisk()
925 blk_start_queue(info->rq); in kick_pending_request_queues()
927 do_blkif_request(info->rq); in kick_pending_request_queues()
952 if (info->rq) in blkif_free()
953 blk_stop_queue(info->rq); in blkif_free()
1184 struct request_queue *rq = info->rq; in blkif_interrupt() local
1190 queue_flag_clear(QUEUE_FLAG_DISCARD, rq); in blkif_interrupt()
1191 queue_flag_clear(QUEUE_FLAG_SECDISCARD, rq); in blkif_interrupt()
1493 blk_queue_max_segments(info->rq, segs); in blkif_recover()
1528 while ((req = blk_fetch_request(info->rq)) != NULL) { in blkif_recover()
1558 blk_requeue_request(info->rq, req); in blkif_recover()