| /linux-4.1.27/drivers/md/ | 
| D | linear.c | 75 	subq = bdev_get_queue(dev0->rdev->bdev);  in linear_mergeable_bvec() 105 		struct request_queue *q = bdev_get_queue(conf->disks[i].rdev->bdev);  in linear_congested() 164 		if (blk_queue_discard(bdev_get_queue(rdev->bdev)))  in linear_conf() 298 			 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) {  in linear_make_request()
  | 
| D | dm-table.c | 294 	q = bdev_get_queue(bdev);  in device_area_is_invalid() 424 	struct request_queue *q = bdev_get_queue(bdev);  in dm_set_device_limits() 889 		struct request_queue *q = bdev_get_queue(dd->dm_dev->bdev);  in dm_table_set_type() 904 			if (!bdev_get_queue(dd->dm_dev->bdev)->mq_ops) {  in dm_table_set_type() 1318 	struct request_queue *q = bdev_get_queue(dev->bdev);  in device_flush_capable() 1370 	struct request_queue *q = bdev_get_queue(dev->bdev);  in device_is_nonrot() 1378 	struct request_queue *q = bdev_get_queue(dev->bdev);  in device_is_not_random() 1386 	struct request_queue *q = bdev_get_queue(dev->bdev);  in queue_supports_sg_merge() 1394 	struct request_queue *q = bdev_get_queue(dev->bdev);  in queue_supports_sg_gaps() 1419 	struct request_queue *q = bdev_get_queue(dev->bdev);  in device_not_write_same_capable() [all …] 
 | 
| D | raid0.c | 36 		struct request_queue *q = bdev_get_queue(devlist[i]->bdev);  in raid0_congested() 382 	subq = bdev_get_queue(rdev->bdev);  in raid0_mergeable_bvec() 445 			if (blk_queue_discard(bdev_get_queue(rdev->bdev)))  in raid0_run() 547 			 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) {  in raid0_make_request()
  | 
| D | dm-mpath.c | 167 			scsi_dh_detach(bdev_get_queue(pgpath->path.dev->bdev));  in free_pgpaths() 425 		clone->q = bdev_get_queue(bdev);  in __multipath_map() 430 		*__clone = blk_get_request(bdev_get_queue(bdev),  in __multipath_map() 580 		q = bdev_get_queue(p->path.dev->bdev);  in parse_path() 1240 		scsi_dh_activate(bdev_get_queue(pgpath->path.dev->bdev),  in activate_path() 1630 	struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev);  in __pgpath_busy()
  | 
| D | dm-linear.c | 137 	struct request_queue *q = bdev_get_queue(lc->dev->bdev);  in linear_merge()
  | 
| D | raid1.c | 615 		nonrot = blk_queue_nonrot(bdev_get_queue(rdev->bdev));  in read_balance() 727 					bdev_get_queue(rdev->bdev);  in raid1_mergeable_bvec() 756 			struct request_queue *q = bdev_get_queue(rdev->bdev);  in raid1_congested() 794 			    !blk_queue_discard(bdev_get_queue(bio->bi_bdev))))  in flush_pending_writes() 1069 		    !blk_queue_discard(bdev_get_queue(bio->bi_bdev))))  in raid1_unplug() 1619 	struct request_queue *q = bdev_get_queue(rdev->bdev);  in raid1_add_disk() 1677 	if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev)))  in raid1_add_disk() 2811 		q = bdev_get_queue(rdev->bdev);  in setup_conf() 2924 		if (blk_queue_discard(bdev_get_queue(rdev->bdev)))  in run()
  | 
| D | dm-log-writes.c | 735 	struct request_queue *q = bdev_get_queue(lc->dev->bdev);  in log_writes_merge() 780 	struct request_queue *q = bdev_get_queue(lc->dev->bdev);  in log_writes_io_hints()
  | 
| D | dm-flakey.c | 394 	struct request_queue *q = bdev_get_queue(fc->dev->bdev);  in flakey_merge()
  | 
| D | dm-stripe.c | 425 	q = bdev_get_queue(sc->stripe[stripe].dev->bdev);  in stripe_merge()
  | 
| D | raid10.c | 735 					bdev_get_queue(rdev->bdev);  in raid10_mergeable_bvec() 747 					bdev_get_queue(rdev->bdev);  in raid10_mergeable_bvec() 928 			struct request_queue *q = bdev_get_queue(rdev->bdev);  in raid10_congested() 958 			    !blk_queue_discard(bdev_get_queue(bio->bi_bdev))))  in flush_pending_writes() 1134 		    !blk_queue_discard(bdev_get_queue(bio->bi_bdev))))  in raid10_unplug() 1780 	struct request_queue *q = bdev_get_queue(rdev->bdev);  in raid10_add_disk() 1850 	if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev)))  in raid10_add_disk() 3655 		q = bdev_get_queue(rdev->bdev);  in run() 3672 		if (blk_queue_discard(bdev_get_queue(rdev->bdev)))  in run()
  | 
| D | multipath.c | 167 			struct request_queue *q = bdev_get_queue(rdev->bdev);  in multipath_congested()
  | 
| D | dm-io.c | 289 	struct request_queue *q = bdev_get_queue(where->bdev);  in do_region()
  | 
| D | dm-thin.c | 2386 	q = bdev_get_queue(pt->data_dev->bdev);  in pool_is_congested() 2410 	struct request_queue *q = bdev_get_queue(pt->data_dev->bdev);  in data_dev_supports_discard() 2428 	struct queue_limits *data_limits = &bdev_get_queue(data_bdev)->limits;  in disable_passdown_if_not_supported() 3580 	struct request_queue *q = bdev_get_queue(pt->data_dev->bdev);  in pool_merge() 3601 		data_limits = &bdev_get_queue(pt->data_dev->bdev)->limits;  in set_discard_limits() 4010 	struct request_queue *q = bdev_get_queue(tc->pool_dev->bdev);  in thin_merge()
  | 
| D | dm-era-target.c | 1382 	struct request_queue *q = bdev_get_queue(dev->bdev);  in dev_is_congested() 1680 	struct request_queue *q = bdev_get_queue(era->origin_dev->bdev);  in era_merge()
  | 
| D | raid5.c | 234 		trace_block_bio_complete(bdev_get_queue(bi->bi_bdev),  in return_io() 1034 				trace_block_bio_remap(bdev_get_queue(bi->bi_bdev),  in ops_run_io() 1080 				trace_block_bio_remap(bdev_get_queue(rbi->bi_bdev),  in ops_run_io() 4778 		trace_block_bio_complete(bdev_get_queue(raid_bi->bi_bdev),  in raid5_align_endio() 4793 	struct request_queue *q = bdev_get_queue(bi->bi_bdev);  in bio_fits_rdev() 4884 			trace_block_bio_remap(bdev_get_queue(align_bi->bi_bdev),  in chunk_aligned_read() 5320 		trace_block_bio_complete(bdev_get_queue(bi->bi_bdev),  in make_request() 5716 		trace_block_bio_complete(bdev_get_queue(raid_bio->bi_bdev),  in retry_aligned_read() 6944 			if (!blk_queue_discard(bdev_get_queue(rdev->bdev)) ||  in run() 6945 			    !bdev_get_queue(rdev->bdev)->  in run()
  | 
| D | dm.c | 986 		     !bdev_get_queue(bio->bi_bdev)->limits.max_write_same_sectors))  in clone_endio() 1451 		trace_block_bio_remap(bdev_get_queue(clone->bi_bdev), clone,  in __map_bio() 2501 	struct request_queue *q = bdev_get_queue(bdev);  in dm_device_merge_is_compulsory()
  | 
| D | dm-verity.c | 655 	struct request_queue *q = bdev_get_queue(v->data_dev->bdev);  in verity_merge()
  | 
| D | dm-raid.c | 1173 		q = bdev_get_queue(rs->dev[i].rdev.bdev);  in configure_discard_support()
  | 
| D | dm-cache-target.c | 1886 	struct request_queue *q = bdev_get_queue(dev->bdev);  in is_congested() 3321 	struct request_queue *q = bdev_get_queue(cache->origin_dev->bdev);  in cache_bvec_merge()
  | 
| D | dm-snap.c | 2328 	struct request_queue *q = bdev_get_queue(o->dev->bdev);  in origin_merge()
  | 
| D | dm-crypt.c | 2025 	struct request_queue *q = bdev_get_queue(cc->dev->bdev);  in crypt_merge()
  | 
| /linux-4.1.27/block/ | 
| D | blk-lib.c | 44 	struct request_queue *q = bdev_get_queue(bdev);  in blkdev_issue_discard() 160 	struct request_queue *q = bdev_get_queue(bdev);  in blkdev_issue_write_same() 304 	struct request_queue *q = bdev_get_queue(bdev);  in blkdev_issue_zeroout()
  | 
| D | ioctl.c | 378 				    queue_max_sectors(bdev_get_queue(bdev)));  in blkdev_ioctl() 381 		return put_ushort(arg, !blk_queue_nonrot(bdev_get_queue(bdev)));  in blkdev_ioctl()
  | 
| D | blk-flush.c | 459 	q = bdev_get_queue(bdev);  in blkdev_issue_flush()
  | 
| D | bio-integrity.c | 271 	q = bdev_get_queue(bio->bi_bdev);  in bio_integrity_prep()
  | 
| D | compat_ioctl.c | 722 				    queue_max_sectors(bdev_get_queue(bdev)));  in compat_blkdev_ioctl() 726 					 !blk_queue_nonrot(bdev_get_queue(bdev)));  in compat_blkdev_ioctl()
  | 
| D | blk-core.c | 91 	struct request_queue *q = bdev_get_queue(bdev);  in blk_get_backing_dev_info() 1705 		trace_block_bio_remap(bdev_get_queue(bio->bi_bdev), bio,  in blk_partition_remap() 1803 	q = bdev_get_queue(bio->bi_bdev);  in generic_make_request_checks() 1945 		struct request_queue *q = bdev_get_queue(bio->bi_bdev);  in generic_make_request()
  | 
| D | blk-settings.c | 668 	struct request_queue *bq = bdev_get_queue(bdev);  in bdev_stack_limits()
  | 
| D | bio.c | 690 	struct request_queue *q = bdev_get_queue(bdev);  in bio_get_nr_vecs() 868 	struct request_queue *q = bdev_get_queue(bio->bi_bdev);  in bio_add_page()
  | 
| /linux-4.1.27/include/linux/ | 
| D | blkdev.h | 863 static inline struct request_queue *bdev_get_queue(struct block_device *bdev)  in bdev_get_queue()  function 1236 	return queue_logical_block_size(bdev_get_queue(bdev));  in bdev_logical_block_size() 1246 	return queue_physical_block_size(bdev_get_queue(bdev));  in bdev_physical_block_size() 1256 	return queue_io_min(bdev_get_queue(bdev));  in bdev_io_min() 1266 	return queue_io_opt(bdev_get_queue(bdev));  in bdev_io_opt() 1287 	struct request_queue *q = bdev_get_queue(bdev);  in bdev_alignment_offset() 1331 	struct request_queue *q = bdev_get_queue(bdev);  in bdev_discard_alignment() 1349 	return queue_discard_zeroes_data(bdev_get_queue(bdev));  in bdev_discard_zeroes_data() 1354 	struct request_queue *q = bdev_get_queue(bdev);  in bdev_write_same()
  | 
| /linux-4.1.27/fs/f2fs/ | 
| D | gc.h | 107 	struct request_queue *q = bdev_get_queue(bdev);  in is_idle()
  | 
| D | segment.h | 703 	struct request_queue *q = bdev_get_queue(bdev);  in max_hw_blocks()
  | 
| D | file.c | 1086 	struct request_queue *q = bdev_get_queue(sb->s_bdev);  in f2fs_ioc_fitrim()
  | 
| D | super.c | 1152 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in f2fs_fill_super()
  | 
| /linux-4.1.27/drivers/target/ | 
| D | target_core_iblock.c | 124 	q = bdev_get_queue(bd);  in iblock_configure_device() 691 		struct request_queue *q = bdev_get_queue(ib_dev->ibd_bd);  in iblock_execute_rw() 800 	struct request_queue *q = bdev_get_queue(bd);  in iblock_get_blocks() 862 	struct request_queue *q = bdev_get_queue(bd);  in iblock_get_write_cache()
  | 
| D | target_core_file.c | 153 		struct request_queue *q = bdev_get_queue(inode->i_bdev);  in fd_configure_device()
  | 
| /linux-4.1.27/fs/jfs/ | 
| D | ioctl.c | 134 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in jfs_ioctl()
  | 
| D | super.c | 376 			struct request_queue *q = bdev_get_queue(sb->s_bdev);  in parse_options() 395 			struct request_queue *q = bdev_get_queue(sb->s_bdev);  in parse_options()
  | 
| /linux-4.1.27/drivers/md/bcache/ | 
| D | request.c | 908 		    blk_queue_discard(bdev_get_queue(dc->bdev)))  in cached_dev_write() 986 		    !blk_queue_discard(bdev_get_queue(dc->bdev)))  in cached_dev_make_request() 1004 	struct request_queue *q = bdev_get_queue(dc->bdev);  in cached_dev_congested() 1015 			q = bdev_get_queue(ca->bdev);  in cached_dev_congested() 1119 		q = bdev_get_queue(ca->bdev);  in flash_dev_congested()
  | 
| D | io.c | 16 	struct request_queue *q = bdev_get_queue(bio->bi_bdev);  in bch_bio_max_sectors()
  | 
| D | sysfs.c | 848 		if (blk_queue_discard(bdev_get_queue(ca->bdev)))  in STORE()
  | 
| D | super.c | 1127 	struct request_queue *q = bdev_get_queue(dc->bdev);  in cached_dev_init() 1892 	if (blk_queue_discard(bdev_get_queue(ca->bdev)))  in register_cache()
  | 
| /linux-4.1.27/fs/xfs/ | 
| D | xfs_discard.c | 159 	struct request_queue	*q = bdev_get_queue(mp->m_ddev_targp->bt_bdev);  in xfs_ioc_trim()
  | 
| /linux-4.1.27/drivers/block/xen-blkback/ | 
| D | xenbus.c | 451 	q = bdev_get_queue(bdev);  in xen_vbd_create() 510 	struct request_queue *q = bdev_get_queue(bdev);  in xen_blkbk_discard()
  | 
| /linux-4.1.27/fs/ext4/ | 
| D | ioctl.c | 600 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in ext4_ioctl()
  | 
| D | super.c | 4270 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in ext4_fill_super()
  | 
| /linux-4.1.27/fs/logfs/ | 
| D | super.c | 122 		sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info;  in logfs_sb_set()
  | 
| /linux-4.1.27/kernel/trace/ | 
| D | blktrace.c | 650 	q = bdev_get_queue(bdev);  in blk_trace_ioctl() 1634 	return bdev_get_queue(bdev);  in blk_trace_get_queue()
  | 
| /linux-4.1.27/fs/ocfs2/ | 
| D | ioctl.c | 929 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in ocfs2_ioctl()
  | 
| /linux-4.1.27/fs/btrfs/ | 
| D | volumes.c | 821 		q = bdev_get_queue(bdev);  in __btrfs_open_devices() 829 		if (!blk_queue_nonrot(bdev_get_queue(bdev)))  in __btrfs_open_devices() 2161 	q = bdev_get_queue(bdev);  in btrfs_init_new_device() 2203 	if (!blk_queue_nonrot(bdev_get_queue(bdev)))  in btrfs_init_new_device() 2364 	q = bdev_get_queue(bdev);  in btrfs_init_dev_replace_tgtdev() 5724 	struct request_queue *q = bdev_get_queue(bdev);  in bio_size_ok()
  | 
| D | ioctl.c | 391 		q = bdev_get_queue(device->bdev);  in btrfs_ioctl_fitrim()
  | 
| /linux-4.1.27/fs/nilfs2/ | 
| D | ioctl.c | 1086 	struct request_queue *q = bdev_get_queue(nilfs->ns_bdev);  in nilfs_ioctl_trim_fs()
  | 
| D | super.c | 1079 	sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info;  in nilfs_fill_super()
  | 
| /linux-4.1.27/fs/ | 
| D | super.c | 940 	s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info;  in set_bdev_super()
  | 
| /linux-4.1.27/mm/ | 
| D | swapfile.c | 2337 	struct request_queue *q = bdev_get_queue(si->bdev);  in swap_discardable() 2435 	if (p->bdev && blk_queue_nonrot(bdev_get_queue(p->bdev))) {  in SYSCALL_DEFINE2()
  | 
| /linux-4.1.27/drivers/block/ | 
| D | pktcdvd.c | 701 	struct request_queue *q = bdev_get_queue(pd->bdev);  in pkt_generic_packet() 2212 	q = bdev_get_queue(pd->bdev);  in pkt_open_dev()
  | 
| /linux-4.1.27/fs/gfs2/ | 
| D | ops_fstype.c | 1224 	s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info;  in set_gfs2_super()
  | 
| D | rgrp.c | 1346 	struct request_queue *q = bdev_get_queue(sdp->sd_vfs->s_bdev);  in gfs2_fitrim()
  | 
| /linux-4.1.27/fs/fat/ | 
| D | inode.c | 1745 		struct request_queue *q = bdev_get_queue(sb->s_bdev);  in fat_fill_super()
  | 
| /linux-4.1.27/drivers/block/drbd/ | 
| D | drbd_main.c | 2395 		q = bdev_get_queue(device->ldev->backing_bdev);  in drbd_congested()
  | 
| D | drbd_receiver.c | 2320 		struct request_queue *q = bdev_get_queue(device->ldev->backing_bdev);  in receive_Data()
  |