/linux-4.1.27/drivers/md/ |
D | linear.c | 75 subq = bdev_get_queue(dev0->rdev->bdev); in linear_mergeable_bvec() 105 struct request_queue *q = bdev_get_queue(conf->disks[i].rdev->bdev); in linear_congested() 164 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in linear_conf() 298 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) { in linear_make_request()
|
D | dm-table.c | 294 q = bdev_get_queue(bdev); in device_area_is_invalid() 424 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits() 889 struct request_queue *q = bdev_get_queue(dd->dm_dev->bdev); in dm_table_set_type() 904 if (!bdev_get_queue(dd->dm_dev->bdev)->mq_ops) { in dm_table_set_type() 1318 struct request_queue *q = bdev_get_queue(dev->bdev); in device_flush_capable() 1370 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_nonrot() 1378 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_not_random() 1386 struct request_queue *q = bdev_get_queue(dev->bdev); in queue_supports_sg_merge() 1394 struct request_queue *q = bdev_get_queue(dev->bdev); in queue_supports_sg_gaps() 1419 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_same_capable() [all …]
|
D | raid0.c | 36 struct request_queue *q = bdev_get_queue(devlist[i]->bdev); in raid0_congested() 382 subq = bdev_get_queue(rdev->bdev); in raid0_mergeable_bvec() 445 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid0_run() 547 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) { in raid0_make_request()
|
D | dm-mpath.c | 167 scsi_dh_detach(bdev_get_queue(pgpath->path.dev->bdev)); in free_pgpaths() 425 clone->q = bdev_get_queue(bdev); in __multipath_map() 430 *__clone = blk_get_request(bdev_get_queue(bdev), in __multipath_map() 580 q = bdev_get_queue(p->path.dev->bdev); in parse_path() 1240 scsi_dh_activate(bdev_get_queue(pgpath->path.dev->bdev), in activate_path() 1630 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in __pgpath_busy()
|
D | dm-linear.c | 137 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in linear_merge()
|
D | raid1.c | 615 nonrot = blk_queue_nonrot(bdev_get_queue(rdev->bdev)); in read_balance() 727 bdev_get_queue(rdev->bdev); in raid1_mergeable_bvec() 756 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid1_congested() 794 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in flush_pending_writes() 1069 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in raid1_unplug() 1619 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid1_add_disk() 1677 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid1_add_disk() 2811 q = bdev_get_queue(rdev->bdev); in setup_conf() 2924 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in run()
|
D | dm-log-writes.c | 735 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in log_writes_merge() 780 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in log_writes_io_hints()
|
D | dm-flakey.c | 394 struct request_queue *q = bdev_get_queue(fc->dev->bdev); in flakey_merge()
|
D | dm-stripe.c | 425 q = bdev_get_queue(sc->stripe[stripe].dev->bdev); in stripe_merge()
|
D | raid10.c | 735 bdev_get_queue(rdev->bdev); in raid10_mergeable_bvec() 747 bdev_get_queue(rdev->bdev); in raid10_mergeable_bvec() 928 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid10_congested() 958 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in flush_pending_writes() 1134 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in raid10_unplug() 1780 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid10_add_disk() 1850 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid10_add_disk() 3655 q = bdev_get_queue(rdev->bdev); in run() 3672 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in run()
|
D | multipath.c | 167 struct request_queue *q = bdev_get_queue(rdev->bdev); in multipath_congested()
|
D | dm-io.c | 289 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
|
D | dm-thin.c | 2386 q = bdev_get_queue(pt->data_dev->bdev); in pool_is_congested() 2410 struct request_queue *q = bdev_get_queue(pt->data_dev->bdev); in data_dev_supports_discard() 2428 struct queue_limits *data_limits = &bdev_get_queue(data_bdev)->limits; in disable_passdown_if_not_supported() 3580 struct request_queue *q = bdev_get_queue(pt->data_dev->bdev); in pool_merge() 3601 data_limits = &bdev_get_queue(pt->data_dev->bdev)->limits; in set_discard_limits() 4010 struct request_queue *q = bdev_get_queue(tc->pool_dev->bdev); in thin_merge()
|
D | dm-era-target.c | 1382 struct request_queue *q = bdev_get_queue(dev->bdev); in dev_is_congested() 1680 struct request_queue *q = bdev_get_queue(era->origin_dev->bdev); in era_merge()
|
D | raid5.c | 234 trace_block_bio_complete(bdev_get_queue(bi->bi_bdev), in return_io() 1034 trace_block_bio_remap(bdev_get_queue(bi->bi_bdev), in ops_run_io() 1080 trace_block_bio_remap(bdev_get_queue(rbi->bi_bdev), in ops_run_io() 4778 trace_block_bio_complete(bdev_get_queue(raid_bi->bi_bdev), in raid5_align_endio() 4793 struct request_queue *q = bdev_get_queue(bi->bi_bdev); in bio_fits_rdev() 4884 trace_block_bio_remap(bdev_get_queue(align_bi->bi_bdev), in chunk_aligned_read() 5320 trace_block_bio_complete(bdev_get_queue(bi->bi_bdev), in make_request() 5716 trace_block_bio_complete(bdev_get_queue(raid_bio->bi_bdev), in retry_aligned_read() 6944 if (!blk_queue_discard(bdev_get_queue(rdev->bdev)) || in run() 6945 !bdev_get_queue(rdev->bdev)-> in run()
|
D | dm.c | 986 !bdev_get_queue(bio->bi_bdev)->limits.max_write_same_sectors)) in clone_endio() 1451 trace_block_bio_remap(bdev_get_queue(clone->bi_bdev), clone, in __map_bio() 2501 struct request_queue *q = bdev_get_queue(bdev); in dm_device_merge_is_compulsory()
|
D | dm-verity.c | 655 struct request_queue *q = bdev_get_queue(v->data_dev->bdev); in verity_merge()
|
D | dm-raid.c | 1173 q = bdev_get_queue(rs->dev[i].rdev.bdev); in configure_discard_support()
|
D | dm-cache-target.c | 1886 struct request_queue *q = bdev_get_queue(dev->bdev); in is_congested() 3321 struct request_queue *q = bdev_get_queue(cache->origin_dev->bdev); in cache_bvec_merge()
|
D | dm-snap.c | 2328 struct request_queue *q = bdev_get_queue(o->dev->bdev); in origin_merge()
|
D | dm-crypt.c | 2025 struct request_queue *q = bdev_get_queue(cc->dev->bdev); in crypt_merge()
|
/linux-4.1.27/block/ |
D | blk-lib.c | 44 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_discard() 160 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_write_same() 304 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_zeroout()
|
D | ioctl.c | 378 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_ioctl() 381 return put_ushort(arg, !blk_queue_nonrot(bdev_get_queue(bdev))); in blkdev_ioctl()
|
D | blk-flush.c | 459 q = bdev_get_queue(bdev); in blkdev_issue_flush()
|
D | bio-integrity.c | 271 q = bdev_get_queue(bio->bi_bdev); in bio_integrity_prep()
|
D | compat_ioctl.c | 722 queue_max_sectors(bdev_get_queue(bdev))); in compat_blkdev_ioctl() 726 !blk_queue_nonrot(bdev_get_queue(bdev))); in compat_blkdev_ioctl()
|
D | blk-core.c | 91 struct request_queue *q = bdev_get_queue(bdev); in blk_get_backing_dev_info() 1705 trace_block_bio_remap(bdev_get_queue(bio->bi_bdev), bio, in blk_partition_remap() 1803 q = bdev_get_queue(bio->bi_bdev); in generic_make_request_checks() 1945 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in generic_make_request()
|
D | blk-settings.c | 668 struct request_queue *bq = bdev_get_queue(bdev); in bdev_stack_limits()
|
D | bio.c | 690 struct request_queue *q = bdev_get_queue(bdev); in bio_get_nr_vecs() 868 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_add_page()
|
/linux-4.1.27/include/linux/ |
D | blkdev.h | 863 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function 1236 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size() 1246 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size() 1256 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min() 1266 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt() 1287 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset() 1331 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment() 1349 return queue_discard_zeroes_data(bdev_get_queue(bdev)); in bdev_discard_zeroes_data() 1354 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
|
/linux-4.1.27/fs/f2fs/ |
D | gc.h | 107 struct request_queue *q = bdev_get_queue(bdev); in is_idle()
|
D | segment.h | 703 struct request_queue *q = bdev_get_queue(bdev); in max_hw_blocks()
|
D | file.c | 1086 struct request_queue *q = bdev_get_queue(sb->s_bdev); in f2fs_ioc_fitrim()
|
D | super.c | 1152 struct request_queue *q = bdev_get_queue(sb->s_bdev); in f2fs_fill_super()
|
/linux-4.1.27/drivers/target/ |
D | target_core_iblock.c | 124 q = bdev_get_queue(bd); in iblock_configure_device() 691 struct request_queue *q = bdev_get_queue(ib_dev->ibd_bd); in iblock_execute_rw() 800 struct request_queue *q = bdev_get_queue(bd); in iblock_get_blocks() 862 struct request_queue *q = bdev_get_queue(bd); in iblock_get_write_cache()
|
D | target_core_file.c | 153 struct request_queue *q = bdev_get_queue(inode->i_bdev); in fd_configure_device()
|
/linux-4.1.27/fs/jfs/ |
D | ioctl.c | 134 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl()
|
D | super.c | 376 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options() 395 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options()
|
/linux-4.1.27/drivers/md/bcache/ |
D | request.c | 908 blk_queue_discard(bdev_get_queue(dc->bdev))) in cached_dev_write() 986 !blk_queue_discard(bdev_get_queue(dc->bdev))) in cached_dev_make_request() 1004 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_congested() 1015 q = bdev_get_queue(ca->bdev); in cached_dev_congested() 1119 q = bdev_get_queue(ca->bdev); in flash_dev_congested()
|
D | io.c | 16 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bch_bio_max_sectors()
|
D | sysfs.c | 848 if (blk_queue_discard(bdev_get_queue(ca->bdev))) in STORE()
|
D | super.c | 1127 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init() 1892 if (blk_queue_discard(bdev_get_queue(ca->bdev))) in register_cache()
|
/linux-4.1.27/fs/xfs/ |
D | xfs_discard.c | 159 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim()
|
/linux-4.1.27/drivers/block/xen-blkback/ |
D | xenbus.c | 451 q = bdev_get_queue(bdev); in xen_vbd_create() 510 struct request_queue *q = bdev_get_queue(bdev); in xen_blkbk_discard()
|
/linux-4.1.27/fs/ext4/ |
D | ioctl.c | 600 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ext4_ioctl()
|
D | super.c | 4270 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ext4_fill_super()
|
/linux-4.1.27/fs/logfs/ |
D | super.c | 122 sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info; in logfs_sb_set()
|
/linux-4.1.27/kernel/trace/ |
D | blktrace.c | 650 q = bdev_get_queue(bdev); in blk_trace_ioctl() 1634 return bdev_get_queue(bdev); in blk_trace_get_queue()
|
/linux-4.1.27/fs/ocfs2/ |
D | ioctl.c | 929 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ocfs2_ioctl()
|
/linux-4.1.27/fs/btrfs/ |
D | volumes.c | 821 q = bdev_get_queue(bdev); in __btrfs_open_devices() 829 if (!blk_queue_nonrot(bdev_get_queue(bdev))) in __btrfs_open_devices() 2161 q = bdev_get_queue(bdev); in btrfs_init_new_device() 2203 if (!blk_queue_nonrot(bdev_get_queue(bdev))) in btrfs_init_new_device() 2364 q = bdev_get_queue(bdev); in btrfs_init_dev_replace_tgtdev() 5724 struct request_queue *q = bdev_get_queue(bdev); in bio_size_ok()
|
D | ioctl.c | 391 q = bdev_get_queue(device->bdev); in btrfs_ioctl_fitrim()
|
/linux-4.1.27/fs/nilfs2/ |
D | ioctl.c | 1086 struct request_queue *q = bdev_get_queue(nilfs->ns_bdev); in nilfs_ioctl_trim_fs()
|
D | super.c | 1079 sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info; in nilfs_fill_super()
|
/linux-4.1.27/fs/ |
D | super.c | 940 s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info; in set_bdev_super()
|
/linux-4.1.27/mm/ |
D | swapfile.c | 2337 struct request_queue *q = bdev_get_queue(si->bdev); in swap_discardable() 2435 if (p->bdev && blk_queue_nonrot(bdev_get_queue(p->bdev))) { in SYSCALL_DEFINE2()
|
/linux-4.1.27/drivers/block/ |
D | pktcdvd.c | 701 struct request_queue *q = bdev_get_queue(pd->bdev); in pkt_generic_packet() 2212 q = bdev_get_queue(pd->bdev); in pkt_open_dev()
|
/linux-4.1.27/fs/gfs2/ |
D | ops_fstype.c | 1224 s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info; in set_gfs2_super()
|
D | rgrp.c | 1346 struct request_queue *q = bdev_get_queue(sdp->sd_vfs->s_bdev); in gfs2_fitrim()
|
/linux-4.1.27/fs/fat/ |
D | inode.c | 1745 struct request_queue *q = bdev_get_queue(sb->s_bdev); in fat_fill_super()
|
/linux-4.1.27/drivers/block/drbd/ |
D | drbd_main.c | 2395 q = bdev_get_queue(device->ldev->backing_bdev); in drbd_congested()
|
D | drbd_receiver.c | 2320 struct request_queue *q = bdev_get_queue(device->ldev->backing_bdev); in receive_Data()
|