Home
last modified time | relevance | path

Searched refs:bdev_get_queue (Results 1 – 61 of 61) sorted by relevance

/linux-4.1.27/drivers/md/
Dlinear.c75 subq = bdev_get_queue(dev0->rdev->bdev); in linear_mergeable_bvec()
105 struct request_queue *q = bdev_get_queue(conf->disks[i].rdev->bdev); in linear_congested()
164 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in linear_conf()
298 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) { in linear_make_request()
Ddm-table.c294 q = bdev_get_queue(bdev); in device_area_is_invalid()
424 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits()
889 struct request_queue *q = bdev_get_queue(dd->dm_dev->bdev); in dm_table_set_type()
904 if (!bdev_get_queue(dd->dm_dev->bdev)->mq_ops) { in dm_table_set_type()
1318 struct request_queue *q = bdev_get_queue(dev->bdev); in device_flush_capable()
1370 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_nonrot()
1378 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_not_random()
1386 struct request_queue *q = bdev_get_queue(dev->bdev); in queue_supports_sg_merge()
1394 struct request_queue *q = bdev_get_queue(dev->bdev); in queue_supports_sg_gaps()
1419 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_same_capable()
[all …]
Draid0.c36 struct request_queue *q = bdev_get_queue(devlist[i]->bdev); in raid0_congested()
382 subq = bdev_get_queue(rdev->bdev); in raid0_mergeable_bvec()
445 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid0_run()
547 !blk_queue_discard(bdev_get_queue(split->bi_bdev)))) { in raid0_make_request()
Ddm-mpath.c167 scsi_dh_detach(bdev_get_queue(pgpath->path.dev->bdev)); in free_pgpaths()
425 clone->q = bdev_get_queue(bdev); in __multipath_map()
430 *__clone = blk_get_request(bdev_get_queue(bdev), in __multipath_map()
580 q = bdev_get_queue(p->path.dev->bdev); in parse_path()
1240 scsi_dh_activate(bdev_get_queue(pgpath->path.dev->bdev), in activate_path()
1630 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in __pgpath_busy()
Ddm-linear.c137 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in linear_merge()
Draid1.c615 nonrot = blk_queue_nonrot(bdev_get_queue(rdev->bdev)); in read_balance()
727 bdev_get_queue(rdev->bdev); in raid1_mergeable_bvec()
756 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid1_congested()
794 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in flush_pending_writes()
1069 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in raid1_unplug()
1619 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid1_add_disk()
1677 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid1_add_disk()
2811 q = bdev_get_queue(rdev->bdev); in setup_conf()
2924 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in run()
Ddm-log-writes.c735 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in log_writes_merge()
780 struct request_queue *q = bdev_get_queue(lc->dev->bdev); in log_writes_io_hints()
Ddm-flakey.c394 struct request_queue *q = bdev_get_queue(fc->dev->bdev); in flakey_merge()
Ddm-stripe.c425 q = bdev_get_queue(sc->stripe[stripe].dev->bdev); in stripe_merge()
Draid10.c735 bdev_get_queue(rdev->bdev); in raid10_mergeable_bvec()
747 bdev_get_queue(rdev->bdev); in raid10_mergeable_bvec()
928 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid10_congested()
958 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in flush_pending_writes()
1134 !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) in raid10_unplug()
1780 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid10_add_disk()
1850 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid10_add_disk()
3655 q = bdev_get_queue(rdev->bdev); in run()
3672 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in run()
Dmultipath.c167 struct request_queue *q = bdev_get_queue(rdev->bdev); in multipath_congested()
Ddm-io.c289 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
Ddm-thin.c2386 q = bdev_get_queue(pt->data_dev->bdev); in pool_is_congested()
2410 struct request_queue *q = bdev_get_queue(pt->data_dev->bdev); in data_dev_supports_discard()
2428 struct queue_limits *data_limits = &bdev_get_queue(data_bdev)->limits; in disable_passdown_if_not_supported()
3580 struct request_queue *q = bdev_get_queue(pt->data_dev->bdev); in pool_merge()
3601 data_limits = &bdev_get_queue(pt->data_dev->bdev)->limits; in set_discard_limits()
4010 struct request_queue *q = bdev_get_queue(tc->pool_dev->bdev); in thin_merge()
Ddm-era-target.c1382 struct request_queue *q = bdev_get_queue(dev->bdev); in dev_is_congested()
1680 struct request_queue *q = bdev_get_queue(era->origin_dev->bdev); in era_merge()
Draid5.c234 trace_block_bio_complete(bdev_get_queue(bi->bi_bdev), in return_io()
1034 trace_block_bio_remap(bdev_get_queue(bi->bi_bdev), in ops_run_io()
1080 trace_block_bio_remap(bdev_get_queue(rbi->bi_bdev), in ops_run_io()
4778 trace_block_bio_complete(bdev_get_queue(raid_bi->bi_bdev), in raid5_align_endio()
4793 struct request_queue *q = bdev_get_queue(bi->bi_bdev); in bio_fits_rdev()
4884 trace_block_bio_remap(bdev_get_queue(align_bi->bi_bdev), in chunk_aligned_read()
5320 trace_block_bio_complete(bdev_get_queue(bi->bi_bdev), in make_request()
5716 trace_block_bio_complete(bdev_get_queue(raid_bio->bi_bdev), in retry_aligned_read()
6944 if (!blk_queue_discard(bdev_get_queue(rdev->bdev)) || in run()
6945 !bdev_get_queue(rdev->bdev)-> in run()
Ddm.c986 !bdev_get_queue(bio->bi_bdev)->limits.max_write_same_sectors)) in clone_endio()
1451 trace_block_bio_remap(bdev_get_queue(clone->bi_bdev), clone, in __map_bio()
2501 struct request_queue *q = bdev_get_queue(bdev); in dm_device_merge_is_compulsory()
Ddm-verity.c655 struct request_queue *q = bdev_get_queue(v->data_dev->bdev); in verity_merge()
Ddm-raid.c1173 q = bdev_get_queue(rs->dev[i].rdev.bdev); in configure_discard_support()
Ddm-cache-target.c1886 struct request_queue *q = bdev_get_queue(dev->bdev); in is_congested()
3321 struct request_queue *q = bdev_get_queue(cache->origin_dev->bdev); in cache_bvec_merge()
Ddm-snap.c2328 struct request_queue *q = bdev_get_queue(o->dev->bdev); in origin_merge()
Ddm-crypt.c2025 struct request_queue *q = bdev_get_queue(cc->dev->bdev); in crypt_merge()
/linux-4.1.27/block/
Dblk-lib.c44 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_discard()
160 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_write_same()
304 struct request_queue *q = bdev_get_queue(bdev); in blkdev_issue_zeroout()
Dioctl.c378 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_ioctl()
381 return put_ushort(arg, !blk_queue_nonrot(bdev_get_queue(bdev))); in blkdev_ioctl()
Dblk-flush.c459 q = bdev_get_queue(bdev); in blkdev_issue_flush()
Dbio-integrity.c271 q = bdev_get_queue(bio->bi_bdev); in bio_integrity_prep()
Dcompat_ioctl.c722 queue_max_sectors(bdev_get_queue(bdev))); in compat_blkdev_ioctl()
726 !blk_queue_nonrot(bdev_get_queue(bdev))); in compat_blkdev_ioctl()
Dblk-core.c91 struct request_queue *q = bdev_get_queue(bdev); in blk_get_backing_dev_info()
1705 trace_block_bio_remap(bdev_get_queue(bio->bi_bdev), bio, in blk_partition_remap()
1803 q = bdev_get_queue(bio->bi_bdev); in generic_make_request_checks()
1945 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in generic_make_request()
Dblk-settings.c668 struct request_queue *bq = bdev_get_queue(bdev); in bdev_stack_limits()
Dbio.c690 struct request_queue *q = bdev_get_queue(bdev); in bio_get_nr_vecs()
868 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_add_page()
/linux-4.1.27/include/linux/
Dblkdev.h863 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function
1236 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()
1246 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()
1256 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()
1266 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()
1287 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
1331 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
1349 return queue_discard_zeroes_data(bdev_get_queue(bdev)); in bdev_discard_zeroes_data()
1354 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
/linux-4.1.27/fs/f2fs/
Dgc.h107 struct request_queue *q = bdev_get_queue(bdev); in is_idle()
Dsegment.h703 struct request_queue *q = bdev_get_queue(bdev); in max_hw_blocks()
Dfile.c1086 struct request_queue *q = bdev_get_queue(sb->s_bdev); in f2fs_ioc_fitrim()
Dsuper.c1152 struct request_queue *q = bdev_get_queue(sb->s_bdev); in f2fs_fill_super()
/linux-4.1.27/drivers/target/
Dtarget_core_iblock.c124 q = bdev_get_queue(bd); in iblock_configure_device()
691 struct request_queue *q = bdev_get_queue(ib_dev->ibd_bd); in iblock_execute_rw()
800 struct request_queue *q = bdev_get_queue(bd); in iblock_get_blocks()
862 struct request_queue *q = bdev_get_queue(bd); in iblock_get_write_cache()
Dtarget_core_file.c153 struct request_queue *q = bdev_get_queue(inode->i_bdev); in fd_configure_device()
/linux-4.1.27/fs/jfs/
Dioctl.c134 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl()
Dsuper.c376 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options()
395 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options()
/linux-4.1.27/drivers/md/bcache/
Drequest.c908 blk_queue_discard(bdev_get_queue(dc->bdev))) in cached_dev_write()
986 !blk_queue_discard(bdev_get_queue(dc->bdev))) in cached_dev_make_request()
1004 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_congested()
1015 q = bdev_get_queue(ca->bdev); in cached_dev_congested()
1119 q = bdev_get_queue(ca->bdev); in flash_dev_congested()
Dio.c16 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bch_bio_max_sectors()
Dsysfs.c848 if (blk_queue_discard(bdev_get_queue(ca->bdev))) in STORE()
Dsuper.c1127 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init()
1892 if (blk_queue_discard(bdev_get_queue(ca->bdev))) in register_cache()
/linux-4.1.27/fs/xfs/
Dxfs_discard.c159 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim()
/linux-4.1.27/drivers/block/xen-blkback/
Dxenbus.c451 q = bdev_get_queue(bdev); in xen_vbd_create()
510 struct request_queue *q = bdev_get_queue(bdev); in xen_blkbk_discard()
/linux-4.1.27/fs/ext4/
Dioctl.c600 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ext4_ioctl()
Dsuper.c4270 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ext4_fill_super()
/linux-4.1.27/fs/logfs/
Dsuper.c122 sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info; in logfs_sb_set()
/linux-4.1.27/kernel/trace/
Dblktrace.c650 q = bdev_get_queue(bdev); in blk_trace_ioctl()
1634 return bdev_get_queue(bdev); in blk_trace_get_queue()
/linux-4.1.27/fs/ocfs2/
Dioctl.c929 struct request_queue *q = bdev_get_queue(sb->s_bdev); in ocfs2_ioctl()
/linux-4.1.27/fs/btrfs/
Dvolumes.c821 q = bdev_get_queue(bdev); in __btrfs_open_devices()
829 if (!blk_queue_nonrot(bdev_get_queue(bdev))) in __btrfs_open_devices()
2161 q = bdev_get_queue(bdev); in btrfs_init_new_device()
2203 if (!blk_queue_nonrot(bdev_get_queue(bdev))) in btrfs_init_new_device()
2364 q = bdev_get_queue(bdev); in btrfs_init_dev_replace_tgtdev()
5724 struct request_queue *q = bdev_get_queue(bdev); in bio_size_ok()
Dioctl.c391 q = bdev_get_queue(device->bdev); in btrfs_ioctl_fitrim()
/linux-4.1.27/fs/nilfs2/
Dioctl.c1086 struct request_queue *q = bdev_get_queue(nilfs->ns_bdev); in nilfs_ioctl_trim_fs()
Dsuper.c1079 sb->s_bdi = &bdev_get_queue(sb->s_bdev)->backing_dev_info; in nilfs_fill_super()
/linux-4.1.27/fs/
Dsuper.c940 s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info; in set_bdev_super()
/linux-4.1.27/mm/
Dswapfile.c2337 struct request_queue *q = bdev_get_queue(si->bdev); in swap_discardable()
2435 if (p->bdev && blk_queue_nonrot(bdev_get_queue(p->bdev))) { in SYSCALL_DEFINE2()
/linux-4.1.27/drivers/block/
Dpktcdvd.c701 struct request_queue *q = bdev_get_queue(pd->bdev); in pkt_generic_packet()
2212 q = bdev_get_queue(pd->bdev); in pkt_open_dev()
/linux-4.1.27/fs/gfs2/
Dops_fstype.c1224 s->s_bdi = &bdev_get_queue(s->s_bdev)->backing_dev_info; in set_gfs2_super()
Drgrp.c1346 struct request_queue *q = bdev_get_queue(sdp->sd_vfs->s_bdev); in gfs2_fitrim()
/linux-4.1.27/fs/fat/
Dinode.c1745 struct request_queue *q = bdev_get_queue(sb->s_bdev); in fat_fill_super()
/linux-4.1.27/drivers/block/drbd/
Ddrbd_main.c2395 q = bdev_get_queue(device->ldev->backing_bdev); in drbd_congested()
Ddrbd_receiver.c2320 struct request_queue *q = bdev_get_queue(device->ldev->backing_bdev); in receive_Data()