/linux-4.1.27/fs/nfs/blocklayout/ |
D | blocklayout.c | 158 disk_addr = (u64)isect << SECTOR_SHIFT; in do_add_page_to_bio() 175 disk_addr >> SECTOR_SHIFT, end_io, par); in do_add_page_to_bio() 251 isect = (sector_t) (f_offset >> SECTOR_SHIFT); in bl_read_pagelist() 277 isect += (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist() 278 extent_length -= (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist() 301 isect += (pg_len >> SECTOR_SHIFT); in bl_read_pagelist() 302 extent_length -= (pg_len >> SECTOR_SHIFT); in bl_read_pagelist() 306 if ((isect << SECTOR_SHIFT) >= header->inode->i_size) { in bl_read_pagelist() 310 header->res.count = (isect << SECTOR_SHIFT) - header->args.offset; in bl_read_pagelist() 351 ext_tree_mark_written(bl, start >> SECTOR_SHIFT, in bl_write_cleanup() [all …]
|
D | blocklayout.h | 43 #define PAGE_CACHE_SECTORS (PAGE_CACHE_SIZE >> SECTOR_SHIFT) 44 #define PAGE_CACHE_SECTOR_SHIFT (PAGE_CACHE_SHIFT - SECTOR_SHIFT) 45 #define SECTOR_SIZE (1 << SECTOR_SHIFT)
|
D | extent_tree.c | 500 p = xdr_encode_hyper(p, be->be_f_offset << SECTOR_SHIFT); in ext_tree_encode_commit() 501 p = xdr_encode_hyper(p, be->be_length << SECTOR_SHIFT); in ext_tree_encode_commit()
|
/linux-4.1.27/drivers/block/zram/ |
D | zram_drv.h | 45 #define SECTOR_SHIFT 9 macro 46 #define SECTORS_PER_PAGE_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) 51 (1 << (ZRAM_LOGICAL_BLOCK_SHIFT - SECTOR_SHIFT))
|
D | zram_drv.c | 336 end = start + (size >> SECTOR_SHIFT); in valid_io_request() 337 bound = zram->disksize >> SECTOR_SHIFT; in valid_io_request() 712 generic_start_io_acct(rw, bvec->bv_len >> SECTOR_SHIFT, in zram_bvec_rw() 856 set_capacity(zram->disk, zram->disksize >> SECTOR_SHIFT); in disksize_store() 931 (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT; in __zram_make_request() 1035 offset = sector & (SECTORS_PER_PAGE - 1) << SECTOR_SHIFT; in zram_rw_page()
|
/linux-4.1.27/drivers/block/ |
D | brd.c | 25 #define SECTOR_SHIFT 9 macro 26 #define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) 198 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd_setup() 205 sector += copy >> SECTOR_SHIFT; in copy_to_brd_setup() 225 sector += PAGE_SIZE >> SECTOR_SHIFT; in discard_from_brd() 238 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd() 251 sector += copy >> SECTOR_SHIFT; in copy_to_brd() 270 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_from_brd() 284 sector += copy >> SECTOR_SHIFT; in copy_from_brd() 356 sector += len >> SECTOR_SHIFT; in brd_make_request()
|
D | rbd.c | 58 #define SECTOR_SHIFT 9 macro 59 #define SECTOR_SIZE (1ULL << SECTOR_SHIFT) 879 if (ondisk->options.order < SECTOR_SHIFT) in rbd_dev_ondisk_valid() 2447 bio_list->bi_iter.bi_sector << SECTOR_SHIFT); in rbd_img_request_fill() 3338 u64 offset = (u64)blk_rq_pos(rq) << SECTOR_SHIFT; in rbd_queue_workfn() 3478 sectors_per_obj = 1 << (rbd_dev->header.obj_order - SECTOR_SHIFT); in rbd_merge_bvec() 3485 ret = (int) (sectors_per_obj - obj_sector_offset) << SECTOR_SHIFT; in rbd_merge_bvec()
|
/linux-4.1.27/drivers/mtd/ |
D | ssfdc.c | 39 #define SECTOR_SHIFT 9 macro 136 cis_sector = (int)(offset >> SECTOR_SHIFT); in get_valid_cis_sector() 157 loff_t offset = (loff_t)sect_no << SECTOR_SHIFT; in read_physical_sector() 311 ssfdc->cis_block = cis_sector / (mtd->erasesize >> SECTOR_SHIFT); in ssfdcr_add_mtd() 323 ssfdc->cylinders = (unsigned short)(((u32)mtd->size >> SECTOR_SHIFT) / in ssfdcr_add_mtd() 375 sectors_per_block = ssfdc->erase_size >> SECTOR_SHIFT; in ssfdcr_readsect()
|
/linux-4.1.27/arch/xtensa/platforms/iss/ |
D | simdisk.c | 24 #define SECTOR_SHIFT 9 macro 75 unsigned long offset = sector << SECTOR_SHIFT; in simdisk_transfer() 76 unsigned long nbytes = nsect << SECTOR_SHIFT; in simdisk_transfer() 112 unsigned len = bvec.bv_len >> SECTOR_SHIFT; in simdisk_xfer_bio() 180 set_capacity(dev->gd, dev->size >> SECTOR_SHIFT); in simdisk_attach()
|
/linux-4.1.27/drivers/md/ |
D | dm-verity.c | 555 ((1 << (v->data_dev_block_bits - SECTOR_SHIFT)) - 1)) { in verity_map() 561 (v->data_dev_block_bits - SECTOR_SHIFT) > v->data_blocks) { in verity_map() 573 io->block = bio->bi_iter.bi_sector >> (v->data_dev_block_bits - SECTOR_SHIFT); in verity_map() 644 ti->len != i_size_read(v->data_dev->bdev->bd_inode) >> SECTOR_SHIFT) in verity_ioctl() 808 (sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT)) in verity_ctr() 809 >> (v->data_dev_block_bits - SECTOR_SHIFT) != num_ll) { in verity_ctr() 816 if (ti->len > (v->data_blocks << (v->data_dev_block_bits - SECTOR_SHIFT))) { in verity_ctr() 823 (sector_t)(num_ll << (v->hash_dev_block_bits - SECTOR_SHIFT)) in verity_ctr() 824 >> (v->hash_dev_block_bits - SECTOR_SHIFT) != num_ll) { in verity_ctr()
|
D | dm-snap-persistent.c | 171 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area() 299 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area() 389 memset(ps->header_area, 0, ps->store->chunk_size << SECTOR_SHIFT); in write_header() 500 ps->store->chunk_size << SECTOR_SHIFT, in read_exceptions() 545 memcpy(ps->area, area, ps->store->chunk_size << SECTOR_SHIFT); in read_exceptions() 626 ps->exceptions_per_area = (ps->store->chunk_size << SECTOR_SHIFT) / in persistent_read_metadata()
|
D | dm-log-writes.c | 305 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 338 return i_size_read(lc->logdev->bdev->bd_inode) >> SECTOR_SHIFT; in logdev_last_sector() 441 lc->sectorsize = 1 << SECTOR_SHIFT; in log_writes_ctr() 725 if (ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in log_writes_ioctl() 784 limits->discard_granularity = 1 << SECTOR_SHIFT; in log_writes_io_hints() 785 limits->max_discard_sectors = (UINT_MAX >> SECTOR_SHIFT); in log_writes_io_hints()
|
D | dm-io.c | 318 dm_sector_div_up(remaining, (PAGE_SIZE >> SECTOR_SHIFT))); in do_region() 328 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region() 337 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region() 514 r = dp_init(io_req, &dp, (unsigned long)where->count << SECTOR_SHIFT); in dm_io()
|
D | dm-cache-metadata.h | 27 #define DM_CACHE_METADATA_MAX_SECTORS_WARNING (16 * (1024 * 1024 * 1024 >> SECTOR_SHIFT))
|
D | dm-bufio.c | 156 static struct kmem_cache *dm_bufio_caches[PAGE_SHIFT - SECTOR_SHIFT]; 157 static char *dm_bufio_cache_names[PAGE_SHIFT - SECTOR_SHIFT]; 564 .count = b->c->block_size >> SECTOR_SHIFT, in use_dmio() 918 (c->sectors_per_block_bits + SECTOR_SHIFT); in __get_memory_limit() 1427 (SECTOR_SHIFT + c->sectors_per_block_bits); in dm_bufio_get_device_size() 1583 BUG_ON(block_size < 1 << SECTOR_SHIFT || in dm_bufio_client_create() 1595 c->sectors_per_block_bits = ffs(block_size) - 1 - SECTOR_SHIFT; in dm_bufio_client_create()
|
D | dm-thin-metadata.h | 24 #define THIN_METADATA_MAX_SECTORS_WARNING (16 * (1024 * 1024 * 1024 >> SECTOR_SHIFT))
|
D | dm-exception-store.h | 193 return i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size()
|
D | dm-table.c | 284 i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in device_area_is_invalid() 286 limits->logical_block_size >> SECTOR_SHIFT; in device_area_is_invalid() 441 (unsigned long long) start << SECTOR_SHIFT); in dm_set_device_limits() 607 limits->logical_block_size >> SECTOR_SHIFT; in validate_hardware_logical_block_alignment() 643 SECTOR_SHIFT) - 1)) in validate_hardware_logical_block_alignment()
|
D | dm-log.c | 444 dm_round_up((LOG_OFFSET << SECTOR_SHIFT) + bitset_size, in create_log_context() 455 lc->header_location.count = buf_size >> SECTOR_SHIFT; in create_log_context() 477 (LOG_OFFSET << SECTOR_SHIFT); in create_log_context()
|
D | dm-linear.c | 127 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in linear_ioctl()
|
D | dm-exception-store.c | 179 if (chunk_size > INT_MAX >> SECTOR_SHIFT) { in dm_exception_store_set_chunk_size()
|
D | dm-era-target.c | 499 disk->metadata_block_size = cpu_to_le32(DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT); in prepare_superblock() 1617 (unsigned) (DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT), in era_status() 1666 return i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size() 1693 uint64_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in era_io_hints() 1702 blk_limits_io_opt(limits, era->sectors_per_block << SECTOR_SHIFT); in era_io_hints()
|
D | dm-thin.c | 44 #define DATA_DEV_BLOCK_SIZE_MIN_SECTORS (64 * 1024 >> SECTOR_SHIFT) 45 #define DATA_DEV_BLOCK_SIZE_MAX_SECTORS (1024 * 1024 * 1024 >> SECTOR_SHIFT) 924 (pool->sectors_per_block << SECTOR_SHIFT); in io_overlaps_block() 1580 bio->bi_iter.bi_size = (tc->origin_size - bio->bi_iter.bi_sector) << SECTOR_SHIFT; in process_cell() 2429 sector_t block_size = pool->sectors_per_block << SECTOR_SHIFT; in disable_passdown_if_not_supported() 2778 return i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size() 3603 pool->sectors_per_block << SECTOR_SHIFT); in set_discard_limits() 3605 limits->discard_granularity = pool->sectors_per_block << SECTOR_SHIFT; in set_discard_limits() 3612 sector_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in pool_io_hints() 3638 blk_limits_io_min(limits, limits->max_sectors << SECTOR_SHIFT); in pool_io_hints() [all …]
|
D | dm-crypt.c | 710 for (i = 0; i < ((1 << SECTOR_SHIFT) / 8); i++) in crypt_iv_tcw_whitening() 852 sg_set_page(&dmreq->sg_in, bv_in.bv_page, 1 << SECTOR_SHIFT, in crypt_convert_block() 856 sg_set_page(&dmreq->sg_out, bv_out.bv_page, 1 << SECTOR_SHIFT, in crypt_convert_block() 859 bio_advance_iter(ctx->bio_in, &ctx->iter_in, 1 << SECTOR_SHIFT); in crypt_convert_block() 860 bio_advance_iter(ctx->bio_out, &ctx->iter_out, 1 << SECTOR_SHIFT); in crypt_convert_block() 869 1 << SECTOR_SHIFT, iv); in crypt_convert_block()
|
D | dm-cache-target.c | 107 #define DATA_DEV_BLOCK_SIZE_MIN_SECTORS (32 * 1024 >> SECTOR_SHIFT) 108 #define DATA_DEV_BLOCK_SIZE_MAX_SECTORS (1024 * 1024 * 1024 >> SECTOR_SHIFT) 1108 (bio->bi_iter.bi_size == (cache->sectors_per_block << SECTOR_SHIFT)); in bio_writes_complete_block() 1962 return i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size() 3337 limits->discard_granularity = cache->discard_block_size << SECTOR_SHIFT; in set_discard_limits() 3343 uint64_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in cache_io_hints() 3351 blk_limits_io_min(limits, cache->sectors_per_block << SECTOR_SHIFT); in cache_io_hints() 3352 blk_limits_io_opt(limits, cache->sectors_per_block << SECTOR_SHIFT); in cache_io_hints()
|
D | dm-flakey.c | 384 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in flakey_ioctl()
|
D | dm-stripe.c | 409 unsigned chunk_size = sc->chunk_size << SECTOR_SHIFT; in stripe_io_hints()
|
D | dm.c | 1421 unsigned bi_size = bio->bi_iter.bi_size >> SECTOR_SHIFT; in dm_accept_partial_bio() 1426 bio->bi_iter.bi_size = n_sectors << SECTOR_SHIFT; in dm_accept_partial_bio() 1738 max_size = (max_sectors << SECTOR_SHIFT) - bvm->bi_size; in dm_merge_bvec() 1764 if (max_size <= biovec->bv_len && !(bvm->bi_size >> SECTOR_SHIFT)) in dm_merge_bvec() 2470 i_size_write(md->bdev->bd_inode, (loff_t)size << SECTOR_SHIFT); in __set_size()
|
D | dm-switch.c | 531 if (ti->len + sctx->path_list[path_nr].start != i_size_read(bdev->bd_inode) >> SECTOR_SHIFT) in switch_ioctl()
|
D | dm-cache-metadata.c | 298 sector_t bdev_size = i_size_read(cmd->bdev->bd_inode) >> SECTOR_SHIFT; in __write_initial_superblock() 484 cmd->bm = dm_block_manager_create(cmd->bdev, DM_CACHE_METADATA_BLOCK_SIZE << SECTOR_SHIFT, in __create_persistent_data_objects()
|
D | dm-kcopyd.c | 34 #define RESERVE_PAGES (DIV_ROUND_UP(SUB_JOB_SIZE << SECTOR_SHIFT, PAGE_SIZE))
|
D | dm-thin-metadata.c | 478 sector_t bdev_size = i_size_read(pmd->bdev->bd_inode) >> SECTOR_SHIFT; in __write_initial_superblock() 685 pmd->bm = dm_block_manager_create(pmd->bdev, THIN_METADATA_BLOCK_SIZE << SECTOR_SHIFT, in __create_persistent_data_objects()
|
D | dm-mpath.c | 1587 if (!bdev || ti->len != i_size_read(bdev->bd_inode) >> SECTOR_SHIFT) { in multipath_ioctl()
|
D | dm-snap.c | 1736 (s->store->chunk_size << SECTOR_SHIFT)) { in snapshot_map()
|
/linux-4.1.27/drivers/md/persistent-data/ |
D | dm-space-map-metadata.h | 12 #define DM_SM_METADATA_BLOCK_SIZE (4096 >> SECTOR_SHIFT)
|
/linux-4.1.27/include/linux/ |
D | device-mapper.h | 548 #define SECTOR_SHIFT 9 macro 600 return (n >> SECTOR_SHIFT); in to_sector() 605 return (n << SECTOR_SHIFT); in to_bytes()
|