Home
last modified time | relevance | path

Searched refs:SECTOR_SHIFT (Results 1 – 36 of 36) sorted by relevance

/linux-4.1.27/fs/nfs/blocklayout/
Dblocklayout.c158 disk_addr = (u64)isect << SECTOR_SHIFT; in do_add_page_to_bio()
175 disk_addr >> SECTOR_SHIFT, end_io, par); in do_add_page_to_bio()
251 isect = (sector_t) (f_offset >> SECTOR_SHIFT); in bl_read_pagelist()
277 isect += (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist()
278 extent_length -= (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist()
301 isect += (pg_len >> SECTOR_SHIFT); in bl_read_pagelist()
302 extent_length -= (pg_len >> SECTOR_SHIFT); in bl_read_pagelist()
306 if ((isect << SECTOR_SHIFT) >= header->inode->i_size) { in bl_read_pagelist()
310 header->res.count = (isect << SECTOR_SHIFT) - header->args.offset; in bl_read_pagelist()
351 ext_tree_mark_written(bl, start >> SECTOR_SHIFT, in bl_write_cleanup()
[all …]
Dblocklayout.h43 #define PAGE_CACHE_SECTORS (PAGE_CACHE_SIZE >> SECTOR_SHIFT)
44 #define PAGE_CACHE_SECTOR_SHIFT (PAGE_CACHE_SHIFT - SECTOR_SHIFT)
45 #define SECTOR_SIZE (1 << SECTOR_SHIFT)
Dextent_tree.c500 p = xdr_encode_hyper(p, be->be_f_offset << SECTOR_SHIFT); in ext_tree_encode_commit()
501 p = xdr_encode_hyper(p, be->be_length << SECTOR_SHIFT); in ext_tree_encode_commit()
/linux-4.1.27/drivers/block/zram/
Dzram_drv.h45 #define SECTOR_SHIFT 9 macro
46 #define SECTORS_PER_PAGE_SHIFT (PAGE_SHIFT - SECTOR_SHIFT)
51 (1 << (ZRAM_LOGICAL_BLOCK_SHIFT - SECTOR_SHIFT))
Dzram_drv.c336 end = start + (size >> SECTOR_SHIFT); in valid_io_request()
337 bound = zram->disksize >> SECTOR_SHIFT; in valid_io_request()
712 generic_start_io_acct(rw, bvec->bv_len >> SECTOR_SHIFT, in zram_bvec_rw()
856 set_capacity(zram->disk, zram->disksize >> SECTOR_SHIFT); in disksize_store()
931 (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT; in __zram_make_request()
1035 offset = sector & (SECTORS_PER_PAGE - 1) << SECTOR_SHIFT; in zram_rw_page()
/linux-4.1.27/drivers/block/
Dbrd.c25 #define SECTOR_SHIFT 9 macro
26 #define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT)
198 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd_setup()
205 sector += copy >> SECTOR_SHIFT; in copy_to_brd_setup()
225 sector += PAGE_SIZE >> SECTOR_SHIFT; in discard_from_brd()
238 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd()
251 sector += copy >> SECTOR_SHIFT; in copy_to_brd()
270 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_from_brd()
284 sector += copy >> SECTOR_SHIFT; in copy_from_brd()
356 sector += len >> SECTOR_SHIFT; in brd_make_request()
Drbd.c58 #define SECTOR_SHIFT 9 macro
59 #define SECTOR_SIZE (1ULL << SECTOR_SHIFT)
879 if (ondisk->options.order < SECTOR_SHIFT) in rbd_dev_ondisk_valid()
2447 bio_list->bi_iter.bi_sector << SECTOR_SHIFT); in rbd_img_request_fill()
3338 u64 offset = (u64)blk_rq_pos(rq) << SECTOR_SHIFT; in rbd_queue_workfn()
3478 sectors_per_obj = 1 << (rbd_dev->header.obj_order - SECTOR_SHIFT); in rbd_merge_bvec()
3485 ret = (int) (sectors_per_obj - obj_sector_offset) << SECTOR_SHIFT; in rbd_merge_bvec()
/linux-4.1.27/drivers/mtd/
Dssfdc.c39 #define SECTOR_SHIFT 9 macro
136 cis_sector = (int)(offset >> SECTOR_SHIFT); in get_valid_cis_sector()
157 loff_t offset = (loff_t)sect_no << SECTOR_SHIFT; in read_physical_sector()
311 ssfdc->cis_block = cis_sector / (mtd->erasesize >> SECTOR_SHIFT); in ssfdcr_add_mtd()
323 ssfdc->cylinders = (unsigned short)(((u32)mtd->size >> SECTOR_SHIFT) / in ssfdcr_add_mtd()
375 sectors_per_block = ssfdc->erase_size >> SECTOR_SHIFT; in ssfdcr_readsect()
/linux-4.1.27/arch/xtensa/platforms/iss/
Dsimdisk.c24 #define SECTOR_SHIFT 9 macro
75 unsigned long offset = sector << SECTOR_SHIFT; in simdisk_transfer()
76 unsigned long nbytes = nsect << SECTOR_SHIFT; in simdisk_transfer()
112 unsigned len = bvec.bv_len >> SECTOR_SHIFT; in simdisk_xfer_bio()
180 set_capacity(dev->gd, dev->size >> SECTOR_SHIFT); in simdisk_attach()
/linux-4.1.27/drivers/md/
Ddm-verity.c555 ((1 << (v->data_dev_block_bits - SECTOR_SHIFT)) - 1)) { in verity_map()
561 (v->data_dev_block_bits - SECTOR_SHIFT) > v->data_blocks) { in verity_map()
573 io->block = bio->bi_iter.bi_sector >> (v->data_dev_block_bits - SECTOR_SHIFT); in verity_map()
644 ti->len != i_size_read(v->data_dev->bdev->bd_inode) >> SECTOR_SHIFT) in verity_ioctl()
808 (sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT)) in verity_ctr()
809 >> (v->data_dev_block_bits - SECTOR_SHIFT) != num_ll) { in verity_ctr()
816 if (ti->len > (v->data_blocks << (v->data_dev_block_bits - SECTOR_SHIFT))) { in verity_ctr()
823 (sector_t)(num_ll << (v->hash_dev_block_bits - SECTOR_SHIFT)) in verity_ctr()
824 >> (v->hash_dev_block_bits - SECTOR_SHIFT) != num_ll) { in verity_ctr()
Ddm-snap-persistent.c171 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area()
299 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area()
389 memset(ps->header_area, 0, ps->store->chunk_size << SECTOR_SHIFT); in write_header()
500 ps->store->chunk_size << SECTOR_SHIFT, in read_exceptions()
545 memcpy(ps->area, area, ps->store->chunk_size << SECTOR_SHIFT); in read_exceptions()
626 ps->exceptions_per_area = (ps->store->chunk_size << SECTOR_SHIFT) / in persistent_read_metadata()
Ddm-log-writes.c305 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block()
338 return i_size_read(lc->logdev->bdev->bd_inode) >> SECTOR_SHIFT; in logdev_last_sector()
441 lc->sectorsize = 1 << SECTOR_SHIFT; in log_writes_ctr()
725 if (ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in log_writes_ioctl()
784 limits->discard_granularity = 1 << SECTOR_SHIFT; in log_writes_io_hints()
785 limits->max_discard_sectors = (UINT_MAX >> SECTOR_SHIFT); in log_writes_io_hints()
Ddm-io.c318 dm_sector_div_up(remaining, (PAGE_SIZE >> SECTOR_SHIFT))); in do_region()
328 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region()
337 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region()
514 r = dp_init(io_req, &dp, (unsigned long)where->count << SECTOR_SHIFT); in dm_io()
Ddm-cache-metadata.h27 #define DM_CACHE_METADATA_MAX_SECTORS_WARNING (16 * (1024 * 1024 * 1024 >> SECTOR_SHIFT))
Ddm-bufio.c156 static struct kmem_cache *dm_bufio_caches[PAGE_SHIFT - SECTOR_SHIFT];
157 static char *dm_bufio_cache_names[PAGE_SHIFT - SECTOR_SHIFT];
564 .count = b->c->block_size >> SECTOR_SHIFT, in use_dmio()
918 (c->sectors_per_block_bits + SECTOR_SHIFT); in __get_memory_limit()
1427 (SECTOR_SHIFT + c->sectors_per_block_bits); in dm_bufio_get_device_size()
1583 BUG_ON(block_size < 1 << SECTOR_SHIFT || in dm_bufio_client_create()
1595 c->sectors_per_block_bits = ffs(block_size) - 1 - SECTOR_SHIFT; in dm_bufio_client_create()
Ddm-thin-metadata.h24 #define THIN_METADATA_MAX_SECTORS_WARNING (16 * (1024 * 1024 * 1024 >> SECTOR_SHIFT))
Ddm-exception-store.h193 return i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size()
Ddm-table.c284 i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in device_area_is_invalid()
286 limits->logical_block_size >> SECTOR_SHIFT; in device_area_is_invalid()
441 (unsigned long long) start << SECTOR_SHIFT); in dm_set_device_limits()
607 limits->logical_block_size >> SECTOR_SHIFT; in validate_hardware_logical_block_alignment()
643 SECTOR_SHIFT) - 1)) in validate_hardware_logical_block_alignment()
Ddm-log.c444 dm_round_up((LOG_OFFSET << SECTOR_SHIFT) + bitset_size, in create_log_context()
455 lc->header_location.count = buf_size >> SECTOR_SHIFT; in create_log_context()
477 (LOG_OFFSET << SECTOR_SHIFT); in create_log_context()
Ddm-linear.c127 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in linear_ioctl()
Ddm-exception-store.c179 if (chunk_size > INT_MAX >> SECTOR_SHIFT) { in dm_exception_store_set_chunk_size()
Ddm-era-target.c499 disk->metadata_block_size = cpu_to_le32(DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT); in prepare_superblock()
1617 (unsigned) (DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT), in era_status()
1666 return i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size()
1693 uint64_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in era_io_hints()
1702 blk_limits_io_opt(limits, era->sectors_per_block << SECTOR_SHIFT); in era_io_hints()
Ddm-thin.c44 #define DATA_DEV_BLOCK_SIZE_MIN_SECTORS (64 * 1024 >> SECTOR_SHIFT)
45 #define DATA_DEV_BLOCK_SIZE_MAX_SECTORS (1024 * 1024 * 1024 >> SECTOR_SHIFT)
924 (pool->sectors_per_block << SECTOR_SHIFT); in io_overlaps_block()
1580 bio->bi_iter.bi_size = (tc->origin_size - bio->bi_iter.bi_sector) << SECTOR_SHIFT; in process_cell()
2429 sector_t block_size = pool->sectors_per_block << SECTOR_SHIFT; in disable_passdown_if_not_supported()
2778 return i_size_read(bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size()
3603 pool->sectors_per_block << SECTOR_SHIFT); in set_discard_limits()
3605 limits->discard_granularity = pool->sectors_per_block << SECTOR_SHIFT; in set_discard_limits()
3612 sector_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in pool_io_hints()
3638 blk_limits_io_min(limits, limits->max_sectors << SECTOR_SHIFT); in pool_io_hints()
[all …]
Ddm-crypt.c710 for (i = 0; i < ((1 << SECTOR_SHIFT) / 8); i++) in crypt_iv_tcw_whitening()
852 sg_set_page(&dmreq->sg_in, bv_in.bv_page, 1 << SECTOR_SHIFT, in crypt_convert_block()
856 sg_set_page(&dmreq->sg_out, bv_out.bv_page, 1 << SECTOR_SHIFT, in crypt_convert_block()
859 bio_advance_iter(ctx->bio_in, &ctx->iter_in, 1 << SECTOR_SHIFT); in crypt_convert_block()
860 bio_advance_iter(ctx->bio_out, &ctx->iter_out, 1 << SECTOR_SHIFT); in crypt_convert_block()
869 1 << SECTOR_SHIFT, iv); in crypt_convert_block()
Ddm-cache-target.c107 #define DATA_DEV_BLOCK_SIZE_MIN_SECTORS (32 * 1024 >> SECTOR_SHIFT)
108 #define DATA_DEV_BLOCK_SIZE_MAX_SECTORS (1024 * 1024 * 1024 >> SECTOR_SHIFT)
1108 (bio->bi_iter.bi_size == (cache->sectors_per_block << SECTOR_SHIFT)); in bio_writes_complete_block()
1962 return i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT; in get_dev_size()
3337 limits->discard_granularity = cache->discard_block_size << SECTOR_SHIFT; in set_discard_limits()
3343 uint64_t io_opt_sectors = limits->io_opt >> SECTOR_SHIFT; in cache_io_hints()
3351 blk_limits_io_min(limits, cache->sectors_per_block << SECTOR_SHIFT); in cache_io_hints()
3352 blk_limits_io_opt(limits, cache->sectors_per_block << SECTOR_SHIFT); in cache_io_hints()
Ddm-flakey.c384 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT) in flakey_ioctl()
Ddm-stripe.c409 unsigned chunk_size = sc->chunk_size << SECTOR_SHIFT; in stripe_io_hints()
Ddm.c1421 unsigned bi_size = bio->bi_iter.bi_size >> SECTOR_SHIFT; in dm_accept_partial_bio()
1426 bio->bi_iter.bi_size = n_sectors << SECTOR_SHIFT; in dm_accept_partial_bio()
1738 max_size = (max_sectors << SECTOR_SHIFT) - bvm->bi_size; in dm_merge_bvec()
1764 if (max_size <= biovec->bv_len && !(bvm->bi_size >> SECTOR_SHIFT)) in dm_merge_bvec()
2470 i_size_write(md->bdev->bd_inode, (loff_t)size << SECTOR_SHIFT); in __set_size()
Ddm-switch.c531 if (ti->len + sctx->path_list[path_nr].start != i_size_read(bdev->bd_inode) >> SECTOR_SHIFT) in switch_ioctl()
Ddm-cache-metadata.c298 sector_t bdev_size = i_size_read(cmd->bdev->bd_inode) >> SECTOR_SHIFT; in __write_initial_superblock()
484 cmd->bm = dm_block_manager_create(cmd->bdev, DM_CACHE_METADATA_BLOCK_SIZE << SECTOR_SHIFT, in __create_persistent_data_objects()
Ddm-kcopyd.c34 #define RESERVE_PAGES (DIV_ROUND_UP(SUB_JOB_SIZE << SECTOR_SHIFT, PAGE_SIZE))
Ddm-thin-metadata.c478 sector_t bdev_size = i_size_read(pmd->bdev->bd_inode) >> SECTOR_SHIFT; in __write_initial_superblock()
685 pmd->bm = dm_block_manager_create(pmd->bdev, THIN_METADATA_BLOCK_SIZE << SECTOR_SHIFT, in __create_persistent_data_objects()
Ddm-mpath.c1587 if (!bdev || ti->len != i_size_read(bdev->bd_inode) >> SECTOR_SHIFT) { in multipath_ioctl()
Ddm-snap.c1736 (s->store->chunk_size << SECTOR_SHIFT)) { in snapshot_map()
/linux-4.1.27/drivers/md/persistent-data/
Ddm-space-map-metadata.h12 #define DM_SM_METADATA_BLOCK_SIZE (4096 >> SECTOR_SHIFT)
/linux-4.1.27/include/linux/
Ddevice-mapper.h548 #define SECTOR_SHIFT 9 macro
600 return (n >> SECTOR_SHIFT); in to_sector()
605 return (n << SECTOR_SHIFT); in to_bytes()