Searched refs:chunk_sectors (Results 1 – 12 of 12) sorted by relevance
/linux-4.1.27/drivers/md/ |
D | raid0.c | 98 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 99 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones() 144 if ((mddev->chunk_sectors << 9) % blksize) { in create_strip_zones() 147 mddev->chunk_sectors << 9, blksize); in create_strip_zones() 314 unsigned int chunk_sects = mddev->chunk_sectors; in map_sector() 356 unsigned int chunk_sectors = mddev->chunk_sectors; in raid0_mergeable_bvec() local 362 if (is_power_of_2(chunk_sectors)) in raid0_mergeable_bvec() 363 max = (chunk_sectors - ((sector & (chunk_sectors-1)) in raid0_mergeable_bvec() 366 max = (chunk_sectors - (sector_div(sector, chunk_sectors) in raid0_mergeable_bvec() 402 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size() [all …]
|
D | linear.c | 152 if (mddev->chunk_sectors) { in linear_conf() 154 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 155 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 321 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
|
D | raid5.c | 768 if (!sector_div(tmp_sec, conf->chunk_sectors)) in stripe_add_to_batch_list() 2549 : conf->chunk_sectors; in raid5_compute_sector() 2745 : conf->chunk_sectors; in compute_blocknr() 3060 previous ? conf->prev_chunk_sectors : conf->chunk_sectors; in stripe_set_idx() 4682 unsigned int chunk_sectors = mddev->chunk_sectors; in raid5_mergeable_bvec() local 4692 if (mddev->new_chunk_sectors < mddev->chunk_sectors) in raid5_mergeable_bvec() 4693 chunk_sectors = mddev->new_chunk_sectors; in raid5_mergeable_bvec() 4694 max = (chunk_sectors - ((sector & (chunk_sectors - 1)) + bio_sectors)) << 9; in raid5_mergeable_bvec() 4705 unsigned int chunk_sectors = mddev->chunk_sectors; in in_chunk_boundary() local 4708 if (mddev->new_chunk_sectors < mddev->chunk_sectors) in in_chunk_boundary() [all …]
|
D | dm-raid.c | 358 if (region_size < rs->md.chunk_sectors) { in validate_region_size() 525 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params() 692 if (rs->md.chunk_sectors) in parse_raid_params() 693 max_io_len = rs->md.chunk_sectors; in parse_raid_params() 838 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync() 937 if (le32_to_cpu(sb->stripe_sectors) != mddev->chunk_sectors) { in super_init_validation() 1462 raid_param_cnt, rs->md.chunk_sectors); in raid_status() 1605 unsigned chunk_size = rs->md.chunk_sectors << 9; in raid_io_hints()
|
D | raid5.h | 434 int chunk_sectors; member
|
D | raid10.c | 692 unsigned int chunk_sectors; in raid10_mergeable_bvec() local 696 chunk_sectors = (conf->geo.chunk_mask & conf->prev.chunk_mask) + 1; in raid10_mergeable_bvec() 703 max = (chunk_sectors - ((sector & (chunk_sectors - 1)) in raid10_mergeable_bvec() 1584 seq_printf(seq, " %dK chunks", mddev->chunk_sectors / 2); in status() 3474 chunk = mddev->chunk_sectors; in setup_geo() 3621 chunk_size = mddev->chunk_sectors << 9; in run() 3624 mddev->chunk_sectors); in run() 3745 ((mddev->chunk_sectors << 9) / PAGE_SIZE); in run() 3886 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid10_takeover_raid0() 4547 ((conf->mddev->chunk_sectors << 9) / PAGE_SIZE); in end_reshape() [all …]
|
D | md.c | 1077 mddev->chunk_sectors = sb->chunk_size >> 9; in super_90_validate() 1106 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_90_validate() 1252 sb->chunk_size = mddev->chunk_sectors << 9; in super_90_sync() 1564 mddev->chunk_sectors = le32_to_cpu(sb->chunksize); in super_1_validate() 1622 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_1_validate() 1709 sb->chunksize = cpu_to_le32(mddev->chunk_sectors); in super_1_sync() 3425 mddev->new_chunk_sectors = mddev->chunk_sectors; in level_store() 3448 mddev->chunk_sectors = mddev->new_chunk_sectors; in level_store() 3640 mddev->chunk_sectors != mddev->new_chunk_sectors) in chunk_size_show() 3643 mddev->chunk_sectors << 9); in chunk_size_show() [all …]
|
D | md.h | 252 int chunk_sectors; member
|
D | raid1.c | 3038 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape() 3041 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
|
/linux-4.1.27/drivers/char/ |
D | ps3flash.c | 38 u64 chunk_sectors; member 50 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors() 130 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 163 sector += priv->chunk_sectors; in ps3flash_read() 199 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 238 sector += priv->chunk_sectors; in ps3flash_write() 388 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
|
/linux-4.1.27/block/ |
D | blk-settings.c | 116 lim->chunk_sectors = 0; in blk_set_default_limits() 291 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument 293 BUG_ON(!is_power_of_2(chunk_sectors)); in blk_queue_chunk_sectors() 294 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors()
|
/linux-4.1.27/include/linux/ |
D | blkdev.h | 289 unsigned int chunk_sectors; member 922 if (!q->limits.chunk_sectors) in blk_max_size_offset() 925 return q->limits.chunk_sectors - in blk_max_size_offset() 926 (offset & (q->limits.chunk_sectors - 1)); in blk_max_size_offset() 936 if (!q->limits.chunk_sectors) in blk_rq_get_max_sectors()
|