Lines Matching refs:chunk_sectors

98 		sector_div(sectors, mddev->chunk_sectors);  in create_strip_zones()
99 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones()
144 if ((mddev->chunk_sectors << 9) % blksize) { in create_strip_zones()
147 mddev->chunk_sectors << 9, blksize); in create_strip_zones()
314 unsigned int chunk_sects = mddev->chunk_sectors; in map_sector()
356 unsigned int chunk_sectors = mddev->chunk_sectors; in raid0_mergeable_bvec() local
362 if (is_power_of_2(chunk_sectors)) in raid0_mergeable_bvec()
363 max = (chunk_sectors - ((sector & (chunk_sectors-1)) in raid0_mergeable_bvec()
366 max = (chunk_sectors - (sector_div(sector, chunk_sectors) in raid0_mergeable_bvec()
402 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size()
414 if (mddev->chunk_sectors == 0) { in raid0_run()
434 blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); in raid0_run()
435 blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); in raid0_run()
436 blk_queue_max_discard_sectors(mddev->queue, mddev->chunk_sectors); in raid0_run()
438 blk_queue_io_min(mddev->queue, mddev->chunk_sectors << 9); in raid0_run()
440 (mddev->chunk_sectors << 9) * mddev->raid_disks); in raid0_run()
472 (mddev->chunk_sectors << 9) / PAGE_SIZE; in raid0_run()
523 unsigned chunk_sects = mddev->chunk_sectors; in raid0_make_request()
557 seq_printf(seq, " %dk chunks", mddev->chunk_sectors / 2); in raid0_status()
586 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid0_takeover_raid45()
626 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid0_takeover_raid10()
669 mddev->chunk_sectors = chunksect; in raid0_takeover_raid1()