Lines Matching refs:chunk_sectors
768 if (!sector_div(tmp_sec, conf->chunk_sectors)) in stripe_add_to_batch_list()
2549 : conf->chunk_sectors; in raid5_compute_sector()
2745 : conf->chunk_sectors; in compute_blocknr()
3060 previous ? conf->prev_chunk_sectors : conf->chunk_sectors; in stripe_set_idx()
4682 unsigned int chunk_sectors = mddev->chunk_sectors; in raid5_mergeable_bvec() local
4692 if (mddev->new_chunk_sectors < mddev->chunk_sectors) in raid5_mergeable_bvec()
4693 chunk_sectors = mddev->new_chunk_sectors; in raid5_mergeable_bvec()
4694 max = (chunk_sectors - ((sector & (chunk_sectors - 1)) + bio_sectors)) << 9; in raid5_mergeable_bvec()
4705 unsigned int chunk_sectors = mddev->chunk_sectors; in in_chunk_boundary() local
4708 if (mddev->new_chunk_sectors < mddev->chunk_sectors) in in_chunk_boundary()
4709 chunk_sectors = mddev->new_chunk_sectors; in in_chunk_boundary()
4710 return chunk_sectors >= in in_chunk_boundary()
4711 ((sector & (chunk_sectors - 1)) + bio_sectors); in in_chunk_boundary()
5074 stripe_sectors = conf->chunk_sectors * in make_discard_request()
5080 logical_sector *= conf->chunk_sectors; in make_discard_request()
5081 last_sector *= conf->chunk_sectors; in make_discard_request()
5374 if (mddev->new_chunk_sectors > mddev->chunk_sectors) in reshape_request()
5377 reshape_sectors = mddev->chunk_sectors; in reshape_request()
6260 sectors &= ~((sector_t)mddev->chunk_sectors - 1); in raid5_size()
6281 max(conf->chunk_sectors, in alloc_scratch_buffer()
6384 conf->scribble_sectors = max(conf->chunk_sectors, in raid5_alloc_percpu()
6522 conf->chunk_sectors = mddev->new_chunk_sectors; in setup_conf()
6569 conf->prev_chunk_sectors = mddev->chunk_sectors; in setup_conf()
6712 sector_div(here_old, mddev->chunk_sectors * in run()
6718 here_old * mddev->chunk_sectors)) { in run()
6730 if (abs(min_offset_diff) >= mddev->chunk_sectors && in run()
6742 here_old * mddev->chunk_sectors) in run()
6744 here_old * mddev->chunk_sectors + (-min_offset_diff))) { in run()
6757 BUG_ON(mddev->chunk_sectors != mddev->new_chunk_sectors); in run()
6839 mddev->dev_sectors &= ~(mddev->chunk_sectors - 1); in run()
6901 ((mddev->chunk_sectors << 9) / PAGE_SIZE); in run()
6905 chunk_size = mddev->chunk_sectors << 9; in run()
6996 mddev->chunk_sectors / 2, mddev->layout); in status()
7202 sectors &= ~((sector_t)mddev->chunk_sectors - 1); in raid5_resize()
7236 if (((mddev->chunk_sectors << 9) / STRIPE_SIZE) * 4 in check_stripe_cache()
7242 ((max(mddev->chunk_sectors, mddev->new_chunk_sectors) << 9) in check_stripe_cache()
7255 mddev->new_chunk_sectors == mddev->chunk_sectors) in check_reshape()
7275 if (mddev->new_chunk_sectors > mddev->chunk_sectors || in check_reshape()
7281 mddev->chunk_sectors) in check_reshape()
7332 conf->prev_chunk_sectors = conf->chunk_sectors; in raid5_start_reshape()
7333 conf->chunk_sectors = mddev->new_chunk_sectors; in raid5_start_reshape()
7408 conf->chunk_sectors = conf->prev_chunk_sectors; in raid5_start_reshape()
7449 int stripe = data_disks * ((conf->chunk_sectors << 9) in end_reshape()
7487 mddev->chunk_sectors = conf->chunk_sectors; in raid5_finish_reshape()
7547 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid45_takeover_raid0()
7647 conf->chunk_sectors = new_chunk ; in raid5_check_reshape()
7648 mddev->chunk_sectors = new_chunk; in raid5_check_reshape()