Searched refs:sector_nr (Results 1 - 4 of 4) sorted by relevance
/linux-4.1.27/drivers/md/ |
H A D | raid10.c | 101 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr, 2891 static sector_t sync_request(struct mddev *mddev, sector_t sector_nr, sync_request() argument 2921 return mddev->dev_sectors - sector_nr; sync_request() 2929 if (sector_nr >= max_sector) { sync_request() 2978 return reshape_request(mddev, sector_nr, skipped); sync_request() 2985 return (max_sector - sector_nr) + sectors_skipped; sync_request() 2995 max_sector > (sector_nr | chunk_mask)) sync_request() 2996 max_sector = (sector_nr | chunk_mask) + 1; sync_request() 3038 sect = raid10_find_virt(conf, sector_nr, i); sync_request() 3240 bitmap_cond_end_sync(mddev->bitmap, sector_nr); sync_request() 3242 if (!bitmap_start_sync(mddev->bitmap, sector_nr, sync_request() 3258 conf->next_resync = sector_nr; sync_request() 3261 r10_bio->sector = sector_nr; sync_request() 3264 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; sync_request() 3347 if (sector_nr + max_sync < max_sector) sync_request() 3348 max_sector = sector_nr + max_sync; sync_request() 3352 if (sector_nr + (len>>9) > max_sector) sync_request() 3353 len = (max_sector - sector_nr) << 9; sync_request() 3375 sector_nr += len>>9; sync_request() 3407 if (sector_nr + max_sync < max_sector) sync_request() 3408 max_sector = sector_nr + max_sync; sync_request() 3410 sectors_skipped += (max_sector - sector_nr); sync_request() 3412 sector_nr = max_sector; sync_request() 4221 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr, reshape_request() argument 4240 * We interpret 'sector_nr' as an address that we want to write to. reshape_request() 4273 if (sector_nr == 0) { reshape_request() 4277 sector_nr = (raid10_size(mddev, 0, 0) reshape_request() 4281 sector_nr = conf->reshape_progress; reshape_request() 4282 if (sector_nr) { reshape_request() 4283 mddev->curr_resync_completed = sector_nr; reshape_request() 4286 return sector_nr; reshape_request() 4290 /* We don't use sector_nr to track where we are up to reshape_request() 4311 sector_nr = last & ~(sector_t)(conf->geo.chunk_mask reshape_request() 4313 if (sector_nr + RESYNC_BLOCK_SIZE/512 < last) reshape_request() 4314 sector_nr = last + 1 - RESYNC_BLOCK_SIZE/512; reshape_request() 4332 sector_nr = conf->reshape_progress; reshape_request() 4333 last = sector_nr | (conf->geo.chunk_mask reshape_request() 4336 if (sector_nr + RESYNC_BLOCK_SIZE/512 <= last) reshape_request() 4337 last = sector_nr + RESYNC_BLOCK_SIZE/512 - 1; reshape_request() 4364 /* Now schedule reads for blocks from sector_nr to last */ reshape_request() 4370 r10_bio->sector = sector_nr; reshape_request() 4372 r10_bio->sectors = last - sector_nr + 1; reshape_request() 4456 sector_nr += len >> 9; reshape_request() 4467 sector_nr += nr_sectors; reshape_request() 4469 if (sector_nr <= last) reshape_request()
|
H A D | raid1.c | 826 static void raise_barrier(struct r1conf *conf, sector_t sector_nr) raise_barrier() argument 836 conf->next_resync = sector_nr; raise_barrier() 2485 static sector_t sync_request(struct mddev *mddev, sector_t sector_nr, int *skipped) sync_request() argument 2505 if (sector_nr >= max_sector) { sync_request() 2527 return max_sector - sector_nr; sync_request() 2532 if (!bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && sync_request() 2539 bitmap_cond_end_sync(mddev->bitmap, sector_nr); sync_request() 2542 raise_barrier(conf, sector_nr); sync_request() 2555 r1_bio->sector = sector_nr; sync_request() 2578 if (is_badblock(rdev, sector_nr, good_sectors, sync_request() 2580 if (first_bad > sector_nr) sync_request() 2581 good_sectors = first_bad - sector_nr; sync_request() 2583 bad_sectors -= (sector_nr - first_bad); sync_request() 2589 if (sector_nr < first_bad) { sync_request() 2616 bio->bi_iter.bi_sector = sector_nr + rdev->data_offset; sync_request() 2634 ok = rdev_set_badblocks(rdev, sector_nr, sync_request() 2671 max_sector = sector_nr + min_bad; sync_request() 2672 rv = max_sector - sector_nr; sync_request() 2680 if (max_sector > sector_nr + good_sectors) sync_request() 2681 max_sector = sector_nr + good_sectors; sync_request() 2687 if (sector_nr + (len>>9) > max_sector) sync_request() 2688 len = (max_sector - sector_nr) << 9; sync_request() 2692 if (!bitmap_start_sync(mddev->bitmap, sector_nr, sync_request() 2724 sector_nr += len>>9; sync_request()
|
H A D | raid5.c | 5328 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr, int *skipped) reshape_request() argument 5352 if (sector_nr == 0) { reshape_request() 5356 sector_nr = raid5_size(mddev, 0, 0) reshape_request() 5360 sector_nr = conf->reshape_progress; reshape_request() 5361 sector_div(sector_nr, new_data_disks); reshape_request() 5362 if (sector_nr) { reshape_request() 5363 mddev->curr_resync_completed = sector_nr; reshape_request() 5366 return sector_nr; reshape_request() 5410 != sector_nr); reshape_request() 5412 BUG_ON(writepos != sector_nr + reshape_sectors); reshape_request() 5413 stripe_addr = sector_nr; reshape_request() 5453 mddev->curr_resync_completed = sector_nr; reshape_request() 5538 sector_nr += reshape_sectors; reshape_request() 5539 if ((sector_nr - mddev->curr_resync_completed) * 2 reshape_request() 5548 mddev->curr_resync_completed = sector_nr; reshape_request() 5567 static inline sector_t sync_request(struct mddev *mddev, sector_t sector_nr, int *skipped) sync_request() argument 5576 if (sector_nr >= max_sector) { sync_request() 5598 return reshape_request(mddev, sector_nr, skipped); sync_request() 5612 sector_t rv = mddev->dev_sectors - sector_nr; sync_request() 5618 !bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && sync_request() 5626 bitmap_cond_end_sync(mddev->bitmap, sector_nr); sync_request() 5628 sh = get_active_stripe(conf, sector_nr, 0, 1, 0); sync_request() 5630 sh = get_active_stripe(conf, sector_nr, 0, 0, 0); sync_request() 5649 bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, still_degraded); sync_request()
|
H A D | md.h | 509 sector_t (*sync_request)(struct mddev *mddev, sector_t sector_nr, int *skipped);
|
Completed in 149 milliseconds