Searched refs:bad_sectors (Results 1 - 5 of 5) sorted by relevance

/linux-4.1.27/drivers/md/
H A Draid1.c434 int bad_sectors; raid1_end_write_request() local
453 &first_bad, &bad_sectors)) { raid1_end_write_request()
553 int bad_sectors; read_balance() local
571 &first_bad, &bad_sectors)) { read_balance()
587 &first_bad, &bad_sectors)) { read_balance()
594 * bad_sectors from another device.. read_balance()
596 bad_sectors -= (this_sector - first_bad); read_balance()
597 if (choose_first && sectors > bad_sectors) read_balance()
598 sectors = bad_sectors; read_balance()
1281 int bad_sectors; make_request() local
1286 &first_bad, &bad_sectors); make_request()
1296 bad_sectors -= (r1_bio->sector - first_bad); make_request()
1297 if (bad_sectors < max_sectors) make_request()
1298 /* mustn't write more than bad_sectors make_request()
1301 max_sectors = bad_sectors; make_request()
1765 int bad_sectors; end_sync_write() local
1790 &first_bad, &bad_sectors) && end_sync_write()
1794 &first_bad, &bad_sectors) end_sync_write()
2123 int bad_sectors; fix_read_error() local
2131 &first_bad, &bad_sectors) == 0 && fix_read_error()
2576 int bad_sectors; sync_request() local
2579 &first_bad, &bad_sectors)) { sync_request()
2583 bad_sectors -= (sector_nr - first_bad); sync_request()
2585 min_bad > bad_sectors) sync_request()
2586 min_bad = bad_sectors; sync_request()
H A Draid10.c488 int bad_sectors; raid10_end_write_request() local
506 &first_bad, &bad_sectors)) { raid10_end_write_request()
816 int bad_sectors; read_balance() local
837 &first_bad, &bad_sectors)) { read_balance()
844 * beyond 'bad_sectors' from another device. read_balance()
846 bad_sectors -= (dev_sector - first_bad); read_balance()
847 if (!do_balance && sectors > bad_sectors) read_balance()
848 sectors = bad_sectors; read_balance()
1346 int bad_sectors; __make_request() local
1351 &first_bad, &bad_sectors); __make_request()
1363 bad_sectors -= (dev_sector - first_bad); __make_request()
1364 if (bad_sectors < max_sectors) __make_request()
1365 /* Mustn't write more than bad_sectors __make_request()
1368 max_sectors = bad_sectors; __make_request()
1988 int bad_sectors; end_sync_write() local
2012 &first_bad, &bad_sectors)) end_sync_write()
2325 int bad_sectors; r10_sync_page_io() local
2327 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) r10_sync_page_io()
2402 int bad_sectors; fix_read_error() local
2410 &first_bad, &bad_sectors) == 0) { fix_read_error()
3097 int bad_sectors; sync_request() local
3107 &first_bad, &bad_sectors)) { sync_request()
3111 bad_sectors -= (sector sync_request()
3113 if (max_sync > bad_sectors) sync_request()
3114 max_sync = bad_sectors; sync_request()
3269 int bad_sectors; sync_request() local
3283 &first_bad, &bad_sectors)) { sync_request()
3287 bad_sectors -= (sector - first_bad); sync_request()
3288 if (max_sync > bad_sectors) sync_request()
3289 max_sync = bad_sectors; sync_request()
H A Dmd.h190 sector_t *first_bad, int *bad_sectors); is_badblock()
192 sector_t *first_bad, int *bad_sectors) is_badblock()
197 first_bad, bad_sectors); is_badblock()
191 is_badblock(struct md_rdev *rdev, sector_t s, int sectors, sector_t *first_bad, int *bad_sectors) is_badblock() argument
H A Draid5.c957 int bad_sectors; ops_run_io() local
959 &first_bad, &bad_sectors); ops_run_io()
2417 int bad_sectors; raid5_end_write_request() local
2451 &first_bad, &bad_sectors)) raid5_end_write_request()
2463 &first_bad, &bad_sectors)) { raid5_end_write_request()
4024 int bad_sectors; analyse_stripe() local
4069 &first_bad, &bad_sectors)) analyse_stripe()
4081 &first_bad, &bad_sectors); analyse_stripe()
4855 int bad_sectors; chunk_aligned_read() local
4866 &first_bad, &bad_sectors)) { chunk_aligned_read()
H A Dmd.c8341 sector_t *first_bad, int *bad_sectors) md_is_badblock()
8400 *bad_sectors = BB_LEN(p[lo]); md_is_badblock()
8340 md_is_badblock(struct badblocks *bb, sector_t s, int sectors, sector_t *first_bad, int *bad_sectors) md_is_badblock() argument

Completed in 237 milliseconds