Searched refs:bad_sectors (Results 1 - 5 of 5) sorted by relevance
/linux-4.1.27/drivers/md/ |
H A D | raid1.c | 434 int bad_sectors; raid1_end_write_request() local 453 &first_bad, &bad_sectors)) { raid1_end_write_request() 553 int bad_sectors; read_balance() local 571 &first_bad, &bad_sectors)) { read_balance() 587 &first_bad, &bad_sectors)) { read_balance() 594 * bad_sectors from another device.. read_balance() 596 bad_sectors -= (this_sector - first_bad); read_balance() 597 if (choose_first && sectors > bad_sectors) read_balance() 598 sectors = bad_sectors; read_balance() 1281 int bad_sectors; make_request() local 1286 &first_bad, &bad_sectors); make_request() 1296 bad_sectors -= (r1_bio->sector - first_bad); make_request() 1297 if (bad_sectors < max_sectors) make_request() 1298 /* mustn't write more than bad_sectors make_request() 1301 max_sectors = bad_sectors; make_request() 1765 int bad_sectors; end_sync_write() local 1790 &first_bad, &bad_sectors) && end_sync_write() 1794 &first_bad, &bad_sectors) end_sync_write() 2123 int bad_sectors; fix_read_error() local 2131 &first_bad, &bad_sectors) == 0 && fix_read_error() 2576 int bad_sectors; sync_request() local 2579 &first_bad, &bad_sectors)) { sync_request() 2583 bad_sectors -= (sector_nr - first_bad); sync_request() 2585 min_bad > bad_sectors) sync_request() 2586 min_bad = bad_sectors; sync_request()
|
H A D | raid10.c | 488 int bad_sectors; raid10_end_write_request() local 506 &first_bad, &bad_sectors)) { raid10_end_write_request() 816 int bad_sectors; read_balance() local 837 &first_bad, &bad_sectors)) { read_balance() 844 * beyond 'bad_sectors' from another device. read_balance() 846 bad_sectors -= (dev_sector - first_bad); read_balance() 847 if (!do_balance && sectors > bad_sectors) read_balance() 848 sectors = bad_sectors; read_balance() 1346 int bad_sectors; __make_request() local 1351 &first_bad, &bad_sectors); __make_request() 1363 bad_sectors -= (dev_sector - first_bad); __make_request() 1364 if (bad_sectors < max_sectors) __make_request() 1365 /* Mustn't write more than bad_sectors __make_request() 1368 max_sectors = bad_sectors; __make_request() 1988 int bad_sectors; end_sync_write() local 2012 &first_bad, &bad_sectors)) end_sync_write() 2325 int bad_sectors; r10_sync_page_io() local 2327 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) r10_sync_page_io() 2402 int bad_sectors; fix_read_error() local 2410 &first_bad, &bad_sectors) == 0) { fix_read_error() 3097 int bad_sectors; sync_request() local 3107 &first_bad, &bad_sectors)) { sync_request() 3111 bad_sectors -= (sector sync_request() 3113 if (max_sync > bad_sectors) sync_request() 3114 max_sync = bad_sectors; sync_request() 3269 int bad_sectors; sync_request() local 3283 &first_bad, &bad_sectors)) { sync_request() 3287 bad_sectors -= (sector - first_bad); sync_request() 3288 if (max_sync > bad_sectors) sync_request() 3289 max_sync = bad_sectors; sync_request()
|
H A D | md.h | 190 sector_t *first_bad, int *bad_sectors); is_badblock() 192 sector_t *first_bad, int *bad_sectors) is_badblock() 197 first_bad, bad_sectors); is_badblock() 191 is_badblock(struct md_rdev *rdev, sector_t s, int sectors, sector_t *first_bad, int *bad_sectors) is_badblock() argument
|
H A D | raid5.c | 957 int bad_sectors; ops_run_io() local 959 &first_bad, &bad_sectors); ops_run_io() 2417 int bad_sectors; raid5_end_write_request() local 2451 &first_bad, &bad_sectors)) raid5_end_write_request() 2463 &first_bad, &bad_sectors)) { raid5_end_write_request() 4024 int bad_sectors; analyse_stripe() local 4069 &first_bad, &bad_sectors)) analyse_stripe() 4081 &first_bad, &bad_sectors); analyse_stripe() 4855 int bad_sectors; chunk_aligned_read() local 4866 &first_bad, &bad_sectors)) { chunk_aligned_read()
|
H A D | md.c | 8341 sector_t *first_bad, int *bad_sectors) md_is_badblock() 8400 *bad_sectors = BB_LEN(p[lo]); md_is_badblock() 8340 md_is_badblock(struct badblocks *bb, sector_t s, int sectors, sector_t *first_bad, int *bad_sectors) md_is_badblock() argument
|
Completed in 237 milliseconds