Searched refs:dev_sectors (Results 1 – 12 of 12) sorted by relevance
48 sector_t dev_sectors; /* temp copy of member
1084 mddev->dev_sectors = ((sector_t)sb->size) * 2; in super_90_validate()1221 sb->size = mddev->dev_sectors / 2; in super_90_sync()1330 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_90_rdev_size_change()1571 mddev->dev_sectors = le64_to_cpu(sb->size); in super_1_validate()1708 sb->size = cpu_to_le64(mddev->dev_sectors); in super_1_sync()1830 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_1_rdev_size_change()2037 if (rdev->sectors && (mddev->dev_sectors == 0 || in bind_rdev_to_array()2038 rdev->sectors < mddev->dev_sectors)) { in bind_rdev_to_array()2047 mddev->dev_sectors = rdev->sectors; in bind_rdev_to_array()2786 + mddev->dev_sectors > rdev->sectors) in new_offset_store()[all …]
2921 return mddev->dev_sectors - sector_nr; in sync_request()2925 max_sector = mddev->dev_sectors; in sync_request()3426 sectors = conf->dev_sectors; in raid10_size()3456 conf->dev_sectors = size << conf->geo.chunk_shift; in calc_sectors()3553 calc_sectors(conf, mddev->dev_sectors); in setup_conf()3567 conf->prev.stride = conf->dev_sectors; in setup_conf()3738 mddev->dev_sectors = conf->dev_sectors; in run()3859 if (sectors > mddev->dev_sectors && in raid10_resize()3865 mddev->dev_sectors = conf->dev_sectors; in raid10_resize()3891 mddev->dev_sectors = size; in raid10_takeover_raid0()
297 return mddev->dev_sectors; in faulty_size()
2504 max_sector = mddev->dev_sectors; in sync_request()2758 return mddev->dev_sectors; in raid1_size()3006 if (sectors > mddev->dev_sectors && in raid1_resize()3007 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()3008 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()3011 mddev->dev_sectors = sectors; in raid1_resize()
258 sector_t dev_sectors; /* used size of member
382 return mddev->dev_sectors; in multipath_size()
5407 BUG_ON((mddev->dev_sectors & in reshape_request()5518 if (last_sector >= mddev->dev_sectors) in reshape_request()5519 last_sector = mddev->dev_sectors - 1; in reshape_request()5571 sector_t max_sector = mddev->dev_sectors; in sync_request()5612 sector_t rv = mddev->dev_sectors - sector_nr; in sync_request()6255 sectors = mddev->dev_sectors; in raid5_size()6839 mddev->dev_sectors &= ~(mddev->chunk_sectors - 1); in run()6840 mddev->resync_max_sectors = mddev->dev_sectors; in run()7215 if (sectors > mddev->dev_sectors && in raid5_resize()7216 mddev->recovery_cp > mddev->dev_sectors) { in raid5_resize()[all …]
728 rs->md.dev_sectors = sectors_per_dev; in parse_raid_params()1596 rs->md.dev_sectors, in raid_iterate_devices()
580 rdev->sectors = mddev->dev_sectors; in raid0_takeover_raid45()
241 < (rdev->data_offset + mddev->dev_sectors in write_sb_page()251 if (rdev->data_offset + mddev->dev_sectors in write_sb_page()
1552 u64 dev_sectors = qc->dev->n_sectors; in ata_scsi_verify_xlat() local1573 if (block >= dev_sectors) in ata_scsi_verify_xlat()1575 if ((block + n_block) > dev_sectors) in ata_scsi_verify_xlat()