raid_map 4024 drivers/md/dm-raid.c .map = raid_map, raid_map 1369 drivers/scsi/hpsa.c h->dev[entry]->raid_map = new_entry->raid_map; raid_map 1688 drivers/scsi/hpsa.c struct raid_map_data *map = &logical_drive->raid_map; raid_map 3300 drivers/scsi/hpsa.c if (fill_cmd(c, HPSA_GET_RAID_MAP, h, &this_device->raid_map, raid_map 3301 drivers/scsi/hpsa.c sizeof(this_device->raid_map), 0, raid_map 3320 drivers/scsi/hpsa.c if (le32_to_cpu(this_device->raid_map.structure_size) > raid_map 3321 drivers/scsi/hpsa.c sizeof(this_device->raid_map)) { raid_map 3325 drivers/scsi/hpsa.c hpsa_debug_map_buff(h, rc, &this_device->raid_map); raid_map 4824 drivers/scsi/hpsa.c struct raid_map_data *map = &dev->raid_map; raid_map 5077 drivers/scsi/hpsa.c struct raid_map_data *map = &dev->raid_map; raid_map 98 drivers/scsi/hpsa.h struct raid_map_data raid_map; /* I/O accelerator RAID map */ raid_map 942 drivers/scsi/smartpqi/smartpqi.h struct raid_map *raid_map; /* RAID bypass map */ raid_map 1109 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_scsi_dev *device, struct raid_map *raid_map) raid_map 1115 drivers/scsi/smartpqi/smartpqi_init.c raid_map_size = get_unaligned_le32(&raid_map->structure_size); raid_map 1117 drivers/scsi/smartpqi/smartpqi_init.c if (raid_map_size < offsetof(struct raid_map, disk_data)) { raid_map 1123 drivers/scsi/smartpqi/smartpqi_init.c if (get_unaligned_le16(&raid_map->layout_map_count) != 2) { raid_map 1128 drivers/scsi/smartpqi/smartpqi_init.c if (get_unaligned_le16(&raid_map->layout_map_count) != 3) { raid_map 1134 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->layout_map_count) > 1) { raid_map 1137 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->strip_size) * raid_map 1138 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->data_disks_per_row); raid_map 1161 drivers/scsi/smartpqi/smartpqi_init.c struct raid_map *raid_map; raid_map 1163 drivers/scsi/smartpqi/smartpqi_init.c raid_map = kmalloc(sizeof(*raid_map), GFP_KERNEL); raid_map 1164 drivers/scsi/smartpqi/smartpqi_init.c if (!raid_map) raid_map 1168 drivers/scsi/smartpqi/smartpqi_init.c device->scsi3addr, raid_map, sizeof(*raid_map), raid_map 1174 drivers/scsi/smartpqi/smartpqi_init.c raid_map_size = get_unaligned_le32(&raid_map->structure_size); raid_map 1176 drivers/scsi/smartpqi/smartpqi_init.c if (raid_map_size > sizeof(*raid_map)) { raid_map 1178 drivers/scsi/smartpqi/smartpqi_init.c kfree(raid_map); raid_map 1180 drivers/scsi/smartpqi/smartpqi_init.c raid_map = kmalloc(raid_map_size, GFP_KERNEL); raid_map 1181 drivers/scsi/smartpqi/smartpqi_init.c if (!raid_map) raid_map 1185 drivers/scsi/smartpqi/smartpqi_init.c device->scsi3addr, raid_map, raid_map_size, raid_map 1190 drivers/scsi/smartpqi/smartpqi_init.c if (get_unaligned_le32(&raid_map->structure_size) raid_map 1195 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le32(&raid_map->structure_size)); raid_map 1200 drivers/scsi/smartpqi/smartpqi_init.c rc = pqi_validate_raid_map(ctrl_info, device, raid_map); raid_map 1204 drivers/scsi/smartpqi/smartpqi_init.c device->raid_map = raid_map; raid_map 1209 drivers/scsi/smartpqi/smartpqi_init.c kfree(raid_map); raid_map 1693 drivers/scsi/smartpqi/smartpqi_init.c kfree(existing_device->raid_map); raid_map 1694 drivers/scsi/smartpqi/smartpqi_init.c existing_device->raid_map = new_device->raid_map; raid_map 1702 drivers/scsi/smartpqi/smartpqi_init.c new_device->raid_map = NULL; raid_map 1708 drivers/scsi/smartpqi/smartpqi_init.c kfree(device->raid_map); raid_map 2238 drivers/scsi/smartpqi/smartpqi_init.c struct pqi_encryption_info *encryption_info, struct raid_map *raid_map, raid_map 2248 drivers/scsi/smartpqi/smartpqi_init.c volume_blk_size = get_unaligned_le32(&raid_map->volume_blk_size); raid_map 2253 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->data_encryption_key_index); raid_map 2268 drivers/scsi/smartpqi/smartpqi_init.c struct raid_map *raid_map; raid_map 2357 drivers/scsi/smartpqi/smartpqi_init.c raid_map = device->raid_map; raid_map 2360 drivers/scsi/smartpqi/smartpqi_init.c if (last_block >= get_unaligned_le64(&raid_map->volume_blk_cnt) || raid_map 2364 drivers/scsi/smartpqi/smartpqi_init.c data_disks_per_row = get_unaligned_le16(&raid_map->data_disks_per_row); raid_map 2365 drivers/scsi/smartpqi/smartpqi_init.c strip_size = get_unaligned_le16(&raid_map->strip_size); raid_map 2366 drivers/scsi/smartpqi/smartpqi_init.c layout_map_count = get_unaligned_le16(&raid_map->layout_map_count); raid_map 2400 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->metadata_disks_per_row); raid_map 2401 drivers/scsi/smartpqi/smartpqi_init.c map_row = ((u32)(first_row >> raid_map->parity_rotation_shift)) % raid_map 2402 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->row_cnt); raid_map 2535 drivers/scsi/smartpqi/smartpqi_init.c ((u32)(first_row >> raid_map->parity_rotation_shift)) % raid_map 2536 drivers/scsi/smartpqi/smartpqi_init.c get_unaligned_le16(&raid_map->row_cnt); raid_map 2539 drivers/scsi/smartpqi/smartpqi_init.c (get_unaligned_le16(&raid_map->row_cnt) * raid_map 2544 drivers/scsi/smartpqi/smartpqi_init.c aio_handle = raid_map->disk_data[map_index].aio_handle; raid_map 2545 drivers/scsi/smartpqi/smartpqi_init.c disk_block = get_unaligned_le64(&raid_map->disk_starting_blk) + raid_map 2551 drivers/scsi/smartpqi/smartpqi_init.c if (raid_map->phys_blk_shift) { raid_map 2552 drivers/scsi/smartpqi/smartpqi_init.c disk_block <<= raid_map->phys_blk_shift; raid_map 2553 drivers/scsi/smartpqi/smartpqi_init.c disk_block_cnt <<= raid_map->phys_blk_shift; raid_map 2578 drivers/scsi/smartpqi/smartpqi_init.c if (get_unaligned_le16(&raid_map->flags) & raid_map 2580 drivers/scsi/smartpqi/smartpqi_init.c pqi_set_encryption_info(&encryption_info, raid_map, raid_map 284 fs/btrfs/raid56.c u64 num = rbio->bbio->raid_map[0]; raid_map 572 fs/btrfs/raid56.c if (last->bbio->raid_map[0] != raid_map 573 fs/btrfs/raid56.c cur->bbio->raid_map[0]) raid_map 685 fs/btrfs/raid56.c if (cur->bbio->raid_map[0] == rbio->bbio->raid_map[0]) { raid_map 1172 fs/btrfs/raid56.c stripe_offset = start - rbio->bbio->raid_map[0]; raid_map 1399 fs/btrfs/raid56.c stripe_start = rbio->bbio->raid_map[i]; raid_map 1900 fs/btrfs/raid56.c if (rbio->bbio->raid_map[failb] == RAID6_Q_STRIPE) { raid_map 1901 fs/btrfs/raid56.c if (rbio->bbio->raid_map[faila] == raid_map 1913 fs/btrfs/raid56.c if (rbio->bbio->raid_map[failb] == RAID5_P_STRIPE) { raid_map 2315 fs/btrfs/raid56.c ASSERT(logical >= rbio->bbio->raid_map[0]); raid_map 2316 fs/btrfs/raid56.c ASSERT(logical + PAGE_SIZE <= rbio->bbio->raid_map[0] + raid_map 2318 fs/btrfs/raid56.c stripe_offset = (int)(logical - rbio->bbio->raid_map[0]); raid_map 1237 fs/btrfs/scrub.c u64 *raid_map, raid_map 1248 fs/btrfs/scrub.c if (raid_map[i] == RAID6_Q_STRIPE || raid_map 1249 fs/btrfs/scrub.c raid_map[i] == RAID5_P_STRIPE) raid_map 1252 fs/btrfs/scrub.c if (logical >= raid_map[i] && raid_map 1253 fs/btrfs/scrub.c logical < raid_map[i] + mapped_length) raid_map 1258 fs/btrfs/scrub.c *stripe_offset = logical - raid_map[i]; raid_map 1357 fs/btrfs/scrub.c bbio->raid_map, raid_map 2176 fs/btrfs/scrub.c if (ret || !bbio || !bbio->raid_map) raid_map 2767 fs/btrfs/scrub.c if (ret || !bbio || !bbio->raid_map) raid_map 5578 fs/btrfs/volumes.c if (parity_smaller(bbio->raid_map[i], raid_map 5579 fs/btrfs/volumes.c bbio->raid_map[i+1])) { raid_map 5581 fs/btrfs/volumes.c l = bbio->raid_map[i]; raid_map 5583 fs/btrfs/volumes.c bbio->raid_map[i] = bbio->raid_map[i+1]; raid_map 5585 fs/btrfs/volumes.c bbio->raid_map[i+1] = l; raid_map 6244 fs/btrfs/volumes.c bbio->raid_map = (u64 *)((void *)bbio->stripes + raid_map 6255 fs/btrfs/volumes.c bbio->raid_map[(i+rot) % num_stripes] = raid_map 6258 fs/btrfs/volumes.c bbio->raid_map[(i+rot) % map->num_stripes] = RAID5_P_STRIPE; raid_map 6260 fs/btrfs/volumes.c bbio->raid_map[(i+rot+1) % num_stripes] = raid_map 6278 fs/btrfs/volumes.c if (bbio->raid_map) raid_map 345 fs/btrfs/volumes.h u64 *raid_map;