origin_dev 393 drivers/md/dm-cache-target.c struct dm_dev *origin_dev; origin_dev 808 drivers/md/dm-cache-target.c bio_set_dev(bio, cache->origin_dev->bdev); origin_dev 1191 drivers/md/dm-cache-target.c o_region.bdev = cache->origin_dev->bdev; origin_dev 2013 drivers/md/dm-cache-target.c if (cache->origin_dev) origin_dev 2014 drivers/md/dm-cache-target.c dm_put_device(cache->ti, cache->origin_dev); origin_dev 2082 drivers/md/dm-cache-target.c struct dm_dev *origin_dev; origin_dev 2102 drivers/md/dm-cache-target.c if (ca->origin_dev) origin_dev 2103 drivers/md/dm-cache-target.c dm_put_device(ca->ti, ca->origin_dev); origin_dev 2171 drivers/md/dm-cache-target.c &ca->origin_dev); origin_dev 2177 drivers/md/dm-cache-target.c ca->origin_sectors = get_dev_size(ca->origin_dev); origin_dev 2457 drivers/md/dm-cache-target.c return is_congested(cache->origin_dev, bdi_bits) || origin_dev 2499 drivers/md/dm-cache-target.c cache->origin_dev = ca->origin_dev; origin_dev 2502 drivers/md/dm-cache-target.c ca->metadata_dev = ca->origin_dev = ca->cache_dev = NULL; origin_dev 3237 drivers/md/dm-cache-target.c format_dev_t(buf, cache->origin_dev->bdev->bd_dev); origin_dev 3430 drivers/md/dm-cache-target.c r = fn(ti, cache->origin_dev, 0, ti->len, data); origin_dev 3448 drivers/md/dm-cache-target.c struct block_device *origin_bdev = cache->origin_dev->bdev; origin_dev 3471 drivers/md/dm-cache-target.c struct block_device *origin_bdev = cache->origin_dev->bdev; origin_dev 1143 drivers/md/dm-era-target.c struct dm_dev *origin_dev; origin_dev 1196 drivers/md/dm-era-target.c bio_set_dev(bio, era->origin_dev->bdev); origin_dev 1387 drivers/md/dm-era-target.c return dev_is_congested(era->origin_dev, bdi_bits); origin_dev 1398 drivers/md/dm-era-target.c if (era->origin_dev) origin_dev 1399 drivers/md/dm-era-target.c dm_put_device(era->ti, era->origin_dev); origin_dev 1450 drivers/md/dm-era-target.c r = dm_get_device(ti, argv[1], FMODE_READ | FMODE_WRITE, &era->origin_dev); origin_dev 1628 drivers/md/dm-era-target.c format_dev_t(buf, era->origin_dev->bdev->bd_dev); origin_dev 1671 drivers/md/dm-era-target.c return fn(ti, era->origin_dev, 0, get_dev_size(era->origin_dev), data); origin_dev 1237 drivers/md/dm-snap.c dev_t origin_dev, cow_dev; origin_dev 1275 drivers/md/dm-snap.c origin_dev = s->origin->bdev->bd_dev; origin_dev 1282 drivers/md/dm-snap.c if (cow_dev && cow_dev == origin_dev) { origin_dev 341 drivers/md/dm-thin.c struct dm_dev *origin_dev; origin_dev 738 drivers/md/dm-thin.c bio_set_dev(bio, tc->origin_dev->bdev); origin_dev 1428 drivers/md/dm-thin.c schedule_copy(tc, virt_block, tc->origin_dev, origin_dev 1433 drivers/md/dm-thin.c schedule_copy(tc, virt_block, tc->origin_dev, origin_dev 1926 drivers/md/dm-thin.c if (tc->origin_dev) origin_dev 1970 drivers/md/dm-thin.c if (bio_data_dir(bio) == READ && tc->origin_dev) { origin_dev 2048 drivers/md/dm-thin.c if (tc->origin_dev) { origin_dev 4173 drivers/md/dm-thin.c if (tc->origin_dev) origin_dev 4174 drivers/md/dm-thin.c dm_put_device(ti, tc->origin_dev); origin_dev 4196 drivers/md/dm-thin.c struct dm_dev *pool_dev, *origin_dev; origin_dev 4228 drivers/md/dm-thin.c r = dm_get_device(ti, argv[2], FMODE_READ, &origin_dev); origin_dev 4233 drivers/md/dm-thin.c tc->origin_dev = origin_dev; origin_dev 4325 drivers/md/dm-thin.c if (tc->origin_dev) origin_dev 4326 drivers/md/dm-thin.c dm_put_device(ti, tc->origin_dev); origin_dev 4404 drivers/md/dm-thin.c if (tc->origin_dev) origin_dev 4405 drivers/md/dm-thin.c tc->origin_size = get_dev_size(tc->origin_dev->bdev); origin_dev 4456 drivers/md/dm-thin.c if (tc->origin_dev) origin_dev 4457 drivers/md/dm-thin.c DMEMIT(" %s", format_dev_t(buf, tc->origin_dev->bdev->bd_dev));