tbio 2012 drivers/md/raid10.c struct bio *tbio, *fbio; tbio 2039 drivers/md/raid10.c tbio = r10_bio->devs[i].bio; tbio 2041 drivers/md/raid10.c if (tbio->bi_end_io != end_sync_read) tbio 2046 drivers/md/raid10.c tpages = get_resync_pages(tbio)->pages; tbio 2081 drivers/md/raid10.c rp = get_resync_pages(tbio); tbio 2082 drivers/md/raid10.c bio_reset(tbio); tbio 2084 drivers/md/raid10.c md_bio_reset_resync_pages(tbio, rp, fbio->bi_iter.bi_size); tbio 2087 drivers/md/raid10.c tbio->bi_private = rp; tbio 2088 drivers/md/raid10.c tbio->bi_iter.bi_sector = r10_bio->devs[i].addr; tbio 2089 drivers/md/raid10.c tbio->bi_end_io = end_sync_write; tbio 2090 drivers/md/raid10.c bio_set_op_attrs(tbio, REQ_OP_WRITE, 0); tbio 2092 drivers/md/raid10.c bio_copy_data(tbio, fbio); tbio 2096 drivers/md/raid10.c md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(tbio)); tbio 2099 drivers/md/raid10.c tbio->bi_opf |= MD_FAILFAST; tbio 2100 drivers/md/raid10.c tbio->bi_iter.bi_sector += conf->mirrors[d].rdev->data_offset; tbio 2101 drivers/md/raid10.c bio_set_dev(tbio, conf->mirrors[d].rdev->bdev); tbio 2102 drivers/md/raid10.c generic_make_request(tbio); tbio 2111 drivers/md/raid10.c tbio = r10_bio->devs[i].repl_bio; tbio 2112 drivers/md/raid10.c if (!tbio || !tbio->bi_end_io) tbio 2116 drivers/md/raid10.c bio_copy_data(tbio, fbio); tbio 2120 drivers/md/raid10.c bio_sectors(tbio)); tbio 2121 drivers/md/raid10.c generic_make_request(tbio);