tbio             2012 drivers/md/raid10.c 	struct bio *tbio, *fbio;
tbio             2039 drivers/md/raid10.c 		tbio = r10_bio->devs[i].bio;
tbio             2041 drivers/md/raid10.c 		if (tbio->bi_end_io != end_sync_read)
tbio             2046 drivers/md/raid10.c 		tpages = get_resync_pages(tbio)->pages;
tbio             2081 drivers/md/raid10.c 		rp = get_resync_pages(tbio);
tbio             2082 drivers/md/raid10.c 		bio_reset(tbio);
tbio             2084 drivers/md/raid10.c 		md_bio_reset_resync_pages(tbio, rp, fbio->bi_iter.bi_size);
tbio             2087 drivers/md/raid10.c 		tbio->bi_private = rp;
tbio             2088 drivers/md/raid10.c 		tbio->bi_iter.bi_sector = r10_bio->devs[i].addr;
tbio             2089 drivers/md/raid10.c 		tbio->bi_end_io = end_sync_write;
tbio             2090 drivers/md/raid10.c 		bio_set_op_attrs(tbio, REQ_OP_WRITE, 0);
tbio             2092 drivers/md/raid10.c 		bio_copy_data(tbio, fbio);
tbio             2096 drivers/md/raid10.c 		md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(tbio));
tbio             2099 drivers/md/raid10.c 			tbio->bi_opf |= MD_FAILFAST;
tbio             2100 drivers/md/raid10.c 		tbio->bi_iter.bi_sector += conf->mirrors[d].rdev->data_offset;
tbio             2101 drivers/md/raid10.c 		bio_set_dev(tbio, conf->mirrors[d].rdev->bdev);
tbio             2102 drivers/md/raid10.c 		generic_make_request(tbio);
tbio             2111 drivers/md/raid10.c 		tbio = r10_bio->devs[i].repl_bio;
tbio             2112 drivers/md/raid10.c 		if (!tbio || !tbio->bi_end_io)
tbio             2116 drivers/md/raid10.c 			bio_copy_data(tbio, fbio);
tbio             2120 drivers/md/raid10.c 			     bio_sectors(tbio));
tbio             2121 drivers/md/raid10.c 		generic_make_request(tbio);