Home
last modified time | relevance | path

Searched refs:bi_rw (Results 1 – 72 of 72) sorted by relevance

/linux-4.1.27/include/linux/
Dbio.h51 #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT)
56 (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \
57 (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \
109 !(bio->bi_rw & REQ_DISCARD)) in bio_has_data()
120 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_is_rw()
128 if (bio->bi_rw & REQ_NOMERGE_FLAGS) in bio_mergeable()
239 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_advance_iter()
267 if (bio->bi_rw & REQ_DISCARD) in bio_segments()
270 if (bio->bi_rw & REQ_WRITE_SAME) in bio_segments()
307 if (bio->bi_rw & REQ_INTEGRITY) in bio_integrity()
Ddm-io.h60 int bi_rw; /* READ|WRITE - not READA */ member
Dblk_types.h50 unsigned long bi_rw; /* bottom bits READ/WRITE, member
Dfs.h2356 #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK))
2361 #define bio_data_dir(bio) ((bio)->bi_rw & 1)
Dblkdev.h246 unsigned long bi_rw; member
/linux-4.1.27/drivers/md/bcache/
Dio.c21 if (bio->bi_rw & REQ_DISCARD) in bch_bio_max_sectors()
29 .bi_rw = bio->bi_rw, in bch_bio_max_sectors()
81 if (!bio_has_data(bio) && !(bio->bi_rw & REQ_DISCARD)) in bch_generic_make_request()
212 unsigned threshold = bio->bi_rw & REQ_WRITE in bch_bbio_count_io_errors()
Drequest.c208 bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); in bch_data_insert_start()
251 n->bi_rw |= REQ_WRITE; in bch_data_insert_start()
375 (bio->bi_rw & REQ_DISCARD)) in check_should_bypass()
380 (bio->bi_rw & REQ_WRITE))) in check_should_bypass()
401 (bio->bi_rw & REQ_WRITE) && in check_should_bypass()
402 (bio->bi_rw & REQ_SYNC)) in check_should_bypass()
651 s->write = (bio->bi_rw & REQ_WRITE) != 0; in search_alloc()
662 s->iop.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; in search_alloc()
790 if (!(bio->bi_rw & REQ_RAHEAD) && in cached_dev_cache_miss()
791 !(bio->bi_rw & REQ_META) && in cached_dev_cache_miss()
[all …]
Dwriteback.h60 return bio->bi_rw & REQ_SYNC || in should_writeback()
Dwriteback.c186 io->bio.bi_rw = WRITE; in write_dirty()
259 io->bio.bi_rw = READ; in read_dirty()
Djournal.c57 bio->bi_rw = READ; in journal_read_bucket()
455 bio->bi_rw = REQ_WRITE|REQ_DISCARD; in do_journal_discard()
627 bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; in journal_write_unlocked()
Dmovinggc.c166 bio->bi_rw = READ; in read_moving()
Dsuper.c238 bio->bi_rw = REQ_SYNC|REQ_META; in __write_super()
374 bio->bi_rw = REQ_SYNC|REQ_META|rw; in uuid_io()
533 bio->bi_rw = REQ_SYNC|REQ_META|rw; in prio_io()
Dbtree.c298 bio->bi_rw = REQ_META|READ_SYNC; in bch_btree_node_read()
400 b->bio->bi_rw = REQ_META|WRITE_SYNC|REQ_FUA; in do_btree_node_write()
/linux-4.1.27/block/
Dblk-merge.c29 if (bio->bi_rw & REQ_DISCARD) in __blk_recalc_rq_segments()
32 if (bio->bi_rw & REQ_WRITE_SAME) in __blk_recalc_rq_segments()
212 if (bio->bi_rw & REQ_DISCARD) { in __blk_bios_map_sg()
227 if (bio->bi_rw & REQ_WRITE_SAME) { in __blk_bios_map_sg()
435 WARN_ON_ONCE((bio->bi_rw & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge()
436 (bio->bi_rw & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge()
437 bio->bi_rw |= ff; in blk_rq_set_mixed_merge()
572 if (!blk_check_merge_flags(rq->cmd_flags, bio->bi_rw)) in blk_rq_merge_ok()
Dbio.c567 bio->bi_rw = bio_src->bi_rw; in __bio_clone_fast()
649 bio->bi_rw = bio_src->bi_rw; in bio_clone_bioset()
653 if (bio->bi_rw & REQ_DISCARD) in bio_clone_bioset()
656 if (bio->bi_rw & REQ_WRITE_SAME) { in bio_clone_bioset()
741 .bi_rw = bio->bi_rw, in __bio_add_page()
802 .bi_rw = bio->bi_rw, in __bio_add_page()
1207 bio->bi_rw |= REQ_WRITE; in bio_copy_user_iov()
1377 bio->bi_rw |= REQ_WRITE; in bio_map_user_iov()
1570 bio->bi_rw |= REQ_WRITE; in bio_copy_kern()
1836 if (bio->bi_rw & REQ_DISCARD) in bio_split()
Dblk-core.c938 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) in blk_rq_should_init_elevator()
1465 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_back_merge()
1487 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_front_merge()
1573 req->cmd_flags |= bio->bi_rw & REQ_COMMON_MASK; in init_request_from_bio()
1574 if (bio->bi_rw & REQ_RAHEAD) in init_request_from_bio()
1585 const bool sync = !!(bio->bi_rw & REQ_SYNC); in blk_queue_bio()
1603 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) { in blk_queue_bio()
1718 bio->bi_rw, in handle_bad_sector()
1842 if ((bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && !q->flush_flags) { in generic_make_request_checks()
1843 bio->bi_rw &= ~(REQ_FLUSH | REQ_FUA); in generic_make_request_checks()
[all …]
Dblk-throttle.c1004 if (!(bio->bi_rw & REQ_THROTTLED)) { in throtl_charge_bio()
1005 bio->bi_rw |= REQ_THROTTLED; in throtl_charge_bio()
1007 bio->bi_iter.bi_size, bio->bi_rw); in throtl_charge_bio()
1491 if (bio->bi_rw & REQ_THROTTLED) in blk_throtl_bio()
1505 bio->bi_iter.bi_size, bio->bi_rw); in blk_throtl_bio()
1592 bio->bi_rw &= ~REQ_THROTTLED; in blk_throtl_bio()
Dblk-map.c205 bio->bi_rw |= REQ_WRITE; in blk_rq_map_kern()
Dblk-mq.c1213 if (rw_is_sync(bio->bi_rw)) in blk_mq_map_request()
1247 const int is_sync = rw_is_sync(bio->bi_rw); in blk_mq_make_request()
1248 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_mq_make_request()
1323 const int is_sync = rw_is_sync(bio->bi_rw); in blk_sq_make_request()
1324 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_sq_make_request()
Dbio-integrity.c82 bio->bi_rw |= REQ_INTEGRITY; in bio_integrity_alloc()
Dcfq-iosched.c900 return bio_data_dir(bio) == READ || (bio->bi_rw & REQ_SYNC); in cfq_bio_sync()
2366 cfqg_stats_update_io_merged(RQ_CFQG(req), bio->bi_rw); in cfq_bio_merged()
/linux-4.1.27/include/trace/events/
Dblock.h275 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
313 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
340 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
407 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
436 bio ? bio->bi_rw : 0, __entry->nr_sector);
570 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
613 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
Dbcache.h30 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
104 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
139 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
170 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
/linux-4.1.27/drivers/md/
Ddm-raid1.c261 .bi_rw = WRITE_FLUSH, in mirror_flush()
528 .bi_rw = READ, in read_async_bio()
611 if (bio->bi_rw & REQ_DISCARD) { in write_callback()
640 .bi_rw = WRITE | (bio->bi_rw & WRITE_FLUSH_FUA), in do_write()
648 if (bio->bi_rw & REQ_DISCARD) { in do_write()
649 io_req.bi_rw |= REQ_DISCARD; in do_write()
687 if ((bio->bi_rw & REQ_FLUSH) || in do_writes()
688 (bio->bi_rw & REQ_DISCARD)) { in do_writes()
1218 if (!(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) in mirror_end_io()
1226 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in mirror_end_io()
Ddm-stats.c446 unsigned long bi_rw, sector_t len, bool merged, in dm_stat_for_entry() argument
449 unsigned long idx = bi_rw & REQ_WRITE; in dm_stat_for_entry()
495 static void __dm_stat_bio(struct dm_stat *s, unsigned long bi_rw, in __dm_stat_bio() argument
525 dm_stat_for_entry(s, entry, bi_rw, fragment_len, in __dm_stat_bio()
533 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw, in dm_stats_account_io() argument
554 ((bi_rw & (REQ_WRITE | REQ_DISCARD)) == in dm_stats_account_io()
558 ACCESS_ONCE(last->last_rw) = bi_rw; in dm_stats_account_io()
564 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration, stats_aux); in dm_stats_account_io()
Ddm-stripe.c287 if (bio->bi_rw & REQ_FLUSH) { in stripe_map()
293 if (unlikely(bio->bi_rw & REQ_DISCARD) || in stripe_map()
294 unlikely(bio->bi_rw & REQ_WRITE_SAME)) { in stripe_map()
361 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in stripe_end_io()
Ddm-stats.h31 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
Dmultipath.c93 else if (!(bio->bi_rw & REQ_RAHEAD)) { in multipath_end_request()
114 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in multipath_make_request()
137 mp_bh->bio.bi_rw |= REQ_FAILFAST_TRANSPORT; in multipath_make_request()
368 bio->bi_rw |= REQ_FAILFAST_TRANSPORT; in multipathd()
Ddm-io.c482 if ((io_req->bi_rw & RW_MASK) == READ) { in dp_init()
520 io_req->bi_rw, &dp, sync_error_bits); in dm_io()
522 return async_io(io_req->client, num_regions, where, io_req->bi_rw, in dm_io()
Draid1.c793 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes()
1068 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid1_unplug()
1089 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in make_request()
1090 const unsigned long do_flush_fua = (bio->bi_rw & (REQ_FLUSH | REQ_FUA)); in make_request()
1091 const unsigned long do_discard = (bio->bi_rw in make_request()
1093 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in make_request()
1200 read_bio->bi_rw = READ | do_sync; in make_request()
1411 mbio->bi_rw = in make_request()
2073 wbio->bi_rw = WRITE; in sync_request_write()
2244 wbio->bi_rw = WRITE; in narrow_write_error()
[all …]
Dlinear.c266 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in linear_make_request()
297 if (unlikely((split->bi_rw & REQ_DISCARD) && in linear_make_request()
Ddm-flakey.c19 (((bio)->bi_rw & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags)
269 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_rw, in corrupt_bio_data()
Ddm-region-hash.c403 if (bio->bi_rw & REQ_FLUSH) { in dm_rh_mark_nosync()
408 if (bio->bi_rw & REQ_DISCARD) in dm_rh_mark_nosync()
531 if (bio->bi_rw & (REQ_FLUSH | REQ_DISCARD)) in dm_rh_inc_pending()
Draid10.c957 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes()
1133 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid10_unplug()
1151 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in __make_request()
1152 const unsigned long do_fua = (bio->bi_rw & REQ_FUA); in __make_request()
1153 const unsigned long do_discard = (bio->bi_rw in __make_request()
1155 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in __make_request()
1248 read_bio->bi_rw = READ | do_sync; in __make_request()
1457 mbio->bi_rw = in __make_request()
1500 mbio->bi_rw = in __make_request()
1544 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in make_request()
[all …]
Ddm.c656 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in start_io_acct()
671 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in end_io_acct()
934 if ((bio->bi_rw & REQ_FLUSH) && bio->bi_iter.bi_size) { in dec_pending()
939 bio->bi_rw &= ~REQ_FLUSH; in dec_pending()
985 if (unlikely(r == -EREMOTEIO && (bio->bi_rw & REQ_WRITE_SAME) && in clone_endio()
1422 BUG_ON(bio->bi_rw & REQ_FLUSH); in dm_accept_partial_bio()
1652 if (unlikely(bio->bi_rw & REQ_DISCARD)) in __split_and_process_non_flush()
1654 else if (unlikely(bio->bi_rw & REQ_WRITE_SAME)) in __split_and_process_non_flush()
1697 if (bio->bi_rw & REQ_FLUSH) { in __split_and_process_bio()
2334 md->flush_bio.bi_rw = WRITE_FLUSH; in alloc_dev()
Ddm-log-writes.c553 bool flush_bio = (bio->bi_rw & REQ_FLUSH); in log_writes_map()
554 bool fua_bio = (bio->bi_rw & REQ_FUA); in log_writes_map()
555 bool discard_bio = (bio->bi_rw & REQ_DISCARD); in log_writes_map()
Ddm-log.c296 lc->io_req.bi_rw = rw; in rw_header()
309 lc->io_req.bi_rw = WRITE_FLUSH; in flush_header()
Draid0.c516 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in raid0_make_request()
546 if (unlikely((split->bi_rw & REQ_DISCARD) && in raid0_make_request()
Ddm-thin.c585 return (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && in bio_triggers_commit()
593 if (bio->bi_rw & REQ_DISCARD) in inc_all_io_entry()
754 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) in __inc_remap_and_issue_cell()
1429 (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA))) in __remap_and_issue_shared_cell()
1818 if (bio->bi_rw & REQ_DISCARD) in process_thin_deferred_bios()
1905 if (cell->holder->bi_rw & REQ_DISCARD) in process_thin_deferred_cells()
2309 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) { in thin_bio_map()
Draid5.c809 if (head->dev[dd_idx].towrite->bi_rw != sh->dev[dd_idx].towrite->bi_rw) in stripe_add_to_batch_list()
996 bi->bi_rw = rw; in ops_run_io()
1004 bi->bi_rw, i); in ops_run_io()
1015 bi->bi_rw |= REQ_NOMERGE; in ops_run_io()
1048 rbi->bi_rw = rw; in ops_run_io()
1056 rbi->bi_rw, i); in ops_run_io()
1089 bi->bi_rw, i, (unsigned long long)sh->sector); in ops_run_io()
1622 if (wbi->bi_rw & REQ_FUA) in ops_run_biodrain()
1624 if (wbi->bi_rw & REQ_SYNC) in ops_run_biodrain()
1626 if (wbi->bi_rw & REQ_DISCARD) in ops_run_biodrain()
[all …]
Ddm-cache-target.c741 !(bio->bi_rw & (REQ_FUA | REQ_FLUSH | REQ_DISCARD))) { in check_if_tick_bio_needed()
782 return bio->bi_rw & (REQ_FLUSH | REQ_FUA); in bio_triggers_commit()
1630 if (bio->bi_rw & REQ_FLUSH) in process_deferred_bios()
1632 else if (bio->bi_rw & REQ_DISCARD) in process_deferred_bios()
2627 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA | REQ_DISCARD)) { in __cache_map()
Ddm-snap.c1669 if (bio->bi_rw & REQ_FLUSH) { in snapshot_map()
1784 if (bio->bi_rw & REQ_FLUSH) { in snapshot_merge_map()
2268 if (unlikely(bio->bi_rw & REQ_FLUSH)) in origin_map()
Ddm-crypt.c1122 clone->bi_rw = io->base_bio->bi_rw; in clone_init()
1893 if (unlikely(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) { in crypt_map()
Ddm-bufio.c556 .bi_rw = rw, in use_dmio()
1299 .bi_rw = WRITE_FLUSH, in dm_bufio_issue_flush()
Ddm-snap-persistent.c237 .bi_rw = rw, in chunk_io()
Ddm-kcopyd.c499 .bi_rw = job->rw, in run_io_job()
Ddm-era-target.c1546 if (!(bio->bi_rw & REQ_FLUSH) && in era_map()
Dmd.c293 bio->bi_rw &= ~REQ_NOMERGE; in md_make_request()
440 bio->bi_rw &= ~REQ_FLUSH; in md_submit_flush_data()
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Dlloop.c215 rw = head->bi_rw; in do_bio_lustrebacked()
217 LASSERT(rw == bio->bi_rw); in do_bio_lustrebacked()
308 rw = first->bi_rw; in loop_get_bio()
310 while (*bio && (*bio)->bi_rw == rw) { in loop_get_bio()
/linux-4.1.27/Documentation/device-mapper/
Ddm-flakey.txt45 <flags>: Perform the replacement only if bio->bi_rw has all the
/linux-4.1.27/drivers/block/
Dpmem.c75 BUG_ON(bio->bi_rw & REQ_DISCARD); in pmem_make_request()
Dumem.c465 } else if ((bio->bi_rw & REQ_WRITE) && in process_page()
538 if (bio->bi_rw & REQ_SYNC || !mm_check_plugged(card)) in mm_make_request()
Dbrd.c340 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in brd_make_request()
Dpktcdvd.c1076 bio->bi_rw = READ; in pkt_gather_data()
1159 pkt->bio->bi_rw = REQ_WRITE; in pkt_start_recovery()
1338 pkt->w_bio->bi_rw = WRITE; in pkt_start_write()
Dxen-blkfront.c1583 submit_bio(cloned_bio->bi_rw, cloned_bio); in blkif_recover()
1592 submit_bio(bio->bi_rw, bio); in blkif_recover()
/linux-4.1.27/drivers/scsi/osd/
Dosd_initiator.c732 bio->bi_rw &= ~REQ_WRITE; in _osd_req_list_objects()
830 WARN_ON(0 == (bio->bi_rw & REQ_WRITE)); in osd_req_write()
845 bio->bi_rw |= REQ_WRITE; /* FIXME: bio_set_dir() */ in osd_req_write_kern()
881 WARN_ON(bio->bi_rw & REQ_WRITE); in osd_req_read()
962 bio->bi_rw |= REQ_WRITE; in _osd_req_finalize_cdb_cont()
1083 bio->bi_rw |= REQ_WRITE; in osd_req_write_sg_kern()
/linux-4.1.27/drivers/block/drbd/
Ddrbd_main.c1605 static u32 bio_flags_to_wire(struct drbd_connection *connection, unsigned long bi_rw) in bio_flags_to_wire() argument
1608 return (bi_rw & REQ_SYNC ? DP_RW_SYNC : 0) | in bio_flags_to_wire()
1609 (bi_rw & REQ_FUA ? DP_FUA : 0) | in bio_flags_to_wire()
1610 (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) | in bio_flags_to_wire()
1611 (bi_rw & REQ_DISCARD ? DP_DISCARD : 0); in bio_flags_to_wire()
1613 return bi_rw & REQ_SYNC ? DP_RW_SYNC : 0; in bio_flags_to_wire()
1638 dp_flags = bio_flags_to_wire(peer_device->connection, req->master_bio->bi_rw); in drbd_send_dblock()
Ddrbd_worker.c179 int is_discard = !!(bio->bi_rw & REQ_DISCARD); in drbd_peer_request_endio()
268 if (bio->bi_rw & REQ_DISCARD) in drbd_request_endio()
Ddrbd_actlog.c162 bio->bi_rw = rw; in _drbd_md_sync_page_io()
Ddrbd_bitmap.c1033 bio->bi_rw |= rw; in bm_page_io_async()
Ddrbd_req.c1116 D_ASSERT(device, req->master_bio->bi_rw & REQ_FLUSH); in drbd_process_write_request()
Ddrbd_receiver.c1420 bio->bi_rw = rw; in drbd_submit_peer_request()
/linux-4.1.27/kernel/trace/
Dblktrace.c787 bio->bi_rw, what, error, 0, NULL); in blk_add_trace_bio()
891 bio->bi_iter.bi_size, bio->bi_rw, BLK_TA_SPLIT, in blk_add_trace_split()
925 bio->bi_rw, BLK_TA_REMAP, in blk_add_trace_bio_remap()
/linux-4.1.27/fs/btrfs/
Dvolumes.c367 btrfsic_submit_bio(cur->bi_rw, cur); in run_scheduled_bios()
5612 if (bio->bi_rw & WRITE) in btrfs_end_bio()
5618 if ((bio->bi_rw & WRITE_FLUSH) == WRITE_FLUSH) in btrfs_end_bio()
5696 bio->bi_rw |= rw; in btrfs_schedule_bio()
5699 if (bio->bi_rw & REQ_SYNC) in btrfs_schedule_bio()
5729 .bi_rw = bio->bi_rw, in bio_size_ok()
Ddisk-io.c716 if (bio->bi_rw & REQ_WRITE) { in end_workqueue_bio()
Dinode.c7691 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in dio_read_error()
7977 btrfs_ino(dip->inode), bio->bi_rw, in btrfs_end_dio_bio()
Dextent_io.c2408 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in bio_readpage_error()
/linux-4.1.27/drivers/target/
Dtarget_core_pscsi.c920 bio->bi_rw |= REQ_WRITE; in pscsi_map_sg()
/linux-4.1.27/drivers/block/rsxx/
Ddma.c708 if (bio->bi_rw & REQ_DISCARD) { in rsxx_dma_queue_bio()
/linux-4.1.27/fs/exofs/
Dore.c881 bio->bi_rw |= REQ_WRITE; in _write_mirror()
/linux-4.1.27/drivers/block/zram/
Dzram_drv.c933 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in __zram_make_request()
/linux-4.1.27/Documentation/block/
Dbiodoc.txt272 requests. Some bits in the bi_rw flags field in the bio structure are
435 unsigned long bi_rw; /* low bits: r/w, high: priority */