Home
last modified time | relevance | path

Searched refs:bi_rw (Results 1 – 73 of 73) sorted by relevance

/linux-4.4.14/block/
Dblk-merge.c173 if ((*bio)->bi_rw & REQ_DISCARD) in blk_queue_split()
175 else if ((*bio)->bi_rw & REQ_WRITE_SAME) in blk_queue_split()
187 split->bi_rw |= REQ_NOMERGE; in blk_queue_split()
213 if (bio->bi_rw & REQ_DISCARD) in __blk_recalc_rq_segments()
216 if (bio->bi_rw & REQ_WRITE_SAME) in __blk_recalc_rq_segments()
389 if (bio->bi_rw & REQ_DISCARD) { in __blk_bios_map_sg()
404 if (bio->bi_rw & REQ_WRITE_SAME) { in __blk_bios_map_sg()
620 WARN_ON_ONCE((bio->bi_rw & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge()
621 (bio->bi_rw & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge()
622 bio->bi_rw |= ff; in blk_rq_set_mixed_merge()
[all …]
Dblk-core.c1024 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) in blk_rq_should_init_elevator()
1546 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_back_merge()
1568 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_front_merge()
1690 req->cmd_flags |= bio->bi_rw & REQ_COMMON_MASK; in init_request_from_bio()
1691 if (bio->bi_rw & REQ_RAHEAD) in init_request_from_bio()
1702 const bool sync = !!(bio->bi_rw & REQ_SYNC); in blk_queue_bio()
1723 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) { in blk_queue_bio()
1843 bio->bi_rw, in handle_bad_sector()
1956 if ((bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && !q->flush_flags) { in generic_make_request_checks()
1957 bio->bi_rw &= ~(REQ_FLUSH | REQ_FUA); in generic_make_request_checks()
[all …]
Dbio.c584 bio->bi_rw = bio_src->bi_rw; in __bio_clone_fast()
666 bio->bi_rw = bio_src->bi_rw; in bio_clone_bioset()
670 if (bio->bi_rw & REQ_DISCARD) in bio_clone_bioset()
673 if (bio->bi_rw & REQ_WRITE_SAME) { in bio_clone_bioset()
1175 bio->bi_rw |= REQ_WRITE; in bio_copy_user_iov()
1345 bio->bi_rw |= REQ_WRITE; in bio_map_user_iov()
1538 bio->bi_rw |= REQ_WRITE; in bio_copy_kern()
1797 if (bio->bi_rw & REQ_DISCARD) in bio_split()
Dblk-map.c227 bio->bi_rw |= REQ_WRITE; in blk_rq_map_kern()
Dblk-throttle.c830 if (!(bio->bi_rw & REQ_THROTTLED)) in throtl_charge_bio()
831 bio->bi_rw |= REQ_THROTTLED; in throtl_charge_bio()
1408 if ((bio->bi_rw & REQ_THROTTLED) || !tg->has_rules[rw]) in blk_throtl_bio()
1487 bio->bi_rw &= ~REQ_THROTTLED; in blk_throtl_bio()
Dblk-mq.c1174 if (rw_is_sync(bio->bi_rw)) in blk_mq_map_request()
1244 const int is_sync = rw_is_sync(bio->bi_rw); in blk_mq_make_request()
1245 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_mq_make_request()
1341 const int is_sync = rw_is_sync(bio->bi_rw); in blk_sq_make_request()
1342 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_sq_make_request()
Dbio-integrity.c87 bio->bi_rw |= REQ_INTEGRITY; in bio_integrity_alloc()
Dcfq-iosched.c901 return bio_data_dir(bio) == READ || (bio->bi_rw & REQ_SYNC); in cfq_bio_sync()
2543 cfqg_stats_update_io_merged(RQ_CFQG(req), bio->bi_rw); in cfq_bio_merged()
/linux-4.4.14/include/linux/
Dbio.h51 #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT)
56 (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \
57 (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \
109 !(bio->bi_rw & REQ_DISCARD)) in bio_has_data()
120 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_is_rw()
128 if (bio->bi_rw & REQ_NOMERGE_FLAGS) in bio_mergeable()
228 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_advance_iter()
256 if (bio->bi_rw & REQ_DISCARD) in bio_segments()
259 if (bio->bi_rw & REQ_WRITE_SAME) in bio_segments()
357 if (bio->bi_rw & REQ_INTEGRITY) in bio_integrity()
Ddm-io.h60 int bi_rw; /* READ|WRITE - not READA */ member
Dblk_types.h51 unsigned long bi_rw; /* bottom bits READ/WRITE, member
Dblk-cgroup.h716 blkg_rwstat_add(&blkg->stat_bytes, bio->bi_rw, in blkcg_bio_issue_check()
718 blkg_rwstat_add(&blkg->stat_ios, bio->bi_rw, 1); in blkcg_bio_issue_check()
Dfs.h2390 #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK))
2395 #define bio_data_dir(bio) ((bio)->bi_rw & 1)
/linux-4.4.14/drivers/md/bcache/
Drequest.c211 bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); in bch_data_insert_start()
256 n->bi_rw |= REQ_WRITE; in bch_data_insert_start()
381 (bio->bi_rw & REQ_DISCARD)) in check_should_bypass()
386 (bio->bi_rw & REQ_WRITE))) in check_should_bypass()
407 (bio->bi_rw & REQ_WRITE) && in check_should_bypass()
408 (bio->bi_rw & REQ_SYNC)) in check_should_bypass()
660 s->write = (bio->bi_rw & REQ_WRITE) != 0; in search_alloc()
671 s->iop.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; in search_alloc()
799 if (!(bio->bi_rw & REQ_RAHEAD) && in cached_dev_cache_miss()
800 !(bio->bi_rw & REQ_META) && in cached_dev_cache_miss()
[all …]
Dwriteback.h60 return bio->bi_rw & REQ_SYNC || in should_writeback()
Dio.c114 unsigned threshold = bio->bi_rw & REQ_WRITE in bch_bbio_count_io_errors()
Dwriteback.c186 io->bio.bi_rw = WRITE; in write_dirty()
259 io->bio.bi_rw = READ; in read_dirty()
Djournal.c57 bio->bi_rw = READ; in journal_read_bucket()
455 bio->bi_rw = REQ_WRITE|REQ_DISCARD; in do_journal_discard()
629 bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; in journal_write_unlocked()
Dmovinggc.c166 bio->bi_rw = READ; in read_moving()
Dsuper.c215 bio->bi_rw = REQ_SYNC|REQ_META; in __write_super()
351 bio->bi_rw = REQ_SYNC|REQ_META|rw; in uuid_io()
510 bio->bi_rw = REQ_SYNC|REQ_META|rw; in prio_io()
Dbtree.c298 bio->bi_rw = REQ_META|READ_SYNC; in bch_btree_node_read()
400 b->bio->bi_rw = REQ_META|WRITE_SYNC|REQ_FUA; in do_btree_node_write()
/linux-4.4.14/include/trace/events/
Dblock.h275 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
313 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
340 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
407 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
436 bio ? bio->bi_rw : 0, __entry->nr_sector);
570 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
613 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
Dbcache.h30 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
104 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
139 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
170 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
/linux-4.4.14/drivers/md/
Ddm-raid1.c263 .bi_rw = WRITE_FLUSH, in mirror_flush()
544 .bi_rw = READ, in read_async_bio()
627 if (bio->bi_rw & REQ_DISCARD) { in write_callback()
657 .bi_rw = WRITE | (bio->bi_rw & WRITE_FLUSH_FUA), in do_write()
665 if (bio->bi_rw & REQ_DISCARD) { in do_write()
666 io_req.bi_rw |= REQ_DISCARD; in do_write()
704 if ((bio->bi_rw & REQ_FLUSH) || in do_writes()
705 (bio->bi_rw & REQ_DISCARD)) { in do_writes()
1253 if (!(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) in mirror_end_io()
1261 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in mirror_end_io()
Ddm-stripe.c289 if (bio->bi_rw & REQ_FLUSH) { in stripe_map()
295 if (unlikely(bio->bi_rw & REQ_DISCARD) || in stripe_map()
296 unlikely(bio->bi_rw & REQ_WRITE_SAME)) { in stripe_map()
363 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in stripe_end_io()
Ddm-stats.h32 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
Dmultipath.c93 else if (!(bio->bi_rw & REQ_RAHEAD)) { in multipath_end_request()
114 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in multipath_make_request()
137 mp_bh->bio.bi_rw |= REQ_FAILFAST_TRANSPORT; in multipath_make_request()
358 bio->bi_rw |= REQ_FAILFAST_TRANSPORT; in multipathd()
Ddm-stats.c517 unsigned long bi_rw, sector_t len, in dm_stat_for_entry() argument
521 unsigned long idx = bi_rw & REQ_WRITE; in dm_stat_for_entry()
587 static void __dm_stat_bio(struct dm_stat *s, unsigned long bi_rw, in __dm_stat_bio() argument
617 dm_stat_for_entry(s, entry, bi_rw, fragment_len, in __dm_stat_bio()
625 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw, in dm_stats_account_io() argument
648 ((bi_rw & (REQ_WRITE | REQ_DISCARD)) == in dm_stats_account_io()
652 ACCESS_ONCE(last->last_rw) = bi_rw; in dm_stats_account_io()
666 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration_jiffies, stats_aux); in dm_stats_account_io()
Dlinear.c224 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in linear_make_request()
255 if (unlikely((split->bi_rw & REQ_DISCARD) && in linear_make_request()
Ddm-io.c483 if ((io_req->bi_rw & RW_MASK) == READ) { in dp_init()
521 io_req->bi_rw, &dp, sync_error_bits); in dm_io()
523 return async_io(io_req->client, num_regions, where, io_req->bi_rw, in dm_io()
Draid1.c762 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes()
1036 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid1_unplug()
1057 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in make_request()
1058 const unsigned long do_flush_fua = (bio->bi_rw & (REQ_FLUSH | REQ_FUA)); in make_request()
1059 const unsigned long do_discard = (bio->bi_rw in make_request()
1061 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in make_request()
1169 read_bio->bi_rw = READ | do_sync; in make_request()
1379 mbio->bi_rw = in make_request()
2033 wbio->bi_rw = WRITE; in sync_request_write()
2204 wbio->bi_rw = WRITE; in narrow_write_error()
[all …]
Ddm-flakey.c19 (((bio)->bi_rw & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags)
271 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_rw, in corrupt_bio_data()
Ddm-region-hash.c401 if (bio->bi_rw & REQ_FLUSH) { in dm_rh_mark_nosync()
406 if (bio->bi_rw & REQ_DISCARD) in dm_rh_mark_nosync()
529 if (bio->bi_rw & (REQ_FLUSH | REQ_DISCARD)) in dm_rh_inc_pending()
Draid10.c868 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes()
1044 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid10_unplug()
1062 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in __make_request()
1063 const unsigned long do_fua = (bio->bi_rw & REQ_FUA); in __make_request()
1064 const unsigned long do_discard = (bio->bi_rw in __make_request()
1066 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in __make_request()
1159 read_bio->bi_rw = READ | do_sync; in __make_request()
1366 mbio->bi_rw = in __make_request()
1409 mbio->bi_rw = in __make_request()
1453 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in make_request()
[all …]
Ddm-log-writes.c555 bool flush_bio = (bio->bi_rw & REQ_FLUSH); in log_writes_map()
556 bool fua_bio = (bio->bi_rw & REQ_FUA); in log_writes_map()
557 bool discard_bio = (bio->bi_rw & REQ_DISCARD); in log_writes_map()
Ddm.c689 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in start_io_acct()
704 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in end_io_acct()
967 if ((bio->bi_rw & REQ_FLUSH) && bio->bi_iter.bi_size) { in dec_pending()
972 bio->bi_rw &= ~REQ_FLUSH; in dec_pending()
1017 if (unlikely(r == -EREMOTEIO && (bio->bi_rw & REQ_WRITE_SAME) && in clone_endio()
1462 BUG_ON(bio->bi_rw & REQ_FLUSH); in dm_accept_partial_bio()
1692 if (unlikely(bio->bi_rw & REQ_DISCARD)) in __split_and_process_non_flush()
1694 else if (unlikely(bio->bi_rw & REQ_WRITE_SAME)) in __split_and_process_non_flush()
1737 if (bio->bi_rw & REQ_FLUSH) { in __split_and_process_bio()
2362 md->flush_bio.bi_rw = WRITE_FLUSH; in alloc_dev()
Draid0.c461 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in raid0_make_request()
491 if (unlikely((split->bi_rw & REQ_DISCARD) && in raid0_make_request()
Ddm-log.c296 lc->io_req.bi_rw = rw; in rw_header()
309 lc->io_req.bi_rw = WRITE_FLUSH; in flush_header()
Ddm-thin.c698 return (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && in bio_triggers_commit()
706 if (bio->bi_rw & REQ_DISCARD) in inc_all_io_entry()
869 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) in __inc_remap_and_issue_cell()
1647 (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA))) in __remap_and_issue_shared_cell()
2036 if (bio->bi_rw & REQ_DISCARD) in process_thin_deferred_bios()
2123 if (cell->holder->bi_rw & REQ_DISCARD) in process_thin_deferred_cells()
2559 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) { in thin_bio_map()
Draid5.c811 if (head->dev[dd_idx].towrite->bi_rw != sh->dev[dd_idx].towrite->bi_rw) in stripe_add_to_batch_list()
1000 bi->bi_rw = rw; in ops_run_io()
1008 bi->bi_rw, i); in ops_run_io()
1019 bi->bi_rw |= REQ_NOMERGE; in ops_run_io()
1052 rbi->bi_rw = rw; in ops_run_io()
1060 rbi->bi_rw, i); in ops_run_io()
1093 bi->bi_rw, i, (unsigned long long)sh->sector); in ops_run_io()
1624 if (wbi->bi_rw & REQ_FUA) in ops_run_biodrain()
1626 if (wbi->bi_rw & REQ_SYNC) in ops_run_biodrain()
1628 if (wbi->bi_rw & REQ_DISCARD) in ops_run_biodrain()
[all …]
Ddm-cache-target.c794 !(bio->bi_rw & (REQ_FUA | REQ_FLUSH | REQ_DISCARD))) { in check_if_tick_bio_needed()
835 return bio->bi_rw & (REQ_FLUSH | REQ_FUA); in bio_triggers_commit()
857 !(bio->bi_rw & REQ_DISCARD)); in accountable_bio()
1073 return bio->bi_rw & (REQ_FLUSH | REQ_FUA | REQ_DISCARD); in discard_or_flush()
1984 if (bio->bi_rw & REQ_FLUSH) in process_deferred_bios()
1986 else if (bio->bi_rw & REQ_DISCARD) in process_deferred_bios()
Ddm-crypt.c1136 clone->bi_rw = io->base_bio->bi_rw; in clone_init()
1915 if (unlikely(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) { in crypt_map()
Draid5-cache.c260 bio->bi_rw = WRITE; in r5l_bio_alloc()
515 bio->bi_rw &= ~REQ_FLUSH; in r5l_handle_flush_request()
Ddm-snap.c1687 if (bio->bi_rw & REQ_FLUSH) { in snapshot_map()
1806 if (bio->bi_rw & REQ_FLUSH) { in snapshot_merge_map()
2292 if (unlikely(bio->bi_rw & REQ_FLUSH)) in origin_map()
Ddm-bufio.c557 .bi_rw = rw, in use_dmio()
1305 .bi_rw = WRITE_FLUSH, in dm_bufio_issue_flush()
Ddm-snap-persistent.c238 .bi_rw = rw, in chunk_io()
Ddm-kcopyd.c499 .bi_rw = job->rw, in run_io_job()
Ddm-era-target.c1545 if (!(bio->bi_rw & REQ_FLUSH) && in era_map()
Dmd.c297 bio->bi_rw &= ~REQ_NOMERGE; in md_make_request()
424 bio->bi_rw &= ~REQ_FLUSH; in md_submit_flush_data()
/linux-4.4.14/drivers/staging/lustre/lustre/llite/
Dlloop.c215 rw = head->bi_rw; in do_bio_lustrebacked()
217 LASSERT(rw == bio->bi_rw); in do_bio_lustrebacked()
308 rw = first->bi_rw; in loop_get_bio()
310 while (*bio && (*bio)->bi_rw == rw) { in loop_get_bio()
/linux-4.4.14/Documentation/device-mapper/
Ddm-flakey.txt45 <flags>: Perform the replacement only if bio->bi_rw has all the
/linux-4.4.14/drivers/scsi/osd/
Dosd_initiator.c732 bio->bi_rw &= ~REQ_WRITE; in _osd_req_list_objects()
830 WARN_ON(0 == (bio->bi_rw & REQ_WRITE)); in osd_req_write()
845 bio->bi_rw |= REQ_WRITE; /* FIXME: bio_set_dir() */ in osd_req_write_kern()
881 WARN_ON(bio->bi_rw & REQ_WRITE); in osd_req_read()
962 bio->bi_rw |= REQ_WRITE; in _osd_req_finalize_cdb_cont()
1083 bio->bi_rw |= REQ_WRITE; in osd_req_write_sg_kern()
/linux-4.4.14/drivers/lightnvm/
Drrpc.c318 bio->bi_rw = READ; in rrpc_move_valid_pages()
336 bio->bi_rw = WRITE; in rrpc_move_valid_pages()
856 if (bio->bi_rw & REQ_DISCARD) { in rrpc_make_rq()
/linux-4.4.14/drivers/block/drbd/
Ddrbd_main.c1605 static u32 bio_flags_to_wire(struct drbd_connection *connection, unsigned long bi_rw) in bio_flags_to_wire() argument
1608 return (bi_rw & REQ_SYNC ? DP_RW_SYNC : 0) | in bio_flags_to_wire()
1609 (bi_rw & REQ_FUA ? DP_FUA : 0) | in bio_flags_to_wire()
1610 (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) | in bio_flags_to_wire()
1611 (bi_rw & REQ_DISCARD ? DP_DISCARD : 0); in bio_flags_to_wire()
1613 return bi_rw & REQ_SYNC ? DP_RW_SYNC : 0; in bio_flags_to_wire()
1638 dp_flags = bio_flags_to_wire(peer_device->connection, req->master_bio->bi_rw); in drbd_send_dblock()
Ddrbd_worker.c178 int is_discard = !!(bio->bi_rw & REQ_DISCARD); in drbd_peer_request_endio()
246 if (bio->bi_rw & REQ_DISCARD) in drbd_request_endio()
Ddrbd_actlog.c162 bio->bi_rw = rw; in _drbd_md_sync_page_io()
Ddrbd_bitmap.c1024 bio->bi_rw |= rw; in bm_page_io_async()
Ddrbd_req.c1117 D_ASSERT(device, req->master_bio->bi_rw & REQ_FLUSH); in drbd_process_write_request()
Ddrbd_receiver.c1421 bio->bi_rw = rw; in drbd_submit_peer_request()
/linux-4.4.14/kernel/trace/
Dblktrace.c779 bio->bi_rw, what, error, 0, NULL); in blk_add_trace_bio()
883 bio->bi_iter.bi_size, bio->bi_rw, BLK_TA_SPLIT, in blk_add_trace_split()
916 bio->bi_rw, BLK_TA_REMAP, bio->bi_error, in blk_add_trace_bio_remap()
/linux-4.4.14/drivers/block/
Dumem.c465 } else if ((bio->bi_rw & REQ_WRITE) && in process_page()
540 if (bio->bi_rw & REQ_SYNC || !mm_check_plugged(card)) in mm_make_request()
Dbrd.c339 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in brd_make_request()
Dpktcdvd.c1077 bio->bi_rw = READ; in pkt_gather_data()
1160 pkt->bio->bi_rw = REQ_WRITE; in pkt_start_recovery()
1339 pkt->w_bio->bi_rw = WRITE; in pkt_start_write()
Dxen-blkfront.c1752 submit_bio(cloned_bio->bi_rw, cloned_bio); in blkif_recover()
1761 submit_bio(bio->bi_rw, bio); in blkif_recover()
/linux-4.4.14/drivers/target/
Dtarget_core_pscsi.c925 bio->bi_rw |= REQ_WRITE; in pscsi_map_sg()
/linux-4.4.14/drivers/block/rsxx/
Ddma.c708 if (bio->bi_rw & REQ_DISCARD) { in rsxx_dma_queue_bio()
/linux-4.4.14/fs/exofs/
Dore.c881 bio->bi_rw |= REQ_WRITE; in _write_mirror()
/linux-4.4.14/fs/btrfs/
Dvolumes.c449 btrfsic_submit_bio(cur->bi_rw, cur); in run_scheduled_bios()
5878 if (bio->bi_rw & WRITE) in btrfs_end_bio()
5884 if ((bio->bi_rw & WRITE_FLUSH) == WRITE_FLUSH) in btrfs_end_bio()
5959 bio->bi_rw |= rw; in btrfs_schedule_bio()
5962 if (bio->bi_rw & REQ_SYNC) in btrfs_schedule_bio()
Ddisk-io.c715 if (bio->bi_rw & REQ_WRITE) { in end_workqueue_bio()
Dinode.c7786 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in dio_read_error()
8068 btrfs_ino(dip->inode), bio->bi_rw, in btrfs_end_dio_bio()
Dextent_io.c2476 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in bio_readpage_error()
/linux-4.4.14/drivers/block/zram/
Dzram_drv.c852 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in __zram_make_request()
/linux-4.4.14/Documentation/block/
Dbiodoc.txt272 requests. Some bits in the bi_rw flags field in the bio structure are
435 unsigned long bi_rw; /* low bits: r/w, high: priority */