/linux-4.1.27/include/linux/ |
D | bio.h | 51 #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT) 56 (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \ 57 (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \ 109 !(bio->bi_rw & REQ_DISCARD)) in bio_has_data() 120 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_is_rw() 128 if (bio->bi_rw & REQ_NOMERGE_FLAGS) in bio_mergeable() 239 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_advance_iter() 267 if (bio->bi_rw & REQ_DISCARD) in bio_segments() 270 if (bio->bi_rw & REQ_WRITE_SAME) in bio_segments() 307 if (bio->bi_rw & REQ_INTEGRITY) in bio_integrity()
|
D | dm-io.h | 60 int bi_rw; /* READ|WRITE - not READA */ member
|
D | blk_types.h | 50 unsigned long bi_rw; /* bottom bits READ/WRITE, member
|
D | fs.h | 2356 #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK)) 2361 #define bio_data_dir(bio) ((bio)->bi_rw & 1)
|
D | blkdev.h | 246 unsigned long bi_rw; member
|
/linux-4.1.27/drivers/md/bcache/ |
D | io.c | 21 if (bio->bi_rw & REQ_DISCARD) in bch_bio_max_sectors() 29 .bi_rw = bio->bi_rw, in bch_bio_max_sectors() 81 if (!bio_has_data(bio) && !(bio->bi_rw & REQ_DISCARD)) in bch_generic_make_request() 212 unsigned threshold = bio->bi_rw & REQ_WRITE in bch_bbio_count_io_errors()
|
D | request.c | 208 bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); in bch_data_insert_start() 251 n->bi_rw |= REQ_WRITE; in bch_data_insert_start() 375 (bio->bi_rw & REQ_DISCARD)) in check_should_bypass() 380 (bio->bi_rw & REQ_WRITE))) in check_should_bypass() 401 (bio->bi_rw & REQ_WRITE) && in check_should_bypass() 402 (bio->bi_rw & REQ_SYNC)) in check_should_bypass() 651 s->write = (bio->bi_rw & REQ_WRITE) != 0; in search_alloc() 662 s->iop.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; in search_alloc() 790 if (!(bio->bi_rw & REQ_RAHEAD) && in cached_dev_cache_miss() 791 !(bio->bi_rw & REQ_META) && in cached_dev_cache_miss() [all …]
|
D | writeback.h | 60 return bio->bi_rw & REQ_SYNC || in should_writeback()
|
D | writeback.c | 186 io->bio.bi_rw = WRITE; in write_dirty() 259 io->bio.bi_rw = READ; in read_dirty()
|
D | journal.c | 57 bio->bi_rw = READ; in journal_read_bucket() 455 bio->bi_rw = REQ_WRITE|REQ_DISCARD; in do_journal_discard() 627 bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; in journal_write_unlocked()
|
D | movinggc.c | 166 bio->bi_rw = READ; in read_moving()
|
D | super.c | 238 bio->bi_rw = REQ_SYNC|REQ_META; in __write_super() 374 bio->bi_rw = REQ_SYNC|REQ_META|rw; in uuid_io() 533 bio->bi_rw = REQ_SYNC|REQ_META|rw; in prio_io()
|
D | btree.c | 298 bio->bi_rw = REQ_META|READ_SYNC; in bch_btree_node_read() 400 b->bio->bi_rw = REQ_META|WRITE_SYNC|REQ_FUA; in do_btree_node_write()
|
/linux-4.1.27/block/ |
D | blk-merge.c | 29 if (bio->bi_rw & REQ_DISCARD) in __blk_recalc_rq_segments() 32 if (bio->bi_rw & REQ_WRITE_SAME) in __blk_recalc_rq_segments() 212 if (bio->bi_rw & REQ_DISCARD) { in __blk_bios_map_sg() 227 if (bio->bi_rw & REQ_WRITE_SAME) { in __blk_bios_map_sg() 435 WARN_ON_ONCE((bio->bi_rw & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge() 436 (bio->bi_rw & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge() 437 bio->bi_rw |= ff; in blk_rq_set_mixed_merge() 572 if (!blk_check_merge_flags(rq->cmd_flags, bio->bi_rw)) in blk_rq_merge_ok()
|
D | bio.c | 567 bio->bi_rw = bio_src->bi_rw; in __bio_clone_fast() 649 bio->bi_rw = bio_src->bi_rw; in bio_clone_bioset() 653 if (bio->bi_rw & REQ_DISCARD) in bio_clone_bioset() 656 if (bio->bi_rw & REQ_WRITE_SAME) { in bio_clone_bioset() 741 .bi_rw = bio->bi_rw, in __bio_add_page() 802 .bi_rw = bio->bi_rw, in __bio_add_page() 1207 bio->bi_rw |= REQ_WRITE; in bio_copy_user_iov() 1377 bio->bi_rw |= REQ_WRITE; in bio_map_user_iov() 1570 bio->bi_rw |= REQ_WRITE; in bio_copy_kern() 1836 if (bio->bi_rw & REQ_DISCARD) in bio_split()
|
D | blk-core.c | 938 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) in blk_rq_should_init_elevator() 1465 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_back_merge() 1487 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_front_merge() 1573 req->cmd_flags |= bio->bi_rw & REQ_COMMON_MASK; in init_request_from_bio() 1574 if (bio->bi_rw & REQ_RAHEAD) in init_request_from_bio() 1585 const bool sync = !!(bio->bi_rw & REQ_SYNC); in blk_queue_bio() 1603 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) { in blk_queue_bio() 1718 bio->bi_rw, in handle_bad_sector() 1842 if ((bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && !q->flush_flags) { in generic_make_request_checks() 1843 bio->bi_rw &= ~(REQ_FLUSH | REQ_FUA); in generic_make_request_checks() [all …]
|
D | blk-throttle.c | 1004 if (!(bio->bi_rw & REQ_THROTTLED)) { in throtl_charge_bio() 1005 bio->bi_rw |= REQ_THROTTLED; in throtl_charge_bio() 1007 bio->bi_iter.bi_size, bio->bi_rw); in throtl_charge_bio() 1491 if (bio->bi_rw & REQ_THROTTLED) in blk_throtl_bio() 1505 bio->bi_iter.bi_size, bio->bi_rw); in blk_throtl_bio() 1592 bio->bi_rw &= ~REQ_THROTTLED; in blk_throtl_bio()
|
D | blk-map.c | 205 bio->bi_rw |= REQ_WRITE; in blk_rq_map_kern()
|
D | blk-mq.c | 1213 if (rw_is_sync(bio->bi_rw)) in blk_mq_map_request() 1247 const int is_sync = rw_is_sync(bio->bi_rw); in blk_mq_make_request() 1248 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_mq_make_request() 1323 const int is_sync = rw_is_sync(bio->bi_rw); in blk_sq_make_request() 1324 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_sq_make_request()
|
D | bio-integrity.c | 82 bio->bi_rw |= REQ_INTEGRITY; in bio_integrity_alloc()
|
D | cfq-iosched.c | 900 return bio_data_dir(bio) == READ || (bio->bi_rw & REQ_SYNC); in cfq_bio_sync() 2366 cfqg_stats_update_io_merged(RQ_CFQG(req), bio->bi_rw); in cfq_bio_merged()
|
/linux-4.1.27/include/trace/events/ |
D | block.h | 275 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 313 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 340 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 407 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 436 bio ? bio->bi_rw : 0, __entry->nr_sector); 570 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 613 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
|
D | bcache.h | 30 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 104 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 139 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 170 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
|
/linux-4.1.27/drivers/md/ |
D | dm-raid1.c | 261 .bi_rw = WRITE_FLUSH, in mirror_flush() 528 .bi_rw = READ, in read_async_bio() 611 if (bio->bi_rw & REQ_DISCARD) { in write_callback() 640 .bi_rw = WRITE | (bio->bi_rw & WRITE_FLUSH_FUA), in do_write() 648 if (bio->bi_rw & REQ_DISCARD) { in do_write() 649 io_req.bi_rw |= REQ_DISCARD; in do_write() 687 if ((bio->bi_rw & REQ_FLUSH) || in do_writes() 688 (bio->bi_rw & REQ_DISCARD)) { in do_writes() 1218 if (!(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) in mirror_end_io() 1226 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in mirror_end_io()
|
D | dm-stats.c | 446 unsigned long bi_rw, sector_t len, bool merged, in dm_stat_for_entry() argument 449 unsigned long idx = bi_rw & REQ_WRITE; in dm_stat_for_entry() 495 static void __dm_stat_bio(struct dm_stat *s, unsigned long bi_rw, in __dm_stat_bio() argument 525 dm_stat_for_entry(s, entry, bi_rw, fragment_len, in __dm_stat_bio() 533 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw, in dm_stats_account_io() argument 554 ((bi_rw & (REQ_WRITE | REQ_DISCARD)) == in dm_stats_account_io() 558 ACCESS_ONCE(last->last_rw) = bi_rw; in dm_stats_account_io() 564 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration, stats_aux); in dm_stats_account_io()
|
D | dm-stripe.c | 287 if (bio->bi_rw & REQ_FLUSH) { in stripe_map() 293 if (unlikely(bio->bi_rw & REQ_DISCARD) || in stripe_map() 294 unlikely(bio->bi_rw & REQ_WRITE_SAME)) { in stripe_map() 361 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in stripe_end_io()
|
D | dm-stats.h | 31 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
|
D | multipath.c | 93 else if (!(bio->bi_rw & REQ_RAHEAD)) { in multipath_end_request() 114 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in multipath_make_request() 137 mp_bh->bio.bi_rw |= REQ_FAILFAST_TRANSPORT; in multipath_make_request() 368 bio->bi_rw |= REQ_FAILFAST_TRANSPORT; in multipathd()
|
D | dm-io.c | 482 if ((io_req->bi_rw & RW_MASK) == READ) { in dp_init() 520 io_req->bi_rw, &dp, sync_error_bits); in dm_io() 522 return async_io(io_req->client, num_regions, where, io_req->bi_rw, in dm_io()
|
D | raid1.c | 793 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes() 1068 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid1_unplug() 1089 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in make_request() 1090 const unsigned long do_flush_fua = (bio->bi_rw & (REQ_FLUSH | REQ_FUA)); in make_request() 1091 const unsigned long do_discard = (bio->bi_rw in make_request() 1093 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in make_request() 1200 read_bio->bi_rw = READ | do_sync; in make_request() 1411 mbio->bi_rw = in make_request() 2073 wbio->bi_rw = WRITE; in sync_request_write() 2244 wbio->bi_rw = WRITE; in narrow_write_error() [all …]
|
D | linear.c | 266 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in linear_make_request() 297 if (unlikely((split->bi_rw & REQ_DISCARD) && in linear_make_request()
|
D | dm-flakey.c | 19 (((bio)->bi_rw & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) 269 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_rw, in corrupt_bio_data()
|
D | dm-region-hash.c | 403 if (bio->bi_rw & REQ_FLUSH) { in dm_rh_mark_nosync() 408 if (bio->bi_rw & REQ_DISCARD) in dm_rh_mark_nosync() 531 if (bio->bi_rw & (REQ_FLUSH | REQ_DISCARD)) in dm_rh_inc_pending()
|
D | raid10.c | 957 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes() 1133 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid10_unplug() 1151 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in __make_request() 1152 const unsigned long do_fua = (bio->bi_rw & REQ_FUA); in __make_request() 1153 const unsigned long do_discard = (bio->bi_rw in __make_request() 1155 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in __make_request() 1248 read_bio->bi_rw = READ | do_sync; in __make_request() 1457 mbio->bi_rw = in __make_request() 1500 mbio->bi_rw = in __make_request() 1544 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in make_request() [all …]
|
D | dm.c | 656 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in start_io_acct() 671 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in end_io_acct() 934 if ((bio->bi_rw & REQ_FLUSH) && bio->bi_iter.bi_size) { in dec_pending() 939 bio->bi_rw &= ~REQ_FLUSH; in dec_pending() 985 if (unlikely(r == -EREMOTEIO && (bio->bi_rw & REQ_WRITE_SAME) && in clone_endio() 1422 BUG_ON(bio->bi_rw & REQ_FLUSH); in dm_accept_partial_bio() 1652 if (unlikely(bio->bi_rw & REQ_DISCARD)) in __split_and_process_non_flush() 1654 else if (unlikely(bio->bi_rw & REQ_WRITE_SAME)) in __split_and_process_non_flush() 1697 if (bio->bi_rw & REQ_FLUSH) { in __split_and_process_bio() 2334 md->flush_bio.bi_rw = WRITE_FLUSH; in alloc_dev()
|
D | dm-log-writes.c | 553 bool flush_bio = (bio->bi_rw & REQ_FLUSH); in log_writes_map() 554 bool fua_bio = (bio->bi_rw & REQ_FUA); in log_writes_map() 555 bool discard_bio = (bio->bi_rw & REQ_DISCARD); in log_writes_map()
|
D | dm-log.c | 296 lc->io_req.bi_rw = rw; in rw_header() 309 lc->io_req.bi_rw = WRITE_FLUSH; in flush_header()
|
D | raid0.c | 516 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in raid0_make_request() 546 if (unlikely((split->bi_rw & REQ_DISCARD) && in raid0_make_request()
|
D | dm-thin.c | 585 return (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && in bio_triggers_commit() 593 if (bio->bi_rw & REQ_DISCARD) in inc_all_io_entry() 754 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) in __inc_remap_and_issue_cell() 1429 (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA))) in __remap_and_issue_shared_cell() 1818 if (bio->bi_rw & REQ_DISCARD) in process_thin_deferred_bios() 1905 if (cell->holder->bi_rw & REQ_DISCARD) in process_thin_deferred_cells() 2309 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) { in thin_bio_map()
|
D | raid5.c | 809 if (head->dev[dd_idx].towrite->bi_rw != sh->dev[dd_idx].towrite->bi_rw) in stripe_add_to_batch_list() 996 bi->bi_rw = rw; in ops_run_io() 1004 bi->bi_rw, i); in ops_run_io() 1015 bi->bi_rw |= REQ_NOMERGE; in ops_run_io() 1048 rbi->bi_rw = rw; in ops_run_io() 1056 rbi->bi_rw, i); in ops_run_io() 1089 bi->bi_rw, i, (unsigned long long)sh->sector); in ops_run_io() 1622 if (wbi->bi_rw & REQ_FUA) in ops_run_biodrain() 1624 if (wbi->bi_rw & REQ_SYNC) in ops_run_biodrain() 1626 if (wbi->bi_rw & REQ_DISCARD) in ops_run_biodrain() [all …]
|
D | dm-cache-target.c | 741 !(bio->bi_rw & (REQ_FUA | REQ_FLUSH | REQ_DISCARD))) { in check_if_tick_bio_needed() 782 return bio->bi_rw & (REQ_FLUSH | REQ_FUA); in bio_triggers_commit() 1630 if (bio->bi_rw & REQ_FLUSH) in process_deferred_bios() 1632 else if (bio->bi_rw & REQ_DISCARD) in process_deferred_bios() 2627 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA | REQ_DISCARD)) { in __cache_map()
|
D | dm-snap.c | 1669 if (bio->bi_rw & REQ_FLUSH) { in snapshot_map() 1784 if (bio->bi_rw & REQ_FLUSH) { in snapshot_merge_map() 2268 if (unlikely(bio->bi_rw & REQ_FLUSH)) in origin_map()
|
D | dm-crypt.c | 1122 clone->bi_rw = io->base_bio->bi_rw; in clone_init() 1893 if (unlikely(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) { in crypt_map()
|
D | dm-bufio.c | 556 .bi_rw = rw, in use_dmio() 1299 .bi_rw = WRITE_FLUSH, in dm_bufio_issue_flush()
|
D | dm-snap-persistent.c | 237 .bi_rw = rw, in chunk_io()
|
D | dm-kcopyd.c | 499 .bi_rw = job->rw, in run_io_job()
|
D | dm-era-target.c | 1546 if (!(bio->bi_rw & REQ_FLUSH) && in era_map()
|
D | md.c | 293 bio->bi_rw &= ~REQ_NOMERGE; in md_make_request() 440 bio->bi_rw &= ~REQ_FLUSH; in md_submit_flush_data()
|
/linux-4.1.27/drivers/staging/lustre/lustre/llite/ |
D | lloop.c | 215 rw = head->bi_rw; in do_bio_lustrebacked() 217 LASSERT(rw == bio->bi_rw); in do_bio_lustrebacked() 308 rw = first->bi_rw; in loop_get_bio() 310 while (*bio && (*bio)->bi_rw == rw) { in loop_get_bio()
|
/linux-4.1.27/Documentation/device-mapper/ |
D | dm-flakey.txt | 45 <flags>: Perform the replacement only if bio->bi_rw has all the
|
/linux-4.1.27/drivers/block/ |
D | pmem.c | 75 BUG_ON(bio->bi_rw & REQ_DISCARD); in pmem_make_request()
|
D | umem.c | 465 } else if ((bio->bi_rw & REQ_WRITE) && in process_page() 538 if (bio->bi_rw & REQ_SYNC || !mm_check_plugged(card)) in mm_make_request()
|
D | brd.c | 340 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in brd_make_request()
|
D | pktcdvd.c | 1076 bio->bi_rw = READ; in pkt_gather_data() 1159 pkt->bio->bi_rw = REQ_WRITE; in pkt_start_recovery() 1338 pkt->w_bio->bi_rw = WRITE; in pkt_start_write()
|
D | xen-blkfront.c | 1583 submit_bio(cloned_bio->bi_rw, cloned_bio); in blkif_recover() 1592 submit_bio(bio->bi_rw, bio); in blkif_recover()
|
/linux-4.1.27/drivers/scsi/osd/ |
D | osd_initiator.c | 732 bio->bi_rw &= ~REQ_WRITE; in _osd_req_list_objects() 830 WARN_ON(0 == (bio->bi_rw & REQ_WRITE)); in osd_req_write() 845 bio->bi_rw |= REQ_WRITE; /* FIXME: bio_set_dir() */ in osd_req_write_kern() 881 WARN_ON(bio->bi_rw & REQ_WRITE); in osd_req_read() 962 bio->bi_rw |= REQ_WRITE; in _osd_req_finalize_cdb_cont() 1083 bio->bi_rw |= REQ_WRITE; in osd_req_write_sg_kern()
|
/linux-4.1.27/drivers/block/drbd/ |
D | drbd_main.c | 1605 static u32 bio_flags_to_wire(struct drbd_connection *connection, unsigned long bi_rw) in bio_flags_to_wire() argument 1608 return (bi_rw & REQ_SYNC ? DP_RW_SYNC : 0) | in bio_flags_to_wire() 1609 (bi_rw & REQ_FUA ? DP_FUA : 0) | in bio_flags_to_wire() 1610 (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) | in bio_flags_to_wire() 1611 (bi_rw & REQ_DISCARD ? DP_DISCARD : 0); in bio_flags_to_wire() 1613 return bi_rw & REQ_SYNC ? DP_RW_SYNC : 0; in bio_flags_to_wire() 1638 dp_flags = bio_flags_to_wire(peer_device->connection, req->master_bio->bi_rw); in drbd_send_dblock()
|
D | drbd_worker.c | 179 int is_discard = !!(bio->bi_rw & REQ_DISCARD); in drbd_peer_request_endio() 268 if (bio->bi_rw & REQ_DISCARD) in drbd_request_endio()
|
D | drbd_actlog.c | 162 bio->bi_rw = rw; in _drbd_md_sync_page_io()
|
D | drbd_bitmap.c | 1033 bio->bi_rw |= rw; in bm_page_io_async()
|
D | drbd_req.c | 1116 D_ASSERT(device, req->master_bio->bi_rw & REQ_FLUSH); in drbd_process_write_request()
|
D | drbd_receiver.c | 1420 bio->bi_rw = rw; in drbd_submit_peer_request()
|
/linux-4.1.27/kernel/trace/ |
D | blktrace.c | 787 bio->bi_rw, what, error, 0, NULL); in blk_add_trace_bio() 891 bio->bi_iter.bi_size, bio->bi_rw, BLK_TA_SPLIT, in blk_add_trace_split() 925 bio->bi_rw, BLK_TA_REMAP, in blk_add_trace_bio_remap()
|
/linux-4.1.27/fs/btrfs/ |
D | volumes.c | 367 btrfsic_submit_bio(cur->bi_rw, cur); in run_scheduled_bios() 5612 if (bio->bi_rw & WRITE) in btrfs_end_bio() 5618 if ((bio->bi_rw & WRITE_FLUSH) == WRITE_FLUSH) in btrfs_end_bio() 5696 bio->bi_rw |= rw; in btrfs_schedule_bio() 5699 if (bio->bi_rw & REQ_SYNC) in btrfs_schedule_bio() 5729 .bi_rw = bio->bi_rw, in bio_size_ok()
|
D | disk-io.c | 716 if (bio->bi_rw & REQ_WRITE) { in end_workqueue_bio()
|
D | inode.c | 7691 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in dio_read_error() 7977 btrfs_ino(dip->inode), bio->bi_rw, in btrfs_end_dio_bio()
|
D | extent_io.c | 2408 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in bio_readpage_error()
|
/linux-4.1.27/drivers/target/ |
D | target_core_pscsi.c | 920 bio->bi_rw |= REQ_WRITE; in pscsi_map_sg()
|
/linux-4.1.27/drivers/block/rsxx/ |
D | dma.c | 708 if (bio->bi_rw & REQ_DISCARD) { in rsxx_dma_queue_bio()
|
/linux-4.1.27/fs/exofs/ |
D | ore.c | 881 bio->bi_rw |= REQ_WRITE; in _write_mirror()
|
/linux-4.1.27/drivers/block/zram/ |
D | zram_drv.c | 933 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in __zram_make_request()
|
/linux-4.1.27/Documentation/block/ |
D | biodoc.txt | 272 requests. Some bits in the bi_rw flags field in the bio structure are 435 unsigned long bi_rw; /* low bits: r/w, high: priority */
|