/linux-4.4.14/block/ |
D | blk-merge.c | 173 if ((*bio)->bi_rw & REQ_DISCARD) in blk_queue_split() 175 else if ((*bio)->bi_rw & REQ_WRITE_SAME) in blk_queue_split() 187 split->bi_rw |= REQ_NOMERGE; in blk_queue_split() 213 if (bio->bi_rw & REQ_DISCARD) in __blk_recalc_rq_segments() 216 if (bio->bi_rw & REQ_WRITE_SAME) in __blk_recalc_rq_segments() 389 if (bio->bi_rw & REQ_DISCARD) { in __blk_bios_map_sg() 404 if (bio->bi_rw & REQ_WRITE_SAME) { in __blk_bios_map_sg() 620 WARN_ON_ONCE((bio->bi_rw & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge() 621 (bio->bi_rw & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge() 622 bio->bi_rw |= ff; in blk_rq_set_mixed_merge() [all …]
|
D | blk-core.c | 1024 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) in blk_rq_should_init_elevator() 1546 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_back_merge() 1568 const int ff = bio->bi_rw & REQ_FAILFAST_MASK; in bio_attempt_front_merge() 1690 req->cmd_flags |= bio->bi_rw & REQ_COMMON_MASK; in init_request_from_bio() 1691 if (bio->bi_rw & REQ_RAHEAD) in init_request_from_bio() 1702 const bool sync = !!(bio->bi_rw & REQ_SYNC); in blk_queue_bio() 1723 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) { in blk_queue_bio() 1843 bio->bi_rw, in handle_bad_sector() 1956 if ((bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && !q->flush_flags) { in generic_make_request_checks() 1957 bio->bi_rw &= ~(REQ_FLUSH | REQ_FUA); in generic_make_request_checks() [all …]
|
D | bio.c | 584 bio->bi_rw = bio_src->bi_rw; in __bio_clone_fast() 666 bio->bi_rw = bio_src->bi_rw; in bio_clone_bioset() 670 if (bio->bi_rw & REQ_DISCARD) in bio_clone_bioset() 673 if (bio->bi_rw & REQ_WRITE_SAME) { in bio_clone_bioset() 1175 bio->bi_rw |= REQ_WRITE; in bio_copy_user_iov() 1345 bio->bi_rw |= REQ_WRITE; in bio_map_user_iov() 1538 bio->bi_rw |= REQ_WRITE; in bio_copy_kern() 1797 if (bio->bi_rw & REQ_DISCARD) in bio_split()
|
D | blk-map.c | 227 bio->bi_rw |= REQ_WRITE; in blk_rq_map_kern()
|
D | blk-throttle.c | 830 if (!(bio->bi_rw & REQ_THROTTLED)) in throtl_charge_bio() 831 bio->bi_rw |= REQ_THROTTLED; in throtl_charge_bio() 1408 if ((bio->bi_rw & REQ_THROTTLED) || !tg->has_rules[rw]) in blk_throtl_bio() 1487 bio->bi_rw &= ~REQ_THROTTLED; in blk_throtl_bio()
|
D | blk-mq.c | 1174 if (rw_is_sync(bio->bi_rw)) in blk_mq_map_request() 1244 const int is_sync = rw_is_sync(bio->bi_rw); in blk_mq_make_request() 1245 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_mq_make_request() 1341 const int is_sync = rw_is_sync(bio->bi_rw); in blk_sq_make_request() 1342 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_sq_make_request()
|
D | bio-integrity.c | 87 bio->bi_rw |= REQ_INTEGRITY; in bio_integrity_alloc()
|
D | cfq-iosched.c | 901 return bio_data_dir(bio) == READ || (bio->bi_rw & REQ_SYNC); in cfq_bio_sync() 2543 cfqg_stats_update_io_merged(RQ_CFQG(req), bio->bi_rw); in cfq_bio_merged()
|
/linux-4.4.14/include/linux/ |
D | bio.h | 51 #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT) 56 (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \ 57 (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \ 109 !(bio->bi_rw & REQ_DISCARD)) in bio_has_data() 120 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_is_rw() 128 if (bio->bi_rw & REQ_NOMERGE_FLAGS) in bio_mergeable() 228 if (bio->bi_rw & BIO_NO_ADVANCE_ITER_MASK) in bio_advance_iter() 256 if (bio->bi_rw & REQ_DISCARD) in bio_segments() 259 if (bio->bi_rw & REQ_WRITE_SAME) in bio_segments() 357 if (bio->bi_rw & REQ_INTEGRITY) in bio_integrity()
|
D | dm-io.h | 60 int bi_rw; /* READ|WRITE - not READA */ member
|
D | blk_types.h | 51 unsigned long bi_rw; /* bottom bits READ/WRITE, member
|
D | blk-cgroup.h | 716 blkg_rwstat_add(&blkg->stat_bytes, bio->bi_rw, in blkcg_bio_issue_check() 718 blkg_rwstat_add(&blkg->stat_ios, bio->bi_rw, 1); in blkcg_bio_issue_check()
|
D | fs.h | 2390 #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK)) 2395 #define bio_data_dir(bio) ((bio)->bi_rw & 1)
|
/linux-4.4.14/drivers/md/bcache/ |
D | request.c | 211 bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); in bch_data_insert_start() 256 n->bi_rw |= REQ_WRITE; in bch_data_insert_start() 381 (bio->bi_rw & REQ_DISCARD)) in check_should_bypass() 386 (bio->bi_rw & REQ_WRITE))) in check_should_bypass() 407 (bio->bi_rw & REQ_WRITE) && in check_should_bypass() 408 (bio->bi_rw & REQ_SYNC)) in check_should_bypass() 660 s->write = (bio->bi_rw & REQ_WRITE) != 0; in search_alloc() 671 s->iop.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; in search_alloc() 799 if (!(bio->bi_rw & REQ_RAHEAD) && in cached_dev_cache_miss() 800 !(bio->bi_rw & REQ_META) && in cached_dev_cache_miss() [all …]
|
D | writeback.h | 60 return bio->bi_rw & REQ_SYNC || in should_writeback()
|
D | io.c | 114 unsigned threshold = bio->bi_rw & REQ_WRITE in bch_bbio_count_io_errors()
|
D | writeback.c | 186 io->bio.bi_rw = WRITE; in write_dirty() 259 io->bio.bi_rw = READ; in read_dirty()
|
D | journal.c | 57 bio->bi_rw = READ; in journal_read_bucket() 455 bio->bi_rw = REQ_WRITE|REQ_DISCARD; in do_journal_discard() 629 bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; in journal_write_unlocked()
|
D | movinggc.c | 166 bio->bi_rw = READ; in read_moving()
|
D | super.c | 215 bio->bi_rw = REQ_SYNC|REQ_META; in __write_super() 351 bio->bi_rw = REQ_SYNC|REQ_META|rw; in uuid_io() 510 bio->bi_rw = REQ_SYNC|REQ_META|rw; in prio_io()
|
D | btree.c | 298 bio->bi_rw = REQ_META|READ_SYNC; in bch_btree_node_read() 400 b->bio->bi_rw = REQ_META|WRITE_SYNC|REQ_FUA; in do_btree_node_write()
|
/linux-4.4.14/include/trace/events/ |
D | block.h | 275 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 313 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 340 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 407 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 436 bio ? bio->bi_rw : 0, __entry->nr_sector); 570 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 613 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
|
D | bcache.h | 30 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 104 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 139 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size); 170 blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_iter.bi_size);
|
/linux-4.4.14/drivers/md/ |
D | dm-raid1.c | 263 .bi_rw = WRITE_FLUSH, in mirror_flush() 544 .bi_rw = READ, in read_async_bio() 627 if (bio->bi_rw & REQ_DISCARD) { in write_callback() 657 .bi_rw = WRITE | (bio->bi_rw & WRITE_FLUSH_FUA), in do_write() 665 if (bio->bi_rw & REQ_DISCARD) { in do_write() 666 io_req.bi_rw |= REQ_DISCARD; in do_write() 704 if ((bio->bi_rw & REQ_FLUSH) || in do_writes() 705 (bio->bi_rw & REQ_DISCARD)) { in do_writes() 1253 if (!(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) in mirror_end_io() 1261 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in mirror_end_io()
|
D | dm-stripe.c | 289 if (bio->bi_rw & REQ_FLUSH) { in stripe_map() 295 if (unlikely(bio->bi_rw & REQ_DISCARD) || in stripe_map() 296 unlikely(bio->bi_rw & REQ_WRITE_SAME)) { in stripe_map() 363 if ((error == -EWOULDBLOCK) && (bio->bi_rw & REQ_RAHEAD)) in stripe_end_io()
|
D | dm-stats.h | 32 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
|
D | multipath.c | 93 else if (!(bio->bi_rw & REQ_RAHEAD)) { in multipath_end_request() 114 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in multipath_make_request() 137 mp_bh->bio.bi_rw |= REQ_FAILFAST_TRANSPORT; in multipath_make_request() 358 bio->bi_rw |= REQ_FAILFAST_TRANSPORT; in multipathd()
|
D | dm-stats.c | 517 unsigned long bi_rw, sector_t len, in dm_stat_for_entry() argument 521 unsigned long idx = bi_rw & REQ_WRITE; in dm_stat_for_entry() 587 static void __dm_stat_bio(struct dm_stat *s, unsigned long bi_rw, in __dm_stat_bio() argument 617 dm_stat_for_entry(s, entry, bi_rw, fragment_len, in __dm_stat_bio() 625 void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw, in dm_stats_account_io() argument 648 ((bi_rw & (REQ_WRITE | REQ_DISCARD)) == in dm_stats_account_io() 652 ACCESS_ONCE(last->last_rw) = bi_rw; in dm_stats_account_io() 666 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration_jiffies, stats_aux); in dm_stats_account_io()
|
D | linear.c | 224 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in linear_make_request() 255 if (unlikely((split->bi_rw & REQ_DISCARD) && in linear_make_request()
|
D | dm-io.c | 483 if ((io_req->bi_rw & RW_MASK) == READ) { in dp_init() 521 io_req->bi_rw, &dp, sync_error_bits); in dm_io() 523 return async_io(io_req->client, num_regions, where, io_req->bi_rw, in dm_io()
|
D | raid1.c | 762 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes() 1036 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid1_unplug() 1057 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in make_request() 1058 const unsigned long do_flush_fua = (bio->bi_rw & (REQ_FLUSH | REQ_FUA)); in make_request() 1059 const unsigned long do_discard = (bio->bi_rw in make_request() 1061 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in make_request() 1169 read_bio->bi_rw = READ | do_sync; in make_request() 1379 mbio->bi_rw = in make_request() 2033 wbio->bi_rw = WRITE; in sync_request_write() 2204 wbio->bi_rw = WRITE; in narrow_write_error() [all …]
|
D | dm-flakey.c | 19 (((bio)->bi_rw & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) 271 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_rw, in corrupt_bio_data()
|
D | dm-region-hash.c | 401 if (bio->bi_rw & REQ_FLUSH) { in dm_rh_mark_nosync() 406 if (bio->bi_rw & REQ_DISCARD) in dm_rh_mark_nosync() 529 if (bio->bi_rw & (REQ_FLUSH | REQ_DISCARD)) in dm_rh_inc_pending()
|
D | raid10.c | 868 if (unlikely((bio->bi_rw & REQ_DISCARD) && in flush_pending_writes() 1044 if (unlikely((bio->bi_rw & REQ_DISCARD) && in raid10_unplug() 1062 const unsigned long do_sync = (bio->bi_rw & REQ_SYNC); in __make_request() 1063 const unsigned long do_fua = (bio->bi_rw & REQ_FUA); in __make_request() 1064 const unsigned long do_discard = (bio->bi_rw in __make_request() 1066 const unsigned long do_same = (bio->bi_rw & REQ_WRITE_SAME); in __make_request() 1159 read_bio->bi_rw = READ | do_sync; in __make_request() 1366 mbio->bi_rw = in __make_request() 1409 mbio->bi_rw = in __make_request() 1453 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in make_request() [all …]
|
D | dm-log-writes.c | 555 bool flush_bio = (bio->bi_rw & REQ_FLUSH); in log_writes_map() 556 bool fua_bio = (bio->bi_rw & REQ_FUA); in log_writes_map() 557 bool discard_bio = (bio->bi_rw & REQ_DISCARD); in log_writes_map()
|
D | dm.c | 689 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in start_io_acct() 704 dm_stats_account_io(&md->stats, bio->bi_rw, bio->bi_iter.bi_sector, in end_io_acct() 967 if ((bio->bi_rw & REQ_FLUSH) && bio->bi_iter.bi_size) { in dec_pending() 972 bio->bi_rw &= ~REQ_FLUSH; in dec_pending() 1017 if (unlikely(r == -EREMOTEIO && (bio->bi_rw & REQ_WRITE_SAME) && in clone_endio() 1462 BUG_ON(bio->bi_rw & REQ_FLUSH); in dm_accept_partial_bio() 1692 if (unlikely(bio->bi_rw & REQ_DISCARD)) in __split_and_process_non_flush() 1694 else if (unlikely(bio->bi_rw & REQ_WRITE_SAME)) in __split_and_process_non_flush() 1737 if (bio->bi_rw & REQ_FLUSH) { in __split_and_process_bio() 2362 md->flush_bio.bi_rw = WRITE_FLUSH; in alloc_dev()
|
D | raid0.c | 461 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in raid0_make_request() 491 if (unlikely((split->bi_rw & REQ_DISCARD) && in raid0_make_request()
|
D | dm-log.c | 296 lc->io_req.bi_rw = rw; in rw_header() 309 lc->io_req.bi_rw = WRITE_FLUSH; in flush_header()
|
D | dm-thin.c | 698 return (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && in bio_triggers_commit() 706 if (bio->bi_rw & REQ_DISCARD) in inc_all_io_entry() 869 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) in __inc_remap_and_issue_cell() 1647 (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA))) in __remap_and_issue_shared_cell() 2036 if (bio->bi_rw & REQ_DISCARD) in process_thin_deferred_bios() 2123 if (cell->holder->bi_rw & REQ_DISCARD) in process_thin_deferred_cells() 2559 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) { in thin_bio_map()
|
D | raid5.c | 811 if (head->dev[dd_idx].towrite->bi_rw != sh->dev[dd_idx].towrite->bi_rw) in stripe_add_to_batch_list() 1000 bi->bi_rw = rw; in ops_run_io() 1008 bi->bi_rw, i); in ops_run_io() 1019 bi->bi_rw |= REQ_NOMERGE; in ops_run_io() 1052 rbi->bi_rw = rw; in ops_run_io() 1060 rbi->bi_rw, i); in ops_run_io() 1093 bi->bi_rw, i, (unsigned long long)sh->sector); in ops_run_io() 1624 if (wbi->bi_rw & REQ_FUA) in ops_run_biodrain() 1626 if (wbi->bi_rw & REQ_SYNC) in ops_run_biodrain() 1628 if (wbi->bi_rw & REQ_DISCARD) in ops_run_biodrain() [all …]
|
D | dm-cache-target.c | 794 !(bio->bi_rw & (REQ_FUA | REQ_FLUSH | REQ_DISCARD))) { in check_if_tick_bio_needed() 835 return bio->bi_rw & (REQ_FLUSH | REQ_FUA); in bio_triggers_commit() 857 !(bio->bi_rw & REQ_DISCARD)); in accountable_bio() 1073 return bio->bi_rw & (REQ_FLUSH | REQ_FUA | REQ_DISCARD); in discard_or_flush() 1984 if (bio->bi_rw & REQ_FLUSH) in process_deferred_bios() 1986 else if (bio->bi_rw & REQ_DISCARD) in process_deferred_bios()
|
D | dm-crypt.c | 1136 clone->bi_rw = io->base_bio->bi_rw; in clone_init() 1915 if (unlikely(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) { in crypt_map()
|
D | raid5-cache.c | 260 bio->bi_rw = WRITE; in r5l_bio_alloc() 515 bio->bi_rw &= ~REQ_FLUSH; in r5l_handle_flush_request()
|
D | dm-snap.c | 1687 if (bio->bi_rw & REQ_FLUSH) { in snapshot_map() 1806 if (bio->bi_rw & REQ_FLUSH) { in snapshot_merge_map() 2292 if (unlikely(bio->bi_rw & REQ_FLUSH)) in origin_map()
|
D | dm-bufio.c | 557 .bi_rw = rw, in use_dmio() 1305 .bi_rw = WRITE_FLUSH, in dm_bufio_issue_flush()
|
D | dm-snap-persistent.c | 238 .bi_rw = rw, in chunk_io()
|
D | dm-kcopyd.c | 499 .bi_rw = job->rw, in run_io_job()
|
D | dm-era-target.c | 1545 if (!(bio->bi_rw & REQ_FLUSH) && in era_map()
|
D | md.c | 297 bio->bi_rw &= ~REQ_NOMERGE; in md_make_request() 424 bio->bi_rw &= ~REQ_FLUSH; in md_submit_flush_data()
|
/linux-4.4.14/drivers/staging/lustre/lustre/llite/ |
D | lloop.c | 215 rw = head->bi_rw; in do_bio_lustrebacked() 217 LASSERT(rw == bio->bi_rw); in do_bio_lustrebacked() 308 rw = first->bi_rw; in loop_get_bio() 310 while (*bio && (*bio)->bi_rw == rw) { in loop_get_bio()
|
/linux-4.4.14/Documentation/device-mapper/ |
D | dm-flakey.txt | 45 <flags>: Perform the replacement only if bio->bi_rw has all the
|
/linux-4.4.14/drivers/scsi/osd/ |
D | osd_initiator.c | 732 bio->bi_rw &= ~REQ_WRITE; in _osd_req_list_objects() 830 WARN_ON(0 == (bio->bi_rw & REQ_WRITE)); in osd_req_write() 845 bio->bi_rw |= REQ_WRITE; /* FIXME: bio_set_dir() */ in osd_req_write_kern() 881 WARN_ON(bio->bi_rw & REQ_WRITE); in osd_req_read() 962 bio->bi_rw |= REQ_WRITE; in _osd_req_finalize_cdb_cont() 1083 bio->bi_rw |= REQ_WRITE; in osd_req_write_sg_kern()
|
/linux-4.4.14/drivers/lightnvm/ |
D | rrpc.c | 318 bio->bi_rw = READ; in rrpc_move_valid_pages() 336 bio->bi_rw = WRITE; in rrpc_move_valid_pages() 856 if (bio->bi_rw & REQ_DISCARD) { in rrpc_make_rq()
|
/linux-4.4.14/drivers/block/drbd/ |
D | drbd_main.c | 1605 static u32 bio_flags_to_wire(struct drbd_connection *connection, unsigned long bi_rw) in bio_flags_to_wire() argument 1608 return (bi_rw & REQ_SYNC ? DP_RW_SYNC : 0) | in bio_flags_to_wire() 1609 (bi_rw & REQ_FUA ? DP_FUA : 0) | in bio_flags_to_wire() 1610 (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) | in bio_flags_to_wire() 1611 (bi_rw & REQ_DISCARD ? DP_DISCARD : 0); in bio_flags_to_wire() 1613 return bi_rw & REQ_SYNC ? DP_RW_SYNC : 0; in bio_flags_to_wire() 1638 dp_flags = bio_flags_to_wire(peer_device->connection, req->master_bio->bi_rw); in drbd_send_dblock()
|
D | drbd_worker.c | 178 int is_discard = !!(bio->bi_rw & REQ_DISCARD); in drbd_peer_request_endio() 246 if (bio->bi_rw & REQ_DISCARD) in drbd_request_endio()
|
D | drbd_actlog.c | 162 bio->bi_rw = rw; in _drbd_md_sync_page_io()
|
D | drbd_bitmap.c | 1024 bio->bi_rw |= rw; in bm_page_io_async()
|
D | drbd_req.c | 1117 D_ASSERT(device, req->master_bio->bi_rw & REQ_FLUSH); in drbd_process_write_request()
|
D | drbd_receiver.c | 1421 bio->bi_rw = rw; in drbd_submit_peer_request()
|
/linux-4.4.14/kernel/trace/ |
D | blktrace.c | 779 bio->bi_rw, what, error, 0, NULL); in blk_add_trace_bio() 883 bio->bi_iter.bi_size, bio->bi_rw, BLK_TA_SPLIT, in blk_add_trace_split() 916 bio->bi_rw, BLK_TA_REMAP, bio->bi_error, in blk_add_trace_bio_remap()
|
/linux-4.4.14/drivers/block/ |
D | umem.c | 465 } else if ((bio->bi_rw & REQ_WRITE) && in process_page() 540 if (bio->bi_rw & REQ_SYNC || !mm_check_plugged(card)) in mm_make_request()
|
D | brd.c | 339 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in brd_make_request()
|
D | pktcdvd.c | 1077 bio->bi_rw = READ; in pkt_gather_data() 1160 pkt->bio->bi_rw = REQ_WRITE; in pkt_start_recovery() 1339 pkt->w_bio->bi_rw = WRITE; in pkt_start_write()
|
D | xen-blkfront.c | 1752 submit_bio(cloned_bio->bi_rw, cloned_bio); in blkif_recover() 1761 submit_bio(bio->bi_rw, bio); in blkif_recover()
|
/linux-4.4.14/drivers/target/ |
D | target_core_pscsi.c | 925 bio->bi_rw |= REQ_WRITE; in pscsi_map_sg()
|
/linux-4.4.14/drivers/block/rsxx/ |
D | dma.c | 708 if (bio->bi_rw & REQ_DISCARD) { in rsxx_dma_queue_bio()
|
/linux-4.4.14/fs/exofs/ |
D | ore.c | 881 bio->bi_rw |= REQ_WRITE; in _write_mirror()
|
/linux-4.4.14/fs/btrfs/ |
D | volumes.c | 449 btrfsic_submit_bio(cur->bi_rw, cur); in run_scheduled_bios() 5878 if (bio->bi_rw & WRITE) in btrfs_end_bio() 5884 if ((bio->bi_rw & WRITE_FLUSH) == WRITE_FLUSH) in btrfs_end_bio() 5959 bio->bi_rw |= rw; in btrfs_schedule_bio() 5962 if (bio->bi_rw & REQ_SYNC) in btrfs_schedule_bio()
|
D | disk-io.c | 715 if (bio->bi_rw & REQ_WRITE) { in end_workqueue_bio()
|
D | inode.c | 7786 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in dio_read_error() 8068 btrfs_ino(dip->inode), bio->bi_rw, in btrfs_end_dio_bio()
|
D | extent_io.c | 2476 BUG_ON(failed_bio->bi_rw & REQ_WRITE); in bio_readpage_error()
|
/linux-4.4.14/drivers/block/zram/ |
D | zram_drv.c | 852 if (unlikely(bio->bi_rw & REQ_DISCARD)) { in __zram_make_request()
|
/linux-4.4.14/Documentation/block/ |
D | biodoc.txt | 272 requests. Some bits in the bi_rw flags field in the bio structure are 435 unsigned long bi_rw; /* low bits: r/w, high: priority */
|