Home
last modified time | relevance | path

Searched refs:REQ_FLUSH (Results 1 – 54 of 54) sorted by relevance

/linux-4.1.27/Documentation/block/
Dwriteback_cache_control.txt23 The REQ_FLUSH flag can be OR ed into the r/w flags of a bio submitted from
27 storage before the flagged bio starts. In addition the REQ_FLUSH flag can be
44 Filesystems can simply set the REQ_FLUSH and REQ_FUA bits and do not have to
46 the Forced Unit Access is implemented. The REQ_FLUSH and REQ_FUA flags
53 These drivers will always see the REQ_FLUSH and REQ_FUA bits as they sit
56 to be implemented for bios with the REQ_FLUSH bit set. For real device
57 drivers that do not have a volatile cache the REQ_FLUSH and REQ_FUA bits
58 on non-empty bios can simply be ignored, and REQ_FLUSH requests without
68 support required, the block layer completes empty REQ_FLUSH requests before
69 entering the driver and strips off the REQ_FLUSH and REQ_FUA bits from
[all …]
/linux-4.1.27/include/linux/
Dblk_types.h215 REQ_DISCARD | REQ_WRITE_SAME | REQ_NOIDLE | REQ_FLUSH | REQ_FUA | \
223 (REQ_NOMERGE | REQ_STARTED | REQ_SOFTBARRIER | REQ_FLUSH | REQ_FUA | REQ_FLUSH_SEQ)
241 #define REQ_FLUSH (1ULL << __REQ_FLUSH) macro
Dfs.h203 #define WRITE_FLUSH (WRITE | REQ_SYNC | REQ_NOIDLE | REQ_FLUSH)
205 #define WRITE_FLUSH_FUA (WRITE | REQ_SYNC | REQ_NOIDLE | REQ_FLUSH | REQ_FUA)
/linux-4.1.27/drivers/block/
Dxen-blkfront.c461 BUG_ON(req->cmd_flags & (REQ_FLUSH | REQ_FUA)); in blkif_queue_request()
473 if (req->cmd_flags & (REQ_FLUSH | REQ_FUA)) { in blkif_queue_request()
482 ((REQ_FLUSH|REQ_FUA))) { in blkif_queue_request()
483 case REQ_FLUSH|REQ_FUA: in blkif_queue_request()
487 case REQ_FLUSH: in blkif_queue_request()
600 ((req->cmd_flags & REQ_FLUSH) && in blkif_request_flush_invalid()
601 !(info->feature_flush & REQ_FLUSH)) || in blkif_request_flush_invalid()
701 switch (feature_flush & ((REQ_FLUSH | REQ_FUA))) { in flush_info()
702 case REQ_FLUSH|REQ_FUA: in flush_info()
704 case REQ_FLUSH: in flush_info()
[all …]
Dps3disk.c199 if (req->cmd_flags & REQ_FLUSH) { in ps3disk_do_request()
259 if (req->cmd_flags & REQ_FLUSH) { in ps3disk_interrupt()
471 blk_queue_flush(queue, REQ_FLUSH); in ps3disk_probe()
Dosdblk.c324 do_flush = rq->cmd_flags & REQ_FLUSH; in osdblk_rq_fn()
440 blk_queue_flush(q, REQ_FLUSH); in osdblk_init_disk()
Dnbd.c469 if (req->cmd_flags & REQ_FLUSH) { in nbd_handle_req()
684 blk_queue_flush(nbd->disk->queue, REQ_FLUSH); in __nbd_ioctl()
Dvirtio_blk.c175 if (req->cmd_flags & REQ_FLUSH) { in virtio_queue_rq()
493 blk_queue_flush(vblk->disk->queue, REQ_FLUSH); in virtblk_update_cache_mode()
Dloop.c400 if (rq->cmd_flags & REQ_FLUSH) in do_req_filebacked()
748 blk_queue_flush(lo->lo_queue, REQ_FLUSH); in loop_set_fd()
Dnvme-core.c892 else if (req->cmd_flags & REQ_FLUSH) in nvme_queue_rq()
2141 blk_queue_flush(ns->queue, REQ_FLUSH | REQ_FUA); in nvme_alloc_ns()
Dskd_main.c629 if (io_flags & REQ_FLUSH) in skd_request_fn()
4415 blk_queue_flush(q, REQ_FLUSH | REQ_FUA); in skd_cons_disk()
/linux-4.1.27/block/
Dblk-flush.c104 if (fflags & REQ_FLUSH) { in blk_flush_policy()
105 if (rq->cmd_flags & REQ_FLUSH) in blk_flush_policy()
382 rq->cmd_flags &= ~REQ_FLUSH; in blk_insert_flush()
Dblk-settings.c840 WARN_ON_ONCE(flush & ~(REQ_FLUSH | REQ_FUA)); in blk_queue_flush()
842 if (WARN_ON_ONCE(!(flush & REQ_FLUSH) && (flush & REQ_FUA))) in blk_queue_flush()
845 q->flush_flags = flush & (REQ_FLUSH | REQ_FUA); in blk_queue_flush()
Dblk-core.c938 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) in blk_rq_should_init_elevator()
1603 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) { in blk_queue_bio()
1842 if ((bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && !q->flush_flags) { in generic_make_request_checks()
1843 bio->bi_rw &= ~(REQ_FLUSH | REQ_FUA); in generic_make_request_checks()
2086 if (rq->cmd_flags & (REQ_FLUSH|REQ_FUA)) in blk_insert_cloned_request()
3179 if (rq->cmd_flags & (REQ_FLUSH | REQ_FUA)) in blk_flush_plug_list()
Dblk-mq.c1248 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_mq_make_request()
1324 const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA); in blk_sq_make_request()
/linux-4.1.27/drivers/mmc/card/
Dqueue.h4 #define MMC_REQ_SPECIAL_MASK (REQ_DISCARD | REQ_FLUSH)
Dblock.c1604 next->cmd_flags & REQ_FLUSH) in mmc_blk_prep_packed_list()
2051 } else if (cmd_flags & REQ_FLUSH) { in mmc_blk_issue_rq()
2188 blk_queue_flush(md->queue.queue, REQ_FLUSH | REQ_FUA); in mmc_blk_alloc_req()
/linux-4.1.27/Documentation/device-mapper/
Dlog-writes.txt17 next REQ_FLUSH request. This is to make it easier for userspace to replay the
22 Once we see a REQ_FLUSH request we splice this list onto the request and once
24 completed WRITEs, at the time the REQ_FLUSH is issued, are added in order to
/linux-4.1.27/drivers/mtd/
Dmtd_blkdevs.c90 if (req->cmd_flags & REQ_FLUSH) in do_blktrans_request()
416 blk_queue_flush(new->rq, REQ_FLUSH); in add_mtd_blktrans_dev()
/linux-4.1.27/drivers/md/
Ddm-region-hash.c403 if (bio->bi_rw & REQ_FLUSH) { in dm_rh_mark_nosync()
531 if (bio->bi_rw & (REQ_FLUSH | REQ_DISCARD)) in dm_rh_inc_pending()
Dlinear.c266 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in linear_make_request()
Ddm.c934 if ((bio->bi_rw & REQ_FLUSH) && bio->bi_iter.bi_size) { in dec_pending()
939 bio->bi_rw &= ~REQ_FLUSH; in dec_pending()
1422 BUG_ON(bio->bi_rw & REQ_FLUSH); in dm_accept_partial_bio()
1697 if (bio->bi_rw & REQ_FLUSH) { in __split_and_process_bio()
2107 if (!(rq->cmd_flags & REQ_FLUSH)) in dm_request_fn()
2718 if (!(rq->cmd_flags & REQ_FLUSH)) in dm_mq_queue_rq()
Ddm-stripe.c287 if (bio->bi_rw & REQ_FLUSH) { in stripe_map()
Dmultipath.c114 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in multipath_make_request()
Ddm-io.c379 if (where[i].count || (rw & REQ_FLUSH)) in dispatch_io()
Ddm-snap.c1669 if (bio->bi_rw & REQ_FLUSH) { in snapshot_map()
1784 if (bio->bi_rw & REQ_FLUSH) { in snapshot_merge_map()
2268 if (unlikely(bio->bi_rw & REQ_FLUSH)) in origin_map()
Ddm-table.c1495 if (dm_table_supports_flush(t, REQ_FLUSH)) { in dm_table_set_restrictions()
1496 flush |= REQ_FLUSH; in dm_table_set_restrictions()
Ddm-raid1.c687 if ((bio->bi_rw & REQ_FLUSH) || in do_writes()
1218 if (!(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) in mirror_end_io()
Ddm-cache-target.c741 !(bio->bi_rw & (REQ_FUA | REQ_FLUSH | REQ_DISCARD))) { in check_if_tick_bio_needed()
782 return bio->bi_rw & (REQ_FLUSH | REQ_FUA); in bio_triggers_commit()
1630 if (bio->bi_rw & REQ_FLUSH) in process_deferred_bios()
2627 if (bio->bi_rw & (REQ_FLUSH | REQ_FUA | REQ_DISCARD)) { in __cache_map()
Draid0.c516 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in raid0_make_request()
Ddm-log-writes.c553 bool flush_bio = (bio->bi_rw & REQ_FLUSH); in log_writes_map()
Ddm-thin.c585 return (bio->bi_rw & (REQ_FLUSH | REQ_FUA)) && in bio_triggers_commit()
754 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) in __inc_remap_and_issue_cell()
1429 (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA))) in __remap_and_issue_shared_cell()
2309 if (bio->bi_rw & (REQ_DISCARD | REQ_FLUSH | REQ_FUA)) { in thin_bio_map()
Ddm-era-target.c1546 if (!(bio->bi_rw & REQ_FLUSH) && in era_map()
Ddm-crypt.c1893 if (unlikely(bio->bi_rw & (REQ_FLUSH | REQ_DISCARD))) { in crypt_map()
Draid1.c1090 const unsigned long do_flush_fua = (bio->bi_rw & (REQ_FLUSH | REQ_FUA)); in make_request()
Dmd.c440 bio->bi_rw &= ~REQ_FLUSH; in md_submit_flush_data()
4927 blk_queue_flush(mddev->queue, REQ_FLUSH | REQ_FUA); in md_alloc()
Draid10.c1544 if (unlikely(bio->bi_rw & REQ_FLUSH)) { in make_request()
Draid5.c5160 if (unlikely(bi->bi_rw & REQ_FLUSH)) { in make_request()
/linux-4.1.27/drivers/ide/
Dide-disk.c434 if (!(rq->cmd_flags & REQ_FLUSH)) in idedisk_prep_fn()
549 flush = REQ_FLUSH; in update_flush()
/linux-4.1.27/drivers/target/
Dtarget_core_iblock.c699 else if (!(q->flush_flags & REQ_FLUSH)) in iblock_execute_rw()
864 return q->flush_flags & REQ_FLUSH; in iblock_get_write_cache()
/linux-4.1.27/drivers/md/bcache/
Drequest.c208 bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); in bch_data_insert_start()
662 s->iop.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; in search_alloc()
914 if (bio->bi_rw & REQ_FLUSH) { in cached_dev_write()
Djournal.c627 bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; in journal_write_unlocked()
Dsuper.c850 blk_queue_flush(q, REQ_FLUSH|REQ_FUA); in bcache_device_init()
/linux-4.1.27/arch/um/drivers/
Dubd_kern.c869 blk_queue_flush(ubd_dev->queue, REQ_FLUSH); in ubd_add()
1293 if (req->cmd_flags & REQ_FLUSH) { in do_ubd_request()
/linux-4.1.27/fs/btrfs/
Dcheck-integrity.c2239 if (block->submit_bio_bh_rw & REQ_FLUSH) { in btrfsic_bio_end_io()
2275 if (block->submit_bio_bh_rw & REQ_FLUSH) { in btrfsic_bh_end_io()
2928 } else if (NULL != dev_state && (rw & REQ_FLUSH)) { in btrfsic_submit_bh()
3026 } else if (NULL != dev_state && (rw & REQ_FLUSH)) { in __btrfsic_submit_bio()
/linux-4.1.27/drivers/block/drbd/
Ddrbd_actlog.c151 rw |= REQ_FUA | REQ_FLUSH; in _drbd_md_sync_page_io()
Ddrbd_req.c1116 D_ASSERT(device, req->master_bio->bi_rw & REQ_FLUSH); in drbd_process_write_request()
Ddrbd_main.c1610 (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) | in bio_flags_to_wire()
2772 blk_queue_flush(q, REQ_FLUSH | REQ_FUA); in drbd_create_device()
Ddrbd_receiver.c2132 (dpf & DP_FLUSH ? REQ_FLUSH : 0) | in wire_flags_to_bio()
/linux-4.1.27/drivers/scsi/
Dsd.c142 flush |= REQ_FLUSH; in sd_set_flush_flag()
1129 else if (rq->cmd_flags & REQ_FLUSH) in sd_init_command()
/linux-4.1.27/fs/xfs/
Dxfs_buf.c1245 rw |= REQ_FLUSH; in _xfs_buf_ioapply()
/linux-4.1.27/fs/jbd/
Djournal.c1062 write_op &= ~(REQ_FUA | REQ_FLUSH); in journal_write_superblock()
/linux-4.1.27/kernel/trace/
Dblktrace.c1791 if (rw & REQ_FLUSH) in blk_fill_rwbs()
/linux-4.1.27/fs/jbd2/
Djournal.c1349 write_op &= ~(REQ_FUA | REQ_FLUSH); in jbd2_write_superblock()