Searched refs:mq_ops (Results 1 – 13 of 13) sorted by relevance
/linux-4.1.27/block/ |
D | blk-flush.c | 134 if (rq->q->mq_ops) { in blk_flush_queue_rq() 205 if (q->mq_ops) in blk_flush_complete_seq() 228 if (q->mq_ops) { in flush_end_io() 239 if (!q->mq_ops) in flush_end_io() 262 WARN_ON(q->mq_ops); in flush_end_io() 266 if (q->mq_ops) in flush_end_io() 313 if (q->mq_ops) { in blk_kick_flush() 347 hctx = q->mq_ops->map_queue(q, ctx->cpu); in mq_flush_data_end_io() 393 if (q->mq_ops) in blk_insert_flush() 409 if (q->mq_ops) { in blk_insert_flush() [all …]
|
D | blk-timeout.c | 162 if (req->q->mq_ops) in blk_abort_request() 197 if (!q->mq_ops && !q->rq_timed_out_fn) in blk_add_timer() 210 if (!q->mq_ops) in blk_add_timer()
|
D | blk-sysfs.c | 54 if (!q->request_fn && !q->mq_ops) in queue_requests_store() 518 if (!q->mq_ops) in blk_release_queue() 561 if (q->mq_ops) in blk_register_queue() 577 if (q->mq_ops) in blk_register_queue() 602 if (q->mq_ops) in blk_unregister_queue()
|
D | blk-mq.c | 267 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_alloc_request() 277 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_alloc_request() 322 hctx = q->mq_ops->map_queue(q, rq->mq_ctx->cpu); in blk_mq_free_request() 593 struct blk_mq_ops *ops = req->q->mq_ops; in blk_mq_rq_timed_out() 820 ret = q->mq_ops->queue_rq(hctx, &bd); in __blk_mq_run_hw_queue() 1044 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_insert_request() 1072 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_insert_requests() 1211 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_map_request() 1226 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_mq_map_request() 1289 ret = q->mq_ops->queue_rq(data.hctx, &bd); in blk_mq_make_request() [all …]
|
D | blk.h | 38 if (!q->mq_ops) in blk_get_flush_queue() 41 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_get_flush_queue()
|
D | blk-exec.c | 68 if (q->mq_ops) { in blk_execute_rq_nowait()
|
D | blk-mq-tag.c | 300 data->hctx = data->q->mq_ops->map_queue(data->q, in bt_get() 636 if (q->mq_ops) { in blk_mq_unique_tag() 637 hctx = q->mq_ops->map_queue(q, rq->mq_ctx->cpu); in blk_mq_unique_tag()
|
D | blk-core.c | 247 if (q->mq_ops) { in blk_sync_queue() 480 if (q->mq_ops) in blk_set_queue_dying() 533 if (q->mq_ops) { in blk_cleanup_queue() 547 if (q->mq_ops) in blk_cleanup_queue() 1213 if (q->mq_ops) in blk_get_request() 1385 if (q->mq_ops) { in __blk_put_request() 1419 if (q->mq_ops) in blk_put_request() 1540 if (q->mq_ops) in blk_attempt_plug_merge() 2067 if (q->mq_ops) { in blk_insert_cloned_request()
|
D | blk-merge.c | 356 return !q->mq_ops && req->special; in req_no_special_merge()
|
/linux-4.1.27/drivers/md/ |
D | dm.c | 1046 return (rq->q->mq_ops ? blk_mq_rq_to_pdu(rq) : rq->special); in tio_from_request() 1068 if (!md->queue->mq_ops && run_queue) in rq_completed() 1087 else if (!md->queue->mq_ops) in free_rq_clone() 1096 if (!md->queue->mq_ops) in free_rq_clone() 1126 if (!rq->q->mq_ops) in dm_end_request() 1138 if (!rq->q->mq_ops) { in dm_unprep_request() 1145 else if (!tio->md->queue->mq_ops) in dm_unprep_request() 1170 if (!rq->q->mq_ops) in dm_requeue_unmapped_original_request() 1201 if (!q->mq_ops) in stop_queue() 1219 if (!q->mq_ops) in start_queue() [all …]
|
D | dm-table.c | 897 if (q->mq_ops) in dm_table_set_type() 904 if (!bdev_get_queue(dd->dm_dev->bdev)->mq_ops) { in dm_table_set_type() 1697 if (queue->mq_ops) in dm_table_run_md_queue_async()
|
/linux-4.1.27/drivers/scsi/ |
D | scsi_lib.c | 166 if (q->mq_ops) { in __scsi_queue_insert() 330 if (q->mq_ops) in scsi_kick_queue() 494 if (q->mq_ops) in scsi_run_queue() 1062 if (q->mq_ops) { in scsi_io_completion() 1129 if (!rq->q->mq_ops) { in scsi_init_io() 1424 if (!q->mq_ops) in scsi_dev_queue_ready() 2962 if (q->mq_ops) { in scsi_internal_device_block() 3014 if (q->mq_ops) { in scsi_internal_device_unblock()
|
/linux-4.1.27/include/linux/ |
D | blkdev.h | 340 struct blk_mq_ops *mq_ops; member 632 return q->request_fn || q->mq_ops; in queue_is_rq_based()
|