Lines Matching refs:q
33 struct request_queue *q; member
313 static void null_queue_bio(struct request_queue *q, struct bio *bio) in null_queue_bio() argument
315 struct nullb *nullb = q->queuedata; in null_queue_bio()
325 static int null_rq_prep_fn(struct request_queue *q, struct request *req) in null_rq_prep_fn() argument
327 struct nullb *nullb = q->queuedata; in null_rq_prep_fn()
341 static void null_request_fn(struct request_queue *q) in null_request_fn() argument
345 while ((rq = blk_fetch_request(q)) != NULL) { in null_request_fn()
348 spin_unlock_irq(q->queue_lock); in null_request_fn()
350 spin_lock_irq(q->queue_lock); in null_request_fn()
402 blk_cleanup_queue(nullb->q); in null_del_dev()
532 nullb->q = blk_mq_init_queue(&nullb->tag_set); in null_add_dev()
533 if (IS_ERR(nullb->q)) { in null_add_dev()
538 nullb->q = blk_alloc_queue_node(GFP_KERNEL, home_node); in null_add_dev()
539 if (!nullb->q) { in null_add_dev()
543 blk_queue_make_request(nullb->q, null_queue_bio); in null_add_dev()
548 nullb->q = blk_init_queue_node(null_request_fn, &nullb->lock, home_node); in null_add_dev()
549 if (!nullb->q) { in null_add_dev()
553 blk_queue_prep_rq(nullb->q, null_rq_prep_fn); in null_add_dev()
554 blk_queue_softirq_done(nullb->q, null_softirq_done_fn); in null_add_dev()
560 nullb->q->queuedata = nullb; in null_add_dev()
561 queue_flag_set_unlocked(QUEUE_FLAG_NONROT, nullb->q); in null_add_dev()
562 queue_flag_clear_unlocked(QUEUE_FLAG_ADD_RANDOM, nullb->q); in null_add_dev()
575 blk_queue_logical_block_size(nullb->q, bs); in null_add_dev()
576 blk_queue_physical_block_size(nullb->q, bs); in null_add_dev()
587 disk->queue = nullb->q; in null_add_dev()
593 blk_cleanup_queue(nullb->q); in null_add_dev()