Lines Matching refs:queue_depth
1465 tags = blk_mq_init_tags(set->queue_depth, set->reserved_tags, in blk_mq_init_rq_map()
1473 tags->rqs = kzalloc_node(set->queue_depth * sizeof(struct request *), in blk_mq_init_rq_map()
1487 left = rq_size * set->queue_depth; in blk_mq_init_rq_map()
1489 for (i = 0; i < set->queue_depth; ) { in blk_mq_init_rq_map()
1523 to_do = min(entries_per_page, set->queue_depth - i); in blk_mq_init_rq_map()
1636 unsigned flush_start_tag = set->queue_depth; in blk_mq_exit_hctx()
1681 unsigned flush_start_tag = set->queue_depth; in blk_mq_init_hctx()
2052 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
2227 depth = set->queue_depth; in blk_mq_alloc_rq_maps()
2233 set->queue_depth >>= 1; in blk_mq_alloc_rq_maps()
2234 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_rq_maps()
2238 } while (set->queue_depth); in blk_mq_alloc_rq_maps()
2240 if (!set->queue_depth || err) { in blk_mq_alloc_rq_maps()
2245 if (depth != set->queue_depth) in blk_mq_alloc_rq_maps()
2247 depth, set->queue_depth); in blk_mq_alloc_rq_maps()
2270 if (!set->queue_depth) in blk_mq_alloc_tag_set()
2272 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) in blk_mq_alloc_tag_set()
2278 if (set->queue_depth > BLK_MQ_MAX_DEPTH) { in blk_mq_alloc_tag_set()
2281 set->queue_depth = BLK_MQ_MAX_DEPTH; in blk_mq_alloc_tag_set()
2291 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set()
2334 if (!set || nr > set->queue_depth) in blk_mq_update_nr_requests()