Lines Matching refs:from_cblock
983 if (!test_and_set_bit(from_cblock(infer_cblock(mq, e)), mq->cache_hit_bits)) { in requeue()
1088 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in end_cache_period()
1369 e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_load_mapping()
1438 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in __remove_cblock()
1472 unsigned nr_clean = from_cblock(mq->cache_size) - q_size(&mq->dirty); in clean_target_met()
1473 unsigned target = from_cblock(mq->cache_size) * CLEAN_TARGET_CRITICAL / 100u; in clean_target_met()
1625 calc_hotspot_params(origin_size, cache_block_size, from_cblock(cache_size), in smq_create()
1630 if (space_init(&mq->es, total_sentinels + mq->nr_hotspot_blocks + from_cblock(cache_size))) { in smq_create()
1648 total_sentinels + mq->nr_hotspot_blocks + from_cblock(cache_size)); in smq_create()
1657 if (from_cblock(cache_size)) { in smq_create()
1658 mq->cache_hit_bits = alloc_bitset(from_cblock(cache_size)); in smq_create()
1663 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in smq_create()
1673 from_cblock(mq->cache_size) / mq->cache_blocks_per_hotspot_block); in smq_create()
1681 if (h_init(&mq->table, &mq->es, from_cblock(cache_size))) in smq_create()