cache_alloc      2280 drivers/md/bcache/super.c 	ret = cache_alloc(ca);
cache_alloc       811 drivers/md/dm-cache-policy-smq.c 	struct entry_alloc cache_alloc;
cache_alloc       988 drivers/md/dm-cache-policy-smq.c 	return to_cblock(get_index(&mq->cache_alloc, e));
cache_alloc      1043 drivers/md/dm-cache-policy-smq.c 	unsigned threshold_level = allocator_empty(&mq->cache_alloc) ?
cache_alloc      1153 drivers/md/dm-cache-policy-smq.c 	nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated;
cache_alloc      1236 drivers/md/dm-cache-policy-smq.c 	if (allocator_empty(&mq->cache_alloc)) {
cache_alloc      1253 drivers/md/dm-cache-policy-smq.c 	e = alloc_entry(&mq->cache_alloc);
cache_alloc      1261 drivers/md/dm-cache-policy-smq.c 		free_entry(&mq->cache_alloc, e);
cache_alloc      1284 drivers/md/dm-cache-policy-smq.c 		if (!allocator_empty(&mq->cache_alloc) && fast_promote)
cache_alloc      1462 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc,
cache_alloc      1475 drivers/md/dm-cache-policy-smq.c 			free_entry(&mq->cache_alloc, e);
cache_alloc      1484 drivers/md/dm-cache-policy-smq.c 			free_entry(&mq->cache_alloc, e);
cache_alloc      1519 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
cache_alloc      1562 drivers/md/dm-cache-policy-smq.c 	e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock));
cache_alloc      1580 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
cache_alloc      1588 drivers/md/dm-cache-policy-smq.c 	free_entry(&mq->cache_alloc, e);
cache_alloc      1595 drivers/md/dm-cache-policy-smq.c 	struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
cache_alloc      1610 drivers/md/dm-cache-policy-smq.c 	r = to_cblock(mq->cache_alloc.nr_allocated);
cache_alloc      1758 drivers/md/dm-cache-policy-smq.c 	init_allocator(&mq->cache_alloc, &mq->es,