cache_alloc 2280 drivers/md/bcache/super.c ret = cache_alloc(ca); cache_alloc 811 drivers/md/dm-cache-policy-smq.c struct entry_alloc cache_alloc; cache_alloc 988 drivers/md/dm-cache-policy-smq.c return to_cblock(get_index(&mq->cache_alloc, e)); cache_alloc 1043 drivers/md/dm-cache-policy-smq.c unsigned threshold_level = allocator_empty(&mq->cache_alloc) ? cache_alloc 1153 drivers/md/dm-cache-policy-smq.c nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; cache_alloc 1236 drivers/md/dm-cache-policy-smq.c if (allocator_empty(&mq->cache_alloc)) { cache_alloc 1253 drivers/md/dm-cache-policy-smq.c e = alloc_entry(&mq->cache_alloc); cache_alloc 1261 drivers/md/dm-cache-policy-smq.c free_entry(&mq->cache_alloc, e); cache_alloc 1284 drivers/md/dm-cache-policy-smq.c if (!allocator_empty(&mq->cache_alloc) && fast_promote) cache_alloc 1462 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, cache_alloc 1475 drivers/md/dm-cache-policy-smq.c free_entry(&mq->cache_alloc, e); cache_alloc 1484 drivers/md/dm-cache-policy-smq.c free_entry(&mq->cache_alloc, e); cache_alloc 1519 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); cache_alloc 1562 drivers/md/dm-cache-policy-smq.c e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); cache_alloc 1580 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); cache_alloc 1588 drivers/md/dm-cache-policy-smq.c free_entry(&mq->cache_alloc, e); cache_alloc 1595 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); cache_alloc 1610 drivers/md/dm-cache-policy-smq.c r = to_cblock(mq->cache_alloc.nr_allocated); cache_alloc 1758 drivers/md/dm-cache-policy-smq.c init_allocator(&mq->cache_alloc, &mq->es,