Lines Matching refs:c
264 struct cache_set *c; member
700 #define btree_bytes(c) ((c)->btree_pages * PAGE_SIZE) argument
702 ((unsigned) (KEY_SIZE(&b->key) >> (b)->c->block_bits))
704 #define btree_default_blocks(c) \ argument
705 ((unsigned) ((PAGE_SECTORS * (c)->btree_pages) >> (c)->block_bits))
707 #define bucket_pages(c) ((c)->sb.bucket_size / PAGE_SECTORS) argument
708 #define bucket_bytes(c) ((c)->sb.bucket_size << 9) argument
709 #define block_bytes(c) ((c)->sb.block_size << 9) argument
711 #define prios_per_bucket(c) \ argument
712 ((bucket_bytes(c) - sizeof(struct prio_set)) / \
714 #define prio_buckets(c) \ argument
715 DIV_ROUND_UP((size_t) (c)->sb.nbuckets, prios_per_bucket(c))
717 static inline size_t sector_to_bucket(struct cache_set *c, sector_t s) in sector_to_bucket() argument
719 return s >> c->bucket_bits; in sector_to_bucket()
722 static inline sector_t bucket_to_sector(struct cache_set *c, size_t b) in bucket_to_sector() argument
724 return ((sector_t) b) << c->bucket_bits; in bucket_to_sector()
727 static inline sector_t bucket_remainder(struct cache_set *c, sector_t s) in bucket_remainder() argument
729 return s & (c->sb.bucket_size - 1); in bucket_remainder()
732 static inline struct cache *PTR_CACHE(struct cache_set *c, in PTR_CACHE() argument
736 return c->cache[PTR_DEV(k, ptr)]; in PTR_CACHE()
739 static inline size_t PTR_BUCKET_NR(struct cache_set *c, in PTR_BUCKET_NR() argument
743 return sector_to_bucket(c, PTR_OFFSET(k, ptr)); in PTR_BUCKET_NR()
746 static inline struct bucket *PTR_BUCKET(struct cache_set *c, in PTR_BUCKET() argument
750 return PTR_CACHE(c, k, ptr)->buckets + PTR_BUCKET_NR(c, k, ptr); in PTR_BUCKET()
759 static inline uint8_t ptr_stale(struct cache_set *c, const struct bkey *k, in ptr_stale() argument
762 return gen_after(PTR_BUCKET(c, k, i)->gen, PTR_GEN(k, i)); in ptr_stale()
765 static inline bool ptr_available(struct cache_set *c, const struct bkey *k, in ptr_available() argument
768 return (PTR_DEV(k, i) < MAX_CACHES_PER_SET) && PTR_CACHE(c, k, i); in ptr_available()
786 if (bch_cache_set_error((b)->c, __VA_ARGS__)) \
790 #define cache_bug(c, ...) \ argument
792 if (bch_cache_set_error(c, __VA_ARGS__)) \
802 #define cache_bug_on(cond, c, ...) \ argument
805 cache_bug(c, __VA_ARGS__); \
808 #define cache_set_err_on(cond, c, ...) \ argument
811 bch_cache_set_error(c, __VA_ARGS__); \
858 static inline void wake_up_allocators(struct cache_set *c) in wake_up_allocators() argument
863 for_each_cache(ca, c, i) in wake_up_allocators()
922 int bch_flash_dev_create(struct cache_set *c, uint64_t size);