Lines Matching refs:key

116 #define btree(fn, key, b, op, ...)					\  argument
120 struct btree *_child = bch_btree_node_get((b)->c, op, key, l, \
191 uint64_t crc = b->key.ptr[0]; in btree_csum_set()
254 bset_sector_offset(&b->keys, i) < KEY_SIZE(&b->key); in bch_btree_node_read_done()
264 bkey_cmp(&b->key, &b->keys.set[0].end) < 0) in bch_btree_node_read_done()
276 err, PTR_BUCKET_NR(b->c, &b->key, 0), in bch_btree_node_read_done()
299 bio->bi_iter.bi_size = KEY_SIZE(&b->key) << 9; in bch_btree_node_read()
305 bch_submit_bbio(bio, b->c, &b->key, 0); in bch_btree_node_read()
322 PTR_BUCKET_NR(b->c, &b->key, 0)); in bch_btree_node_read()
390 BKEY_PADDED(key) k; in do_btree_node_write()
419 bkey_copy(&k.key, &b->key); in do_btree_node_write()
420 SET_PTR_OFFSET(&k.key, 0, PTR_OFFSET(&k.key, 0) + in do_btree_node_write()
432 bch_submit_bbio(b->bio, b->c, &k.key, 0); in do_btree_node_write()
439 bch_submit_bbio(b->bio, b->c, &k.key, 0); in do_btree_node_write()
472 &PTR_CACHE(b->c, &b->key, 0)->btree_sectors_written); in __bch_btree_node_write()
576 b->key.ptr[0] = 0; in mca_bucket_free()
829 if (PTR_HASH(c, &b->key) == PTR_HASH(c, k)) in mca_find()
929 bkey_copy(&b->key, k); in mca_alloc()
998 if (PTR_HASH(c, &b->key) != PTR_HASH(c, k)) { in bch_btree_node_get()
1060 bch_bucket_free(b->c, &b->key); in btree_node_free()
1069 BKEY_PADDED(key) k; in __bch_btree_node_alloc()
1074 if (__bch_bucket_alloc_set(c, RESERVE_BTREE, &k.key, 1, wait)) in __bch_btree_node_alloc()
1077 bkey_put(c, &k.key); in __bch_btree_node_alloc()
1078 SET_KEY_SIZE(&k.key, c->btree_pages * PAGE_SECTORS); in __bch_btree_node_alloc()
1080 b = mca_alloc(c, op, &k.key, level); in __bch_btree_node_alloc()
1099 bch_bucket_free(c, &k.key); in __bch_btree_node_alloc()
1121 bkey_copy_key(&n->key, &b->key); in btree_node_alloc_replacement()
1136 bkey_copy(k, &b->key); in make_btree_freeing_key()
1141 bch_inc_gen(PTR_CACHE(b->c, &b->key, i), in make_btree_freeing_key()
1142 PTR_BUCKET(b->c, &b->key, i))); in make_btree_freeing_key()
1272 bkey_cmp(&b->key, &t->end) < 0, in btree_gc_mark_node()
1377 last = &r->b->key; in btree_gc_coalesce()
1384 bkey_copy_key(&new_nodes[i]->key, last); in btree_gc_coalesce()
1401 bkey_u64s(&new_nodes[i]->key))) in btree_gc_coalesce()
1405 bch_keylist_add(&keylist, &new_nodes[i]->key); in btree_gc_coalesce()
1420 if (__bch_keylist_realloc(&keylist, bkey_u64s(&r[i].b->key))) in btree_gc_coalesce()
1485 bch_keylist_add(&keys, &n->key); in btree_gc_rewrite_node()
1561 bkey_copy_key(&b->c->gc_done, &last->b->key); in btree_gc_recurse()
1617 __bch_btree_mark_key(b->c, b->level + 1, &b->key); in bch_btree_gc_root()
1625 bkey_copy_key(&b->c->gc_done, &b->key); in bch_btree_gc_root()
1688 for (j = 0; j < KEY_PTRS(&w->key); j++) in bch_btree_gc_finish()
1689 SET_GC_MARK(PTR_BUCKET(c, &w->key, j), in bch_btree_gc_finish()
1818 bch_initial_mark_key(b->c, b->level + 1, &b->key); in bch_btree_check_recurse()
1891 BUG_ON(bkey_cmp(k, &b->key) > 0); in btree_insert_key()
1931 if (bkey_cmp(k, &b->key) <= 0) { in bch_btree_insert_keys()
1937 } else if (bkey_cmp(&START_KEY(k), &b->key) < 0) { in bch_btree_insert_keys()
1938 BKEY_PADDED(key) temp; in bch_btree_insert_keys()
1939 bkey_copy(&temp.key, insert_keys->keys); in bch_btree_insert_keys()
1941 bch_cut_back(&b->key, &temp.key); in bch_btree_insert_keys()
1942 bch_cut_front(&b->key, insert_keys->keys); in bch_btree_insert_keys()
1944 ret |= btree_insert_key(b, &temp.key, replace_key); in bch_btree_insert_keys()
2016 bkey_copy_key(&n1->key, in btree_split()
2027 bkey_copy_key(&n2->key, &b->key); in btree_split()
2029 bch_keylist_add(&parent_keys, &n2->key); in btree_split()
2040 bch_keylist_add(&parent_keys, &n1->key); in btree_split()
2047 bkey_copy_key(&n3->key, &MAX_KEY); in btree_split()
2076 bkey_put(b->c, &n2->key); in btree_split()
2080 bkey_put(b->c, &n1->key); in btree_split()
2154 uint64_t btree_ptr = b->key.ptr[0]; in bch_btree_insert_check_key()
2165 if (b->key.ptr[0] != btree_ptr || in bch_btree_insert_check_key()
2253 for (i = 0; i < KEY_PTRS(&b->key); i++) in bch_btree_set_root()
2254 BUG_ON(PTR_BUCKET(b->c, &b->key, i)->prio != BTREE_PRIO); in bch_btree_set_root()
2324 ret = fn(op, b, &KEY(KEY_INODE(&b->key), in bch_btree_map_keys_recurse()
2325 KEY_OFFSET(&b->key), 0)); in bch_btree_map_keys_recurse()
2341 if (bkey_cmp(&l->key, &START_KEY(&r->key)) <= 0) in keybuf_cmp()
2343 if (bkey_cmp(&START_KEY(&l->key), &r->key) >= 0) in keybuf_cmp()
2351 return clamp_t(int64_t, bkey_cmp(&l->key, &r->key), -1, 1); in keybuf_nonoverlapping_cmp()
2389 bkey_copy(&w->key, k); in refill_keybuf_fn()
2433 buf->start = START_KEY(&w->key); in bch_refill_keybuf()
2436 buf->end = w->key; in bch_refill_keybuf()
2463 s.key = *start; in bch_keybuf_check_overlapping()
2472 while (w && bkey_cmp(&START_KEY(&w->key), end) < 0) { in bch_keybuf_check_overlapping()