/linux-4.1.27/drivers/md/bcache/ |
H A D | extents.c | 56 if (KEY_SIZE(k) + r > c->sb.bucket_size || __ptr_invalid() 77 if (KEY_SIZE(k) + r > c->sb.bucket_size) bch_ptr_status() 91 if (!KEY_SIZE(k)) bch_ptr_status() 103 p("%llu:%llu len %llu -> [", KEY_INODE(k), KEY_START(k), KEY_SIZE(k)); bch_extent_to_text() 152 if (!KEY_PTRS(k) || !KEY_SIZE(k) || KEY_DIRTY(k)) __bch_btree_ptr_invalid() 277 if (!KEY_SIZE(i->k)) { bch_extent_sort_fixup() 332 BUG_ON(!KEY_SIZE(insert)); bch_extent_insert_fixup() 340 if (KEY_SIZE(k)) bch_extent_insert_fixup() 350 old_size = KEY_SIZE(k); bch_extent_insert_fixup() 360 if (replace_key && KEY_SIZE(k)) { bch_extent_insert_fixup() 405 KEY_SIZE(insert)); bch_extent_insert_fixup() 455 bch_subtract_dirty(k, c, old_offset, old_size - KEY_SIZE(k)); bch_extent_insert_fixup() 462 } else if (sectors_found < KEY_SIZE(insert)) { bch_extent_insert_fixup() 464 (KEY_SIZE(insert) - sectors_found)); bch_extent_insert_fixup() 472 KEY_SIZE(insert)); bch_extent_insert_fixup() 481 if (!KEY_SIZE(k)) __bch_extent_invalid() 484 if (KEY_SIZE(k) > KEY_OFFSET(k)) __bch_extent_invalid() 558 btree_bug_on(stale && KEY_DIRTY(k) && KEY_SIZE(k), bch_extent_bad() 587 if (l->ptr[i] + PTR(0, KEY_SIZE(l), 0) != r->ptr[i] || bch_extent_merge() 592 * overflow KEY_SIZE bch_extent_merge() 594 if (KEY_SIZE(l) + KEY_SIZE(r) > USHRT_MAX) { bch_extent_merge() 595 SET_KEY_OFFSET(l, KEY_OFFSET(l) + USHRT_MAX - KEY_SIZE(l)); bch_extent_merge() 609 SET_KEY_OFFSET(l, KEY_OFFSET(l) + KEY_SIZE(r)); bch_extent_merge() 610 SET_KEY_SIZE(l, KEY_SIZE(l) + KEY_SIZE(r)); bch_extent_merge()
|
H A D | debug.c | 25 (void *) i < (void *) (start) + (KEY_SIZE(&b->key) << 9) &&\ 54 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; bch_btree_verify() 60 memcpy(ondisk, sorted, KEY_SIZE(&v->key) << 9); bch_btree_verify()
|
H A D | movinggc.c | 87 bio->bi_iter.bi_size = KEY_SIZE(&io->w->key) << 9; moving_init() 88 bio->bi_max_vecs = DIV_ROUND_UP(KEY_SIZE(&io->w->key), moving_init() 152 * DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS), read_moving()
|
H A D | writeback.c | 114 bio->bi_iter.bi_size = KEY_SIZE(&w->key) << 9; dirty_init() 115 bio->bi_max_vecs = DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS); dirty_init() 247 * DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS), read_dirty() 270 delay = writeback_delay(dc, KEY_SIZE(&w->key)); read_dirty() 486 KEY_START(k), KEY_SIZE(k)); sectors_dirty_init_fn()
|
H A D | bset.c | 60 ret += KEY_SIZE(k); __bch_count_data() 206 SET_PTR_OFFSET(k, i, PTR_OFFSET(k, i) + KEY_SIZE(k) - len); __bch_cut_front() 208 BUG_ON(len > KEY_SIZE(k)); __bch_cut_front() 227 BUG_ON(len > KEY_SIZE(k)); __bch_cut_back() 827 BUG_ON(b->ops->is_extents && !KEY_SIZE(k)); bch_btree_insert_key() 850 KEY_PTRS(m) == KEY_PTRS(k) && !KEY_SIZE(m)) bch_btree_insert_key()
|
H A D | btree.c | 254 bset_sector_offset(&b->keys, i) < KEY_SIZE(&b->key); bch_btree_node_read_done() 299 bio->bi_iter.bi_size = KEY_SIZE(&b->key) << 9; bch_btree_node_read() 583 return ilog2(KEY_SIZE(k) / PAGE_SECTORS ?: 1); btree_order() 1214 GC_SECTORS_USED(g) + KEY_SIZE(k), __bch_btree_mark_key() 1266 gc->data += KEY_SIZE(k); btree_gc_mark_node() 2374 if (!KEY_SIZE(k)) /* end key */ refill_keybuf_fn()
|
H A D | alloc.c | 563 * allocated space, and updates KEY_SIZE(k) and KEY_OFFSET(k) (to point to the 566 * May allocate fewer sectors than @sectors, KEY_SIZE(k) indicates how many
|
H A D | request.c | 231 n = bio_next_split(bio, KEY_SIZE(k), GFP_NOIO, split); bch_data_insert_start() 528 if (!KEY_SIZE(k)) cache_lookup_fn()
|
H A D | bcache.h | 702 ((unsigned) (KEY_SIZE(&b->key) >> (b)->c->block_bits))
|
H A D | super.c | 375 bio->bi_iter.bi_size = KEY_SIZE(k) << 9; uuid_io()
|
/linux-4.1.27/fs/reiserfs/ |
H A D | ibalance.c | 146 RFALSE(B_FREE_SPACE(cur) < count * (KEY_SIZE + DC_SIZE), internal_insert_childs() 148 B_FREE_SPACE(cur), count * (KEY_SIZE + DC_SIZE)); internal_insert_childs() 167 (nr - to) * KEY_SIZE + (nr + 1 + count) * DC_SIZE); internal_insert_childs() 170 memcpy(ih, inserted, KEY_SIZE); internal_insert_childs() 172 memcpy(ih + 1, inserted + 1, KEY_SIZE); internal_insert_childs() 178 KEY_SIZE)); internal_insert_childs() 190 dc_size(t_dc) + (count * (DC_SIZE + KEY_SIZE))); internal_insert_childs() 248 (nr - first_i - del_num) * KEY_SIZE + (nr + 1 - internal_delete_pointers_items() 255 (del_num * (KEY_SIZE + DC_SIZE))); internal_delete_pointers_items() 266 dc_size(t_dc) - (del_num * (KEY_SIZE + DC_SIZE))); internal_delete_pointers_items() 353 KEY_SIZE * (nr_dest - dest_order) + DC_SIZE * (nr_dest + internal_copy_pointers_items() 357 memcpy(key, internal_key(src, src_order), KEY_SIZE * (cpy_num - 1)); internal_copy_pointers_items() 362 blkh_free_space(blkh) - (KEY_SIZE * (cpy_num - 1) + internal_copy_pointers_items() 375 dc_size(t_dc) + (KEY_SIZE * (cpy_num - 1) + internal_copy_pointers_items() 449 RFALSE(B_FREE_SPACE(dest) < KEY_SIZE, internal_insert_key() 458 (nr - dest_position_before) * KEY_SIZE + (nr + 1) * DC_SIZE); internal_insert_key() 461 memcpy(key, internal_key(src, src_position), KEY_SIZE); internal_insert_key() 466 set_blkh_free_space(blkh, blkh_free_space(blkh) - KEY_SIZE); internal_insert_key() 473 put_dc_size(t_dc, dc_size(t_dc) + KEY_SIZE); internal_insert_key() 640 insert_num = tb->insert_size[h] / ((int)(DC_SIZE + KEY_SIZE)); balance_internal_when_delete() 769 memcpy(internal_key(tb->CFL[h], tb->lkey[h]), key, KEY_SIZE); replace_lkey() 784 memcpy(internal_key(tb->CFR[h], tb->rkey[h]), key, KEY_SIZE); replace_rkey() 837 insert_num = tb->insert_size[h] / ((int)(KEY_SIZE + DC_SIZE)); balance_internal() 1066 KEY_SIZE); balance_internal() 1079 KEY_SIZE); balance_internal() 1113 KEY_SIZE); balance_internal() 1156 memcpy(new_insert_key_addr, &new_insert_key, KEY_SIZE); balance_internal()
|
H A D | fix_node.c | 66 vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE); create_virtual_node() 205 tb->lnum[h] = cur_free / (DC_SIZE + KEY_SIZE); check_left() 291 tb->rnum[h] = cur_free / (DC_SIZE + KEY_SIZE); check_right() 432 i = ((to - from) * (KEY_SIZE + DC_SIZE) + DC_SIZE); get_num_ver() 1295 + ((h) ? KEY_SIZE : 0)) { can_node_be_removed() 1800 (DC_SIZE + KEY_SIZE); dc_check_balance_internal() 1817 (DC_SIZE + KEY_SIZE); dc_check_balance_internal() 1862 KEY_SIZE); dc_check_balance_internal() 1877 KEY_SIZE); dc_check_balance_internal() 2695 KEY_SIZE) * (tb->blknum[h] - 1) + fix_nodes() 2701 (DC_SIZE + KEY_SIZE) * (tb->blknum[h] - 1); fix_nodes()
|
H A D | do_balan.c | 1260 memcpy(insert_key + i, leaf_key(tb->S_new[i], 0), KEY_SIZE); balance_leaf_new_nodes() 1582 KEY_SIZE); replace_key() local 1585 KEY_SIZE); replace_key() local
|
H A D | stree.c | 481 if (nr > (blocksize - BLKH_SIZE - DC_SIZE) / (KEY_SIZE + DC_SIZE)) { is_internal() 487 used_space = BLKH_SIZE + KEY_SIZE * nr + DC_SIZE * (nr + 1); is_internal() 750 KEY_SIZE, search_by_key()
|
H A D | prints.c | 741 bh->b_size - BLKH_SIZE - KEY_SIZE * B_NR_ITEMS(bh) - check_internal_block_head()
|
H A D | reiserfs.h | 1328 #define KEY_SIZE (sizeof(struct reiserfs_key)) macro 2049 ((bh)->b_data + BLKH_SIZE + B_NR_ITEMS(bh) * KEY_SIZE + DC_SIZE * (n_pos))) 2064 #define MAX_NR_KEY(bh) ( (MAX_CHILD_SIZE(bh)-DC_SIZE)/(KEY_SIZE+DC_SIZE) ) 2995 memcpy(to, from, KEY_SIZE); copy_key()
|
H A D | lbalance.c | 73 memcpy(&new_ih.ih_key, &ih->ih_key, KEY_SIZE); leaf_copy_dir_entries()
|
H A D | inode.c | 1522 memset(INODE_PKEY(inode), 0, KEY_SIZE); reiserfs_make_bad_inode() local 1970 memcpy(INODE_PKEY(inode), &ih.ih_key, KEY_SIZE); reiserfs_new_inode() local
|
/linux-4.1.27/include/uapi/linux/ |
H A D | bcache.h | 50 KEY_FIELD(KEY_SIZE, high, 20, KEY_SIZE_BITS) 82 #define KEY_START(k) (KEY_OFFSET(k) - KEY_SIZE(k))
|
/linux-4.1.27/include/trace/events/ |
H A D | bcache.h | 54 __entry->size = KEY_SIZE(k); 329 __entry->size = KEY_SIZE(k);
|