clump_size 151 fs/hfs/btree.h __be32 clump_size; /* (F) clump size. not usually used. */ clump_size 209 fs/hfs/hfs_fs.h __be32 log_size, __be32 phys_size, u32 clump_size); clump_size 266 fs/hfs/inode.c __be32 __log_size, __be32 phys_size, u32 clump_size) clump_size 283 fs/hfs/inode.c HFS_I(inode)->clump_blocks = clump_size / HFS_SB(sb)->alloc_blksz; clump_size 79 fs/hfsplus/btree.c u32 clump_size; clump_size 101 fs/hfsplus/btree.c clump_size = sectors << 2; /* 0.8 % */ clump_size 102 fs/hfsplus/btree.c if (clump_size < (8 * node_size)) clump_size 103 fs/hfsplus/btree.c clump_size = 8 * node_size; clump_size 112 fs/hfsplus/btree.c clump_size = clumptbl[column + (i) * 3] * 1024 * 1024; clump_size 119 fs/hfsplus/btree.c clump_size /= mod; clump_size 120 fs/hfsplus/btree.c clump_size *= mod; clump_size 126 fs/hfsplus/btree.c if (clump_size == 0) clump_size 127 fs/hfsplus/btree.c clump_size = mod; clump_size 129 fs/hfsplus/btree.c return clump_size; clump_size 98 fs/hfsplus/hfsplus_raw.h __be32 clump_size; clump_size 178 fs/hfsplus/hfsplus_raw.h __be32 clump_size; clump_size 471 fs/hfsplus/inode.c be32_to_cpu(fork->clump_size) >> sbi->alloc_blksz_shift; clump_size 54 fs/hfsplus/xattr.c u32 clump_size, clump_size 68 fs/hfsplus/xattr.c clump_size, node_size); clump_size 85 fs/hfsplus/xattr.c head->clump_size = cpu_to_be32(clump_size); clump_size 127 fs/hfsplus/xattr.c u32 clump_size; clump_size 179 fs/hfsplus/xattr.c clump_size = hfsplus_calc_btree_clump_size(sb->s_blocksize, clump_size 185 fs/hfsplus/xattr.c hip->clump_blocks = clump_size >> sbi->alloc_blksz_shift; clump_size 212 fs/hfsplus/xattr.c hfsplus_init_header_node(attr_file, clump_size, buf, node_size);