clump_size        151 fs/hfs/btree.h 	__be32 clump_size;	/* (F) clump size. not usually used. */
clump_size        209 fs/hfs/hfs_fs.h 			__be32 log_size, __be32 phys_size, u32 clump_size);
clump_size        266 fs/hfs/inode.c 			 __be32 __log_size, __be32 phys_size, u32 clump_size)
clump_size        283 fs/hfs/inode.c 	HFS_I(inode)->clump_blocks = clump_size / HFS_SB(sb)->alloc_blksz;
clump_size         79 fs/hfsplus/btree.c 	u32 clump_size;
clump_size        101 fs/hfsplus/btree.c 		clump_size = sectors << 2;	/*  0.8 %  */
clump_size        102 fs/hfsplus/btree.c 		if (clump_size < (8 * node_size))
clump_size        103 fs/hfsplus/btree.c 			clump_size = 8 * node_size;
clump_size        112 fs/hfsplus/btree.c 		clump_size = clumptbl[column + (i) * 3] * 1024 * 1024;
clump_size        119 fs/hfsplus/btree.c 	clump_size /= mod;
clump_size        120 fs/hfsplus/btree.c 	clump_size *= mod;
clump_size        126 fs/hfsplus/btree.c 	if (clump_size == 0)
clump_size        127 fs/hfsplus/btree.c 		clump_size = mod;
clump_size        129 fs/hfsplus/btree.c 	return clump_size;
clump_size         98 fs/hfsplus/hfsplus_raw.h 	__be32 clump_size;
clump_size        178 fs/hfsplus/hfsplus_raw.h 	__be32 clump_size;
clump_size        471 fs/hfsplus/inode.c 		be32_to_cpu(fork->clump_size) >> sbi->alloc_blksz_shift;
clump_size         54 fs/hfsplus/xattr.c 					u32 clump_size,
clump_size         68 fs/hfsplus/xattr.c 		clump_size, node_size);
clump_size         85 fs/hfsplus/xattr.c 	head->clump_size = cpu_to_be32(clump_size);
clump_size        127 fs/hfsplus/xattr.c 	u32 clump_size;
clump_size        179 fs/hfsplus/xattr.c 	clump_size = hfsplus_calc_btree_clump_size(sb->s_blocksize,
clump_size        185 fs/hfsplus/xattr.c 	hip->clump_blocks = clump_size >> sbi->alloc_blksz_shift;
clump_size        212 fs/hfsplus/xattr.c 	hfsplus_init_header_node(attr_file, clump_size, buf, node_size);