btree1 276 fs/hpfs/anode.c struct bplus_header *btree1 = btree; btree1 287 fs/hpfs/anode.c while (bp_internal(btree1)) { btree1 288 fs/hpfs/anode.c ano = le32_to_cpu(btree1->u.internal[pos].down); btree1 294 fs/hpfs/anode.c btree1 = &anode->btree; btree1 298 fs/hpfs/anode.c for (i = 0; i < btree1->n_used_nodes; i++) btree1 299 fs/hpfs/anode.c hpfs_free_sectors(s, le32_to_cpu(btree1->u.external[i].disk_secno), le32_to_cpu(btree1->u.external[i].length)); btree1 310 fs/hpfs/anode.c btree1 = &anode->btree; btree1 311 fs/hpfs/anode.c } else btree1 = btree; btree1 312 fs/hpfs/anode.c for (i = 0; i < btree1->n_used_nodes; i++) { btree1 313 fs/hpfs/anode.c if (le32_to_cpu(btree1->u.internal[i].down) == oano) { btree1 314 fs/hpfs/anode.c if ((pos = i + 1) < btree1->n_used_nodes) btree1 775 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_node_entry *btree1; btree1 793 fs/xfs/libxfs/xfs_da_btree.c btree1 = dp->d_ops->node_tree_p(node1); btree1 801 fs/xfs/libxfs/xfs_da_btree.c ((be32_to_cpu(btree2[0].hashval) < be32_to_cpu(btree1[0].hashval)) || btree1 803 fs/xfs/libxfs/xfs_da_btree.c be32_to_cpu(btree1[nodehdr1.count - 1].hashval)))) { btree1 809 fs/xfs/libxfs/xfs_da_btree.c btree1 = dp->d_ops->node_tree_p(node1); btree1 839 fs/xfs/libxfs/xfs_da_btree.c btree_s = &btree1[nodehdr1.count - count]; btree1 851 fs/xfs/libxfs/xfs_da_btree.c btree_d = &btree1[nodehdr1.count]; btree1 891 fs/xfs/libxfs/xfs_da_btree.c btree1 = dp->d_ops->node_tree_p(node1); btree1 894 fs/xfs/libxfs/xfs_da_btree.c blk1->hashval = be32_to_cpu(btree1[nodehdr1.count - 1].hashval); btree1 1674 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_node_entry *btree1; btree1 1683 fs/xfs/libxfs/xfs_da_btree.c btree1 = dp->d_ops->node_tree_p(node1); btree1 1687 fs/xfs/libxfs/xfs_da_btree.c ((be32_to_cpu(btree2[0].hashval) < be32_to_cpu(btree1[0].hashval)) || btree1 1689 fs/xfs/libxfs/xfs_da_btree.c be32_to_cpu(btree1[node1hdr.count - 1].hashval)))) {