rb_entry 41 arch/arm/xen/p2m.c entry = rb_entry(parent, struct xen_p2m_entry, rbnode_phys); rb_entry 71 arch/arm/xen/p2m.c entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys); rb_entry 132 arch/arm/xen/p2m.c p2m_entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys); rb_entry 60 arch/powerpc/kernel/eeh_cache.c piar = rb_entry(n, struct pci_io_addr_range, rb_node); rb_entry 106 arch/powerpc/kernel/eeh_cache.c piar = rb_entry(n, struct pci_io_addr_range, rb_node); rb_entry 128 arch/powerpc/kernel/eeh_cache.c piar = rb_entry(parent, struct pci_io_addr_range, rb_node); rb_entry 229 arch/powerpc/kernel/eeh_cache.c piar = rb_entry(n, struct pci_io_addr_range, rb_node); rb_entry 278 arch/powerpc/kernel/eeh_cache.c piar = rb_entry(n, struct pci_io_addr_range, rb_node); rb_entry 570 arch/powerpc/perf/hv-24x7.c it = rb_entry(*new, struct event_uniq, node); rb_entry 323 arch/sh/kernel/dwarf.c cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node); rb_entry 359 arch/sh/kernel/dwarf.c fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node); rb_entry 840 arch/sh/kernel/dwarf.c cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node); rb_entry 926 arch/sh/kernel/dwarf.c fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node); rb_entry 51 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(node, struct memtype, rb); rb_entry 70 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(node, struct memtype, rb); rb_entry 112 arch/x86/mm/pat_rbtree.c match = rb_entry(node, struct memtype, rb); rb_entry 141 arch/x86/mm/pat_rbtree.c match = rb_entry(node, struct memtype, rb); rb_entry 172 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(*node, struct memtype, rb); rb_entry 261 arch/x86/mm/pat_rbtree.c struct memtype *this = rb_entry(node, struct memtype, rb); rb_entry 564 block/bfq-iosched.c bfqq = rb_entry(parent, struct bfq_queue, pos_node); rb_entry 2434 block/bfq-iosched.c __bfqq = rb_entry(parent, struct bfq_queue, pos_node); rb_entry 2445 block/bfq-iosched.c __bfqq = rb_entry(node, struct bfq_queue, pos_node); rb_entry 27 block/bfq-wf2q.c return rb_entry(node, struct bfq_entity, rb_node); rb_entry 323 block/bfq-wf2q.c entity = rb_entry(node, struct bfq_entity, rb_node); rb_entry 382 block/bfq-wf2q.c entry = rb_entry(parent, struct bfq_entity, rb_node); rb_entry 411 block/bfq-wf2q.c child = rb_entry(node, struct bfq_entity, rb_node); rb_entry 427 block/bfq-wf2q.c struct bfq_entity *entity = rb_entry(node, struct bfq_entity, rb_node); rb_entry 1365 block/bfq-wf2q.c entry = rb_entry(node, struct bfq_entity, rb_node); rb_entry 1371 block/bfq-wf2q.c entry = rb_entry(node->rb_left, rb_entry 98 block/blk-throttle.c #define rb_entry_tg(node) rb_entry((node), struct throtl_grp, rb_node) rb_entry 263 block/elevator.c __rq = rb_entry(parent, struct request, rb_node); rb_entry 290 block/elevator.c rq = rb_entry(n, struct request, rb_node); rb_entry 954 drivers/android/binder.c thread = rb_entry(n, struct binder_thread, rb_node); rb_entry 1073 drivers/android/binder.c node = rb_entry(n, struct binder_node, rb_node); rb_entry 1120 drivers/android/binder.c node = rb_entry(parent, struct binder_node, rb_node); rb_entry 1401 drivers/android/binder.c ref = rb_entry(n, struct binder_ref, rb_node_desc); rb_entry 1448 drivers/android/binder.c ref = rb_entry(parent, struct binder_ref, rb_node_node); rb_entry 1469 drivers/android/binder.c ref = rb_entry(n, struct binder_ref, rb_node_desc); rb_entry 1478 drivers/android/binder.c ref = rb_entry(parent, struct binder_ref, rb_node_desc); rb_entry 4639 drivers/android/binder.c thread = rb_entry(parent, struct binder_thread, rb_node); rb_entry 4984 drivers/android/binder.c struct binder_node *node = rb_entry(n, struct binder_node, rb_entry 5312 drivers/android/binder.c struct binder_thread *thread = rb_entry(n, struct binder_thread, rb_node); rb_entry 5447 drivers/android/binder.c thread = rb_entry(n, struct binder_thread, rb_node); rb_entry 5459 drivers/android/binder.c node = rb_entry(n, struct binder_node, rb_node); rb_entry 5479 drivers/android/binder.c ref = rb_entry(n, struct binder_ref, rb_node_desc); rb_entry 5721 drivers/android/binder.c print_binder_thread_ilocked(m, rb_entry(n, struct binder_thread, rb_entry 5725 drivers/android/binder.c struct binder_node *node = rb_entry(n, struct binder_node, rb_entry 5755 drivers/android/binder.c print_binder_ref_olocked(m, rb_entry(n, rb_entry 5907 drivers/android/binder.c struct binder_ref *ref = rb_entry(n, struct binder_ref, rb_entry 87 drivers/android/binder_alloc.c buffer = rb_entry(parent, struct binder_buffer, rb_node); rb_entry 112 drivers/android/binder_alloc.c buffer = rb_entry(parent, struct binder_buffer, rb_node); rb_entry 137 drivers/android/binder_alloc.c buffer = rb_entry(n, struct binder_buffer, rb_node); rb_entry 393 drivers/android/binder_alloc.c buffer = rb_entry(n, struct binder_buffer, rb_node); rb_entry 417 drivers/android/binder_alloc.c buffer = rb_entry(n, struct binder_buffer, rb_node); rb_entry 426 drivers/android/binder_alloc.c buffer = rb_entry(n, struct binder_buffer, rb_node); rb_entry 444 drivers/android/binder_alloc.c buffer = rb_entry(best_fit, struct binder_buffer, rb_node); rb_entry 751 drivers/android/binder_alloc.c buffer = rb_entry(n, struct binder_buffer, rb_node); rb_entry 828 drivers/android/binder_alloc.c rb_entry(n, struct binder_buffer, rb_node)); rb_entry 80 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rb_entry 107 drivers/base/regmap/regcache-rbtree.c rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node); rb_entry 150 drivers/base/regmap/regcache-rbtree.c n = rb_entry(node, struct regcache_rbtree_node, node); rb_entry 227 drivers/base/regmap/regcache-rbtree.c rbtree_node = rb_entry(next, struct regcache_rbtree_node, node); rb_entry 405 drivers/base/regmap/regcache-rbtree.c rbnode_tmp = rb_entry(node, struct regcache_rbtree_node, rb_entry 477 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rb_entry 517 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rb_entry 625 drivers/base/regmap/regmap-debugfs.c range_node = rb_entry(next, struct regmap_range_node, node); rb_entry 539 drivers/base/regmap/regmap.c rb_entry(*new, struct regmap_range_node, node); rb_entry 563 drivers/base/regmap/regmap.c rb_entry(node, struct regmap_range_node, node); rb_entry 583 drivers/base/regmap/regmap.c range_node = rb_entry(next, struct regmap_range_node, node); rb_entry 12 drivers/block/drbd/drbd_interval.c struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb); rb_entry 34 drivers/block/drbd/drbd_interval.c rb_entry(*new, struct drbd_interval, rb); rb_entry 75 drivers/block/drbd/drbd_interval.c rb_entry(node, struct drbd_interval, rb); rb_entry 122 drivers/block/drbd/drbd_interval.c rb_entry(node, struct drbd_interval, rb); rb_entry 151 drivers/block/drbd/drbd_interval.c i = rb_entry(node, struct drbd_interval, rb); rb_entry 628 drivers/block/pktcdvd.c return rb_entry(n, struct pkt_rb_node, rb_node); rb_entry 654 drivers/block/pktcdvd.c tmp = rb_entry(n, struct pkt_rb_node, rb_node); rb_entry 685 drivers/block/pktcdvd.c tmp = rb_entry(parent, struct pkt_rb_node, rb_node); rb_entry 1186 drivers/block/pktcdvd.c first_node = rb_entry(n, struct pkt_rb_node, rb_node); rb_entry 1204 drivers/block/pktcdvd.c node = rb_entry(n, struct pkt_rb_node, rb_node); rb_entry 257 drivers/dma-buf/sw_sync.c other = rb_entry(parent, typeof(*pt), node); rb_entry 278 drivers/dma-buf/sw_sync.c parent ? &rb_entry(parent, typeof(*pt), node)->link : &obj->pt_list); rb_entry 180 drivers/gpu/drm/drm_mm.c parent = rb_entry(rb, struct drm_mm_node, rb); rb_entry 199 drivers/gpu/drm/drm_mm.c parent = rb_entry(rb, struct drm_mm_node, rb); rb_entry 220 drivers/gpu/drm/drm_mm.c if (x < expr(rb_entry(rb, struct drm_mm_node, member))) \ rb_entry 234 drivers/gpu/drm/drm_mm.c return rb_entry(rb, struct drm_mm_node, rb_hole_size)->hole_size; rb_entry 296 drivers/gpu/drm/drm_mm.c return rb_entry(rb, struct drm_mm_node, rb_hole_size)->hole_size; rb_entry 306 drivers/gpu/drm/drm_mm.c rb_entry(rb, struct drm_mm_node, rb_hole_size); rb_entry 118 drivers/gpu/drm/drm_prime.c pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); rb_entry 133 drivers/gpu/drm/drm_prime.c pos = rb_entry(rb, struct drm_prime_member, handle_rb); rb_entry 154 drivers/gpu/drm/drm_prime.c member = rb_entry(rb, struct drm_prime_member, handle_rb); rb_entry 176 drivers/gpu/drm/drm_prime.c member = rb_entry(rb, struct drm_prime_member, dmabuf_rb); rb_entry 199 drivers/gpu/drm/drm_prime.c member = rb_entry(rb, struct drm_prime_member, dmabuf_rb); rb_entry 152 drivers/gpu/drm/drm_vma_manager.c node = rb_entry(iter, struct drm_mm_node, rb); rb_entry 282 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); rb_entry 335 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); rb_entry 376 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); rb_entry 22 drivers/gpu/drm/i915/gt/intel_engine_user.c rb_entry(p, typeof(*it), uabi_node); rb_entry 266 drivers/gpu/drm/i915/gt/intel_lrc.c return rb_entry(rb, struct i915_priolist, node); rb_entry 356 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); rb_entry 1071 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); rb_entry 1169 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); rb_entry 2615 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); rb_entry 3603 drivers/gpu/drm/i915/gt/intel_lrc.c other = rb_entry(rb, typeof(*other), rb); rb_entry 3952 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_priolist *p = rb_entry(rb, typeof(*p), node); rb_entry 3975 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); rb_entry 77 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c return rb_entry(rb, struct i915_priolist, node); rb_entry 226 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, dma_addr_node); rb_entry 244 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, gfn_node); rb_entry 276 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(parent, struct gvt_dma, gfn_node); rb_entry 291 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(parent, struct gvt_dma, dma_addr_node); rb_entry 326 drivers/gpu/drm/i915/gvt/kvmgt.c dma = rb_entry(node, struct gvt_dma, gfn_node); rb_entry 212 drivers/gpu/drm/i915/i915_active.c node = rb_entry(parent, struct active_node, node); rb_entry 526 drivers/gpu/drm/i915/i915_active.c rb_entry(p, struct active_node, node); rb_entry 546 drivers/gpu/drm/i915/i915_active.c rb_entry(p, struct active_node, node); rb_entry 582 drivers/gpu/drm/i915/i915_active.c return rb_entry(p, struct active_node, node); rb_entry 682 drivers/gpu/drm/i915/i915_active.c it = rb_entry(parent, struct active_node, node); rb_entry 271 drivers/gpu/drm/i915/i915_debugfs.c vma = rb_entry(p, typeof(*vma), obj_node); rb_entry 40 drivers/gpu/drm/i915/i915_scheduler.c return rb_entry(rb, struct i915_priolist, node); rb_entry 186 drivers/gpu/drm/i915/i915_vma.c pos = rb_entry(rb, struct i915_vma, obj_node); rb_entry 241 drivers/gpu/drm/i915/i915_vma.c struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); rb_entry 37 drivers/gpu/drm/nouveau/nvkm/core/object.c object = rb_entry(node, typeof(*object), node); rb_entry 71 drivers/gpu/drm/nouveau/nvkm/core/object.c struct nvkm_object *this = rb_entry(*ptr, typeof(*this), node); rb_entry 805 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree); rb_entry 847 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree); rb_entry 867 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); rb_entry 987 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); rb_entry 1689 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *this = rb_entry(temp, typeof(*this), tree); rb_entry 1705 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *this = rb_entry(node, typeof(*this), tree); rb_entry 671 drivers/infiniband/core/cm.c cur_cm_id_priv = rb_entry(parent, struct cm_id_private, rb_entry 701 drivers/infiniband/core/cm.c cm_id_priv = rb_entry(node, struct cm_id_private, service_node); rb_entry 732 drivers/infiniband/core/cm.c cur_timewait_info = rb_entry(parent, struct cm_timewait_info, rb_entry 758 drivers/infiniband/core/cm.c timewait_info = rb_entry(node, struct cm_timewait_info, rb_entry 785 drivers/infiniband/core/cm.c cur_timewait_info = rb_entry(parent, struct cm_timewait_info, rb_entry 815 drivers/infiniband/core/cm.c cur_cm_id_priv = rb_entry(parent, struct cm_id_private, rb_entry 137 drivers/infiniband/core/multicast.c group = rb_entry(node, struct mcast_group, node); rb_entry 161 drivers/infiniband/core/multicast.c cur_group = rb_entry(parent, struct mcast_group, node); rb_entry 779 drivers/infiniband/core/multicast.c group = rb_entry(node, struct mcast_group, node); rb_entry 91 drivers/infiniband/core/umem_odp.c rb_entry(node, struct ib_umem_odp, interval_tree.rb); rb_entry 494 drivers/infiniband/core/uverbs_cmd.c scan = rb_entry(parent, struct xrcd_table_entry, node); rb_entry 519 drivers/infiniband/core/uverbs_cmd.c entry = rb_entry(p, struct xrcd_table_entry, node); rb_entry 149 drivers/infiniband/hw/hfi1/mmu_rb.c rbnode = rb_entry(node, struct mmu_rb_node, node); rb_entry 151 drivers/infiniband/hw/mlx4/cm.c rb_entry(node, struct id_map_entry, node); rb_entry 209 drivers/infiniband/hw/mlx4/cm.c ent = rb_entry(parent, struct id_map_entry, node); rb_entry 410 drivers/infiniband/hw/mlx4/cm.c rb_entry(rb_first(sl_id_map), rb_entry 422 drivers/infiniband/hw/mlx4/cm.c rb_entry(nd, struct id_map_entry, node); rb_entry 170 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(node, struct mcast_group, node); rb_entry 193 drivers/infiniband/hw/mlx4/mcg.c cur_group = rb_entry(parent, struct mcast_group, node); rb_entry 1103 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(p, struct mcast_group, node); rb_entry 1232 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(p, struct mcast_group, node); rb_entry 123 drivers/infiniband/hw/mlx5/odp.c odp = rb_entry(rb, struct ib_umem_odp, interval_tree.rb); rb_entry 152 drivers/infiniband/hw/mlx5/odp.c odp = rb_entry(rb, struct ib_umem_odp, interval_tree.rb); rb_entry 584 drivers/infiniband/hw/mlx5/odp.c rb_entry(node, struct ib_umem_odp, interval_tree.rb); rb_entry 147 drivers/infiniband/hw/qib/qib_user_sdma.c sdma_rb_node = rb_entry(node, struct qib_user_sdma_rb_node, rb_entry 167 drivers/infiniband/hw/qib/qib_user_sdma.c got = rb_entry(*node, struct qib_user_sdma_rb_node, node); rb_entry 161 drivers/infiniband/sw/rdmavt/mcast.c mcast = rb_entry(n, struct rvt_mcast, rb_node); rb_entry 207 drivers/infiniband/sw/rdmavt/mcast.c tmcast = rb_entry(pn, struct rvt_mcast, rb_node); rb_entry 367 drivers/infiniband/sw/rdmavt/mcast.c mcast = rb_entry(n, struct rvt_mcast, rb_node); rb_entry 291 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(parent, struct rxe_pool_entry, node); rb_entry 319 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(parent, struct rxe_pool_entry, node); rb_entry 491 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(node, struct rxe_pool_entry, node); rb_entry 523 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(node, struct rxe_pool_entry, node); rb_entry 563 drivers/infiniband/ulp/ipoib/ipoib_main.c path = rb_entry(n, struct ipoib_path, rb_node); rb_entry 589 drivers/infiniband/ulp/ipoib/ipoib_main.c tpath = rb_entry(pn, struct ipoib_path, rb_node); rb_entry 660 drivers/infiniband/ulp/ipoib/ipoib_main.c path = rb_entry(n, struct ipoib_path, rb_node); rb_entry 168 drivers/infiniband/ulp/ipoib/ipoib_multicast.c mcast = rb_entry(n, struct ipoib_mcast, rb_node); rb_entry 193 drivers/infiniband/ulp/ipoib/ipoib_multicast.c tmcast = rb_entry(pn, struct ipoib_mcast, rb_node); rb_entry 1028 drivers/infiniband/ulp/ipoib/ipoib_multicast.c mcast = rb_entry(n, struct ipoib_mcast, rb_node); rb_entry 139 drivers/iommu/iova.c cached_iova = rb_entry(iovad->cached32_node, struct iova, node); rb_entry 147 drivers/iommu/iova.c cached_iova = rb_entry(iovad->cached_node, struct iova, node); rb_entry 162 drivers/iommu/iova.c struct iova *this = rb_entry(*new, struct iova, node); rb_entry 200 drivers/iommu/iova.c curr_iova = rb_entry(curr, struct iova, node); rb_entry 206 drivers/iommu/iova.c curr_iova = rb_entry(curr, struct iova, node); rb_entry 325 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); rb_entry 608 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); rb_entry 676 drivers/iommu/iova.c iova = rb_entry(node, struct iova, node); rb_entry 713 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); rb_entry 2145 drivers/macintosh/via-pmu.c struct rb_entry *rp = &pp->rb_buf[pp->rb_put]; rb_entry 2202 drivers/macintosh/via-pmu.c struct rb_entry *rp = &pp->rb_buf[i]; rb_entry 120 drivers/md/dm-bio-prison-v1.c rb_entry(*new, struct dm_bio_prison_cell, node); rb_entry 124 drivers/md/dm-bio-prison-v2.c rb_entry(*new, struct dm_bio_prison_cell_v2, node); rb_entry 1509 drivers/md/dm-crypt.c #define crypt_io_from_node(node) rb_entry((node), struct dm_crypt_io, rb_node) rb_entry 40 drivers/md/dm-dust.c struct badblock *bblk = rb_entry(node, struct badblock, node); rb_entry 61 drivers/md/dm-dust.c bblk = rb_entry(parent, struct badblock, node); rb_entry 1752 drivers/md/dm-snap.c pe = rb_entry(next, struct dm_snap_pending_exception, rb_entry 1768 drivers/md/dm-snap.c pe2 = rb_entry(*p, struct dm_snap_pending_exception, out_of_order_node); rb_entry 2108 drivers/md/dm-thin.c #define thin_pbd(node) rb_entry((node), struct dm_thin_endio_hook, rb_node) rb_entry 79 drivers/mtd/mtdswap.c #define MTDSWAP_ECNT_MIN(rbroot) (rb_entry(rb_first(rbroot), struct swap_eb, \ rb_entry 81 drivers/mtd/mtdswap.c #define MTDSWAP_ECNT_MAX(rbroot) (rb_entry(rb_last(rbroot), struct swap_eb, \ rb_entry 211 drivers/mtd/mtdswap.c cur = rb_entry(parent, struct swap_eb, rb); rb_entry 438 drivers/mtd/mtdswap.c median = rb_entry(medrb, struct swap_eb, rb)->erase_count; rb_entry 576 drivers/mtd/mtdswap.c eb = rb_entry(rb_first(clean_root), struct swap_eb, rb); rb_entry 861 drivers/mtd/mtdswap.c eb = rb_entry(rb_first(rp), struct swap_eb, rb); rb_entry 113 drivers/mtd/ubi/attach.c av = rb_entry(parent, struct ubi_ainf_volume, rb); rb_entry 595 drivers/mtd/ubi/attach.c aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); rb_entry 1289 drivers/mtd/ubi/attach.c aeb = rb_entry(this, struct ubi_ainf_peb, u.rb); rb_entry 1346 drivers/mtd/ubi/attach.c av = rb_entry(rb, struct ubi_ainf_volume, rb); rb_entry 220 drivers/mtd/ubi/eba.c le = rb_entry(p, struct ubi_ltree_entry, rb); rb_entry 286 drivers/mtd/ubi/eba.c le1 = rb_entry(parent, struct ubi_ltree_entry, rb); rb_entry 379 drivers/mtd/ubi/fastmap-wl.c e = rb_entry(rb_next(root->rb_node), rb_entry 213 drivers/mtd/ubi/fastmap.c tmp_aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); rb_entry 252 drivers/mtd/ubi/fastmap.c aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); rb_entry 148 drivers/mtd/ubi/wl.c e1 = rb_entry(parent, struct ubi_wl_entry, u.rb); rb_entry 244 drivers/mtd/ubi/wl.c e1 = rb_entry(p, struct ubi_wl_entry, u.rb); rb_entry 325 drivers/mtd/ubi/wl.c e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); rb_entry 332 drivers/mtd/ubi/wl.c e1 = rb_entry(p, struct ubi_wl_entry, u.rb); rb_entry 366 drivers/mtd/ubi/wl.c first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); rb_entry 367 drivers/mtd/ubi/wl.c last = rb_entry(rb_last(root), struct ubi_wl_entry, u.rb); rb_entry 370 drivers/mtd/ubi/wl.c e = rb_entry(root->rb_node, struct ubi_wl_entry, u.rb); rb_entry 731 drivers/mtd/ubi/wl.c e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb); rb_entry 752 drivers/mtd/ubi/wl.c e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, u.rb); rb_entry 1036 drivers/mtd/ubi/wl.c e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb); rb_entry 1597 drivers/mtd/ubi/wl.c e = rb_entry(rb, struct ubi_wl_entry, u.rb); rb_entry 240 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_common *res = rb_entry(node, struct res_common, rb_entry 259 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_common *this = rb_entry(*new, struct res_common, rb_entry 5036 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c fs_rule = rb_entry(p, struct res_fs_rule, com.node); rb_entry 87 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c tfp = rb_entry(parent, struct fw_page, rb_node); rb_entry 122 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c tfp = rb_entry(tmp, struct fw_page, rb_node); rb_entry 360 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c fwp = rb_entry(p, struct fw_page, rb_node); rb_entry 531 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c fwp = rb_entry(p, struct fw_page, rb_node); rb_entry 140 drivers/vfio/vfio_iommu_type1.c struct vfio_dma *dma = rb_entry(node, struct vfio_dma, node); rb_entry 160 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(parent, struct vfio_dma, node); rb_entry 186 drivers/vfio/vfio_iommu_type1.c vpfn = rb_entry(node, struct vfio_pfn, node); rb_entry 207 drivers/vfio/vfio_iommu_type1.c vpfn = rb_entry(parent, struct vfio_pfn, node); rb_entry 1203 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); rb_entry 1859 drivers/vfio/vfio_iommu_type1.c vfio_remove_dma(iommu, rb_entry(node, struct vfio_dma, node)); rb_entry 1871 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); rb_entry 1875 drivers/vfio/vfio_iommu_type1.c struct vfio_pfn *vpfn = rb_entry(p, struct vfio_pfn, rb_entry 1893 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); rb_entry 116 drivers/xen/evtchn.c this = rb_entry(*new, struct user_evtchn, node); rb_entry 148 drivers/xen/evtchn.c evtchn = rb_entry(node, struct user_evtchn, node); rb_entry 678 drivers/xen/evtchn.c evtchn = rb_entry(node, struct user_evtchn, node); rb_entry 86 fs/afs/cell.c cell = rb_entry(p, struct afs_cell, net_node); rb_entry 265 fs/afs/cell.c cursor = rb_entry(parent, struct afs_cell, net_node); rb_entry 771 fs/afs/cell.c rb_entry(cursor, struct afs_cell, net_node); rb_entry 118 fs/afs/server.c server = rb_entry(p, struct afs_server, uuid_rb); rb_entry 161 fs/afs/server.c server = rb_entry(p, struct afs_server, uuid_rb); rb_entry 455 fs/afs/server.c rb_entry(cursor, struct afs_server, uuid_rb); rb_entry 240 fs/btrfs/backref.c ref = rb_entry(parent, struct prelim_ref, rbnode); rb_entry 635 fs/btrfs/backref.c ref = rb_entry(rnode, struct prelim_ref, rbnode); rb_entry 723 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); rb_entry 777 fs/btrfs/backref.c node = rb_entry(n, struct btrfs_delayed_ref_node, rb_entry 1254 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); rb_entry 163 fs/btrfs/block-group.c cache = rb_entry(parent, struct btrfs_block_group_cache, rb_entry 202 fs/btrfs/block-group.c cache = rb_entry(n, struct btrfs_block_group_cache, rb_entry 269 fs/btrfs/block-group.c cache = rb_entry(node, struct btrfs_block_group_cache, rb_entry 3121 fs/btrfs/block-group.c block_group = rb_entry(n, struct btrfs_block_group_cache, rb_entry 381 fs/btrfs/ctree.c tm = rb_entry(node, struct tree_mod_elem, node); rb_entry 413 fs/btrfs/ctree.c cur = rb_entry(*new, struct tree_mod_elem, node); rb_entry 693 fs/btrfs/ctree.c cur = rb_entry(node, struct tree_mod_elem, node); rb_entry 1282 fs/btrfs/ctree.c tm = rb_entry(next, struct tree_mod_elem, node); rb_entry 353 fs/btrfs/delayed-inode.c delayed_item = rb_entry(node, struct btrfs_delayed_item, rb_entry 371 fs/btrfs/delayed-inode.c *prev = rb_entry(node, struct btrfs_delayed_item, rb_entry 383 fs/btrfs/delayed-inode.c *next = rb_entry(node, struct btrfs_delayed_item, rb_entry 421 fs/btrfs/delayed-inode.c item = rb_entry(parent_node, struct btrfs_delayed_item, rb_entry 516 fs/btrfs/delayed-inode.c item = rb_entry(p, struct btrfs_delayed_item, rb_node); rb_entry 529 fs/btrfs/delayed-inode.c item = rb_entry(p, struct btrfs_delayed_item, rb_node); rb_entry 542 fs/btrfs/delayed-inode.c next = rb_entry(p, struct btrfs_delayed_item, rb_node); rb_entry 288 fs/btrfs/delayed-ref.c ins = rb_entry(node, struct btrfs_delayed_ref_head, href_node); rb_entry 292 fs/btrfs/delayed-ref.c entry = rb_entry(parent_node, struct btrfs_delayed_ref_head, rb_entry 323 fs/btrfs/delayed-ref.c entry = rb_entry(parent_node, struct btrfs_delayed_ref_node, rb_entry 351 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node); rb_entry 372 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node); rb_entry 386 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, rb_entry 443 fs/btrfs/delayed-ref.c next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); rb_entry 508 fs/btrfs/delayed-ref.c ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); rb_entry 563 fs/btrfs/delayed-ref.c head = rb_entry(node, struct btrfs_delayed_ref_head, rb_entry 4287 fs/btrfs/disk-io.c head = rb_entry(node, struct btrfs_delayed_ref_head, rb_entry 4294 fs/btrfs/disk-io.c ref = rb_entry(n, struct btrfs_delayed_ref_node, rb_entry 1725 fs/btrfs/extent-tree.c ref = rb_entry(rb_first_cached(&head->ref_tree), rb_entry 2089 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); rb_entry 2094 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); rb_entry 2100 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); rb_entry 2201 fs/btrfs/extent-tree.c head = rb_entry(node, struct btrfs_delayed_ref_head, rb_entry 2295 fs/btrfs/extent-tree.c ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); rb_entry 277 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 346 fs/btrfs/extent_io.c entry = rb_entry(parent, struct tree_entry, rb_node); rb_entry 395 fs/btrfs/extent_io.c entry = rb_entry(prev, struct tree_entry, rb_node); rb_entry 415 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); rb_entry 422 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); rb_entry 425 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); rb_entry 473 fs/btrfs/extent_io.c other = rb_entry(other_node, struct extent_state, rb_node); rb_entry 488 fs/btrfs/extent_io.c other = rb_entry(other_node, struct extent_state, rb_node); rb_entry 538 fs/btrfs/extent_io.c found = rb_entry(node, struct extent_state, rb_node); rb_entry 588 fs/btrfs/extent_io.c return rb_entry(next, struct extent_state, rb_node); rb_entry 730 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 861 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 992 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1231 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1505 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1599 fs/btrfs/extent_io.c state = rb_entry(prev, struct extent_state, rb_node); rb_entry 1610 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1646 fs/btrfs/extent_io.c state = rb_entry(prev, struct extent_state, rb_entry 1661 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1706 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 1990 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 2038 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 2066 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 2097 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 105 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); rb_entry 120 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); rb_entry 127 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); rb_entry 130 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); rb_entry 156 fs/btrfs/extent_map.c entry = rb_entry(n, struct extent_map, rb_node); rb_entry 172 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); rb_entry 179 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); rb_entry 182 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); rb_entry 250 fs/btrfs/extent_map.c merge = rb_entry(rb, struct extent_map, rb_node); rb_entry 269 fs/btrfs/extent_map.c merge = rb_entry(rb, struct extent_map, rb_node); rb_entry 433 fs/btrfs/extent_map.c em = rb_entry(rb_node, struct extent_map, rb_node); rb_entry 93 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); rb_entry 226 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); rb_entry 240 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); rb_entry 260 fs/btrfs/file.c defrag = rb_entry(node, struct inode_defrag, rb_node); rb_entry 645 fs/btrfs/free-space-cache.c e = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 950 fs/btrfs/free-space-cache.c e = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 1458 fs/btrfs/free-space-cache.c info = rb_entry(parent, struct btrfs_free_space, offset_index); rb_entry 1521 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 1545 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 1559 fs/btrfs/free-space-cache.c prev = rb_entry(n, struct btrfs_free_space, rb_entry 1577 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, rb_entry 1591 fs/btrfs/free-space-cache.c prev = rb_entry(n, struct btrfs_free_space, rb_entry 1618 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 1825 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 1944 fs/btrfs/free-space-cache.c bitmap_info = rb_entry(next, struct btrfs_free_space, rb_entry 2086 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2182 fs/btrfs/free-space-cache.c left_info = rb_entry(rb_prev(&right_info->offset_index), rb_entry 2498 fs/btrfs/free-space-cache.c info = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 2560 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2588 fs/btrfs/free-space-cache.c info = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2770 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2781 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, rb_entry 2795 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, rb_entry 2946 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2956 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 2987 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); rb_entry 3233 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, rb_entry 3444 fs/btrfs/free-space-cache.c entry = rb_entry(rb_first(&ctl->free_space_offset), rb_entry 3705 fs/btrfs/free-space-cache.c tmp = rb_entry(n, struct btrfs_free_space, rb_entry 3719 fs/btrfs/free-space-cache.c tmp = rb_entry(n, struct btrfs_free_space, rb_entry 1162 fs/btrfs/free-space-tree.c block_group = rb_entry(node, struct btrfs_block_group_cache, rb_entry 271 fs/btrfs/inode-map.c info = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 308 fs/btrfs/inode-map.c info = rb_entry(n, struct btrfs_free_space, offset_index); rb_entry 2523 fs/btrfs/inode.c entry = rb_entry(parent, struct sa_defrag_extent_backref, node); rb_entry 2951 fs/btrfs/inode.c backref = rb_entry(node, struct sa_defrag_extent_backref, node); rb_entry 4460 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); rb_entry 4471 fs/btrfs/inode.c entry = rb_entry(prev, struct btrfs_inode, rb_node); rb_entry 4480 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); rb_entry 5438 fs/btrfs/inode.c em = rb_entry(node, struct extent_map, rb_node); rb_entry 5476 fs/btrfs/inode.c state = rb_entry(node, struct extent_state, rb_node); rb_entry 5783 fs/btrfs/inode.c entry = rb_entry(parent, struct btrfs_inode, rb_node); rb_entry 10753 fs/btrfs/inode.c entry = rb_entry(parent, struct btrfs_swapfile_pin, node); rb_entry 10783 fs/btrfs/inode.c sp = rb_entry(node, struct btrfs_swapfile_pin, node); rb_entry 40 fs/btrfs/ordered-data.c entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); rb_entry 77 fs/btrfs/ordered-data.c entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); rb_entry 95 fs/btrfs/ordered-data.c prev_entry = rb_entry(test, struct btrfs_ordered_extent, rb_entry 103 fs/btrfs/ordered-data.c prev_entry = rb_entry(prev, struct btrfs_ordered_extent, rb_entry 109 fs/btrfs/ordered-data.c prev_entry = rb_entry(test, struct btrfs_ordered_extent, rb_entry 150 fs/btrfs/ordered-data.c entry = rb_entry(tree->last, struct btrfs_ordered_extent, rb_entry 324 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 396 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 725 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 755 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 792 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 874 fs/btrfs/ordered-data.c test = rb_entry(prev, struct btrfs_ordered_extent, rb_entry 881 fs/btrfs/ordered-data.c test = rb_entry(node, struct btrfs_ordered_extent, rb_node); rb_entry 177 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); rb_entry 198 fs/btrfs/qgroup.c qgroup = rb_entry(parent, struct btrfs_qgroup, node); rb_entry 520 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); rb_entry 1558 fs/btrfs/qgroup.c entry = rb_entry(parent_node, struct btrfs_qgroup_extent_record, rb_entry 2521 fs/btrfs/qgroup.c record = rb_entry(node, struct btrfs_qgroup_extent_record, rb_entry 3300 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); rb_entry 3902 fs/btrfs/qgroup.c entry = rb_entry(parent, struct btrfs_qgroup_swapped_block, rb_entry 3975 fs/btrfs/qgroup.c block = rb_entry(node, struct btrfs_qgroup_swapped_block, node); rb_entry 84 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct block_entry, node); rb_entry 105 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); rb_entry 125 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct root_entry, node); rb_entry 171 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct ref_entry, node); rb_entry 194 fs/btrfs/ref-verify.c entry = rb_entry(n, struct root_entry, node); rb_entry 240 fs/btrfs/ref-verify.c re = rb_entry(n, struct root_entry, node); rb_entry 246 fs/btrfs/ref-verify.c ref = rb_entry(n, struct ref_entry, node); rb_entry 644 fs/btrfs/ref-verify.c ref = rb_entry(n, struct ref_entry, node); rb_entry 652 fs/btrfs/ref-verify.c re = rb_entry(n, struct root_entry, node); rb_entry 904 fs/btrfs/ref-verify.c be = rb_entry(n, struct block_entry, node); rb_entry 924 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); rb_entry 951 fs/btrfs/ref-verify.c be = rb_entry(n, struct block_entry, node); rb_entry 291 fs/btrfs/relocation.c entry = rb_entry(parent, struct tree_entry, rb_node); rb_entry 312 fs/btrfs/relocation.c entry = rb_entry(n, struct tree_entry, rb_node); rb_entry 328 fs/btrfs/relocation.c struct backref_node *bnode = rb_entry(rb_node, struct backref_node, rb_entry 588 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); rb_entry 864 fs/btrfs/relocation.c upper = rb_entry(rb_node, struct backref_node, rb_entry 998 fs/btrfs/relocation.c upper = rb_entry(rb_node, struct backref_node, rb_entry 1219 fs/btrfs/relocation.c node = rb_entry(rb_node, struct backref_node, rb_node); rb_entry 1230 fs/btrfs/relocation.c node = rb_entry(rb_node, struct backref_node, rb_entry 1334 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); rb_entry 1365 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); rb_entry 1564 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); rb_entry 1575 fs/btrfs/relocation.c entry = rb_entry(prev, struct btrfs_inode, rb_node); rb_entry 1584 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); rb_entry 2589 fs/btrfs/relocation.c block = rb_entry(rb_node, struct tree_block, rb_node); rb_entry 333 fs/btrfs/scrub.c entry = rb_entry(parent, struct full_stripe_lock, node); rb_entry 376 fs/btrfs/scrub.c entry = rb_entry(node, struct full_stripe_lock, node); rb_entry 2867 fs/btrfs/send.c entry = rb_entry(parent, struct orphan_dir_info, node); rb_entry 2896 fs/btrfs/send.c entry = rb_entry(n, struct orphan_dir_info, node); rb_entry 3042 fs/btrfs/send.c entry = rb_entry(parent, struct waiting_dir_move, node); rb_entry 3065 fs/btrfs/send.c entry = rb_entry(n, struct waiting_dir_move, node); rb_entry 3112 fs/btrfs/send.c entry = rb_entry(parent, struct pending_dir_move, node); rb_entry 3160 fs/btrfs/send.c entry = rb_entry(n, struct pending_dir_move, node); rb_entry 7313 fs/btrfs/send.c pm = rb_entry(n, struct pending_dir_move, node); rb_entry 7330 fs/btrfs/send.c dm = rb_entry(n, struct waiting_dir_move, node); rb_entry 7341 fs/btrfs/send.c odi = rb_entry(n, struct orphan_dir_info, node); rb_entry 17 fs/btrfs/tests/extent-map-tests.c em = rb_entry(node, struct extent_map, rb_node); rb_entry 124 fs/btrfs/ulist.c u = rb_entry(n, struct ulist_node, rb_node); rb_entry 152 fs/btrfs/ulist.c cur = rb_entry(parent, struct ulist_node, rb_node); rb_entry 1873 fs/btrfs/volumes.c em = rb_entry(n, struct extent_map, rb_node); rb_entry 7801 fs/btrfs/volumes.c em = rb_entry(node, struct extent_map, rb_node); rb_entry 7919 fs/btrfs/volumes.c sp = rb_entry(node, struct btrfs_swapfile_pin, node); rb_entry 97 fs/cachefiles/namei.c object = rb_entry(p, struct cachefiles_object, active_node); rb_entry 160 fs/cachefiles/namei.c xobject = rb_entry(_parent, rb_entry 904 fs/cachefiles/namei.c object = rb_entry(_n, struct cachefiles_object, active_node); rb_entry 1829 fs/ceph/addr.c perm = rb_entry(*p, struct ceph_pool_perm, node); rb_entry 1863 fs/ceph/addr.c perm = rb_entry(parent, struct ceph_pool_perm, node); rb_entry 2060 fs/ceph/addr.c perm = rb_entry(n, struct ceph_pool_perm, node); rb_entry 439 fs/ceph/caps.c cap = rb_entry(n, struct ceph_cap, ci_node); rb_entry 472 fs/ceph/caps.c cap = rb_entry(parent, struct ceph_cap, ci_node); rb_entry 799 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 830 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 880 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 907 fs/ceph/caps.c cap = rb_entry(q, struct ceph_cap, rb_entry 931 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 995 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 1249 fs/ceph/caps.c struct ceph_cap *cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 1924 fs/ceph/caps.c cap = rb_entry(p, struct ceph_cap, ci_node); rb_entry 61 fs/ceph/debugfs.c req = rb_entry(rp, struct ceph_mds_request, r_node); rb_entry 126 fs/ceph/inode.c frag = rb_entry(parent, struct ceph_inode_frag, node); rb_entry 162 fs/ceph/inode.c rb_entry(n, struct ceph_inode_frag, node); rb_entry 341 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); rb_entry 371 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); rb_entry 402 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); rb_entry 574 fs/ceph/inode.c frag = rb_entry(n, struct ceph_inode_frag, node); rb_entry 813 fs/ceph/mds_client.c rb_entry(p, struct ceph_mds_request, r_node); rb_entry 1003 fs/ceph/mds_client.c cap = rb_entry(rb_first(&ci->i_caps), struct ceph_cap, ci_node); rb_entry 1299 fs/ceph/mds_client.c req = rb_entry(p, struct ceph_mds_request, r_node); rb_entry 2110 fs/ceph/mds_client.c return rb_entry(rb_first(&mdsc->request_tree), rb_entry 2651 fs/ceph/mds_client.c req = rb_entry(p, struct ceph_mds_request, r_node); rb_entry 3233 fs/ceph/mds_client.c req = rb_entry(p, struct ceph_mds_request, r_node); rb_entry 3522 fs/ceph/mds_client.c rb_entry(p, struct ceph_snap_realm, node); rb_entry 4267 fs/ceph/mds_client.c nextreq = rb_entry(n, struct ceph_mds_request, r_node); rb_entry 186 fs/ceph/quota.c qri = rb_entry(node, struct ceph_quotarealm_inode, node); rb_entry 92 fs/ceph/snap.c r = rb_entry(parent, struct ceph_snap_realm, node); rb_entry 147 fs/ceph/snap.c r = rb_entry(n, struct ceph_snap_realm, node); rb_entry 1021 fs/ceph/snap.c exist = rb_entry(*p, struct ceph_snapid_map, node); rb_entry 1059 fs/ceph/snap.c exist = rb_entry(*p, struct ceph_snapid_map, node); rb_entry 1143 fs/ceph/snap.c sm = rb_entry(p, struct ceph_snapid_map, node); rb_entry 448 fs/ceph/xattr.c xattr = rb_entry(parent, struct ceph_inode_xattr, node); rb_entry 544 fs/ceph/xattr.c xattr = rb_entry(parent, struct ceph_inode_xattr, node); rb_entry 607 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); rb_entry 631 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); rb_entry 780 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); rb_entry 5174 fs/cifs/connect.c tlink = rb_entry(node, struct tcon_link, tl_rbnode); rb_entry 5341 fs/cifs/connect.c tlink = rb_entry(node, struct tcon_link, tl_rbnode); rb_entry 5361 fs/cifs/connect.c tlink = rb_entry(*new, struct tcon_link, tl_rbnode); rb_entry 5489 fs/cifs/connect.c tlink = rb_entry(tmp, struct tcon_link, tl_rbnode); rb_entry 455 fs/dlm/debug_fs.c r = rb_entry(node, struct dlm_rsb, res_hashnode); rb_entry 487 fs/dlm/debug_fs.c r = rb_entry(node, struct dlm_rsb, res_hashnode); rb_entry 521 fs/dlm/debug_fs.c r = rb_entry(next, struct dlm_rsb, res_hashnode); rb_entry 552 fs/dlm/debug_fs.c r = rb_entry(next, struct dlm_rsb, res_hashnode); rb_entry 453 fs/dlm/lock.c r = rb_entry(node, struct dlm_rsb, res_hashnode); rb_entry 477 fs/dlm/lock.c struct dlm_rsb *cur = rb_entry(*newn, struct dlm_rsb, rb_entry 1092 fs/dlm/lock.c r = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 1669 fs/dlm/lock.c r = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 5478 fs/dlm/lock.c r = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 823 fs/dlm/lockspace.c rsb = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 829 fs/dlm/lockspace.c rsb = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 909 fs/dlm/recover.c r = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 946 fs/dlm/recover.c r = rb_entry(n, struct dlm_rsb, res_hashnode); rb_entry 835 fs/eventpoll.c epi = rb_entry(rbp, struct epitem, rbn); rb_entry 851 fs/eventpoll.c epi = rb_entry(rbp, struct epitem, rbn); rb_entry 954 fs/eventpoll.c struct epitem *epi = rb_entry(rbp, struct epitem, rbn); rb_entry 1057 fs/eventpoll.c epi = rb_entry(rbp, struct epitem, rbn); rb_entry 1079 fs/eventpoll.c epi = rb_entry(rbp, struct epitem, rbn); rb_entry 1347 fs/eventpoll.c epic = rb_entry(parent, struct epitem, rbn); rb_entry 1975 fs/eventpoll.c epi = rb_entry(rbp, struct epitem, rbn); rb_entry 215 fs/ext2/balloc.c rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node); rb_entry 302 fs/ext2/balloc.c rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node); rb_entry 319 fs/ext2/balloc.c rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node); rb_entry 345 fs/ext2/balloc.c this = rb_entry(parent, struct ext2_reserve_window_node, rsv_node); rb_entry 813 fs/ext2/balloc.c rsv = rb_entry(next,struct ext2_reserve_window_node,rsv_node); rb_entry 1055 fs/ext2/balloc.c next_rsv = rb_entry(next, struct ext2_reserve_window_node, rsv_node); rb_entry 77 fs/ext4/block_validity.c entry = rb_entry(parent, struct ext4_system_zone, node); rb_entry 88 fs/ext4/block_validity.c new_entry = rb_entry(new_node, struct ext4_system_zone, rb_entry 110 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); rb_entry 122 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); rb_entry 141 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); rb_entry 175 fs/ext4/block_validity.c entry = rb_entry(n, struct ext4_system_zone, node); rb_entry 472 fs/ext4/dir.c fname = rb_entry(parent, struct fname, rb_hash); rb_entry 595 fs/ext4/dir.c fname = rb_entry(info->curr_node, struct fname, rb_hash); rb_entry 603 fs/ext4/dir.c fname = rb_entry(info->curr_node, struct fname, rb_entry 188 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 217 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 231 fs/ext4/extents_status.c return node ? rb_entry(node, struct extent_status, rb_node) : rb_entry 287 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); rb_entry 539 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); rb_entry 563 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); rb_entry 757 fs/ext4/extents_status.c es = rb_entry(parent, struct extent_status, rb_node); rb_entry 931 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); rb_entry 955 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_entry 1012 fs/ext4/extents_status.c rc->left_es = node ? rb_entry(node, rb_entry 1123 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); rb_entry 1135 fs/ext4/extents_status.c return node ? rb_entry(node, struct pending_reservation, rb_entry 1198 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 1205 fs/ext4/extents_status.c es = node ? rb_entry(node, struct extent_status, rb_entry 1218 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_entry 1263 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_entry 1366 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 1381 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 1708 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 1757 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 1779 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); rb_entry 1829 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); rb_entry 1864 fs/ext4/extents_status.c pr = rb_entry(parent, struct pending_reservation, rb_node); rb_entry 2057 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); rb_entry 3524 fs/ext4/mballoc.c entry = rb_entry(n, struct ext4_free_data, efd_node); rb_entry 4677 fs/ext4/mballoc.c entry = rb_entry(parent, struct ext4_free_data, efd_node); rb_entry 4697 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); rb_entry 4704 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); rb_entry 18 fs/f2fs/extent_cache.c static struct rb_entry *__lookup_rb_tree_fast(struct rb_entry *cached_re, rb_entry 30 fs/f2fs/extent_cache.c static struct rb_entry *__lookup_rb_tree_slow(struct rb_root_cached *root, rb_entry 34 fs/f2fs/extent_cache.c struct rb_entry *re; rb_entry 37 fs/f2fs/extent_cache.c re = rb_entry(node, struct rb_entry, rb_node); rb_entry 49 fs/f2fs/extent_cache.c struct rb_entry *f2fs_lookup_rb_tree(struct rb_root_cached *root, rb_entry 50 fs/f2fs/extent_cache.c struct rb_entry *cached_re, unsigned int ofs) rb_entry 52 fs/f2fs/extent_cache.c struct rb_entry *re; rb_entry 67 fs/f2fs/extent_cache.c struct rb_entry *re; rb_entry 71 fs/f2fs/extent_cache.c re = rb_entry(*parent, struct rb_entry, rb_node); rb_entry 95 fs/f2fs/extent_cache.c struct rb_entry *f2fs_lookup_rb_tree_ret(struct rb_root_cached *root, rb_entry 96 fs/f2fs/extent_cache.c struct rb_entry *cached_re, rb_entry 98 fs/f2fs/extent_cache.c struct rb_entry **prev_entry, rb_entry 99 fs/f2fs/extent_cache.c struct rb_entry **next_entry, rb_entry 106 fs/f2fs/extent_cache.c struct rb_entry *re = cached_re; rb_entry 126 fs/f2fs/extent_cache.c re = rb_entry(*pnode, struct rb_entry, rb_node); rb_entry 142 fs/f2fs/extent_cache.c re = rb_entry(parent, struct rb_entry, rb_node); rb_entry 146 fs/f2fs/extent_cache.c *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry 151 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry 158 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry 163 fs/f2fs/extent_cache.c *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry 173 fs/f2fs/extent_cache.c struct rb_entry *cur_re, *next_re; rb_entry 183 fs/f2fs/extent_cache.c cur_re = rb_entry(cur, struct rb_entry, rb_node); rb_entry 184 fs/f2fs/extent_cache.c next_re = rb_entry(next, struct rb_entry, rb_node); rb_entry 309 fs/f2fs/extent_cache.c en = rb_entry(node, struct extent_node, rb_node); rb_entry 399 fs/f2fs/extent_cache.c (struct rb_entry *)et->cached_en, pgofs); rb_entry 533 fs/f2fs/extent_cache.c (struct rb_entry *)et->cached_en, fofs, rb_entry 534 fs/f2fs/extent_cache.c (struct rb_entry **)&prev_en, rb_entry 535 fs/f2fs/extent_cache.c (struct rb_entry **)&next_en, rb_entry 3531 fs/f2fs/f2fs.h struct rb_entry *f2fs_lookup_rb_tree(struct rb_root_cached *root, rb_entry 3532 fs/f2fs/f2fs.h struct rb_entry *cached_re, unsigned int ofs); rb_entry 3537 fs/f2fs/f2fs.h struct rb_entry *f2fs_lookup_rb_tree_ret(struct rb_root_cached *root, rb_entry 3538 fs/f2fs/f2fs.h struct rb_entry *cached_re, unsigned int ofs, rb_entry 3539 fs/f2fs/f2fs.h struct rb_entry **prev_entry, struct rb_entry **next_entry, rb_entry 1308 fs/f2fs/segment.c (struct rb_entry **)&prev_dc, rb_entry 1309 fs/f2fs/segment.c (struct rb_entry **)&next_dc, rb_entry 1419 fs/f2fs/segment.c (struct rb_entry **)&prev_dc, rb_entry 1420 fs/f2fs/segment.c (struct rb_entry **)&next_dc, rb_entry 2765 fs/f2fs/segment.c (struct rb_entry **)&prev_dc, rb_entry 2766 fs/f2fs/segment.c (struct rb_entry **)&next_dc, rb_entry 53 fs/fscache/object-list.c xobj = rb_entry(parent, struct fscache_object, objlist_link); rb_entry 108 fs/fscache/object-list.c obj = rb_entry(p, struct fscache_object, objlist_link); rb_entry 2063 fs/fuse/dev.c ff = rb_entry(p, struct fuse_file, polled_node); rb_entry 2938 fs/fuse/file.c ff = rb_entry(last, struct fuse_file, polled_node); rb_entry 3034 fs/fuse/file.c ff = rb_entry(*link, struct fuse_file, polled_node); rb_entry 516 fs/gfs2/rgrp.c cur = rb_entry(n, struct gfs2_rgrpd, rd_node); rb_entry 553 fs/gfs2/rgrp.c rgd = rb_entry(n, struct gfs2_rgrpd, rd_node); rb_entry 580 fs/gfs2/rgrp.c rgd = rb_entry(n, struct gfs2_rgrpd, rd_node); rb_entry 714 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rb_entry 727 fs/gfs2/rgrp.c rgd = rb_entry(n, struct gfs2_rgrpd, rd_node); rb_entry 872 fs/gfs2/rgrp.c struct gfs2_rgrpd *cur = rb_entry(*newn, struct gfs2_rgrpd, rb_entry 1503 fs/gfs2/rgrp.c rb_entry(*newn, struct gfs2_blkreserv, rs_node); rb_entry 1627 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rb_entry 1643 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rb_entry 2279 fs/gfs2/rgrp.c trs = rb_entry(n, struct gfs2_blkreserv, rs_node); rb_entry 131 fs/jffs2/nodelist.c base = rb_entry(parent, struct jffs2_node_frag, rb); rb_entry 537 fs/jffs2/nodelist.c frag = rb_entry(next, struct jffs2_node_frag, rb); rb_entry 334 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); rb_entry 344 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); rb_entry 347 fs/jffs2/nodelist.h #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) rb_entry 348 fs/jffs2/nodelist.h #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) rb_entry 349 fs/jffs2/nodelist.h #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) rb_entry 350 fs/jffs2/nodelist.h #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) rb_entry 351 fs/jffs2/nodelist.h #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) rb_entry 354 fs/jffs2/nodelist.h #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) rb_entry 355 fs/jffs2/nodelist.h #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) rb_entry 356 fs/jffs2/nodelist.h #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) rb_entry 357 fs/jffs2/nodelist.h #define tn_left(tn) rb_entry((tn)->rb.rb_left, struct jffs2_tmp_dnode_info, rb) rb_entry 358 fs/jffs2/nodelist.h #define tn_right(tn) rb_entry((tn)->rb.rb_right, struct jffs2_tmp_dnode_info, rb) rb_entry 360 fs/jffs2/nodelist.h #define tn_last(list) rb_entry(rb_last(list), struct jffs2_tmp_dnode_info, rb) rb_entry 361 fs/jffs2/nodelist.h #define tn_first(list) rb_entry(rb_first(list), struct jffs2_tmp_dnode_info, rb) rb_entry 185 fs/jffs2/readinode.c tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb); rb_entry 344 fs/jffs2/readinode.c insert_point = rb_entry(parent, struct jffs2_tmp_dnode_info, rb); rb_entry 432 fs/jffs2/readinode.c this_tn = rb_entry(parent, struct jffs2_tmp_dnode_info, rb); rb_entry 25 fs/kernfs/dir.c #define rb_to_kn(X) rb_entry((X), struct kernfs_node, rb) rb_entry 15 fs/nfs/blocklayout/extent_tree.c return rb_entry(node, struct pnfs_block_extent, be_node); rb_entry 2273 fs/nfs/dir.c entry = rb_entry(n, struct nfs_access_entry, rb_node); rb_entry 2305 fs/nfs/dir.c rb_entry(n, struct nfs_access_entry, rb_node); rb_entry 2403 fs/nfs/dir.c entry = rb_entry(parent, struct nfs_access_entry, rb_node); rb_entry 152 fs/nfs/fscache.c xkey = rb_entry(parent, struct nfs_fscache_key, node); rb_entry 213 fs/nfs/nfs4state.c sp = rb_entry(pos, struct nfs4_state_owner, so_server_node); rb_entry 410 fs/nfs/nfs4state.c sp = rb_entry(parent, struct nfs4_state_owner, so_server_node); rb_entry 438 fs/nfs/nfs4state.c sp = rb_entry(parent, struct nfs4_state_owner, so_server_node); rb_entry 1720 fs/nfs/nfs4state.c sp = rb_entry(pos, struct nfs4_state_owner, so_server_node); rb_entry 1771 fs/nfs/nfs4state.c sp = rb_entry(pos, struct nfs4_state_owner, so_server_node); rb_entry 1879 fs/nfs/nfs4state.c sp = rb_entry(pos, rb_entry 2640 fs/nfsd/nfs4state.c clp = rb_entry(*new, struct nfs4_client, cl_namenode); rb_entry 2661 fs/nfsd/nfs4state.c clp = rb_entry(node, struct nfs4_client, cl_namenode); rb_entry 350 fs/nfsd/nfscache.c rp = rb_entry(parent, struct svc_cacherep, c_node); rb_entry 724 fs/nilfs2/the_nilfs.c root = rb_entry(n, struct nilfs_root, rb_node); rb_entry 763 fs/nilfs2/the_nilfs.c root = rb_entry(parent, struct nilfs_root, rb_node); rb_entry 79 fs/ocfs2/cluster/nodemanager.c node = rb_entry(parent, struct o2nm_node, nd_ip_node); rb_entry 762 fs/ocfs2/cluster/tcp.c nmh = rb_entry(parent, struct o2net_msg_handler, nh_node); rb_entry 200 fs/ocfs2/refcounttree.c tree = rb_entry(n, struct ocfs2_refcount_tree, rf_node); rb_entry 225 fs/ocfs2/refcounttree.c tmp = rb_entry(parent, struct ocfs2_refcount_tree, rb_entry 535 fs/ocfs2/refcounttree.c tree = rb_entry(node, struct ocfs2_refcount_tree, rf_node); rb_entry 90 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); rb_entry 145 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); rb_entry 280 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); rb_entry 321 fs/ocfs2/reservations.c tmp = rb_entry(parent, struct ocfs2_alloc_reservation, r_node); rb_entry 372 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); rb_entry 530 fs/ocfs2/reservations.c next_resv = rb_entry(next, struct ocfs2_alloc_reservation, rb_entry 567 fs/ocfs2/reservations.c next_resv = rb_entry(next, rb_entry 610 fs/ocfs2/reservations.c prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation, rb_entry 138 fs/ocfs2/uptodate.c item = rb_entry(node, struct ocfs2_meta_cache_item, c_node); rb_entry 216 fs/ocfs2/uptodate.c item = rb_entry(n, struct ocfs2_meta_cache_item, c_node); rb_entry 320 fs/ocfs2/uptodate.c tmp = rb_entry(parent, struct ocfs2_meta_cache_item, c_node); rb_entry 63 fs/overlayfs/readdir.c return rb_entry(n, struct ovl_cache_entry, node); rb_entry 75 fs/proc/generic.c struct proc_dir_entry *de = rb_entry(node, rb_entry 98 fs/proc/generic.c struct proc_dir_entry *this = rb_entry(*new, rb_entry 78 fs/proc/nommu.c return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb)); rb_entry 122 fs/proc/proc_sysctl.c ctl_node = rb_entry(node, struct ctl_node, node); rb_entry 156 fs/proc/proc_sysctl.c parent_node = rb_entry(parent, struct ctl_node, node); rb_entry 389 fs/proc/proc_sysctl.c ctl_node = rb_entry(node, struct ctl_node, node); rb_entry 30 fs/proc/task_nommu.c vma = rb_entry(p, struct vm_area_struct, vm_rb); rb_entry 91 fs/proc/task_nommu.c vma = rb_entry(p, struct vm_area_struct, vm_rb); rb_entry 109 fs/proc/task_nommu.c vma = rb_entry(p, struct vm_area_struct, vm_rb); rb_entry 195 fs/proc/task_nommu.c return nommu_vma_show(m, rb_entry(p, struct vm_area_struct, vm_rb)); rb_entry 621 fs/ubifs/debug.c bud = rb_entry(rb, struct ubifs_bud, rb); rb_entry 709 fs/ubifs/debug.c bud = rb_entry(rb, struct ubifs_bud, rb); rb_entry 1804 fs/ubifs/debug.c fscki = rb_entry(parent, struct fsck_inode, rb); rb_entry 1882 fs/ubifs/debug.c fscki = rb_entry(p, struct fsck_inode, rb); rb_entry 2151 fs/ubifs/debug.c fscki = rb_entry(this, struct fsck_inode, rb); rb_entry 38 fs/ubifs/log.c bud = rb_entry(p, struct ubifs_bud, rb); rb_entry 71 fs/ubifs/log.c bud = rb_entry(p, struct ubifs_bud, rb); rb_entry 122 fs/ubifs/log.c b = rb_entry(parent, struct ubifs_bud, rb); rb_entry 311 fs/ubifs/log.c bud = rb_entry(p1, struct ubifs_bud, rb); rb_entry 568 fs/ubifs/log.c dr = rb_entry(parent, struct done_ref, rb); rb_entry 67 fs/ubifs/orphan.c o = rb_entry(parent, struct ubifs_orphan, rb); rb_entry 103 fs/ubifs/orphan.c o = rb_entry(p, struct ubifs_orphan, rb); rb_entry 588 fs/ubifs/orphan.c o = rb_entry(parent, struct ubifs_orphan, rb); rb_entry 869 fs/ubifs/orphan.c o = rb_entry(parent, struct check_orphan, rb); rb_entry 891 fs/ubifs/orphan.c o = rb_entry(p, struct check_orphan, rb); rb_entry 1254 fs/ubifs/recovery.c e = rb_entry(parent, struct size_entry, rb); rb_entry 1287 fs/ubifs/recovery.c e = rb_entry(p, struct size_entry, rb); rb_entry 1536 fs/ubifs/recovery.c e = rb_entry(this, struct size_entry, rb); rb_entry 84 fs/ubifs/tnc.c o = rb_entry(parent, struct ubifs_old_idx, rb); rb_entry 170 fs/ubifs/tnc_commit.c o = rb_entry(p, struct ubifs_old_idx, rb); rb_entry 51 fs/xfs/xfs_extent_busy.c busyp = rb_entry(parent, struct xfs_extent_busy, rb_node); rb_entry 100 fs/xfs/xfs_extent_busy.c busyp = rb_entry(rbp, struct xfs_extent_busy, rb_node); rb_entry 300 fs/xfs/xfs_extent_busy.c rb_entry(rbp, struct xfs_extent_busy, rb_node); rb_entry 353 fs/xfs/xfs_extent_busy.c rb_entry(rbp, struct xfs_extent_busy, rb_node); rb_entry 199 include/linux/ceph/libceph.h type *cur = rb_entry(*n, type, nodefld); \ rb_entry 233 include/linux/ceph/libceph.h type *cur = rb_entry(n, type, nodefld); \ rb_entry 164 include/linux/elevator.h #define rb_entry_rq(node) rb_entry((node), struct request, rb_node) rb_entry 48 include/linux/interval_tree_generic.h parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \ rb_entry 89 include/linux/interval_tree_generic.h ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \ rb_entry 108 include/linux/interval_tree_generic.h node = rb_entry(node->ITRB.rb_right, \ rb_entry 140 include/linux/interval_tree_generic.h node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \ rb_entry 144 include/linux/interval_tree_generic.h leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ rb_entry 165 include/linux/interval_tree_generic.h ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \ rb_entry 177 include/linux/interval_tree_generic.h node = rb_entry(rb, ITSTRUCT, ITRB); \ rb_entry 673 include/linux/perf_event.h struct list_head rb_entry; rb_entry 89 include/linux/rbtree.h ____ptr ? rb_entry(____ptr, type, member) : NULL; \ rb_entry 80 include/linux/rbtree_augmented.h RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ rb_entry 89 include/linux/rbtree_augmented.h RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ rb_entry 90 include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ rb_entry 96 include/linux/rbtree_augmented.h RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ rb_entry 97 include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ rb_entry 127 include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ rb_entry 132 include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \ rb_entry 38 include/linux/timerqueue.h return rb_entry(leftmost, struct timerqueue_node, node); rb_entry 141 ipc/mqueue.c leaf = rb_entry(parent, struct posix_msg_tree_node, rb_node); rb_entry 212 ipc/mqueue.c leaf = rb_entry(parent, struct posix_msg_tree_node, rb_node); rb_entry 5481 kernel/events/core.c list_del_rcu(&event->rb_entry); rb_entry 5495 kernel/events/core.c list_add_rcu(&event->rb_entry, &rb->event_list); rb_entry 5532 kernel/events/core.c list_for_each_entry_rcu(event, &rb->event_list, rb_entry) rb_entry 5644 kernel/events/core.c list_for_each_entry_rcu(event, &rb->event_list, rb_entry) { rb_entry 6989 kernel/events/core.c list_for_each_entry_rcu(iter, &event->rb->event_list, rb_entry) { rb_entry 10455 kernel/events/core.c INIT_LIST_HEAD(&event->rb_entry); rb_entry 647 kernel/events/uprobes.c uprobe = rb_entry(n, struct uprobe, rb_node); rb_entry 684 kernel/events/uprobes.c u = rb_entry(parent, struct uprobe, rb_node); rb_entry 1277 kernel/events/uprobes.c struct uprobe *u = rb_entry(n, struct uprobe, rb_node); rb_entry 1316 kernel/events/uprobes.c u = rb_entry(t, struct uprobe, rb_node); rb_entry 1323 kernel/events/uprobes.c u = rb_entry(t, struct uprobe, rb_node); rb_entry 282 kernel/locking/rtmutex.c entry = rb_entry(parent, struct rt_mutex_waiter, tree_entry); rb_entry 315 kernel/locking/rtmutex.c entry = rb_entry(parent, struct rt_mutex_waiter, pi_tree_entry); rb_entry 59 kernel/locking/rtmutex_common.h w = rb_entry(leftmost, struct rt_mutex_waiter, tree_entry); rb_entry 73 kernel/locking/rtmutex_common.h return rb_entry(p->pi_waiters.rb_leftmost, rb_entry 138 kernel/power/swap.c ext = rb_entry(*new, struct swsusp_extent, node); rb_entry 204 kernel/power/swap.c ext = rb_entry(node, struct swsusp_extent, node); rb_entry 48 kernel/power/wakelock.c wl = rb_entry(node, struct wakelock, node); rb_entry 161 kernel/power/wakelock.c wl = rb_entry(*node, struct wakelock, node); rb_entry 457 kernel/sched/deadline.c entry = rb_entry(parent, struct task_struct, rb_entry 487 kernel/sched/deadline.c dl_rq->earliest_dl.next = rb_entry(next_node, rb_entry 1361 kernel/sched/deadline.c entry = rb_entry(leftmost, struct sched_dl_entity, rb_node); rb_entry 1414 kernel/sched/deadline.c entry = rb_entry(parent, struct sched_dl_entity, rb_node); rb_entry 1773 kernel/sched/deadline.c return rb_entry(left, struct sched_dl_entity, rb_node); rb_entry 1862 kernel/sched/deadline.c p = rb_entry(next_node, struct task_struct, pushable_dl_tasks); rb_entry 2028 kernel/sched/deadline.c p = rb_entry(rq->dl.pushable_dl_tasks_root.rb_leftmost, rb_entry 546 kernel/sched/fair.c se = rb_entry(leftmost, struct sched_entity, run_node); rb_entry 577 kernel/sched/fair.c entry = rb_entry(parent, struct sched_entity, run_node); rb_entry 607 kernel/sched/fair.c return rb_entry(left, struct sched_entity, run_node); rb_entry 617 kernel/sched/fair.c return rb_entry(next, struct sched_entity, run_node); rb_entry 628 kernel/sched/fair.c return rb_entry(last, struct sched_entity, run_node); rb_entry 39 lib/rbtree_test.c if (key < rb_entry(parent, struct test_node, rb)->key) rb_entry 57 lib/rbtree_test.c if (key < rb_entry(parent, struct test_node, rb)->key) rb_entry 95 lib/rbtree_test.c parent = rb_entry(rb_parent, struct test_node, rb); rb_entry 120 lib/rbtree_test.c parent = rb_entry(rb_parent, struct test_node, rb); rb_entry 198 lib/rbtree_test.c struct test_node *node = rb_entry(rb, struct test_node, rb); rb_entry 224 lib/rbtree_test.c struct test_node *node = rb_entry(rb, struct test_node, rb); rb_entry 227 lib/rbtree_test.c subtree = rb_entry(node->rb.rb_left, struct test_node, rb_entry 233 lib/rbtree_test.c subtree = rb_entry(node->rb.rb_right, struct test_node, rb_entry 39 lib/timerqueue.c ptr = rb_entry(parent, struct timerqueue_node, node); rb_entry 419 mm/backing-dev.c congested = rb_entry(parent, struct bdi_writeback_congested, rb_entry 780 mm/backing-dev.c rb_entry(rbn, struct bdi_writeback_congested, rb_node); rb_entry 895 mm/backing-dev.c bdi = rb_entry(parent, struct backing_dev_info, rb_node); rb_entry 925 mm/backing-dev.c bdi = rb_entry(*p, struct backing_dev_info, rb_node); rb_entry 42 mm/interval_tree.c parent = rb_entry(prev->shared.rb.rb_right, rb_entry 47 mm/interval_tree.c parent = rb_entry(parent->shared.rb.rb_left, rb_entry 385 mm/kmemleak.c rb_entry(rb, struct kmemleak_object, rb_node); rb_entry 629 mm/kmemleak.c parent = rb_entry(rb_parent, struct kmemleak_object, rb_node); rb_entry 946 mm/ksm.c stable_node = rb_entry(root_stable_tree[nid].rb_node, rb_entry 1578 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); rb_entry 1825 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); rb_entry 1942 mm/ksm.c tree_rmap_item = rb_entry(*new, struct rmap_item, node); rb_entry 2775 mm/ksm.c stable_node = rb_entry(node, struct stable_node, node); rb_entry 540 mm/memcontrol.c mz_node = rb_entry(parent, struct mem_cgroup_per_node, rb_entry 659 mm/memcontrol.c mz = rb_entry(mctz->rb_rightmost, rb_entry 2293 mm/mempolicy.c struct sp_node *p = rb_entry(n, struct sp_node, nd); rb_entry 2309 mm/mempolicy.c w = rb_entry(prev, struct sp_node, nd); rb_entry 2314 mm/mempolicy.c return rb_entry(n, struct sp_node, nd); rb_entry 2329 mm/mempolicy.c nd = rb_entry(parent, struct sp_node, nd); rb_entry 2543 mm/mempolicy.c n = rb_entry(next, struct sp_node, nd); rb_entry 2654 mm/mempolicy.c n = rb_entry(next, struct sp_node, nd); rb_entry 312 mm/mmap.c subtree_gap = rb_entry(vma->vm_rb.rb_left, rb_entry 318 mm/mmap.c subtree_gap = rb_entry(vma->vm_rb.rb_right, rb_entry 335 mm/mmap.c vma = rb_entry(nd, struct vm_area_struct, vm_rb); rb_entry 380 mm/mmap.c vma = rb_entry(nd, struct vm_area_struct, vm_rb); rb_entry 539 mm/mmap.c vma_tmp = rb_entry(__rb_parent, struct vm_area_struct, vm_rb); rb_entry 554 mm/mmap.c *pprev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); rb_entry 1908 mm/mmap.c vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); rb_entry 1917 mm/mmap.c rb_entry(vma->vm_rb.rb_left, rb_entry 1937 mm/mmap.c rb_entry(vma->vm_rb.rb_right, rb_entry 1950 mm/mmap.c vma = rb_entry(rb_parent(prev), rb_entry 2012 mm/mmap.c vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); rb_entry 2021 mm/mmap.c rb_entry(vma->vm_rb.rb_right, rb_entry 2041 mm/mmap.c rb_entry(vma->vm_rb.rb_left, rb_entry 2054 mm/mmap.c vma = rb_entry(rb_parent(prev), rb_entry 2250 mm/mmap.c tmp = rb_entry(rb_node, struct vm_area_struct, vm_rb); rb_entry 2283 mm/mmap.c *pprev = rb_node ? rb_entry(rb_node, struct vm_area_struct, vm_rb) : NULL; rb_entry 465 mm/nommu.c last = rb_entry(lastp, struct vm_region, vm_rb); rb_entry 470 mm/nommu.c region = rb_entry(p, struct vm_region, vm_rb); rb_entry 471 mm/nommu.c last = rb_entry(lastp, struct vm_region, vm_rb); rb_entry 500 mm/nommu.c pregion = rb_entry(parent, struct vm_region, vm_rb); rb_entry 613 mm/nommu.c pvma = rb_entry(parent, struct vm_area_struct, vm_rb); rb_entry 642 mm/nommu.c prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); rb_entry 1170 mm/nommu.c pregion = rb_entry(rb, struct vm_region, vm_rb); rb_entry 158 mm/swapfile.c return rb_entry(rb, struct swap_extent, rb_node); rb_entry 164 mm/swapfile.c return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL; rb_entry 212 mm/swapfile.c se = rb_entry(rb, struct swap_extent, rb_node); rb_entry 2288 mm/swapfile.c struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node); rb_entry 2328 mm/swapfile.c se = rb_entry(parent, struct swap_extent, rb_node); rb_entry 285 mm/util.c next = rb_entry(rb_parent, rb_entry 421 mm/vmalloc.c va = rb_entry(n, struct vmap_area, rb_node); rb_entry 461 mm/vmalloc.c tmp_va = rb_entry(*link, struct vmap_area, rb_node); rb_entry 496 mm/vmalloc.c list = &rb_entry(parent, struct vmap_area, rb_node)->list; rb_entry 509 mm/vmalloc.c head = &rb_entry(parent, struct vmap_area, rb_node)->list; rb_entry 567 mm/vmalloc.c va = rb_entry(n, struct vmap_area, rb_node); rb_entry 572 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); rb_entry 587 mm/vmalloc.c va = rb_entry(n, struct vmap_area, rb_node); rb_entry 631 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); rb_entry 803 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); rb_entry 828 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); rb_entry 3159 mm/vmalloc.c tmp = rb_entry(n, struct vmap_area, rb_node); rb_entry 273 mm/zswap.c entry = rb_entry(node, struct zswap_entry, rbnode); rb_entry 296 mm/zswap.c myentry = rb_entry(parent, struct zswap_entry, rbnode); rb_entry 25 net/6lowpan/nhc.c struct lowpan_nhc *this = rb_entry(*new, struct lowpan_nhc, rb_entry 67 net/6lowpan/nhc.c struct lowpan_nhc *nhc = rb_entry(node, struct lowpan_nhc, rb_entry 155 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); rb_entry 176 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); rb_entry 388 net/802/garp.c attr = rb_entry(node, struct garp_attr, node); rb_entry 244 net/802/mrp.c attr = rb_entry(parent, struct mrp_attr, node); rb_entry 265 net/802/mrp.c attr = rb_entry(parent, struct mrp_attr, node); rb_entry 573 net/802/mrp.c attr = rb_entry(node, struct mrp_attr, node); rb_entry 125 net/ceph/auth_x.c th = rb_entry(parent, struct ceph_x_ticket_handler, node); rb_entry 773 net/ceph/auth_x.c rb_entry(p, struct ceph_x_ticket_handler, node); rb_entry 71 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_pool_info, node); rb_entry 92 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); rb_entry 103 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); rb_entry 110 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); rb_entry 121 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); rb_entry 160 net/ceph/debugfs.c req = rb_entry(rp, struct ceph_mon_generic_request, node); rb_entry 236 net/ceph/debugfs.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 262 net/ceph/debugfs.c rb_entry(n, struct ceph_osd_linger_request, node); rb_entry 323 net/ceph/debugfs.c rb_entry(n, struct ceph_osd_backoff, id_node); rb_entry 348 net/ceph/debugfs.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 356 net/ceph/debugfs.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 364 net/ceph/debugfs.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 961 net/ceph/mon_client.c req = rb_entry(p, struct ceph_mon_generic_request, node); rb_entry 1159 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 1163 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_request, r_node); rb_entry 1173 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_request, r_node); rb_entry 1318 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 1328 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); rb_entry 1369 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 1461 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); rb_entry 1845 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_backoff, spg_node); rb_entry 1873 net/ceph/osd_client.c rb_entry(rb_first(&osd->o_backoff_mappings), rb_entry 1878 net/ceph/osd_client.c rb_entry(rb_first(&spg->backoffs), rb_entry 3289 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 3294 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_request, r_node); rb_entry 3312 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_linger_request, node); rb_entry 3331 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_request, r_node); rb_entry 3710 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); rb_entry 3763 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); rb_entry 3799 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 3857 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); rb_entry 3882 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 3909 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 3924 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 4076 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 4089 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); rb_entry 4315 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_request, r_node); rb_entry 4545 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 4550 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_request, r_node); rb_entry 5121 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); rb_entry 5212 net/ceph/osd_client.c struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds), rb_entry 180 net/ceph/osdmap.c rb_entry(rb_first(&c->choose_args), rb_entry 647 net/ceph/osdmap.c pi = rb_entry(parent, struct ceph_pg_pool_info, node); rb_entry 667 net/ceph/osdmap.c pi = rb_entry(n, struct ceph_pg_pool_info, node); rb_entry 705 net/ceph/osdmap.c rb_entry(rbp, struct ceph_pg_pool_info, node); rb_entry 953 net/ceph/osdmap.c rb_entry(rb_first(&map->pg_temp), rb_entry 960 net/ceph/osdmap.c rb_entry(rb_first(&map->primary_temp), rb_entry 967 net/ceph/osdmap.c rb_entry(rb_first(&map->pg_upmap), rb_entry 974 net/ceph/osdmap.c rb_entry(rb_first(&map->pg_upmap_items), rb_entry 981 net/ceph/osdmap.c rb_entry(rb_first(&map->pg_pools), rb_entry 21 net/ceph/string_table.c exist = rb_entry(*p, struct ceph_string, node); rb_entry 56 net/ceph/string_table.c exist = rb_entry(*p, struct ceph_string, node); rb_entry 3098 net/core/skbuff.c struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); rb_entry 238 net/ipv4/inet_fragment.c struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); rb_entry 118 net/ipv4/inetpeer.c p = rb_entry(parent, struct inet_peer, rb_node); rb_entry 301 net/ipv4/inetpeer.c struct inet_peer *peer = rb_entry(p, struct inet_peer, rb_node); rb_entry 156 net/ipv4/nexthop.c nh = rb_entry(parent, struct nexthop, rb_node); rb_entry 1028 net/ipv4/nexthop.c nh = rb_entry(parent, struct nexthop, rb_node); rb_entry 1092 net/ipv4/nexthop.c nh = rb_entry(node, struct nexthop, rb_node); rb_entry 1737 net/ipv4/nexthop.c nh = rb_entry(node, struct nexthop, rb_node); rb_entry 322 net/netfilter/nf_conncount.c rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node); rb_entry 404 net/netfilter/nf_conncount.c rbconn = rb_entry(parent, struct nf_conncount_rb, node); rb_entry 458 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); rb_entry 474 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); rb_entry 565 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); rb_entry 63 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); rb_entry 144 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); rb_entry 231 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); rb_entry 324 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); rb_entry 363 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); rb_entry 399 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); rb_entry 477 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); rb_entry 115 net/rds/cong.c map = rb_entry(parent, struct rds_cong_map, m_rb_node); rb_entry 406 net/rds/cong.c map = rb_entry(node, struct rds_cong_map, m_rb_node); rb_entry 74 net/rds/rdma.c mr = rb_entry(parent, struct rds_mr, r_rb_node); rb_entry 137 net/rds/rdma.c mr = rb_entry(node, struct rds_mr, r_rb_node); rb_entry 114 net/rxrpc/call_accept.c xcall = rb_entry(parent, struct rxrpc_call, sock_node); rb_entry 488 net/rxrpc/call_accept.c call = rb_entry(parent, struct rxrpc_call, sock_node); rb_entry 526 net/rxrpc/call_accept.c call = rb_entry(parent, struct rxrpc_call, sock_node); rb_entry 74 net/rxrpc/call_object.c call = rb_entry(p, struct rxrpc_call, sock_node); rb_entry 258 net/rxrpc/call_object.c xcall = rb_entry(parent, struct rxrpc_call, sock_node); rb_entry 307 net/rxrpc/conn_client.c conn = rb_entry(p, struct rxrpc_connection, client_node); rb_entry 374 net/rxrpc/conn_client.c conn = rb_entry(parent, struct rxrpc_connection, client_node); rb_entry 42 net/rxrpc/conn_service.c conn = rb_entry(p, struct rxrpc_connection, service_node); rb_entry 76 net/rxrpc/conn_service.c cursor = rb_entry(parent, rb_entry 171 net/sched/sch_fq.c aux = rb_entry(parent, struct fq_flow, rate_node); rb_entry 214 net/sched/sch_fq.c f = rb_entry(parent, struct fq_flow, fq_node); rb_entry 293 net/sched/sch_fq.c f = rb_entry(parent, struct fq_flow, fq_node); rb_entry 472 net/sched/sch_fq.c struct fq_flow *f = rb_entry(p, struct fq_flow, rate_node); rb_entry 637 net/sched/sch_fq.c f = rb_entry(p, struct fq_flow, fq_node); rb_entry 667 net/sched/sch_fq.c of = rb_entry(op, struct fq_flow, fq_node); rb_entry 680 net/sched/sch_fq.c nf = rb_entry(parent, struct fq_flow, fq_node); rb_entry 194 net/sched/sch_hfsc.c cl1 = rb_entry(parent, struct hfsc_class, el_node); rb_entry 225 net/sched/sch_hfsc.c p = rb_entry(n, struct hfsc_class, el_node); rb_entry 243 net/sched/sch_hfsc.c return rb_entry(n, struct hfsc_class, el_node); rb_entry 259 net/sched/sch_hfsc.c cl1 = rb_entry(parent, struct hfsc_class, vt_node); rb_entry 289 net/sched/sch_hfsc.c p = rb_entry(n, struct hfsc_class, vt_node); rb_entry 328 net/sched/sch_hfsc.c cl1 = rb_entry(parent, struct hfsc_class, cf_node); rb_entry 660 net/sched/sch_hfsc.c p = rb_entry(n, struct hfsc_class, cf_node); rb_entry 683 net/sched/sch_hfsc.c max_cl = rb_entry(n, struct hfsc_class, vt_node); rb_entry 282 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, node[prio]); rb_entry 316 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, pq_node); rb_entry 718 net/sched/sch_htb.c cl = rb_entry(p, struct htb_class, pq_node); rb_entry 747 net/sched/sch_htb.c rb_entry(n, struct htb_class, node[prio]); rb_entry 807 net/sched/sch_htb.c cl = rb_entry(*sp->pptr, struct htb_class, node[prio]); rb_entry 74 net/smc/smc_core.c struct smc_connection *cur = rb_entry(*link, rb_entry 475 net/smc/smc_core.c conn = rb_entry(node, struct smc_connection, alert_node); rb_entry 265 net/smc/smc_core.h struct smc_connection *cur = rb_entry(node, rb_entry 910 net/sunrpc/xprt.c req = rb_entry(n, struct rpc_rqst, rq_recv); rb_entry 934 net/sunrpc/xprt.c req = rb_entry(n, struct rpc_rqst, rq_recv); rb_entry 972 net/wireless/scan.c tbss = rb_entry(parent, struct cfg80211_internal_bss, rbn); rb_entry 1001 net/wireless/scan.c bss = rb_entry(n, struct cfg80211_internal_bss, rbn); rb_entry 906 net/xfrm/xfrm_policy.c node = rb_entry(*p, struct xfrm_pol_inexact_node, node); rb_entry 960 net/xfrm/xfrm_policy.c node = rb_entry(rnode, struct xfrm_pol_inexact_node, node); rb_entry 989 net/xfrm/xfrm_policy.c node = rb_entry(*p, struct xfrm_pol_inexact_node, node); rb_entry 1057 net/xfrm/xfrm_policy.c node = rb_entry(rn, struct xfrm_pol_inexact_node, node); rb_entry 1920 net/xfrm/xfrm_policy.c node = rb_entry(parent, struct xfrm_pol_inexact_node, node); rb_entry 649 security/apparmor/label.c struct aa_label *this = rb_entry(*new, struct aa_label, node); rb_entry 702 security/apparmor/label.c struct aa_label *this = rb_entry(node, struct aa_label, node); rb_entry 1961 security/apparmor/label.c struct aa_label *this = rb_entry(node, struct aa_label, node); rb_entry 1994 security/apparmor/label.c label = rb_entry(node, struct aa_label, node); rb_entry 39 security/integrity/iint.c iint = rb_entry(n, struct integrity_iint_cache, rb_node); rb_entry 114 security/integrity/iint.c test_iint = rb_entry(parent, struct integrity_iint_cache, rb_entry 213 security/keys/gc.c key = rb_entry(cursor, struct key, serial_node); rb_entry 63 security/keys/key.c user = rb_entry(parent, struct key_user, node); rb_entry 154 security/keys/key.c xkey = rb_entry(parent, struct key, serial_node); rb_entry 185 security/keys/key.c xkey = rb_entry(parent, struct key, serial_node); rb_entry 665 security/keys/key.c key = rb_entry(n, struct key, serial_node); rb_entry 70 security/keys/proc.c struct key *key = rb_entry(n, struct key, serial_node); rb_entry 85 security/keys/proc.c struct key *key = rb_entry(n, struct key, serial_node); rb_entry 108 security/keys/proc.c minkey = rb_entry(n, struct key, serial_node); rb_entry 131 security/keys/proc.c struct key *key = rb_entry(n, struct key, serial_node); rb_entry 156 security/keys/proc.c struct key *key = rb_entry(_p, struct key, serial_node); rb_entry 255 security/keys/proc.c struct key_user *user = rb_entry(n, struct key_user, node); rb_entry 306 security/keys/proc.c struct key_user *user = rb_entry(_p, struct key_user, node); rb_entry 77 tools/include/linux/rbtree.h ____ptr ? rb_entry(____ptr, type, member) : NULL; \ rb_entry 82 tools/include/linux/rbtree_augmented.h RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ rb_entry 91 tools/include/linux/rbtree_augmented.h RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ rb_entry 92 tools/include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ rb_entry 98 tools/include/linux/rbtree_augmented.h RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ rb_entry 99 tools/include/linux/rbtree_augmented.h RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ rb_entry 129 tools/include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ rb_entry 134 tools/include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \ rb_entry 109 tools/lib/lockdep/preload.c l = rb_entry(*node, struct lock_lookup, node); rb_entry 184 tools/lib/lockdep/preload.c return rb_entry(*node, struct lock_lookup, node); rb_entry 317 tools/perf/builtin-annotate.c struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2112 tools/perf/builtin-c2c.c he = rb_entry(next, struct hist_entry, rb_node); rb_entry 2236 tools/perf/builtin-c2c.c struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2304 tools/perf/builtin-c2c.c struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node); rb_entry 523 tools/perf/builtin-diff.c struct hist_entry *he = rb_entry(next, struct hist_entry, rb_node_in); rb_entry 679 tools/perf/builtin-diff.c struct hist_entry *he_pair = rb_entry(next, struct hist_entry, rb_entry 710 tools/perf/builtin-diff.c struct hist_entry *he = rb_entry(next, struct hist_entry, rb_entry 747 tools/perf/builtin-diff.c he = rb_entry(next, struct hist_entry, rb_node_in); rb_entry 94 tools/perf/builtin-kmem.c data = rb_entry(*node, struct alloc_stat, node); rb_entry 139 tools/perf/builtin-kmem.c data = rb_entry(*node, struct alloc_stat, node); rb_entry 222 tools/perf/builtin-kmem.c data = rb_entry(node, struct alloc_stat, node); rb_entry 452 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); rb_entry 502 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); rb_entry 557 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); rb_entry 1004 tools/perf/builtin-kmem.c struct alloc_stat *data = rb_entry(next, struct alloc_stat, rb_entry 1077 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); rb_entry 1119 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); rb_entry 1264 tools/perf/builtin-kmem.c this = rb_entry(*new, struct alloc_stat, node); rb_entry 1295 tools/perf/builtin-kmem.c data = rb_entry(node, struct alloc_stat, node); rb_entry 1311 tools/perf/builtin-kmem.c this = rb_entry(*new, struct page_stat, node); rb_entry 1342 tools/perf/builtin-kmem.c data = rb_entry(node, struct page_stat, node); rb_entry 732 tools/perf/builtin-report.c struct map *map = rb_entry(nd, struct map, rb_node); rb_entry 799 tools/perf/builtin-report.c task->thread = rb_entry(nd, struct thread, rb_node); rb_entry 1469 tools/perf/builtin-sched.c data = rb_entry(node, struct work_atoms, node); rb_entry 2785 tools/perf/builtin-sched.c chain = rb_entry(rb_node, struct callchain_node, rb_node); rb_entry 3133 tools/perf/builtin-sched.c data = rb_entry(node, struct work_atoms, node); rb_entry 3159 tools/perf/builtin-sched.c work_list = rb_entry(next, struct work_atoms, node); rb_entry 411 tools/perf/builtin-top.c n = rb_entry(next, struct hist_entry, rb_node); rb_entry 3673 tools/perf/builtin-trace.c struct int_node *source = rb_entry(nd, struct int_node, rb_node); rb_entry 3761 tools/perf/builtin-trace.c entry->thread = rb_entry(nd, struct thread, rb_node); rb_entry 179 tools/perf/tests/hists_common.c he = rb_entry(node, struct hist_entry, rb_node_in); rb_entry 206 tools/perf/tests/hists_common.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 143 tools/perf/tests/hists_cumulate.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 204 tools/perf/tests/hists_cumulate.c node && (he = rb_entry(node, struct hist_entry, rb_node)); rb_entry 225 tools/perf/tests/hists_cumulate.c cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); rb_entry 159 tools/perf/tests/hists_link.c he = rb_entry(node, struct hist_entry, rb_node_in); rb_entry 211 tools/perf/tests/hists_link.c he = rb_entry(node, struct hist_entry, rb_node_in); rb_entry 109 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 167 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 173 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 179 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 185 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 191 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 197 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 203 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 209 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 215 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 267 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 272 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 321 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 327 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 333 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 339 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 345 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 399 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 405 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 411 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 417 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 423 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 429 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 435 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 441 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 447 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 502 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 510 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 517 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 524 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 531 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 538 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 545 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 552 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 559 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 566 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 115 tools/perf/tests/vmlinux-kallsyms.c sym = rb_entry(nd, struct symbol, rb_node); rb_entry 254 tools/perf/ui/browsers/annotate.c l = rb_entry(parent, struct annotation_line, rb_node); rb_entry 293 tools/perf/ui/browsers/annotate.c struct annotation_line * pos = rb_entry(nd, struct annotation_line, rb_node); rb_entry 68 tools/perf/ui/browsers/hists.c rb_entry(nd, struct hist_entry, rb_node); rb_entry 179 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node); rb_entry 262 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 286 tools/perf/ui/browsers/hists.c child = rb_entry(node, struct hist_entry, rb_node); rb_entry 330 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node); rb_entry 370 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 467 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node); rb_entry 508 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 524 tools/perf/ui/browsers/hists.c child = rb_entry(nd, struct hist_entry, rb_node); rb_entry 583 tools/perf/ui/browsers/hists.c he = rb_entry(nd, struct hist_entry, rb_node); rb_entry 675 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(browser->b.top, rb_entry 853 tools/perf/ui/browsers/hists.c child = rb_entry(node, struct callchain_node, rb_node); rb_entry 873 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); rb_entry 977 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); rb_entry 1066 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); rb_entry 1777 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1815 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1838 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1886 tools/perf/ui/browsers/hists.c h = rb_entry(browser->top, struct hist_entry, rb_node); rb_entry 1908 tools/perf/ui/browsers/hists.c h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1930 tools/perf/ui/browsers/hists.c h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1967 tools/perf/ui/browsers/hists.c h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 1976 tools/perf/ui/browsers/hists.c h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2102 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2792 tools/perf/ui/browsers/hists.c he = rb_entry(nd, struct hist_entry, rb_node); rb_entry 27 tools/perf/ui/browsers/map.c struct symbol *sym = rb_entry(nd, struct symbol, rb_node); rb_entry 121 tools/perf/ui/browsers/map.c struct symbol *pos = rb_entry(nd, struct symbol, rb_node); rb_entry 109 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 173 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 233 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); rb_entry 359 tools/perf/ui/gtk/hists.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 425 tools/perf/ui/gtk/hists.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 129 tools/perf/ui/stdio/hist.c child = rb_entry(node, struct callchain_node, rb_node); rb_entry 212 tools/perf/ui/stdio/hist.c cnode = rb_entry(node, struct callchain_node, rb_node); rb_entry 231 tools/perf/ui/stdio/hist.c cnode = rb_entry(node, struct callchain_node, rb_node); rb_entry 313 tools/perf/ui/stdio/hist.c chain = rb_entry(rb_node, struct callchain_node, rb_node); rb_entry 365 tools/perf/ui/stdio/hist.c chain = rb_entry(rb_node, struct callchain_node, rb_node); rb_entry 831 tools/perf/ui/stdio/hist.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2125 tools/perf/util/annotate.c iter = rb_entry(parent, struct annotation_line, rb_node); rb_entry 2172 tools/perf/util/annotate.c iter = rb_entry(parent, struct annotation_line, rb_node); rb_entry 2193 tools/perf/util/annotate.c al = rb_entry(node, struct annotation_line, rb_node); rb_entry 2222 tools/perf/util/annotate.c al = rb_entry(node, struct annotation_line, rb_node); rb_entry 23 tools/perf/util/block-range.c struct block_range *entry = rb_entry(rb, struct block_range, node); rb_entry 41 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); rb_entry 90 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); rb_entry 117 tools/perf/util/block-range.c next = rb_entry(n, struct block_range, node); rb_entry 43 tools/perf/util/block-range.h return rb_entry(n, struct block_range, node); rb_entry 374 tools/perf/util/build-id.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 408 tools/perf/util/build-id.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 862 tools/perf/util/build-id.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 880 tools/perf/util/build-id.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 95 tools/perf/util/call-path.c cp = rb_entry(node_parent, struct call_path, rb_node); rb_entry 384 tools/perf/util/callchain.c rnode = rb_entry(parent, struct callchain_node, rb_node); rb_entry 421 tools/perf/util/callchain.c child = rb_entry(n, struct callchain_node, rb_node_in); rb_entry 453 tools/perf/util/callchain.c child = rb_entry(n, struct callchain_node, rb_node_in); rb_entry 483 tools/perf/util/callchain.c child = rb_entry(n, struct callchain_node, rb_node_in); rb_entry 548 tools/perf/util/callchain.c child = rb_entry(n, struct callchain_node, rb_node_in); rb_entry 843 tools/perf/util/callchain.c first = rb_entry(p, struct callchain_node, rb_node_in); rb_entry 885 tools/perf/util/callchain.c rnode = rb_entry(parent, struct callchain_node, rb_node_in); rb_entry 1263 tools/perf/util/callchain.c child = rb_entry(n, struct callchain_node, rb_node_in); rb_entry 70 tools/perf/util/comm.c iter = rb_entry(parent, struct comm_str, rb_node); rb_entry 763 tools/perf/util/dso.c cache = rb_entry(next, struct dso_cache, rb_node); rb_entry 782 tools/perf/util/dso.c cache = rb_entry(parent, struct dso_cache, rb_node); rb_entry 810 tools/perf/util/dso.c cache = rb_entry(parent, struct dso_cache, rb_node); rb_entry 1309 tools/perf/util/dso.c struct symbol *pos = rb_entry(nd, struct symbol, rb_node); rb_entry 53 tools/perf/util/dsos.c struct dso *this = rb_entry(*p, struct dso, rb_node); rb_entry 29 tools/perf/util/env.c node = rb_entry(parent, struct bpf_prog_info_node, rb_node); rb_entry 57 tools/perf/util/env.c node = rb_entry(n, struct bpf_prog_info_node, rb_node); rb_entry 84 tools/perf/util/env.c node = rb_entry(parent, struct btf_node, rb_node); rb_entry 111 tools/perf/util/env.c node = rb_entry(n, struct btf_node, rb_node); rb_entry 140 tools/perf/util/env.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); rb_entry 154 tools/perf/util/env.c node = rb_entry(next, struct btf_node, rb_node); rb_entry 936 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); rb_entry 985 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); rb_entry 1546 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); rb_entry 1570 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); rb_entry 233 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); rb_entry 323 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); rb_entry 366 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); rb_entry 382 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); rb_entry 396 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node); rb_entry 581 tools/perf/util/hist.c he = rb_entry(parent, struct hist_entry, rb_node_in); rb_entry 1409 tools/perf/util/hist.c iter = rb_entry(parent, struct hist_entry, rb_node_in); rb_entry 1528 tools/perf/util/hist.c iter = rb_entry(parent, struct hist_entry, rb_node_in); rb_entry 1606 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node_in); rb_entry 1691 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 1711 tools/perf/util/hist.c iter = rb_entry(parent, struct hist_entry, rb_node); rb_entry 1745 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node_in); rb_entry 1811 tools/perf/util/hist.c iter = rb_entry(parent, struct hist_entry, rb_node); rb_entry 1869 tools/perf/util/hist.c n = rb_entry(next, struct hist_entry, rb_node_in); rb_entry 1930 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); rb_entry 1934 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node); rb_entry 1941 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); rb_entry 1960 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); rb_entry 1983 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); rb_entry 1987 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); rb_entry 2101 tools/perf/util/hist.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2122 tools/perf/util/hist.c iter = rb_entry(parent, struct hist_entry, rb_node); rb_entry 2140 tools/perf/util/hist.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2163 tools/perf/util/hist.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2207 tools/perf/util/hist.c struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); rb_entry 2295 tools/perf/util/hist.c he = rb_entry(parent, struct hist_entry, rb_node_in); rb_entry 2340 tools/perf/util/hist.c he = rb_entry(parent, struct hist_entry, rb_node_in); rb_entry 2383 tools/perf/util/hist.c struct hist_entry *iter = rb_entry(n, struct hist_entry, rb_node_in); rb_entry 2407 tools/perf/util/hist.c iter = rb_entry(n, struct hist_entry, rb_node_in); rb_entry 2432 tools/perf/util/hist.c pos = rb_entry(nd, struct hist_entry, rb_node_in); rb_entry 2463 tools/perf/util/hist.c pos = rb_entry(nd, struct hist_entry, rb_node_in); rb_entry 2480 tools/perf/util/hist.c pos = rb_entry(nd, struct hist_entry, rb_node_in); rb_entry 2539 tools/perf/util/hist.c pos = rb_entry(nd, struct hist_entry, rb_node_in); rb_entry 2564 tools/perf/util/hist.c pos = rb_entry(nd, struct hist_entry, rb_node_in); rb_entry 2750 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node_in); rb_entry 49 tools/perf/util/intlist.h return rn ? rb_entry(rn, struct int_node, rb_node) : NULL; rb_entry 57 tools/perf/util/intlist.h return rn ? rb_entry(rn, struct int_node, rb_node) : NULL; rb_entry 198 tools/perf/util/machine.c struct thread *t = rb_entry(nd, struct thread, rb_node); rb_entry 278 tools/perf/util/machine.c pos = rb_entry(parent, struct machine, rb_node); rb_entry 300 tools/perf/util/machine.c struct machine *machine = rb_entry(nd, struct machine, rb_node); rb_entry 318 tools/perf/util/machine.c machine = rb_entry(parent, struct machine, rb_node); rb_entry 372 tools/perf/util/machine.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 386 tools/perf/util/machine.c machine = rb_entry(node, struct machine, rb_node); rb_entry 509 tools/perf/util/machine.c th = rb_entry(parent, struct thread, rb_node); rb_entry 809 tools/perf/util/machine.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 829 tools/perf/util/machine.c struct machine *pos = rb_entry(nd, struct machine, rb_node); rb_entry 870 tools/perf/util/machine.c struct thread *pos = rb_entry(nd, struct thread, rb_node); rb_entry 1183 tools/perf/util/machine.c struct machine *pos = rb_entry(next, struct machine, rb_node); rb_entry 2552 tools/perf/util/machine.c thread = rb_entry(nd, struct thread, rb_node); rb_entry 2579 tools/perf/util/machine.c struct machine *machine = rb_entry(nd, struct machine, rb_node); rb_entry 312 tools/perf/util/map.c struct symbol *sym = rb_entry(nd, struct symbol, rb_node); rb_entry 322 tools/perf/util/map.c struct symbol *sym = rb_entry(nd, struct symbol, rb_node); rb_entry 601 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); rb_entry 615 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node_name); rb_entry 695 tools/perf/util/map.c struct map *pos = rb_entry(nd, struct map, rb_node); rb_entry 747 tools/perf/util/map.c struct map *pos = rb_entry(nd, struct map, rb_node); rb_entry 790 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); rb_entry 803 tools/perf/util/map.c struct map *pos = rb_entry(next, struct map, rb_node); rb_entry 920 tools/perf/util/map.c m = rb_entry(parent, struct map, rb_node); rb_entry 941 tools/perf/util/map.c m = rb_entry(parent, struct map, rb_node_name); rb_entry 986 tools/perf/util/map.c m = rb_entry(p, struct map, rb_node); rb_entry 1006 tools/perf/util/map.c return rb_entry(first, struct map, rb_node); rb_entry 1015 tools/perf/util/map.c return rb_entry(next, struct map, rb_node); rb_entry 25 tools/perf/util/mem2node.c e = rb_entry(parent, struct phys_entry, rb_node); rb_entry 125 tools/perf/util/mem2node.c entry = rb_entry(parent, struct phys_entry, rb_node); rb_entry 3538 tools/perf/util/probe-event.c struct symbol_name_rb_node *pos = rb_entry(nd, struct symbol_name_rb_node, rb_node); rb_entry 66 tools/perf/util/rb_resort.h a = rb_entry(nda, struct __name##_sorted_entry, rb_node); \ rb_entry 67 tools/perf/util/rb_resort.h b = rb_entry(ndb, struct __name##_sorted_entry, rb_node); \ rb_entry 129 tools/perf/util/rb_resort.h __name##_entry = rb_entry(__nd, struct __name##_sorted_entry, \ rb_entry 632 tools/perf/util/srcline.c i = rb_entry(parent, struct srcline_node, rb_node); rb_entry 649 tools/perf/util/srcline.c struct srcline_node *i = rb_entry(n, struct srcline_node, rb_entry 669 tools/perf/util/srcline.c pos = rb_entry(next, struct srcline_node, rb_node); rb_entry 716 tools/perf/util/srcline.c i = rb_entry(parent, struct inline_node, rb_node); rb_entry 733 tools/perf/util/srcline.c struct inline_node *i = rb_entry(n, struct inline_node, rb_entry 753 tools/perf/util/srcline.c pos = rb_entry(next, struct inline_node, rb_node); rb_entry 61 tools/perf/util/strlist.h return rn ? rb_entry(rn, struct str_node, rb_node) : NULL; rb_entry 69 tools/perf/util/strlist.h return rn ? rb_entry(rn, struct str_node, rb_node) : NULL; rb_entry 196 tools/perf/util/symbol.c curr = rb_entry(nd, struct symbol, rb_node); rb_entry 199 tools/perf/util/symbol.c next = rb_entry(nd, struct symbol, rb_node); rb_entry 227 tools/perf/util/symbol.c curr = rb_entry(prevnd, struct symbol, rb_node); rb_entry 231 tools/perf/util/symbol.c curr = rb_entry(nd, struct symbol, rb_node); rb_entry 310 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); rb_entry 339 tools/perf/util/symbol.c s = rb_entry(parent, struct symbol, rb_node); rb_entry 366 tools/perf/util/symbol.c struct symbol *s = rb_entry(n, struct symbol, rb_node); rb_entry 384 tools/perf/util/symbol.c return rb_entry(n, struct symbol, rb_node); rb_entry 394 tools/perf/util/symbol.c return rb_entry(n, struct symbol, rb_node); rb_entry 404 tools/perf/util/symbol.c return rb_entry(n, struct symbol, rb_node); rb_entry 420 tools/perf/util/symbol.c s = rb_entry(parent, struct symbol_name_rb_node, rb_node); rb_entry 438 tools/perf/util/symbol.c struct symbol *pos = rb_entry(nd, struct symbol, rb_node); rb_entry 475 tools/perf/util/symbol.c s = rb_entry(n, struct symbol_name_rb_node, rb_node); rb_entry 494 tools/perf/util/symbol.c tmp = rb_entry(n, struct symbol_name_rb_node, rb_node); rb_entry 552 tools/perf/util/symbol.c return n ? &rb_entry(n, struct symbol_name_rb_node, rb_node)->sym : NULL; rb_entry 722 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); rb_entry 780 tools/perf/util/symbol.c pos = rb_entry(next, struct symbol, rb_node); rb_entry 933 tools/perf/util/symbol.c m = rb_entry(parent, struct module_info, rb_node); rb_entry 949 tools/perf/util/symbol.c mi = rb_entry(next, struct module_info, rb_node); rb_entry 966 tools/perf/util/symbol.c m = rb_entry(n, struct module_info, rb_node); rb_entry 1035 tools/perf/util/symbol.c from_m = rb_entry(from_node, struct module_info, rb_node); rb_entry 1036 tools/perf/util/symbol.c to_m = rb_entry(to_node, struct module_info, rb_node); rb_entry 1798 tools/perf/util/symbol.c map = rb_entry(node, struct map, rb_node_name); rb_entry 71 tools/perf/util/symbol.h nd && (pos = rb_entry(nd, struct symbol, rb_node)); \ rb_entry 70 tools/perf/util/symbol_fprintf.c pos = rb_entry(nd, struct symbol_name_rb_node, rb_node);