rb_entry_safe 286 drivers/gpu/drm/drm_mm.c return rb_entry_safe(rb, struct drm_mm_node, rb_hole_size); rb_entry_safe 291 drivers/gpu/drm/drm_mm.c return rb_entry_safe(rb, struct drm_mm_node, rb_hole_addr); rb_entry_safe 1806 drivers/gpu/drm/i915/i915_drv.h rb_entry_safe(rb, struct intel_engine_cs, uabi_node) rb_entry_safe 146 fs/f2fs/extent_cache.c *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry_safe 151 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry_safe 158 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry_safe 163 fs/f2fs/extent_cache.c *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); rb_entry_safe 578 fs/f2fs/extent_cache.c next_en = rb_entry_safe(node, struct extent_node, rb_entry_safe 1379 fs/f2fs/segment.c next_dc = rb_entry_safe(node, struct discard_cmd, rb_node); rb_entry_safe 1448 fs/f2fs/segment.c dc = rb_entry_safe(node, struct discard_cmd, rb_node); rb_entry_safe 2803 fs/f2fs/segment.c dc = rb_entry_safe(node, struct discard_cmd, rb_node); rb_entry_safe 58 fs/proc/generic.c return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry, rb_entry_safe 64 fs/proc/generic.c return rb_entry_safe(rb_next(&dir->subdir_node), struct proc_dir_entry, rb_entry_safe 110 include/linux/rbtree.h for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \ rb_entry_safe 111 include/linux/rbtree.h pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \ rb_entry_safe 3398 include/linux/skbuff.h #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode) rb_entry_safe 1649 kernel/events/core.c next = rb_entry_safe(rb_next(&event->group_node), typeof(*event), group_node); rb_entry_safe 1660 kernel/events/core.c for (event = rb_entry_safe(rb_first(&((groups)->tree)), \ rb_entry_safe 1662 kernel/events/core.c event = rb_entry_safe(rb_next(&event->group_node), \ rb_entry_safe 3799 kernel/events/core.c event = rb_entry_safe(rb_first(&ctx->flexible_groups.tree), rb_entry_safe 385 mm/vmalloc.c va = rb_entry_safe(node, struct vmap_area, rb_node); rb_entry_safe 3137 mm/vmalloc.c return rb_entry_safe(n, struct vmap_area, rb_node); rb_entry_safe 98 tools/include/linux/rbtree.h for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \ rb_entry_safe 99 tools/include/linux/rbtree.h pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \