/linux-4.1.27/tools/perf/tests/ |
D | vmlinux-kallsyms.c | 110 for (nd = rb_first(&vmlinux_map->dso->symbols[type]); nd; nd = rb_next(nd)) { in test__vmlinux_matches_kallsyms() 187 for (nd = rb_first(&vmlinux.kmaps.maps[type]); nd; nd = rb_next(nd)) { in test__vmlinux_matches_kallsyms() 207 for (nd = rb_first(&vmlinux.kmaps.maps[type]); nd; nd = rb_next(nd)) { in test__vmlinux_matches_kallsyms() 231 for (nd = rb_first(&kallsyms.kmaps.maps[type]); in test__vmlinux_matches_kallsyms()
|
D | hists_output.c | 104 node = rb_first(root_out); in del_hist_entries() 163 node = rb_first(root); in test1() 263 node = rb_first(root); in test2() 317 node = rb_first(root); in test3() 395 node = rb_first(root); in test4() 498 node = rb_first(root); in test5()
|
D | hists_common.c | 165 node = rb_first(root); in print_hists_in() 192 node = rb_first(root); in print_hists_out()
|
D | hists_cumulate.c | 138 node = rb_first(root_out); in del_hist_entries() 200 for (node = rb_first(root), i = 0; in do_test() 220 cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); in do_test()
|
D | hists_link.c | 161 node = rb_first(root); in __validate_match() 213 node = rb_first(root); in __validate_link()
|
/linux-4.1.27/net/ceph/ |
D | debugfs.c | 68 for (n = rb_first(&map->pg_pools); n; n = rb_next(n)) { in osdmap_show() 87 for (n = rb_first(&map->pg_temp); n; n = rb_next(n)) { in osdmap_show() 98 for (n = rb_first(&map->primary_temp); n; n = rb_next(n)) { in osdmap_show() 125 for (rp = rb_first(&monc->generic_request_tree); rp; rp = rb_next(rp)) { in monc_show() 148 for (p = rb_first(&osdc->requests); p; p = rb_next(p)) { in osdc_show()
|
D | osdmap.c | 522 for (rbp = rb_first(&map->pg_pools); rbp; rbp = rb_next(rbp)) { in ceph_pg_poolid_by_name() 661 rb_entry(rb_first(&map->pg_temp), in ceph_osdmap_destroy() 668 rb_entry(rb_first(&map->primary_temp), in ceph_osdmap_destroy() 675 rb_entry(rb_first(&map->pg_pools), in ceph_osdmap_destroy()
|
D | auth_x.c | 650 while ((p = rb_first(&xi->ticket_handlers)) != NULL) { in ceph_x_destroy()
|
D | osd_client.c | 1077 struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds), in remove_all_osds() 1938 for (p = rb_first(&osdc->osds); p; p = n) { in reset_changed_osds() 1969 for (p = rb_first(&osdc->requests); p; ) { in kick_requests()
|
D | mon_client.c | 715 for (p = rb_first(&monc->generic_request_tree); p; p = rb_next(p)) { in __resend_generic_request()
|
/linux-4.1.27/tools/perf/util/ |
D | map.c | 232 struct rb_node *nd = rb_first(symbols); in map__fixup_start() 429 struct rb_node *next = rb_first(maps); in maps__delete() 502 struct rb_node *next = rb_first(root); in map_groups__flush() 543 for (nd = rb_first(&mg->maps[type]); nd; nd = rb_next(nd)) { in map_groups__find_symbol_by_name() 580 for (nd = rb_first(&mg->maps[type]); nd; nd = rb_next(nd)) { in __map_groups__fprintf_maps() 638 struct rb_node *next = rb_first(root); in map_groups__fixup_overlappings() 709 for (nd = rb_first(&parent->maps[type]); nd; nd = rb_next(nd)) { in map_groups__clone() 766 struct rb_node *first = rb_first(maps); in maps__first()
|
D | rblist.c | 107 struct rb_node *pos, *next = rb_first(&rblist->entries); in rblist__delete() 122 for (node = rb_first(&rblist->entries); node; node = rb_next(node)) { in rblist__entry()
|
D | symbol.c | 151 nd = rb_first(symbols); in symbols__fixup_duplicate() 179 struct rb_node *nd, *prevnd = rb_first(symbols); in symbols__fixup_end() 203 struct rb_node *nd, *prevnd = rb_first(&mg->maps[type]); in __map_groups__fixup_end() 288 struct rb_node *next = rb_first(symbols); in symbols__delete() 342 struct rb_node *n = rb_first(symbols); in symbols__first() 390 for (nd = rb_first(source); nd; nd = rb_next(nd)) { in symbols__sort_by_name() 485 for (nd = rb_first(&dso->symbol_names[type]); nd; nd = rb_next(nd)) { in dso__fprintf_symbols_by_name() 638 struct rb_node *next = rb_first(root); in dso__split_kallsyms_for_kcore() 694 struct rb_node *next = rb_first(root); in dso__split_kallsyms() 863 struct rb_node *next = rb_first(modules); in delete_modules() [all …]
|
D | callchain.c | 301 n = rb_first(&node->rb_root_in); in __sort_chain_flat() 331 n = rb_first(&node->rb_root_in); in __sort_chain_graph_abs() 362 n = rb_first(&node->rb_root_in); in __sort_chain_graph_rel() 425 n = rb_first(&new->rb_root_in); in create_child() 711 n = rb_first(&src->rb_root_in); in merge_chain_branch() 864 n = rb_first(&node->rb_root_in); in free_callchain_node()
|
D | hist.c | 164 struct rb_node *next = rb_first(&hists->entries); in hists__output_recalc_col_len() 263 struct rb_node *next = rb_first(&hists->entries); in hists__decay_entries() 279 struct rb_node *next = rb_first(&hists->entries); in hists__delete_entries() 1032 next = rb_first(root); in hists__collapse_resort() 1140 next = rb_first(root); in hists__output_resort() 1201 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_dso() 1235 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_thread() 1267 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_symbol() 1378 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in hists__match() 1403 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in hists__link()
|
D | build-id.c | 226 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__write_buildid_table() 265 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in dsos__hit_all() 511 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__cache_build_ids() 532 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__read_build_ids()
|
D | machine.c | 94 struct rb_node *nd = rb_first(&machine->threads); in machine__delete_threads() 173 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_symbol_filter() 186 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_comm_exec() 258 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__process_guests() 285 for (node = rb_first(&machines->guests); node; node = rb_next(node)) { in machines__set_id_hdr_size() 525 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos() 547 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos_buildid() 578 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__fprintf() 772 struct rb_node *next = rb_first(&machines->guests); in machines__destroy_kernel_maps() 1757 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__for_each_thread()
|
D | strlist.h | 47 struct rb_node *rn = rb_first(&slist->rblist.entries); in strlist__first()
|
D | intlist.h | 47 struct rb_node *rn = rb_first(&ilist->rblist.entries); in intlist__first()
|
D | symbol.h | 71 for (nd = rb_first(symbols); \
|
D | dso.c | 499 struct rb_node *next = rb_first(root); in dso_cache__free() 1122 for (nd = rb_first(&dso->symbols[type]); nd; nd = rb_next(nd)) { in dso__fprintf()
|
D | annotate.c | 1154 node = rb_first(src_root); in resort_source_line() 1261 node = rb_first(root); in print_summary()
|
D | event.c | 348 for (nd = rb_first(&kmaps->maps[MAP__FUNCTION]); in perf_event__synthesize_modules()
|
D | probe-event.c | 174 for (nd = rb_first(&grp->maps[MAP__FUNCTION]); nd; nd = rb_next(nd)) { in kernel_get_module_map()
|
/linux-4.1.27/fs/proc/ |
D | task_nommu.c | 26 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_mem() 87 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_vsize() 105 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_statm() 244 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) in m_start()
|
D | nommu.c | 92 for (p = rb_first(&nommu_region_tree); p; p = rb_next(p)) in nommu_region_list_start()
|
D | generic.c | 43 return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry, in pde_subdir_first()
|
D | proc_sysctl.c | 367 ctl_node = first_usable_entry(rb_first(&dir->root)); in first_entry()
|
/linux-4.1.27/drivers/base/regmap/ |
D | regcache-rbtree.c | 153 for (node = rb_first(&rbtree_ctx->root); node != NULL; in rbtree_show() 240 next = rb_first(&rbtree_ctx->root); in regcache_rbtree_exit() 416 for (node = rb_first(&rbtree_ctx->root); node; in regcache_rbtree_write() 467 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_sync() 507 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_drop()
|
D | regmap-debugfs.c | 540 next = rb_first(&map->range_tree); in regmap_debugfs_init()
|
D | regmap.c | 419 next = rb_first(&map->range_tree); in regmap_range_exit()
|
/linux-4.1.27/tools/perf/ui/browsers/ |
D | hists.c | 57 for (nd = rb_first(&hists->entries); in hist_browser__get_folding() 164 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__count_rows_rb_tree() 206 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__count_rows() 228 struct rb_node *nd = rb_first(&node->rb_root); in callchain_node__init_have_children_rb_tree() 230 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__init_have_children_rb_tree() 267 struct rb_node *nd = rb_first(root); in callchain__init_have_children() 270 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in callchain__init_have_children() 314 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__set_folding_rb_tree() 355 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__set_folding() 381 for (nd = rb_first(&hists->entries); in __hist_browser__set_folding() [all …]
|
D | map.c | 116 for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) { in map__browse()
|
D | annotate.c | 722 nd = rb_first(&browser->entries); in annotate_browser__run()
|
/linux-4.1.27/drivers/iommu/ |
D | iova.c | 369 node = rb_first(&iovad->rbroot); in put_iova_domain() 374 node = rb_first(&iovad->rbroot); in put_iova_domain() 445 for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) { in reserve_iova() 482 for (node = rb_first(&from->rbroot); node; node = rb_next(node)) { in copy_reserved_iova()
|
D | arm-smmu.c | 1814 for (node = rb_first(&smmu->masters); node; node = rb_next(node)) { in arm_smmu_device_dt_probe() 1843 for (node = rb_first(&smmu->masters); node; node = rb_next(node)) { in arm_smmu_device_remove()
|
/linux-4.1.27/tools/perf/ui/gtk/ |
D | hists.c | 96 bool has_single_node = (rb_first(root) == rb_last(root)); in perf_gtk__add_callchain() 98 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in perf_gtk__add_callchain() 223 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in perf_gtk__show_hists()
|
/linux-4.1.27/tools/perf/ui/stdio/ |
D | hist.c | 95 node = rb_first(root); in __callchain__fprintf_graph() 176 node = rb_first(root); in callchain__fprintf_graph() 243 struct rb_node *rb_node = rb_first(tree); in callchain__fprintf_flat() 461 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__fprintf()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | eeh_cache.c | 113 n = rb_first(&cache->rb_root); in eeh_addr_cache_print() 239 n = rb_first(&pci_io_addr_cache_root.rb_root); in __eeh_addr_cache_rmv_dev()
|
/linux-4.1.27/fs/ocfs2/ |
D | reservations.c | 96 node = rb_first(&resmap->m_reservations); in ocfs2_dump_resv() 151 node = rb_first(&resmap->m_reservations); in ocfs2_check_resmap() 378 node = rb_first(&resmap->m_reservations); in ocfs2_find_resv_lhs() 537 next = rb_first(root); in __ocfs2_resv_find_window()
|
/linux-4.1.27/kernel/trace/ |
D | trace_stat.c | 189 node = rb_first(&session->stat_root); in stat_seq_start() 204 return rb_first(&session->stat_root); in stat_seq_next()
|
/linux-4.1.27/include/linux/ |
D | rbtree.h | 68 extern struct rb_node *rb_first(const struct rb_root *);
|
/linux-4.1.27/lib/ |
D | rbtree_test.c | 143 for (rb = rb_first(&root); rb; rb = rb_next(rb)) { in check() 169 for (rb = rb_first(&root); rb; rb = rb_next(rb)) { in check_augmented()
|
D | rbtree.c | 418 struct rb_node *rb_first(const struct rb_root *root) in rb_first() function 429 EXPORT_SYMBOL(rb_first);
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | cm.c | 439 while (rb_first(sl_id_map)) { in mlx4_ib_cm_paravirt_clean() 441 rb_entry(rb_first(sl_id_map), in mlx4_ib_cm_paravirt_clean() 450 nd = rb_first(sl_id_map); in mlx4_ib_cm_paravirt_clean()
|
D | mcg.c | 1087 for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) in _mlx4_ib_mcg_port_cleanup() 1101 while ((p = rb_first(&ctx->mcg_table)) != NULL) { in _mlx4_ib_mcg_port_cleanup() 1232 for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) { in clean_vf_mcast()
|
/linux-4.1.27/fs/logfs/ |
D | gc.c | 241 cand = rb_entry(rb_first(&list->rb_tree), struct gc_candidate, rb_node); in get_best_cand() 352 return rb_entry(rb_first(&list->rb_tree), struct gc_candidate, rb_node); in first_in_list() 597 cand = rb_entry(rb_first(&super->s_free_list.rb_tree), in logfs_journal_wl_pass()
|
/linux-4.1.27/drivers/staging/android/ion/ |
D | ion.c | 689 for (n = rb_first(&client->handles); n; n = rb_next(n)) { in ion_debug_client_show() 727 for (node = rb_first(root); node; node = rb_next(node)) { in ion_get_client_serial() 833 while ((n = rb_first(&client->handles))) { in ion_client_destroy() 1384 for (n = rb_first(&client->handles); n; n = rb_next(n)) { in ion_debug_heap_total() 1406 for (n = rb_first(&dev->clients); n; n = rb_next(n)) { in ion_debug_heap_show() 1427 for (n = rb_first(&dev->buffers); n; n = rb_next(n)) { in ion_debug_heap_show()
|
/linux-4.1.27/tools/perf/ |
D | builtin-kmem.c | 528 next = rb_first(root); in __print_slab_result() 583 struct rb_node *next = rb_first(root); in __print_page_result() 744 node = rb_first(root); in __sort_slab_result() 785 node = rb_first(root); in __sort_page_result()
|
D | builtin-annotate.c | 118 struct rb_node *nd = rb_first(&hists->entries), *next; in hists__find_annotations()
|
D | builtin-diff.c | 434 next = rb_first(root); in hists__baseline_only() 456 next = rb_first(root); in hists__precompute()
|
D | builtin-lock.c | 766 node = rb_first(&thread_stats); in dump_threads()
|
D | builtin-sched.c | 1282 node = rb_first(&sched->atom_root); in perf_sched__sort_lat() 1560 next = rb_first(&sched->sorted_atom_root); in perf_sched__lat()
|
D | builtin-top.c | 352 next = rb_first(&hists->entries); in perf_top__prompt_symbol()
|
D | builtin-kvm.c | 532 struct rb_node *node = rb_first(result); in pop_from_result()
|
/linux-4.1.27/drivers/android/ |
D | binder.c | 1054 for (n = rb_first(&proc->refs_by_desc); n != NULL; n = rb_next(n)) { in binder_get_ref_for_node() 2993 for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) { in binder_deferred_flush() 3084 while ((n = rb_first(&proc->threads))) { in binder_deferred_release() 3094 while ((n = rb_first(&proc->nodes))) { in binder_deferred_release() 3104 while ((n = rb_first(&proc->refs_by_desc))) { in binder_deferred_release() 3116 while ((n = rb_first(&proc->allocated_buffers))) { in binder_deferred_release() 3368 for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) in print_binder_proc() 3371 for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) { in print_binder_proc() 3378 for (n = rb_first(&proc->refs_by_desc); in print_binder_proc() 3384 for (n = rb_first(&proc->allocated_buffers); n != NULL; n = rb_next(n)) in print_binder_proc() [all …]
|
/linux-4.1.27/fs/ext3/ |
D | dir.c | 468 info->curr_node = rb_first(&info->root); in ext3_dx_readdir() 490 info->curr_node = rb_first(&info->root); in ext3_dx_readdir()
|
D | balloc.c | 215 n = rb_first(root); in __rsv_window_dump()
|
/linux-4.1.27/fs/dlm/ |
D | debug_fs.c | 457 for (node = rb_first(tree); node; node = rb_next(node)) { in table_seq_start() 489 node = rb_first(tree); in table_seq_start() 554 next = rb_first(tree); in table_seq_next()
|
D | recover.c | 906 for (n = rb_first(&ls->ls_rsbtbl[i].keep); n; n = rb_next(n)) { in dlm_create_root_list() 942 for (n = rb_first(&ls->ls_rsbtbl[i].toss); n; n = next) { in dlm_clear_toss()
|
D | lockspace.c | 817 while ((n = rb_first(&ls->ls_rsbtbl[i].keep))) { in release_lockspace() 823 while ((n = rb_first(&ls->ls_rsbtbl[i].toss))) { in release_lockspace()
|
D | lock.c | 1094 for (n = rb_first(&ls->ls_rsbtbl[i].keep); n; n = rb_next(n)) { in dlm_dump_rsb_hash() 1670 for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) { in shrink_bucket() 5476 for (n = rb_first(&ls->ls_rsbtbl[bucket].keep); n; n = rb_next(n)) { in find_grant_rsb()
|
/linux-4.1.27/drivers/mtd/ubi/ |
D | wl.c | 317 e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_wl_entry() 358 first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_mean_wl_entry() 711 e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb); 732 e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, u.rb); 982 e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
|
D | ubi.h | 979 for (rb = rb_first(root), \
|
D | fastmap.c | 422 for (node = rb_first(&ai->volumes); node; node = rb_next(node)) { in unmap_peb() 425 for (node2 = rb_first(&av->root); node2; in unmap_peb()
|
D | attach.c | 638 while ((rb = rb_first(&av->root))) { in ubi_remove_av()
|
/linux-4.1.27/fs/jffs2/ |
D | nodelist.h | 329 struct rb_node *node = rb_first(root); in frag_first() 361 #define tn_first(list) rb_entry(rb_first(list), struct jffs2_tmp_dnode_info, rb)
|
/linux-4.1.27/kernel/power/ |
D | wakelock.c | 45 for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) { in pm_show_wakelocks()
|
/linux-4.1.27/fs/btrfs/ |
D | delayed-ref.c | 227 n = rb_first(root); in find_ref_head() 357 node = rb_first(&head->ref_root); in btrfs_merge_delayed_refs() 367 node = rb_first(&head->ref_root); in btrfs_merge_delayed_refs()
|
D | free-space-cache.c | 648 for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { in merge_space_tree() 914 struct rb_node *node = rb_first(&ctl->free_space_offset); in write_cache_extent_entries() 927 node = rb_first(&cluster->root); in write_cache_extent_entries() 949 node = rb_first(&cluster->root); in write_cache_extent_entries() 2028 node = rb_first(&cluster->root); in insert_into_bitmap() 2432 for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { in btrfs_dump_free_space() 2491 node = rb_first(&cluster->root); in __btrfs_return_cluster_to_free_space() 2700 node = rb_first(&cluster->root); in btrfs_alloc_from_cluster() 3354 entry = rb_entry(rb_first(&ctl->free_space_offset), in btrfs_find_ino_for_alloc()
|
D | inode-map.c | 261 n = rb_first(rbroot); in btrfs_unpin_free_ino()
|
D | relocation.c | 2436 while ((rb_node = rb_first(blocks))) { 2965 rb_node = rb_first(blocks); 2973 rb_node = rb_first(blocks); 2984 rb_node = rb_first(blocks); 2998 if (ret != -EAGAIN || rb_node == rb_first(blocks))
|
D | delayed-inode.c | 509 p = rb_first(&delayed_node->ins_root); in __btrfs_first_delayed_insertion_item() 522 p = rb_first(&delayed_node->del_root); in __btrfs_first_delayed_deletion_item()
|
D | qgroup.c | 463 while ((n = rb_first(&fs_info->qgroup_tree))) { in btrfs_free_qgroup_config() 2888 for (n = rb_first(&fs_info->qgroup_tree); n; n = rb_next(n)) { in qgroup_rescan_zero_tracking()
|
D | file.c | 270 node = rb_first(&fs_info->defrag_inodes); in btrfs_cleanup_defrag_inodes() 278 node = rb_first(&fs_info->defrag_inodes); in btrfs_cleanup_defrag_inodes()
|
D | extent-tree.c | 2337 node = rb_first(&head->ref_root); in select_delayed_ref() 2485 if (rb_first(&locked_ref->ref_root) || in __btrfs_run_delayed_refs() 2590 n = rb_first(root); in find_middle() 2830 node = rb_first(&delayed_refs->href_root); in btrfs_run_delayed_refs() 2937 node = rb_first(&head->ref_root); in check_delayed_ref() 6402 if (rb_first(&head->ref_root)) in check_ref_cleanup()
|
D | backref.c | 586 n = rb_first(&head->ref_root); in __add_delayed_refs()
|
D | send.c | 5912 n = rb_first(&sctx->pending_dir_moves); in btrfs_ioctl_send() 5929 n = rb_first(&sctx->waiting_dir_moves); in btrfs_ioctl_send() 5940 n = rb_first(&sctx->orphan_dirs); in btrfs_ioctl_send()
|
D | disk-io.c | 4092 while ((node = rb_first(&delayed_refs->href_root)) != NULL) { in btrfs_destroy_delayed_refs() 4109 while ((node = rb_first(&head->ref_root)) != NULL) { in btrfs_destroy_delayed_refs()
|
D | transaction.c | 89 node = rb_first(&tree->state); in clear_btree_io_tree()
|
D | inode.c | 2634 node = rb_first(&new->root); in relink_file_extents() 5025 node = rb_first(&map_tree->map); in evict_inode_truncate_pages() 5044 node = rb_first(&io_tree->state); in evict_inode_truncate_pages()
|
D | ctree.c | 425 for (node = rb_first(tm_root); node; node = next) { in btrfs_put_tree_mod_seq()
|
/linux-4.1.27/net/sched/ |
D | sch_fq.c | 412 while ((p = rb_first(&q->delayed)) != NULL) { in fq_check_throttled() 532 while ((p = rb_first(root)) != NULL) { in fq_reset() 562 while ((op = rb_first(oroot)) != NULL) { in fq_rehash()
|
D | sch_netem.c | 367 while ((p = rb_first(&q->t_root))) { in tfifo_reset() 530 struct rb_node *p = rb_first(&q->t_root); in netem_drop() 569 p = rb_first(&q->t_root); in netem_dequeue()
|
D | sch_hfsc.c | 232 for (n = rb_first(&q->eligible); n != NULL; n = rb_next(n)) { in eltree_get_mindl() 248 n = rb_first(&q->eligible); in eltree_get_minel() 296 for (n = rb_first(&cl->vt_tree); n != NULL; n = rb_next(n)) { in vttree_firstfit() 661 struct rb_node *n = rb_first(&cl->cf_tree); in update_cfmin()
|
D | sch_htb.c | 702 struct rb_node *p = rb_first(wait_pq); in htb_do_events()
|
/linux-4.1.27/drivers/mtd/ |
D | mtdswap.c | 92 #define MTDSWAP_ECNT_MIN(rbroot) (rb_entry(rb_first(rbroot), struct swap_eb, \ 256 p = rb_first(root); in mtdswap_rb_index() 625 eb = rb_entry(rb_first(clean_root), struct swap_eb, rb); in mtdswap_map_free_block() 910 eb = rb_entry(rb_first(rp), struct swap_eb, rb); in mtdswap_pick_gc_eblk() 1238 min[i] = rb_entry(rb_first(root), struct swap_eb, in mtdswap_show()
|
/linux-4.1.27/fs/ |
D | eventpoll.c | 756 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_free() 772 while ((rbp = rb_first(&ep->rbr)) != NULL) { in ep_free() 879 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_show_fdinfo() 1691 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_loop_check_proc()
|
/linux-4.1.27/fs/ext4/ |
D | block_validity.c | 128 node = rb_first(&sbi->system_blks); in debug_print_tree()
|
D | dir.c | 550 info->curr_node = rb_first(&info->root); in ext4_dx_readdir() 572 info->curr_node = rb_first(&info->root); in ext4_dx_readdir()
|
D | extents_status.c | 182 node = rb_first(&tree->root); in ext4_es_print_tree()
|
D | mballoc.c | 3437 n = rb_first(&(grp->bb_free_root)); in ext4_mb_generate_from_freelist()
|
/linux-4.1.27/arch/x86/mm/ |
D | pat_rbtree.c | 235 node = rb_first(&memtype_rbroot); in rbt_memtype_copy_nth_element()
|
/linux-4.1.27/security/keys/ |
D | gc.c | 216 cursor = rb_first(&key_serial_tree); in key_garbage_collector()
|
D | proc.c | 290 struct rb_node *n = rb_first(r); in key_user_first()
|
/linux-4.1.27/tools/perf/ui/ |
D | browser.c | 117 nd = rb_first(root); in ui_browser__rb_tree_seek() 146 browser->top = rb_first(browser->entries); in ui_browser__rb_tree_refresh()
|
/linux-4.1.27/net/netfilter/ |
D | nft_rbtree.c | 184 for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) { in nft_rbtree_walk()
|
D | xt_connlimit.c | 405 while ((node = rb_first(r)) != NULL) { in destroy_tree()
|
/linux-4.1.27/fs/ceph/ |
D | caps.c | 347 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_get_cap_mds() 675 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued() 706 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued_other() 756 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued_mask() 782 for (q = rb_first(&ci->i_caps); q != p; in __ceph_caps_issued_mask() 807 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_revoking_other() 867 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_mds_wanted() 1103 p = rb_first(&ci->i_caps); in ceph_queue_caps_release() 1631 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in ceph_check_caps()
|
D | debugfs.c | 60 for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) { in mdsc_show()
|
D | xattr.c | 518 p = rb_first(&ci->i_xattrs.index); in __copy_xattr_names() 541 p = rb_first(&ci->i_xattrs.index); in __ceph_destroy_xattrs() 687 p = rb_first(&ci->i_xattrs.index); in __ceph_build_xattrs_blob()
|
D | mds_client.c | 793 cap = rb_entry(rb_first(&ci->i_caps), struct ceph_cap, ci_node); in __choose_mds() 1040 p = rb_first(&mdsc->request_tree); in cleanup_session_requests() 1696 return rb_entry(rb_first(&mdsc->request_tree), in __get_oldest_req() 2205 struct rb_node *p = rb_first(&mdsc->request_tree); in kick_requests() 2699 p = rb_first(&mdsc->request_tree); in replay_unsafe_requests() 2930 for (p = rb_first(&mdsc->snap_realms); p; p = rb_next(p)) { in send_mds_reconnect()
|
D | inode.c | 322 rb_node = rb_first(&ci->i_fragtree); in ceph_fill_fragtree() 336 rb_node = rb_first(&ci->i_fragtree); in ceph_fill_fragtree() 509 while ((n = rb_first(&ci->i_fragtree)) != NULL) { in ceph_destroy_inode()
|
/linux-4.1.27/Documentation/ |
D | rbtree.txt | 171 struct rb_node *rb_first(struct rb_root *tree); 176 To start iterating, call rb_first() or rb_last() with a pointer to the root 190 for (node = rb_first(&mytree); node; node = rb_next(node))
|
/linux-4.1.27/net/rds/ |
D | cong.c | 397 while ((node = rb_first(&rds_cong_tree))) { in rds_cong_exit()
|
D | rdma.c | 137 while ((node = rb_first(&rs->rs_rdma_keys))) { in rds_rdma_drop_keys()
|
/linux-4.1.27/net/rxrpc/ |
D | ar-connevent.c | 38 for (p = rb_first(&conn->calls); p; p = rb_next(p)) { in rxrpc_abort_calls()
|
D | ar-call.c | 790 for (p = rb_first(&rx->calls); p; p = rb_next(p)) { in rxrpc_release_calls_on_socket()
|
/linux-4.1.27/fs/afs/ |
D | callback.c | 384 vnode = rb_entry(rb_first(&server->cb_promises),
|
/linux-4.1.27/drivers/vfio/ |
D | vfio_iommu_type1.c | 650 n = rb_first(&iommu->dma_list); in vfio_iommu_replay() 845 while ((node = rb_first(&iommu->dma_list))) in vfio_iommu_unmap_unpin_all()
|
/linux-4.1.27/fs/gfs2/ |
D | rgrp.c | 547 n = rb_first(&sdp->sd_rindex_tree); in gfs2_rgrpd_get_first() 569 n = rb_first(&sdp->sd_rindex_tree); in gfs2_rgrpd_get_next() 712 while ((n = rb_first(&rgd->rd_rstree))) { in return_all_reservations() 725 while ((n = rb_first(&sdp->sd_rindex_tree))) { in gfs2_clear_rgrpd() 2216 for (n = rb_first(&rgd->rd_rstree); n; n = rb_next(&trs->rs_node)) { in gfs2_rgrp_dump()
|
/linux-4.1.27/fs/nfs/ |
D | nfs4state.c | 177 for (pos = rb_first(&server->state_owners); in nfs4_get_renew_cred_server_locked() 1560 for (pos = rb_first(&server->state_owners); in nfs4_reset_seqids() 1610 for (pos = rb_first(&server->state_owners); in nfs4_clear_reclaim_server() 1721 for (pos = rb_first(&server->state_owners); in nfs4_do_reclaim()
|
D | dir.c | 2195 while ((n = rb_first(root_node)) != NULL) { in __nfs_access_zap_cache()
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx5/core/ |
D | pagealloc.c | 488 p = rb_first(&dev->priv.page_root); in mlx5_reclaim_startup_pages()
|
/linux-4.1.27/fs/nfs/blocklayout/ |
D | extent_tree.c | 20 struct rb_node *node = rb_first(root); in ext_tree_first()
|
/linux-4.1.27/fs/ubifs/ |
D | log.c | 307 p = rb_first(&c->buds); in remove_buds()
|
D | debug.c | 622 for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_budg() 710 for (rb = rb_first((struct rb_root *)&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_lprop() 2151 struct rb_node *this = rb_first(&fsckd->inodes); in check_inodes()
|
D | recovery.c | 1477 struct rb_node *this = rb_first(&c->size_tree); in ubifs_recover_size()
|
/linux-4.1.27/drivers/md/bcache/ |
D | util.h | 549 container_of_or_null(rb_first(root), type, member)
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | trace.c | 128 for (n = rb_first(&mm->mm_rb); n; n = rb_next(n)) { in decode_address()
|
/linux-4.1.27/net/802/ |
D | garp.c | 388 for (node = rb_first(&app->gid); in garp_gid_event()
|
D | mrp.c | 574 for (node = rb_first(&app->mad); in mrp_mad_event()
|
/linux-4.1.27/mm/ |
D | nommu.c | 592 lastp = rb_first(&nommu_region_tree); in validate_nommu_regions() 1348 for (rb = rb_first(&nommu_region_tree); rb; rb = rb_next(rb)) { in do_mmap_pgoff()
|
D | ksm.c | 2000 node = rb_first(root_stable_tree + nid); in ksm_check_stable_tree() 2010 node = rb_first(root_stable_tree + nid); in ksm_check_stable_tree()
|
D | mmap.c | 383 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in browse_rb() 426 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in validate_mm_rb()
|
D | mempolicy.c | 2504 next = rb_first(&p->root); in mpol_free_shared_policy()
|
/linux-4.1.27/block/ |
D | cfq-iosched.c | 1171 root->left = rb_first(&root->rb); in cfq_rb_first() 1182 root->left = rb_first(&root->rb); in cfq_rb_first_group() 1223 rbnext = rb_first(&cfqq->sort_list); in cfq_find_next_rq() 3382 BUG_ON(rb_first(&cfqq->sort_list)); in cfq_put_queue()
|
D | blk-throttle.c | 560 parent_sq->first_pending = rb_first(&parent_sq->pending_tree); in throtl_rb_first()
|
/linux-4.1.27/drivers/infiniband/core/ |
D | multicast.c | 762 for (node = rb_first(&port->table); node; node = rb_next(node)) { in mcast_groups_event()
|
/linux-4.1.27/fs/f2fs/ |
D | data.c | 551 node = rb_first(&et->root); in __free_extent_tree() 745 struct rb_node *node = rb_first(&et->root); in f2fs_preserve_extent_tree()
|
/linux-4.1.27/fs/kernfs/ |
D | dir.c | 950 rbn = rb_first(&pos->dir.children); in kernfs_leftmost_descendant()
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_multicast.c | 961 n = rb_first(&priv->multicast_tree); in ipoib_mcast_iter_next()
|
D | ipoib_main.c | 358 n = rb_first(&priv->path_tree); in ipoib_path_iter_next()
|
/linux-4.1.27/fs/ext2/ |
D | balloc.c | 207 n = rb_first(root); in __rsv_window_dump()
|
/linux-4.1.27/fs/cifs/ |
D | connect.c | 3818 while ((node = rb_first(root))) { in cifs_umount() 4102 struct rb_node *node = rb_first(root); in cifs_prune_tlinks() 4114 node = rb_first(root); in cifs_prune_tlinks()
|
/linux-4.1.27/drivers/block/xen-blkback/ |
D | blkback.c | 177 for ((pos) = container_of(rb_first((rbtree)), typeof(*(pos)), node), \
|
/linux-4.1.27/drivers/block/ |
D | pktcdvd.c | 1216 n = rb_first(&pd->bio_queue); in pkt_handle_queue() 1234 n = rb_first(&pd->bio_queue); in pkt_handle_queue()
|
/linux-4.1.27/drivers/md/ |
D | dm-crypt.c | 1224 io = crypt_io_from_node(rb_first(&write_tree)); in dmcrypt_write()
|
D | dm-thin.c | 1741 for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) { in __extract_sorted_bios()
|
/linux-4.1.27/fs/fuse/ |
D | dev.c | 2151 p = rb_first(&fc->polled_files); in end_polls()
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
D | osc_cache.c | 169 return rb_extent(rb_first(&obj->oo_root)); in first_extent()
|