nodes 703 arch/arm/mach-sunxi/mc_smp.c int (*get_smp_nodes)(struct sunxi_mc_smp_nodes *nodes); nodes 707 arch/arm/mach-sunxi/mc_smp.c static void __init sunxi_mc_smp_put_nodes(struct sunxi_mc_smp_nodes *nodes) nodes 709 arch/arm/mach-sunxi/mc_smp.c of_node_put(nodes->prcm_node); nodes 710 arch/arm/mach-sunxi/mc_smp.c of_node_put(nodes->cpucfg_node); nodes 711 arch/arm/mach-sunxi/mc_smp.c of_node_put(nodes->sram_node); nodes 712 arch/arm/mach-sunxi/mc_smp.c of_node_put(nodes->r_cpucfg_node); nodes 713 arch/arm/mach-sunxi/mc_smp.c memset(nodes, 0, sizeof(*nodes)); nodes 716 arch/arm/mach-sunxi/mc_smp.c static int __init sun9i_a80_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) nodes 718 arch/arm/mach-sunxi/mc_smp.c nodes->prcm_node = of_find_compatible_node(NULL, NULL, nodes 720 arch/arm/mach-sunxi/mc_smp.c if (!nodes->prcm_node) { nodes 725 arch/arm/mach-sunxi/mc_smp.c nodes->cpucfg_node = of_find_compatible_node(NULL, NULL, nodes 727 arch/arm/mach-sunxi/mc_smp.c if (!nodes->cpucfg_node) { nodes 732 arch/arm/mach-sunxi/mc_smp.c nodes->sram_node = of_find_compatible_node(NULL, NULL, nodes 734 arch/arm/mach-sunxi/mc_smp.c if (!nodes->sram_node) { nodes 742 arch/arm/mach-sunxi/mc_smp.c static int __init sun8i_a83t_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) nodes 744 arch/arm/mach-sunxi/mc_smp.c nodes->prcm_node = of_find_compatible_node(NULL, NULL, nodes 746 arch/arm/mach-sunxi/mc_smp.c if (!nodes->prcm_node) { nodes 751 arch/arm/mach-sunxi/mc_smp.c nodes->cpucfg_node = of_find_compatible_node(NULL, NULL, nodes 753 arch/arm/mach-sunxi/mc_smp.c if (!nodes->cpucfg_node) { nodes 758 arch/arm/mach-sunxi/mc_smp.c nodes->r_cpucfg_node = of_find_compatible_node(NULL, NULL, nodes 760 arch/arm/mach-sunxi/mc_smp.c if (!nodes->r_cpucfg_node) { nodes 782 arch/arm/mach-sunxi/mc_smp.c struct sunxi_mc_smp_nodes nodes = { 0 }; nodes 826 arch/arm/mach-sunxi/mc_smp.c ret = sunxi_mc_smp_data[i].get_smp_nodes(&nodes); nodes 834 arch/arm/mach-sunxi/mc_smp.c prcm_base = of_iomap(nodes.prcm_node, 0); nodes 841 arch/arm/mach-sunxi/mc_smp.c cpucfg_base = of_io_request_and_map(nodes.cpucfg_node, 0, nodes 851 arch/arm/mach-sunxi/mc_smp.c r_cpucfg_base = of_io_request_and_map(nodes.r_cpucfg_node, nodes 860 arch/arm/mach-sunxi/mc_smp.c sram_b_smp_base = of_io_request_and_map(nodes.sram_node, 0, nodes 878 arch/arm/mach-sunxi/mc_smp.c sunxi_mc_smp_put_nodes(&nodes); nodes 897 arch/arm/mach-sunxi/mc_smp.c of_address_to_resource(nodes.r_cpucfg_node, 0, &res); nodes 900 arch/arm/mach-sunxi/mc_smp.c of_address_to_resource(nodes.sram_node, 0, &res); nodes 905 arch/arm/mach-sunxi/mc_smp.c of_address_to_resource(nodes.cpucfg_node, 0, &res); nodes 910 arch/arm/mach-sunxi/mc_smp.c sunxi_mc_smp_put_nodes(&nodes); nodes 447 arch/powerpc/oprofile/cell/spu_task_sync.c int nodes = 0; nodes 450 arch/powerpc/oprofile/cell/spu_task_sync.c if (tmp > nodes) nodes 451 arch/powerpc/oprofile/cell/spu_task_sync.c nodes++; nodes 453 arch/powerpc/oprofile/cell/spu_task_sync.c return nodes; nodes 291 arch/s390/numa/mode_emu.c static struct toptree *toptree_new(int id, int nodes) nodes 299 arch/s390/numa/mode_emu.c for (nid = 0; nid < nodes; nid++) { nodes 491 arch/s390/numa/mode_emu.c static int emu_setup_nodes_adjust(int nodes) nodes 497 arch/s390/numa/mode_emu.c if (nodes_max >= nodes) nodes 498 arch/s390/numa/mode_emu.c return nodes; nodes 499 arch/s390/numa/mode_emu.c pr_warn("Not enough memory for %d nodes, reducing node count\n", nodes); nodes 53 arch/sparc/kernel/cpumap.c struct cpuinfo_node nodes[0]; nodes 197 arch/sparc/kernel/cpumap.c new_tree = kzalloc(struct_size(new_tree, nodes, n), GFP_ATOMIC); nodes 211 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[n]; nodes 252 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[level_rover[level]]; nodes 277 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[n]; nodes 299 arch/sparc/kernel/cpumap.c struct cpuinfo_node *node = &t->nodes[node_index]; nodes 302 arch/sparc/kernel/cpumap.c top_level = t->nodes[root_index].level; nodes 314 arch/sparc/kernel/cpumap.c node = &t->nodes[node->parent_index]; nodes 340 arch/sparc/kernel/cpumap.c for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; nodes 342 arch/sparc/kernel/cpumap.c new_index = t->nodes[index].rover; nodes 368 arch/sparc/kernel/cpumap.c for (i = 0; i < cpuinfo_tree->nodes[0].num_cpus; i++) nodes 404 arch/sparc/kernel/cpumap.c root_node = &cpuinfo_tree->nodes[0]; nodes 830 arch/x86/events/amd/ibs.c int nodes; nodes 833 arch/x86/events/amd/ibs.c nodes = 0; nodes 841 arch/x86/events/amd/ibs.c ++nodes; nodes 853 arch/x86/events/amd/ibs.c if (!nodes) { nodes 3028 arch/x86/include/asm/uv/uv_mmrs.h unsigned long nodes:64; /* RW */ nodes 173 arch/x86/kernel/apic/apic_numachip.c u32 nodes = 1; nodes 180 arch/x86/kernel/apic/apic_numachip.c nodes = ((val >> 3) & 7) + 1; nodes 183 arch/x86/kernel/apic/apic_numachip.c c->phys_proc_id = node / nodes; nodes 1204 drivers/acpi/acpica/aclocal.h u32 nodes; nodes 654 drivers/acpi/acpica/dbnames.c info->nodes++; nodes 724 drivers/acpi/acpica/dbnames.c info.nodes, info.objects); nodes 466 drivers/android/binder.c struct rb_root nodes; nodes 1067 drivers/android/binder.c struct rb_node *n = proc->nodes.rb_node; nodes 1108 drivers/android/binder.c struct rb_node **p = &proc->nodes.rb_node; nodes 1140 drivers/android/binder.c rb_insert_color(&node->rb_node, &proc->nodes); nodes 1280 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); nodes 4332 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); nodes 4983 drivers/android/binder.c for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) { nodes 5413 drivers/android/binder.c int threads, nodes, incoming_refs, outgoing_refs, active_transactions; nodes 5454 drivers/android/binder.c nodes = 0; nodes 5456 drivers/android/binder.c while ((n = rb_first(&proc->nodes))) { nodes 5460 drivers/android/binder.c nodes++; nodes 5467 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); nodes 5493 drivers/android/binder.c __func__, proc->pid, threads, nodes, incoming_refs, nodes 5724 drivers/android/binder.c for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) { nodes 5898 drivers/android/binder.c for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) nodes 140 drivers/base/regmap/regcache-rbtree.c int nodes = 0; nodes 159 drivers/base/regmap/regcache-rbtree.c nodes++; nodes 163 drivers/base/regmap/regcache-rbtree.c if (nodes) nodes 164 drivers/base/regmap/regcache-rbtree.c average = registers / nodes; nodes 169 drivers/base/regmap/regcache-rbtree.c nodes, registers, average, mem_size); nodes 764 drivers/base/swnode.c int software_node_register_nodes(const struct software_node *nodes) nodes 769 drivers/base/swnode.c for (i = 0; nodes[i].name; i++) { nodes 770 drivers/base/swnode.c ret = software_node_register(&nodes[i]); nodes 772 drivers/base/swnode.c software_node_unregister_nodes(nodes); nodes 787 drivers/base/swnode.c void software_node_unregister_nodes(const struct software_node *nodes) nodes 792 drivers/base/swnode.c for (i = 0; nodes[i].name; i++) { nodes 793 drivers/base/swnode.c swnode = software_node_to_swnode(&nodes[i]); nodes 113 drivers/clk/zynqmp/clk-gate-zynqmp.c const struct clock_topology *nodes) nodes 127 drivers/clk/zynqmp/clk-gate-zynqmp.c init.flags = nodes->flag; nodes 132 drivers/clk/zynqmp/clk-gate-zynqmp.c gate->flags = nodes->type_flag; nodes 109 drivers/clk/zynqmp/clk-mux-zynqmp.c const struct clock_topology *nodes) nodes 121 drivers/clk/zynqmp/clk-mux-zynqmp.c if (nodes->type_flag & CLK_MUX_READ_ONLY) nodes 125 drivers/clk/zynqmp/clk-mux-zynqmp.c init.flags = nodes->flag; nodes 128 drivers/clk/zynqmp/clk-mux-zynqmp.c mux->flags = nodes->type_flag; nodes 38 drivers/clk/zynqmp/clk-zynqmp.h const struct clock_topology *nodes); nodes 43 drivers/clk/zynqmp/clk-zynqmp.h const struct clock_topology *nodes); nodes 49 drivers/clk/zynqmp/clk-zynqmp.h const struct clock_topology *nodes); nodes 54 drivers/clk/zynqmp/clk-zynqmp.h const struct clock_topology *nodes); nodes 60 drivers/clk/zynqmp/clk-zynqmp.h const struct clock_topology *nodes); nodes 123 drivers/clk/zynqmp/clkc.c const struct clock_topology *nodes) nodes 288 drivers/clk/zynqmp/clkc.c const struct clock_topology *nodes) nodes 308 drivers/clk/zynqmp/clkc.c nodes->flag, mult, nodes 562 drivers/clk/zynqmp/clkc.c struct clock_topology *nodes; nodes 565 drivers/clk/zynqmp/clkc.c nodes = clock[clk_id].node; nodes 576 drivers/clk/zynqmp/clkc.c clk_type_postfix[nodes[j].type]); nodes 581 drivers/clk/zynqmp/clkc.c if (!clk_topology[nodes[j].type]) nodes 584 drivers/clk/zynqmp/clkc.c hw = (*clk_topology[nodes[j].type])(clk_out, clk_dev_id, nodes 587 drivers/clk/zynqmp/clkc.c &nodes[j]); nodes 192 drivers/clk/zynqmp/divider.c const struct clock_topology *nodes) nodes 207 drivers/clk/zynqmp/divider.c init.flags = nodes->flag & ~CLK_FRAC; nodes 212 drivers/clk/zynqmp/divider.c div->is_frac = !!(nodes->flag & CLK_FRAC); nodes 213 drivers/clk/zynqmp/divider.c div->flags = nodes->type_flag; nodes 216 drivers/clk/zynqmp/divider.c div->div_type = nodes->type; nodes 303 drivers/clk/zynqmp/pll.c const struct clock_topology *nodes) nodes 312 drivers/clk/zynqmp/pll.c init.flags = nodes->flag; nodes 123 drivers/gpu/drm/amd/amdgpu/amdgpu_psp.h struct psp_xgmi_node_info nodes[AMDGPU_XGMI_MAX_CONNECTED_NODES]; nodes 596 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct drm_mm_node *nodes = mem->mm_node; nodes 605 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (nodes->size != mem->num_pages) nodes 608 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c return ((nodes->start + nodes->size) << PAGE_SHIFT) nodes 1545 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct drm_mm_node *nodes; nodes 1554 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c nodes = amdgpu_find_mm_node(&abo->tbo.mem, &offset); nodes 1555 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c pos = (nodes->start << PAGE_SHIFT) + offset; nodes 1588 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (pos >= (nodes->start + nodes->size) << PAGE_SHIFT) { nodes 1589 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c ++nodes; nodes 1590 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c pos = (nodes->start << PAGE_SHIFT); nodes 1563 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct drm_mm_node *nodes, nodes 1603 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (nodes) { nodes 1604 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c while (pfn >= nodes->size) { nodes 1605 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c pfn -= nodes->size; nodes 1606 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ++nodes; nodes 1615 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (nodes) { nodes 1616 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c addr = nodes->start << PAGE_SHIFT; nodes 1617 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c max_entries = (nodes->size - pfn) * nodes 1658 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (nodes && nodes->size == pfn) { nodes 1660 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ++nodes; nodes 1690 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct drm_mm_node *nodes; nodes 1698 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c nodes = NULL; nodes 1704 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c nodes = mem->mm_node; nodes 1734 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c mapping, flags, bo_adev, nodes, nodes 219 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct drm_mm_node *nodes = mem->mm_node; nodes 229 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) nodes 230 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c usage += amdgpu_vram_mgr_vis_size(adev, nodes); nodes 275 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct drm_mm_node *nodes; nodes 308 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c nodes = kvmalloc_array((uint32_t)num_nodes, sizeof(*nodes), nodes 310 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c if (!nodes) { nodes 326 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c r = drm_mm_insert_node_in_range(mm, &nodes[i], pages, nodes 333 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); nodes 334 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c amdgpu_vram_mgr_virt_start(mem, &nodes[i]); nodes 345 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c r = drm_mm_insert_node_in_range(mm, &nodes[i], nodes 352 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); nodes 353 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c amdgpu_vram_mgr_virt_start(mem, &nodes[i]); nodes 360 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c mem->mm_node = nodes; nodes 366 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c drm_mm_remove_node(&nodes[i]); nodes 370 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c kvfree(nodes); nodes 389 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct drm_mm_node *nodes = mem->mm_node; nodes 398 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c pages -= nodes->size; nodes 399 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c drm_mm_remove_node(nodes); nodes 400 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c usage += nodes->size << PAGE_SHIFT; nodes 401 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c vis_usage += amdgpu_vram_mgr_vis_size(adev, nodes); nodes 402 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c ++nodes; nodes 321 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) nodes 322 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c return top->nodes[i].num_hops; nodes 370 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c top_info->nodes[count++].node_id = entry->node_id; nodes 379 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c top_info->nodes[count - 1].node_id = nodes 754 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].node_id = topology->nodes[i].node_id; nodes 755 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].num_hops = topology->nodes[i].num_hops; nodes 756 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].is_sharing_enabled = topology->nodes[i].is_sharing_enabled; nodes 757 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].sdma_engine = topology->nodes[i].sdma_engine; nodes 769 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology->nodes[i].node_id = topology_info_output->nodes[i].node_id; nodes 770 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology->nodes[i].num_hops = topology_info_output->nodes[i].num_hops; nodes 771 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology->nodes[i].is_sharing_enabled = topology_info_output->nodes[i].is_sharing_enabled; nodes 772 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology->nodes[i].sdma_engine = topology_info_output->nodes[i].sdma_engine; nodes 796 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].node_id = topology->nodes[i].node_id; nodes 797 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].num_hops = topology->nodes[i].num_hops; nodes 798 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].is_sharing_enabled = 1; nodes 799 drivers/gpu/drm/amd/amdgpu/psp_v11_0.c topology_info_input->nodes[i].sdma_engine = topology->nodes[i].sdma_engine; nodes 92 drivers/gpu/drm/amd/amdgpu/ta_xgmi_if.h struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; nodes 97 drivers/gpu/drm/amd/amdgpu/ta_xgmi_if.h struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; nodes 102 drivers/gpu/drm/amd/amdgpu/ta_xgmi_if.h struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; nodes 913 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c uint32_t nodes = 0; nodes 957 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpu_id = pdd->dev->id; nodes 958 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].lds_base = pdd->lds_base; nodes 959 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].lds_limit = pdd->lds_limit; nodes 960 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpuvm_base = pdd->gpuvm_base; nodes 961 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpuvm_limit = pdd->gpuvm_limit; nodes 962 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].scratch_base = pdd->scratch_base; nodes 963 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].scratch_limit = pdd->scratch_limit; nodes 979 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c nodes++; nodes 982 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c } while (pdd && (nodes < args->num_of_nodes)); nodes 985 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c args->num_of_nodes = nodes; nodes 989 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c (nodes * sizeof(struct kfd_process_device_apertures))); nodes 204 drivers/gpu/drm/i915/gt/intel_lrc.c } nodes[I915_NUM_ENGINES]; nodes 356 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); nodes 1071 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); nodes 1169 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); nodes 2615 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); nodes 3430 drivers/gpu/drm/i915/gt/intel_lrc.c struct rb_node *node = &ve->nodes[sibling->id].rb; nodes 3566 drivers/gpu/drm/i915/gt/intel_lrc.c struct ve_node * const node = &ve->nodes[sibling->id]; nodes 3794 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(RB_EMPTY_NODE(&ve->nodes[sibling->id].rb)); nodes 3795 drivers/gpu/drm/i915/gt/intel_lrc.c RB_CLEAR_NODE(&ve->nodes[sibling->id].rb); nodes 3975 drivers/gpu/drm/i915/gt/intel_lrc.c rb_entry(rb, typeof(*ve), nodes[engine->id].rb); nodes 21 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h struct list_head nodes; nodes 48 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h list_for_each_entry(node, &mm->nodes, nl_entry) { nodes 26 drivers/gpu/drm/nouveau/nvkm/core/mm.c #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ nodes 36 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(node, &mm->nodes, nl_entry) { nodes 246 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); nodes 255 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&node->nl_entry, &mm->nodes); nodes 259 drivers/gpu/drm/nouveau/nvkm/core/mm.c INIT_LIST_HEAD(&mm->nodes); nodes 275 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&node->nl_entry, &mm->nodes); nodes 286 drivers/gpu/drm/nouveau/nvkm/core/mm.c int nodes = 0; nodes 291 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(node, &mm->nodes, nl_entry) { nodes 293 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (++nodes > mm->heap_nodes) { nodes 300 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) { nodes 267 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node nodes[2]; nodes 276 drivers/gpu/drm/selftests/test-drm_mm.c memset(nodes, 0, sizeof(nodes)); nodes 277 drivers/gpu/drm/selftests/test-drm_mm.c nodes[0].start = 512; nodes 278 drivers/gpu/drm/selftests/test-drm_mm.c nodes[0].size = 1024; nodes 279 drivers/gpu/drm/selftests/test-drm_mm.c ret = drm_mm_reserve_node(&mm, &nodes[0]); nodes 282 drivers/gpu/drm/selftests/test-drm_mm.c nodes[0].start, nodes[0].size); nodes 286 drivers/gpu/drm/selftests/test-drm_mm.c nodes[1].size = 1024; nodes 287 drivers/gpu/drm/selftests/test-drm_mm.c nodes[1].start = 4096 - 512 - nodes[1].size; nodes 288 drivers/gpu/drm/selftests/test-drm_mm.c ret = drm_mm_reserve_node(&mm, &nodes[1]); nodes 291 drivers/gpu/drm/selftests/test-drm_mm.c nodes[1].start, nodes[1].size); nodes 375 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node tmp, *nodes, *node, *next; nodes 393 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(count, sizeof(*nodes))); nodes 394 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 404 drivers/gpu/drm/selftests/test-drm_mm.c nodes[n].start = order[n] * size; nodes 405 drivers/gpu/drm/selftests/test-drm_mm.c nodes[n].size = size; nodes 407 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, &nodes[n]); nodes 410 drivers/gpu/drm/selftests/test-drm_mm.c n, nodes[n].start); nodes 415 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_node_allocated(&nodes[n])) { nodes 417 drivers/gpu/drm/selftests/test-drm_mm.c n, nodes[n].start); nodes 421 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_reserve_fail(&mm, &nodes[n])) nodes 437 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&nodes[order[n]]); nodes 438 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, &nodes[order[n]]); nodes 441 drivers/gpu/drm/selftests/test-drm_mm.c n, nodes[n].start); nodes 466 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 471 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 492 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 573 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; nodes 583 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vmalloc(array_size(count, sizeof(*nodes))); nodes 584 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 598 drivers/gpu/drm/selftests/test-drm_mm.c node = replace ? &tmp : &nodes[n]; nodes 607 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_replace_node(&tmp, &nodes[n]); nodes 614 drivers/gpu/drm/selftests/test-drm_mm.c if (!assert_node(&nodes[n], &mm, size, 0, n)) { nodes 620 drivers/gpu/drm/selftests/test-drm_mm.c if (tmp.start != nodes[n].start) { nodes 623 drivers/gpu/drm/selftests/test-drm_mm.c nodes[n].start, nodes[n].size); nodes 639 drivers/gpu/drm/selftests/test-drm_mm.c u64 addr = nodes[n].start; nodes 641 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&nodes[n]); nodes 642 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n], size, 0, n, mode)) { nodes 648 drivers/gpu/drm/selftests/test-drm_mm.c if (nodes[n].start != addr) { nodes 650 drivers/gpu/drm/selftests/test-drm_mm.c mode->name, n, addr, nodes[n].start); nodes 661 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 666 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 697 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 880 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; nodes 893 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(count, sizeof(*nodes))); nodes 894 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 905 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert_in_range(&mm, &nodes[n], nodes 924 drivers/gpu/drm/selftests/test-drm_mm.c u64 addr = nodes[n].start; nodes 926 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&nodes[n]); nodes 927 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert_in_range(&mm, &nodes[n], nodes 934 drivers/gpu/drm/selftests/test-drm_mm.c if (nodes[n].start != addr) { nodes 936 drivers/gpu/drm/selftests/test-drm_mm.c mode->name, n, addr, nodes[n].start); nodes 959 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 1041 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; nodes 1050 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(max_count, sizeof(*nodes))); nodes 1051 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 1062 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[i], nodes 1085 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 1188 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes, nodes 1198 drivers/gpu/drm/selftests/test-drm_mm.c e = &nodes[order ? order[i] : i]; nodes 1236 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes) nodes 1246 drivers/gpu/drm/selftests/test-drm_mm.c e = &nodes[n]; nodes 1254 drivers/gpu/drm/selftests/test-drm_mm.c e = &nodes[n]; nodes 1270 drivers/gpu/drm/selftests/test-drm_mm.c e = &nodes[n]; nodes 1278 drivers/gpu/drm/selftests/test-drm_mm.c return assert_continuous(mm, nodes[0].node.size); nodes 1283 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes) nodes 1293 drivers/gpu/drm/selftests/test-drm_mm.c e = &nodes[n]; nodes 1327 drivers/gpu/drm/selftests/test-drm_mm.c return assert_continuous(mm, nodes[0].node.size); nodes 1332 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes, nodes 1350 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, count, false, nodes 1393 drivers/gpu/drm/selftests/test-drm_mm.c if (!assert_continuous(mm, nodes[0].node.size)) { nodes 1407 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes; nodes 1420 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(size, sizeof(*nodes))); nodes 1421 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 1431 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node(&mm, &nodes[n].node, 1); nodes 1440 drivers/gpu/drm/selftests/test-drm_mm.c if (!evict_nothing(&mm, size, nodes)) { nodes 1444 drivers/gpu/drm/selftests/test-drm_mm.c if (!evict_everything(&mm, size, nodes)) { nodes 1453 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1467 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1485 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1506 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 1520 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes; nodes 1530 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(size, sizeof(*nodes))); nodes 1531 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 1541 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node(&mm, &nodes[n].node, 1); nodes 1553 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1566 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1583 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, size, nodes 1603 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 1621 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; nodes 1631 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(count, sizeof(*nodes))); nodes 1632 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 1647 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n], nodes 1654 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_hole_follows(&nodes[n])) { nodes 1656 drivers/gpu/drm/selftests/test-drm_mm.c n, nodes[n].start, size); nodes 1670 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 1678 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 1722 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 1735 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; nodes 1744 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(count, sizeof(*nodes))); nodes 1745 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 1760 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n], nodes 1777 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 1785 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; nodes 1822 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 2085 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes, nodes 2104 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, count, true, nodes 2159 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes; nodes 2171 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(total_size, sizeof(*nodes))); nodes 2172 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 2183 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n].node, nodes 2195 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2208 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2225 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2247 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 2262 drivers/gpu/drm/selftests/test-drm_mm.c struct evict_node *nodes; nodes 2272 drivers/gpu/drm/selftests/test-drm_mm.c nodes = vzalloc(array_size(total_size, sizeof(*nodes))); nodes 2273 drivers/gpu/drm/selftests/test-drm_mm.c if (!nodes) nodes 2284 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n].node, nodes 2296 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2309 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2326 drivers/gpu/drm/selftests/test-drm_mm.c nodes, order, total_size, nodes 2348 drivers/gpu/drm/selftests/test-drm_mm.c vfree(nodes); nodes 105 drivers/hwmon/applesmc.c struct applesmc_dev_attr *nodes; /* dynamic node array */ nodes 1084 drivers/hwmon/applesmc.c for (grp = groups; grp->nodes; grp++) { nodes 1085 drivers/hwmon/applesmc.c for (node = grp->nodes; node->sda.dev_attr.attr.name; node++) nodes 1088 drivers/hwmon/applesmc.c kfree(grp->nodes); nodes 1089 drivers/hwmon/applesmc.c grp->nodes = NULL; nodes 1104 drivers/hwmon/applesmc.c grp->nodes = kcalloc(num + 1, sizeof(*node), GFP_KERNEL); nodes 1105 drivers/hwmon/applesmc.c if (!grp->nodes) { nodes 1110 drivers/hwmon/applesmc.c node = &grp->nodes[i]; nodes 76 drivers/interconnect/core.c list_for_each_entry(n, &provider->nodes, node_list) { nodes 268 drivers/interconnect/core.c return icc_data->nodes[idx]; nodes 725 drivers/interconnect/core.c list_add_tail(&node->node_list, &provider->nodes); nodes 760 drivers/interconnect/core.c INIT_LIST_HEAD(&provider->nodes); nodes 787 drivers/interconnect/core.c if (!list_empty(&provider->nodes)) { nodes 143 drivers/interconnect/qcom/qcs404.c struct qcom_icc_node **nodes; nodes 244 drivers/interconnect/qcom/qcs404.c .nodes = qcs404_bimc_nodes, nodes 298 drivers/interconnect/qcom/qcs404.c .nodes = qcs404_pcnoc_nodes, nodes 326 drivers/interconnect/qcom/qcs404.c .nodes = qcs404_snoc_nodes, nodes 356 drivers/interconnect/qcom/qcs404.c list_for_each_entry(n, &provider->nodes, node_list) nodes 429 drivers/interconnect/qcom/qcs404.c qnodes = desc->nodes; nodes 436 drivers/interconnect/qcom/qcs404.c data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), nodes 456 drivers/interconnect/qcom/qcs404.c INIT_LIST_HEAD(&provider->nodes); nodes 489 drivers/interconnect/qcom/qcs404.c data->nodes[i] = node; nodes 497 drivers/interconnect/qcom/qcs404.c list_for_each_entry_safe(node, tmp, &provider->nodes, node_list) { nodes 513 drivers/interconnect/qcom/qcs404.c list_for_each_entry_safe(n, tmp, &provider->nodes, node_list) { nodes 121 drivers/interconnect/qcom/sdm845.c struct qcom_icc_node *nodes[]; nodes 125 drivers/interconnect/qcom/sdm845.c struct qcom_icc_node **nodes; nodes 130 drivers/interconnect/qcom/sdm845.c struct qcom_icc_node **nodes; nodes 283 drivers/interconnect/qcom/sdm845.c .nodes = { __VA_ARGS__ }, \ nodes 480 drivers/interconnect/qcom/sdm845.c .nodes = rsc_hlos_nodes, nodes 521 drivers/interconnect/qcom/sdm845.c qn = bcm->nodes[i]; nodes 604 drivers/interconnect/qcom/sdm845.c temp = bcm->nodes[i]->sum_avg[bucket] * bcm->aux_data.width; nodes 605 drivers/interconnect/qcom/sdm845.c do_div(temp, bcm->nodes[i]->buswidth * bcm->nodes[i]->channels); nodes 608 drivers/interconnect/qcom/sdm845.c temp = bcm->nodes[i]->max_peak[bucket] * bcm->aux_data.width; nodes 609 drivers/interconnect/qcom/sdm845.c do_div(temp, bcm->nodes[i]->buswidth); nodes 786 drivers/interconnect/qcom/sdm845.c qnodes = desc->nodes; nodes 793 drivers/interconnect/qcom/sdm845.c data = devm_kzalloc(&pdev->dev, struct_size(data, nodes, num_nodes), nodes 804 drivers/interconnect/qcom/sdm845.c INIT_LIST_HEAD(&provider->nodes); nodes 837 drivers/interconnect/qcom/sdm845.c data->nodes[i] = node; nodes 858 drivers/interconnect/qcom/sdm845.c list_for_each_entry(node, &provider->nodes, node_list) { nodes 873 drivers/interconnect/qcom/sdm845.c list_for_each_entry_safe(n, tmp, &provider->nodes, node_list) { nodes 473 drivers/md/bcache/bcache.h size_t nodes; nodes 1323 drivers/md/bcache/btree.c gc->nodes++; nodes 1372 drivers/md/bcache/btree.c unsigned int i, nodes = 0, keys = 0, blocks; nodes 1386 drivers/md/bcache/btree.c while (nodes < GC_MERGE_NODES && !IS_ERR_OR_NULL(r[nodes].b)) nodes 1387 drivers/md/bcache/btree.c keys += r[nodes++].keys; nodes 1391 drivers/md/bcache/btree.c if (nodes < 2 || nodes 1393 drivers/md/bcache/btree.c block_bytes(b->c)) > blocks * (nodes - 1)) nodes 1396 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) { nodes 1411 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) nodes 1414 drivers/md/bcache/btree.c for (i = nodes - 1; i > 0; --i) { nodes 1480 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) nodes 1491 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) { nodes 1502 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) { nodes 1509 drivers/md/bcache/btree.c memmove(r, r + 1, sizeof(r[0]) * (nodes - 1)); nodes 1510 drivers/md/bcache/btree.c r[nodes - 1].b = ERR_PTR(-EINTR); nodes 1512 drivers/md/bcache/btree.c trace_bcache_btree_gc_coalesce(nodes); nodes 1513 drivers/md/bcache/btree.c gc->nodes--; nodes 1528 drivers/md/bcache/btree.c for (i = 0; i < nodes; i++) nodes 1602 drivers/md/bcache/btree.c min_nodes = c->gc_stats.nodes / MAX_GC_TIMES; nodes 1676 drivers/md/bcache/btree.c gc->nodes >= gc->nodes_pre + btree_gc_min_nodes(b->c)) { nodes 1677 drivers/md/bcache/btree.c gc->nodes_pre = gc->nodes; nodes 1940 drivers/md/bcache/btree.c b->c->gc_stats.nodes++; nodes 592 drivers/md/bcache/sysfs.c size_t nodes; nodes 600 drivers/md/bcache/sysfs.c op->nodes++; nodes 626 drivers/md/bcache/sysfs.c op.nodes, nodes 695 drivers/md/bcache/sysfs.c (c->gc_stats.nodes ?: 1) * btree_bytes(c)); nodes 726 drivers/md/bcache/sysfs.c sysfs_print(btree_nodes, c->gc_stats.nodes); nodes 474 drivers/md/md-bitmap.c sb->nodes = cpu_to_le32(bitmap->mddev->bitmap_info.nodes); nodes 573 drivers/md/md-bitmap.c bitmap->mddev->bitmap_info.nodes = 0; nodes 587 drivers/md/md-bitmap.c int nodes = 0; nodes 649 drivers/md/md-bitmap.c nodes = le32_to_cpu(sb->nodes); nodes 688 drivers/md/md-bitmap.c if (!nodes && (events < bitmap->mddev->events)) { nodes 708 drivers/md/md-bitmap.c if (err == 0 && nodes && (bitmap->cluster_slot < 0)) { nodes 709 drivers/md/md-bitmap.c err = md_setup_cluster(bitmap->mddev, nodes); nodes 726 drivers/md/md-bitmap.c bitmap->mddev->bitmap_info.nodes = nodes; nodes 1914 drivers/md/md-bitmap.c md_cluster_ops->load_bitmaps(mddev, mddev->bitmap_info.nodes); nodes 136 drivers/md/md-bitmap.h __le32 nodes; /* 68 the maximum number of nodes in cluster. */ nodes 852 drivers/md/md-cluster.c static int join(struct mddev *mddev, int nodes) nodes 880 drivers/md/md-cluster.c if (nodes < cinfo->slot_number) { nodes 882 drivers/md/md-cluster.c cinfo->slot_number, nodes); nodes 1135 drivers/md/md-cluster.c for (i = 0; i < mddev->bitmap_info.nodes; i++) { nodes 1185 drivers/md/md-cluster.c int node_num = mddev->bitmap_info.nodes; nodes 1480 drivers/md/md-cluster.c kcalloc(mddev->bitmap_info.nodes - 1, nodes 1488 drivers/md/md-cluster.c for (slot = 0; slot < mddev->bitmap_info.nodes; slot++) { nodes 1515 drivers/md/md-cluster.c for (i = 0; i < mddev->bitmap_info.nodes - 1; i++) { nodes 1538 drivers/md/md-cluster.c for (sn = 0; sn < mddev->bitmap_info.nodes; sn++) { nodes 13 drivers/md/md-cluster.h int (*join)(struct mddev *mddev, int nodes); nodes 6000 drivers/md/md.c mddev->bitmap_info.nodes = 0; nodes 7189 drivers/md/md.c if (mddev->bitmap_info.nodes) { nodes 7198 drivers/md/md.c mddev->bitmap_info.nodes = 0; nodes 8193 drivers/md/md.c int md_setup_cluster(struct mddev *mddev, int nodes) nodes 8206 drivers/md/md.c return md_cluster_ops->join(mddev, nodes); nodes 465 drivers/md/md.h int nodes; /* Maximum number of nodes in the cluster */ nodes 687 drivers/md/md.h extern int md_setup_cluster(struct mddev *mddev, int nodes); nodes 770 drivers/md/md.h return mddev->cluster_info && mddev->bitmap_info.nodes > 1; nodes 67 drivers/md/persistent-data/dm-btree-internal.h struct dm_block *nodes[2]; nodes 80 drivers/md/persistent-data/dm-btree-internal.h struct dm_block *nodes[2]; nodes 131 drivers/md/persistent-data/dm-btree-spine.c s->nodes[0] = NULL; nodes 132 drivers/md/persistent-data/dm-btree-spine.c s->nodes[1] = NULL; nodes 140 drivers/md/persistent-data/dm-btree-spine.c unlock_block(s->info, s->nodes[i]); nodes 151 drivers/md/persistent-data/dm-btree-spine.c unlock_block(s->info, s->nodes[0]); nodes 152 drivers/md/persistent-data/dm-btree-spine.c s->nodes[0] = s->nodes[1]; nodes 156 drivers/md/persistent-data/dm-btree-spine.c r = bn_read_lock(s->info, new_child, s->nodes + s->count); nodes 167 drivers/md/persistent-data/dm-btree-spine.c unlock_block(s->info, s->nodes[s->count]); nodes 175 drivers/md/persistent-data/dm-btree-spine.c block = s->nodes[s->count - 1]; nodes 193 drivers/md/persistent-data/dm-btree-spine.c unlock_block(s->info, s->nodes[i]); nodes 205 drivers/md/persistent-data/dm-btree-spine.c unlock_block(s->info, s->nodes[0]); nodes 206 drivers/md/persistent-data/dm-btree-spine.c s->nodes[0] = s->nodes[1]; nodes 210 drivers/md/persistent-data/dm-btree-spine.c r = bn_shadow(s->info, b, vt, s->nodes + s->count); nodes 213 drivers/md/persistent-data/dm-btree-spine.c s->root = dm_block_location(s->nodes[0]); nodes 225 drivers/md/persistent-data/dm-btree-spine.c return s->nodes[s->count - 1]; nodes 232 drivers/md/persistent-data/dm-btree-spine.c return s->count == 2 ? s->nodes[0] : NULL; nodes 590 drivers/md/persistent-data/dm-btree.c s->nodes[1] = left; nodes 593 drivers/md/persistent-data/dm-btree.c s->nodes[1] = right; nodes 999 drivers/md/persistent-data/dm-btree.c struct cursor_node *n = c->nodes + c->depth - 1; nodes 1014 drivers/md/persistent-data/dm-btree.c struct cursor_node *n = c->nodes + c->depth - 1; nodes 1023 drivers/md/persistent-data/dm-btree.c struct cursor_node *n = c->nodes + c->depth; nodes 1046 drivers/md/persistent-data/dm-btree.c unlock_block(c->info, c->nodes[c->depth].b); nodes 1058 drivers/md/persistent-data/dm-btree.c n = c->nodes + c->depth - 1; nodes 1079 drivers/md/persistent-data/dm-btree.c n = c->nodes + c->depth - 1; nodes 1151 drivers/md/persistent-data/dm-btree.c struct cursor_node *n = c->nodes + c->depth - 1; nodes 200 drivers/md/persistent-data/dm-btree.h struct cursor_node nodes[DM_BTREE_CURSOR_MAX_DEPTH]; nodes 1418 drivers/misc/mic/scif/scif_api.c int scif_get_node_ids(u16 *nodes, int len, u16 *self) nodes 1434 drivers/misc/mic/scif/scif_api.c nodes[offset++] = node; nodes 261 drivers/misc/mic/scif/scif_fd.c u16 *nodes; nodes 271 drivers/misc/mic/scif/scif_fd.c nodes = kmalloc_array(entries, sizeof(u16), GFP_KERNEL); nodes 272 drivers/misc/mic/scif/scif_fd.c if (entries && !nodes) { nodes 276 drivers/misc/mic/scif/scif_fd.c node_ids.len = scif_get_node_ids(nodes, entries, &self); nodes 278 drivers/misc/mic/scif/scif_fd.c unodes = (void __user *)node_ids.nodes; nodes 279 drivers/misc/mic/scif/scif_fd.c if (copy_to_user(unodes, nodes, sizeof(u16) * entries)) { nodes 295 drivers/misc/mic/scif/scif_fd.c kfree(nodes); nodes 169 drivers/misc/sgi-gru/grufile.c info.nodes = num_online_nodes(); nodes 170 drivers/misc/sgi-gru/grufile.c info.blades = info.nodes / nodesperblade; nodes 148 drivers/misc/sgi-gru/grulib.h int nodes; nodes 130 drivers/net/wireless/ath/ath9k/dynack.c list_for_each_entry(an, &da->nodes, list) nodes 329 drivers/net/wireless/ath/ath9k/dynack.c list_add_tail(&an->list, &da->nodes); nodes 370 drivers/net/wireless/ath/ath9k/dynack.c list_for_each_entry(an, &da->nodes, list) nodes 392 drivers/net/wireless/ath/ath9k/dynack.c INIT_LIST_HEAD(&da->nodes); nodes 74 drivers/net/wireless/ath/ath9k/dynack.h struct list_head nodes; nodes 456 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor nodes[BRCMF_FWS_MAC_DESC_TABLE_SIZE]; nodes 754 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[0]; nodes 755 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c for (i = 0; i < ARRAY_SIZE(fws->desc.nodes); i++) { nodes 866 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c table = &fws->desc.nodes[0]; nodes 867 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c for (i = 0; i < ARRAY_SIZE(fws->desc.nodes); i++) nodes 1015 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[mac_handle & 0x1F]; nodes 1072 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[mac_handle & 0x1F]; nodes 1146 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[data[1] & 0x1F]; nodes 49 drivers/platform/x86/intel_cht_int33fe.c static const struct software_node nodes[]; nodes 52 drivers/platform/x86/intel_cht_int33fe.c &nodes[INT33FE_NODE_PI3USB30532] nodes 56 drivers/platform/x86/intel_cht_int33fe.c &nodes[INT33FE_NODE_DISPLAYPORT] nodes 135 drivers/platform/x86/intel_cht_int33fe.c static const struct software_node nodes[] = { nodes 140 drivers/platform/x86/intel_cht_int33fe.c { "connector", &nodes[0], usb_connector_props, usb_connector_refs }, nodes 149 drivers/platform/x86/intel_cht_int33fe.c fwnode = software_node_fwnode(&nodes[INT33FE_NODE_DISPLAYPORT]); nodes 174 drivers/platform/x86/intel_cht_int33fe.c software_node_unregister_nodes(nodes); nodes 192 drivers/platform/x86/intel_cht_int33fe.c ret = software_node_register_nodes(nodes); nodes 234 drivers/platform/x86/intel_cht_int33fe.c fwnode = software_node_fwnode(&nodes[INT33FE_NODE_MAX17047]); nodes 333 drivers/platform/x86/intel_cht_int33fe.c fwnode = software_node_fwnode(&nodes[INT33FE_NODE_FUSB302]); nodes 351 drivers/platform/x86/intel_cht_int33fe.c fwnode = software_node_fwnode(&nodes[INT33FE_NODE_PI3USB30532]); nodes 32 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.width; nodes 34 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.height; nodes 43 drivers/staging/media/ipu3/ipu3-v4l2.c try_fmt->code = imgu_pipe->nodes[i].pad_fmt.code; nodes 79 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_pipe->queue_enabled[i] = imgu_pipe->nodes[i].enabled; nodes 91 drivers/staging/media/ipu3/ipu3-v4l2.c &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp : NULL; nodes 96 drivers/staging/media/ipu3/ipu3-v4l2.c if (imgu_pipe->nodes[IMGU_NODE_VF].enabled) nodes 137 drivers/staging/media/ipu3/ipu3-v4l2.c fmt->format = imgu_pipe->nodes[pad].pad_fmt; nodes 167 drivers/staging/media/ipu3/ipu3-v4l2.c mf = &imgu_pipe->nodes[pad].pad_fmt; nodes 286 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_pipe->nodes[pad].enabled = flags & MEDIA_LNK_FL_ENABLED; nodes 430 drivers/staging/media/ipu3/ipu3-v4l2.c node = &imgu->imgu_pipe[p].nodes[i]; nodes 666 drivers/staging/media/ipu3/ipu3-v4l2.c pipe, i, imgu_pipe->nodes[i].enabled); nodes 668 drivers/staging/media/ipu3/ipu3-v4l2.c if (imgu_pipe->nodes[IMGU_NODE_VF].enabled) nodes 686 drivers/staging/media/ipu3/ipu3-v4l2.c fmts[i] = kmemdup(&imgu_pipe->nodes[inode].vdev_fmt.fmt.pix_mp, nodes 694 drivers/staging/media/ipu3/ipu3-v4l2.c fmts[i] = &imgu_pipe->nodes[inode].vdev_fmt.fmt.pix_mp; nodes 699 drivers/staging/media/ipu3/ipu3-v4l2.c !imgu_pipe->nodes[inode].enabled) nodes 712 drivers/staging/media/ipu3/ipu3-v4l2.c pad_fmt = imgu_pipe->nodes[IMGU_NODE_IN].pad_fmt; nodes 740 drivers/staging/media/ipu3/ipu3-v4l2.c f->fmt = imgu_pipe->nodes[node].vdev_fmt.fmt; nodes 1110 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_sd->subdev_pads[i].flags = imgu_pipe->nodes[i].output ? nodes 1160 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = &imgu_pipe->nodes[node_num]; nodes 1282 drivers/staging/media/ipu3/ipu3-v4l2.c video_unregister_device(&imgu_pipe->nodes[i].vdev); nodes 1283 drivers/staging/media/ipu3/ipu3-v4l2.c media_entity_cleanup(&imgu_pipe->nodes[i].vdev.entity); nodes 1284 drivers/staging/media/ipu3/ipu3-v4l2.c mutex_destroy(&imgu_pipe->nodes[i].lock); nodes 120 drivers/staging/media/ipu3/ipu3.c if (!imgu_pipe->nodes[IMGU_NODE_VF].enabled && nodes 128 drivers/staging/media/ipu3/ipu3.c meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta; nodes 129 drivers/staging/media/ipu3/ipu3.c mpix = &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp; nodes 214 drivers/staging/media/ipu3/ipu3.c list_for_each_entry(buf, &imgu_pipe->nodes[node].buffers, vid_buf.list) { nodes 249 drivers/staging/media/ipu3/ipu3.c !imgu_pipe->nodes[IMGU_NODE_VF].enabled) { nodes 254 drivers/staging/media/ipu3/ipu3.c imgu_pipe->nodes[node].enabled) { nodes 259 drivers/staging/media/ipu3/ipu3.c if (list_empty(&imgu_pipe->nodes[node].buffers)) nodes 262 drivers/staging/media/ipu3/ipu3.c ivb = list_first_entry(&imgu_pipe->nodes[node].buffers, nodes 330 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->nodes[node].buffers, nodes 446 drivers/staging/media/ipu3/ipu3.c imgu_pipe->nodes[i].name = imgu_node_map[i].name; nodes 447 drivers/staging/media/ipu3/ipu3.c imgu_pipe->nodes[i].output = i < IMGU_QUEUE_FIRST_INPUT; nodes 448 drivers/staging/media/ipu3/ipu3.c imgu_pipe->nodes[i].enabled = false; nodes 452 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->nodes[i].vdev_fmt.fmt.pix_mp; nodes 453 drivers/staging/media/ipu3/ipu3.c atomic_set(&imgu_pipe->nodes[i].sequence, 0); nodes 547 drivers/staging/media/ipu3/ipu3.c if (!imgu_pipe->nodes[node].output) { nodes 552 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->nodes[node].sequence); nodes 113 drivers/staging/media/ipu3/ipu3.h struct imgu_video_device nodes[IMGU_NODE_NUM]; nodes 88 drivers/staging/rtl8712/rtl871x_event.h struct event_node nodes[C2HEVENT_SZ]; nodes 96 drivers/staging/rtl8723bs/include/rtw_event.h struct event_node nodes[C2HEVENT_SZ]; nodes 155 drivers/target/iscsi/iscsi_target_stat.c CONFIGFS_ATTR_RO(iscsi_stat_instance_, nodes); nodes 429 fs/btrfs/backref.c eb = path->nodes[level]; nodes 441 fs/btrfs/backref.c if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { nodes 449 fs/btrfs/backref.c eb = path->nodes[0]; nodes 567 fs/btrfs/backref.c eb = path->nodes[level]; nodes 574 fs/btrfs/backref.c eb = path->nodes[level]; nodes 895 fs/btrfs/backref.c leaf = path->nodes[0]; nodes 1019 fs/btrfs/backref.c leaf = path->nodes[0]; nodes 1212 fs/btrfs/backref.c leaf = path->nodes[0]; nodes 1551 fs/btrfs/backref.c leaf = path->nodes[0]; nodes 1654 fs/btrfs/backref.c eb = path->nodes[0]; nodes 1659 fs/btrfs/backref.c path->nodes[0] = NULL; nodes 1717 fs/btrfs/backref.c btrfs_item_key_to_cpu(path->nodes[0], found_key, path->slots[0]); nodes 1730 fs/btrfs/backref.c eb = path->nodes[0]; nodes 2029 fs/btrfs/backref.c eb = btrfs_clone_extent_buffer(path->nodes[0]); nodes 2088 fs/btrfs/backref.c eb = btrfs_clone_extent_buffer(path->nodes[0]); nodes 540 fs/btrfs/block-group.c leaf = path->nodes[0]; nodes 574 fs/btrfs/block-group.c leaf = path->nodes[0]; nodes 1447 fs/btrfs/block-group.c leaf = path->nodes[0]; nodes 1732 fs/btrfs/block-group.c leaf = path->nodes[0]; nodes 2137 fs/btrfs/block-group.c leaf = path->nodes[0]; nodes 67 fs/btrfs/ctree.c if (!p->nodes[i] || !p->locks[i]) nodes 75 fs/btrfs/ctree.c btrfs_set_lock_blocking_read(p->nodes[i]); nodes 78 fs/btrfs/ctree.c btrfs_set_lock_blocking_write(p->nodes[i]); nodes 105 fs/btrfs/ctree.c if (!p->nodes[i]) nodes 108 fs/btrfs/ctree.c btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); nodes 111 fs/btrfs/ctree.c free_extent_buffer(p->nodes[i]); nodes 112 fs/btrfs/ctree.c p->nodes[i] = NULL; nodes 1859 fs/btrfs/ctree.c mid = path->nodes[level]; nodes 1868 fs/btrfs/ctree.c parent = path->nodes[level + 1]; nodes 1907 fs/btrfs/ctree.c path->nodes[level] = NULL; nodes 2037 fs/btrfs/ctree.c path->nodes[level] = left; nodes 2051 fs/btrfs/ctree.c btrfs_node_blockptr(path->nodes[level], path->slots[level])) nodes 2059 fs/btrfs/ctree.c if (path->nodes[level] != left) nodes 2087 fs/btrfs/ctree.c mid = path->nodes[level]; nodes 2091 fs/btrfs/ctree.c parent = path->nodes[level + 1]; nodes 2133 fs/btrfs/ctree.c path->nodes[level] = left; nodes 2189 fs/btrfs/ctree.c path->nodes[level] = right; nodes 2229 fs/btrfs/ctree.c if (!path->nodes[level]) nodes 2232 fs/btrfs/ctree.c node = path->nodes[level]; nodes 2285 fs/btrfs/ctree.c parent = path->nodes[level + 1]; nodes 2344 fs/btrfs/ctree.c if (!path->nodes[i]) nodes 2354 fs/btrfs/ctree.c t = path->nodes[i]; nodes 2364 fs/btrfs/ctree.c t = path->nodes[i]; nodes 2394 fs/btrfs/ctree.c if (!path->nodes[i]) nodes 2398 fs/btrfs/ctree.c btrfs_tree_unlock_rw(path->nodes[i], path->locks[i]); nodes 2538 fs/btrfs/ctree.c b = p->nodes[level]; nodes 2557 fs/btrfs/ctree.c b = p->nodes[level]; nodes 2604 fs/btrfs/ctree.c eb = path->nodes[0]; nodes 2609 fs/btrfs/ctree.c eb = path->nodes[0]; nodes 2695 fs/btrfs/ctree.c p->nodes[level] = b; nodes 2748 fs/btrfs/ctree.c WARN_ON(p->nodes[0] != NULL); nodes 2810 fs/btrfs/ctree.c p->nodes[level + 1])) { nodes 2822 fs/btrfs/ctree.c p->nodes[level + 1], nodes 2830 fs/btrfs/ctree.c p->nodes[level] = b; nodes 2851 fs/btrfs/ctree.c btrfs_tree_unlock_rw(p->nodes[u], p->locks[u]); nodes 2875 fs/btrfs/ctree.c b = p->nodes[level]; nodes 2924 fs/btrfs/ctree.c p->nodes[level] = b; nodes 2990 fs/btrfs/ctree.c WARN_ON(p->nodes[0] != NULL); nodes 3008 fs/btrfs/ctree.c p->nodes[level] = b; nodes 3062 fs/btrfs/ctree.c p->nodes[level] = b; nodes 3110 fs/btrfs/ctree.c leaf = p->nodes[0]; nodes 3134 fs/btrfs/ctree.c leaf = p->nodes[0]; nodes 3174 fs/btrfs/ctree.c if (!path->nodes[i]) nodes 3176 fs/btrfs/ctree.c t = path->nodes[i]; nodes 3181 fs/btrfs/ctree.c btrfs_mark_buffer_dirty(path->nodes[i]); nodes 3201 fs/btrfs/ctree.c eb = path->nodes[0]; nodes 3397 fs/btrfs/ctree.c BUG_ON(path->nodes[level]); nodes 3398 fs/btrfs/ctree.c BUG_ON(path->nodes[level-1] != root->node); nodes 3400 fs/btrfs/ctree.c lower = path->nodes[level-1]; nodes 3433 fs/btrfs/ctree.c path->nodes[level] = c; nodes 3455 fs/btrfs/ctree.c BUG_ON(!path->nodes[level]); nodes 3456 fs/btrfs/ctree.c btrfs_assert_tree_locked(path->nodes[level]); nodes 3457 fs/btrfs/ctree.c lower = path->nodes[level]; nodes 3506 fs/btrfs/ctree.c c = path->nodes[level]; nodes 3524 fs/btrfs/ctree.c c = path->nodes[level]; nodes 3567 fs/btrfs/ctree.c path->nodes[level] = split; nodes 3636 fs/btrfs/ctree.c struct extent_buffer *left = path->nodes[0]; nodes 3637 fs/btrfs/ctree.c struct extent_buffer *upper = path->nodes[1]; nodes 3749 fs/btrfs/ctree.c if (btrfs_header_nritems(path->nodes[0]) == 0) nodes 3750 fs/btrfs/ctree.c btrfs_clean_tree_block(path->nodes[0]); nodes 3751 fs/btrfs/ctree.c btrfs_tree_unlock(path->nodes[0]); nodes 3752 fs/btrfs/ctree.c free_extent_buffer(path->nodes[0]); nodes 3753 fs/btrfs/ctree.c path->nodes[0] = right; nodes 3782 fs/btrfs/ctree.c struct extent_buffer *left = path->nodes[0]; nodes 3790 fs/btrfs/ctree.c if (!path->nodes[1]) nodes 3794 fs/btrfs/ctree.c upper = path->nodes[1]; nodes 3798 fs/btrfs/ctree.c btrfs_assert_tree_locked(path->nodes[1]); nodes 3836 fs/btrfs/ctree.c path->nodes[0] = right; nodes 3865 fs/btrfs/ctree.c struct extent_buffer *right = path->nodes[0]; nodes 3987 fs/btrfs/ctree.c btrfs_tree_unlock(path->nodes[0]); nodes 3988 fs/btrfs/ctree.c free_extent_buffer(path->nodes[0]); nodes 3989 fs/btrfs/ctree.c path->nodes[0] = left; nodes 4016 fs/btrfs/ctree.c struct extent_buffer *right = path->nodes[0]; nodes 4026 fs/btrfs/ctree.c if (!path->nodes[1]) nodes 4033 fs/btrfs/ctree.c btrfs_assert_tree_locked(path->nodes[1]); nodes 4035 fs/btrfs/ctree.c left = btrfs_read_node_slot(path->nodes[1], slot - 1); nodes 4054 fs/btrfs/ctree.c path->nodes[1], slot - 1, &left); nodes 4128 fs/btrfs/ctree.c btrfs_tree_unlock(path->nodes[0]); nodes 4129 fs/btrfs/ctree.c free_extent_buffer(path->nodes[0]); nodes 4130 fs/btrfs/ctree.c path->nodes[0] = right; nodes 4163 fs/btrfs/ctree.c if (slot < btrfs_header_nritems(path->nodes[0])) nodes 4164 fs/btrfs/ctree.c space_needed -= btrfs_leaf_free_space(path->nodes[0]); nodes 4177 fs/btrfs/ctree.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 4185 fs/btrfs/ctree.c if (btrfs_leaf_free_space(path->nodes[0]) >= data_size) nodes 4192 fs/btrfs/ctree.c space_needed -= btrfs_leaf_free_space(path->nodes[0]); nodes 4230 fs/btrfs/ctree.c l = path->nodes[0]; nodes 4237 fs/btrfs/ctree.c if (data_size && path->nodes[1]) { nodes 4256 fs/btrfs/ctree.c l = path->nodes[0]; nodes 4263 fs/btrfs/ctree.c if (!path->nodes[1]) { nodes 4270 fs/btrfs/ctree.c l = path->nodes[0]; nodes 4329 fs/btrfs/ctree.c btrfs_tree_unlock(path->nodes[0]); nodes 4330 fs/btrfs/ctree.c free_extent_buffer(path->nodes[0]); nodes 4331 fs/btrfs/ctree.c path->nodes[0] = right; nodes 4338 fs/btrfs/ctree.c btrfs_tree_unlock(path->nodes[0]); nodes 4339 fs/btrfs/ctree.c free_extent_buffer(path->nodes[0]); nodes 4340 fs/btrfs/ctree.c path->nodes[0] = right; nodes 4366 fs/btrfs/ctree.c if (btrfs_leaf_free_space(path->nodes[0]) >= data_size) nodes 4382 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4409 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4415 fs/btrfs/ctree.c if (btrfs_leaf_free_space(path->nodes[0]) >= ins_len) nodes 4452 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4555 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4566 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4593 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4689 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4765 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4884 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 4902 fs/btrfs/ctree.c struct extent_buffer *parent = path->nodes[level]; nodes 4986 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 5054 fs/btrfs/ctree.c if (path->nodes[0] == leaf && nodes 5073 fs/btrfs/ctree.c if (path->nodes[0] == leaf) nodes 5098 fs/btrfs/ctree.c btrfs_item_key_to_cpu(path->nodes[0], &key, 0); nodes 5117 fs/btrfs/ctree.c btrfs_item_key(path->nodes[0], &found_key, 0); nodes 5170 fs/btrfs/ctree.c WARN_ON(path->nodes[level]); nodes 5171 fs/btrfs/ctree.c path->nodes[level] = cur; nodes 5246 fs/btrfs/ctree.c path->nodes[level - 1] = cur; nodes 5278 fs/btrfs/ctree.c if (!path->nodes[level]) nodes 5282 fs/btrfs/ctree.c c = path->nodes[level]; nodes 5289 fs/btrfs/ctree.c !path->nodes[level + 1]) nodes 5312 fs/btrfs/ctree.c c = path->nodes[level]; nodes 5358 fs/btrfs/ctree.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 5362 fs/btrfs/ctree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nritems - 1); nodes 5381 fs/btrfs/ctree.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 5414 fs/btrfs/ctree.c if (!path->nodes[level]) { nodes 5420 fs/btrfs/ctree.c c = path->nodes[level]; nodes 5473 fs/btrfs/ctree.c c = path->nodes[level]; nodes 5478 fs/btrfs/ctree.c path->nodes[level] = next; nodes 5538 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 5580 fs/btrfs/ctree.c leaf = path->nodes[0]; nodes 347 fs/btrfs/ctree.h struct extent_buffer *nodes[BTRFS_MAX_LEVEL]; nodes 2608 fs/btrfs/ctree.h if (p->slots[0] >= btrfs_header_nritems(p->nodes[0])) nodes 183 fs/btrfs/delayed-inode.c root->nodes++; nodes 195 fs/btrfs/delayed-inode.c root->nodes--; nodes 712 fs/btrfs/delayed-inode.c BUG_ON(!path->nodes[0]); nodes 714 fs/btrfs/delayed-inode.c leaf = path->nodes[0]; nodes 819 fs/btrfs/delayed-inode.c leaf = path->nodes[0]; nodes 863 fs/btrfs/delayed-inode.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 886 fs/btrfs/delayed-inode.c BUG_ON(!path->nodes[0]); nodes 888 fs/btrfs/delayed-inode.c leaf = path->nodes[0]; nodes 1044 fs/btrfs/delayed-inode.c leaf = path->nodes[0]; nodes 1096 fs/btrfs/delayed-inode.c leaf = path->nodes[0]; nodes 34 fs/btrfs/delayed-inode.h int nodes; /* for delayed nodes */ nodes 81 fs/btrfs/delayed-inode.h delayed_root->nodes = 0; nodes 78 fs/btrfs/dev-replace.c eb = path->nodes[0]; nodes 315 fs/btrfs/dev-replace.c btrfs_item_size_nr(path->nodes[0], path->slots[0]) < sizeof(*ptr)) { nodes 349 fs/btrfs/dev-replace.c eb = path->nodes[0]; nodes 43 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 83 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 95 fs/btrfs/dir-item.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 144 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 249 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 314 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 324 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 382 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 418 fs/btrfs/dir-item.c leaf = path->nodes[0]; nodes 193 fs/btrfs/export.c leaf = path->nodes[0]; nodes 273 fs/btrfs/export.c leaf = path->nodes[0]; nodes 198 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 208 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 514 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 524 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 581 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 608 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 642 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 687 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 788 fs/btrfs/extent-tree.c if (!path->nodes[level]) nodes 791 fs/btrfs/extent-tree.c btrfs_header_nritems(path->nodes[level])) nodes 794 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[level], key, nodes 797 fs/btrfs/extent-tree.c btrfs_node_key_to_cpu(path->nodes[level], key, nodes 878 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 902 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 1027 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 1108 fs/btrfs/extent-tree.c struct extent_buffer *leaf = path->nodes[0]; nodes 1464 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 1596 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 1618 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 2350 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 2992 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 3040 fs/btrfs/extent-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 3063 fs/btrfs/extent-tree.c btrfs_print_leaf(path->nodes[0]); nodes 3072 fs/btrfs/extent-tree.c btrfs_print_leaf(path->nodes[0]); nodes 3084 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 4222 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 4246 fs/btrfs/extent-tree.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 4307 fs/btrfs/extent-tree.c leaf = path->nodes[0]; nodes 4621 fs/btrfs/extent-tree.c eb = path->nodes[wc->level]; nodes 4690 fs/btrfs/extent-tree.c struct extent_buffer *eb = path->nodes[level]; nodes 4810 fs/btrfs/extent-tree.c generation = btrfs_node_ptr_generation(path->nodes[level], nodes 4823 fs/btrfs/extent-tree.c bytenr = btrfs_node_blockptr(path->nodes[level], path->slots[level]); nodes 4824 fs/btrfs/extent-tree.c btrfs_node_key_to_cpu(path->nodes[level], &first_key, nodes 4864 fs/btrfs/extent-tree.c btrfs_node_key_to_cpu(path->nodes[level], &key, nodes 4908 fs/btrfs/extent-tree.c path->nodes[level] = next; nodes 4920 fs/btrfs/extent-tree.c parent = path->nodes[level]->start; nodes 4923 fs/btrfs/extent-tree.c btrfs_header_owner(path->nodes[level])); nodes 4925 fs/btrfs/extent-tree.c btrfs_header_owner(path->nodes[level])) { nodes 5014 fs/btrfs/extent-tree.c struct extent_buffer *eb = path->nodes[level]; nodes 5095 fs/btrfs/extent-tree.c parent = path->nodes[level + 1]->start; nodes 5097 fs/btrfs/extent-tree.c btrfs_header_owner(path->nodes[level + 1])) nodes 5131 fs/btrfs/extent-tree.c btrfs_header_nritems(path->nodes[level])) nodes 5153 fs/btrfs/extent-tree.c path->slots[level] = btrfs_header_nritems(path->nodes[level]); nodes 5154 fs/btrfs/extent-tree.c while (level < max_level && path->nodes[level]) { nodes 5157 fs/btrfs/extent-tree.c btrfs_header_nritems(path->nodes[level])) { nodes 5168 fs/btrfs/extent-tree.c btrfs_tree_unlock_rw(path->nodes[level], nodes 5172 fs/btrfs/extent-tree.c free_extent_buffer(path->nodes[level]); nodes 5173 fs/btrfs/extent-tree.c path->nodes[level] = NULL; nodes 5248 fs/btrfs/extent-tree.c path->nodes[level] = btrfs_lock_root_node(root); nodes 5249 fs/btrfs/extent-tree.c btrfs_set_lock_blocking_write(path->nodes[level]); nodes 5278 fs/btrfs/extent-tree.c btrfs_tree_lock(path->nodes[level]); nodes 5279 fs/btrfs/extent-tree.c btrfs_set_lock_blocking_write(path->nodes[level]); nodes 5283 fs/btrfs/extent-tree.c path->nodes[level]->start, nodes 5295 fs/btrfs/extent-tree.c btrfs_tree_unlock(path->nodes[level]); nodes 5331 fs/btrfs/extent-tree.c btrfs_node_key_to_cpu(path->nodes[wc->drop_level], nodes 5459 fs/btrfs/extent-tree.c path->nodes[parent_level] = parent; nodes 5464 fs/btrfs/extent-tree.c path->nodes[level] = node; nodes 4694 fs/btrfs/extent_io.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, path->slots[0]); nodes 60 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 102 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 249 fs/btrfs/file-item.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 253 fs/btrfs/file-item.c item_size = btrfs_item_size_nr(path->nodes[0], nodes 258 fs/btrfs/file-item.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 270 fs/btrfs/file-item.c read_extent_buffer(path->nodes[0], csum, nodes 340 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 353 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 360 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 380 fs/btrfs/file-item.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 400 fs/btrfs/file-item.c read_extent_buffer(path->nodes[0], nodes 554 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 628 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 774 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 788 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 798 fs/btrfs/file-item.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 807 fs/btrfs/file-item.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, slot); nodes 834 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 902 fs/btrfs/file-item.c leaf = path->nodes[0]; nodes 922 fs/btrfs/file-item.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 944 fs/btrfs/file-item.c struct extent_buffer *leaf = path->nodes[0]; nodes 793 fs/btrfs/file.c leaf = path->nodes[0]; nodes 802 fs/btrfs/file.c leaf = path->nodes[0]; nodes 813 fs/btrfs/file.c leaf = path->nodes[0]; nodes 898 fs/btrfs/file.c leaf = path->nodes[0]; nodes 1052 fs/btrfs/file.c leaf = path->nodes[0]; nodes 1185 fs/btrfs/file.c leaf = path->nodes[0]; nodes 1283 fs/btrfs/file.c leaf = path->nodes[0]; nodes 2320 fs/btrfs/file.c leaf = path->nodes[0]; nodes 2493 fs/btrfs/file.c leaf = path->nodes[0]; nodes 69 fs/btrfs/free-space-cache.c leaf = path->nodes[0]; nodes 152 fs/btrfs/free-space-cache.c leaf = path->nodes[0]; nodes 181 fs/btrfs/free-space-cache.c leaf = path->nodes[0]; nodes 701 fs/btrfs/free-space-cache.c leaf = path->nodes[0]; nodes 1018 fs/btrfs/free-space-cache.c leaf = path->nodes[0]; nodes 65 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 103 fs/btrfs/free-space-tree.c return btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 220 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 263 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 302 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 359 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 408 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 473 fs/btrfs/free-space-tree.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 474 fs/btrfs/free-space-tree.c extent_count = btrfs_free_space_extent_count(path->nodes[0], info); nodes 477 fs/btrfs/free-space-tree.c btrfs_set_free_space_extent_count(path->nodes[0], info, extent_count); nodes 478 fs/btrfs/free-space-tree.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 502 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 527 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 563 fs/btrfs/free-space-tree.c if (p->slots[0] + 1 < btrfs_header_nritems(p->nodes[0])) { nodes 568 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(p->nodes[0], &key, p->slots[0]); nodes 614 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 654 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 716 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 800 fs/btrfs/free-space-tree.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 893 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 933 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 993 fs/btrfs/free-space-tree.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 1090 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1206 fs/btrfs/free-space-tree.c nr = btrfs_header_nritems(path->nodes[0]); nodes 1351 fs/btrfs/free-space-tree.c leaf = path->nodes[0]; nodes 1419 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1497 fs/btrfs/free-space-tree.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1560 fs/btrfs/free-space-tree.c extent_count = btrfs_free_space_extent_count(path->nodes[0], info); nodes 1561 fs/btrfs/free-space-tree.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 93 fs/btrfs/inode-item.c return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], nodes 135 fs/btrfs/inode-item.c extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], nodes 143 fs/btrfs/inode-item.c leaf = path->nodes[0]; nodes 207 fs/btrfs/inode-item.c ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name, nodes 214 fs/btrfs/inode-item.c leaf = path->nodes[0]; nodes 277 fs/btrfs/inode-item.c if (btrfs_find_name_in_ext_backref(path->nodes[0], nodes 289 fs/btrfs/inode-item.c leaf = path->nodes[0]; nodes 295 fs/btrfs/inode-item.c btrfs_set_inode_extref_name_len(path->nodes[0], extref, name_len); nodes 296 fs/btrfs/inode-item.c btrfs_set_inode_extref_index(path->nodes[0], extref, index); nodes 297 fs/btrfs/inode-item.c btrfs_set_inode_extref_parent(path->nodes[0], extref, ref_objectid); nodes 300 fs/btrfs/inode-item.c write_extent_buffer(path->nodes[0], name, ptr, name_len); nodes 301 fs/btrfs/inode-item.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 336 fs/btrfs/inode-item.c ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], nodes 341 fs/btrfs/inode-item.c old_size = btrfs_item_size_nr(path->nodes[0], path->slots[0]); nodes 343 fs/btrfs/inode-item.c ref = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 346 fs/btrfs/inode-item.c btrfs_set_inode_ref_name_len(path->nodes[0], ref, name_len); nodes 347 fs/btrfs/inode-item.c btrfs_set_inode_ref_index(path->nodes[0], ref, index); nodes 352 fs/btrfs/inode-item.c if (btrfs_find_name_in_backref(path->nodes[0], nodes 361 fs/btrfs/inode-item.c ref = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 363 fs/btrfs/inode-item.c btrfs_set_inode_ref_name_len(path->nodes[0], ref, name_len); nodes 364 fs/btrfs/inode-item.c btrfs_set_inode_ref_index(path->nodes[0], ref, index); nodes 367 fs/btrfs/inode-item.c write_extent_buffer(path->nodes[0], name, ptr, name_len); nodes 368 fs/btrfs/inode-item.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 418 fs/btrfs/inode-item.c leaf = path->nodes[0]; nodes 70 fs/btrfs/inode-map.c leaf = path->nodes[0]; nodes 81 fs/btrfs/inode-map.c leaf = path->nodes[0]; nodes 549 fs/btrfs/inode-map.c l = path->nodes[0]; nodes 205 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 1386 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 1396 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 1406 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 2405 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 2590 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 2805 fs/btrfs/inode.c extent = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 2808 fs/btrfs/inode.c if (btrfs_file_extent_generation(path->nodes[0], extent) != nodes 2843 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 2873 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 3024 fs/btrfs/inode.c l = path->nodes[0]; nodes 3588 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 3827 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 4055 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 4324 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 4355 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 4408 fs/btrfs/inode.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &key); nodes 4431 fs/btrfs/inode.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 4811 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 5684 fs/btrfs/inode.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, location); nodes 5694 fs/btrfs/inode.c *type = btrfs_dir_type(path->nodes[0], di); nodes 5737 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 6137 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 6331 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 6531 fs/btrfs/inode.c inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 6533 fs/btrfs/inode.c memzero_extent_buffer(path->nodes[0], (unsigned long)inode_item, nodes 6535 fs/btrfs/inode.c fill_inode_item(trans, path->nodes[0], inode_item, inode); nodes 6538 fs/btrfs/inode.c ref = btrfs_item_ptr(path->nodes[0], path->slots[0] + 1, nodes 6540 fs/btrfs/inode.c btrfs_set_inode_ref_name_len(path->nodes[0], ref, name_len); nodes 6541 fs/btrfs/inode.c btrfs_set_inode_ref_index(path->nodes[0], ref, *index); nodes 6543 fs/btrfs/inode.c write_extent_buffer(path->nodes[0], name, ptr, name_len); nodes 6546 fs/btrfs/inode.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 6969 fs/btrfs/inode.c struct extent_buffer *leaf = path->nodes[0]; nodes 7089 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 7144 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 7483 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 10424 fs/btrfs/inode.c leaf = path->nodes[0]; nodes 1081 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2049 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2343 fs/btrfs/ioctl.c l = path->nodes[0]; nodes 2436 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2463 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2514 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2701 fs/btrfs/ioctl.c btrfs_header_nritems(path->nodes[0])) { nodes 2711 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 2783 fs/btrfs/ioctl.c btrfs_header_nritems(path->nodes[0])) { nodes 2793 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 3391 fs/btrfs/ioctl.c if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { nodes 3398 fs/btrfs/ioctl.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 3413 fs/btrfs/ioctl.c ei = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 3419 fs/btrfs/ioctl.c if (btrfs_file_extent_type(path->nodes[0], ei) == nodes 3423 fs/btrfs/ioctl.c ext_len = btrfs_file_extent_num_bytes(path->nodes[0], ei); nodes 3431 fs/btrfs/ioctl.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 3474 fs/btrfs/ioctl.c write_extent_buffer(path->nodes[0], inline_data, nodes 3475 fs/btrfs/ioctl.c btrfs_item_ptr_offset(path->nodes[0], nodes 3555 fs/btrfs/ioctl.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 3561 fs/btrfs/ioctl.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 3569 fs/btrfs/ioctl.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 3571 fs/btrfs/ioctl.c leaf = path->nodes[0]; nodes 4022 fs/btrfs/ioctl.c btrfs_set_dir_item_key(path->nodes[0], di, &disk_key); nodes 4023 fs/btrfs/ioctl.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 138 fs/btrfs/props.c leaf = path->nodes[0]; nodes 371 fs/btrfs/qgroup.c l = path->nodes[0]; nodes 461 fs/btrfs/qgroup.c l = path->nodes[0]; nodes 551 fs/btrfs/qgroup.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 620 fs/btrfs/qgroup.c leaf = path->nodes[0]; nodes 639 fs/btrfs/qgroup.c leaf = path->nodes[0]; nodes 728 fs/btrfs/qgroup.c l = path->nodes[0]; nodes 774 fs/btrfs/qgroup.c l = path->nodes[0]; nodes 816 fs/btrfs/qgroup.c l = path->nodes[0]; nodes 857 fs/btrfs/qgroup.c leaf = path->nodes[0]; nodes 946 fs/btrfs/qgroup.c leaf = path->nodes[0]; nodes 973 fs/btrfs/qgroup.c leaf = path->nodes[0]; nodes 1702 fs/btrfs/qgroup.c eb = path->nodes[level]; nodes 1717 fs/btrfs/qgroup.c path->nodes[level] = NULL; nodes 1732 fs/btrfs/qgroup.c eb = path->nodes[root_level]; nodes 1810 fs/btrfs/qgroup.c btrfs_node_key_to_cpu(dst_path->nodes[dst_level], &key, 0); nodes 1812 fs/btrfs/qgroup.c btrfs_item_key_to_cpu(dst_path->nodes[dst_level], &key, 0); nodes 1816 fs/btrfs/qgroup.c src_path->nodes[root_level] = src_eb; nodes 1825 fs/btrfs/qgroup.c if (src_path->nodes[cur_level] == NULL) { nodes 1832 fs/btrfs/qgroup.c eb = src_path->nodes[cur_level + 1]; nodes 1849 fs/btrfs/qgroup.c src_path->nodes[cur_level] = eb; nodes 1858 fs/btrfs/qgroup.c btrfs_node_key_to_cpu(dst_path->nodes[cur_level], nodes 1860 fs/btrfs/qgroup.c btrfs_node_key_to_cpu(src_path->nodes[cur_level], nodes 1863 fs/btrfs/qgroup.c btrfs_item_key_to_cpu(dst_path->nodes[cur_level], nodes 1865 fs/btrfs/qgroup.c btrfs_item_key_to_cpu(src_path->nodes[cur_level], nodes 1880 fs/btrfs/qgroup.c ret = btrfs_qgroup_trace_extent(trans, src_path->nodes[dst_level]->start, nodes 1885 fs/btrfs/qgroup.c dst_path->nodes[dst_level]->start, nodes 1892 fs/btrfs/qgroup.c ret = btrfs_qgroup_trace_leaf_items(trans, src_path->nodes[0]); nodes 1895 fs/btrfs/qgroup.c ret = btrfs_qgroup_trace_leaf_items(trans, dst_path->nodes[0]); nodes 1947 fs/btrfs/qgroup.c if (dst_path->nodes[cur_level] == NULL) { nodes 1968 fs/btrfs/qgroup.c eb = dst_path->nodes[cur_level + 1]; nodes 1989 fs/btrfs/qgroup.c dst_path->nodes[cur_level] = eb; nodes 2004 fs/btrfs/qgroup.c eb = dst_path->nodes[cur_level]; nodes 2026 fs/btrfs/qgroup.c btrfs_tree_unlock_rw(dst_path->nodes[cur_level], nodes 2028 fs/btrfs/qgroup.c free_extent_buffer(dst_path->nodes[cur_level]); nodes 2029 fs/btrfs/qgroup.c dst_path->nodes[cur_level] = NULL; nodes 2072 fs/btrfs/qgroup.c dst_path->nodes[level] = dst_eb; nodes 2131 fs/btrfs/qgroup.c path->nodes[root_level] = root_eb; nodes 2137 fs/btrfs/qgroup.c if (path->nodes[level] == NULL) { nodes 2147 fs/btrfs/qgroup.c eb = path->nodes[level + 1]; nodes 2164 fs/btrfs/qgroup.c path->nodes[level] = eb; nodes 2180 fs/btrfs/qgroup.c path->nodes[level]); nodes 3038 fs/btrfs/qgroup.c for (i = 1; i < BTRFS_MAX_LEVEL && path->nodes[i]; i++) { nodes 3039 fs/btrfs/qgroup.c if (path->slots[i] != btrfs_header_nritems(path->nodes[i]) - 1) nodes 3088 fs/btrfs/qgroup.c btrfs_item_key_to_cpu(path->nodes[0], &found, nodes 3089 fs/btrfs/qgroup.c btrfs_header_nritems(path->nodes[0]) - 1); nodes 3092 fs/btrfs/qgroup.c scratch_leaf = btrfs_clone_extent_buffer(path->nodes[0]); nodes 435 fs/btrfs/ref-verify.c struct extent_buffer *leaf = path->nodes[0]; nodes 499 fs/btrfs/ref-verify.c struct extent_buffer *leaf = path->nodes[0]; nodes 561 fs/btrfs/ref-verify.c block_bytenr = btrfs_node_blockptr(path->nodes[level], nodes 563 fs/btrfs/ref-verify.c gen = btrfs_node_ptr_generation(path->nodes[level], nodes 565 fs/btrfs/ref-verify.c btrfs_node_key_to_cpu(path->nodes[level], &first_key, nodes 577 fs/btrfs/ref-verify.c path->nodes[level-1] = eb; nodes 596 fs/btrfs/ref-verify.c if (!path->nodes[l]) nodes 601 fs/btrfs/ref-verify.c btrfs_header_nritems(path->nodes[l])) { nodes 606 fs/btrfs/ref-verify.c btrfs_tree_unlock_rw(path->nodes[l], path->locks[l]); nodes 607 fs/btrfs/ref-verify.c free_extent_buffer(path->nodes[l]); nodes 608 fs/btrfs/ref-verify.c path->nodes[l] = NULL; nodes 994 fs/btrfs/ref-verify.c path->nodes[level] = eb; nodes 770 fs/btrfs/relocation.c eb = path1->nodes[0]; nodes 781 fs/btrfs/relocation.c eb = path1->nodes[0]; nodes 924 fs/btrfs/relocation.c eb = path2->nodes[level]; nodes 941 fs/btrfs/relocation.c if (!path2->nodes[level]) { nodes 957 fs/btrfs/relocation.c eb = path2->nodes[level]; nodes 1636 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 1791 fs/btrfs/relocation.c btrfs_node_key(path->nodes[level], &key2, path->slots[level]); nodes 1832 fs/btrfs/relocation.c btrfs_node_key_to_cpu(path->nodes[lowest_level], &key, slot); nodes 1878 fs/btrfs/relocation.c eb = path->nodes[level]; nodes 1932 fs/btrfs/relocation.c btrfs_node_key_to_cpu(path->nodes[level], &key, nodes 1957 fs/btrfs/relocation.c path->nodes[level], path->slots[level], nodes 1968 fs/btrfs/relocation.c btrfs_set_node_blockptr(path->nodes[level], nodes 1970 fs/btrfs/relocation.c btrfs_set_node_ptr_generation(path->nodes[level], nodes 1972 fs/btrfs/relocation.c btrfs_mark_buffer_dirty(path->nodes[level]); nodes 1975 fs/btrfs/relocation.c blocksize, path->nodes[level]->start); nodes 1988 fs/btrfs/relocation.c blocksize, path->nodes[level]->start); nodes 2026 fs/btrfs/relocation.c free_extent_buffer(path->nodes[i]); nodes 2027 fs/btrfs/relocation.c path->nodes[i] = NULL; nodes 2030 fs/btrfs/relocation.c for (i = *level; i < BTRFS_MAX_LEVEL && path->nodes[i]; i++) { nodes 2031 fs/btrfs/relocation.c eb = path->nodes[i]; nodes 2042 fs/btrfs/relocation.c free_extent_buffer(path->nodes[i]); nodes 2043 fs/btrfs/relocation.c path->nodes[i] = NULL; nodes 2068 fs/btrfs/relocation.c eb = path->nodes[i]; nodes 2098 fs/btrfs/relocation.c path->nodes[i - 1] = eb; nodes 2182 fs/btrfs/relocation.c if (!path->nodes[level]) nodes 2185 fs/btrfs/relocation.c btrfs_header_nritems(path->nodes[level])) { nodes 2186 fs/btrfs/relocation.c btrfs_node_key_to_cpu(path->nodes[level], key, nodes 2293 fs/btrfs/relocation.c path->nodes[level] = reloc_root->node; nodes 2308 fs/btrfs/relocation.c btrfs_node_key_to_cpu(path->nodes[level], &next_key, nodes 2358 fs/btrfs/relocation.c btrfs_node_key_to_cpu(path->nodes[level], &key, nodes 2372 fs/btrfs/relocation.c btrfs_node_key(path->nodes[level], &root_item->drop_progress, nodes 2854 fs/btrfs/relocation.c upper->eb = path->nodes[upper->level]; nodes 2855 fs/btrfs/relocation.c path->nodes[upper->level] = NULL; nodes 2857 fs/btrfs/relocation.c BUG_ON(upper->eb != path->nodes[upper->level]); nodes 3461 fs/btrfs/relocation.c eb = path->nodes[0]; nodes 3545 fs/btrfs/relocation.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 3562 fs/btrfs/relocation.c btrfs_print_leaf(path->nodes[0]); nodes 3714 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 3742 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 3832 fs/btrfs/relocation.c eb = path->nodes[0]; nodes 3865 fs/btrfs/relocation.c eb = path->nodes[0]; nodes 3874 fs/btrfs/relocation.c eb = path->nodes[0]; nodes 3941 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 3946 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 4125 fs/btrfs/relocation.c ei = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 4127 fs/btrfs/relocation.c item_size = btrfs_item_size_nr(path->nodes[0], path->slots[0]); nodes 4129 fs/btrfs/relocation.c flags = btrfs_extent_flags(path->nodes[0], ei); nodes 4252 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 4532 fs/btrfs/relocation.c leaf = path->nodes[0]; nodes 90 fs/btrfs/root-tree.c l = path->nodes[0]; nodes 150 fs/btrfs/root-tree.c l = path->nodes[0]; nodes 181 fs/btrfs/root-tree.c l = path->nodes[0]; nodes 193 fs/btrfs/root-tree.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 238 fs/btrfs/root-tree.c leaf = path->nodes[0]; nodes 245 fs/btrfs/root-tree.c leaf = path->nodes[0]; nodes 376 fs/btrfs/root-tree.c leaf = path->nodes[0]; nodes 452 fs/btrfs/root-tree.c leaf = path->nodes[0]; nodes 675 fs/btrfs/scrub.c eb = swarn->path->nodes[0]; nodes 767 fs/btrfs/scrub.c eb = path->nodes[0]; nodes 2899 fs/btrfs/scrub.c l = path->nodes[0]; nodes 3251 fs/btrfs/scrub.c l = path->nodes[0]; nodes 3505 fs/btrfs/scrub.c btrfs_header_nritems(path->nodes[0])) { nodes 3518 fs/btrfs/scrub.c l = path->nodes[0]; nodes 842 fs/btrfs/send.c ii = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 845 fs/btrfs/send.c *size = btrfs_inode_size(path->nodes[0], ii); nodes 847 fs/btrfs/send.c *gen = btrfs_inode_generation(path->nodes[0], ii); nodes 849 fs/btrfs/send.c *mode = btrfs_inode_mode(path->nodes[0], ii); nodes 851 fs/btrfs/send.c *uid = btrfs_inode_uid(path->nodes[0], ii); nodes 853 fs/btrfs/send.c *gid = btrfs_inode_gid(path->nodes[0], ii); nodes 855 fs/btrfs/send.c *rdev = btrfs_inode_rdev(path->nodes[0], ii); nodes 893 fs/btrfs/send.c struct extent_buffer *eb = path->nodes[0]; nodes 1041 fs/btrfs/send.c eb = path->nodes[0]; nodes 1164 fs/btrfs/send.c btrfs_item_key_to_cpu(p->nodes[0], &found_key, p->slots[0]); nodes 1308 fs/btrfs/send.c struct extent_buffer *eb = path->nodes[0]; nodes 1369 fs/btrfs/send.c ei = btrfs_item_ptr(tmp_path->nodes[0], tmp_path->slots[0], nodes 1378 fs/btrfs/send.c if (btrfs_extent_refs(tmp_path->nodes[0], ei) > SEND_MAX_EXTENT_REFS) { nodes 1516 fs/btrfs/send.c ei = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 1518 fs/btrfs/send.c type = btrfs_file_extent_type(path->nodes[0], ei); nodes 1519 fs/btrfs/send.c compression = btrfs_file_extent_compression(path->nodes[0], ei); nodes 1524 fs/btrfs/send.c len = btrfs_file_extent_ram_bytes(path->nodes[0], ei); nodes 1526 fs/btrfs/send.c ret = fs_path_add_from_extent_buffer(dest, path->nodes[0], off, len); nodes 1720 fs/btrfs/send.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &key); nodes 1726 fs/btrfs/send.c *found_type = btrfs_dir_type(path->nodes[0], di); nodes 1759 fs/btrfs/send.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 1770 fs/btrfs/send.c iref = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 1772 fs/btrfs/send.c len = btrfs_inode_ref_name_len(path->nodes[0], iref); nodes 1773 fs/btrfs/send.c ret = fs_path_add_from_extent_buffer(name, path->nodes[0], nodes 1779 fs/btrfs/send.c extref = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 1781 fs/btrfs/send.c len = btrfs_inode_extref_name_len(path->nodes[0], extref); nodes 1782 fs/btrfs/send.c ret = fs_path_add_from_extent_buffer(name, path->nodes[0], nodes 1784 fs/btrfs/send.c parent_dir = btrfs_inode_extref_parent(path->nodes[0], extref); nodes 2378 fs/btrfs/send.c leaf = path->nodes[0]; nodes 2555 fs/btrfs/send.c eb = path->nodes[0]; nodes 2696 fs/btrfs/send.c eb = path->nodes[0]; nodes 2965 fs/btrfs/send.c if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { nodes 2973 fs/btrfs/send.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 2979 fs/btrfs/send.c di = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 2981 fs/btrfs/send.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &loc); nodes 3499 fs/btrfs/send.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &di_key); nodes 3610 fs/btrfs/send.c struct extent_buffer *leaf = path->nodes[0]; nodes 4450 fs/btrfs/send.c eb = path->nodes[0]; nodes 4756 fs/btrfs/send.c eb = path->nodes[0]; nodes 5175 fs/btrfs/send.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0] - 1); nodes 5182 fs/btrfs/send.c struct extent_buffer *leaf = path->nodes[0]; nodes 5344 fs/btrfs/send.c ei = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 5346 fs/btrfs/send.c type = btrfs_file_extent_type(path->nodes[0], ei); nodes 5348 fs/btrfs/send.c len = btrfs_file_extent_ram_bytes(path->nodes[0], ei); nodes 5356 fs/btrfs/send.c len = btrfs_file_extent_num_bytes(path->nodes[0], ei); nodes 5374 fs/btrfs/send.c disk_byte = btrfs_file_extent_disk_bytenr(path->nodes[0], ei); nodes 5375 fs/btrfs/send.c data_offset = btrfs_file_extent_offset(path->nodes[0], ei); nodes 5413 fs/btrfs/send.c eb = left_path->nodes[0]; nodes 5462 fs/btrfs/send.c eb = path->nodes[0]; nodes 5545 fs/btrfs/send.c eb = path->nodes[0]; nodes 5599 fs/btrfs/send.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 5603 fs/btrfs/send.c fi = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 5605 fs/btrfs/send.c type = btrfs_file_extent_type(path->nodes[0], fi); nodes 5607 fs/btrfs/send.c u64 size = btrfs_file_extent_ram_bytes(path->nodes[0], fi); nodes 5612 fs/btrfs/send.c btrfs_file_extent_num_bytes(path->nodes[0], fi); nodes 5644 fs/btrfs/send.c struct extent_buffer *leaf = path->nodes[0]; nodes 5712 fs/btrfs/send.c fi = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 5714 fs/btrfs/send.c type = btrfs_file_extent_type(path->nodes[0], fi); nodes 5716 fs/btrfs/send.c u64 size = btrfs_file_extent_ram_bytes(path->nodes[0], fi); nodes 5721 fs/btrfs/send.c btrfs_file_extent_num_bytes(path->nodes[0], fi); nodes 5775 fs/btrfs/send.c ei = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 5777 fs/btrfs/send.c type = btrfs_file_extent_type(path->nodes[0], ei); nodes 5792 fs/btrfs/send.c if (btrfs_file_extent_disk_bytenr(path->nodes[0], ei) == 0) { nodes 5836 fs/btrfs/send.c eb = path->nodes[0]; nodes 6070 fs/btrfs/send.c struct extent_buffer *eb = path->nodes[0]; nodes 6141 fs/btrfs/send.c left_ii = btrfs_item_ptr(sctx->left_path->nodes[0], nodes 6144 fs/btrfs/send.c left_gen = btrfs_inode_generation(sctx->left_path->nodes[0], nodes 6147 fs/btrfs/send.c right_ii = btrfs_item_ptr(sctx->right_path->nodes[0], nodes 6150 fs/btrfs/send.c right_gen = btrfs_inode_generation(sctx->right_path->nodes[0], nodes 6154 fs/btrfs/send.c right_ii = btrfs_item_ptr(sctx->right_path->nodes[0], nodes 6158 fs/btrfs/send.c right_gen = btrfs_inode_generation(sctx->right_path->nodes[0], nodes 6189 fs/btrfs/send.c nlinks = btrfs_inode_nlink(sctx->left_path->nodes[0], left_ii); nodes 6203 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6205 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6207 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6215 fs/btrfs/send.c sctx->right_path->nodes[0], right_ii); nodes 6217 fs/btrfs/send.c sctx->right_path->nodes[0], right_ii); nodes 6234 fs/btrfs/send.c sctx->right_path->nodes[0], right_ii); nodes 6236 fs/btrfs/send.c sctx->right_path->nodes[0], right_ii); nodes 6249 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6251 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6253 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6283 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6285 fs/btrfs/send.c sctx->left_path->nodes[0], left_ii); nodes 6426 fs/btrfs/send.c leaf = path->nodes[0]; nodes 6529 fs/btrfs/send.c eb = path->nodes[0]; nodes 6560 fs/btrfs/send.c eb = btrfs_read_node_slot(path->nodes[*level], path->slots[*level]); nodes 6564 fs/btrfs/send.c path->nodes[*level - 1] = eb; nodes 6575 fs/btrfs/send.c nritems = btrfs_header_nritems(path->nodes[*level]); nodes 6585 fs/btrfs/send.c free_extent_buffer(path->nodes[*level]); nodes 6586 fs/btrfs/send.c path->nodes[*level] = NULL; nodes 6590 fs/btrfs/send.c nritems = btrfs_header_nritems(path->nodes[*level]); nodes 6614 fs/btrfs/send.c btrfs_item_key_to_cpu(path->nodes[*level], key, nodes 6617 fs/btrfs/send.c btrfs_node_key_to_cpu(path->nodes[*level], key, nodes 6631 fs/btrfs/send.c len1 = btrfs_item_size_nr(left_path->nodes[0], left_path->slots[0]); nodes 6632 fs/btrfs/send.c len2 = btrfs_item_size_nr(right_path->nodes[0], right_path->slots[0]); nodes 6636 fs/btrfs/send.c off1 = btrfs_item_ptr_offset(left_path->nodes[0], left_path->slots[0]); nodes 6637 fs/btrfs/send.c off2 = btrfs_item_ptr_offset(right_path->nodes[0], nodes 6640 fs/btrfs/send.c read_extent_buffer(left_path->nodes[0], tmp_buf, off1, len1); nodes 6642 fs/btrfs/send.c cmp = memcmp_extent_buffer(right_path->nodes[0], tmp_buf, off2, len1); nodes 6747 fs/btrfs/send.c left_path->nodes[left_level] = nodes 6749 fs/btrfs/send.c if (!left_path->nodes[left_level]) { nodes 6757 fs/btrfs/send.c right_path->nodes[right_level] = nodes 6759 fs/btrfs/send.c if (!right_path->nodes[right_level]) { nodes 6767 fs/btrfs/send.c btrfs_item_key_to_cpu(left_path->nodes[left_level], nodes 6770 fs/btrfs/send.c btrfs_node_key_to_cpu(left_path->nodes[left_level], nodes 6773 fs/btrfs/send.c btrfs_item_key_to_cpu(right_path->nodes[right_level], nodes 6776 fs/btrfs/send.c btrfs_node_key_to_cpu(right_path->nodes[right_level], nodes 6855 fs/btrfs/send.c WARN_ON(!extent_buffer_uptodate(left_path->nodes[0])); nodes 6877 fs/btrfs/send.c left_path->nodes[left_level], nodes 6880 fs/btrfs/send.c right_path->nodes[right_level], nodes 6883 fs/btrfs/send.c left_path->nodes[left_level], nodes 6886 fs/btrfs/send.c right_path->nodes[right_level], nodes 1060 fs/btrfs/super.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1063 fs/btrfs/super.c root_ref = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 1065 fs/btrfs/super.c len = btrfs_root_ref_name_len(path->nodes[0], root_ref); nodes 1071 fs/btrfs/super.c read_extent_buffer(path->nodes[0], ptr + 1, nodes 1074 fs/btrfs/super.c dirid = btrfs_root_ref_dirid(path->nodes[0], root_ref); nodes 1109 fs/btrfs/super.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1112 fs/btrfs/super.c inode_ref = btrfs_item_ptr(path->nodes[0], nodes 1115 fs/btrfs/super.c len = btrfs_inode_ref_name_len(path->nodes[0], nodes 1122 fs/btrfs/super.c read_extent_buffer(path->nodes[0], ptr + 1, nodes 1179 fs/btrfs/super.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &location); nodes 51 fs/btrfs/tests/extent-buffer-tests.c path->nodes[0] = eb = alloc_dummy_extent_buffer(fs_info, nodesize); nodes 40 fs/btrfs/tests/free-space-tree-tests.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 41 fs/btrfs/tests/free-space-tree-tests.c extent_count = btrfs_free_space_extent_count(path->nodes[0], info); nodes 53 fs/btrfs/tests/free-space-tree-tests.c while (++path->slots[0] < btrfs_header_nritems(path->nodes[0])) { nodes 54 fs/btrfs/tests/free-space-tree-tests.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 85 fs/btrfs/tests/free-space-tree-tests.c if (btrfs_header_nritems(path->nodes[0]) != num_extents + 1 || nodes 90 fs/btrfs/tests/free-space-tree-tests.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 125 fs/btrfs/tests/free-space-tree-tests.c flags = btrfs_free_space_flags(path->nodes[0], info); nodes 29 fs/btrfs/tests/inode-tests.c path.nodes[0] = leaf; nodes 60 fs/btrfs/tests/inode-tests.c path.nodes[0] = leaf; nodes 47 fs/btrfs/tests/qgroup-tests.c leaf = path->nodes[0]; nodes 97 fs/btrfs/tests/qgroup-tests.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 99 fs/btrfs/tests/qgroup-tests.c refs = btrfs_extent_refs(path->nodes[0], item); nodes 100 fs/btrfs/tests/qgroup-tests.c btrfs_set_extent_refs(path->nodes[0], item, refs + 1); nodes 181 fs/btrfs/tests/qgroup-tests.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 183 fs/btrfs/tests/qgroup-tests.c refs = btrfs_extent_refs(path->nodes[0], item); nodes 184 fs/btrfs/tests/qgroup-tests.c btrfs_set_extent_refs(path->nodes[0], item, refs - 1); nodes 90 fs/btrfs/tree-defrag.c if (!path->nodes[1]) { nodes 101 fs/btrfs/tree-defrag.c path->nodes[1], 0, nodes 117 fs/btrfs/tree-defrag.c path->slots[1] = btrfs_header_nritems(path->nodes[1]); nodes 371 fs/btrfs/tree-log.c u32 dst_size = btrfs_item_size_nr(path->nodes[0], nodes 391 fs/btrfs/tree-log.c dst_ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); nodes 392 fs/btrfs/tree-log.c read_extent_buffer(path->nodes[0], dst_copy, dst_ptr, nodes 418 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 420 fs/btrfs/tree-log.c nbytes = btrfs_inode_nbytes(path->nodes[0], item); nodes 465 fs/btrfs/tree-log.c found_size = btrfs_item_size_nr(path->nodes[0], nodes 474 fs/btrfs/tree-log.c dst_ptr = btrfs_item_ptr_offset(path->nodes[0], nodes 494 fs/btrfs/tree-log.c struct extent_buffer *dst_eb = path->nodes[0]; nodes 518 fs/btrfs/tree-log.c S_ISDIR(btrfs_inode_mode(path->nodes[0], dst_item))) { nodes 520 fs/btrfs/tree-log.c saved_i_size = btrfs_inode_size(path->nodes[0], nodes 525 fs/btrfs/tree-log.c copy_extent_buffer(path->nodes[0], eb, dst_ptr, nodes 531 fs/btrfs/tree-log.c btrfs_set_inode_size(path->nodes[0], dst_item, saved_i_size); nodes 538 fs/btrfs/tree-log.c if (btrfs_inode_generation(path->nodes[0], dst_item) == 0) { nodes 539 fs/btrfs/tree-log.c btrfs_set_inode_generation(path->nodes[0], dst_item, nodes 544 fs/btrfs/tree-log.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 642 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 681 fs/btrfs/tree-log.c dest_offset = btrfs_item_ptr_offset(path->nodes[0], nodes 683 fs/btrfs/tree-log.c copy_extent_buffer(path->nodes[0], eb, dest_offset, nodes 863 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 913 fs/btrfs/tree-log.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &location); nodes 922 fs/btrfs/tree-log.c btrfs_dir_item_key_to_cpu(path->nodes[0], di, &location); nodes 966 fs/btrfs/tree-log.c ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); nodes 969 fs/btrfs/tree-log.c if (btrfs_find_name_in_ext_backref(path->nodes[0], nodes 978 fs/btrfs/tree-log.c item_size = btrfs_item_size_nr(path->nodes[0], path->slots[0]); nodes 982 fs/btrfs/tree-log.c found_name_len = btrfs_inode_ref_name_len(path->nodes[0], ref); nodes 985 fs/btrfs/tree-log.c ret = memcmp_extent_buffer(path->nodes[0], name, nodes 1028 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 1095 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 1251 fs/btrfs/tree-log.c eb = path->nodes[0]; nodes 1338 fs/btrfs/tree-log.c ret = !!btrfs_find_name_in_ext_backref(path->nodes[0], nodes 1341 fs/btrfs/tree-log.c ret = !!btrfs_find_name_in_backref(path->nodes[0], path->slots[0], nodes 1379 fs/btrfs/tree-log.c btrfs_dir_item_key_to_cpu(path->nodes[0], dir_item, &key); nodes 1618 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 1667 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &key, nodes 1672 fs/btrfs/tree-log.c ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); nodes 1673 fs/btrfs/tree-log.c ptr_end = ptr + btrfs_item_size_nr(path->nodes[0], nodes 1679 fs/btrfs/tree-log.c name_len = btrfs_inode_ref_name_len(path->nodes[0], nodes 1779 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 1992 fs/btrfs/tree-log.c btrfs_dir_item_key_to_cpu(path->nodes[0], dst_di, &found_key); nodes 1997 fs/btrfs/tree-log.c btrfs_dir_type(path->nodes[0], dst_di) == log_type) { nodes 2166 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 2172 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 2174 fs/btrfs/tree-log.c found_end = btrfs_dir_log_end(path->nodes[0], item); nodes 2185 fs/btrfs/tree-log.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 2193 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); nodes 2199 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 2201 fs/btrfs/tree-log.c found_end = btrfs_dir_log_end(path->nodes[0], item); nodes 2237 fs/btrfs/tree-log.c eb = path->nodes[0]; nodes 2342 fs/btrfs/tree-log.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 2350 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &key, i); nodes 2356 fs/btrfs/tree-log.c di = btrfs_item_ptr(path->nodes[0], i, struct btrfs_dir_item); nodes 2357 fs/btrfs/tree-log.c total_size = btrfs_item_size_nr(path->nodes[0], i); nodes 2360 fs/btrfs/tree-log.c u16 name_len = btrfs_dir_name_len(path->nodes[0], di); nodes 2361 fs/btrfs/tree-log.c u16 data_len = btrfs_dir_data_len(path->nodes[0], di); nodes 2370 fs/btrfs/tree-log.c read_extent_buffer(path->nodes[0], name, nodes 2477 fs/btrfs/tree-log.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 2485 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 2703 fs/btrfs/tree-log.c cur = path->nodes[*level]; nodes 2716 fs/btrfs/tree-log.c parent = path->nodes[*level]; nodes 2771 fs/btrfs/tree-log.c if (path->nodes[*level-1]) nodes 2772 fs/btrfs/tree-log.c free_extent_buffer(path->nodes[*level-1]); nodes 2773 fs/btrfs/tree-log.c path->nodes[*level-1] = next; nodes 2781 fs/btrfs/tree-log.c path->slots[*level] = btrfs_header_nritems(path->nodes[*level]); nodes 2798 fs/btrfs/tree-log.c for (i = *level; i < BTRFS_MAX_LEVEL - 1 && path->nodes[i]; i++) { nodes 2800 fs/btrfs/tree-log.c if (slot + 1 < btrfs_header_nritems(path->nodes[i])) { nodes 2807 fs/btrfs/tree-log.c if (path->nodes[*level] == root->node) nodes 2808 fs/btrfs/tree-log.c parent = path->nodes[*level]; nodes 2810 fs/btrfs/tree-log.c parent = path->nodes[*level + 1]; nodes 2813 fs/btrfs/tree-log.c ret = wc->process_func(root, path->nodes[*level], wc, nodes 2814 fs/btrfs/tree-log.c btrfs_header_generation(path->nodes[*level]), nodes 2822 fs/btrfs/tree-log.c next = path->nodes[*level]; nodes 2838 fs/btrfs/tree-log.c path->nodes[*level]->start, nodes 2839 fs/btrfs/tree-log.c path->nodes[*level]->len); nodes 2843 fs/btrfs/tree-log.c free_extent_buffer(path->nodes[*level]); nodes 2844 fs/btrfs/tree-log.c path->nodes[*level] = NULL; nodes 2872 fs/btrfs/tree-log.c path->nodes[level] = log->node; nodes 2895 fs/btrfs/tree-log.c if (path->nodes[orig_level]) { nodes 2896 fs/btrfs/tree-log.c ret = wc->process_func(log, path->nodes[orig_level], wc, nodes 2897 fs/btrfs/tree-log.c btrfs_header_generation(path->nodes[orig_level]), nodes 2904 fs/btrfs/tree-log.c next = path->nodes[orig_level]; nodes 3468 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 3470 fs/btrfs/tree-log.c i_size = btrfs_inode_size(path->nodes[0], item); nodes 3475 fs/btrfs/tree-log.c btrfs_set_inode_size(path->nodes[0], item, i_size); nodes 3476 fs/btrfs/tree-log.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 3553 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 3555 fs/btrfs/tree-log.c btrfs_set_dir_log_end(path->nodes[0], item, last_offset); nodes 3556 fs/btrfs/tree-log.c btrfs_mark_buffer_dirty(path->nodes[0]); nodes 3615 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &tmp, nodes 3627 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); nodes 3631 fs/btrfs/tree-log.c path->nodes[0], path->slots[0], nodes 3659 fs/btrfs/tree-log.c src = path->nodes[0]; nodes 3720 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); nodes 3725 fs/btrfs/tree-log.c if (btrfs_header_generation(path->nodes[0]) != trans->transid) { nodes 3727 fs/btrfs/tree-log.c path->nodes[0], path->slots[0], nodes 3827 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 3835 fs/btrfs/tree-log.c ret = btrfs_bin_search(path->nodes[0], &found_key, 0, nodes 3923 fs/btrfs/tree-log.c inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 3925 fs/btrfs/tree-log.c fill_inode_item(trans, path->nodes[0], inode_item, &inode->vfs_inode, nodes 3966 fs/btrfs/tree-log.c struct extent_buffer *src = src_path->nodes[0]; nodes 3997 fs/btrfs/tree-log.c dst_offset = btrfs_item_ptr_offset(dst_path->nodes[0], nodes 4003 fs/btrfs/tree-log.c inode_item = btrfs_item_ptr(dst_path->nodes[0], nodes 4006 fs/btrfs/tree-log.c fill_inode_item(trans, dst_path->nodes[0], inode_item, nodes 4011 fs/btrfs/tree-log.c copy_extent_buffer(dst_path->nodes[0], src, dst_offset, nodes 4061 fs/btrfs/tree-log.c btrfs_mark_buffer_dirty(dst_path->nodes[0]); nodes 4178 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 4277 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 4297 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 4485 fs/btrfs/tree-log.c item = btrfs_item_ptr(path->nodes[0], path->slots[0], nodes 4487 fs/btrfs/tree-log.c *size_ret = btrfs_inode_size(path->nodes[0], item); nodes 4538 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 4611 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 4614 fs/btrfs/tree-log.c if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { nodes 4622 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 4658 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 4799 fs/btrfs/tree-log.c btrfs_dir_item_key_to_cpu(search_path->nodes[0], nodes 4962 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 5199 fs/btrfs/tree-log.c ret = btrfs_check_ref_name_override(path->nodes[0], nodes 5267 fs/btrfs/tree-log.c nritems = btrfs_header_nritems(path->nodes[0]); nodes 5270 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &min_key, nodes 5593 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 5700 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 5803 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, path->slots[0]); nodes 5808 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 5839 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 5847 fs/btrfs/tree-log.c leaf = path->nodes[0]; nodes 5930 fs/btrfs/tree-log.c struct extent_buffer *leaf = path->nodes[0]; nodes 6212 fs/btrfs/tree-log.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 49 fs/btrfs/ulist.c INIT_LIST_HEAD(&ulist->nodes); nodes 66 fs/btrfs/ulist.c list_for_each_entry_safe(node, next, &ulist->nodes, list) { nodes 70 fs/btrfs/ulist.c INIT_LIST_HEAD(&ulist->nodes); nodes 212 fs/btrfs/ulist.c list_add_tail(&node->list, &ulist->nodes); nodes 265 fs/btrfs/ulist.c if (list_empty(&ulist->nodes)) nodes 267 fs/btrfs/ulist.c if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) nodes 272 fs/btrfs/ulist.c uiter->cur_list = ulist->nodes.next; nodes 41 fs/btrfs/ulist.h struct list_head nodes; nodes 53 fs/btrfs/uuid-tree.c eb = path->nodes[0]; nodes 116 fs/btrfs/uuid-tree.c eb = path->nodes[0]; nodes 125 fs/btrfs/uuid-tree.c eb = path->nodes[0]; nodes 187 fs/btrfs/uuid-tree.c eb = path->nodes[0]; nodes 282 fs/btrfs/uuid-tree.c leaf = path->nodes[0]; nodes 1654 fs/btrfs/volumes.c l = path->nodes[0]; nodes 1788 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 1798 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 1846 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 1914 fs/btrfs/volumes.c btrfs_item_key_to_cpu(path->nodes[0], &found_key, nodes 1951 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 2547 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 2555 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 2837 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 3183 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 3289 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 3760 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 4342 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 4497 fs/btrfs/volumes.c eb = path->nodes[0]; nodes 4802 fs/btrfs/volumes.c l = path->nodes[0]; nodes 7273 fs/btrfs/volumes.c leaf = path->nodes[0]; nodes 7403 fs/btrfs/volumes.c eb = path->nodes[0]; nodes 7455 fs/btrfs/volumes.c btrfs_item_size_nr(path->nodes[0], path->slots[0]) < sizeof(*ptr)) { nodes 7480 fs/btrfs/volumes.c eb = path->nodes[0]; nodes 7846 fs/btrfs/volumes.c if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { nodes 7857 fs/btrfs/volumes.c struct extent_buffer *leaf = path->nodes[0]; nodes 48 fs/btrfs/xattr.c leaf = path->nodes[0]; nodes 141 fs/btrfs/xattr.c btrfs_assert_tree_locked(path->nodes[0]); nodes 169 fs/btrfs/xattr.c struct extent_buffer *leaf = path->nodes[0]; nodes 285 fs/btrfs/xattr.c leaf = path->nodes[0]; nodes 780 fs/dlm/config.c struct dlm_config_node *nodes, *node; nodes 796 fs/dlm/config.c nodes = kcalloc(count, sizeof(struct dlm_config_node), GFP_NOFS); nodes 797 fs/dlm/config.c if (!nodes) { nodes 802 fs/dlm/config.c node = nodes; nodes 814 fs/dlm/config.c *nodes_out = nodes; nodes 136 fs/dlm/dlm_internal.h struct dlm_config_node *nodes; nodes 518 fs/dlm/member.c if (rv->nodes[i].nodeid == nodeid) nodes 519 fs/dlm/member.c return &rv->nodes[i]; nodes 562 fs/dlm/member.c node = &rv->nodes[i]; nodes 672 fs/dlm/member.c struct dlm_config_node *nodes = NULL; nodes 679 fs/dlm/member.c error = dlm_config_nodes(ls->ls_name, &nodes, &count); nodes 694 fs/dlm/member.c rv->nodes = nodes; nodes 704 fs/dlm/member.c kfree(rv_old->nodes); nodes 713 fs/dlm/member.c kfree(nodes); nodes 269 fs/dlm/recoverd.c kfree(rv->nodes); nodes 406 fs/hpfs/anode.c int i, j, nodes; nodes 428 fs/hpfs/anode.c nodes = btree->n_used_nodes + btree->n_free_nodes; nodes 438 fs/hpfs/anode.c btree->n_free_nodes = nodes - btree->n_used_nodes; nodes 453 fs/hpfs/anode.c nodes = btree->n_used_nodes + btree->n_free_nodes; nodes 472 fs/hpfs/anode.c btree->n_free_nodes = nodes - btree->n_used_nodes; nodes 248 fs/jffs2/build.c raw = ic->nodes; nodes 223 fs/jffs2/erase.c prev = &ic->nodes; nodes 259 fs/jffs2/erase.c this = ic->nodes; nodes 284 fs/jffs2/erase.c if (ic->nodes == (void *)ic && ic->pino_nlink == 0) nodes 549 fs/jffs2/gc.c frag->node->raw = f->inocache->nodes; nodes 924 fs/jffs2/gc.c for (raw = f->inocache->nodes; raw != (void *)f->inocache; raw = raw->next_in_ino) { nodes 627 fs/jffs2/nodelist.c ref->next_in_ino = ic->nodes; nodes 628 fs/jffs2/nodelist.c ic->nodes = ref; nodes 167 fs/jffs2/nodelist.h struct jffs2_raw_node_ref *nodes; nodes 810 fs/jffs2/nodemgmt.c for (p = &ic->nodes; (*p) != ref; p = &((*p)->next_in_ino)) nodes 826 fs/jffs2/nodemgmt.c if (ic->nodes == (void *)ic && ic->pino_nlink == 0) nodes 986 fs/jffs2/readinode.c valid_ref = jffs2_first_valid_node(f->inocache->nodes); nodes 1364 fs/jffs2/readinode.c f->inocache->nodes = (struct jffs2_raw_node_ref *)f->inocache; nodes 1426 fs/jffs2/readinode.c if (f->inocache->nodes == (void *)f->inocache) nodes 981 fs/jffs2/scan.c ic->nodes = (void *)ic; nodes 489 fs/jffs2/wbuf.c struct jffs2_raw_node_ref **p = &ic->nodes; nodes 37 fs/jffs2/write.c f->inocache->nodes = (struct jffs2_raw_node_ref *)f->inocache; nodes 1070 fs/ocfs2/cluster/heartbeat.c static int o2hb_highest_node(unsigned long *nodes, int numbits) nodes 1072 fs/ocfs2/cluster/heartbeat.c return find_last_bit(nodes, numbits); nodes 1075 fs/ocfs2/cluster/heartbeat.c static int o2hb_lowest_node(unsigned long *nodes, int numbits) nodes 1077 fs/ocfs2/cluster/heartbeat.c return find_first_bit(nodes, numbits); nodes 226 fs/ubifs/auth.c snod = list_first_entry(&sleb->nodes, struct ubifs_scan_node, list); nodes 802 fs/ubifs/debug.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 833 fs/ubifs/debug.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 240 fs/ubifs/gc.c list_for_each_entry_safe(snod, tmp, &sleb->nodes, list) { nodes 283 fs/ubifs/gc.c list_sort(c, &sleb->nodes, &data_nodes_cmp); nodes 286 fs/ubifs/gc.c err = dbg_check_data_nodes_order(c, &sleb->nodes); nodes 360 fs/ubifs/gc.c list_for_each_entry_safe(snod, tmp, &sleb->nodes, list) { nodes 439 fs/ubifs/gc.c if (list_empty(&sleb->nodes) && list_empty(&nondata)) nodes 454 fs/ubifs/gc.c list_splice_tail(&nondata, &sleb->nodes); nodes 545 fs/ubifs/gc.c ubifs_assert(c, !list_empty(&sleb->nodes)); nodes 546 fs/ubifs/gc.c snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list); nodes 553 fs/ubifs/gc.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 663 fs/ubifs/log.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 1113 fs/ubifs/lprops.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 106 fs/ubifs/master.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, nodes 124 fs/ubifs/master.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, list); nodes 638 fs/ubifs/orphan.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 950 fs/ubifs/orphan.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 497 fs/ubifs/recovery.c if (!list_empty(&sleb->nodes)) { nodes 500 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, nodes 564 fs/ubifs/recovery.c while (!list_empty(&sleb->nodes)) { nodes 568 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, nodes 595 fs/ubifs/recovery.c if (!list_empty(&sleb->nodes)) { nodes 596 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.prev, struct ubifs_scan_node, nodes 878 fs/ubifs/recovery.c snod = list_entry(sleb->nodes.next, nodes 610 fs/ubifs/replay.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 726 fs/ubifs/replay.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 1017 fs/ubifs/replay.c snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list); nodes 1069 fs/ubifs/replay.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 138 fs/ubifs/scan.c INIT_LIST_HEAD(&sleb->nodes); nodes 213 fs/ubifs/scan.c list_add_tail(&snod->list, &sleb->nodes); nodes 359 fs/ubifs/scan.c head = &sleb->nodes; nodes 251 fs/ubifs/tnc_commit.c list_for_each_entry(snod, &sleb->nodes, list) { nodes 322 fs/ubifs/ubifs.h struct list_head nodes; nodes 424 fs/unicode/mkutf8data.c int nodes, singletons, leaves; nodes 426 fs/unicode/mkutf8data.c nodes = singletons = leaves = 0; nodes 444 fs/unicode/mkutf8data.c nodes += 1; nodes 486 fs/unicode/mkutf8data.c nodes, leaves, singletons); nodes 1198 fs/unicode/mkutf8data.c int nodes[4]; nodes 1201 fs/unicode/mkutf8data.c nodes[0] = nodes[1] = nodes[2] = nodes[3] = 0; nodes 1242 fs/unicode/mkutf8data.c nodes[offlen]++; nodes 1255 fs/unicode/mkutf8data.c nodes[0]++; nodes 1262 fs/unicode/mkutf8data.c nodes[0]++; nodes 1316 fs/unicode/mkutf8data.c nodes[0] + nodes[1] + nodes[2] + nodes[3], nodes 1317 fs/unicode/mkutf8data.c nodes[0], nodes[1], nodes[2], nodes[3]); nodes 806 include/linux/compat.h const int __user *nodes, nodes 25 include/linux/interconnect-provider.h struct icc_node *nodes[]; nodes 48 include/linux/interconnect-provider.h struct list_head nodes; nodes 51 include/linux/mempolicy.h nodemask_t nodes; /* interleave/bind */ nodes 1024 include/linux/mmzone.h nodemask_t *nodes); nodes 1040 include/linux/mmzone.h nodemask_t *nodes) nodes 1042 include/linux/mmzone.h if (likely(!nodes && zonelist_zone_idx(z) <= highest_zoneidx)) nodes 1044 include/linux/mmzone.h return __next_zones_zonelist(z, highest_zoneidx, nodes); nodes 1065 include/linux/mmzone.h nodemask_t *nodes) nodes 1068 include/linux/mmzone.h highest_zoneidx, nodes); nodes 428 include/linux/property.h int software_node_register_nodes(const struct software_node *nodes); nodes 429 include/linux/property.h void software_node_unregister_nodes(const struct software_node *nodes); nodes 1045 include/linux/scif.h int scif_get_node_ids(u16 *nodes, int len, u16 *self); nodes 896 include/linux/syscalls.h const int __user *nodes, nodes 297 include/trace/events/bcache.h TP_PROTO(unsigned nodes), nodes 298 include/trace/events/bcache.h TP_ARGS(nodes), nodes 301 include/trace/events/bcache.h __field(unsigned, nodes ) nodes 305 include/trace/events/bcache.h __entry->nodes = nodes; nodes 308 include/trace/events/bcache.h TP_printk("coalesced %u nodes", __entry->nodes) nodes 193 include/uapi/linux/scif_ioctl.h __u64 nodes; nodes 86 kernel/irq/affinity.c int n, nodes = 0; nodes 92 kernel/irq/affinity.c nodes++; nodes 95 kernel/irq/affinity.c return nodes; nodes 255 kernel/irq/affinity.c unsigned int i, n, nodes, cpus_per_vec, extra_vecs, done = 0; nodes 264 kernel/irq/affinity.c nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); nodes 270 kernel/irq/affinity.c if (numvecs <= nodes) { nodes 359 kernel/power/snapshot.c struct list_head nodes; /* Radix Tree inner nodes */ nodes 447 kernel/power/snapshot.c &zone->nodes); nodes 470 kernel/power/snapshot.c &zone->nodes); nodes 513 kernel/power/snapshot.c INIT_LIST_HEAD(&zone->nodes); nodes 541 kernel/power/snapshot.c list_for_each_entry(node, &zone->nodes, list) nodes 918 kernel/power/snapshot.c list_for_each_entry(node, &zone->nodes, list) nodes 1180 kernel/power/snapshot.c unsigned int rtree, nodes; nodes 1182 kernel/power/snapshot.c rtree = nodes = DIV_ROUND_UP(zone->spanned_pages, BM_BITS_PER_BLOCK); nodes 1185 kernel/power/snapshot.c while (nodes > 1) { nodes 1186 kernel/power/snapshot.c nodes = DIV_ROUND_UP(nodes, BM_ENTRIES_PER_LEVEL); nodes 1187 kernel/power/snapshot.c rtree += nodes; nodes 2042 kernel/sched/fair.c nodemask_t nodes; nodes 2079 kernel/sched/fair.c nodes = node_online_map; nodes 2089 kernel/sched/fair.c for_each_node_mask(a, nodes) { nodes 2095 kernel/sched/fair.c for_each_node_mask(b, nodes) { nodes 2099 kernel/sched/fair.c node_clear(b, nodes); nodes 2118 kernel/sched/fair.c nodes = max_group; nodes 24 lib/interval_tree_test.c static struct interval_tree_node *nodes = NULL; nodes 49 lib/interval_tree_test.c nodes[i].start = a; nodes 50 lib/interval_tree_test.c nodes[i].last = b; nodes 68 lib/interval_tree_test.c nodes = kmalloc_array(nnodes, sizeof(struct interval_tree_node), nodes 70 lib/interval_tree_test.c if (!nodes) nodes 75 lib/interval_tree_test.c kfree(nodes); nodes 88 lib/interval_tree_test.c interval_tree_insert(nodes + j, &root); nodes 90 lib/interval_tree_test.c interval_tree_remove(nodes + j, &root); nodes 102 lib/interval_tree_test.c interval_tree_insert(nodes + j, &root); nodes 124 lib/interval_tree_test.c kfree(nodes); nodes 705 lib/objagg.c struct objagg_tmp_node *nodes; nodes 737 lib/objagg.c struct objagg_tmp_node *node = &graph->nodes[index]; nodes 748 lib/objagg.c node = &graph->nodes[j]; nodes 765 lib/objagg.c node = &graph->nodes[i]; nodes 791 lib/objagg.c graph->nodes = kcalloc(nodes_count, sizeof(*graph->nodes), GFP_KERNEL); nodes 792 lib/objagg.c if (!graph->nodes) nodes 804 lib/objagg.c node = &graph->nodes[i++]; nodes 815 lib/objagg.c pnode = &graph->nodes[i]; nodes 816 lib/objagg.c node = &graph->nodes[j]; nodes 828 lib/objagg.c kfree(graph->nodes); nodes 837 lib/objagg.c kfree(graph->nodes); nodes 860 lib/objagg.c node = &graph->nodes[index]; nodes 874 lib/objagg.c node = &graph->nodes[j]; nodes 72 lib/radix-tree.c struct radix_tree_node *nodes; nodes 274 lib/radix-tree.c ret = rtp->nodes; nodes 275 lib/radix-tree.c rtp->nodes = ret->parent; nodes 353 lib/radix-tree.c node->parent = rtp->nodes; nodes 354 lib/radix-tree.c rtp->nodes = node; nodes 1595 lib/radix-tree.c node = rtp->nodes; nodes 1596 lib/radix-tree.c rtp->nodes = node->parent; nodes 28 lib/rbtree_test.c static struct test_node *nodes = NULL; nodes 153 lib/rbtree_test.c nodes[i].key = prandom_u32_state(&rnd); nodes 154 lib/rbtree_test.c nodes[i].val = prandom_u32_state(&rnd); nodes 248 lib/rbtree_test.c nodes = kmalloc_array(nnodes, sizeof(*nodes), GFP_KERNEL); nodes 249 lib/rbtree_test.c if (!nodes) nodes 261 lib/rbtree_test.c insert(nodes + j, &root); nodes 263 lib/rbtree_test.c erase(nodes + j, &root); nodes 277 lib/rbtree_test.c insert_cached(nodes + j, &root); nodes 279 lib/rbtree_test.c erase_cached(nodes + j, &root); nodes 290 lib/rbtree_test.c insert(nodes + i, &root); nodes 330 lib/rbtree_test.c erase(nodes + i, &root); nodes 337 lib/rbtree_test.c insert(nodes + j, &root); nodes 341 lib/rbtree_test.c erase(nodes + j, &root); nodes 354 lib/rbtree_test.c insert_augmented(nodes + j, &root); nodes 356 lib/rbtree_test.c erase_augmented(nodes + j, &root); nodes 369 lib/rbtree_test.c insert_augmented_cached(nodes + j, &root); nodes 371 lib/rbtree_test.c erase_augmented_cached(nodes + j, &root); nodes 384 lib/rbtree_test.c insert_augmented(nodes + j, &root); nodes 388 lib/rbtree_test.c erase_augmented(nodes + j, &root); nodes 393 lib/rbtree_test.c kfree(nodes); nodes 150 mm/mempolicy.c int (*create)(struct mempolicy *pol, const nodemask_t *nodes); nodes 151 mm/mempolicy.c void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes); nodes 167 mm/mempolicy.c static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) nodes 169 mm/mempolicy.c if (nodes_empty(*nodes)) nodes 171 mm/mempolicy.c pol->v.nodes = *nodes; nodes 175 mm/mempolicy.c static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) nodes 177 mm/mempolicy.c if (!nodes) nodes 179 mm/mempolicy.c else if (nodes_empty(*nodes)) nodes 182 mm/mempolicy.c pol->v.preferred_node = first_node(*nodes); nodes 186 mm/mempolicy.c static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) nodes 188 mm/mempolicy.c if (nodes_empty(*nodes)) nodes 190 mm/mempolicy.c pol->v.nodes = *nodes; nodes 204 mm/mempolicy.c const nodemask_t *nodes, struct nodemask_scratch *nsc) nodes 215 mm/mempolicy.c VM_BUG_ON(!nodes); nodes 216 mm/mempolicy.c if (pol->mode == MPOL_PREFERRED && nodes_empty(*nodes)) nodes 217 mm/mempolicy.c nodes = NULL; /* explicit local allocation */ nodes 220 mm/mempolicy.c mpol_relative_nodemask(&nsc->mask2, nodes, &nsc->mask1); nodes 222 mm/mempolicy.c nodes_and(nsc->mask2, *nodes, nsc->mask1); nodes 225 mm/mempolicy.c pol->w.user_nodemask = *nodes; nodes 231 mm/mempolicy.c if (nodes) nodes 243 mm/mempolicy.c nodemask_t *nodes) nodes 248 mm/mempolicy.c mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE); nodes 251 mm/mempolicy.c if (nodes && !nodes_empty(*nodes)) nodes 255 mm/mempolicy.c VM_BUG_ON(!nodes); nodes 263 mm/mempolicy.c if (nodes_empty(*nodes)) { nodes 269 mm/mempolicy.c if (!nodes_empty(*nodes) || nodes 274 mm/mempolicy.c } else if (nodes_empty(*nodes)) nodes 294 mm/mempolicy.c static void mpol_rebind_default(struct mempolicy *pol, const nodemask_t *nodes) nodes 298 mm/mempolicy.c static void mpol_rebind_nodemask(struct mempolicy *pol, const nodemask_t *nodes) nodes 303 mm/mempolicy.c nodes_and(tmp, pol->w.user_nodemask, *nodes); nodes 305 mm/mempolicy.c mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); nodes 307 mm/mempolicy.c nodes_remap(tmp, pol->v.nodes,pol->w.cpuset_mems_allowed, nodes 308 mm/mempolicy.c *nodes); nodes 309 mm/mempolicy.c pol->w.cpuset_mems_allowed = *nodes; nodes 313 mm/mempolicy.c tmp = *nodes; nodes 315 mm/mempolicy.c pol->v.nodes = tmp; nodes 319 mm/mempolicy.c const nodemask_t *nodes) nodes 326 mm/mempolicy.c if (node_isset(node, *nodes)) { nodes 332 mm/mempolicy.c mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); nodes 337 mm/mempolicy.c *nodes); nodes 338 mm/mempolicy.c pol->w.cpuset_mems_allowed = *nodes; nodes 681 mm/mempolicy.c nodemask_t *nodes, unsigned long flags, nodes 687 mm/mempolicy.c .nmask = nodes, nodes 793 mm/mempolicy.c nodemask_t *nodes) nodes 802 mm/mempolicy.c new = mpol_new(mode, flags, nodes); nodes 809 mm/mempolicy.c ret = mpol_set_nodemask(new, nodes, scratch); nodes 832 mm/mempolicy.c static void get_policy_nodemask(struct mempolicy *p, nodemask_t *nodes) nodes 834 mm/mempolicy.c nodes_clear(*nodes); nodes 842 mm/mempolicy.c *nodes = p->v.nodes; nodes 846 mm/mempolicy.c node_set(p->v.preferred_node, *nodes); nodes 932 mm/mempolicy.c *policy = next_node_in(current->il_prev, pol->v.nodes); nodes 1327 mm/mempolicy.c static int get_nodes(nodemask_t *nodes, const unsigned long __user *nmask, nodes 1336 mm/mempolicy.c nodes_clear(*nodes); nodes 1381 mm/mempolicy.c if (copy_from_user(nodes_addr(*nodes), nmask, nlongs*sizeof(unsigned long))) nodes 1383 mm/mempolicy.c nodes_addr(*nodes)[nlongs-1] &= endmask; nodes 1389 mm/mempolicy.c nodemask_t *nodes) nodes 1401 mm/mempolicy.c return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0; nodes 1408 mm/mempolicy.c nodemask_t nodes; nodes 1420 mm/mempolicy.c err = get_nodes(&nodes, nmask, maxnode); nodes 1423 mm/mempolicy.c return do_mbind(start, len, mode, mode_flags, &nodes, flags); nodes 1438 mm/mempolicy.c nodemask_t nodes; nodes 1447 mm/mempolicy.c err = get_nodes(&nodes, nmask, maxnode); nodes 1450 mm/mempolicy.c return do_set_mempolicy(mode, flags, &nodes); nodes 1564 mm/mempolicy.c nodemask_t nodes; nodes 1571 mm/mempolicy.c err = do_get_mempolicy(&pval, &nodes, addr, flags); nodes 1580 mm/mempolicy.c err = copy_nodes_to_user(nmask, maxnode, &nodes); nodes 1786 mm/mempolicy.c if (!nodes_intersects(policy->v.nodes, node_states[N_HIGH_MEMORY])) nodes 1801 mm/mempolicy.c cpuset_nodemask_valid_mems_allowed(&policy->v.nodes)) nodes 1802 mm/mempolicy.c return &policy->v.nodes; nodes 1831 mm/mempolicy.c next = next_node_in(me->il_prev, policy->v.nodes); nodes 1874 mm/mempolicy.c &policy->v.nodes); nodes 1890 mm/mempolicy.c unsigned nnodes = nodes_weight(pol->v.nodes); nodes 1898 mm/mempolicy.c nid = first_node(pol->v.nodes); nodes 1900 mm/mempolicy.c nid = next_node(nid, pol->v.nodes); nodes 1956 mm/mempolicy.c *nodemask = &(*mpol)->v.nodes; nodes 1999 mm/mempolicy.c *mask = mempolicy->v.nodes; nodes 2045 mm/mempolicy.c ret = nodes_intersects(mempolicy->v.nodes, *mask); nodes 2262 mm/mempolicy.c return !!nodes_equal(a->v.nodes, b->v.nodes); nodes 2422 mm/mempolicy.c if (node_isset(curnid, pol->v.nodes)) nodes 2427 mm/mempolicy.c &pol->v.nodes); nodes 2630 mm/mempolicy.c npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE); nodes 2800 mm/mempolicy.c nodemask_t nodes; nodes 2811 mm/mempolicy.c if (nodelist_parse(nodelist, nodes)) nodes 2813 mm/mempolicy.c if (!nodes_subset(nodes, node_states[N_MEMORY])) nodes 2816 mm/mempolicy.c nodes_clear(nodes); nodes 2835 mm/mempolicy.c if (nodes_empty(nodes)) nodes 2844 mm/mempolicy.c nodes = node_states[N_MEMORY]; nodes 2883 mm/mempolicy.c new = mpol_new(mode, mode_flags, &nodes); nodes 2892 mm/mempolicy.c new->v.nodes = nodes; nodes 2894 mm/mempolicy.c new->v.preferred_node = first_node(nodes); nodes 2902 mm/mempolicy.c new->w.user_nodemask = nodes; nodes 2931 mm/mempolicy.c nodemask_t nodes = NODE_MASK_NONE; nodes 2947 mm/mempolicy.c node_set(pol->v.preferred_node, nodes); nodes 2951 mm/mempolicy.c nodes = pol->v.nodes; nodes 2973 mm/mempolicy.c if (!nodes_empty(nodes)) nodes 2975 mm/mempolicy.c nodemask_pr_args(&nodes)); nodes 1597 mm/migrate.c const int __user *nodes, nodes 1615 mm/migrate.c if (get_user(node, nodes + i)) nodes 1788 mm/migrate.c const int __user *nodes, nodes 1834 mm/migrate.c if (nodes) nodes 1836 mm/migrate.c nodes, status, flags); nodes 1850 mm/migrate.c const int __user *, nodes, nodes 1853 mm/migrate.c return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags); nodes 1859 mm/migrate.c const int __user *, nodes, nodes 1874 mm/migrate.c return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags); nodes 46 mm/mmzone.c static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes) nodes 49 mm/mmzone.c return node_isset(zonelist_node_idx(zref), *nodes); nodes 58 mm/mmzone.c nodemask_t *nodes) nodes 64 mm/mmzone.c if (unlikely(nodes == NULL)) nodes 69 mm/mmzone.c (z->zone && !zref_in_nodemask(z, nodes))) nodes 4503 mm/slub.c nodemask_t nodes; nodes 4580 mm/slub.c node_set(page_to_nid(virt_to_page(track)), l->nodes); nodes 4610 mm/slub.c nodes_clear(l->nodes); nodes 4611 mm/slub.c node_set(page_to_nid(virt_to_page(track)), l->nodes); nodes 4698 mm/slub.c if (nr_online_nodes > 1 && !nodes_empty(l->nodes) && nodes 4702 mm/slub.c nodemask_pr_args(&l->nodes)); nodes 4812 mm/slub.c unsigned long *nodes; nodes 4814 mm/slub.c nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); nodes 4815 mm/slub.c if (!nodes) nodes 4840 mm/slub.c nodes[node] += x; nodes 4852 mm/slub.c nodes[node] += x; nodes 4882 mm/slub.c nodes[node] += x; nodes 4898 mm/slub.c nodes[node] += x; nodes 4904 mm/slub.c if (nodes[node]) nodes 4906 mm/slub.c node, nodes[node]); nodes 4908 mm/slub.c kfree(nodes); nodes 397 mm/workingset.c unsigned long nodes; nodes 400 mm/workingset.c nodes = list_lru_shrink_count(&shadow_nodes, sc); nodes 441 mm/workingset.c if (!nodes) nodes 444 mm/workingset.c if (nodes <= max_nodes) nodes 446 mm/workingset.c return nodes - max_nodes; nodes 550 net/tipc/name_table.c u32 upper, struct tipc_nlist *nodes) nodes 571 net/tipc/name_table.c tipc_nlist_add(nodes, p->node); nodes 115 net/tipc/name_table.h u32 upper, struct tipc_nlist *nodes); nodes 16 sound/hda/hdac_sysfs.c struct kobject **nodes; nodes 328 sound/hda/hdac_sysfs.c if (tree->nodes) { nodes 329 sound/hda/hdac_sysfs.c for (p = tree->nodes; *p; p++) nodes 331 sound/hda/hdac_sysfs.c kfree(tree->nodes); nodes 375 sound/hda/hdac_sysfs.c tree->nodes = kcalloc(codec->num_nodes + 1, sizeof(*tree->nodes), nodes 377 sound/hda/hdac_sysfs.c if (!tree->nodes) nodes 382 sound/hda/hdac_sysfs.c &tree->nodes[i]); nodes 437 sound/hda/hdac_sysfs.c tree->nodes = kcalloc(num_nodes + 1, sizeof(*tree->nodes), GFP_KERNEL); nodes 438 sound/hda/hdac_sysfs.c if (!tree->nodes) { nodes 446 sound/hda/hdac_sysfs.c free_widget_node(codec->widgets->nodes[i], nodes 454 sound/hda/hdac_sysfs.c &tree->nodes[i]); nodes 456 sound/hda/hdac_sysfs.c tree->nodes[i] = nodes 457 sound/hda/hdac_sysfs.c codec->widgets->nodes[nid - codec->start_nid]; nodes 461 sound/hda/hdac_sysfs.c kfree(codec->widgets->nodes); nodes 743 sound/pci/hda/hda_proc.c int i, nodes; nodes 759 sound/pci/hda/hda_proc.c nodes = snd_hda_get_sub_nodes(codec, fg, &nid); nodes 760 sound/pci/hda/hda_proc.c if (! nid || nodes < 0) { nodes 770 sound/pci/hda/hda_proc.c for (i = 0; i < nodes; i++, nid++) { nodes 1821 sound/pci/hda/patch_hdmi.c int i, nodes; nodes 1823 sound/pci/hda/patch_hdmi.c nodes = snd_hda_get_sub_nodes(codec, codec->core.afg, &nid); nodes 1824 sound/pci/hda/patch_hdmi.c if (!nid || nodes < 0) { nodes 1829 sound/pci/hda/patch_hdmi.c for (i = 0; i < nodes; i++, nid++) { nodes 878 tools/perf/bench/numa.c int nodes; nodes 896 tools/perf/bench/numa.c nodes = 0; nodes 899 tools/perf/bench/numa.c nodes += node_present[n]; nodes 901 tools/perf/bench/numa.c return nodes; nodes 945 tools/perf/bench/numa.c unsigned int nodes = count_process_nodes(p); nodes 947 tools/perf/bench/numa.c if (!nodes) { nodes 952 tools/perf/bench/numa.c nodes_min = min(nodes, nodes_min); nodes 953 tools/perf/bench/numa.c nodes_max = max(nodes, nodes_max); nodes 969 tools/perf/bench/numa.c int nodes[MAX_NR_NODES]; nodes 984 tools/perf/bench/numa.c nodes[node] = 0; nodes 1001 tools/perf/bench/numa.c nodes[node]++; nodes 1015 tools/perf/bench/numa.c nr = nodes[node]; nodes 1040 tools/perf/bench/numa.c nr = nodes[node]; nodes 87 tools/perf/builtin-c2c.c unsigned long **nodes; nodes 1097 tools/perf/builtin-c2c.c bitmap_and(set, c2c_he->cpuset, c2c.nodes[node], c2c.cpus_cnt); nodes 2032 tools/perf/builtin-c2c.c unsigned long **nodes; nodes 2046 tools/perf/builtin-c2c.c nodes = zalloc(sizeof(unsigned long *) * c2c.nodes_cnt); nodes 2047 tools/perf/builtin-c2c.c if (!nodes) nodes 2050 tools/perf/builtin-c2c.c c2c.nodes = nodes; nodes 2069 tools/perf/builtin-c2c.c nodes[node] = set; nodes 49 tools/perf/tests/mem2node.c struct memory_node nodes[3]; nodes 51 tools/perf/tests/mem2node.c .memory_nodes = (struct memory_node *) &nodes[0], nodes 52 tools/perf/tests/mem2node.c .nr_memory_nodes = ARRAY_SIZE(nodes), nodes 57 tools/perf/tests/mem2node.c for (i = 0; i < ARRAY_SIZE(nodes); i++) { nodes 58 tools/perf/tests/mem2node.c nodes[i].node = test_nodes[i].node; nodes 59 tools/perf/tests/mem2node.c nodes[i].size = 10; nodes 62 tools/perf/tests/mem2node.c (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); nodes 74 tools/perf/tests/mem2node.c for (i = 0; i < ARRAY_SIZE(nodes); i++) nodes 75 tools/perf/tests/mem2node.c zfree(&nodes[i].set); nodes 324 tools/perf/util/cputopo.c tp = zalloc(sizeof(*tp) + sizeof(tp->nodes[0])*nr); nodes 331 tools/perf/util/cputopo.c if (load_numa_node(&tp->nodes[i], node_map->map[i])) { nodes 350 tools/perf/util/cputopo.c zfree(&tp->nodes[i].cpus); nodes 25 tools/perf/util/cputopo.h struct numa_topology_node nodes[0]; nodes 687 tools/perf/util/header.c struct numa_topology_node *n = &tp->nodes[i]; nodes 1259 tools/perf/util/header.c static int build_mem_topology(struct memory_node *nodes, u64 size, u64 *cntp) nodes 1295 tools/perf/util/header.c ret = memory_node__read(&nodes[cnt++], idx); nodes 1302 tools/perf/util/header.c qsort(nodes, cnt, sizeof(nodes[0]), memory_node__sort); nodes 1327 tools/perf/util/header.c static struct memory_node nodes[MAX_MEMORY_NODES]; nodes 1336 tools/perf/util/header.c ret = build_mem_topology(&nodes[0], MAX_MEMORY_NODES, &nr); nodes 1353 tools/perf/util/header.c struct memory_node *n = &nodes[i]; nodes 1872 tools/perf/util/header.c struct memory_node *nodes; nodes 1875 tools/perf/util/header.c nodes = ff->ph->env.memory_nodes; nodes 1882 tools/perf/util/header.c memory_node__fprintf(&nodes[i], ff->ph->env.memory_bsize, fp); nodes 2331 tools/perf/util/header.c struct numa_node *nodes, *n; nodes 2339 tools/perf/util/header.c nodes = zalloc(sizeof(*nodes) * nr); nodes 2340 tools/perf/util/header.c if (!nodes) nodes 2344 tools/perf/util/header.c n = &nodes[i]; nodes 2367 tools/perf/util/header.c ff->ph->env.numa_nodes = nodes; nodes 2371 tools/perf/util/header.c free(nodes); nodes 2596 tools/perf/util/header.c struct memory_node *nodes; nodes 2612 tools/perf/util/header.c nodes = zalloc(sizeof(*nodes) * nr); nodes 2613 tools/perf/util/header.c if (!nodes) nodes 2631 tools/perf/util/header.c nodes[i] = n; nodes 2635 tools/perf/util/header.c ff->ph->env.memory_nodes = nodes; nodes 2641 tools/perf/util/header.c free(nodes); nodes 48 tools/perf/util/mem2node.c struct memory_node *n, *nodes = &env->memory_nodes[0]; nodes 57 tools/perf/util/mem2node.c n = &nodes[i]; nodes 68 tools/perf/util/mem2node.c n = &nodes[i]; nodes 61 tools/testing/radix-tree/test.h struct radix_tree_node *nodes; nodes 231 tools/testing/selftests/net/reuseport_bpf_numa.c int *rcv_fd, nodes; nodes 236 tools/testing/selftests/net/reuseport_bpf_numa.c nodes = numa_max_node() + 1; nodes 238 tools/testing/selftests/net/reuseport_bpf_numa.c rcv_fd = calloc(nodes, sizeof(int)); nodes 243 tools/testing/selftests/net/reuseport_bpf_numa.c test(rcv_fd, nodes, AF_INET, SOCK_DGRAM); nodes 246 tools/testing/selftests/net/reuseport_bpf_numa.c test(rcv_fd, nodes, AF_INET6, SOCK_DGRAM); nodes 249 tools/testing/selftests/net/reuseport_bpf_numa.c test(rcv_fd, nodes, AF_INET, SOCK_STREAM); nodes 252 tools/testing/selftests/net/reuseport_bpf_numa.c test(rcv_fd, nodes, AF_INET6, SOCK_STREAM);