node 13 arch/alpha/include/asm/topology.h int node; node 18 arch/alpha/include/asm/topology.h node = alpha_mv.cpuid_to_nid(cpu); node 21 arch/alpha/include/asm/topology.h BUG_ON(node < 0); node 24 arch/alpha/include/asm/topology.h return node; node 29 arch/alpha/include/asm/topology.h static const struct cpumask *cpumask_of_node(int node) node 33 arch/alpha/include/asm/topology.h if (node == NUMA_NO_NODE) node 36 arch/alpha/include/asm/topology.h cpumask_clear(&node_to_cpumask_map[node]); node 39 arch/alpha/include/asm/topology.h if (cpu_to_node(cpu) == node) node 40 arch/alpha/include/asm/topology.h cpumask_set_cpu(cpu, node_to_cpumask_map[node]); node 43 arch/alpha/include/asm/topology.h return &node_to_cpumask_map[node]; node 362 arch/alpha/kernel/core_marvel.c marvel_io7_present(gct6_node *node) node 366 arch/alpha/kernel/core_marvel.c if (node->type != GCT_TYPE_HOSE || node 367 arch/alpha/kernel/core_marvel.c node->subtype != GCT_SUBTYPE_IO_PORT_MODULE) node 370 arch/alpha/kernel/core_marvel.c pe = (node->id >> 8) & 0xff; node 14 arch/alpha/kernel/gct.c gct6_find_nodes(gct6_node *node, gct6_search_struct *search) node 20 arch/alpha/kernel/gct.c if (node->magic != GCT_NODE_MAGIC) { node 29 arch/alpha/kernel/gct.c if (node->type != wanted->type) node 31 arch/alpha/kernel/gct.c if (node->subtype != wanted->subtype) node 36 arch/alpha/kernel/gct.c wanted->callout(node); node 40 arch/alpha/kernel/gct.c if (node->next) node 41 arch/alpha/kernel/gct.c status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search); node 44 arch/alpha/kernel/gct.c if (node->child) node 45 arch/alpha/kernel/gct.c status |= gct6_find_nodes(GCT_NODE_PTR(node->child), search); node 308 arch/alpha/kernel/pci.c list_for_each_entry(child_bus, &b->children, node) node 317 arch/alpha/kernel/pci.c list_for_each_entry(b, &pci_root_buses, node) node 108 arch/arc/plat-hsdk/platform.c int node, ret; node 111 arch/arc/plat-hsdk/platform.c node = fdt_path_offset(fdt, path); node 112 arch/arc/plat-hsdk/platform.c if (node < 0) node 115 arch/arc/plat-hsdk/platform.c prop = fdt_getprop(fdt, node, "dma-coherent", &ret); node 124 arch/arc/plat-hsdk/platform.c ret = fdt_delprop(fdt, node, "dma-coherent"); node 128 arch/arc/plat-hsdk/platform.c ret = fdt_setprop(fdt, node, "dma-coherent", NULL, 0); node 46 arch/arm/common/dmabounce.c struct list_head node; node 158 arch/arm/common/dmabounce.c list_add(&buf->node, &device_info->safe_buffers); node 173 arch/arm/common/dmabounce.c list_for_each_entry(b, &device_info->safe_buffers, node) node 193 arch/arm/common/dmabounce.c list_del(&buf->node); node 42 arch/arm/include/asm/mach/pci.h struct list_head node; node 11 arch/arm/include/asm/traps.h struct list_head node; node 63 arch/arm/kernel/bios32.c list_for_each_entry(bus, &pci_root_buses, node) node 516 arch/arm/kernel/bios32.c list_add(&sys->node, head); node 537 arch/arm/kernel/bios32.c list_for_each_entry(sys, &head, node) { node 553 arch/arm/kernel/bios32.c list_for_each_entry(child, &bus->children, node) node 35 arch/arm/kernel/devtree.c static int __init set_smp_ops_by_method(struct device_node *node) node 40 arch/arm/kernel/devtree.c if (of_property_read_string(node, "enable-method", &method)) node 52 arch/arm/kernel/devtree.c static inline int set_smp_ops_by_method(struct device_node *node) node 406 arch/arm/kernel/traps.c list_add(&hook->node, &undef_hook); node 415 arch/arm/kernel/traps.c list_del(&hook->node); node 427 arch/arm/kernel/traps.c list_for_each_entry(hook, &undef_hook, node) node 97 arch/arm/mach-actions/platsmp.c struct device_node *node; node 99 arch/arm/mach-actions/platsmp.c node = of_find_compatible_node(NULL, NULL, "actions,s500-timer"); node 100 arch/arm/mach-actions/platsmp.c if (!node) { node 105 arch/arm/mach-actions/platsmp.c timer_base_addr = of_iomap(node, 0); node 111 arch/arm/mach-actions/platsmp.c node = of_find_compatible_node(NULL, NULL, "actions,s500-sps"); node 112 arch/arm/mach-actions/platsmp.c if (!node) { node 117 arch/arm/mach-actions/platsmp.c sps_base_addr = of_iomap(node, 0); node 124 arch/arm/mach-actions/platsmp.c node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu"); node 125 arch/arm/mach-actions/platsmp.c if (!node) { node 130 arch/arm/mach-actions/platsmp.c scu_base_addr = of_iomap(node, 0); node 576 arch/arm/mach-at91/pm.c struct device_node *node; node 579 arch/arm/mach-at91/pm.c for_each_compatible_node(node, NULL, "mmio-sram") { node 580 arch/arm/mach-at91/pm.c pdev = of_find_device_by_node(node); node 582 arch/arm/mach-at91/pm.c of_node_put(node); node 45 arch/arm/mach-bcm/bcm_kona_smc.c struct device_node *node; node 52 arch/arm/mach-bcm/bcm_kona_smc.c node = of_find_matching_node(NULL, bcm_kona_smc_ids); node 53 arch/arm/mach-bcm/bcm_kona_smc.c if (!node) node 56 arch/arm/mach-bcm/bcm_kona_smc.c prop_val = of_get_address(node, 0, &prop_size, NULL); node 41 arch/arm/mach-exynos/exynos.c struct device_node *node; node 43 arch/arm/mach-exynos/exynos.c for_each_compatible_node(node, NULL, "samsung,exynos4210-sysram") { node 44 arch/arm/mach-exynos/exynos.c if (!of_device_is_available(node)) node 46 arch/arm/mach-exynos/exynos.c sysram_base_addr = of_iomap(node, 0); node 47 arch/arm/mach-exynos/exynos.c sysram_base_phys = of_translate_address(node, node 48 arch/arm/mach-exynos/exynos.c of_get_address(node, 0, NULL, NULL)); node 52 arch/arm/mach-exynos/exynos.c for_each_compatible_node(node, NULL, "samsung,exynos4210-sysram-ns") { node 53 arch/arm/mach-exynos/exynos.c if (!of_device_is_available(node)) node 55 arch/arm/mach-exynos/exynos.c sysram_ns_base_addr = of_iomap(node, 0); node 60 arch/arm/mach-exynos/exynos.c static int __init exynos_fdt_map_chipid(unsigned long node, const char *uname, node 67 arch/arm/mach-exynos/exynos.c if (!of_flat_dt_is_compatible(node, "samsung,exynos4210-chipid")) node 70 arch/arm/mach-exynos/exynos.c reg = of_get_flat_dt_prop(node, "reg", &len); node 237 arch/arm/mach-exynos/mcpm-exynos.c struct device_node *node; node 241 arch/arm/mach-exynos/mcpm-exynos.c node = of_find_matching_node(NULL, exynos_dt_mcpm_match); node 242 arch/arm/mach-exynos/mcpm-exynos.c if (!node) node 244 arch/arm/mach-exynos/mcpm-exynos.c of_node_put(node); node 249 arch/arm/mach-exynos/mcpm-exynos.c node = of_find_compatible_node(NULL, NULL, node 251 arch/arm/mach-exynos/mcpm-exynos.c if (!node) node 254 arch/arm/mach-exynos/mcpm-exynos.c ns_sram_base_addr = of_iomap(node, 0); node 255 arch/arm/mach-exynos/mcpm-exynos.c of_node_put(node); node 189 arch/arm/mach-exynos/suspend.c static int __init exynos_pmu_irq_init(struct device_node *node, node 195 arch/arm/mach-exynos/suspend.c pr_err("%pOF: no parent, giving up\n", node); node 201 arch/arm/mach-exynos/suspend.c pr_err("%pOF: unable to obtain parent domain\n", node); node 205 arch/arm/mach-exynos/suspend.c pmu_base_addr = of_iomap(node, 0); node 208 arch/arm/mach-exynos/suspend.c pr_err("%pOF: failed to find exynos pmu register\n", node); node 213 arch/arm/mach-exynos/suspend.c node, &exynos_pmu_domain_ops, node 225 arch/arm/mach-exynos/suspend.c of_node_clear_flag(node, OF_POPULATED); node 145 arch/arm/mach-hisi/hotplug.c struct device_node *node; node 147 arch/arm/mach-hisi/hotplug.c node = of_find_compatible_node(NULL, NULL, "hisilicon,sysctrl"); node 148 arch/arm/mach-hisi/hotplug.c if (!node) { node 153 arch/arm/mach-hisi/hotplug.c ctrl_base = of_iomap(node, 0); node 154 arch/arm/mach-hisi/hotplug.c of_node_put(node); node 153 arch/arm/mach-hisi/platsmp.c struct device_node *node; node 159 arch/arm/mach-hisi/platsmp.c node = of_find_compatible_node(NULL, NULL, "hisilicon,hip01-sysctrl"); node 160 arch/arm/mach-hisi/platsmp.c if (WARN_ON(!node)) node 162 arch/arm/mach-hisi/platsmp.c ctrl_base = of_iomap(node, 0); node 228 arch/arm/mach-imx/gpc.c static int __init imx_gpc_init(struct device_node *node, node 235 arch/arm/mach-imx/gpc.c pr_err("%pOF: no parent, giving up\n", node); node 241 arch/arm/mach-imx/gpc.c pr_err("%pOF: unable to obtain parent domain\n", node); node 245 arch/arm/mach-imx/gpc.c gpc_base = of_iomap(node, 0); node 250 arch/arm/mach-imx/gpc.c node, &imx_gpc_domain_ops, node 265 arch/arm/mach-imx/gpc.c of_node_clear_flag(node, OF_POPULATED); node 104 arch/arm/mach-imx/mmdc.c struct hlist_node node; node 214 arch/arm/mach-imx/mmdc.c static int mmdc_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 216 arch/arm/mach-imx/mmdc.c struct mmdc_pmu *pmu_mmdc = hlist_entry_safe(node, struct mmdc_pmu, node); node 463 arch/arm/mach-imx/mmdc.c cpuhp_state_remove_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node); node 512 arch/arm/mach-imx/mmdc.c cpuhp_state_add_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node); node 523 arch/arm/mach-imx/mmdc.c cpuhp_state_remove_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node); node 272 arch/arm/mach-imx/pm-imx5.c struct device_node *node; node 281 arch/arm/mach-imx/pm-imx5.c node = of_find_compatible_node(NULL, NULL, "mmio-sram"); node 282 arch/arm/mach-imx/pm-imx5.c if (!node) { node 287 arch/arm/mach-imx/pm-imx5.c pdev = of_find_device_by_node(node); node 316 arch/arm/mach-imx/pm-imx5.c of_node_put(node); node 439 arch/arm/mach-imx/pm-imx6.c struct device_node *node; node 443 arch/arm/mach-imx/pm-imx6.c node = of_find_compatible_node(NULL, NULL, compat); node 444 arch/arm/mach-imx/pm-imx6.c if (!node) node 447 arch/arm/mach-imx/pm-imx6.c ret = of_address_to_resource(node, 0, &res); node 457 arch/arm/mach-imx/pm-imx6.c of_node_put(node); node 464 arch/arm/mach-imx/pm-imx6.c struct device_node *node; node 479 arch/arm/mach-imx/pm-imx6.c node = of_find_compatible_node(NULL, NULL, "mmio-sram"); node 480 arch/arm/mach-imx/pm-imx6.c if (!node) { node 485 arch/arm/mach-imx/pm-imx6.c pdev = of_find_device_by_node(node); node 584 arch/arm/mach-imx/pm-imx6.c of_node_put(node); node 64 arch/arm/mach-meson/platsmp.c static struct device_node *node; node 67 arch/arm/mach-meson/platsmp.c node = of_find_compatible_node(NULL, NULL, sram_compatible); node 68 arch/arm/mach-meson/platsmp.c if (!node) { node 73 arch/arm/mach-meson/platsmp.c sram_base = of_iomap(node, 0); node 87 arch/arm/mach-meson/platsmp.c node = of_find_compatible_node(NULL, NULL, scu_compatible); node 88 arch/arm/mach-meson/platsmp.c if (!node) { node 93 arch/arm/mach-meson/platsmp.c scu_base = of_iomap(node, 0); node 33 arch/arm/mach-mmp/sram.c struct list_head node; node 48 arch/arm/mach-mmp/sram.c list_for_each_entry(info, &sram_bank_list, node) node 54 arch/arm/mach-mmp/sram.c if (&info->node == &sram_bank_list) node 107 arch/arm/mach-mmp/sram.c list_add(&info->node, &sram_bank_list); node 134 arch/arm/mach-mmp/sram.c list_del(&info->node); node 69 arch/arm/mach-mvebu/board-v7.c static int __init mvebu_scan_mem(unsigned long node, const char *uname, node 72 arch/arm/mach-mvebu/board-v7.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 79 arch/arm/mach-mvebu/board-v7.c reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l); node 81 arch/arm/mach-mvebu/board-v7.c reg = of_get_flat_dt_prop(node, "reg", &l); node 117 arch/arm/mach-mvebu/platsmp.c struct device_node *node; node 136 arch/arm/mach-mvebu/platsmp.c node = of_find_compatible_node(NULL, NULL, "marvell,bootrom"); node 137 arch/arm/mach-mvebu/platsmp.c if (!node) node 140 arch/arm/mach-mvebu/platsmp.c err = of_address_to_resource(node, 0, &res); node 141 arch/arm/mach-mvebu/platsmp.c of_node_put(node); node 800 arch/arm/mach-omap1/clock.c if (clk->node.next || clk->node.prev) node 809 arch/arm/mach-omap1/clock.c list_add(&clk->node, &clocks); node 825 arch/arm/mach-omap1/clock.c list_del(&clk->node); node 834 arch/arm/mach-omap1/clock.c list_for_each_entry(clkp, &clocks, node) node 854 arch/arm/mach-omap1/clock.c list_for_each_entry(c, &clocks, node) { node 873 arch/arm/mach-omap1/clock.c list_for_each_entry(c, &clocks, node) node 889 arch/arm/mach-omap1/clock.c list_for_each_entry(c, &clocks, node) node 941 arch/arm/mach-omap1/clock.c list_for_each_entry(ck, &clocks, node) { node 977 arch/arm/mach-omap1/clock.c list_for_each_entry(c, &clocks, node) { node 1022 arch/arm/mach-omap1/clock.c list_for_each_entry(c, &clocks, node) node 139 arch/arm/mach-omap1/clock.h struct list_head node; node 55 arch/arm/mach-omap2/clockdomain.c list_for_each_entry(temp_clkdm, &clkdm_list, node) { node 92 arch/arm/mach-omap2/clockdomain.c list_add(&clkdm->node, &clkdm_list); node 485 arch/arm/mach-omap2/clockdomain.c list_for_each_entry(clkdm, &clkdm_list, node) { node 520 arch/arm/mach-omap2/clockdomain.c list_for_each_entry(temp_clkdm, &clkdm_list, node) { node 553 arch/arm/mach-omap2/clockdomain.c list_for_each_entry(clkdm, &clkdm_list, node) { node 140 arch/arm/mach-omap2/clockdomain.h struct list_head node; node 172 arch/arm/mach-omap2/display.c struct device_node *node; node 211 arch/arm/mach-omap2/display.c node = of_find_node_by_name(NULL, "omap4_padconf_global"); node 212 arch/arm/mach-omap2/display.c if (node) node 213 arch/arm/mach-omap2/display.c omap4_dsi_mux_syscon = syscon_node_to_regmap(node); node 228 arch/arm/mach-omap2/display.c struct device_node *node; node 232 arch/arm/mach-omap2/display.c node = of_find_compatible_node(NULL, NULL, node 234 arch/arm/mach-omap2/display.c if (node) node 235 arch/arm/mach-omap2/display.c return node; node 244 arch/arm/mach-omap2/display.c struct device_node *node; node 249 arch/arm/mach-omap2/display.c node = omapdss_find_dss_of_node(); node 250 arch/arm/mach-omap2/display.c if (!node) node 253 arch/arm/mach-omap2/display.c if (!of_device_is_available(node)) { node 254 arch/arm/mach-omap2/display.c of_node_put(node); node 258 arch/arm/mach-omap2/display.c pdev = of_find_device_by_node(node); node 265 arch/arm/mach-omap2/display.c r = of_platform_populate(node, NULL, NULL, &pdev->dev); node 545 arch/arm/mach-omap2/omap-wakeupgen.c static int __init wakeupgen_init(struct device_node *node, node 554 arch/arm/mach-omap2/omap-wakeupgen.c pr_err("%pOF: no parent, giving up\n", node); node 560 arch/arm/mach-omap2/omap-wakeupgen.c pr_err("%pOF: unable to obtain parent domain\n", node); node 570 arch/arm/mach-omap2/omap-wakeupgen.c wakeupgen_base = of_iomap(node, 0); node 588 arch/arm/mach-omap2/omap-wakeupgen.c node, &wakeupgen_domain_ops, node 137 arch/arm/mach-omap2/omap_device.c struct device_node *node = pdev->dev.of_node; node 143 arch/arm/mach-omap2/omap_device.c oh_cnt = of_property_count_strings(node, "ti,hwmods"); node 150 arch/arm/mach-omap2/omap_device.c ret = of_property_read_string_index(node, "ti,hwmods", 0, &oh_name); node 157 arch/arm/mach-omap2/omap_device.c !omap_hwmod_parse_module_range(NULL, node, &res)) node 167 arch/arm/mach-omap2/omap_device.c of_property_read_string_index(node, "ti,hwmods", i, &oh_name); node 201 arch/arm/mach-omap2/omap_hwmod.c struct device_node *node; node 748 arch/arm/mach-omap2/omap_hwmod.c provider->node = np; node 825 arch/arm/mach-omap2/omap_hwmod.c clkspec.np = provider->node; node 834 arch/arm/mach-omap2/omap_hwmod.c clkspec.args[0], provider->node); node 907 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(os, &oh->slave_ports, node) { node 1019 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(os, &oh->slave_ports, node) { node 1074 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(os, &oh->slave_ports, node) { node 1159 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(os, &oh->slave_ports, node) { node 1388 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(temp_oh, &omap_hwmod_list, node) { node 2453 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(os, &oh->slave_ports, node) { node 2653 arch/arm/mach-omap2/omap_hwmod.c list_add_tail(&oh->node, &omap_hwmod_list); node 2686 arch/arm/mach-omap2/omap_hwmod.c list_add(&oi->node, &oi->slave->slave_ports); node 3058 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(temp_oh, &omap_hwmod_list, node) { node 4014 arch/arm/mach-omap2/omap_hwmod.c list_for_each_entry(temp_oh, &omap_hwmod_list, node) { node 245 arch/arm/mach-omap2/omap_hwmod.h struct list_head node; node 597 arch/arm/mach-omap2/omap_hwmod.h struct list_head node; node 2822 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c struct device_node *node; node 2828 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c node = of_get_child_by_name(bus, dev_name); node 2829 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c available = of_device_is_available(node); node 2830 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c of_node_put(node); node 61 arch/arm/mach-omap2/pm34xx.c struct list_head node; node 314 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) node 317 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 330 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 367 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 385 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 396 arch/arm/mach-omap2/pm34xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 417 arch/arm/mach-omap2/pm34xx.c list_add(&pwrst->node, &pwrst_list); node 561 arch/arm/mach-omap2/pm34xx.c list_for_each_entry_safe(pwrst, tmp, &pwrst_list, node) { node 562 arch/arm/mach-omap2/pm34xx.c list_del(&pwrst->node); node 34 arch/arm/mach-omap2/pm44xx.c struct list_head node; node 59 arch/arm/mach-omap2/pm44xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 65 arch/arm/mach-omap2/pm44xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 82 arch/arm/mach-omap2/pm44xx.c list_for_each_entry(pwrst, &pwrst_list, node) { node 154 arch/arm/mach-omap2/pm44xx.c list_add(&pwrst->node, &pwrst_list); node 71 arch/arm/mach-omap2/powerdomain.c list_for_each_entry(temp_pwrdm, &pwrdm_list, node) { node 122 arch/arm/mach-omap2/powerdomain.c list_add(&pwrdm->node, &pwrdm_list); node 372 arch/arm/mach-omap2/powerdomain.c list_for_each_entry(temp_p, &pwrdm_list, node) node 446 arch/arm/mach-omap2/powerdomain.c list_for_each_entry(temp_pwrdm, &pwrdm_list, node) { node 124 arch/arm/mach-omap2/powerdomain.h struct list_head node; node 246 arch/arm/mach-omap2/voltage.c list_for_each_entry(voltdm, &voltdm_list, node) { node 281 arch/arm/mach-omap2/voltage.c list_for_each_entry(temp_voltdm, &voltdm_list, node) { node 296 arch/arm/mach-omap2/voltage.c list_add(&voltdm->node, &voltdm_list); node 65 arch/arm/mach-omap2/voltage.h struct list_head node; node 143 arch/arm/mach-pxa/irq.c pxa_init_irq_common(struct device_node *node, int irq_nr, node 149 arch/arm/mach-pxa/irq.c pxa_irq_domain = irq_domain_add_legacy(node, irq_nr, node 235 arch/arm/mach-pxa/irq.c struct device_node *node; node 239 arch/arm/mach-pxa/irq.c node = of_find_matching_node(NULL, intc_ids); node 240 arch/arm/mach-pxa/irq.c if (!node) { node 245 arch/arm/mach-pxa/irq.c ret = of_property_read_u32(node, "marvell,intc-nr-irqs", node 252 arch/arm/mach-pxa/irq.c ret = of_address_to_resource(node, 0, &res); node 259 arch/arm/mach-pxa/irq.c if (of_find_property(node, "marvell,intc-priority", NULL)) node 268 arch/arm/mach-pxa/irq.c pxa_init_irq_common(node, pxa_internal_irq_nr, fn); node 156 arch/arm/mach-pxa/pxa25x.c pxa25x_dt_init_irq(struct device_node *node, struct device_node *parent) node 238 arch/arm/mach-pxa/pxa27x.c pxa27x_dt_init_irq(struct device_node *node, struct device_node *parent) node 361 arch/arm/mach-pxa/pxa3xx.c pxa3xx_dt_init_irq(struct device_node *node, struct device_node *parent) node 55 arch/arm/mach-qcom/platsmp.c struct device_node *node; node 58 arch/arm/mach-qcom/platsmp.c node = of_find_compatible_node(NULL, NULL, "qcom,gcc-msm8660"); node 59 arch/arm/mach-qcom/platsmp.c if (!node) { node 64 arch/arm/mach-qcom/platsmp.c base = of_iomap(node, 0); node 65 arch/arm/mach-qcom/platsmp.c of_node_put(node); node 166 arch/arm/mach-rockchip/platsmp.c static int __init rockchip_smp_prepare_sram(struct device_node *node) node 174 arch/arm/mach-rockchip/platsmp.c ret = of_address_to_resource(node, 0, &res); node 177 arch/arm/mach-rockchip/platsmp.c __func__, node); node 210 arch/arm/mach-rockchip/platsmp.c struct device_node *node; node 219 arch/arm/mach-rockchip/platsmp.c node = of_find_node_by_path("/cpus"); node 221 arch/arm/mach-rockchip/platsmp.c pmu = syscon_regmap_lookup_by_phandle(node, "rockchip,pmu"); node 222 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 232 arch/arm/mach-rockchip/platsmp.c node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-pmu"); node 233 arch/arm/mach-rockchip/platsmp.c if (!node) { node 238 arch/arm/mach-rockchip/platsmp.c pmu_base = of_iomap(node, 0); node 239 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 260 arch/arm/mach-rockchip/platsmp.c struct device_node *node; node 263 arch/arm/mach-rockchip/platsmp.c node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-smp-sram"); node 264 arch/arm/mach-rockchip/platsmp.c if (!node) { node 269 arch/arm/mach-rockchip/platsmp.c sram_base_addr = of_iomap(node, 0); node 272 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 277 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 282 arch/arm/mach-rockchip/platsmp.c if (rockchip_smp_prepare_sram(node)) { node 283 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 290 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 291 arch/arm/mach-rockchip/platsmp.c node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu"); node 292 arch/arm/mach-rockchip/platsmp.c if (!node) { node 297 arch/arm/mach-rockchip/platsmp.c scu_base_addr = of_iomap(node, 0); node 300 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 319 arch/arm/mach-rockchip/platsmp.c of_node_put(node); node 21 arch/arm/mach-s5pv210/s5pv210.c static int __init s5pv210_fdt_map_sys(unsigned long node, const char *uname, node 28 arch/arm/mach-s5pv210/s5pv210.c if (!of_flat_dt_is_compatible(node, "samsung,s5pv210-clock")) node 31 arch/arm/mach-s5pv210/s5pv210.c reg = of_get_flat_dt_prop(node, "reg", &len); node 131 arch/arm/mach-shmobile/setup-rcar-gen2.c static int __init rcar_gen2_scan_mem(unsigned long node, const char *uname, node 134 arch/arm/mach-shmobile/setup-rcar-gen2.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 144 arch/arm/mach-shmobile/setup-rcar-gen2.c reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l); node 146 arch/arm/mach-shmobile/setup-rcar-gen2.c reg = of_get_flat_dt_prop(node, "reg", &l); node 90 arch/arm/mach-sunxi/mc_smp.c struct device_node *node; node 94 arch/arm/mach-sunxi/mc_smp.c node = of_cpu_device_node_get(cpu); node 97 arch/arm/mach-sunxi/mc_smp.c if (!node) node 98 arch/arm/mach-sunxi/mc_smp.c node = of_get_cpu_node(cpu, NULL); node 100 arch/arm/mach-sunxi/mc_smp.c if (!node) { node 111 arch/arm/mach-sunxi/mc_smp.c is_compatible = of_device_is_compatible(node, "arm,cortex-a15"); node 112 arch/arm/mach-sunxi/mc_smp.c of_node_put(node); node 783 arch/arm/mach-sunxi/mc_smp.c struct device_node *node; node 792 arch/arm/mach-sunxi/mc_smp.c node = of_cpu_device_node_get(0); node 793 arch/arm/mach-sunxi/mc_smp.c if (!node) node 805 arch/arm/mach-sunxi/mc_smp.c ret = of_property_match_string(node, "enable-method", node 813 arch/arm/mach-sunxi/mc_smp.c of_node_put(node); node 44 arch/arm/mach-sunxi/platsmp.c struct device_node *node; node 46 arch/arm/mach-sunxi/platsmp.c node = of_find_compatible_node(NULL, NULL, "allwinner,sun6i-a31-prcm"); node 47 arch/arm/mach-sunxi/platsmp.c if (!node) { node 52 arch/arm/mach-sunxi/platsmp.c prcm_membase = of_iomap(node, 0); node 53 arch/arm/mach-sunxi/platsmp.c of_node_put(node); node 59 arch/arm/mach-sunxi/platsmp.c node = of_find_compatible_node(NULL, NULL, node 61 arch/arm/mach-sunxi/platsmp.c if (!node) { node 66 arch/arm/mach-sunxi/platsmp.c cpucfg_membase = of_iomap(node, 0); node 67 arch/arm/mach-sunxi/platsmp.c of_node_put(node); node 129 arch/arm/mach-sunxi/platsmp.c struct device_node *node; node 131 arch/arm/mach-sunxi/platsmp.c node = of_find_compatible_node(NULL, NULL, "allwinner,sun8i-a23-prcm"); node 132 arch/arm/mach-sunxi/platsmp.c if (!node) { node 137 arch/arm/mach-sunxi/platsmp.c prcm_membase = of_iomap(node, 0); node 138 arch/arm/mach-sunxi/platsmp.c of_node_put(node); node 144 arch/arm/mach-sunxi/platsmp.c node = of_find_compatible_node(NULL, NULL, node 146 arch/arm/mach-sunxi/platsmp.c if (!node) { node 151 arch/arm/mach-sunxi/platsmp.c cpucfg_membase = of_iomap(node, 0); node 152 arch/arm/mach-sunxi/platsmp.c of_node_put(node); node 135 arch/arm/mach-vexpress/dcscb.c struct device_node *node; node 142 arch/arm/mach-vexpress/dcscb.c node = of_find_compatible_node(NULL, NULL, "arm,rtsm,dcscb"); node 143 arch/arm/mach-vexpress/dcscb.c if (!node) node 145 arch/arm/mach-vexpress/dcscb.c dcscb_base = of_iomap(node, 0); node 368 arch/arm/mm/cache-feroceon-l2.c struct device_node *node; node 376 arch/arm/mm/cache-feroceon-l2.c node = of_find_matching_node(NULL, feroceon_ids); node 377 arch/arm/mm/cache-feroceon-l2.c if (node && of_device_is_compatible(node, "marvell,kirkwood-cache")) { node 378 arch/arm/mm/cache-feroceon-l2.c base = of_iomap(node, 0); node 288 arch/arm/mm/cache-tauros2.c struct device_node *node; node 292 arch/arm/mm/cache-tauros2.c node = of_find_matching_node(NULL, tauros2_ids); node 293 arch/arm/mm/cache-tauros2.c if (!node) { node 296 arch/arm/mm/cache-tauros2.c ret = of_property_read_u32(node, "marvell,tauros2-cache-features", &f); node 44 arch/arm/plat-pxa/ssp.c list_for_each_entry(ssp, &ssp_list, node) { node 54 arch/arm/plat-pxa/ssp.c if (&ssp->node == &ssp_list) node 68 arch/arm/plat-pxa/ssp.c list_for_each_entry(ssp, &ssp_list, node) { node 78 arch/arm/plat-pxa/ssp.c if (&ssp->node == &ssp_list) node 173 arch/arm/plat-pxa/ssp.c list_add(&ssp->node, &ssp_list); node 190 arch/arm/plat-pxa/ssp.c list_del(&ssp->node); node 207 arch/arm/xen/enlighten.c static int __init fdt_find_hyper_node(unsigned long node, const char *uname, node 216 arch/arm/xen/enlighten.c if (of_flat_dt_is_compatible(node, hyper_node.compat)) node 219 arch/arm/xen/enlighten.c s = of_get_flat_dt_prop(node, "compatible", &len); node 232 arch/arm/xen/enlighten.c if ((of_get_flat_dt_subnode_by_name(node, "uefi") > 0) && node 77 arch/arm64/include/asm/debug-monitors.h struct list_head node; node 88 arch/arm64/include/asm/debug-monitors.h struct list_head node; node 23 arch/arm64/include/asm/numa.h const struct cpumask *cpumask_of_node(int node); node 26 arch/arm64/include/asm/numa.h static inline const struct cpumask *cpumask_of_node(int node) node 28 arch/arm64/include/asm/numa.h return node_to_cpumask_map[node]; node 17 arch/arm64/include/asm/traps.h struct list_head node; node 18 arch/arm64/include/asm/vmap_stack.h static inline unsigned long *arch_alloc_vmap_stack(size_t stack_size, int node) node 24 arch/arm64/include/asm/vmap_stack.h THREADINFO_GFP, PAGE_KERNEL, 0, node, node 64 arch/arm64/kernel/acpi.c static int __init dt_scan_depth1_nodes(unsigned long node, node 79 arch/arm64/kernel/acpi.c of_flat_dt_is_compatible(node, "xen,xen")) node 52 arch/arm64/kernel/acpi_numa.c int cpu, pxm, node; node 65 arch/arm64/kernel/acpi_numa.c node = pxm_to_node(pxm); node 77 arch/arm64/kernel/acpi_numa.c acpi_early_node_map[cpu] = node; node 79 arch/arm64/kernel/acpi_numa.c cpu_logical_map(cpu), node); node 94 arch/arm64/kernel/acpi_numa.c int pxm, node; node 110 arch/arm64/kernel/acpi_numa.c node = acpi_map_pxm_to_node(pxm); node 112 arch/arm64/kernel/acpi_numa.c if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) { node 118 arch/arm64/kernel/acpi_numa.c node_set(node, numa_nodes_parsed); node 52 arch/arm64/kernel/armv8_deprecated.c struct list_head node; node 126 arch/arm64/kernel/armv8_deprecated.c list_for_each_entry(insn, &insn_emulation, node) { node 197 arch/arm64/kernel/armv8_deprecated.c list_add(&insn->node, &insn_emulation); node 243 arch/arm64/kernel/armv8_deprecated.c list_for_each_entry(insn, &insn_emulation, node) { node 160 arch/arm64/kernel/debug-monitors.c static void register_debug_hook(struct list_head *node, struct list_head *list) node 163 arch/arm64/kernel/debug-monitors.c list_add_rcu(node, list); node 168 arch/arm64/kernel/debug-monitors.c static void unregister_debug_hook(struct list_head *node) node 171 arch/arm64/kernel/debug-monitors.c list_del_rcu(node); node 178 arch/arm64/kernel/debug-monitors.c register_debug_hook(&hook->node, &user_step_hook); node 183 arch/arm64/kernel/debug-monitors.c unregister_debug_hook(&hook->node); node 188 arch/arm64/kernel/debug-monitors.c register_debug_hook(&hook->node, &kernel_step_hook); node 193 arch/arm64/kernel/debug-monitors.c unregister_debug_hook(&hook->node); node 214 arch/arm64/kernel/debug-monitors.c list_for_each_entry_rcu(hook, list, node) { node 282 arch/arm64/kernel/debug-monitors.c register_debug_hook(&hook->node, &user_break_hook); node 287 arch/arm64/kernel/debug-monitors.c unregister_debug_hook(&hook->node); node 292 arch/arm64/kernel/debug-monitors.c register_debug_hook(&hook->node, &kernel_break_hook); node 297 arch/arm64/kernel/debug-monitors.c unregister_debug_hook(&hook->node); node 312 arch/arm64/kernel/debug-monitors.c list_for_each_entry_rcu(hook, list, node) { node 27 arch/arm64/kernel/kaslr.c int node, len; node 31 arch/arm64/kernel/kaslr.c node = fdt_path_offset(fdt, "/chosen"); node 32 arch/arm64/kernel/kaslr.c if (node < 0) node 35 arch/arm64/kernel/kaslr.c prop = fdt_getprop_w(fdt, node, "kaslr-seed", &len); node 49 arch/arm64/kernel/kaslr.c int node; node 52 arch/arm64/kernel/kaslr.c node = fdt_path_offset(fdt, "/chosen"); node 53 arch/arm64/kernel/kaslr.c if (node < 0) node 56 arch/arm64/kernel/kaslr.c prop = fdt_getprop(fdt, node, "bootargs", NULL); node 204 arch/arm64/kernel/pci.c list_for_each_entry(child, &bus->children, node) node 291 arch/arm64/kernel/traps.c list_add(&hook->node, &undef_hook); node 300 arch/arm64/kernel/traps.c list_del(&hook->node); node 340 arch/arm64/kernel/traps.c list_for_each_entry(hook, &undef_hook, node) node 123 arch/arm64/mm/init.c static int __init early_init_dt_scan_elfcorehdr(unsigned long node, node 132 arch/arm64/mm/init.c reg = of_get_flat_dt_prop(node, "linux,elfcorehdr", &len); node 276 arch/arm64/mm/init.c static int __init early_init_dt_scan_usablemem(unsigned long node, node 286 arch/arm64/mm/init.c reg = of_get_flat_dt_prop(node, "linux,usable-memory-range", &len); node 34 arch/arm64/mm/kasan_init.c static phys_addr_t __init kasan_alloc_zeroed_page(int node) node 38 arch/arm64/mm/kasan_init.c MEMBLOCK_ALLOC_KASAN, node); node 41 arch/arm64/mm/kasan_init.c __func__, PAGE_SIZE, PAGE_SIZE, node, node 47 arch/arm64/mm/kasan_init.c static phys_addr_t __init kasan_alloc_raw_page(int node) node 51 arch/arm64/mm/kasan_init.c MEMBLOCK_ALLOC_KASAN, node); node 54 arch/arm64/mm/kasan_init.c __func__, PAGE_SIZE, PAGE_SIZE, node, node 60 arch/arm64/mm/kasan_init.c static pte_t *__init kasan_pte_offset(pmd_t *pmdp, unsigned long addr, int node, node 66 arch/arm64/mm/kasan_init.c : kasan_alloc_zeroed_page(node); node 74 arch/arm64/mm/kasan_init.c static pmd_t *__init kasan_pmd_offset(pud_t *pudp, unsigned long addr, int node, node 80 arch/arm64/mm/kasan_init.c : kasan_alloc_zeroed_page(node); node 87 arch/arm64/mm/kasan_init.c static pud_t *__init kasan_pud_offset(pgd_t *pgdp, unsigned long addr, int node, node 93 arch/arm64/mm/kasan_init.c : kasan_alloc_zeroed_page(node); node 101 arch/arm64/mm/kasan_init.c unsigned long end, int node, bool early) node 104 arch/arm64/mm/kasan_init.c pte_t *ptep = kasan_pte_offset(pmdp, addr, node, early); node 109 arch/arm64/mm/kasan_init.c : kasan_alloc_raw_page(node); node 118 arch/arm64/mm/kasan_init.c unsigned long end, int node, bool early) node 121 arch/arm64/mm/kasan_init.c pmd_t *pmdp = kasan_pmd_offset(pudp, addr, node, early); node 125 arch/arm64/mm/kasan_init.c kasan_pte_populate(pmdp, addr, next, node, early); node 130 arch/arm64/mm/kasan_init.c unsigned long end, int node, bool early) node 133 arch/arm64/mm/kasan_init.c pud_t *pudp = kasan_pud_offset(pgdp, addr, node, early); node 137 arch/arm64/mm/kasan_init.c kasan_pmd_populate(pudp, addr, next, node, early); node 142 arch/arm64/mm/kasan_init.c int node, bool early) node 150 arch/arm64/mm/kasan_init.c kasan_pud_populate(pgdp, addr, next, node, early); node 168 arch/arm64/mm/kasan_init.c int node) node 170 arch/arm64/mm/kasan_init.c kasan_pgd_populate(start & PAGE_MASK, PAGE_ALIGN(end), node, false); node 730 arch/arm64/mm/mmu.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 733 arch/arm64/mm/mmu.c return vmemmap_populate_basepages(start, end, node); node 736 arch/arm64/mm/mmu.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 748 arch/arm64/mm/mmu.c pgdp = vmemmap_pgd_populate(addr, node); node 752 arch/arm64/mm/mmu.c pudp = vmemmap_pud_populate(pgdp, addr, node); node 760 arch/arm64/mm/mmu.c p = vmemmap_alloc_block_buf(PMD_SIZE, node); node 766 arch/arm64/mm/mmu.c vmemmap_verify((pte_t *)pmdp, node, addr, next); node 47 arch/arm64/mm/numa.c const struct cpumask *cpumask_of_node(int node) node 49 arch/arm64/mm/numa.c if (WARN_ON(node >= nr_node_ids)) node 52 arch/arm64/mm/numa.c if (WARN_ON(node_to_cpumask_map[node] == NULL)) node 55 arch/arm64/mm/numa.c return node_to_cpumask_map[node]; node 99 arch/arm64/mm/numa.c int node; node 106 arch/arm64/mm/numa.c for (node = 0; node < nr_node_ids; node++) { node 107 arch/arm64/mm/numa.c alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]); node 108 arch/arm64/mm/numa.c cpumask_clear(node_to_cpumask_map[node]); node 80 arch/c6x/include/asm/clock.h struct list_head node; node 95 arch/c6x/kernel/setup.c struct device_node *node; node 189 arch/c6x/kernel/setup.c for_each_of_cpu_node(node) node 192 arch/c6x/kernel/setup.c node = of_find_node_by_name(NULL, "soc"); node 193 arch/c6x/kernel/setup.c if (node) { node 194 arch/c6x/kernel/setup.c if (of_property_read_string(node, "model", &c6x_soc_name)) node 196 arch/c6x/kernel/setup.c of_node_put(node); node 426 arch/c6x/platforms/cache.c struct device_node *node; node 428 arch/c6x/platforms/cache.c node = of_find_compatible_node(NULL, NULL, "ti,c64x+cache"); node 429 arch/c6x/platforms/cache.c if (!node) node 432 arch/c6x/platforms/cache.c cache_base = of_iomap(node, 0); node 434 arch/c6x/platforms/cache.c of_node_put(node); node 99 arch/c6x/platforms/dscr.c void (*init)(struct device_node *node); node 281 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstat(struct device_node *node, node 287 arch/c6x/platforms/dscr.c err = of_property_read_u32_array(node, "ti,dscr-devstat", &val, 1); node 293 arch/c6x/platforms/dscr.c static void __init dscr_parse_silicon_rev(struct device_node *node, node 299 arch/c6x/platforms/dscr.c err = of_property_read_u32_array(node, "ti,dscr-silicon-rev", vals, 3); node 324 arch/c6x/platforms/dscr.c static void __init dscr_parse_mac_fuse(struct device_node *node, node 330 arch/c6x/platforms/dscr.c err = of_property_read_u32_array(node, "ti,dscr-mac-fuse-regs", node 343 arch/c6x/platforms/dscr.c static void __init dscr_parse_rmii_resets(struct device_node *node, node 350 arch/c6x/platforms/dscr.c p = of_get_property(node, "ti,dscr-rmii-resets", &size); node 365 arch/c6x/platforms/dscr.c static void __init dscr_parse_privperm(struct device_node *node, node 371 arch/c6x/platforms/dscr.c err = of_property_read_u32_array(node, "ti,dscr-privperm", vals, 2); node 393 arch/c6x/platforms/dscr.c static void __init dscr_parse_locked_regs(struct device_node *node, node 400 arch/c6x/platforms/dscr.c p = of_get_property(node, "ti,dscr-locked-regs", &size); node 428 arch/c6x/platforms/dscr.c static void __init dscr_parse_kick_regs(struct device_node *node, node 434 arch/c6x/platforms/dscr.c err = of_property_read_u32_array(node, "ti,dscr-kick-regs", vals, 4); node 466 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstate_ctl_regs(struct device_node *node, node 473 arch/c6x/platforms/dscr.c p = of_get_property(node, "ti,dscr-devstate-ctl-regs", &size); node 522 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstate_stat_regs(struct device_node *node, node 529 arch/c6x/platforms/dscr.c p = of_get_property(node, "ti,dscr-devstate-stat-regs", &size); node 569 arch/c6x/platforms/dscr.c struct device_node *node; node 574 arch/c6x/platforms/dscr.c node = of_find_matching_node(NULL, dscr_ids); node 575 arch/c6x/platforms/dscr.c if (!node) node 578 arch/c6x/platforms/dscr.c base = of_iomap(node, 0); node 580 arch/c6x/platforms/dscr.c of_node_put(node); node 586 arch/c6x/platforms/dscr.c dscr_parse_devstat(node, base); node 587 arch/c6x/platforms/dscr.c dscr_parse_silicon_rev(node, base); node 588 arch/c6x/platforms/dscr.c dscr_parse_mac_fuse(node, base); node 589 arch/c6x/platforms/dscr.c dscr_parse_rmii_resets(node, base); node 590 arch/c6x/platforms/dscr.c dscr_parse_locked_regs(node, base); node 591 arch/c6x/platforms/dscr.c dscr_parse_kick_regs(node, base); node 592 arch/c6x/platforms/dscr.c dscr_parse_devstate_ctl_regs(node, base); node 593 arch/c6x/platforms/dscr.c dscr_parse_devstate_stat_regs(node, base); node 594 arch/c6x/platforms/dscr.c dscr_parse_privperm(node, base); node 44 arch/c6x/platforms/emif.c struct device_node *node; node 49 arch/c6x/platforms/emif.c node = of_find_matching_node(NULL, emifa_match); node 50 arch/c6x/platforms/emif.c if (!node) node 53 arch/c6x/platforms/emif.c regs = of_iomap(node, 0); node 58 arch/c6x/platforms/emif.c err = of_property_read_u32_array(node, "ti,dscr-dev-enable", &val, 1); node 63 arch/c6x/platforms/emif.c p = of_get_property(node, "ti,emifa-ce-config", &len); node 72 arch/c6x/platforms/emif.c err = of_property_read_u32_array(node, "ti,emifa-burst-priority", &val, 1); node 76 arch/c6x/platforms/emif.c err = of_property_read_u32_array(node, "ti,emifa-async-wait-control", &val, 1); node 81 arch/c6x/platforms/emif.c of_node_put(node); node 166 arch/c6x/platforms/pll.c list_add_tail(&clk->node, &clocks); node 193 arch/c6x/platforms/pll.c list_del(&clk->node); node 324 arch/c6x/platforms/pll.c INIT_LIST_HEAD(&clk->node); node 412 arch/c6x/platforms/pll.c list_for_each_entry(clk, &clocks, node) node 29 arch/c6x/platforms/plldata.c .node = LIST_HEAD_INIT(clkin1.node), node 168 arch/c6x/platforms/plldata.c static void __init c6455_setup_clocks(struct device_node *node) node 206 arch/c6x/platforms/plldata.c static void __init c6457_setup_clocks(struct device_node *node) node 256 arch/c6x/platforms/plldata.c static void __init c6472_setup_clocks(struct device_node *node) node 305 arch/c6x/platforms/plldata.c static void __init c6474_setup_clocks(struct device_node *node) node 354 arch/c6x/platforms/plldata.c static void __init c6678_setup_clocks(struct device_node *node) node 422 arch/c6x/platforms/plldata.c struct device_node *node; node 427 arch/c6x/platforms/plldata.c node = of_find_matching_node(NULL, c6x_clkc_match); node 428 arch/c6x/platforms/plldata.c if (!node) node 431 arch/c6x/platforms/plldata.c pll->base = of_iomap(node, 0); node 435 arch/c6x/platforms/plldata.c err = of_property_read_u32(node, "clock-frequency", &val); node 438 arch/c6x/platforms/plldata.c node, (int)val / 1000000); node 443 arch/c6x/platforms/plldata.c err = of_property_read_u32(node, "ti,c64x+pll-bypass-delay", &val); node 448 arch/c6x/platforms/plldata.c err = of_property_read_u32(node, "ti,c64x+pll-reset-delay", &val); node 453 arch/c6x/platforms/plldata.c err = of_property_read_u32(node, "ti,c64x+pll-lock-delay", &val); node 459 arch/c6x/platforms/plldata.c id = of_match_node(c6x_clkc_match, node); node 462 arch/c6x/platforms/plldata.c __setup_clocks(node); node 466 arch/c6x/platforms/plldata.c of_node_put(node); node 1296 arch/csky/kernel/perf_event.c struct device_node *node = pdev->dev.of_node; node 1305 arch/csky/kernel/perf_event.c if (of_property_read_u32(node, "count-width", node 138 arch/csky/kernel/smp.c struct device_node *node = NULL; node 141 arch/csky/kernel/smp.c for_each_of_cpu_node(node) { node 142 arch/csky/kernel/smp.c if (!of_device_is_available(node)) node 145 arch/csky/kernel/smp.c if (of_property_read_u32(node, "reg", &cpu)) node 212 arch/ia64/hp/common/sba_iommu.c unsigned int node; /* node where this IOC lives */ node 1116 arch/ia64/hp/common/sba_iommu.c int node = -1; node 1122 arch/ia64/hp/common/sba_iommu.c node = ioc->node; node 1125 arch/ia64/hp/common/sba_iommu.c page = alloc_pages_node(node, flags, get_order(size)); node 1878 arch/ia64/hp/common/sba_iommu.c if (ioc->node != NUMA_NO_NODE) node 1879 arch/ia64/hp/common/sba_iommu.c seq_printf(s, "NUMA node : %d\n", ioc->node); node 1970 arch/ia64/hp/common/sba_iommu.c unsigned int node; node 1972 arch/ia64/hp/common/sba_iommu.c node = acpi_get_node(handle); node 1973 arch/ia64/hp/common/sba_iommu.c if (node != NUMA_NO_NODE && !node_online(node)) node 1974 arch/ia64/hp/common/sba_iommu.c node = NUMA_NO_NODE; node 1976 arch/ia64/hp/common/sba_iommu.c ioc->node = node; node 69 arch/ia64/include/asm/meminit.h static inline int vmemmap_find_next_valid_pfn(int node, int i) node 28 arch/ia64/include/asm/nodedata.h short node; node 50 arch/ia64/include/asm/pci.h int node; /* nearest node with memory or NUMA_NO_NODE for global allocation */ node 455 arch/ia64/include/asm/sal.h node : 1, node 474 arch/ia64/include/asm/sal.h u16 node; node 58 arch/ia64/include/asm/thread_info.h #define alloc_thread_stack_node(tsk, node) \ node 63 arch/ia64/include/asm/thread_info.h #define alloc_thread_stack_node(tsk, node) ((unsigned long *) 0) node 83 arch/ia64/include/asm/thread_info.h #define alloc_task_struct_node(node) \ node 85 arch/ia64/include/asm/thread_info.h struct page *page = alloc_pages_node(node, GFP_KERNEL | __GFP_COMP, \ node 28 arch/ia64/include/asm/topology.h #define cpumask_of_node(node) ((node) == -1 ? \ node 30 arch/ia64/include/asm/topology.h &node_to_cpu_mask[node]) node 35 arch/ia64/include/asm/topology.h #define pcibus_to_node(bus) PCI_CONTROLLER(bus)->node node 834 arch/ia64/kernel/acpi.c int node; node 862 arch/ia64/kernel/acpi.c node = acpi_get_node(handle); node 863 arch/ia64/kernel/acpi.c if (node == NUMA_NO_NODE || !node_online(node) || node 864 arch/ia64/kernel/acpi.c cpumask_empty(cpumask_of_node(node))) node 868 arch/ia64/kernel/acpi.c map_iosapic_to_node(gsi_base, node); node 122 arch/ia64/kernel/iosapic.c unsigned short node; /* numa node association via pxm */ node 655 arch/ia64/kernel/iosapic.c iosapic_lists[iosapic_index].node == MAX_NUMNODES) node 658 arch/ia64/kernel/iosapic.c cpu_mask = cpumask_of_node(iosapic_lists[iosapic_index].node); node 1066 arch/ia64/kernel/iosapic.c iosapic_lists[index].node = MAX_NUMNODES; node 1123 arch/ia64/kernel/iosapic.c void map_iosapic_to_node(unsigned int gsi_base, int node) node 1133 arch/ia64/kernel/iosapic.c iosapic_lists[index].node = node; node 175 arch/ia64/kernel/msi_ia64.c int dmar_alloc_hwirq(int id, int node, void *arg) node 59 arch/ia64/kernel/numa.c int cpu, i, node; node 61 arch/ia64/kernel/numa.c for(node=0; node < MAX_NUMNODES; node++) node 62 arch/ia64/kernel/numa.c cpumask_clear(&node_to_cpu_mask[node]); node 65 arch/ia64/kernel/numa.c node = NUMA_NO_NODE; node 68 arch/ia64/kernel/numa.c node = node_cpuid[i].nid; node 71 arch/ia64/kernel/numa.c map_cpu_to_node(cpu, node); node 55 arch/ia64/mm/discontig.c #define NODEDATA_ALIGN(addr, node) \ node 57 arch/ia64/mm/discontig.c (((node)*PERCPU_PAGE_SIZE) & (MAX_NODE_ALIGN_OFFSET - 1))) node 72 arch/ia64/mm/discontig.c int node) node 79 arch/ia64/mm/discontig.c if (!mem_data[node].min_pfn) { node 80 arch/ia64/mm/discontig.c mem_data[node].min_pfn = spfn; node 81 arch/ia64/mm/discontig.c mem_data[node].max_pfn = epfn; node 83 arch/ia64/mm/discontig.c mem_data[node].min_pfn = min(spfn, mem_data[node].min_pfn); node 84 arch/ia64/mm/discontig.c mem_data[node].max_pfn = max(epfn, mem_data[node].max_pfn); node 98 arch/ia64/mm/discontig.c static int __meminit early_nr_cpus_node(int node) node 103 arch/ia64/mm/discontig.c if (node == node_cpuid[cpu].nid) node 113 arch/ia64/mm/discontig.c static unsigned long __meminit compute_pernodesize(int node) node 117 arch/ia64/mm/discontig.c cpus = early_nr_cpus_node(node); node 119 arch/ia64/mm/discontig.c pernodesize += node * L1_CACHE_BYTES; node 136 arch/ia64/mm/discontig.c static void *per_cpu_node_setup(void *cpu_data, int node) node 144 arch/ia64/mm/discontig.c if (node != node_cpuid[cpu].nid) node 189 arch/ia64/mm/discontig.c int node, prev_node, unit, nr_units; node 205 arch/ia64/mm/discontig.c for_each_node(node) node 207 arch/ia64/mm/discontig.c if (node == node_cpuid[cpu].nid) node 234 arch/ia64/mm/discontig.c node = node_cpuid[cpu].nid; node 236 arch/ia64/mm/discontig.c if (node == prev_node) { node 240 arch/ia64/mm/discontig.c prev_node = node; node 259 arch/ia64/mm/discontig.c static void __init fill_pernode(int node, unsigned long pernode, node 263 arch/ia64/mm/discontig.c int cpus = early_nr_cpus_node(node); node 265 arch/ia64/mm/discontig.c mem_data[node].pernode_addr = pernode; node 266 arch/ia64/mm/discontig.c mem_data[node].pernode_size = pernodesize; node 271 arch/ia64/mm/discontig.c pernode += node * L1_CACHE_BYTES; node 273 arch/ia64/mm/discontig.c pgdat_list[node] = __va(pernode); node 276 arch/ia64/mm/discontig.c mem_data[node].node_data = __va(pernode); node 280 arch/ia64/mm/discontig.c cpu_data = per_cpu_node_setup(cpu_data, node); node 314 arch/ia64/mm/discontig.c int node) node 326 arch/ia64/mm/discontig.c if (spfn < mem_data[node].min_pfn || epfn > mem_data[node].max_pfn) node 330 arch/ia64/mm/discontig.c if (mem_data[node].pernode_addr) node 337 arch/ia64/mm/discontig.c pernodesize = compute_pernodesize(node); node 338 arch/ia64/mm/discontig.c pernode = NODEDATA_ALIGN(start, node); node 342 arch/ia64/mm/discontig.c fill_pernode(node, pernode, pernodesize); node 357 arch/ia64/mm/discontig.c int node; node 359 arch/ia64/mm/discontig.c for_each_online_node(node) { node 360 arch/ia64/mm/discontig.c if (node_isset(node, memory_less_mask)) node 364 arch/ia64/mm/discontig.c size = mem_data[node].pernode_size; node 365 arch/ia64/mm/discontig.c base = __pa(mem_data[node].pernode_addr); node 373 arch/ia64/mm/discontig.c int node; node 383 arch/ia64/mm/discontig.c for_each_node(node) { node 384 arch/ia64/mm/discontig.c if (pgdat_list[node]) { node 385 arch/ia64/mm/discontig.c dst = LOCAL_DATA_ADDR(pgdat_list[node])->pg_data_ptrs; node 400 arch/ia64/mm/discontig.c int cpu, node; node 407 arch/ia64/mm/discontig.c node = node_cpuid[cpu].nid; node 409 arch/ia64/mm/discontig.c mem_data[node].node_data; node 415 arch/ia64/mm/discontig.c node = node_cpuid[cpu].nid; node 418 arch/ia64/mm/discontig.c cpu0_cpu_info->node_data = mem_data[node].node_data; node 434 arch/ia64/mm/discontig.c int bestnode = NUMA_NO_NODE, node, anynode = 0; node 436 arch/ia64/mm/discontig.c for_each_online_node(node) { node 437 arch/ia64/mm/discontig.c if (node_isset(node, memory_less_mask)) node 439 arch/ia64/mm/discontig.c else if (node_distance(nid, node) < best) { node 440 arch/ia64/mm/discontig.c best = node_distance(nid, node); node 441 arch/ia64/mm/discontig.c bestnode = node; node 443 arch/ia64/mm/discontig.c anynode = node; node 469 arch/ia64/mm/discontig.c int node; node 471 arch/ia64/mm/discontig.c for_each_node_mask(node, memory_less_mask) { node 472 arch/ia64/mm/discontig.c pernodesize = compute_pernodesize(node); node 473 arch/ia64/mm/discontig.c pernode = memory_less_node_alloc(node, pernodesize); node 474 arch/ia64/mm/discontig.c fill_pernode(node, __pa(pernode), pernodesize); node 488 arch/ia64/mm/discontig.c int node; node 507 arch/ia64/mm/discontig.c for_each_online_node(node) node 508 arch/ia64/mm/discontig.c if (mem_data[node].min_pfn) node 509 arch/ia64/mm/discontig.c node_clear(node, memory_less_mask); node 599 arch/ia64/mm/discontig.c int node; node 615 arch/ia64/mm/discontig.c for_each_online_node(node) { node 616 arch/ia64/mm/discontig.c pfn_offset = mem_data[node].min_pfn; node 619 arch/ia64/mm/discontig.c NODE_DATA(node)->node_mem_map = vmem_map + pfn_offset; node 621 arch/ia64/mm/discontig.c if (mem_data[node].max_pfn > max_pfn) node 622 arch/ia64/mm/discontig.c max_pfn = mem_data[node].max_pfn; node 656 arch/ia64/mm/discontig.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 659 arch/ia64/mm/discontig.c return vmemmap_populate_basepages(start, end, node); node 373 arch/ia64/mm/init.c int vmemmap_find_next_valid_pfn(int node, int i) node 377 arch/ia64/mm/init.c pg_data_t *pgdat = NODE_DATA(node); node 431 arch/ia64/mm/init.c int node; node 442 arch/ia64/mm/init.c node = paddr_to_nid(__pa(start)); node 447 arch/ia64/mm/init.c pud = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); node 455 arch/ia64/mm/init.c pmd = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); node 463 arch/ia64/mm/init.c pte = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node); node 472 arch/ia64/mm/init.c node); node 483 arch/ia64/mm/init.c __func__, PAGE_SIZE, PAGE_SIZE, node); node 305 arch/ia64/pci/pci.c info->controller.node = acpi_get_node(device->handle); node 116 arch/m68k/include/asm/dvma.h int node; /* Prom node for this DMA device */ node 278 arch/m68k/include/asm/openprom.h int (*no_nextnode)(int node); node 279 arch/m68k/include/asm/openprom.h int (*no_child)(int node); node 280 arch/m68k/include/asm/openprom.h int (*no_proplen)(int node, char *name); node 281 arch/m68k/include/asm/openprom.h int (*no_getprop)(int node, char *name, char *val); node 282 arch/m68k/include/asm/openprom.h int (*no_setprop)(int node, char *name, char *val, int len); node 283 arch/m68k/include/asm/openprom.h char * (*no_nextprop)(int node, char *name); node 219 arch/m68k/include/asm/oplib.h extern int prom_getsibling(int node); node 233 arch/m68k/include/asm/oplib.h extern int prom_getint(int node, char *property); node 236 arch/m68k/include/asm/oplib.h extern int prom_getintdefault(int node, char *property, int defval); node 239 arch/m68k/include/asm/oplib.h extern int prom_getbool(int node, char *prop); node 242 arch/m68k/include/asm/oplib.h extern void prom_getstring(int node, char *prop, char *buf, int bufsize); node 255 arch/m68k/include/asm/oplib.h extern char *prom_firstprop(int node); node 260 arch/m68k/include/asm/oplib.h extern char *prom_nextprop(int node, char *prev_property); node 263 arch/m68k/include/asm/oplib.h extern int prom_node_has_property(int node, char *property); node 268 arch/m68k/include/asm/oplib.h extern int prom_setprop(int node, char *prop_name, char *prop_value, node 288 arch/m68k/include/asm/oplib.h extern void prom_apply_generic_ranges(int node, int parent, node 19 arch/m68k/include/asm/pgalloc.h extern void m68k_setup_node(int node); node 60 arch/m68k/mm/init.c void __init m68k_setup_node(int node) node 63 arch/m68k/mm/init.c struct m68k_mem_info *info = m68k_memory + node; node 70 arch/m68k/mm/init.c pr_warn("overlap at %u for chunk %u\n", i, node); node 71 arch/m68k/mm/init.c pg_data_table[i] = pg_data_map + node; node 74 arch/m68k/mm/init.c node_set_online(node); node 115 arch/m68k/mm/motorola.c static void __init map_node(int node) node 124 arch/m68k/mm/motorola.c size = m68k_memory[node].size; node 125 arch/m68k/mm/motorola.c physaddr = m68k_memory[node].addr; node 133 arch/m68k/sun3x/prom.c int prom_getintdefault(int node, char *property, int deflt) node 138 arch/m68k/sun3x/prom.c int prom_getbool (int node, char *prop) node 130 arch/microblaze/include/asm/pci-bridge.h struct device_node *node); node 132 arch/microblaze/pci/pci-common.c struct pci_controller *pci_find_hose_for_OF_device(struct device_node *node) node 134 arch/microblaze/pci/pci-common.c while (node) { node 137 arch/microblaze/pci/pci-common.c if (hose->dn == node) node 139 arch/microblaze/pci/pci-common.c node = node->parent; node 741 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &bus->children, node) node 883 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &pci_root_buses, node) node 893 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &pci_root_buses, node) node 967 arch/microblaze/pci/pci-common.c struct device_node *node = hose->dn; node 969 arch/microblaze/pci/pci-common.c pr_debug("PCI: Scanning PHB %pOF\n", node); node 45 arch/mips/cavium-octeon/octeon-irq.c int node; node 1147 arch/mips/cavium-octeon/octeon-irq.c struct device_node *node, node 1157 arch/mips/cavium-octeon/octeon-irq.c if (irq_domain_get_of_node(d) != node) node 1184 arch/mips/cavium-octeon/octeon-irq.c node, node 1196 arch/mips/cavium-octeon/octeon-irq.c struct device_node *node, node 1891 arch/mips/cavium-octeon/octeon-irq.c struct device_node *node, node 2161 arch/mips/cavium-octeon/octeon-irq.c struct device_node *node, node 2334 arch/mips/cavium-octeon/octeon-irq.c struct device_node *node, node 2547 arch/mips/cavium-octeon/octeon-irq.c ciu3_info->node); node 2553 arch/mips/cavium-octeon/octeon-irq.c cd->ciu_node = ciu3_info->node; node 2850 arch/mips/cavium-octeon/octeon-irq.c int node; node 2857 arch/mips/cavium-octeon/octeon-irq.c node = 0; /* of_node_to_nid(ciu_node); */ node 2858 arch/mips/cavium-octeon/octeon-irq.c ciu3_info = kzalloc_node(sizeof(*ciu3_info), GFP_KERNEL, node); node 2871 arch/mips/cavium-octeon/octeon-irq.c ciu3_info->node = node; node 2881 arch/mips/cavium-octeon/octeon-irq.c if (node == cvmx_get_node_num()) { node 2886 arch/mips/cavium-octeon/octeon-irq.c i = irq_alloc_descs_from(OCTEON_IRQ_MBOX0, 8, node); node 2903 arch/mips/cavium-octeon/octeon-irq.c octeon_ciu3_info_per_node[node] = ciu3_info; node 2905 arch/mips/cavium-octeon/octeon-irq.c if (node == cvmx_get_node_num()) { node 2908 arch/mips/cavium-octeon/octeon-irq.c if (node == 0) node 2972 arch/mips/cavium-octeon/octeon-irq.c struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block) node 2976 arch/mips/cavium-octeon/octeon-irq.c ciu3_info = octeon_ciu3_info_per_node[node & CVMX_NODE_MASK]; node 591 arch/mips/cavium-octeon/octeon-platform.c static void __init octeon_fdt_rm_ethernet(int node) node 595 arch/mips/cavium-octeon/octeon-platform.c phy_handle = fdt_getprop(initial_boot_params, node, "phy-handle", NULL); node 603 arch/mips/cavium-octeon/octeon-platform.c fdt_nop_node(initial_boot_params, node); node 235 arch/mips/cavium-octeon/octeon-usb.c struct device_node *node = dev->of_node; node 238 arch/mips/cavium-octeon/octeon-usb.c if (of_find_property(node, "power", &len) != NULL) { node 240 arch/mips/cavium-octeon/octeon-usb.c of_property_read_u32_array(node, "power", gpio_pwr, 3); node 244 arch/mips/cavium-octeon/octeon-usb.c of_property_read_u32_array(node, "power", gpio_pwr, 2); node 500 arch/mips/cavium-octeon/octeon-usb.c struct device_node *node; node 508 arch/mips/cavium-octeon/octeon-usb.c node = NULL; node 510 arch/mips/cavium-octeon/octeon-usb.c node = of_find_node_by_name(node, "uctl"); node 511 arch/mips/cavium-octeon/octeon-usb.c if (!node) node 514 arch/mips/cavium-octeon/octeon-usb.c if (of_device_is_compatible(node, compat_node_name)) { node 515 arch/mips/cavium-octeon/octeon-usb.c pdev = of_find_device_by_node(node); node 545 arch/mips/cavium-octeon/octeon-usb.c } while (node != NULL); node 22 arch/mips/include/asm/clock.h struct list_head node; node 46 arch/mips/include/asm/mach-ip27/kernel-entry-init.h li t0, 0x1c000 # Offset of text into node memory node 18 arch/mips/include/asm/mach-ip27/topology.h #define cpumask_of_node(node) ((node) == -1 ? \ node 20 arch/mips/include/asm/mach-ip27/topology.h &hub_data(node)->h_cpus) node 8 arch/mips/include/asm/mach-loongson64/topology.h #define cpumask_of_node(node) (&__node_data[(node)]->cpumask) node 72 arch/mips/include/asm/mach-netlogic/multi-node.h void nlm_node_init(int node); node 94 arch/mips/include/asm/netlogic/common.h uint64_t nlm_pci_irqmask(int node); node 95 arch/mips/include/asm/netlogic/common.h void nlm_setup_pic_irq(int node, int picirq, int irq, int irt); node 96 arch/mips/include/asm/netlogic/common.h void nlm_set_pic_extra_ack(int node, int irq, void (*xack)(struct irq_data *)); node 99 arch/mips/include/asm/netlogic/common.h void nlm_dispatch_msi(int node, int lirq); node 100 arch/mips/include/asm/netlogic/common.h void nlm_dispatch_msix(int node, int msixirq); node 106 arch/mips/include/asm/netlogic/common.h static inline int nlm_irq_to_xirq(int node, int irq) node 108 arch/mips/include/asm/netlogic/common.h return node * NR_IRQS / NLM_NR_NODES + irq; node 180 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_get_bridge_pcibase(node) nlm_pcicfg_base(cpu_is_xlp9xx() ? \ node 181 arch/mips/include/asm/netlogic/xlp-hal/bridge.h XLP9XX_IO_BRIDGE_OFFSET(node) : XLP_IO_BRIDGE_OFFSET(node)) node 182 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_get_bridge_regbase(node) \ node 183 arch/mips/include/asm/netlogic/xlp-hal/bridge.h (nlm_get_bridge_pcibase(node) + XLP_IO_PCI_HDRSZ) node 50 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_DEV(node, dev) ((dev) + (node) * 8) node 53 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_HDR_OFFSET(node, bus, dev, fn) \ node 54 arch/mips/include/asm/netlogic/xlp-hal/iomap.h XLP_IO_PCI_OFFSET(bus, XLP_IO_DEV(node, dev), fn) node 56 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_BRIDGE_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 0) node 58 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 1) node 59 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 2) node 60 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC2_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 3) node 61 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PIC_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 4) node 63 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 1, i) node 64 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 1, 0) node 65 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 1, 1) node 66 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE2_OFFSET(node) XLP_HDR_OFFSET(node, 0, 1, 2) node 67 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE3_OFFSET(node) XLP_HDR_OFFSET(node, 0, 1, 3) node 69 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 2, i) node 70 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_EHCI0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 0) node 71 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 1) node 72 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 2) node 73 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_EHCI1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 3) node 74 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI2_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 4) node 75 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI3_OFFSET(node) XLP_HDR_OFFSET(node, 0, 2, 5) node 77 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SATA_OFFSET(node) XLP_HDR_OFFSET(node, 0, 3, 2) node 80 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 4, i) node 81 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 4, 1) node 82 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 4, 2) node 83 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI2_OFFSET(node) XLP_HDR_OFFSET(node, 0, 4, 3) node 85 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NAE_OFFSET(node) XLP_HDR_OFFSET(node, 0, 3, 0) node 86 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_POE_OFFSET(node) XLP_HDR_OFFSET(node, 0, 3, 1) node 88 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CMS_OFFSET(node) XLP_HDR_OFFSET(node, 0, 4, 0) node 90 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_DMA_OFFSET(node) XLP_HDR_OFFSET(node, 0, 5, 1) node 91 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SEC_OFFSET(node) XLP_HDR_OFFSET(node, 0, 5, 2) node 92 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CMP_OFFSET(node) XLP_HDR_OFFSET(node, 0, 5, 3) node 94 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 6, i) node 95 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 0) node 96 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 1) node 97 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 6, 2 + i) node 98 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 2) node 99 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 3) node 100 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_GPIO_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 4) node 102 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_I2C_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 7) node 105 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SYS_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 5) node 106 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_JTAG_OFFSET(node) XLP_HDR_OFFSET(node, 0, 6, 6) node 109 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NOR_OFFSET(node) XLP_HDR_OFFSET(node, 0, 7, 0) node 110 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NAND_OFFSET(node) XLP_HDR_OFFSET(node, 0, 7, 1) node 111 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SPI_OFFSET(node) XLP_HDR_OFFSET(node, 0, 7, 2) node 112 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_MMC_OFFSET(node) XLP_HDR_OFFSET(node, 0, 7, 3) node 118 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_BRIDGE_OFFSET(node) XLP_IO_PCI_OFFSET(0, 0, node) node 119 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PIC_OFFSET(node) XLP9XX_HDR_OFFSET(node, 2, 0) node 120 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_UART_OFFSET(node) XLP9XX_HDR_OFFSET(node, 2, 2) node 121 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SYS_OFFSET(node) XLP9XX_HDR_OFFSET(node, 6, 0) node 122 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_FUSE_OFFSET(node) XLP9XX_HDR_OFFSET(node, 6, 1) node 123 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_CLOCK_OFFSET(node) XLP9XX_HDR_OFFSET(node, 6, 2) node 124 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_POWER_OFFSET(node) XLP9XX_HDR_OFFSET(node, 6, 3) node 125 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_JTAG_OFFSET(node) XLP9XX_HDR_OFFSET(node, 6, 4) node 127 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE_OFFSET(node, i) XLP9XX_HDR_OFFSET(node, 1, i) node 128 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE0_OFFSET(node) XLP9XX_HDR_OFFSET(node, 1, 0) node 129 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE2_OFFSET(node) XLP9XX_HDR_OFFSET(node, 1, 2) node 130 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE3_OFFSET(node) XLP9XX_HDR_OFFSET(node, 1, 3) node 133 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_OFFSET(node, i) XLP9XX_HDR_OFFSET(node, 4, i) node 134 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_XHCI0_OFFSET(node) XLP9XX_HDR_OFFSET(node, 4, 1) node 135 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_XHCI1_OFFSET(node) XLP9XX_HDR_OFFSET(node, 4, 2) node 138 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SATA_OFFSET(node) XLP9XX_HDR_OFFSET(node, 3, 2) node 141 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_NOR_OFFSET(node) XLP9XX_HDR_OFFSET(node, 7, 0) node 142 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_NAND_OFFSET(node) XLP9XX_HDR_OFFSET(node, 7, 1) node 143 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SPI_OFFSET(node) XLP9XX_HDR_OFFSET(node, 7, 2) node 144 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_MMC_OFFSET(node) XLP9XX_HDR_OFFSET(node, 7, 3) node 203 arch/mips/include/asm/netlogic/xlp-hal/iomap.h static inline int xlp9xx_get_socbus(int node) node 207 arch/mips/include/asm/netlogic/xlp-hal/iomap.h if (node == 0) node 209 arch/mips/include/asm/netlogic/xlp-hal/iomap.h socbridge = nlm_pcicfg_base(XLP9XX_IO_BRIDGE_OFFSET(node)); node 101 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h #define nlm_get_pcie_base(node, inst) nlm_pcicfg_base(cpu_is_xlp9xx() ? \ node 102 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h XLP9XX_IO_PCIE_OFFSET(node, inst) : XLP_IO_PCIE_OFFSET(node, inst)) node 105 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h void xlp_init_node_msi_irqs(int node, int link); node 107 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h static inline void xlp_init_node_msi_irqs(int node, int link) {} node 219 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_get_pic_pcibase(node) nlm_pcicfg_base(cpu_is_xlp9xx() ? \ node 220 arch/mips/include/asm/netlogic/xlp-hal/pic.h XLP9XX_IO_PIC_OFFSET(node) : XLP_IO_PIC_OFFSET(node)) node 221 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_get_pic_regbase(node) (nlm_get_pic_pcibase(node) + XLP_IO_PCI_HDRSZ) node 196 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_sys_pcibase(node) nlm_pcicfg_base(cpu_is_xlp9xx() ? \ node 197 arch/mips/include/asm/netlogic/xlp-hal/sys.h XLP9XX_IO_SYS_OFFSET(node) : XLP_IO_SYS_OFFSET(node)) node 198 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_sys_regbase(node) (nlm_get_sys_pcibase(node) + XLP_IO_PCI_HDRSZ) node 201 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_fuse_pcibase(node) \ node 202 arch/mips/include/asm/netlogic/xlp-hal/sys.h nlm_pcicfg_base(XLP9XX_IO_FUSE_OFFSET(node)) node 203 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_fuse_regbase(node) \ node 204 arch/mips/include/asm/netlogic/xlp-hal/sys.h (nlm_get_fuse_pcibase(node) + XLP_IO_PCI_HDRSZ) node 206 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_clock_pcibase(node) \ node 207 arch/mips/include/asm/netlogic/xlp-hal/sys.h nlm_pcicfg_base(XLP9XX_IO_CLOCK_OFFSET(node)) node 208 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_clock_regbase(node) \ node 209 arch/mips/include/asm/netlogic/xlp-hal/sys.h (nlm_get_clock_pcibase(node) + XLP_IO_PCI_HDRSZ) node 211 arch/mips/include/asm/netlogic/xlp-hal/sys.h unsigned int nlm_get_pic_frequency(int node); node 96 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_get_uart_pcibase(node, inst) \ node 97 arch/mips/include/asm/netlogic/xlp-hal/uart.h nlm_pcicfg_base(cpu_is_xlp9xx() ? XLP9XX_IO_UART_OFFSET(node) : \ node 98 arch/mips/include/asm/netlogic/xlp-hal/uart.h XLP_IO_UART_OFFSET(node, inst)) node 99 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_get_uart_regbase(node, inst) \ node 100 arch/mips/include/asm/netlogic/xlp-hal/uart.h (nlm_get_uart_pcibase(node, inst) + XLP_IO_PCI_HDRSZ) node 93 arch/mips/include/asm/netlogic/xlp-hal/xlp.h int nlm_get_dram_map(int node, uint64_t *dram_map, int nentries); node 367 arch/mips/include/asm/octeon/cvmx.h static inline void cvmx_write_csr_node(uint64_t node, uint64_t csr_addr, node 372 arch/mips/include/asm/octeon/cvmx.h node_addr = (node & CVMX_NODE_MASK) << CVMX_NODE_IO_SHIFT; node 380 arch/mips/include/asm/octeon/cvmx.h static inline uint64_t cvmx_read_csr_node(uint64_t node, uint64_t csr_addr) node 385 arch/mips/include/asm/octeon/cvmx.h (node & CVMX_NODE_MASK) << CVMX_NODE_IO_SHIFT; node 313 arch/mips/include/asm/octeon/octeon.h struct device_node *node, node 364 arch/mips/include/asm/octeon/octeon.h struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block); node 74 arch/mips/include/asm/pci.h struct device_node *node); node 77 arch/mips/include/asm/pci.h struct device_node *node) {} node 680 arch/mips/include/asm/r4kcache.h static inline void blast_##pfx##cache##lsize##_node(long node) \ node 682 arch/mips/include/asm/r4kcache.h unsigned long start = CAC_BASE | nid_to_addrbase(node); \ node 130 arch/mips/include/asm/sn/sn0/hubio.h #define IIO_IGFX_INIT(widget, node, cpu, valid) (\ node 132 arch/mips/include/asm/sn/sn0/hubio.h (((node) & IIO_IGFX_N_NUM_MASK) << IIO_IGFX_N_NUM_SHIFT) | \ node 915 arch/mips/include/asm/sn/sn0/hubio.h node : 9, node 216 arch/mips/kernel/relocate.c int node, len; node 219 arch/mips/kernel/relocate.c node = fdt_path_offset(initial_boot_params, "/chosen"); node 220 arch/mips/kernel/relocate.c if (node >= 0) { node 221 arch/mips/kernel/relocate.c prop = fdt_getprop_w(initial_boot_params, node, node 238 arch/mips/kernel/smp.c struct device_node *node; node 240 arch/mips/kernel/smp.c node = of_irq_find_parent(of_root); node 241 arch/mips/kernel/smp.c ipidomain = irq_find_matching_host(node, DOMAIN_BUS_IPI); node 248 arch/mips/kernel/smp.c if (node && !ipidomain) node 295 arch/mips/kernel/smp.c struct device_node *node; node 297 arch/mips/kernel/smp.c node = of_irq_find_parent(of_root); node 298 arch/mips/kernel/smp.c ipidomain = irq_find_matching_host(node, DOMAIN_BUS_IPI); node 305 arch/mips/kernel/smp.c if (node && !ipidomain) node 337 arch/mips/lantiq/irq.c int __init icu_of_init(struct device_node *node, struct device_node *parent) node 345 arch/mips/lantiq/irq.c if (of_address_to_resource(node, vpe, &res)) node 379 arch/mips/lantiq/irq.c ltq_domain = irq_domain_add_linear(node, node 60 arch/mips/loongson64/lemote-2f/clock.c list_for_each_entry(clkp, &clock_list, node) { node 119 arch/mips/loongson64/loongson-3/numa.c static void __init szmem(unsigned int node) node 128 arch/mips/loongson64/loongson-3/numa.c if (node_id != node) node 146 arch/mips/loongson64/loongson-3/numa.c PFN_PHYS(end_pfn - start_pfn), node); node 158 arch/mips/loongson64/loongson-3/numa.c PFN_PHYS(end_pfn - start_pfn), node); node 170 arch/mips/loongson64/loongson-3/numa.c static void __init node_mem_init(unsigned int node) node 175 arch/mips/loongson64/loongson-3/numa.c node_addrspace_offset = nid_to_addroffset(node); node 177 arch/mips/loongson64/loongson-3/numa.c node, node_addrspace_offset); node 179 arch/mips/loongson64/loongson-3/numa.c get_pfn_range_for_nid(node, &start_pfn, &end_pfn); node 181 arch/mips/loongson64/loongson-3/numa.c node, start_pfn, end_pfn); node 183 arch/mips/loongson64/loongson-3/numa.c __node_data[node] = prealloc__node_data + node; node 185 arch/mips/loongson64/loongson-3/numa.c NODE_DATA(node)->node_start_pfn = start_pfn; node 186 arch/mips/loongson64/loongson-3/numa.c NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; node 188 arch/mips/loongson64/loongson-3/numa.c if (node == 0) { node 208 arch/mips/loongson64/loongson-3/numa.c unsigned int node, cpu, active_cpu = 0; node 213 arch/mips/loongson64/loongson-3/numa.c for (node = 0; node < loongson_sysconf.nr_nodes; node++) { node 214 arch/mips/loongson64/loongson-3/numa.c if (node_online(node)) { node 215 arch/mips/loongson64/loongson-3/numa.c szmem(node); node 216 arch/mips/loongson64/loongson-3/numa.c node_mem_init(node); node 217 arch/mips/loongson64/loongson-3/numa.c cpumask_clear(&__node_data[(node)]->cpumask); node 224 arch/mips/loongson64/loongson-3/numa.c node = cpu / loongson_sysconf.cores_per_node; node 225 arch/mips/loongson64/loongson-3/numa.c if (node >= num_online_nodes()) node 226 arch/mips/loongson64/loongson-3/numa.c node = 0; node 231 arch/mips/loongson64/loongson-3/numa.c cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask); node 232 arch/mips/loongson64/loongson-3/numa.c pr_info("NUMA: set cpumask cpu %d on node %d\n", active_cpu, node); node 456 arch/mips/loongson64/loongson-3/smp.c register long cpuid, core, node, count; node 508 arch/mips/loongson64/loongson-3/smp.c : [core] "=&r" (core), [node] "=&r" (node), node 518 arch/mips/loongson64/loongson-3/smp.c register long cpuid, core, node, count; node 591 arch/mips/loongson64/loongson-3/smp.c : [core] "=&r" (core), [node] "=&r" (node), node 601 arch/mips/loongson64/loongson-3/smp.c register long cpuid, core, node, count; node 655 arch/mips/loongson64/loongson-3/smp.c : [core] "=&r" (core), [node] "=&r" (node), node 462 arch/mips/mm/c-r4k.c static void (*r4k_blast_scache_node)(long node); node 210 arch/mips/mti-malta/malta-time.c struct device_node *node; node 214 arch/mips/mti-malta/malta-time.c node = of_find_compatible_node(NULL, NULL, "mti,gic-timer"); node 215 arch/mips/mti-malta/malta-time.c if (!node) { node 220 arch/mips/mti-malta/malta-time.c if (of_update_property(node, &gic_frequency_prop) < 0) node 81 arch/mips/netlogic/common/irq.c struct nlm_soc_info *node; node 93 arch/mips/netlogic/common/irq.c spin_lock_irqsave(&pd->node->piclock, flags); node 94 arch/mips/netlogic/common/irq.c nlm_pic_enable_irt(pd->node->picbase, pd->irt); node 95 arch/mips/netlogic/common/irq.c spin_unlock_irqrestore(&pd->node->piclock, flags); node 104 arch/mips/netlogic/common/irq.c spin_lock_irqsave(&pd->node->piclock, flags); node 105 arch/mips/netlogic/common/irq.c nlm_pic_disable_irt(pd->node->picbase, pd->irt); node 106 arch/mips/netlogic/common/irq.c spin_unlock_irqrestore(&pd->node->piclock, flags); node 130 arch/mips/netlogic/common/irq.c nlm_pic_ack(pd->node->picbase, pd->irt); node 184 arch/mips/netlogic/common/irq.c void nlm_setup_pic_irq(int node, int picirq, int irq, int irt) node 189 arch/mips/netlogic/common/irq.c xirq = nlm_irq_to_xirq(node, irq); node 194 arch/mips/netlogic/common/irq.c pic_data->node = nlm_get_node(node); node 199 arch/mips/netlogic/common/irq.c void nlm_set_pic_extra_ack(int node, int irq, void (*xack)(struct irq_data *)) node 204 arch/mips/netlogic/common/irq.c xirq = nlm_irq_to_xirq(node, irq); node 211 arch/mips/netlogic/common/irq.c static void nlm_init_node_irqs(int node) node 216 arch/mips/netlogic/common/irq.c pr_info("Init IRQ for node %d\n", node); node 217 arch/mips/netlogic/common/irq.c nodep = nlm_get_node(node); node 228 arch/mips/netlogic/common/irq.c node * nlm_threads_per_node(), 0); node 229 arch/mips/netlogic/common/irq.c nlm_setup_pic_irq(node, i, i, irt); node 235 arch/mips/netlogic/common/irq.c int cpu, node; node 238 arch/mips/netlogic/common/irq.c node = hwtid / nlm_threads_per_node(); node 240 arch/mips/netlogic/common/irq.c if (cpu == 0 && node != 0) node 241 arch/mips/netlogic/common/irq.c nlm_init_node_irqs(node); node 242 arch/mips/netlogic/common/irq.c write_c0_eimr(nlm_get_node(node)->irqmask); node 248 arch/mips/netlogic/common/irq.c int i, node; node 250 arch/mips/netlogic/common/irq.c node = nlm_nodeid(); node 265 arch/mips/netlogic/common/irq.c nlm_dispatch_msi(node, i); node 269 arch/mips/netlogic/common/irq.c nlm_dispatch_msix(node, i); node 275 arch/mips/netlogic/common/irq.c do_IRQ(nlm_irq_to_xirq(node, i)); node 283 arch/mips/netlogic/common/irq.c static int __init xlp_of_pic_init(struct device_node *node, node 292 arch/mips/netlogic/common/irq.c ret = of_address_to_resource(node, 0, &res); node 294 arch/mips/netlogic/common/irq.c pr_err("PIC %pOFn: reg property not found!\n", node); node 308 arch/mips/netlogic/common/irq.c node, bus); node 315 arch/mips/netlogic/common/irq.c node, socid); node 321 arch/mips/netlogic/common/irq.c pr_err("PIC %pOFn: node %d does not exist!\n", node, socid); node 325 arch/mips/netlogic/common/irq.c xlp_pic_domain = irq_domain_add_legacy(node, n_picirqs, node 329 arch/mips/netlogic/common/irq.c pr_err("PIC %pOFn: Creating legacy domain failed!\n", node); node 171 arch/mips/netlogic/common/smp.c int num_cpus, i, ncore, node; node 193 arch/mips/netlogic/common/smp.c node = nlm_hwtid_to_node(i); node 194 arch/mips/netlogic/common/smp.c cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask); node 145 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_get_sata_pcibase(node) \ node 146 arch/mips/netlogic/xlp/ahci-init-xlp2.c nlm_pcicfg_base(XLP9XX_IO_SATA_OFFSET(node)) node 147 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_get_sata_regbase(node) \ node 148 arch/mips/netlogic/xlp/ahci-init-xlp2.c (nlm_get_sata_pcibase(node) + 0x100) node 253 arch/mips/netlogic/xlp/ahci-init-xlp2.c static void nlm_sata_firmware_init(int node) node 260 arch/mips/netlogic/xlp/ahci-init-xlp2.c regbase = nlm_get_sata_regbase(node); node 335 arch/mips/netlogic/xlp/ahci-init-xlp2.c int node; node 339 arch/mips/netlogic/xlp/ahci-init-xlp2.c for (node = 0; node < NLM_NR_NODES; node++) node 340 arch/mips/netlogic/xlp/ahci-init-xlp2.c if (nlm_node_present(node)) node 341 arch/mips/netlogic/xlp/ahci-init-xlp2.c nlm_sata_firmware_init(node); node 349 arch/mips/netlogic/xlp/ahci-init-xlp2.c int node; node 351 arch/mips/netlogic/xlp/ahci-init-xlp2.c node = data->irq / NLM_IRQS_PER_NODE; node 352 arch/mips/netlogic/xlp/ahci-init-xlp2.c regbase = nlm_get_sata_regbase(node); node 367 arch/mips/netlogic/xlp/ahci-init-xlp2.c int node; node 370 arch/mips/netlogic/xlp/ahci-init-xlp2.c node = xlp_socdev_to_node(dev); node 371 arch/mips/netlogic/xlp/ahci-init-xlp2.c regbase = nlm_get_sata_regbase(node); node 380 arch/mips/netlogic/xlp/ahci-init-xlp2.c dev->irq = nlm_irq_to_xirq(node, PIC_SATA_IRQ); node 381 arch/mips/netlogic/xlp/ahci-init-xlp2.c nlm_set_pic_extra_ack(node, PIC_SATA_IRQ, nlm_sata_intr_ack); node 89 arch/mips/netlogic/xlp/ahci-init.c #define nlm_get_sata_pcibase(node) \ node 90 arch/mips/netlogic/xlp/ahci-init.c nlm_pcicfg_base(XLP_IO_SATA_OFFSET(node)) node 92 arch/mips/netlogic/xlp/ahci-init.c #define nlm_get_sata_regbase(node) \ node 93 arch/mips/netlogic/xlp/ahci-init.c (nlm_get_sata_pcibase(node) + 0x900) node 111 arch/mips/netlogic/xlp/ahci-init.c static void nlm_sata_firmware_init(int node) node 118 arch/mips/netlogic/xlp/ahci-init.c regbase = nlm_get_sata_regbase(node); node 153 arch/mips/netlogic/xlp/ahci-init.c int node = 0; node 157 arch/mips/netlogic/xlp/ahci-init.c nlm_sata_firmware_init(node); node 185 arch/mips/netlogic/xlp/ahci-init.c int node = 0; /* XLP3XX does not support multi-node */ node 187 arch/mips/netlogic/xlp/ahci-init.c regbase = nlm_get_sata_regbase(node); node 201 arch/mips/netlogic/xlp/ahci-init.c nlm_set_pic_extra_ack(node, PIC_SATA_IRQ, nlm_sata_intr_ack); node 52 arch/mips/netlogic/xlp/nlm_hal.c void nlm_node_init(int node) node 56 arch/mips/netlogic/xlp/nlm_hal.c nodep = nlm_get_node(node); node 57 arch/mips/netlogic/xlp/nlm_hal.c if (node == 0) node 59 arch/mips/netlogic/xlp/nlm_hal.c nodep->sysbase = nlm_get_sys_regbase(node); node 60 arch/mips/netlogic/xlp/nlm_hal.c nodep->picbase = nlm_get_pic_regbase(node); node 63 arch/mips/netlogic/xlp/nlm_hal.c nodep->socbus = xlp9xx_get_socbus(node); node 228 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp2_get_core_frequency(int node, int core) node 234 arch/mips/netlogic/xlp/nlm_hal.c clockbase = nlm_get_clock_regbase(node); node 240 arch/mips/netlogic/xlp/nlm_hal.c sysbase = nlm_get_node(node)->sysbase; node 274 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp_get_core_frequency(int node, int core) node 280 arch/mips/netlogic/xlp/nlm_hal.c sysbase = nlm_get_node(node)->sysbase; node 295 arch/mips/netlogic/xlp/nlm_hal.c unsigned int nlm_get_core_frequency(int node, int core) node 298 arch/mips/netlogic/xlp/nlm_hal.c return nlm_xlp2_get_core_frequency(node, core); node 300 arch/mips/netlogic/xlp/nlm_hal.c return nlm_xlp_get_core_frequency(node, core); node 308 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp2_get_pic_frequency(int node) node 314 arch/mips/netlogic/xlp/nlm_hal.c sysbase = nlm_get_node(node)->sysbase; node 315 arch/mips/netlogic/xlp/nlm_hal.c clockbase = nlm_get_clock_regbase(node); node 449 arch/mips/netlogic/xlp/nlm_hal.c unsigned int nlm_get_pic_frequency(int node) node 452 arch/mips/netlogic/xlp/nlm_hal.c return nlm_xlp2_get_pic_frequency(node); node 466 arch/mips/netlogic/xlp/nlm_hal.c int nlm_get_dram_map(int node, uint64_t *dram_map, int nentries) node 488 arch/mips/netlogic/xlp/nlm_hal.c if (node >= 0) { node 492 arch/mips/netlogic/xlp/nlm_hal.c if (n != node) node 87 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_xlpii_get_usb_pcibase(node, inst) \ node 89 arch/mips/netlogic/xlp/usb-init-xlp2.c XLP9XX_IO_USB_OFFSET(node, inst) : \ node 90 arch/mips/netlogic/xlp/usb-init-xlp2.c XLP2XX_IO_USB_OFFSET(node, inst)) node 91 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_xlpii_get_usb_regbase(node, inst) \ node 92 arch/mips/netlogic/xlp/usb-init-xlp2.c (nlm_xlpii_get_usb_pcibase(node, inst) + XLP_IO_PCI_HDRSZ) node 118 arch/mips/netlogic/xlp/usb-init-xlp2.c int node, irq; node 122 arch/mips/netlogic/xlp/usb-init-xlp2.c node = data->irq / NLM_IRQS_PER_NODE; node 126 arch/mips/netlogic/xlp/usb-init-xlp2.c port_addr = nlm_xlpii_get_usb_regbase(node, 1); node 129 arch/mips/netlogic/xlp/usb-init-xlp2.c port_addr = nlm_xlpii_get_usb_regbase(node, 2); node 132 arch/mips/netlogic/xlp/usb-init-xlp2.c port_addr = nlm_xlpii_get_usb_regbase(node, 3); node 135 arch/mips/netlogic/xlp/usb-init-xlp2.c pr_err("No matching USB irq %d node %d!\n", irq, node); node 141 arch/mips/netlogic/xlp/usb-init-xlp2.c static void nlm_xlpii_usb_hw_reset(int node, int port) node 147 arch/mips/netlogic/xlp/usb-init-xlp2.c port_addr = nlm_xlpii_get_usb_regbase(node, port); node 180 arch/mips/netlogic/xlp/usb-init-xlp2.c pci_base = nlm_xlpii_get_usb_pcibase(node, port); node 210 arch/mips/netlogic/xlp/usb-init-xlp2.c int node; node 229 arch/mips/netlogic/xlp/usb-init-xlp2.c for (node = 0; node < NLM_NR_NODES; node++) { node 230 arch/mips/netlogic/xlp/usb-init-xlp2.c if (!nlm_node_present(node)) node 232 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_xlpii_usb_hw_reset(node, 1); node 233 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_xlpii_usb_hw_reset(node, 2); node 234 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_xlpii_usb_hw_reset(node, 3); node 235 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_0_IRQ, xlp9xx_usb_ack); node 236 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_1_IRQ, xlp9xx_usb_ack); node 237 arch/mips/netlogic/xlp/usb-init-xlp2.c nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_2_IRQ, xlp9xx_usb_ack); node 249 arch/mips/netlogic/xlp/usb-init-xlp2.c int node; node 251 arch/mips/netlogic/xlp/usb-init-xlp2.c node = xlp_socdev_to_node(dev); node 256 arch/mips/netlogic/xlp/usb-init-xlp2.c dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_0_IRQ); node 259 arch/mips/netlogic/xlp/usb-init-xlp2.c dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_1_IRQ); node 262 arch/mips/netlogic/xlp/usb-init-xlp2.c dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_2_IRQ); node 65 arch/mips/netlogic/xlp/usb-init.c #define nlm_get_usb_pcibase(node, inst) \ node 66 arch/mips/netlogic/xlp/usb-init.c nlm_pcicfg_base(XLP_IO_USB_OFFSET(node, inst)) node 67 arch/mips/netlogic/xlp/usb-init.c #define nlm_get_usb_regbase(node, inst) \ node 68 arch/mips/netlogic/xlp/usb-init.c (nlm_get_usb_pcibase(node, inst) + XLP_IO_PCI_HDRSZ) node 70 arch/mips/netlogic/xlp/usb-init.c static void nlm_usb_intr_en(int node, int port) node 75 arch/mips/netlogic/xlp/usb-init.c port_addr = nlm_get_usb_regbase(node, port); node 82 arch/mips/netlogic/xlp/usb-init.c static void nlm_usb_hw_reset(int node, int port) node 88 arch/mips/netlogic/xlp/usb-init.c port_addr = nlm_get_usb_regbase(node, port); node 53 arch/mips/netlogic/xlp/wakeup.c static int xlp_wakeup_core(uint64_t sysbase, int node, int core) node 113 arch/mips/pci/msi-xlp.c struct nlm_soc_info *node; node 227 arch/mips/pci/msi-xlp.c nlm_pic_ack(md->node->picbase, node 293 arch/mips/pci/msi-xlp.c static int xlp_setup_msi(uint64_t lnkbase, int node, int link, node 304 arch/mips/pci/msi-xlp.c xirq = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0)); node 306 arch/mips/pci/msi-xlp.c msiaddr = MSI_LINK_ADDR(node, link); node 316 arch/mips/pci/msi-xlp.c nlm_setup_pic_irq(node, lirq, lirq, irt); node 317 arch/mips/pci/msi-xlp.c nlm_pic_init_irt(nlm_get_node(node)->picbase, irt, lirq, node 318 arch/mips/pci/msi-xlp.c node * nlm_threads_per_node(), 1 /*en */); node 400 arch/mips/pci/msi-xlp.c static int xlp_setup_msix(uint64_t lnkbase, int node, int link, node 411 arch/mips/pci/msi-xlp.c xirq = nlm_irq_to_xirq(node, nlm_link_msixirq(link, 0)); node 413 arch/mips/pci/msi-xlp.c msixaddr = MSIX_LINK_ADDR(node, link); node 448 arch/mips/pci/msi-xlp.c int node, link, slot; node 457 arch/mips/pci/msi-xlp.c node = slot / 8; node 458 arch/mips/pci/msi-xlp.c lnkbase = nlm_get_pcie_base(node, link); node 461 arch/mips/pci/msi-xlp.c return xlp_setup_msix(lnkbase, node, link, desc); node 463 arch/mips/pci/msi-xlp.c return xlp_setup_msi(lnkbase, node, link, desc); node 466 arch/mips/pci/msi-xlp.c void __init xlp_init_node_msi_irqs(int node, int link) node 472 arch/mips/pci/msi-xlp.c pr_info("[%d %d] Init node PCI IRT\n", node, link); node 473 arch/mips/pci/msi-xlp.c nodep = nlm_get_node(node); node 481 arch/mips/pci/msi-xlp.c md->node = nodep; node 482 arch/mips/pci/msi-xlp.c md->lnkbase = nlm_get_pcie_base(node, link); node 485 arch/mips/pci/msi-xlp.c irq = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0)); node 493 arch/mips/pci/msi-xlp.c val = ((node * nlm_threads_per_node()) << 7 | node 505 arch/mips/pci/msi-xlp.c node * nlm_threads_per_node(), 1); node 509 arch/mips/pci/msi-xlp.c irq = nlm_irq_to_xirq(node, nlm_link_msixirq(link, i)); node 515 arch/mips/pci/msi-xlp.c void nlm_dispatch_msi(int node, int lirq) node 522 arch/mips/pci/msi-xlp.c irqbase = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0)); node 539 arch/mips/pci/msi-xlp.c nlm_pic_ack(md->node->picbase, node 542 arch/mips/pci/msi-xlp.c nlm_pic_ack(md->node->picbase, PIC_IRT_PCIE_LINK_INDEX(link)); node 545 arch/mips/pci/msi-xlp.c void nlm_dispatch_msix(int node, int lirq) node 552 arch/mips/pci/msi-xlp.c irqbase = nlm_irq_to_xirq(node, nlm_link_msixirq(link, 0)); node 95 arch/mips/pci/pci-lantiq.c struct device_node *node = pdev->dev.of_node; node 114 arch/mips/pci/pci-lantiq.c bus_clk = of_get_property(node, "lantiq,bus-clock", NULL); node 120 arch/mips/pci/pci-lantiq.c if (of_find_property(node, "lantiq,external-clock", NULL)) node 126 arch/mips/pci/pci-lantiq.c reset_gpio = of_get_named_gpio(node, "gpio-reset", 0); node 150 arch/mips/pci/pci-lantiq.c req_mask = of_get_property(node, "req-mask", NULL); node 131 arch/mips/pci/pci-legacy.c list_for_each_entry(child, &bus->children, node) node 138 arch/mips/pci/pci-legacy.c void pci_load_of_ranges(struct pci_controller *hose, struct device_node *node) node 143 arch/mips/pci/pci-legacy.c pr_info("PCI host bridge %pOF ranges:\n", node); node 144 arch/mips/pci/pci-legacy.c hose->of_node = node; node 146 arch/mips/pci/pci-legacy.c if (of_pci_range_parser_init(&parser, node)) node 170 arch/mips/pci/pci-legacy.c of_pci_range_to_resource(&range, node, res); node 211 arch/mips/pci/pci-xlp.c int lnkfunc, node; node 222 arch/mips/pci/pci-xlp.c node = xlp_socdev_to_node(lnkdev); node 224 arch/mips/pci/pci-xlp.c return nlm_irq_to_xirq(node, PIC_PCIE_LINK_LEGACY_IRQ(lnkfunc)); node 239 arch/mips/pci/pci-xlp.c static void xlp_config_pci_bswap(int node, int link) node 244 arch/mips/pci/pci-xlp.c nbubase = nlm_get_bridge_regbase(node); node 245 arch/mips/pci/pci-xlp.c lnkbase = nlm_get_pcie_base(node, link); node 286 arch/mips/pci/pci-xlp.c static inline void xlp_config_pci_bswap(int node, int link) {} node 26 arch/mips/pic32/pic32mzda/time.c static struct device_node *node; node 29 arch/mips/pic32/pic32mzda/time.c node = of_find_matching_node(NULL, pic32_infra_match); node 31 arch/mips/pic32/pic32mzda/time.c if (WARN_ON(!node)) node 34 arch/mips/pic32/pic32mzda/time.c irq = irq_of_parse_and_map(node, 0); node 32 arch/mips/ralink/common.h __iomem void *plat_of_remap_node(const char *node); node 146 arch/mips/ralink/irq.c static int __init intc_of_init(struct device_node *node, node 153 arch/mips/ralink/irq.c if (!of_property_read_u32_array(node, "ralink,intc-registers", node 157 arch/mips/ralink/irq.c irq = irq_of_parse_and_map(node, 0); node 161 arch/mips/ralink/irq.c if (of_address_to_resource(node, 0, &res)) node 179 arch/mips/ralink/irq.c domain = irq_domain_add_legacy(node, RALINK_INTC_IRQ_COUNT, node 29 arch/mips/ralink/of.c __iomem void *plat_of_remap_node(const char *node) node 34 arch/mips/ralink/of.c np = of_find_compatible_node(NULL, NULL, node); node 36 arch/mips/ralink/of.c panic("Failed to find %s node", node); node 39 arch/mips/ralink/of.c panic("Failed to get resource for %s", node); node 44 arch/mips/ralink/of.c panic("Failed to request resources for %s", node); node 56 arch/mips/ralink/of.c static int __init early_init_dt_find_memory(unsigned long node, node 151 arch/mips/sgi-ip27/ip27-irq.c desc->irq_common_data.node = info->nasid; node 264 arch/mips/sgi-ip27/ip27-memory.c static unsigned long __init slot_psize_compute(cnodeid_t node, int slot) node 271 arch/mips/sgi-ip27/ip27-memory.c nasid = COMPACT_TO_NASID_NODEID(node); node 353 arch/mips/sgi-ip27/ip27-memory.c cnodeid_t node; node 355 arch/mips/sgi-ip27/ip27-memory.c for_each_online_node(node) { node 358 arch/mips/sgi-ip27/ip27-memory.c slot_psize = slot_psize_compute(node, slot); node 373 arch/mips/sgi-ip27/ip27-memory.c slot, node); node 377 arch/mips/sgi-ip27/ip27-memory.c memblock_add_node(PFN_PHYS(slot_getbasepfn(node, slot)), node 378 arch/mips/sgi-ip27/ip27-memory.c PFN_PHYS(slot_psize), node); node 383 arch/mips/sgi-ip27/ip27-memory.c static void __init node_mem_init(cnodeid_t node) node 385 arch/mips/sgi-ip27/ip27-memory.c unsigned long slot_firstpfn = slot_getbasepfn(node, 0); node 386 arch/mips/sgi-ip27/ip27-memory.c unsigned long slot_freepfn = node_getfirstfree(node); node 389 arch/mips/sgi-ip27/ip27-memory.c get_pfn_range_for_nid(node, &start_pfn, &end_pfn); node 394 arch/mips/sgi-ip27/ip27-memory.c __node_data[node] = __va(slot_freepfn << PAGE_SHIFT); node 395 arch/mips/sgi-ip27/ip27-memory.c memset(__node_data[node], 0, PAGE_SIZE); node 397 arch/mips/sgi-ip27/ip27-memory.c NODE_DATA(node)->node_start_pfn = start_pfn; node 398 arch/mips/sgi-ip27/ip27-memory.c NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; node 400 arch/mips/sgi-ip27/ip27-memory.c cpumask_clear(&hub_data(node)->h_cpus); node 405 arch/mips/sgi-ip27/ip27-memory.c free_bootmem_with_active_regions(node, end_pfn); node 410 arch/mips/sgi-ip27/ip27-memory.c sparse_memory_present_with_active_regions(node); node 430 arch/mips/sgi-ip27/ip27-memory.c cnodeid_t node; node 436 arch/mips/sgi-ip27/ip27-memory.c for (node = 0; node < MAX_COMPACT_NODES; node++) { node 437 arch/mips/sgi-ip27/ip27-memory.c if (node_online(node)) { node 438 arch/mips/sgi-ip27/ip27-memory.c node_mem_init(node); node 441 arch/mips/sgi-ip27/ip27-memory.c __node_data[node] = &null_node; node 214 arch/mips/sgi-ip27/ip27-nmi.c for_each_online_node(node) node 215 arch/mips/sgi-ip27/ip27-nmi.c if (NODEPDA(node)->dump_count == 0) node 217 arch/mips/sgi-ip27/ip27-nmi.c if (node == MAX_NUMNODES) node 220 arch/mips/sgi-ip27/ip27-nmi.c for_each_online_node(node) node 221 arch/mips/sgi-ip27/ip27-nmi.c if (NODEPDA(node)->dump_count == 0) { node 222 arch/mips/sgi-ip27/ip27-nmi.c cpu = cpumask_first(cpumask_of_node(node)); node 223 arch/mips/sgi-ip27/ip27-nmi.c for (n=0; n < CNODE_NUM_CPUS(node); cpu++, n++) { node 36 arch/mips/sgi-ip27/ip27-smp.c cnodeid_t node = get_cpu_cnode(cpu); node 37 arch/mips/sgi-ip27/ip27-smp.c nasid_t nasid = COMPACT_TO_NASID_NODEID(node); node 40 arch/mips/sgi-ip27/ip27-smp.c sn_cpu_info[cpunum].p_nodeid = node; node 1132 arch/nds32/kernel/perf_event_cpu.c struct device_node *node = pdev->dev.of_node; node 1146 arch/nds32/kernel/perf_event_cpu.c if (node && of_id) { node 65 arch/nios2/kernel/irq.c struct device_node *node; node 67 arch/nios2/kernel/irq.c node = of_find_compatible_node(NULL, NULL, "altr,nios2-1.0"); node 68 arch/nios2/kernel/irq.c if (!node) node 69 arch/nios2/kernel/irq.c node = of_find_compatible_node(NULL, NULL, "altr,nios2-1.1"); node 71 arch/nios2/kernel/irq.c BUG_ON(!node); node 73 arch/nios2/kernel/irq.c domain = irq_domain_add_linear(node, NIOS2_CPU_NR_IRQS, &irq_ops, NULL); node 77 arch/nios2/kernel/irq.c of_node_put(node); node 22 arch/powerpc/boot/cuboot-8xx.c void *node; node 28 arch/powerpc/boot/cuboot-8xx.c node = finddevice("/soc/cpm"); node 29 arch/powerpc/boot/cuboot-8xx.c if (node) node 30 arch/powerpc/boot/cuboot-8xx.c setprop(node, "clock-frequency", &bd.bi_busfreq, 4); node 32 arch/powerpc/boot/cuboot-8xx.c node = finddevice("/soc/cpm/brg"); node 33 arch/powerpc/boot/cuboot-8xx.c if (node) node 34 arch/powerpc/boot/cuboot-8xx.c setprop(node, "clock-frequency", &bd.bi_busfreq, 4); node 127 arch/powerpc/boot/cuboot-pq2.c void *node, *parent_node; node 130 arch/powerpc/boot/cuboot-pq2.c node = finddevice("/pci"); node 131 arch/powerpc/boot/cuboot-pq2.c if (!node || !dt_is_compatible(node, "fsl,pq2-pci")) node 135 arch/powerpc/boot/cuboot-pq2.c if (!dt_xlate_reg(node, i, node 143 arch/powerpc/boot/cuboot-pq2.c dt_get_reg_format(node, &naddr, &nsize); node 147 arch/powerpc/boot/cuboot-pq2.c parent_node = get_parent(node); node 155 arch/powerpc/boot/cuboot-pq2.c len = getprop(node, "ranges", pci_ranges_buf, node 244 arch/powerpc/boot/cuboot-pq2.c void *node; node 250 arch/powerpc/boot/cuboot-pq2.c node = finddevice("/soc/cpm"); node 251 arch/powerpc/boot/cuboot-pq2.c if (node) node 252 arch/powerpc/boot/cuboot-pq2.c setprop(node, "clock-frequency", &bd.bi_cpmfreq, 4); node 254 arch/powerpc/boot/cuboot-pq2.c node = finddevice("/soc/cpm/brg"); node 255 arch/powerpc/boot/cuboot-pq2.c if (node) node 256 arch/powerpc/boot/cuboot-pq2.c setprop(node, "clock-frequency", &bd.bi_brgfreq, 4); node 132 arch/powerpc/boot/devtree.c void dt_get_reg_format(void *node, u32 *naddr, u32 *nsize) node 134 arch/powerpc/boot/devtree.c if (getprop(node, "#address-cells", naddr, 4) != 4) node 136 arch/powerpc/boot/devtree.c if (getprop(node, "#size-cells", nsize, 4) != 4) node 228 arch/powerpc/boot/devtree.c static int dt_xlate(void *node, int res, int reglen, unsigned long *addr, node 238 arch/powerpc/boot/devtree.c parent = get_parent(node); node 264 arch/powerpc/boot/devtree.c node = parent; node 266 arch/powerpc/boot/devtree.c parent = get_parent(node); node 272 arch/powerpc/boot/devtree.c buflen = getprop(node, "ranges", prop_buf, node 313 arch/powerpc/boot/devtree.c int dt_xlate_reg(void *node, int res, unsigned long *addr, unsigned long *size) node 317 arch/powerpc/boot/devtree.c reglen = getprop(node, "reg", prop_buf, sizeof(prop_buf)) / 4; node 318 arch/powerpc/boot/devtree.c return dt_xlate(node, res, reglen, addr, size); node 321 arch/powerpc/boot/devtree.c int dt_xlate_addr(void *node, u32 *buf, int buflen, unsigned long *xlated_addr) node 328 arch/powerpc/boot/devtree.c return dt_xlate(node, 0, buflen / 4, xlated_addr, NULL); node 331 arch/powerpc/boot/devtree.c int dt_is_compatible(void *node, const char *compat) node 336 arch/powerpc/boot/devtree.c len = getprop(node, "compatible", buf, MAX_PROP_LEN); node 350 arch/powerpc/boot/devtree.c int dt_get_virtual_reg(void *node, void **addr, int nres) node 355 arch/powerpc/boot/devtree.c n = getprop(node, "virtual-reg", addr, nres * 4); node 360 arch/powerpc/boot/devtree.c if (!dt_xlate_reg(node, n, &xaddr, NULL)) node 57 arch/powerpc/boot/mpc8xx.c void *node; node 61 arch/powerpc/boot/mpc8xx.c node = finddevice("/soc/cpm"); node 62 arch/powerpc/boot/mpc8xx.c if (node) node 63 arch/powerpc/boot/mpc8xx.c setprop(node, "clock-frequency", &sysclk, 4); node 65 arch/powerpc/boot/mpc8xx.c node = finddevice("/soc/cpm/brg"); node 66 arch/powerpc/boot/mpc8xx.c if (node) node 67 arch/powerpc/boot/mpc8xx.c setprop(node, "clock-frequency", &sysclk, 4); node 96 arch/powerpc/boot/ops.h int dt_xlate_reg(void *node, int res, unsigned long *addr, unsigned long *size); node 97 arch/powerpc/boot/ops.h int dt_xlate_addr(void *node, u32 *buf, int buflen, unsigned long *xlated_addr); node 98 arch/powerpc/boot/ops.h int dt_is_compatible(void *node, const char *compat); node 99 arch/powerpc/boot/ops.h void dt_get_reg_format(void *node, u32 *naddr, u32 *nsize); node 100 arch/powerpc/boot/ops.h int dt_get_virtual_reg(void *node, void **addr, int nres); node 108 arch/powerpc/boot/planetcore.c void *node, *chosen; node 114 arch/powerpc/boot/planetcore.c node = find_node_by_prop_value_str(NULL, "linux,planetcore-label", node 116 arch/powerpc/boot/planetcore.c if (!node) node 119 arch/powerpc/boot/planetcore.c path = get_path(node, prop_buf, MAX_PROP_LEN); node 77 arch/powerpc/boot/pq2.c void *node; node 81 arch/powerpc/boot/pq2.c node = finddevice("/soc/cpm"); node 82 arch/powerpc/boot/pq2.c if (node) node 83 arch/powerpc/boot/pq2.c setprop(node, "clock-frequency", &sysfreq, 4); node 85 arch/powerpc/boot/pq2.c node = finddevice("/soc/cpm/brg"); node 86 arch/powerpc/boot/pq2.c if (node) node 87 arch/powerpc/boot/pq2.c setprop(node, "clock-frequency", &brgfreq, 4); node 24 arch/powerpc/boot/redboot-83xx.c void *node; node 30 arch/powerpc/boot/redboot-83xx.c node = finddevice("/soc/cpm/brg"); node 31 arch/powerpc/boot/redboot-83xx.c if (node) { node 34 arch/powerpc/boot/redboot-83xx.c setprop(node, "clock-frequency", &bd.bi_busfreq, 4); node 23 arch/powerpc/boot/redboot-8xx.c void *node; node 29 arch/powerpc/boot/redboot-8xx.c node = finddevice("/soc/cpm/brg"); node 30 arch/powerpc/boot/redboot-8xx.c if (node) { node 33 arch/powerpc/boot/redboot-8xx.c setprop(node, "clock-frequency", &bd.bi_busfreq, 4); node 30 arch/powerpc/boot/simpleboot.c int node, size, i; node 37 arch/powerpc/boot/simpleboot.c node = fdt_path_offset(_dtb_start, "/"); node 38 arch/powerpc/boot/simpleboot.c if (node < 0) node 40 arch/powerpc/boot/simpleboot.c na = fdt_getprop(_dtb_start, node, "#address-cells", &size); node 43 arch/powerpc/boot/simpleboot.c ns = fdt_getprop(_dtb_start, node, "#size-cells", &size); node 48 arch/powerpc/boot/simpleboot.c node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type", node 50 arch/powerpc/boot/simpleboot.c if (node < 0) node 52 arch/powerpc/boot/simpleboot.c reg = fdt_getprop(_dtb_start, node, "reg", &size); node 69 arch/powerpc/boot/simpleboot.c node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type", node 71 arch/powerpc/boot/simpleboot.c if (!node) node 73 arch/powerpc/boot/simpleboot.c timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size); node 97 arch/powerpc/boot/treeboot-akebono.c int node, size; node 147 arch/powerpc/boot/treeboot-akebono.c node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type", node 149 arch/powerpc/boot/treeboot-akebono.c if (!node) node 151 arch/powerpc/boot/treeboot-akebono.c timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size); node 84 arch/powerpc/boot/treeboot-currituck.c int node, size; node 103 arch/powerpc/boot/treeboot-currituck.c node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type", node 105 arch/powerpc/boot/treeboot-currituck.c if (!node) node 107 arch/powerpc/boot/treeboot-currituck.c timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size); node 87 arch/powerpc/include/asm/cell-pmu.h extern void cbe_sync_irq(int node); node 321 arch/powerpc/include/asm/cell-regs.h extern u32 cbe_node_to_cpu(int node); node 97 arch/powerpc/include/asm/drmem.h void __init walk_drmem_lmbs_early(unsigned long node, node 156 arch/powerpc/include/asm/fadump-internal.h extern void rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node); node 159 arch/powerpc/include/asm/fadump-internal.h rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) { } node 163 arch/powerpc/include/asm/fadump-internal.h extern void opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node); node 166 arch/powerpc/include/asm/fadump-internal.h opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) { } node 31 arch/powerpc/include/asm/fadump.h extern int early_init_dt_scan_fw_dump(unsigned long node, const char *uname, node 32 arch/powerpc/include/asm/hvcserver.h struct list_head node; node 8 arch/powerpc/include/asm/i8259.h extern void i8259_init(struct device_node *node, unsigned long intack_addr); node 265 arch/powerpc/include/asm/iommu.h unsigned long mask, gfp_t flag, int node); node 72 arch/powerpc/include/asm/ipic.h extern struct ipic * ipic_init(struct device_node *node, unsigned int flags); node 100 arch/powerpc/include/asm/kvm_book3s_64.h #define for_each_nest_rmap_safe(pos, node, rmapp) \ node 101 arch/powerpc/include/asm/kvm_book3s_64.h for ((pos) = llist_entry((node), typeof(*(pos)), list); \ node 102 arch/powerpc/include/asm/kvm_book3s_64.h (node) && \ node 103 arch/powerpc/include/asm/kvm_book3s_64.h (*(rmapp) = ((RMAP_NESTED_IS_SINGLE_ENTRY & ((u64) (node))) ? \ node 104 arch/powerpc/include/asm/kvm_book3s_64.h ((u64) (node)) : ((pos)->rmap))) && \ node 105 arch/powerpc/include/asm/kvm_book3s_64.h (((node) = ((RMAP_NESTED_IS_SINGLE_ENTRY & ((u64) (node))) ? \ node 108 arch/powerpc/include/asm/kvm_book3s_64.h (pos) = llist_entry((node), typeof(*(pos)), list)) node 277 arch/powerpc/include/asm/mpc52xx.h extern unsigned int mpc52xx_get_xtal_freq(struct device_node *node); node 335 arch/powerpc/include/asm/mpc52xx.h extern int __init mpc52xx_add_bridge(struct device_node *node); node 14 arch/powerpc/include/asm/mpc5xxx.h extern unsigned long mpc5xxx_get_bus_frequency(struct device_node *node); node 259 arch/powerpc/include/asm/mpic.h struct device_node *node; node 427 arch/powerpc/include/asm/mpic.h extern struct mpic *mpic_alloc(struct device_node *node, node 309 arch/powerpc/include/asm/opal.h extern int early_init_dt_scan_opal(unsigned long node, const char *uname, node 311 arch/powerpc/include/asm/opal.h extern int early_init_dt_scan_recoverable_ranges(unsigned long node, node 56 arch/powerpc/include/asm/pci-bridge.h int node; node 171 arch/powerpc/include/asm/pci-bridge.h extern int pci_device_from_OF_node(struct device_node *node, node 264 arch/powerpc/include/asm/pci-bridge.h #define PHB_SET_NODE(PHB, NODE) ((PHB)->node = (NODE)) node 266 arch/powerpc/include/asm/pci-bridge.h #define PHB_SET_NODE(PHB, NODE) ((PHB)->node = NUMA_NO_NODE) node 273 arch/powerpc/include/asm/pci-bridge.h struct device_node* node); node 99 arch/powerpc/include/asm/pci.h extern struct pci_dev *of_create_pci_dev(struct device_node *node, node 106 arch/powerpc/include/asm/pci.h extern void of_scan_bus(struct device_node *node, struct pci_bus *bus); node 107 arch/powerpc/include/asm/pci.h extern void of_rescan_bus(struct device_node *node, struct pci_bus *bus); node 146 arch/powerpc/include/asm/pmac_feature.h static inline long pmac_call_feature(int selector, struct device_node* node, node 151 arch/powerpc/include/asm/pmac_feature.h return ppc_md.feature_call(selector, node, param, value); node 55 arch/powerpc/include/asm/pmac_low_i2c.h extern struct pmac_i2c_bus *pmac_i2c_find_bus(struct device_node *node); node 129 arch/powerpc/include/asm/pmac_pfunc.h struct device_node *node; node 42 arch/powerpc/include/asm/pmi.h struct list_head node; node 513 arch/powerpc/include/asm/ps3.h void ps3_sync_irq(int node); node 383 arch/powerpc/include/asm/rtas.h extern int early_init_dt_scan_rtas(unsigned long node, node 120 arch/powerpc/include/asm/spu.h u32 node; node 20 arch/powerpc/include/asm/topology.h #define cpumask_of_node(node) ((node) == -1 ? \ node 22 arch/powerpc/include/asm/topology.h node_to_cpumask_map[node]) node 48 arch/powerpc/include/asm/topology.h static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) node 50 arch/powerpc/include/asm/topology.h numa_cpu_lookup_table[cpu] = node; node 86 arch/powerpc/include/asm/topology.h static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) {} node 26 arch/powerpc/include/asm/tsi108_pci.h extern void tsi108_pci_int_init(struct device_node *node); node 15 arch/powerpc/include/asm/ultravisor.h int early_init_dt_scan_ultravisor(unsigned long node, const char *uname, node 91 arch/powerpc/include/asm/xics.h int (*host_match)(struct ics *ics, struct device_node *node); node 317 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_lookup_by_node(const struct device_node *node) node 323 arch/powerpc/kernel/cacheinfo.c if (iter->ofnode != node) node 353 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode_unified(struct device_node *node, int level) node 355 arch/powerpc/kernel/cacheinfo.c pr_debug("creating L%d ucache for %pOF\n", level, node); node 357 arch/powerpc/kernel/cacheinfo.c return new_cache(cache_is_unified_d(node), level, node); node 360 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode_split(struct device_node *node, node 366 arch/powerpc/kernel/cacheinfo.c node); node 368 arch/powerpc/kernel/cacheinfo.c dcache = new_cache(CACHE_TYPE_DATA, level, node); node 369 arch/powerpc/kernel/cacheinfo.c icache = new_cache(CACHE_TYPE_INSTRUCTION, level, node); node 383 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode(struct device_node *node, int level) node 387 arch/powerpc/kernel/cacheinfo.c if (cache_node_is_unified(node)) node 388 arch/powerpc/kernel/cacheinfo.c cache = cache_do_one_devnode_unified(node, level); node 390 arch/powerpc/kernel/cacheinfo.c cache = cache_do_one_devnode_split(node, level); node 395 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_lookup_or_instantiate(struct device_node *node, node 400 arch/powerpc/kernel/cacheinfo.c cache = cache_lookup_by_node(node); node 407 arch/powerpc/kernel/cacheinfo.c cache = cache_do_one_devnode(node, level); node 51 arch/powerpc/kernel/dt_cpu_ftrs.c unsigned long node; node 795 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init fdt_find_cpu_features(unsigned long node, const char *uname, node 798 arch/powerpc/kernel/dt_cpu_ftrs.c if (of_flat_dt_is_compatible(node, "ibm,powerpc-cpu-features") node 799 arch/powerpc/kernel/dt_cpu_ftrs.c && of_get_flat_dt_prop(node, "isa", NULL)) node 833 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init process_cpufeatures_node(unsigned long node, node 842 arch/powerpc/kernel/dt_cpu_ftrs.c f->node = node; node 846 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "isa", &len); node 853 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "usable-privilege", &len); node 860 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "hv-support", &len); node 866 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "os-support", &len); node 872 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "hfscr-bit-nr", &len); node 877 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "fscr-bit-nr", &len); node 882 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "hwcap-bit-nr", &len); node 939 arch/powerpc/kernel/dt_cpu_ftrs.c if (!of_get_flat_dt_prop(node, "dependencies", &len)) { node 959 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(f->node, "dependencies", &len); node 974 arch/powerpc/kernel/dt_cpu_ftrs.c if (of_get_flat_dt_phandle(d->node) == phandle) { node 990 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init scan_cpufeatures_subnodes(unsigned long node, node 996 arch/powerpc/kernel/dt_cpu_ftrs.c process_cpufeatures_node(node, uname, *count); node 1003 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init count_cpufeatures_subnodes(unsigned long node, node 1014 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init dt_cpu_ftrs_scan_callback(unsigned long node, const char node 1022 arch/powerpc/kernel/dt_cpu_ftrs.c if (!of_flat_dt_is_compatible(node, "ibm,powerpc-cpu-features")) node 1025 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "isa", NULL); node 1033 arch/powerpc/kernel/dt_cpu_ftrs.c of_scan_flat_dt_subnodes(node, count_cpufeatures_subnodes, node 1046 arch/powerpc/kernel/dt_cpu_ftrs.c of_scan_flat_dt_subnodes(node, scan_cpufeatures_subnodes, &count); node 1055 arch/powerpc/kernel/dt_cpu_ftrs.c prop = of_get_flat_dt_prop(node, "display-name", NULL); node 23 arch/powerpc/kernel/epapr_paravirt.c static int __init early_init_dt_scan_epapr(unsigned long node, node 31 arch/powerpc/kernel/epapr_paravirt.c insts = of_get_flat_dt_prop(node, "hcall-instructions", &len); node 47 arch/powerpc/kernel/epapr_paravirt.c if (of_get_flat_dt_prop(node, "has-idle", NULL)) node 108 arch/powerpc/kernel/fadump.c int __init early_init_dt_scan_fw_dump(unsigned long node, const char *uname, node 115 arch/powerpc/kernel/fadump.c rtas_fadump_dt_scan(&fw_dump, node); node 120 arch/powerpc/kernel/fadump.c opal_fadump_dt_scan(&fw_dump, node); node 1503 arch/powerpc/kernel/fadump.c int __init early_init_dt_scan_fw_dump(unsigned long node, const char *uname, node 1509 arch/powerpc/kernel/fadump.c opal_fadump_dt_scan(&fw_dump, node); node 854 arch/powerpc/kernel/iommu.c unsigned long mask, gfp_t flag, int node) node 880 arch/powerpc/kernel/iommu.c page = alloc_pages_node(node, flag, order); node 235 arch/powerpc/kernel/machine_kexec.c static void __init export_crashk_values(struct device_node *node) node 239 arch/powerpc/kernel/machine_kexec.c of_remove_property(node, of_find_property(node, node 241 arch/powerpc/kernel/machine_kexec.c of_remove_property(node, of_find_property(node, node 246 arch/powerpc/kernel/machine_kexec.c of_add_property(node, &crashk_base_prop); node 248 arch/powerpc/kernel/machine_kexec.c of_add_property(node, &crashk_size_prop); node 256 arch/powerpc/kernel/machine_kexec.c of_update_property(node, &memory_limit_prop); node 261 arch/powerpc/kernel/machine_kexec.c struct device_node *node; node 263 arch/powerpc/kernel/machine_kexec.c node = of_find_node_by_path("/chosen"); node 264 arch/powerpc/kernel/machine_kexec.c if (!node) node 268 arch/powerpc/kernel/machine_kexec.c of_remove_property(node, of_find_property(node, kernel_end_prop.name, NULL)); node 272 arch/powerpc/kernel/machine_kexec.c of_add_property(node, &kernel_end_prop); node 274 arch/powerpc/kernel/machine_kexec.c export_crashk_values(node); node 276 arch/powerpc/kernel/machine_kexec.c of_node_put(node); node 40 arch/powerpc/kernel/machine_kexec_64.c struct device_node *node; node 54 arch/powerpc/kernel/machine_kexec_64.c for_each_node_by_type(node, "pci") { node 55 arch/powerpc/kernel/machine_kexec_64.c basep = of_get_property(node, "linux,tce-base", NULL); node 56 arch/powerpc/kernel/machine_kexec_64.c sizep = of_get_property(node, "linux,tce-size", NULL); node 394 arch/powerpc/kernel/machine_kexec_64.c struct device_node *node; node 400 arch/powerpc/kernel/machine_kexec_64.c node = of_find_node_by_path("/chosen"); node 401 arch/powerpc/kernel/machine_kexec_64.c if (!node) node 405 arch/powerpc/kernel/machine_kexec_64.c of_remove_property(node, of_find_property(node, htab_base_prop.name, NULL)); node 406 arch/powerpc/kernel/machine_kexec_64.c of_remove_property(node, of_find_property(node, htab_size_prop.name, NULL)); node 409 arch/powerpc/kernel/machine_kexec_64.c of_add_property(node, &htab_base_prop); node 411 arch/powerpc/kernel/machine_kexec_64.c of_add_property(node, &htab_size_prop); node 413 arch/powerpc/kernel/machine_kexec_64.c of_node_put(node); node 339 arch/powerpc/kernel/pci-common.c struct pci_controller* pci_find_hose_for_OF_device(struct device_node* node) node 341 arch/powerpc/kernel/pci-common.c while(node) { node 344 arch/powerpc/kernel/pci-common.c if (hose->dn == node) node 346 arch/powerpc/kernel/pci-common.c node = node->parent; node 1226 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &bus->children, node) node 1359 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &pci_root_buses, node) node 1371 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &pci_root_buses, node) node 1413 arch/powerpc/kernel/pci-common.c list_for_each_entry(child_bus, &bus->children, node) node 1596 arch/powerpc/kernel/pci-common.c struct device_node *node = hose->dn; node 1599 arch/powerpc/kernel/pci-common.c pr_debug("PCI: Scanning PHB %pOF\n", node); node 1625 arch/powerpc/kernel/pci-common.c if (node && hose->controller_ops.probe_mode) node 1629 arch/powerpc/kernel/pci-common.c of_scan_bus(node, bus); node 1646 arch/powerpc/kernel/pci-common.c list_for_each_entry(child, &bus->children, node) node 29 arch/powerpc/kernel/pci-hotplug.c list_for_each_entry(tmp, &bus->children, node) { node 85 arch/powerpc/kernel/pci-hotplug.c list_for_each_entry(child_bus, &bus->children, node) node 75 arch/powerpc/kernel/pci_32.c make_one_node_map(struct device_node* node, u8 pci_bus) node 82 arch/powerpc/kernel/pci_32.c bus_range = of_get_property(node, "bus-range", &len); node 85 arch/powerpc/kernel/pci_32.c "assuming it starts at 0\n", node); node 90 arch/powerpc/kernel/pci_32.c for_each_child_of_node(node, node) { node 94 arch/powerpc/kernel/pci_32.c class_code = of_get_property(node, "class-code", NULL); node 98 arch/powerpc/kernel/pci_32.c reg = of_get_property(node, "reg", NULL); node 107 arch/powerpc/kernel/pci_32.c make_one_node_map(node, dev->subordinate->number); node 134 arch/powerpc/kernel/pci_32.c struct device_node* node = hose->dn; node 136 arch/powerpc/kernel/pci_32.c if (!node) node 138 arch/powerpc/kernel/pci_32.c make_one_node_map(node, hose->first_busno); node 161 arch/powerpc/kernel/pci_32.c int pci_device_from_OF_node(struct device_node *node, u8 *bus, u8 *devfn) node 168 arch/powerpc/kernel/pci_32.c if (!pci_find_hose_for_OF_device(node)) node 171 arch/powerpc/kernel/pci_32.c reg = of_get_property(node, "reg", &size); node 233 arch/powerpc/kernel/pci_64.c list_for_each_entry(tmp_bus, &pci_root_buses, node) { node 266 arch/powerpc/kernel/pci_64.c return phb->node; node 117 arch/powerpc/kernel/pci_of_scan.c static void of_pci_parse_addrs(struct device_node *node, struct pci_dev *dev) node 128 arch/powerpc/kernel/pci_of_scan.c addrs = of_get_property(node, "assigned-addresses", &proplen); node 130 arch/powerpc/kernel/pci_of_scan.c addrs = of_get_property(node, "reg", &proplen); node 175 arch/powerpc/kernel/pci_of_scan.c struct pci_dev *of_create_pci_dev(struct device_node *node, node 185 arch/powerpc/kernel/pci_of_scan.c of_node_get_device_type(node)); node 187 arch/powerpc/kernel/pci_of_scan.c dev->dev.of_node = of_node_get(node); node 196 arch/powerpc/kernel/pci_of_scan.c dev->vendor = get_int_prop(node, "vendor-id", 0xffff); node 197 arch/powerpc/kernel/pci_of_scan.c dev->device = get_int_prop(node, "device-id", 0xffff); node 198 arch/powerpc/kernel/pci_of_scan.c dev->subsystem_vendor = get_int_prop(node, "subsystem-vendor-id", 0); node 199 arch/powerpc/kernel/pci_of_scan.c dev->subsystem_device = get_int_prop(node, "subsystem-id", 0); node 205 arch/powerpc/kernel/pci_of_scan.c dev->class = get_int_prop(node, "class-code", 0); node 206 arch/powerpc/kernel/pci_of_scan.c dev->revision = get_int_prop(node, "revision-id", 0); node 218 arch/powerpc/kernel/pci_of_scan.c if (of_node_is_type(node, "pci") || of_node_is_type(node, "pciex")) { node 223 arch/powerpc/kernel/pci_of_scan.c } else if (of_node_is_type(node, "cardbus")) { node 232 arch/powerpc/kernel/pci_of_scan.c of_pci_parse_addrs(node, dev); node 252 arch/powerpc/kernel/pci_of_scan.c struct device_node *node = dev->dev.of_node; node 262 arch/powerpc/kernel/pci_of_scan.c pr_debug("of_scan_pci_bridge(%pOF)\n", node); node 265 arch/powerpc/kernel/pci_of_scan.c busrange = of_get_property(node, "bus-range", &len); node 268 arch/powerpc/kernel/pci_of_scan.c node); node 271 arch/powerpc/kernel/pci_of_scan.c ranges = of_get_property(node, "ranges", &len); node 274 arch/powerpc/kernel/pci_of_scan.c node); node 285 arch/powerpc/kernel/pci_of_scan.c node); node 313 arch/powerpc/kernel/pci_of_scan.c " for bridge %pOF\n", node); node 319 arch/powerpc/kernel/pci_of_scan.c " for bridge %pOF\n", node); node 342 arch/powerpc/kernel/pci_of_scan.c of_scan_bus(node, bus); node 395 arch/powerpc/kernel/pci_of_scan.c static void __of_scan_bus(struct device_node *node, struct pci_bus *bus, node 402 arch/powerpc/kernel/pci_of_scan.c node, bus->number); node 405 arch/powerpc/kernel/pci_of_scan.c for_each_child_of_node(node, child) { node 429 arch/powerpc/kernel/pci_of_scan.c void of_scan_bus(struct device_node *node, struct pci_bus *bus) node 431 arch/powerpc/kernel/pci_of_scan.c __of_scan_bus(node, bus, 0); node 443 arch/powerpc/kernel/pci_of_scan.c void of_rescan_bus(struct device_node *node, struct pci_bus *bus) node 445 arch/powerpc/kernel/pci_of_scan.c __of_scan_bus(node, bus, 1); node 180 arch/powerpc/kernel/prom.c static void __init scan_features(unsigned long node, const unsigned char *ftrs, node 219 arch/powerpc/kernel/prom.c static void __init check_cpu_pa_features(unsigned long node) node 224 arch/powerpc/kernel/prom.c pa_ftrs = of_get_flat_dt_prop(node, "ibm,pa-features", &tablelen); node 228 arch/powerpc/kernel/prom.c scan_features(node, pa_ftrs, tablelen, node 233 arch/powerpc/kernel/prom.c static void __init init_mmu_slb_size(unsigned long node) node 237 arch/powerpc/kernel/prom.c slb_size_ptr = of_get_flat_dt_prop(node, "slb-size", NULL) ? : node 238 arch/powerpc/kernel/prom.c of_get_flat_dt_prop(node, "ibm,slb-size", NULL); node 244 arch/powerpc/kernel/prom.c #define init_mmu_slb_size(node) do { } while(0) node 269 arch/powerpc/kernel/prom.c static inline void identical_pvr_fixup(unsigned long node) node 272 arch/powerpc/kernel/prom.c const char *model = of_get_flat_dt_prop(node, "model", NULL); node 288 arch/powerpc/kernel/prom.c #define identical_pvr_fixup(node) do { } while(0) node 291 arch/powerpc/kernel/prom.c static void __init check_cpu_feature_properties(unsigned long node) node 298 arch/powerpc/kernel/prom.c prop = of_get_flat_dt_prop(node, fp->name, NULL); node 306 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_cpus(unsigned long node, node 310 arch/powerpc/kernel/prom.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 323 arch/powerpc/kernel/prom.c intserv = of_get_flat_dt_prop(node, "ibm,ppc-interrupt-server#s", &len); node 325 arch/powerpc/kernel/prom.c intserv = of_get_flat_dt_prop(node, "reg", &len); node 373 arch/powerpc/kernel/prom.c prop = of_get_flat_dt_prop(node, "cpu-version", NULL); node 377 arch/powerpc/kernel/prom.c check_cpu_feature_properties(node); node 378 arch/powerpc/kernel/prom.c check_cpu_pa_features(node); node 381 arch/powerpc/kernel/prom.c identical_pvr_fixup(node); node 382 arch/powerpc/kernel/prom.c init_mmu_slb_size(node); node 396 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_chosen_ppc(unsigned long node, node 403 arch/powerpc/kernel/prom.c if (early_init_dt_scan_chosen(node, uname, depth, data) == 0) node 408 arch/powerpc/kernel/prom.c if (of_get_flat_dt_prop(node, "linux,iommu-off", NULL) != NULL) node 410 arch/powerpc/kernel/prom.c if (of_get_flat_dt_prop(node, "linux,iommu-force-on", NULL) != NULL) node 415 arch/powerpc/kernel/prom.c lprop = of_get_flat_dt_prop(node, "linux,memory-limit", NULL); node 420 arch/powerpc/kernel/prom.c lprop = of_get_flat_dt_prop(node, "linux,tce-alloc-start", NULL); node 423 arch/powerpc/kernel/prom.c lprop = of_get_flat_dt_prop(node, "linux,tce-alloc-end", NULL); node 429 arch/powerpc/kernel/prom.c lprop = of_get_flat_dt_prop(node, "linux,crashkernel-base", NULL); node 433 arch/powerpc/kernel/prom.c lprop = of_get_flat_dt_prop(node, "linux,crashkernel-size", NULL); node 525 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_memory_ppc(unsigned long node, node 532 arch/powerpc/kernel/prom.c walk_drmem_lmbs_early(node, early_init_drmem_lmb); node 537 arch/powerpc/kernel/prom.c return early_init_dt_scan_memory(node, uname, depth, data); node 618 arch/powerpc/kernel/prom_init.c phandle node; node 620 arch/powerpc/kernel/prom_init.c if ((node = *nodep) != 0 node 621 arch/powerpc/kernel/prom_init.c && (*nodep = call_prom("child", 1, 1, node)) != 0) node 623 arch/powerpc/kernel/prom_init.c if ((*nodep = call_prom("peer", 1, 1, node)) != 0) node 626 arch/powerpc/kernel/prom_init.c if ((node = call_prom("parent", 1, 1, node)) == 0) node 628 arch/powerpc/kernel/prom_init.c if ((*nodep = call_prom("peer", 1, 1, node)) != 0) node 633 arch/powerpc/kernel/prom_init.c static inline int __init prom_getprop(phandle node, const char *pname, node 636 arch/powerpc/kernel/prom_init.c return call_prom("getprop", 4, 1, node, ADDR(pname), node 640 arch/powerpc/kernel/prom_init.c static inline int __init prom_getproplen(phandle node, const char *pname) node 642 arch/powerpc/kernel/prom_init.c return call_prom("getproplen", 2, 1, node, ADDR(pname)); node 671 arch/powerpc/kernel/prom_init.c static int __init prom_setprop(phandle node, const char *nodename, node 677 arch/powerpc/kernel/prom_init.c return call_prom("setprop", 4, 1, node, ADDR(pname), node 1165 arch/powerpc/kernel/prom_init.c phandle node; node 1170 arch/powerpc/kernel/prom_init.c for (node = 0; prom_next_node(&node); ) { node 1172 arch/powerpc/kernel/prom_init.c prom_getprop(node, "device_type", type, sizeof(type)); node 1181 arch/powerpc/kernel/prom_init.c plen = prom_getproplen(node, "ibm,ppc-interrupt-server#s"); node 1594 arch/powerpc/kernel/prom_init.c phandle node; node 1617 arch/powerpc/kernel/prom_init.c for (node = 0; prom_next_node(&node); ) { node 1619 arch/powerpc/kernel/prom_init.c prom_getprop(node, "device_type", type, sizeof(type)); node 1626 arch/powerpc/kernel/prom_init.c prom_getprop(node, "name", type, sizeof(type)); node 1631 arch/powerpc/kernel/prom_init.c plen = prom_getprop(node, "reg", regbuf, sizeof(regbuf)); node 1641 arch/powerpc/kernel/prom_init.c call_prom("package-to-path", 3, 1, node, prom_scratch, node 1921 arch/powerpc/kernel/prom_init.c phandle node; node 1941 arch/powerpc/kernel/prom_init.c for (node = 0; prom_next_node(&node); ) { node 1945 arch/powerpc/kernel/prom_init.c prom_getprop(node, "compatible", node 1947 arch/powerpc/kernel/prom_init.c prom_getprop(node, "device_type", type, sizeof(type)); node 1948 arch/powerpc/kernel/prom_init.c prom_getprop(node, "model", model, sizeof(model)); node 1966 arch/powerpc/kernel/prom_init.c if (prom_getprop(node, "tce-table-minalign", &minalign, node 1969 arch/powerpc/kernel/prom_init.c if (prom_getprop(node, "tce-table-minsize", &minsize, node 1992 arch/powerpc/kernel/prom_init.c if (call_prom("package-to-path", 3, 1, node, node 1998 arch/powerpc/kernel/prom_init.c prom_setprop(node, path, "linux,tce-base", &base, sizeof(base)); node 1999 arch/powerpc/kernel/prom_init.c prom_setprop(node, path, "linux,tce-size", &minsize, sizeof(minsize)); node 2002 arch/powerpc/kernel/prom_init.c prom_debug("\tnode = 0x%x\n", node); node 2069 arch/powerpc/kernel/prom_init.c phandle node; node 2105 arch/powerpc/kernel/prom_init.c for (node = 0; prom_next_node(&node); ) { node 2110 arch/powerpc/kernel/prom_init.c prom_getprop(node, "device_type", type, sizeof(type)); node 2115 arch/powerpc/kernel/prom_init.c if (prom_getprop(node, "status", type, sizeof(type)) > 0) node 2120 arch/powerpc/kernel/prom_init.c prom_getprop(node, "reg", ®, sizeof(reg)); node 2134 arch/powerpc/kernel/prom_init.c call_prom("start-cpu", 3, 0, node, node 2321 arch/powerpc/kernel/prom_init.c phandle node; node 2346 arch/powerpc/kernel/prom_init.c for (node = 0; prom_next_node(&node); ) { node 2348 arch/powerpc/kernel/prom_init.c prom_getprop(node, "device_type", type, sizeof(type)); node 2360 arch/powerpc/kernel/prom_init.c if (call_prom("package-to-path", 3, 1, node, path, node 2373 arch/powerpc/kernel/prom_init.c prom_setprop(node, path, "linux,opened", NULL, 0); node 2392 arch/powerpc/kernel/prom_init.c if (prom_getprop(node, "linux,boot-display", NULL, 0) != node 2397 arch/powerpc/kernel/prom_init.c prom_getprop(node, "width", &width, 4); node 2398 arch/powerpc/kernel/prom_init.c prom_getprop(node, "height", &height, 4); node 2399 arch/powerpc/kernel/prom_init.c prom_getprop(node, "linebytes", &pitch, 4); node 2400 arch/powerpc/kernel/prom_init.c prom_getprop(node, "address", &addr, 4); node 2467 arch/powerpc/kernel/prom_init.c static void __init scan_dt_build_strings(phandle node, node 2482 arch/powerpc/kernel/prom_init.c if (call_prom("nextprop", 3, 1, node, prev_name, namep) != 1) { node 2508 arch/powerpc/kernel/prom_init.c child = call_prom("child", 1, 1, node); node 2515 arch/powerpc/kernel/prom_init.c static void __init scan_dt_build_struct(phandle node, unsigned long *mem_start, node 2532 arch/powerpc/kernel/prom_init.c l = call_prom("package-to-path", 3, 1, node, namep, room); node 2538 arch/powerpc/kernel/prom_init.c call_prom("package-to-path", 3, 1, node, namep, l); node 2559 arch/powerpc/kernel/prom_init.c call_prom("package-to-path", 3, 1, node, path, sizeof(prom_scratch) - 1); node 2565 arch/powerpc/kernel/prom_init.c if (call_prom("nextprop", 3, 1, node, prev_name, node 2585 arch/powerpc/kernel/prom_init.c l = call_prom("getproplen", 2, 1, node, pname); node 2598 arch/powerpc/kernel/prom_init.c call_prom("getprop", 4, 1, node, pname, valp, l); node 2615 arch/powerpc/kernel/prom_init.c *(__be32 *)valp = cpu_to_be32(node); node 2620 arch/powerpc/kernel/prom_init.c child = call_prom("child", 1, 1, node); node 2915 arch/powerpc/kernel/prom_init.c u32 node; node 2920 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/builtin/ethernet")); node 2921 arch/powerpc/kernel/prom_init.c if (!PHANDLE_VALID(node)) node 2925 arch/powerpc/kernel/prom_init.c rv = prom_getprop(node, "phy-handle", prop, sizeof(prop)); node 2935 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/builtin/mdio")); node 2936 arch/powerpc/kernel/prom_init.c if (!PHANDLE_VALID(node)) { node 2956 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, node 2958 arch/powerpc/kernel/prom_init.c if (!PHANDLE_VALID(node)) { node 2982 arch/powerpc/kernel/prom_init.c u32 node; node 2987 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/")); node 2988 arch/powerpc/kernel/prom_init.c if (!PHANDLE_VALID(node)) node 2991 arch/powerpc/kernel/prom_init.c rv = prom_getprop(node, "model", prop, sizeof(prop)); node 3000 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/")); node 3001 arch/powerpc/kernel/prom_init.c rv = prom_getprop(node, "device_type", prop, sizeof(prop)); node 3003 arch/powerpc/kernel/prom_init.c prom_setprop(node, "/", "device_type", "efika", sizeof("efika")); node 3007 arch/powerpc/kernel/prom_init.c rv = prom_getprop(node, "CODEGEN,description", prop, sizeof(prop)); node 3009 arch/powerpc/kernel/prom_init.c prom_setprop(node, "/", "CODEGEN,description", node 3014 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/builtin/bestcomm")); node 3015 arch/powerpc/kernel/prom_init.c if (PHANDLE_VALID(node)) { node 3016 arch/powerpc/kernel/prom_init.c len = prom_getproplen(node, "interrupts"); node 3019 arch/powerpc/kernel/prom_init.c prom_setprop(node, "/builtin/bestcom", "interrupts", node 3025 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR("/builtin/sound")); node 3026 arch/powerpc/kernel/prom_init.c if (PHANDLE_VALID(node)) { node 3027 arch/powerpc/kernel/prom_init.c rv = prom_getprop(node, "interrupts", prop, sizeof(prop)); node 3030 arch/powerpc/kernel/prom_init.c prom_setprop(node, "/builtin/sound", "interrupts", node 3053 arch/powerpc/kernel/prom_init.c phandle iob, node; node 3070 arch/powerpc/kernel/prom_init.c node = call_prom("finddevice", 1, 1, ADDR(pci_name)); node 3073 arch/powerpc/kernel/prom_init.c for( ; prom_next_node(&node); ) { node 3075 arch/powerpc/kernel/prom_init.c if (!PHANDLE_VALID(node)) node 3078 arch/powerpc/kernel/prom_init.c rval = prom_getproplen(node, "interrupts"); node 3082 arch/powerpc/kernel/prom_init.c prom_getprop(node, "interrupts", &interrupts, sizeof(interrupts)); node 3096 arch/powerpc/kernel/prom_init.c prom_setprop(node, pci_name, "interrupts", interrupts, node 3098 arch/powerpc/kernel/prom_init.c prom_setprop(node, pci_name, "interrupt-parent", &parent, node 1200 arch/powerpc/kernel/rtas.c int __init early_init_dt_scan_rtas(unsigned long node, node 1208 arch/powerpc/kernel/rtas.c basep = of_get_flat_dt_prop(node, "linux,rtas-base", NULL); node 1209 arch/powerpc/kernel/rtas.c entryp = of_get_flat_dt_prop(node, "linux,rtas-entry", NULL); node 1210 arch/powerpc/kernel/rtas.c sizep = of_get_flat_dt_prop(node, "rtas-size", NULL); node 1219 arch/powerpc/kernel/rtas.c basep = of_get_flat_dt_prop(node, "put-term-char", NULL); node 1223 arch/powerpc/kernel/rtas.c basep = of_get_flat_dt_prop(node, "get-term-char", NULL); node 1016 arch/powerpc/kernel/sysfs.c struct node *node = node_devices[nid]; node 1017 arch/powerpc/kernel/sysfs.c return sysfs_create_link(&node->dev.kobj, &dev->kobj, node 1024 arch/powerpc/kernel/sysfs.c struct node *node = node_devices[nid]; node 1025 arch/powerpc/kernel/sysfs.c sysfs_remove_link(&node->dev.kobj, kobject_name(&dev->kobj)); node 673 arch/powerpc/kernel/vdso.c unsigned long cpu, node, val; node 682 arch/powerpc/kernel/vdso.c node = cpu_to_node(cpu); node 683 arch/powerpc/kernel/vdso.c WARN_ON_ONCE(node > 0xffff); node 685 arch/powerpc/kernel/vdso.c val = (cpu & 0xfff) | ((node & 0xffff) << 16); node 5452 arch/powerpc/kvm/book3s_hv.c int node = cpu_to_node(first_cpu); node 5460 arch/powerpc/kvm/book3s_hv.c GFP_KERNEL, node); node 354 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_seg_sizes(unsigned long node, node 358 arch/powerpc/mm/book3s64/hash_utils.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 366 arch/powerpc/mm/book3s64/hash_utils.c prop = of_get_flat_dt_prop(node, "ibm,processor-segment-sizes", &size); node 410 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_page_sizes(unsigned long node, node 414 arch/powerpc/mm/book3s64/hash_utils.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 422 arch/powerpc/mm/book3s64/hash_utils.c prop = of_get_flat_dt_prop(node, "ibm,segment-page-sizes", &size); node 493 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_hugepage_blocks(unsigned long node, node 496 arch/powerpc/mm/book3s64/hash_utils.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 511 arch/powerpc/mm/book3s64/hash_utils.c page_count_prop = of_get_flat_dt_prop(node, "ibm,expected#pages", NULL); node 515 arch/powerpc/mm/book3s64/hash_utils.c addr_prop = of_get_flat_dt_prop(node, "reg", NULL); node 712 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_pftsize(unsigned long node, node 716 arch/powerpc/mm/book3s64/hash_utils.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 723 arch/powerpc/mm/book3s64/hash_utils.c prop = of_get_flat_dt_prop(node, "ibm,pft-size", NULL); node 423 arch/powerpc/mm/book3s64/radix_pgtable.c static int __init radix_dt_scan_page_sizes(unsigned long node, node 431 arch/powerpc/mm/book3s64/radix_pgtable.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 438 arch/powerpc/mm/book3s64/radix_pgtable.c prop = of_get_flat_dt_prop(node, "ibm,mmu-pid-bits", &size); node 443 arch/powerpc/mm/book3s64/radix_pgtable.c prop = of_get_flat_dt_prop(node, "ibm,processor-radix-AP-encodings", &size); node 268 arch/powerpc/mm/drmem.c void __init walk_drmem_lmbs_early(unsigned long node, node 274 arch/powerpc/mm/drmem.c prop = of_get_flat_dt_prop(node, "ibm,lmb-size", &len); node 280 arch/powerpc/mm/drmem.c usm = of_get_flat_dt_prop(node, "linux,drconf-usable-memory", &len); node 282 arch/powerpc/mm/drmem.c prop = of_get_flat_dt_prop(node, "ibm,dynamic-memory", &len); node 286 arch/powerpc/mm/drmem.c prop = of_get_flat_dt_prop(node, "ibm,dynamic-memory-v2", node 129 arch/powerpc/mm/init_64.c static __meminit struct vmemmap_backing * vmemmap_list_alloc(int node) node 143 arch/powerpc/mm/init_64.c next = vmemmap_alloc_block(PAGE_SIZE, node); node 158 arch/powerpc/mm/init_64.c int node) node 162 arch/powerpc/mm/init_64.c vmem_back = vmemmap_list_alloc(node); node 190 arch/powerpc/mm/init_64.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 198 arch/powerpc/mm/init_64.c pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node); node 218 arch/powerpc/mm/init_64.c p = vmemmap_alloc_block_buf(page_size, node); node 222 arch/powerpc/mm/init_64.c vmemmap_list_populate(__pa(p), start, node); node 71 arch/powerpc/mm/numa.c unsigned int node; node 78 arch/powerpc/mm/numa.c for_each_node(node) node 79 arch/powerpc/mm/numa.c alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]); node 140 arch/powerpc/mm/numa.c static void map_cpu_to_node(int cpu, int node) node 142 arch/powerpc/mm/numa.c update_numa_cpu_lookup_table(cpu, node); node 144 arch/powerpc/mm/numa.c dbg("adding cpu %d to node %d\n", cpu, node); node 146 arch/powerpc/mm/numa.c if (!(cpumask_test_cpu(cpu, node_to_cpumask_map[node]))) node 147 arch/powerpc/mm/numa.c cpumask_set_cpu(cpu, node_to_cpumask_map[node]); node 153 arch/powerpc/mm/numa.c int node = numa_cpu_lookup_table[cpu]; node 155 arch/powerpc/mm/numa.c dbg("removing cpu %lu from node %d\n", cpu, node); node 157 arch/powerpc/mm/numa.c if (cpumask_test_cpu(cpu, node_to_cpumask_map[node])) { node 158 arch/powerpc/mm/numa.c cpumask_clear_cpu(cpu, node_to_cpumask_map[node]); node 161 arch/powerpc/mm/numa.c cpu, node); node 505 arch/powerpc/mm/numa.c static void verify_cpu_node_mapping(int cpu, int node) node 518 arch/powerpc/mm/numa.c if (cpu_to_node(sibling) != node) { node 767 arch/powerpc/mm/numa.c unsigned int node; node 773 arch/powerpc/mm/numa.c for_each_online_node(node) { node 774 arch/powerpc/mm/numa.c pr_info("Node %d CPUs:", node); node 783 arch/powerpc/mm/numa.c node_to_cpumask_map[node])) { node 136 arch/powerpc/oprofile/cell/spu_profiler.c int cpu, node, k, num_samples, spu_num; node 145 arch/powerpc/oprofile/cell/spu_profiler.c node = cbe_cpu_to_node(cpu); node 166 arch/powerpc/oprofile/cell/spu_profiler.c spu_num = k + (node * SPUS_PER_NODE); node 200 arch/powerpc/oprofile/op_model_cell.c static void pm_rtas_reset_signals(u32 node) node 215 arch/powerpc/oprofile/op_model_cell.c pm_signal_local.cpu = node; node 235 arch/powerpc/oprofile/op_model_cell.c static int pm_rtas_activate_signals(u32 node, u32 count) node 254 arch/powerpc/oprofile/op_model_cell.c pm_signal_local[i].cpu = node; node 588 arch/powerpc/oprofile/op_model_cell.c int node; node 620 arch/powerpc/oprofile/op_model_cell.c node = cbe_cpu_to_node(cpu); node 621 arch/powerpc/oprofile/op_model_cell.c cur_phys_spu = (node * NUM_SPUS_PER_NODE) node 623 arch/powerpc/oprofile/op_model_cell.c nxt_phys_spu = (node * NUM_SPUS_PER_NODE) node 1076 arch/powerpc/oprofile/op_model_cell.c static int pm_rtas_activate_spu_profiling(u32 node) node 1086 arch/powerpc/oprofile/op_model_cell.c pm_signal_local[i].cpu = node; node 528 arch/powerpc/perf/hv-24x7.c struct rb_node node; node 570 arch/powerpc/perf/hv-24x7.c it = rb_entry(*new, struct event_uniq, node); node 599 arch/powerpc/perf/hv-24x7.c rb_link_node(&data->node, parent, new); node 600 arch/powerpc/perf/hv-24x7.c rb_insert_color(&data->node, root); node 614 arch/powerpc/perf/hv-24x7.c rbtree_postorder_for_each_entry_safe(pos, n, root, node) node 209 arch/powerpc/perf/imc-pmu.c static int update_events_in_group(struct device_node *node, struct imc_pmu *pmu) node 219 arch/powerpc/perf/imc-pmu.c if (!of_property_read_u32(node, "events", &handle)) node 232 arch/powerpc/perf/imc-pmu.c if (of_property_read_string(node, "events-prefix", &prefix)) node 236 arch/powerpc/perf/imc-pmu.c if (of_property_read_string(node, "scale", &g_scale)) node 239 arch/powerpc/perf/imc-pmu.c if (of_property_read_string(node, "unit", &g_unit)) node 243 arch/powerpc/perf/imc-pmu.c of_property_read_u32(node, "reg", &base_reg); node 638 arch/powerpc/platforms/4xx/pci.c struct device_node *node; node 1244 arch/powerpc/platforms/4xx/pci.c port->node); node 1525 arch/powerpc/platforms/4xx/pci.c if (of_device_is_compatible(port->node, node 1745 arch/powerpc/platforms/4xx/pci.c if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx")) node 1750 arch/powerpc/platforms/4xx/pci.c port->node, "ibm,plb-pciex-476fpe") || node 1752 arch/powerpc/platforms/4xx/pci.c port->node, "ibm,plb-pciex-476gtr")) node 1802 arch/powerpc/platforms/4xx/pci.c port->node); node 1881 arch/powerpc/platforms/4xx/pci.c if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx") || node 1883 arch/powerpc/platforms/4xx/pci.c port->node, "ibm,plb-pciex-476fpe") || node 1885 arch/powerpc/platforms/4xx/pci.c port->node, "ibm,plb-pciex-476gtr")) node 1925 arch/powerpc/platforms/4xx/pci.c if (of_get_property(port->node, "primary", NULL)) node 1929 arch/powerpc/platforms/4xx/pci.c bus_range = of_get_property(port->node, "bus-range", NULL); node 1932 arch/powerpc/platforms/4xx/pci.c hose = pcibios_alloc_controller(port->node); node 1965 arch/powerpc/platforms/4xx/pci.c port->node); node 1977 arch/powerpc/platforms/4xx/pci.c port->node); node 1982 arch/powerpc/platforms/4xx/pci.c pr_debug("PCIE %pOF, bus %d..%d\n", port->node, node 2007 arch/powerpc/platforms/4xx/pci.c pci_process_bridge_OF_ranges(hose, port->node, primary); node 2027 arch/powerpc/platforms/4xx/pci.c pval = of_get_property(port->node, "vendor-id", NULL); node 2038 arch/powerpc/platforms/4xx/pci.c pval = of_get_property(port->node, "device-id", NULL); node 2050 arch/powerpc/platforms/4xx/pci.c if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx")) node 2113 arch/powerpc/platforms/4xx/pci.c port->node = of_node_get(np); node 2128 arch/powerpc/platforms/4xx/pci.c if (of_node_is_type(port->node, "pci-endpoint")) { node 2130 arch/powerpc/platforms/4xx/pci.c } else if (of_node_is_type(port->node, "pci")) { node 228 arch/powerpc/platforms/4xx/uic.c static struct uic * __init uic_init_one(struct device_node *node) node 234 arch/powerpc/platforms/4xx/uic.c BUG_ON(! of_device_is_compatible(node, "ibm,uic")); node 241 arch/powerpc/platforms/4xx/uic.c indexp = of_get_property(node, "cell-index", &len); node 244 arch/powerpc/platforms/4xx/uic.c "cell-index property\n", node); node 249 arch/powerpc/platforms/4xx/uic.c dcrreg = of_get_property(node, "dcr-reg", &len); node 252 arch/powerpc/platforms/4xx/uic.c "dcr-reg property\n", node); node 257 arch/powerpc/platforms/4xx/uic.c uic->irqhost = irq_domain_add_linear(node, NR_UIC_INTS, &uic_host_ops, node 123 arch/powerpc/platforms/512x/mpc5121_ads_cpld.c cpld_pic_host_match(struct irq_domain *h, struct device_node *node, node 126 arch/powerpc/platforms/512x/mpc5121_ads_cpld.c return cpld_pic_node == node; node 214 arch/powerpc/platforms/52xx/mpc52xx_common.c unsigned int mpc52xx_get_xtal_freq(struct device_node *node) node 222 arch/powerpc/platforms/52xx/mpc52xx_common.c freq = mpc5xxx_get_bus_frequency(node); node 241 arch/powerpc/platforms/52xx/mpc52xx_gpt.c mpc52xx_gpt_irq_setup(struct mpc52xx_gpt_priv *gpt, struct device_node *node) node 247 arch/powerpc/platforms/52xx/mpc52xx_gpt.c cascade_virq = irq_of_parse_and_map(node, 0); node 251 arch/powerpc/platforms/52xx/mpc52xx_gpt.c gpt->irqhost = irq_domain_add_linear(node, 1, &mpc52xx_gpt_irq_ops, gpt); node 321 arch/powerpc/platforms/52xx/mpc52xx_gpt.c mpc52xx_gpt_gpio_setup(struct mpc52xx_gpt_priv *gpt, struct device_node *node) node 327 arch/powerpc/platforms/52xx/mpc52xx_gpt.c if (!of_find_property(node, "gpio-controller", NULL)) node 330 arch/powerpc/platforms/52xx/mpc52xx_gpt.c gpt->gc.label = kasprintf(GFP_KERNEL, "%pOF", node); node 342 arch/powerpc/platforms/52xx/mpc52xx_gpt.c gpt->gc.of_node = node; node 364 arch/powerpc/platforms/52xx/mpc52xx_pci.c mpc52xx_add_bridge(struct device_node *node) node 372 arch/powerpc/platforms/52xx/mpc52xx_pci.c pr_debug("Adding MPC52xx PCI host bridge %pOF\n", node); node 376 arch/powerpc/platforms/52xx/mpc52xx_pci.c if (of_address_to_resource(node, 0, &rsrc) != 0) { node 377 arch/powerpc/platforms/52xx/mpc52xx_pci.c printk(KERN_ERR "Can't get %pOF resources\n", node); node 381 arch/powerpc/platforms/52xx/mpc52xx_pci.c bus_range = of_get_property(node, "bus-range", &len); node 384 arch/powerpc/platforms/52xx/mpc52xx_pci.c node); node 396 arch/powerpc/platforms/52xx/mpc52xx_pci.c hose = pcibios_alloc_controller(node); node 409 arch/powerpc/platforms/52xx/mpc52xx_pci.c pci_process_bridge_OF_ranges(hose, node, 1); node 114 arch/powerpc/platforms/82xx/ep8248e.c struct device_node *node; node 117 arch/powerpc/platforms/82xx/ep8248e.c node = of_get_parent(ofdev->dev.of_node); node 118 arch/powerpc/platforms/82xx/ep8248e.c of_node_put(node); node 119 arch/powerpc/platforms/82xx/ep8248e.c if (node != ep8248e_bcsr_node) node 27 arch/powerpc/platforms/82xx/pq2ads-pci-pic.c struct device_node *node; node 128 arch/powerpc/platforms/85xx/mpc85xx_ds.c struct device_node *node; node 132 arch/powerpc/platforms/85xx/mpc85xx_ds.c node = of_find_node_by_name(NULL, "uli1575"); node 133 arch/powerpc/platforms/85xx/mpc85xx_ds.c while ((pci_with_uli = of_get_parent(node))) { node 134 arch/powerpc/platforms/85xx/mpc85xx_ds.c of_node_put(node); node 135 arch/powerpc/platforms/85xx/mpc85xx_ds.c node = pci_with_uli; node 73 arch/powerpc/platforms/85xx/sgy_cts1000.c struct device_node *node = pdev->dev.of_node; node 77 arch/powerpc/platforms/85xx/sgy_cts1000.c if (!node) node 81 arch/powerpc/platforms/85xx/sgy_cts1000.c halt_node = of_find_matching_node(node, child_match); node 161 arch/powerpc/platforms/cell/cbe_regs.c u32 cbe_node_to_cpu(int node) node 163 arch/powerpc/platforms/cell/cbe_regs.c return cpumask_first(&cbe_local_mask[node]); node 41 arch/powerpc/platforms/cell/interrupt.c struct device_node *node; node 52 arch/powerpc/platforms/cell/interrupt.c unsigned char node = bits.source >> 4; node 59 arch/powerpc/platforms/cell/interrupt.c return (node << IIC_IRQ_NODE_SHIFT) | (class << 4) | unit; node 207 arch/powerpc/platforms/cell/interrupt.c static int iic_host_match(struct irq_domain *h, struct device_node *node, node 210 arch/powerpc/platforms/cell/interrupt.c return of_device_is_compatible(node, node 236 arch/powerpc/platforms/cell/interrupt.c unsigned int node, ext, unit, class; node 248 arch/powerpc/platforms/cell/interrupt.c node = intspec[0] >> 24; node 254 arch/powerpc/platforms/cell/interrupt.c if (node > 1) node 258 arch/powerpc/platforms/cell/interrupt.c *out_hwirq = (node << IIC_IRQ_NODE_SHIFT); node 278 arch/powerpc/platforms/cell/interrupt.c struct device_node *node) node 290 arch/powerpc/platforms/cell/interrupt.c iic->node = of_node_get(node); node 294 arch/powerpc/platforms/cell/interrupt.c hw_cpu, iic->target_id, node); node 301 arch/powerpc/platforms/cell/interrupt.c unsigned int node, cascade, found = 0; node 330 arch/powerpc/platforms/cell/interrupt.c node = np[0] >> 1; node 332 arch/powerpc/platforms/cell/interrupt.c cascade = node << IIC_IRQ_NODE_SHIFT; node 346 arch/powerpc/platforms/cell/interrupt.c (node << 4) /* dest node */ | node 383 arch/powerpc/platforms/cell/interrupt.c int node = cpu >> 1; node 387 arch/powerpc/platforms/cell/interrupt.c CBE_IIC_IR_DEST_NODE(node); node 683 arch/powerpc/platforms/cell/iommu.c int node; node 688 arch/powerpc/platforms/cell/iommu.c for_each_online_node(node) { node 689 arch/powerpc/platforms/cell/iommu.c if (cell_iommu_find_ioc(node, &base)) node 696 arch/powerpc/platforms/cell/iommu.c pr_debug("iommu: cleaning up iommu on node %d\n", node); node 370 arch/powerpc/platforms/cell/pmu.c int rc, node; node 372 arch/powerpc/platforms/cell/pmu.c for_each_online_node(node) { node 374 arch/powerpc/platforms/cell/pmu.c (node << IIC_IRQ_NODE_SHIFT)); node 377 arch/powerpc/platforms/cell/pmu.c node); node 385 arch/powerpc/platforms/cell/pmu.c node); node 394 arch/powerpc/platforms/cell/pmu.c void cbe_sync_irq(int node) node 400 arch/powerpc/platforms/cell/pmu.c | (node << IIC_IRQ_NODE_SHIFT)); node 404 arch/powerpc/platforms/cell/pmu.c "for node %d\n", irq, node); node 158 arch/powerpc/platforms/cell/setup.c int node; node 175 arch/powerpc/platforms/cell/setup.c for_each_online_node(node) { node 176 arch/powerpc/platforms/cell/setup.c if (cbe_get_cpu_mic_tm_regs(cbe_node_to_cpu(node)) == NULL) node 178 arch/powerpc/platforms/cell/setup.c platform_device_register_simple("cbe-mic", node, NULL, 0); node 556 arch/powerpc/platforms/cell/spu_base.c sysfs_add_device_to_node(&spu->dev, spu->node); node 595 arch/powerpc/platforms/cell/spu_base.c mutex_lock(&cbe_spu_info[spu->node].list_mutex); node 596 arch/powerpc/platforms/cell/spu_base.c list_add(&spu->cbe_list, &cbe_spu_info[spu->node].spus); node 597 arch/powerpc/platforms/cell/spu_base.c cbe_spu_info[spu->node].n_spus++; node 598 arch/powerpc/platforms/cell/spu_base.c mutex_unlock(&cbe_spu_info[spu->node].list_mutex); node 83 arch/powerpc/platforms/cell/spu_manage.c nid = spu->node; node 121 arch/powerpc/platforms/cell/spu_manage.c struct device_node *node = spu->devnode; node 126 arch/powerpc/platforms/cell/spu_manage.c spu->name = of_get_property(node, "name", NULL); node 130 arch/powerpc/platforms/cell/spu_manage.c prop = of_get_property(node, "local-store", NULL); node 137 arch/powerpc/platforms/cell/spu_manage.c spu_map_prop_old(spu, node, "local-store"); node 141 arch/powerpc/platforms/cell/spu_manage.c prop = of_get_property(node, "problem", NULL); node 146 arch/powerpc/platforms/cell/spu_manage.c spu->problem = spu_map_prop_old(spu, node, "problem"); node 150 arch/powerpc/platforms/cell/spu_manage.c spu->priv2 = spu_map_prop_old(spu, node, "priv2"); node 155 arch/powerpc/platforms/cell/spu_manage.c spu->priv1 = spu_map_prop_old(spu, node, "priv1"); node 266 arch/powerpc/platforms/cell/spu_manage.c struct device_node *node; node 270 arch/powerpc/platforms/cell/spu_manage.c for_each_node_by_type(node, "spe") { node 271 arch/powerpc/platforms/cell/spu_manage.c ret = fn(node); node 274 arch/powerpc/platforms/cell/spu_manage.c __func__, node); node 275 arch/powerpc/platforms/cell/spu_manage.c of_node_put(node); node 292 arch/powerpc/platforms/cell/spu_manage.c spu->node = of_node_to_nid(spe); node 293 arch/powerpc/platforms/cell/spu_manage.c if (spu->node >= MAX_NUMNODES) { node 295 arch/powerpc/platforms/cell/spu_manage.c " node number too big\n", spe, spu->node); node 364 arch/powerpc/platforms/cell/spu_manage.c static struct spu *spu_lookup_reg(int node, u32 reg) node 369 arch/powerpc/platforms/cell/spu_manage.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 379 arch/powerpc/platforms/cell/spu_manage.c int node, i; node 383 arch/powerpc/platforms/cell/spu_manage.c for (node = 0; node < MAX_NUMNODES; node++) { node 387 arch/powerpc/platforms/cell/spu_manage.c spu = spu_lookup_reg(node, reg); node 68 arch/powerpc/platforms/cell/spu_priv1_mmio.c if (nr_cpus_node(spu->node)) { node 69 arch/powerpc/platforms/cell/spu_priv1_mmio.c const struct cpumask *spumask = cpumask_of_node(spu->node), node 306 arch/powerpc/platforms/cell/spufs/inode.c int count, node; node 353 arch/powerpc/platforms/cell/spufs/inode.c for (node = 0; node < MAX_NUMNODES; node++) { node 354 arch/powerpc/platforms/cell/spufs/inode.c if ((cbe_spu_info[node].n_spus - atomic_read( node 355 arch/powerpc/platforms/cell/spufs/inode.c &cbe_spu_info[node].reserved_spus)) >= count) node 359 arch/powerpc/platforms/cell/spufs/inode.c if (node == MAX_NUMNODES) { node 139 arch/powerpc/platforms/cell/spufs/sched.c int node; node 142 arch/powerpc/platforms/cell/spufs/sched.c node = ctx->spu->node; node 147 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 149 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 155 arch/powerpc/platforms/cell/spufs/sched.c static int __node_allowed(struct spu_context *ctx, int node) node 157 arch/powerpc/platforms/cell/spufs/sched.c if (nr_cpus_node(node)) { node 158 arch/powerpc/platforms/cell/spufs/sched.c const struct cpumask *mask = cpumask_of_node(node); node 167 arch/powerpc/platforms/cell/spufs/sched.c static int node_allowed(struct spu_context *ctx, int node) node 172 arch/powerpc/platforms/cell/spufs/sched.c rval = __node_allowed(ctx, node); node 180 arch/powerpc/platforms/cell/spufs/sched.c int node; node 188 arch/powerpc/platforms/cell/spufs/sched.c for_each_online_node(node) { node 191 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 192 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 201 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 217 arch/powerpc/platforms/cell/spufs/sched.c atomic_inc(&cbe_spu_info[spu->node].reserved_spus); node 253 arch/powerpc/platforms/cell/spufs/sched.c BUG_ON(!mutex_is_locked(&cbe_spu_info[spu->node].list_mutex)); node 296 arch/powerpc/platforms/cell/spufs/sched.c int node, n; node 302 arch/powerpc/platforms/cell/spufs/sched.c node = cpu_to_node(raw_smp_processor_id()); node 303 arch/powerpc/platforms/cell/spufs/sched.c for (n = 0; n < MAX_NUMNODES; n++, node++) { node 315 arch/powerpc/platforms/cell/spufs/sched.c node = (node < MAX_NUMNODES) ? node : 0; node 316 arch/powerpc/platforms/cell/spufs/sched.c if (!node_allowed(ctx, node)) node 320 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 321 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 328 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 332 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 335 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 339 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 368 arch/powerpc/platforms/cell/spufs/sched.c static struct spu *ctx_location(struct spu *ref, int offset, int node) node 375 arch/powerpc/platforms/cell/spufs/sched.c BUG_ON(spu->node != node); node 383 arch/powerpc/platforms/cell/spufs/sched.c BUG_ON(spu->node != node); node 433 arch/powerpc/platforms/cell/spufs/sched.c atomic_dec(&cbe_spu_info[spu->node].reserved_spus); node 562 arch/powerpc/platforms/cell/spufs/sched.c int node, n; node 572 arch/powerpc/platforms/cell/spufs/sched.c node = aff_ref_spu->node; node 574 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 575 arch/powerpc/platforms/cell/spufs/sched.c spu = ctx_location(aff_ref_spu, ctx->aff_offset, node); node 578 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 585 arch/powerpc/platforms/cell/spufs/sched.c node = cpu_to_node(raw_smp_processor_id()); node 586 arch/powerpc/platforms/cell/spufs/sched.c for (n = 0; n < MAX_NUMNODES; n++, node++) { node 587 arch/powerpc/platforms/cell/spufs/sched.c node = (node < MAX_NUMNODES) ? node : 0; node 588 arch/powerpc/platforms/cell/spufs/sched.c if (!node_allowed(ctx, node)) node 591 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 592 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 596 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 605 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 621 arch/powerpc/platforms/cell/spufs/sched.c int node, n; node 633 arch/powerpc/platforms/cell/spufs/sched.c node = cpu_to_node(raw_smp_processor_id()); node 634 arch/powerpc/platforms/cell/spufs/sched.c for (n = 0; n < MAX_NUMNODES; n++, node++) { node 635 arch/powerpc/platforms/cell/spufs/sched.c node = (node < MAX_NUMNODES) ? node : 0; node 636 arch/powerpc/platforms/cell/spufs/sched.c if (!node_allowed(ctx, node)) node 639 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 640 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) { node 651 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 685 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 686 arch/powerpc/platforms/cell/spufs/sched.c cbe_spu_info[node].nr_active--; node 688 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 707 arch/powerpc/platforms/cell/spufs/sched.c int node = spu->node; node 712 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 715 arch/powerpc/platforms/cell/spufs/sched.c cbe_spu_info[node].nr_active++; node 719 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 753 arch/powerpc/platforms/cell/spufs/sched.c int node = spu->node; node 755 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 756 arch/powerpc/platforms/cell/spufs/sched.c cbe_spu_info[node].nr_active--; node 762 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 825 arch/powerpc/platforms/cell/spufs/sched.c static struct spu_context *grab_runnable_context(int prio, int node) node 837 arch/powerpc/platforms/cell/spufs/sched.c if (__node_allowed(ctx, node)) { node 856 arch/powerpc/platforms/cell/spufs/sched.c new = grab_runnable_context(max_prio, spu->node); node 929 arch/powerpc/platforms/cell/spufs/sched.c new = grab_runnable_context(ctx->prio + 1, spu->node); node 957 arch/powerpc/platforms/cell/spufs/sched.c int nr_active = 0, node; node 959 arch/powerpc/platforms/cell/spufs/sched.c for (node = 0; node < MAX_NUMNODES; node++) node 960 arch/powerpc/platforms/cell/spufs/sched.c nr_active += cbe_spu_info[node].nr_active; node 997 arch/powerpc/platforms/cell/spufs/sched.c int node; node 1002 arch/powerpc/platforms/cell/spufs/sched.c for (node = 0; node < MAX_NUMNODES; node++) { node 1003 arch/powerpc/platforms/cell/spufs/sched.c struct mutex *mtx = &cbe_spu_info[node].list_mutex; node 1006 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, node 1032 arch/powerpc/platforms/cell/spufs/sched.c int node; node 1053 arch/powerpc/platforms/cell/spufs/sched.c node = spu->node; node 1055 arch/powerpc/platforms/cell/spufs/sched.c atomic_dec(&cbe_spu_info[node].busy_spus); node 1057 arch/powerpc/platforms/cell/spufs/sched.c atomic_inc(&cbe_spu_info[node].busy_spus); node 1129 arch/powerpc/platforms/cell/spufs/sched.c int node; node 1137 arch/powerpc/platforms/cell/spufs/sched.c for (node = 0; node < MAX_NUMNODES; node++) { node 1138 arch/powerpc/platforms/cell/spufs/sched.c mutex_lock(&cbe_spu_info[node].list_mutex); node 1139 arch/powerpc/platforms/cell/spufs/sched.c list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) node 1142 arch/powerpc/platforms/cell/spufs/sched.c mutex_unlock(&cbe_spu_info[node].list_mutex); node 259 arch/powerpc/platforms/chrp/setup.c struct device_node *node; node 267 arch/powerpc/platforms/chrp/setup.c node = of_find_node_by_path("/"); node 268 arch/powerpc/platforms/chrp/setup.c if (!node) node 270 arch/powerpc/platforms/chrp/setup.c property = of_get_property(node, "model", NULL); node 279 arch/powerpc/platforms/chrp/setup.c of_node_put(node); node 280 arch/powerpc/platforms/chrp/setup.c node = of_find_node_by_path(property); node 281 arch/powerpc/platforms/chrp/setup.c if (!node) node 283 arch/powerpc/platforms/chrp/setup.c if (!of_node_is_type(node, "serial")) node 290 arch/powerpc/platforms/chrp/setup.c if (of_node_name_eq(node, "failsafe") || of_node_name_eq(node, "serial")) node 293 arch/powerpc/platforms/chrp/setup.c of_node_put(node); node 35 arch/powerpc/platforms/maple/pci.c static int __init fixup_one_level_bus_range(struct device_node *node, int higher) node 37 arch/powerpc/platforms/maple/pci.c for (; node != 0;node = node->sibling) { node 43 arch/powerpc/platforms/maple/pci.c class_code = of_get_property(node, "class-code", NULL); node 47 arch/powerpc/platforms/maple/pci.c bus_range = of_get_property(node, "bus-range", &len); node 52 arch/powerpc/platforms/maple/pci.c higher = fixup_one_level_bus_range(node->child, higher); node 30 arch/powerpc/platforms/pasemi/misc.c static int __init find_i2c_driver(struct device_node *node, node 36 arch/powerpc/platforms/pasemi/misc.c if (!of_device_is_compatible(node, i2c_devices[i].of_device)) node 50 arch/powerpc/platforms/pasemi/misc.c struct device_node *node; node 59 arch/powerpc/platforms/pasemi/misc.c node = NULL; node 60 arch/powerpc/platforms/pasemi/misc.c while ((node = of_get_next_child(adap_node, node))) { node 65 arch/powerpc/platforms/pasemi/misc.c addr = of_get_property(node, "reg", &len); node 72 arch/powerpc/platforms/pasemi/misc.c info.irq = irq_of_parse_and_map(node, 0); node 76 arch/powerpc/platforms/pasemi/misc.c if (find_i2c_driver(node, &info) < 0) node 90 arch/powerpc/platforms/powermac/bootx_init.c unsigned long node, node 93 arch/powerpc/platforms/powermac/bootx_init.c struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node); node 217 arch/powerpc/platforms/powermac/bootx_init.c unsigned long node, node 220 arch/powerpc/platforms/powermac/bootx_init.c struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node); node 240 arch/powerpc/platforms/powermac/bootx_init.c bootx_node_chosen = node; node 242 arch/powerpc/platforms/powermac/bootx_init.c if (node == bootx_info->dispDeviceRegEntryOffset) { node 275 arch/powerpc/platforms/powermac/bootx_init.c unsigned long node, node 278 arch/powerpc/platforms/powermac/bootx_init.c struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node); node 318 arch/powerpc/platforms/powermac/bootx_init.c if (node == bootx_node_chosen && !strcmp(namep, "bootargs")) node 329 arch/powerpc/platforms/powermac/bootx_init.c if (node == bootx_node_chosen) { node 334 arch/powerpc/platforms/powermac/bootx_init.c else if (node == bootx_info->dispDeviceRegEntryOffset) node 115 arch/powerpc/platforms/powermac/feature.c typedef long (*feature_call)(struct device_node *node, long param, long value); node 136 arch/powerpc/platforms/powermac/feature.c static inline int simple_feature_tweak(struct device_node *node, int type, node 142 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, type); node 158 arch/powerpc/platforms/powermac/feature.c static long ohare_htw_scc_enable(struct device_node *node, long param, node 168 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 171 arch/powerpc/platforms/powermac/feature.c if (of_node_name_eq(node, "ch-a")) node 173 arch/powerpc/platforms/powermac/feature.c else if (of_node_name_eq(node, "ch-b")) node 251 arch/powerpc/platforms/powermac/feature.c static long ohare_floppy_enable(struct device_node *node, long param, node 254 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 258 arch/powerpc/platforms/powermac/feature.c static long ohare_mesh_enable(struct device_node *node, long param, long value) node 260 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 264 arch/powerpc/platforms/powermac/feature.c static long ohare_ide_enable(struct device_node *node, long param, long value) node 272 arch/powerpc/platforms/powermac/feature.c simple_feature_tweak(node, macio_ohare, node 274 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 277 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 284 arch/powerpc/platforms/powermac/feature.c static long ohare_ide_reset(struct device_node *node, long param, long value) node 288 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 291 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_ohare, node 298 arch/powerpc/platforms/powermac/feature.c static long ohare_sleep_state(struct device_node *node, long param, long value) node 313 arch/powerpc/platforms/powermac/feature.c static long heathrow_modem_enable(struct device_node *node, long param, node 320 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, macio_unknown); node 357 arch/powerpc/platforms/powermac/feature.c static long heathrow_floppy_enable(struct device_node *node, long param, node 360 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 366 arch/powerpc/platforms/powermac/feature.c static long heathrow_mesh_enable(struct device_node *node, long param, node 372 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, macio_unknown); node 395 arch/powerpc/platforms/powermac/feature.c static long heathrow_ide_enable(struct device_node *node, long param, node 400 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 403 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 410 arch/powerpc/platforms/powermac/feature.c static long heathrow_ide_reset(struct device_node *node, long param, node 415 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 418 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 425 arch/powerpc/platforms/powermac/feature.c static long heathrow_bmac_enable(struct device_node *node, long param, node 431 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 454 arch/powerpc/platforms/powermac/feature.c static long heathrow_sound_enable(struct device_node *node, long param, node 467 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 581 arch/powerpc/platforms/powermac/feature.c static long heathrow_sleep_state(struct device_node *node, long param, node 598 arch/powerpc/platforms/powermac/feature.c static long core99_scc_enable(struct device_node *node, long param, long value) node 605 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 608 arch/powerpc/platforms/powermac/feature.c if (of_node_name_eq(node, "ch-a")) node 610 arch/powerpc/platforms/powermac/feature.c else if (of_node_name_eq(node, "ch-b")) node 699 arch/powerpc/platforms/powermac/feature.c core99_modem_enable(struct device_node *node, long param, long value) node 706 arch/powerpc/platforms/powermac/feature.c if (node == NULL) { node 709 arch/powerpc/platforms/powermac/feature.c node = macio_chips[0].of_node; node 711 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 751 arch/powerpc/platforms/powermac/feature.c pangea_modem_enable(struct device_node *node, long param, long value) node 758 arch/powerpc/platforms/powermac/feature.c if (node == NULL) { node 762 arch/powerpc/platforms/powermac/feature.c node = macio_chips[0].of_node; node 764 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 806 arch/powerpc/platforms/powermac/feature.c core99_ata100_enable(struct device_node *node, long value) node 826 arch/powerpc/platforms/powermac/feature.c if (pci_device_from_OF_node(node, &pbus, &pid) == 0) node 841 arch/powerpc/platforms/powermac/feature.c core99_ide_enable(struct device_node *node, long param, long value) node 848 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 851 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 854 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 857 arch/powerpc/platforms/powermac/feature.c return core99_ata100_enable(node, value); node 864 arch/powerpc/platforms/powermac/feature.c core99_ide_reset(struct device_node *node, long param, long value) node 868 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 871 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 874 arch/powerpc/platforms/powermac/feature.c return simple_feature_tweak(node, macio_unknown, node 882 arch/powerpc/platforms/powermac/feature.c core99_gmac_enable(struct device_node *node, long param, long value) node 899 arch/powerpc/platforms/powermac/feature.c core99_gmac_phy_reset(struct device_node *node, long param, long value) node 924 arch/powerpc/platforms/powermac/feature.c core99_sound_chip_enable(struct device_node *node, long param, long value) node 929 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 954 arch/powerpc/platforms/powermac/feature.c core99_airport_enable(struct device_node *node, long param, long value) node 960 arch/powerpc/platforms/powermac/feature.c macio = macio_find(node, 0); node 967 arch/powerpc/platforms/powermac/feature.c if (node != macio->of_node && node 968 arch/powerpc/platforms/powermac/feature.c (!node->parent || node->parent != macio->of_node)) node 1041 arch/powerpc/platforms/powermac/feature.c core99_reset_cpu(struct device_node *node, long param, long value) node 1082 arch/powerpc/platforms/powermac/feature.c core99_usb_enable(struct device_node *node, long param, long value) node 1095 arch/powerpc/platforms/powermac/feature.c prop = of_get_property(node, "AAPL,clock-id", NULL); node 1235 arch/powerpc/platforms/powermac/feature.c core99_firewire_enable(struct device_node *node, long param, long value) node 1262 arch/powerpc/platforms/powermac/feature.c core99_firewire_cable_power(struct device_node *node, long param, long value) node 1293 arch/powerpc/platforms/powermac/feature.c intrepid_aack_delay_enable(struct device_node *node, long param, long value) node 1314 arch/powerpc/platforms/powermac/feature.c core99_read_gpio(struct device_node *node, long param, long value) node 1323 arch/powerpc/platforms/powermac/feature.c core99_write_gpio(struct device_node *node, long param, long value) node 1332 arch/powerpc/platforms/powermac/feature.c static long g5_gmac_enable(struct device_node *node, long param, long value) node 1337 arch/powerpc/platforms/powermac/feature.c if (node == NULL) node 1346 arch/powerpc/platforms/powermac/feature.c k2_skiplist[0] = node; node 1357 arch/powerpc/platforms/powermac/feature.c static long g5_fw_enable(struct device_node *node, long param, long value) node 1362 arch/powerpc/platforms/powermac/feature.c if (node == NULL) node 1371 arch/powerpc/platforms/powermac/feature.c k2_skiplist[1] = node; node 1382 arch/powerpc/platforms/powermac/feature.c static long g5_mpic_enable(struct device_node *node, long param, long value) node 1385 arch/powerpc/platforms/powermac/feature.c struct device_node *parent = of_get_parent(node); node 1402 arch/powerpc/platforms/powermac/feature.c static long g5_eth_phy_reset(struct device_node *node, long param, long value) node 1412 arch/powerpc/platforms/powermac/feature.c phy = of_get_next_child(node, NULL); node 1432 arch/powerpc/platforms/powermac/feature.c static long g5_i2s_enable(struct device_node *node, long param, long value) node 1458 arch/powerpc/platforms/powermac/feature.c if (strncmp(node->name, "i2s-", 4)) node 1460 arch/powerpc/platforms/powermac/feature.c cell = node->name[4] - 'a'; node 1491 arch/powerpc/platforms/powermac/feature.c static long g5_reset_cpu(struct device_node *node, long param, long value) node 1822 arch/powerpc/platforms/powermac/feature.c core99_sleep_state(struct device_node *node, long param, long value) node 1855 arch/powerpc/platforms/powermac/feature.c generic_dev_can_wake(struct device_node *node, long param, long value) node 1866 arch/powerpc/platforms/powermac/feature.c static long generic_get_mb_info(struct device_node *node, long param, long value) node 2369 arch/powerpc/platforms/powermac/feature.c struct device_node *node; node 2391 arch/powerpc/platforms/powermac/feature.c node = (struct device_node*)va_arg(args, void*); node 2396 arch/powerpc/platforms/powermac/feature.c return func(node, param, value); node 2617 arch/powerpc/platforms/powermac/feature.c struct device_node* node; node 2624 arch/powerpc/platforms/powermac/feature.c for_each_node_by_name(node, name) { node 2627 arch/powerpc/platforms/powermac/feature.c if (of_device_is_compatible(node, compat)) node 2630 arch/powerpc/platforms/powermac/feature.c if (!node) node 2635 arch/powerpc/platforms/powermac/feature.c if (macio_chips[i].of_node == node) node 2641 arch/powerpc/platforms/powermac/feature.c printk(KERN_ERR "pmac_feature: %pOF skipped\n", node); node 2644 arch/powerpc/platforms/powermac/feature.c addrp = of_get_pci_address(node, 0, &size, NULL); node 2647 arch/powerpc/platforms/powermac/feature.c node); node 2650 arch/powerpc/platforms/powermac/feature.c addr = of_translate_address(node, addrp); node 2653 arch/powerpc/platforms/powermac/feature.c node); node 2659 arch/powerpc/platforms/powermac/feature.c node); node 2663 arch/powerpc/platforms/powermac/feature.c const u32 *did = of_get_property(node, "device-id", NULL); node 2671 arch/powerpc/platforms/powermac/feature.c macio_chips[i].of_node = node; node 2676 arch/powerpc/platforms/powermac/feature.c revp = of_get_property(node, "revision-id", NULL); node 955 arch/powerpc/platforms/powermac/low_i2c.c struct pmac_i2c_bus *pmac_i2c_find_bus(struct device_node *node) node 957 arch/powerpc/platforms/powermac/low_i2c.c struct device_node *p = of_node_get(node); node 1235 arch/powerpc/platforms/powermac/low_i2c.c bus = pmac_i2c_find_bus(func->node); node 1238 arch/powerpc/platforms/powermac/low_i2c.c func->node); node 1243 arch/powerpc/platforms/powermac/low_i2c.c func->node); node 1258 arch/powerpc/platforms/powermac/low_i2c.c inst->addr = pmac_i2c_get_dev_addr(func->node); node 59 arch/powerpc/platforms/powermac/pci.c static int __init fixup_one_level_bus_range(struct device_node *node, int higher) node 61 arch/powerpc/platforms/powermac/pci.c for (; node; node = node->sibling) { node 67 arch/powerpc/platforms/powermac/pci.c class_code = of_get_property(node, "class-code", NULL); node 71 arch/powerpc/platforms/powermac/pci.c bus_range = of_get_property(node, "bus-range", &len); node 76 arch/powerpc/platforms/powermac/pci.c higher = fixup_one_level_bus_range(node->child, higher); node 952 arch/powerpc/platforms/powermac/pci.c struct device_node* node; node 956 arch/powerpc/platforms/powermac/pci.c node = pci_device_to_OF_node(dev); node 963 arch/powerpc/platforms/powermac/pci.c && !node) { node 969 arch/powerpc/platforms/powermac/pci.c if (!node) node 972 arch/powerpc/platforms/powermac/pci.c uninorth_child = node->parent && node 973 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node->parent, "uni-north"); node 978 arch/powerpc/platforms/powermac/pci.c if (uninorth_child && of_node_name_eq(node, "firewire") && node 979 arch/powerpc/platforms/powermac/pci.c (of_device_is_compatible(node, "pci106b,18") || node 980 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node, "pci106b,30") || node 981 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node, "pci11c1,5811"))) { node 982 arch/powerpc/platforms/powermac/pci.c pmac_call_feature(PMAC_FTR_1394_CABLE_POWER, node, 0, 1); node 983 arch/powerpc/platforms/powermac/pci.c pmac_call_feature(PMAC_FTR_1394_ENABLE, node, 0, 1); node 986 arch/powerpc/platforms/powermac/pci.c if (uninorth_child && of_node_name_eq(node, "ethernet") && node 987 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node, "gmac")) { node 988 arch/powerpc/platforms/powermac/pci.c pmac_call_feature(PMAC_FTR_GMAC_ENABLE, node, 0, 1); node 1016 arch/powerpc/platforms/powermac/pci.c struct device_node *node = pci_device_to_OF_node(dev); node 1021 arch/powerpc/platforms/powermac/pci.c if (dev->class == PCI_CLASS_SERIAL_USB_OHCI && !node) node 1235 arch/powerpc/platforms/powermac/pci.c struct device_node *node = pci_bus_to_OF_node(bus); node 1241 arch/powerpc/platforms/powermac/pci.c if (bus->self == NULL && (of_device_is_compatible(node, "u3-agp") || node 1242 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node, "u4-pcie") || node 1243 arch/powerpc/platforms/powermac/pci.c of_device_is_compatible(node, "u3-ht"))) node 29 arch/powerpc/platforms/powermac/pfunc_base.c unsigned int irq = irq_of_parse_and_map(func->node, 0); node 32 arch/powerpc/platforms/powermac/pfunc_base.c return request_irq(irq, macio_gpio_irq, 0, func->node->name, func); node 37 arch/powerpc/platforms/powermac/pfunc_base.c unsigned int irq = irq_of_parse_and_map(func->node, 0); node 59 arch/powerpc/platforms/powermac/pfunc_base.c tmp, func->node, addr); node 541 arch/powerpc/platforms/powermac/pfunc_core.c struct device_node *node; node 573 arch/powerpc/platforms/powermac/pfunc_core.c if (dev->node == np) node 652 arch/powerpc/platforms/powermac/pfunc_core.c func->node = dev->node; node 688 arch/powerpc/platforms/powermac/pfunc_core.c for (pp = dev->node->properties; pp != 0; pp = pp->next) { node 729 arch/powerpc/platforms/powermac/pfunc_core.c dev->node = of_node_get(np); node 944 arch/powerpc/platforms/powermac/pfunc_core.c DBG(" ** pmf_call_one(%pOF/%s) **\n", dev->node, func->name); node 267 arch/powerpc/platforms/powermac/pic.c static int pmac_pic_host_match(struct irq_domain *h, struct device_node *node, node 335 arch/powerpc/platforms/powermac/setup.c void note_scsi_host(struct device_node *node, void *host) node 162 arch/powerpc/platforms/powernv/npu-dma.c pnv_pci_link_table_and_group(phb->hose->node, num, node 31 arch/powerpc/platforms/powernv/opal-fadump.c void __init opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) node 39 arch/powerpc/platforms/powernv/opal-fadump.c dn = of_get_flat_dt_subnode_by_name(node, "dump"); node 623 arch/powerpc/platforms/powernv/opal-fadump.c void __init opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) node 635 arch/powerpc/platforms/powernv/opal-fadump.c dn = of_get_flat_dt_subnode_by_name(node, "dump"); node 52 arch/powerpc/platforms/powernv/opal-imc.c static void export_imc_mode_and_cmd(struct device_node *node, node 65 arch/powerpc/platforms/powernv/opal-imc.c if (of_property_read_u32(node, "cb_offset", &cb_offset)) node 93 arch/powerpc/platforms/powernv/opal-imc.c static int imc_get_mem_addr_nest(struct device_node *node, node 101 arch/powerpc/platforms/powernv/opal-imc.c nr_chips = of_property_count_u32_elems(node, "chip-id"); node 115 arch/powerpc/platforms/powernv/opal-imc.c if (of_property_read_u32_array(node, "chip-id", chipid_arr, nr_chips)) node 118 arch/powerpc/platforms/powernv/opal-imc.c if (of_property_read_u64_array(node, "base-addr", base_addr_arr, node 231 arch/powerpc/platforms/powernv/opal-imc.c struct device_node *node; node 234 arch/powerpc/platforms/powernv/opal-imc.c for_each_compatible_node(node, NULL, IMC_DTB_UNIT_COMPAT) { node 235 arch/powerpc/platforms/powernv/opal-imc.c if (of_property_read_u32(node, "type", &type)) node 137 arch/powerpc/platforms/powernv/opal-irqchip.c static int opal_event_match(struct irq_domain *h, struct device_node *node, node 140 arch/powerpc/platforms/powernv/opal-irqchip.c return irq_domain_get_of_node(h) == node; node 106 arch/powerpc/platforms/powernv/opal-msglog.c struct memcons *memcons_init(struct device_node *node, const char *mc_prop_name) node 111 arch/powerpc/platforms/powernv/opal-msglog.c if (of_property_read_u64(node, mc_prop_name, &mcaddr)) { node 144 arch/powerpc/platforms/powernv/opal-powercap.c struct device_node *powercap, *node; node 165 arch/powerpc/platforms/powernv/opal-powercap.c for_each_child_of_node(powercap, node) { node 170 arch/powerpc/platforms/powernv/opal-powercap.c if (!of_property_read_u32(node, "powercap-min", &min)) { node 175 arch/powerpc/platforms/powernv/opal-powercap.c if (!of_property_read_u32(node, "powercap-max", &max)) { node 180 arch/powerpc/platforms/powernv/opal-powercap.c if (!of_property_read_u32(node, "powercap-current", &cur)) { node 198 arch/powerpc/platforms/powernv/opal-powercap.c pcaps[i].pg.name = kasprintf(GFP_KERNEL, "%pOFn", node); node 44 arch/powerpc/platforms/powernv/opal-prd.c struct device_node *parent, *node; node 56 arch/powerpc/platforms/powernv/opal-prd.c for_each_child_of_node(parent, node) { node 61 arch/powerpc/platforms/powernv/opal-prd.c addrp = of_get_address(node, 0, &range_size, NULL); node 66 arch/powerpc/platforms/powernv/opal-prd.c label = of_get_property(node, "ibm,prd-label", NULL); node 77 arch/powerpc/platforms/powernv/opal-prd.c of_node_put(node); node 125 arch/powerpc/platforms/powernv/opal-psr.c struct device_node *psr, *node; node 146 arch/powerpc/platforms/powernv/opal-psr.c for_each_child_of_node(psr, node) { node 147 arch/powerpc/platforms/powernv/opal-psr.c if (of_property_read_u32(node, "handle", node 152 arch/powerpc/platforms/powernv/opal-psr.c if (of_property_read_string(node, "label", node 162 arch/powerpc/platforms/powernv/opal-sensor-groups.c struct device_node *sg, *node; node 181 arch/powerpc/platforms/powernv/opal-sensor-groups.c for_each_child_of_node(sg, node) { node 185 arch/powerpc/platforms/powernv/opal-sensor-groups.c ops = of_get_property(node, "ops", &len); node 207 arch/powerpc/platforms/powernv/opal-sensor-groups.c if (of_property_read_u32(node, "sensor-group-id", &sgid)) { node 212 arch/powerpc/platforms/powernv/opal-sensor-groups.c if (!of_property_read_u32(node, "ibm,chip-id", &chipid)) node 213 arch/powerpc/platforms/powernv/opal-sensor-groups.c sprintf(sgs[i].name, "%pOFn%d", node, chipid); node 215 arch/powerpc/platforms/powernv/opal-sensor-groups.c sprintf(sgs[i].name, "%pOFn", node); node 98 arch/powerpc/platforms/powernv/opal.c int __init early_init_dt_scan_opal(unsigned long node, node 107 arch/powerpc/platforms/powernv/opal.c basep = of_get_flat_dt_prop(node, "opal-base-address", &basesz); node 108 arch/powerpc/platforms/powernv/opal.c entryp = of_get_flat_dt_prop(node, "opal-entry-address", &entrysz); node 109 arch/powerpc/platforms/powernv/opal.c sizep = of_get_flat_dt_prop(node, "opal-runtime-size", &runtimesz); node 125 arch/powerpc/platforms/powernv/opal.c if (of_flat_dt_is_compatible(node, "ibm,opal-v3")) { node 135 arch/powerpc/platforms/powernv/opal.c int __init early_init_dt_scan_recoverable_ranges(unsigned long node, node 144 arch/powerpc/platforms/powernv/opal.c prop = of_get_flat_dt_prop(node, "mcheck-recoverable-ranges", &psize); node 387 arch/powerpc/platforms/powernv/pci-ioda-tce.c long pnv_pci_link_table_and_group(int node, int num, node 397 arch/powerpc/platforms/powernv/pci-ioda-tce.c node); node 1807 arch/powerpc/platforms/powernv/pci-ioda.c table_pages = alloc_pages_node(pe->phb->hose->node, GFP_KERNEL, node 2220 arch/powerpc/platforms/powernv/pci-ioda.c tbl = pnv_pci_table_alloc(phb->hose->node); node 2226 arch/powerpc/platforms/powernv/pci-ioda.c pnv_pci_link_table_and_group(phb->hose->node, 0, tbl, &pe->table_group); node 2244 arch/powerpc/platforms/powernv/pci-ioda.c tce_mem = alloc_pages_node(phb->hose->node, GFP_KERNEL, node 2279 arch/powerpc/platforms/powernv/pci-ioda.c iommu_init_table(tbl, phb->hose->node, 0, 0); node 2327 arch/powerpc/platforms/powernv/pci-ioda.c pnv_pci_link_table_and_group(phb->hose->node, num, node 2368 arch/powerpc/platforms/powernv/pci-ioda.c int nid = pe->phb->hose->node; node 2449 arch/powerpc/platforms/powernv/pci-ioda.c iommu_init_table(tbl, pe->phb->hose->node, res_start, res_end); node 3176 arch/powerpc/platforms/powernv/pci-ioda.c list_for_each_entry(child, &bus->children, node) node 241 arch/powerpc/platforms/powernv/pci.h extern long pnv_pci_link_table_and_group(int node, int num, node 36 arch/powerpc/platforms/powernv/powernv.h struct memcons *memcons_init(struct device_node *node, const char *mc_prop_name); node 21 arch/powerpc/platforms/powernv/ultravisor.c int __init early_init_dt_scan_ultravisor(unsigned long node, const char *uname, node 24 arch/powerpc/platforms/powernv/ultravisor.c if (!of_flat_dt_is_compatible(node, "ibm,ultravisor")) node 48 arch/powerpc/platforms/powernv/ultravisor.c struct device_node *node; node 53 arch/powerpc/platforms/powernv/ultravisor.c node = of_find_compatible_node(NULL, NULL, "ibm,uv-firmware"); node 54 arch/powerpc/platforms/powernv/ultravisor.c if (!node) node 57 arch/powerpc/platforms/powernv/ultravisor.c uv_memcons = memcons_init(node, "memcons"); node 49 arch/powerpc/platforms/powernv/vas.c INIT_LIST_HEAD(&vinst->node); node 82 arch/powerpc/platforms/powernv/vas.c list_add(&vinst->node, &vas_instances); node 112 arch/powerpc/platforms/powernv/vas.c vinst = list_entry(ent, struct vas_instance, node); node 308 arch/powerpc/platforms/powernv/vas.h struct list_head node; node 258 arch/powerpc/platforms/ps3/os-area.c static void os_area_set_property(struct device_node *node, node 262 arch/powerpc/platforms/ps3/os-area.c struct property *tmp = of_find_property(node, prop->name, NULL); node 266 arch/powerpc/platforms/ps3/os-area.c of_remove_property(node, tmp); node 269 arch/powerpc/platforms/ps3/os-area.c result = of_add_property(node, prop); node 281 arch/powerpc/platforms/ps3/os-area.c static void __init os_area_get_property(struct device_node *node, node 284 arch/powerpc/platforms/ps3/os-area.c const struct property *tmp = of_find_property(node, prop->name, NULL); node 675 arch/powerpc/platforms/ps3/os-area.c struct device_node *node; node 680 arch/powerpc/platforms/ps3/os-area.c node = of_find_node_by_path("/"); node 681 arch/powerpc/platforms/ps3/os-area.c if (node) { node 682 arch/powerpc/platforms/ps3/os-area.c os_area_set_property(node, &property_rtc_diff); node 683 arch/powerpc/platforms/ps3/os-area.c of_node_put(node); node 774 arch/powerpc/platforms/ps3/os-area.c struct device_node *node; node 778 arch/powerpc/platforms/ps3/os-area.c node = of_find_node_by_path("/"); node 780 arch/powerpc/platforms/ps3/os-area.c if (!saved_params.valid && node) { node 782 arch/powerpc/platforms/ps3/os-area.c os_area_get_property(node, &property_rtc_diff); node 783 arch/powerpc/platforms/ps3/os-area.c os_area_get_property(node, &property_av_multi_out); node 789 arch/powerpc/platforms/ps3/os-area.c if (node) { node 790 arch/powerpc/platforms/ps3/os-area.c os_area_set_property(node, &property_rtc_diff); node 791 arch/powerpc/platforms/ps3/os-area.c os_area_set_property(node, &property_av_multi_out); node 792 arch/powerpc/platforms/ps3/os-area.c of_node_put(node); node 141 arch/powerpc/platforms/pseries/firmware.c static int __init probe_fw_features(unsigned long node, const char *uname, int node 153 arch/powerpc/platforms/pseries/firmware.c prop = of_get_flat_dt_prop(node, "ibm,hypertas-functions", node 164 arch/powerpc/platforms/pseries/firmware.c prop = of_get_flat_dt_prop(node, "ibm,architecture-vec-5", node 72 arch/powerpc/platforms/pseries/hvcserver.c pi = list_entry(element, struct hvcs_partner_info, node); node 183 arch/powerpc/platforms/pseries/hvcserver.c list_add_tail(&(next_partner_info->node), head); node 42 arch/powerpc/platforms/pseries/iommu.c static struct iommu_table_group *iommu_pseries_alloc_group(int node) node 48 arch/powerpc/platforms/pseries/iommu.c node); node 52 arch/powerpc/platforms/pseries/iommu.c tbl = kzalloc_node(sizeof(struct iommu_table), GFP_KERNEL, node); node 484 arch/powerpc/platforms/pseries/iommu.c struct device_node *node; node 488 arch/powerpc/platforms/pseries/iommu.c node = phb->dn; node 490 arch/powerpc/platforms/pseries/iommu.c basep = of_get_property(node, "linux,tce-base", NULL); node 491 arch/powerpc/platforms/pseries/iommu.c sizep = of_get_property(node, "linux,tce-size", NULL); node 622 arch/powerpc/platforms/pseries/iommu.c pci->table_group = iommu_pseries_alloc_group(pci->phb->node); node 627 arch/powerpc/platforms/pseries/iommu.c iommu_init_table(tbl, pci->phb->node, 0, 0); node 704 arch/powerpc/platforms/pseries/iommu.c ppci->table_group = iommu_pseries_alloc_group(ppci->phb->node); node 709 arch/powerpc/platforms/pseries/iommu.c iommu_init_table(tbl, ppci->phb->node, 0, 0); node 734 arch/powerpc/platforms/pseries/iommu.c PCI_DN(dn)->table_group = iommu_pseries_alloc_group(phb->node); node 738 arch/powerpc/platforms/pseries/iommu.c iommu_init_table(tbl, phb->node, 0, 0); node 1183 arch/powerpc/platforms/pseries/iommu.c pci->table_group = iommu_pseries_alloc_group(pci->phb->node); node 1188 arch/powerpc/platforms/pseries/iommu.c iommu_init_table(tbl, pci->phb->node, 0, 0); node 306 arch/powerpc/platforms/pseries/papr_scm.c static inline int papr_scm_node(int node) node 311 arch/powerpc/platforms/pseries/papr_scm.c if ((node == NUMA_NO_NODE) || node_online(node)) node 312 arch/powerpc/platforms/pseries/papr_scm.c return node; node 316 arch/powerpc/platforms/pseries/papr_scm.c dist = node_distance(node, nid); node 242 arch/powerpc/platforms/pseries/reconfig.c struct device_node *node; node 245 arch/powerpc/platforms/pseries/reconfig.c if ((node = of_find_node_by_path(buf))) node 246 arch/powerpc/platforms/pseries/reconfig.c rv = pSeries_reconfig_remove_node(node); node 248 arch/powerpc/platforms/pseries/reconfig.c of_node_put(node); node 490 arch/powerpc/platforms/pseries/rtas-fadump.c void __init rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) node 500 arch/powerpc/platforms/pseries/rtas-fadump.c token = of_get_flat_dt_prop(node, "ibm,configure-kernel-dump", NULL); node 515 arch/powerpc/platforms/pseries/rtas-fadump.c fdm_active = of_get_flat_dt_prop(node, "ibm,kernel-dump", NULL); node 528 arch/powerpc/platforms/pseries/rtas-fadump.c sections = of_get_flat_dt_prop(node, "ibm,configure-kernel-dump-sizes", node 456 arch/powerpc/platforms/pseries/setup.c struct device_node *node; node 460 arch/powerpc/platforms/pseries/setup.c for_each_child_of_node(root, node) { node 461 arch/powerpc/platforms/pseries/setup.c if (!of_node_is_type(node, "pci") && node 462 arch/powerpc/platforms/pseries/setup.c !of_node_is_type(node, "pciex")) node 465 arch/powerpc/platforms/pseries/setup.c phb = pcibios_alloc_controller(node); node 469 arch/powerpc/platforms/pseries/setup.c pci_process_bridge_OF_ranges(phb, node, 0); node 230 arch/powerpc/sysdev/cpm2_pic.c void cpm2_pic_init(struct device_node *node) node 264 arch/powerpc/sysdev/cpm2_pic.c cpm2_pic_host = irq_domain_add_linear(node, 64, &cpm2_pic_host_ops, NULL); node 15 arch/powerpc/sysdev/dcr.c static struct device_node *find_dcr_parent(struct device_node *node) node 20 arch/powerpc/sysdev/dcr.c for (par = of_node_get(node); par;) { node 180 arch/powerpc/sysdev/ehv_pic.c static int ehv_pic_host_match(struct irq_domain *h, struct device_node *node, node 185 arch/powerpc/sysdev/ehv_pic.c return of_node == NULL || of_node == node; node 186 arch/powerpc/sysdev/fsl_lbc.c struct device_node *node) node 198 arch/powerpc/sysdev/fsl_lbc.c if (of_device_is_compatible(node, "fsl,elbc")) node 181 arch/powerpc/sysdev/fsl_pci.c struct device_node *node; node 183 arch/powerpc/sysdev/fsl_pci.c node = of_find_node_by_type(NULL, "memory"); node 184 arch/powerpc/sysdev/fsl_pci.c if (!node) { node 189 arch/powerpc/sysdev/fsl_pci.c return of_property_read_bool(node, "linux,usable-memory"); node 1310 arch/powerpc/sysdev/fsl_pci.c struct device_node *node; node 1313 arch/powerpc/sysdev/fsl_pci.c node = pdev->dev.of_node; node 1314 arch/powerpc/sysdev/fsl_pci.c ret = fsl_add_bridge(pdev, fsl_pci_primary == node); node 125 arch/powerpc/sysdev/fsl_rio.h struct device_node *node); node 328 arch/powerpc/sysdev/fsl_rmu.c &fsl_dbell->mport[i]->dbells, node) { node 1065 arch/powerpc/sysdev/fsl_rmu.c int fsl_rio_setup_rmu(struct rio_mport *mport, struct device_node *node) node 1079 arch/powerpc/sysdev/fsl_rmu.c if (!node) { node 1089 arch/powerpc/sysdev/fsl_rmu.c aw = of_n_addr_cells(node); node 1090 arch/powerpc/sysdev/fsl_rmu.c msg_addr = of_get_property(node, "reg", &mlen); node 1093 arch/powerpc/sysdev/fsl_rmu.c node); node 1102 arch/powerpc/sysdev/fsl_rmu.c rmu->txirq = irq_of_parse_and_map(node, 0); node 1103 arch/powerpc/sysdev/fsl_rmu.c rmu->rxirq = irq_of_parse_and_map(node, 1); node 1105 arch/powerpc/sysdev/fsl_rmu.c node, rmu->txirq, rmu->rxirq); node 102 arch/powerpc/sysdev/fsl_soc.c struct device_node *node; node 107 arch/powerpc/sysdev/fsl_soc.c node = of_find_compatible_node(NULL, NULL, "fsl,cpm-brg"); node 108 arch/powerpc/sysdev/fsl_soc.c if (node) { node 109 arch/powerpc/sysdev/fsl_soc.c of_property_read_u32(node, "clock-frequency", &brgfreq); node 110 arch/powerpc/sysdev/fsl_soc.c of_node_put(node); node 115 arch/powerpc/sysdev/fsl_soc.c node = of_find_node_by_type(NULL, "cpm"); node 116 arch/powerpc/sysdev/fsl_soc.c if (!node) node 117 arch/powerpc/sysdev/fsl_soc.c node = of_find_compatible_node(NULL, NULL, "fsl,qe"); node 118 arch/powerpc/sysdev/fsl_soc.c if (!node) node 119 arch/powerpc/sysdev/fsl_soc.c node = of_find_node_by_type(NULL, "qe"); node 121 arch/powerpc/sysdev/fsl_soc.c if (node) { node 122 arch/powerpc/sysdev/fsl_soc.c of_property_read_u32(node, "brg-frequency", &brgfreq); node 124 arch/powerpc/sysdev/fsl_soc.c if (!of_property_read_u32(node, "bus-frequency", node 127 arch/powerpc/sysdev/fsl_soc.c of_node_put(node); node 138 arch/powerpc/sysdev/fsl_soc.c struct device_node *node; node 143 arch/powerpc/sysdev/fsl_soc.c node = of_find_node_by_type(NULL, "serial"); node 144 arch/powerpc/sysdev/fsl_soc.c if (node) { node 145 arch/powerpc/sysdev/fsl_soc.c of_property_read_u32(node, "current-speed", &fs_baudrate); node 146 arch/powerpc/sysdev/fsl_soc.c of_node_put(node); node 161 arch/powerpc/sysdev/i8259.c static int i8259_host_match(struct irq_domain *h, struct device_node *node, node 165 arch/powerpc/sysdev/i8259.c return of_node == NULL || of_node == node; node 223 arch/powerpc/sysdev/i8259.c void i8259_init(struct device_node *node, unsigned long intack_addr) node 263 arch/powerpc/sysdev/i8259.c i8259_host = irq_domain_add_legacy_isa(node, &i8259_host_ops, NULL); node 671 arch/powerpc/sysdev/ipic.c static int ipic_host_match(struct irq_domain *h, struct device_node *node, node 676 arch/powerpc/sysdev/ipic.c return of_node == NULL || of_node == node; node 699 arch/powerpc/sysdev/ipic.c struct ipic * __init ipic_init(struct device_node *node, unsigned int flags) node 705 arch/powerpc/sysdev/ipic.c ret = of_address_to_resource(node, 0, &res); node 713 arch/powerpc/sysdev/ipic.c ipic->irqhost = irq_domain_add_linear(node, NR_IPIC_INTS, node 15 arch/powerpc/sysdev/mpc5xxx_clocks.c unsigned long mpc5xxx_get_bus_frequency(struct device_node *node) node 19 arch/powerpc/sysdev/mpc5xxx_clocks.c of_node_get(node); node 20 arch/powerpc/sysdev/mpc5xxx_clocks.c while (node) { node 21 arch/powerpc/sysdev/mpc5xxx_clocks.c p_bus_freq = of_get_property(node, "bus-frequency", NULL); node 25 arch/powerpc/sysdev/mpc5xxx_clocks.c node = of_get_next_parent(node); node 27 arch/powerpc/sysdev/mpc5xxx_clocks.c of_node_put(node); node 326 arch/powerpc/sysdev/mpic.c phys_addr_t phys_addr = dcr_resource_start(mpic->node, 0); node 327 arch/powerpc/sysdev/mpic.c rb->dhost = dcr_map(mpic->node, phys_addr + offset, size); node 992 arch/powerpc/sysdev/mpic.c static int mpic_host_match(struct irq_domain *h, struct device_node *node, node 997 arch/powerpc/sysdev/mpic.c return of_node == NULL || of_node == node; node 1213 arch/powerpc/sysdev/mpic.c struct mpic * __init mpic_alloc(struct device_node *node, node 1239 arch/powerpc/sysdev/mpic.c if (node) { node 1240 arch/powerpc/sysdev/mpic.c node = of_node_get(node); node 1242 arch/powerpc/sysdev/mpic.c node = of_find_matching_node(NULL, mpic_device_id); node 1243 arch/powerpc/sysdev/mpic.c if (!node) node 1250 arch/powerpc/sysdev/mpic.c if (of_property_read_bool(node, "dcr-reg")) { node 1254 arch/powerpc/sysdev/mpic.c if (of_address_to_resource(node, 0, &r)) node 1261 arch/powerpc/sysdev/mpic.c if (of_get_property(node, "big-endian", NULL)) node 1263 arch/powerpc/sysdev/mpic.c if (of_get_property(node, "pic-no-reset", NULL)) node 1265 arch/powerpc/sysdev/mpic.c if (of_get_property(node, "single-cpu-affinity", NULL)) node 1267 arch/powerpc/sysdev/mpic.c if (of_device_is_compatible(node, "fsl,mpic")) { node 1278 arch/powerpc/sysdev/mpic.c mpic->node = node; node 1323 arch/powerpc/sysdev/mpic.c psrc = of_get_property(mpic->node, "protected-sources", &psize); node 1470 arch/powerpc/sysdev/mpic.c of_property_read_u32(mpic->node, "last-interrupt-source", &last_irq); node 1487 arch/powerpc/sysdev/mpic.c mpic->irqhost = irq_domain_add_linear(mpic->node, node 1530 arch/powerpc/sysdev/mpic.c of_node_put(node); node 1650 arch/powerpc/sysdev/mpic.c int virq = irq_of_parse_and_map(mpic->node, 0); node 1653 arch/powerpc/sysdev/mpic.c mpic->node, virq); node 133 arch/powerpc/sysdev/mpic_msgr.c static int mpic_msgr_block_number(struct device_node *node) node 149 arch/powerpc/sysdev/mpic_msgr.c if (node == of_find_node_by_path(prop->value)) node 66 arch/powerpc/sysdev/mpic_timer.c struct list_head node; node 202 arch/powerpc/sysdev/mpic_timer.c list_for_each_entry(priv, &timer_group_list, node) { node 509 arch/powerpc/sysdev/mpic_timer.c list_add_tail(&priv->node, &timer_group_list); node 527 arch/powerpc/sysdev/mpic_timer.c list_for_each_entry(priv, &timer_group_list, node) { node 23 arch/powerpc/sysdev/of_rtc.c struct device_node *node; node 30 arch/powerpc/sysdev/of_rtc.c for_each_compatible_node(node, NULL, node 38 arch/powerpc/sysdev/of_rtc.c node); node 42 arch/powerpc/sysdev/of_rtc.c err = of_address_to_resource(node, 0, res); node 46 arch/powerpc/sysdev/of_rtc.c node); node 51 arch/powerpc/sysdev/of_rtc.c node, plat_name, node 104 arch/powerpc/sysdev/pmi.c list_for_each_entry(handler, &data->handler, node) { node 184 arch/powerpc/sysdev/pmi.c list_for_each_entry_safe(handler, tmp, &data->handler, node) node 185 arch/powerpc/sysdev/pmi.c list_del(&handler->node); node 246 arch/powerpc/sysdev/pmi.c list_add_tail(&handler->node, &data->handler); node 261 arch/powerpc/sysdev/pmi.c list_del(&handler->node); node 403 arch/powerpc/sysdev/tsi108_pci.c void __init tsi108_pci_int_init(struct device_node *node) node 407 arch/powerpc/sysdev/tsi108_pci.c pci_irq_host = irq_domain_add_legacy_isa(node, &pci_irq_domain_ops, NULL); node 164 arch/powerpc/sysdev/xics/ics-opal.c static int ics_opal_host_match(struct ics *ics, struct device_node *node) node 30 arch/powerpc/sysdev/xics/ics-rtas.c static int ics_rtas_host_match(struct ics *ics, struct device_node *node); node 207 arch/powerpc/sysdev/xics/ics-rtas.c static int ics_rtas_host_match(struct ics *ics, struct device_node *node) node 213 arch/powerpc/sysdev/xics/ics-rtas.c return !of_device_is_compatible(node, "chrp,iic"); node 307 arch/powerpc/sysdev/xics/xics-common.c static int xics_host_match(struct irq_domain *h, struct device_node *node, node 313 arch/powerpc/sysdev/xics/xics-common.c if (ics->host_match(ics, node)) node 1293 arch/powerpc/sysdev/xive/common.c static int xive_irq_domain_match(struct irq_domain *h, struct device_node *node, node 1296 arch/powerpc/sysdev/xive/common.c return xive_ops->match(node); node 243 arch/powerpc/sysdev/xive/native.c static bool xive_native_match(struct device_node *node) node 245 arch/powerpc/sysdev/xive/native.c return of_device_is_compatible(node, "ibm,opal-xive-vc"); node 541 arch/powerpc/sysdev/xive/spapr.c static bool xive_spapr_match(struct device_node *node) node 4072 arch/powerpc/xmon/xmon.c DUMP_FIELD(spu, "0x%x", node); node 77 arch/riscv/include/asm/processor.h int riscv_of_processor_hartid(struct device_node *node); node 12 arch/riscv/kernel/cacheinfo.c struct device_node *node, node 15 arch/riscv/kernel/cpu.c int riscv_of_processor_hartid(struct device_node *node) node 20 arch/riscv/kernel/cpu.c if (!of_device_is_compatible(node, "riscv")) { node 25 arch/riscv/kernel/cpu.c if (of_property_read_u32(node, "reg", &hart)) { node 30 arch/riscv/kernel/cpu.c if (!of_device_is_available(node)) { node 35 arch/riscv/kernel/cpu.c if (of_property_read_string(node, "riscv,isa", &isa)) { node 131 arch/riscv/kernel/cpu.c struct device_node *node = of_get_cpu_node(cpu_id, NULL); node 136 arch/riscv/kernel/cpu.c if (!of_property_read_string(node, "riscv,isa", &isa)) node 138 arch/riscv/kernel/cpu.c if (!of_property_read_string(node, "mmu-type", &mmu)) node 140 arch/riscv/kernel/cpu.c if (!of_property_read_string(node, "compatible", &compat) node 144 arch/riscv/kernel/cpu.c of_node_put(node); node 22 arch/riscv/kernel/cpufeature.c struct device_node *node; node 36 arch/riscv/kernel/cpufeature.c for_each_of_cpu_node(node) { node 39 arch/riscv/kernel/cpufeature.c if (riscv_of_processor_hartid(node) < 0) node 42 arch/riscv/kernel/cpufeature.c if (of_property_read_string(node, "riscv,isa", &isa)) { node 469 arch/riscv/kernel/perf_event.c struct device_node *node = of_find_node_by_type(NULL, "pmu"); node 474 arch/riscv/kernel/perf_event.c if (node) { node 475 arch/riscv/kernel/perf_event.c of_id = of_match_node(riscv_pmu_of_ids, node); node 479 arch/riscv/kernel/perf_event.c of_node_put(node); node 464 arch/riscv/mm/init.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 467 arch/riscv/mm/init.c return vmemmap_populate_basepages(start, end, node); node 79 arch/s390/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node) node 81 arch/s390/include/asm/topology.h return &node_to_cpumask_map[node]; node 62 arch/s390/lib/spinlock.c struct spin_wait *node; node 65 arch/s390/lib/spinlock.c node = per_cpu_ptr(&spin_wait[0], cpu); node 66 arch/s390/lib/spinlock.c for (ix = 0; ix < 4; ix++, node++) { node 67 arch/s390/lib/spinlock.c memset(node, 0, sizeof(*node)); node 68 arch/s390/lib/spinlock.c node->node_id = ((cpu + 1) << _Q_TAIL_CPU_OFFSET) + node 106 arch/s390/lib/spinlock.c static inline int arch_spin_yield_target(int lock, struct spin_wait *node) node 110 arch/s390/lib/spinlock.c if (node == NULL || node->prev == NULL) node 112 arch/s390/lib/spinlock.c while (node->prev) node 113 arch/s390/lib/spinlock.c node = node->prev; node 114 arch/s390/lib/spinlock.c return node->node_id >> _Q_TAIL_CPU_OFFSET; node 119 arch/s390/lib/spinlock.c struct spin_wait *node, *next; node 125 arch/s390/lib/spinlock.c node = this_cpu_ptr(&spin_wait[ix]); node 126 arch/s390/lib/spinlock.c node->prev = node->next = NULL; node 127 arch/s390/lib/spinlock.c node_id = node->node_id; node 156 arch/s390/lib/spinlock.c node->prev = arch_spin_decode_tail(tail_id); node 157 arch/s390/lib/spinlock.c WRITE_ONCE(node->prev->next, node); node 161 arch/s390/lib/spinlock.c owner = arch_spin_yield_target(old, node); node 168 arch/s390/lib/spinlock.c while (READ_ONCE(node->prev) != NULL) { node 173 arch/s390/lib/spinlock.c owner = arch_spin_yield_target(old, node); node 202 arch/s390/lib/spinlock.c while ((next = READ_ONCE(node->next)) == NULL) node 213 arch/s390/mm/vmem.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 267 arch/s390/mm/vmem.c new_page = vmemmap_alloc_block(PMD_SIZE, node); node 287 arch/s390/mm/vmem.c new_page = vmemmap_alloc_block(PAGE_SIZE, node); node 86 arch/s390/numa/mode_emu.c static int cores_pinned(struct toptree *node) node 88 arch/s390/numa/mode_emu.c return emu_cores->per_node[node->id]; node 164 arch/s390/numa/mode_emu.c static int dist_node_to_core(struct toptree *node, struct toptree *core) node 169 arch/s390/numa/mode_emu.c toptree_for_each(core_node, node, CORE) node 193 arch/s390/numa/mode_emu.c struct toptree *node, *node_best = NULL; node 199 arch/s390/numa/mode_emu.c toptree_for_each(node, numa, NODE) { node 201 arch/s390/numa/mode_emu.c if (core_pinned_to_node_id(core) == node->id) { node 202 arch/s390/numa/mode_emu.c node_best = node; node 206 arch/s390/numa/mode_emu.c if (cores_pinned(node) >= cores_target) node 208 arch/s390/numa/mode_emu.c dist_cur = dist_node_to_core(node, core); node 211 arch/s390/numa/mode_emu.c node_best = node; node 223 arch/s390/numa/mode_emu.c struct toptree *node, *core, *tmp; node 226 arch/s390/numa/mode_emu.c node = node_for_core(numa, core, extra); node 227 arch/s390/numa/mode_emu.c if (!node) node 229 arch/s390/numa/mode_emu.c toptree_move(core, node); node 230 arch/s390/numa/mode_emu.c pin_core_to_node(core->id, node->id); node 237 arch/s390/numa/mode_emu.c static void move_level_to_numa_node(struct toptree *node, struct toptree *phys, node 244 arch/s390/numa/mode_emu.c cores_free = cores_target - toptree_count(node, CORE); node 247 arch/s390/numa/mode_emu.c toptree_move(cur, node); node 250 arch/s390/numa/mode_emu.c toptree_move(cur, node); node 263 arch/s390/numa/mode_emu.c struct toptree *node; node 265 arch/s390/numa/mode_emu.c toptree_for_each(node, numa, NODE) node 266 arch/s390/numa/mode_emu.c move_level_to_numa_node(node, phys, level, perfect); node 353 arch/s390/numa/mode_emu.c struct toptree *phys, *node, *drawer, *book, *mc, *core; node 361 arch/s390/numa/mode_emu.c node = toptree_get_child(phys, 0); node 362 arch/s390/numa/mode_emu.c drawer = toptree_get_child(node, top->drawer_id); node 248 arch/sh/include/asm/dwarf.h struct rb_node node; node 266 arch/sh/include/asm/dwarf.h struct rb_node node; node 9 arch/sh/include/asm/topology.h #define cpumask_of_node(node) ((void)node, cpu_online_mask) node 18 arch/sh/kernel/cpu/sh2/probe.c static int __init scan_cache(unsigned long node, const char *uname, node 21 arch/sh/kernel/cpu/sh2/probe.c if (!of_flat_dt_is_compatible(node, "jcore,cache")) node 24 arch/sh/kernel/cpu/sh2/probe.c j2_ccr_base = (u32 __iomem *)of_flat_dt_translate_address(node); node 323 arch/sh/kernel/dwarf.c cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node); node 359 arch/sh/kernel/dwarf.c fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node); node 840 arch/sh/kernel/dwarf.c cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node); node 852 arch/sh/kernel/dwarf.c rb_link_node(&cie->node, parent, rb_node); node 853 arch/sh/kernel/dwarf.c rb_insert_color(&cie->node, &cie_root); node 926 arch/sh/kernel/dwarf.c fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node); node 944 arch/sh/kernel/dwarf.c rb_link_node(&fde->node, parent, rb_node); node 945 arch/sh/kernel/dwarf.c rb_insert_color(&fde->node, &fde_root); node 1004 arch/sh/kernel/dwarf.c rbtree_postorder_for_each_entry_safe(fde, next_fde, &fde_root, node) node 1007 arch/sh/kernel/dwarf.c rbtree_postorder_for_each_entry_safe(cie, next_cie, &cie_root, node) node 1140 arch/sh/kernel/dwarf.c rb_erase(&cie->node, &cie_root); node 1150 arch/sh/kernel/dwarf.c rb_erase(&fde->node, &fde_root); node 21 arch/sparc/include/asm/fb.h struct device_node *node; node 26 arch/sparc/include/asm/fb.h node = dev->of_node; node 27 arch/sparc/include/asm/fb.h if (node && node 28 arch/sparc/include/asm/fb.h node == of_console_device) node 40 arch/sparc/include/asm/mdesc.h u64 node, const char *name, int *lenp); node 41 arch/sparc/include/asm/mdesc.h const char *mdesc_node_name(struct mdesc_handle *hp, u64 node); node 67 arch/sparc/include/asm/mdesc.h void (*add)(struct mdesc_handle *handle, u64 node, node 69 arch/sparc/include/asm/mdesc.h void (*remove)(struct mdesc_handle *handle, u64 node, node 90 arch/sparc/include/asm/mdesc.h int mdesc_get_node_info(struct mdesc_handle *hp, u64 node, node 174 arch/sparc/include/asm/openprom.h phandle (*no_nextnode)(phandle node); node 175 arch/sparc/include/asm/openprom.h phandle (*no_child)(phandle node); node 176 arch/sparc/include/asm/openprom.h int (*no_proplen)(phandle node, const char *name); node 177 arch/sparc/include/asm/openprom.h int (*no_getprop)(phandle node, const char *name, char *val); node 178 arch/sparc/include/asm/openprom.h int (*no_setprop)(phandle node, const char *name, char *val, int len); node 179 arch/sparc/include/asm/openprom.h char * (*no_nextprop)(phandle node, char *name); node 120 arch/sparc/include/asm/oplib_32.h phandle prom_getsibling(phandle node); node 134 arch/sparc/include/asm/oplib_32.h int prom_getint(phandle node, char *property); node 137 arch/sparc/include/asm/oplib_32.h int prom_getintdefault(phandle node, char *property, int defval); node 140 arch/sparc/include/asm/oplib_32.h int prom_getbool(phandle node, char *prop); node 143 arch/sparc/include/asm/oplib_32.h void prom_getstring(phandle node, char *prop, char *buf, int bufsize); node 153 arch/sparc/include/asm/oplib_32.h char *prom_nextprop(phandle node, char *prev_property, char *buffer); node 161 arch/sparc/include/asm/oplib_32.h int prom_setprop(phandle node, const char *prop_name, char *prop_value, node 172 arch/sparc/include/asm/oplib_32.h void prom_apply_generic_ranges(phandle node, phandle parent, node 188 arch/sparc/include/asm/oplib_64.h phandle prom_getsibling(phandle node); node 202 arch/sparc/include/asm/oplib_64.h int prom_getint(phandle node, const char *property); node 205 arch/sparc/include/asm/oplib_64.h int prom_getintdefault(phandle node, const char *property, int defval); node 208 arch/sparc/include/asm/oplib_64.h int prom_getbool(phandle node, const char *prop); node 211 arch/sparc/include/asm/oplib_64.h void prom_getstring(phandle node, const char *prop, char *buf, node 225 arch/sparc/include/asm/oplib_64.h char *prom_firstprop(phandle node, char *buffer); node 230 arch/sparc/include/asm/oplib_64.h char *prom_nextprop(phandle node, const char *prev_property, char *buf); node 233 arch/sparc/include/asm/oplib_64.h int prom_node_has_property(phandle node, const char *property); node 241 arch/sparc/include/asm/oplib_64.h int prom_setprop(phandle node, const char *prop_name, char *prop_value, node 33 arch/sparc/include/asm/prom.h int of_set_property(struct device_node *node, const char *name, void *val, int len); node 14 arch/sparc/include/asm/topology_64.h #define cpumask_of_node(node) ((node) == -1 ? \ node 16 arch/sparc/include/asm/topology_64.h &numa_cpumask_lookup_table[node]) node 362 arch/sparc/include/asm/vio.h struct list_head node; node 32 arch/sparc/kernel/auxio_32.c phandle node, auxio_nd; node 43 arch/sparc/kernel/auxio_32.c node = prom_getchild(prom_root_node); node 44 arch/sparc/kernel/auxio_32.c auxio_nd = prom_searchsiblings(node, "auxiliary-io"); node 46 arch/sparc/kernel/auxio_32.c node = prom_searchsiblings(node, "obio"); node 47 arch/sparc/kernel/auxio_32.c node = prom_getchild(node); node 48 arch/sparc/kernel/auxio_32.c auxio_nd = prom_searchsiblings(node, "auxio"); node 54 arch/sparc/kernel/auxio_32.c if(prom_searchsiblings(node, "leds")) { node 115 arch/sparc/kernel/auxio_32.c phandle node; node 119 arch/sparc/kernel/auxio_32.c node = prom_getchild(prom_root_node); node 120 arch/sparc/kernel/auxio_32.c node = prom_searchsiblings(node, "obio"); node 121 arch/sparc/kernel/auxio_32.c node = prom_getchild(node); node 122 arch/sparc/kernel/auxio_32.c node = prom_searchsiblings(node, "power"); node 123 arch/sparc/kernel/auxio_32.c if (node == 0 || (s32)node == -1) node 127 arch/sparc/kernel/auxio_32.c if (prom_getproperty(node, "reg", (char *)®s, sizeof(regs)) <= 0) node 43 arch/sparc/kernel/btext.c static int __init btext_initialize(phandle node) node 49 arch/sparc/kernel/btext.c if (prom_getproperty(node, "width", (char *)&width, 4) < 0) node 51 arch/sparc/kernel/btext.c if (prom_getproperty(node, "height", (char *)&height, 4) < 0) node 53 arch/sparc/kernel/btext.c if (prom_getproperty(node, "depth", (char *)&depth, 4) < 0) node 57 arch/sparc/kernel/btext.c if (prom_getproperty(node, "linebytes", (char *)&prop, 4) >= 0 && node 64 arch/sparc/kernel/btext.c if (prom_getproperty(node, "address", (char *)&prop, 4) >= 0) node 312 arch/sparc/kernel/btext.c phandle node; node 316 arch/sparc/kernel/btext.c node = prom_inst2pkg(prom_stdout); node 317 arch/sparc/kernel/btext.c if (prom_getproperty(node, "device_type", type, 32) < 0) node 322 arch/sparc/kernel/btext.c ret = btext_initialize(node); node 188 arch/sparc/kernel/cpumap.c struct cpuinfo_node *node; node 211 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[n]; node 218 arch/sparc/kernel/cpumap.c node->id = id; node 219 arch/sparc/kernel/cpumap.c node->level = level; node 220 arch/sparc/kernel/cpumap.c node->num_cpus = 1; node 222 arch/sparc/kernel/cpumap.c node->parent_index = (level > CPUINFO_LVL_ROOT) node 225 arch/sparc/kernel/cpumap.c node->child_start = node->child_end = node->rover = node 229 arch/sparc/kernel/cpumap.c prev_id[level] = node->id; node 252 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[level_rover[level]]; node 253 arch/sparc/kernel/cpumap.c node->num_cpus = num_cpus[level]; node 257 arch/sparc/kernel/cpumap.c node->num_cpus++; node 261 arch/sparc/kernel/cpumap.c node->parent_index = -1; node 263 arch/sparc/kernel/cpumap.c node->parent_index = node 267 arch/sparc/kernel/cpumap.c node->child_end = node 270 arch/sparc/kernel/cpumap.c node->child_end = node 277 arch/sparc/kernel/cpumap.c node = &new_tree->nodes[n]; node 278 arch/sparc/kernel/cpumap.c node->id = id; node 279 arch/sparc/kernel/cpumap.c node->level = level; node 282 arch/sparc/kernel/cpumap.c node->child_start = node->child_end = node 283 arch/sparc/kernel/cpumap.c node->rover = node 299 arch/sparc/kernel/cpumap.c struct cpuinfo_node *node = &t->nodes[node_index]; node 303 arch/sparc/kernel/cpumap.c for (level = node->level; level >= top_level; level--) { node 304 arch/sparc/kernel/cpumap.c node->rover++; node 305 arch/sparc/kernel/cpumap.c if (node->rover <= node->child_end) node 308 arch/sparc/kernel/cpumap.c node->rover = node->child_start; node 314 arch/sparc/kernel/cpumap.c node = &t->nodes[node->parent_index]; node 92 arch/sparc/kernel/mdesc.c static int get_vdev_port_node_info(struct mdesc_handle *md, u64 node, node 98 arch/sparc/kernel/mdesc.c static int get_ds_port_node_info(struct mdesc_handle *md, u64 node, node 290 arch/sparc/kernel/mdesc.c u64 node; node 312 arch/sparc/kernel/mdesc.c mdesc_for_each_node_by_name(cur_mdesc, node, client->node_name) node 313 arch/sparc/kernel/mdesc.c client->add(cur_mdesc, node, client->node_name); node 318 arch/sparc/kernel/mdesc.c static const u64 *parent_cfg_handle(struct mdesc_handle *hp, u64 node) node 324 arch/sparc/kernel/mdesc.c mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) { node 337 arch/sparc/kernel/mdesc.c static int get_vdev_port_node_info(struct mdesc_handle *md, u64 node, node 350 arch/sparc/kernel/mdesc.c idp = mdesc_get_property(md, node, "id", NULL); node 351 arch/sparc/kernel/mdesc.c name = mdesc_get_property(md, node, "name", NULL); node 352 arch/sparc/kernel/mdesc.c parent_cfg_hdlp = parent_cfg_handle(md, node); node 391 arch/sparc/kernel/mdesc.c static int get_ds_port_node_info(struct mdesc_handle *md, u64 node, node 397 arch/sparc/kernel/mdesc.c idp = mdesc_get_property(md, node, "id", NULL); node 583 arch/sparc/kernel/mdesc.c int mdesc_get_node_info(struct mdesc_handle *hp, u64 node, node 589 arch/sparc/kernel/mdesc.c if (hp == NULL || node == MDESC_NODE_NULL || node 602 arch/sparc/kernel/mdesc.c rv = get_info_func(hp, node, node_info); node 657 arch/sparc/kernel/mdesc.c const void *mdesc_get_property(struct mdesc_handle *hp, u64 node, node 665 arch/sparc/kernel/mdesc.c if (node == MDESC_NODE_NULL || node >= last_node) node 668 arch/sparc/kernel/mdesc.c ep = node_block(&hp->mdesc) + node; node 739 arch/sparc/kernel/mdesc.c const char *mdesc_node_name(struct mdesc_handle *hp, u64 node) node 745 arch/sparc/kernel/mdesc.c if (node == MDESC_NODE_NULL || node >= last_node) node 748 arch/sparc/kernel/mdesc.c ep = base + node; node 860 arch/sparc/kernel/mdesc.c static void find_back_node_value(struct mdesc_handle *hp, u64 node, node 871 arch/sparc/kernel/mdesc.c mdesc_for_each_arc(arc, hp, node, MDESC_ARC_TYPE_BACK) { node 882 arch/sparc/kernel/mdesc.c static void __mark_core_id(struct mdesc_handle *hp, u64 node, node 885 arch/sparc/kernel/mdesc.c const u64 *id = mdesc_get_property(hp, node, "id", NULL); node 891 arch/sparc/kernel/mdesc.c static void __mark_max_cache_id(struct mdesc_handle *hp, u64 node, node 894 arch/sparc/kernel/mdesc.c const u64 *id = mdesc_get_property(hp, node, "id", NULL); node 16 arch/sparc/kernel/of_device_common.c unsigned int irq_of_parse_and_map(struct device_node *node, int index) node 18 arch/sparc/kernel/of_device_common.c struct platform_device *op = of_find_device_by_node(node); node 27 arch/sparc/kernel/of_device_common.c int of_address_to_resource(struct device_node *node, int index, node 30 arch/sparc/kernel/of_device_common.c struct platform_device *op = of_find_device_by_node(node); node 40 arch/sparc/kernel/of_device_common.c void __iomem *of_iomap(struct device_node *node, int index) node 42 arch/sparc/kernel/of_device_common.c struct platform_device *op = of_find_device_by_node(node); node 206 arch/sparc/kernel/pci.c struct device_node *node, node 213 arch/sparc/kernel/pci.c addrs = of_get_property(node, "assigned-addresses", &proplen); node 264 arch/sparc/kernel/pci.c struct device_node *node, node 276 arch/sparc/kernel/pci.c op = of_find_device_by_node(node); node 285 arch/sparc/kernel/pci.c if (of_node_name_eq(node, "ebus")) node 290 arch/sparc/kernel/pci.c devfn, of_node_get_device_type(node)); node 292 arch/sparc/kernel/pci.c dev->sysdata = node; node 295 arch/sparc/kernel/pci.c dev->dev.of_node = of_node_get(node); node 301 arch/sparc/kernel/pci.c dev->vendor = of_getintprop_default(node, "vendor-id", 0xffff); node 302 arch/sparc/kernel/pci.c dev->device = of_getintprop_default(node, "device-id", 0xffff); node 304 arch/sparc/kernel/pci.c of_getintprop_default(node, "subsystem-vendor-id", 0); node 306 arch/sparc/kernel/pci.c of_getintprop_default(node, "subsystem-id", 0); node 334 arch/sparc/kernel/pci.c if (of_node_name_eq(node, "pci")) { node 338 arch/sparc/kernel/pci.c } else if (of_node_is_type(node, "cardbus")) { node 352 arch/sparc/kernel/pci.c pci_parse_of_addrs(sd->op, node, dev); node 411 arch/sparc/kernel/pci.c struct device_node *node, node 417 arch/sparc/kernel/pci.c struct device_node *node, node 429 arch/sparc/kernel/pci.c pci_info(dev, "of_scan_pci_bridge(%pOF)\n", node); node 432 arch/sparc/kernel/pci.c busrange = of_get_property(node, "bus-range", &len); node 435 arch/sparc/kernel/pci.c node); node 443 arch/sparc/kernel/pci.c ranges = of_get_property(node, "ranges", &len); node 446 arch/sparc/kernel/pci.c const char *model = of_get_property(node, "model", NULL); node 454 arch/sparc/kernel/pci.c node); node 510 arch/sparc/kernel/pci.c " for bridge %pOF\n", node); node 516 arch/sparc/kernel/pci.c " for bridge %pOF\n", node); node 539 arch/sparc/kernel/pci.c pci_of_scan_bus(pbm, node, bus); node 543 arch/sparc/kernel/pci.c struct device_node *node, node 553 arch/sparc/kernel/pci.c node, bus->number); node 557 arch/sparc/kernel/pci.c while ((child = of_get_next_child(node, child)) != NULL) { node 618 arch/sparc/kernel/pci.c list_for_each_entry(child_bus, &bus->children, node) node 685 arch/sparc/kernel/pci.c list_for_each_entry(child_bus, &bus->children, node) node 693 arch/sparc/kernel/pci.c struct device_node *node = pbm->op->dev.of_node; node 696 arch/sparc/kernel/pci.c printk("PCI: Scanning PBM %pOF\n", node); node 712 arch/sparc/kernel/pci.c printk(KERN_ERR "Failed to create bus for %pOF\n", node); node 717 arch/sparc/kernel/pci.c pci_of_scan_bus(pbm, node, bus); node 1070 arch/sparc/kernel/pci.c list_for_each_entry(bus, &pbus->children, node) node 1074 arch/sparc/kernel/pci.c static void pci_bus_slot_names(struct device_node *node, struct pci_bus *bus) node 1084 arch/sparc/kernel/pci.c prop = of_get_property(node, "slot-names", &len); node 1093 arch/sparc/kernel/pci.c node, mask); node 1124 arch/sparc/kernel/pci.c struct device_node *node; node 1137 arch/sparc/kernel/pci.c node = pbus->self->dev.of_node; node 1143 arch/sparc/kernel/pci.c node = pbm->op->dev.of_node; node 1146 arch/sparc/kernel/pci.c pci_bus_slot_names(node, pbus); node 497 arch/sparc/kernel/pci_common.c list_for_each_entry(bus, &pbus->children, node) node 520 arch/sparc/kernel/pci_common.c list_for_each_entry(bus, &pbus->children, node) node 544 arch/sparc/kernel/pci_common.c list_for_each_entry(bus, &pbus->children, node) node 293 arch/sparc/kernel/pcic.c phandle node; node 302 arch/sparc/kernel/pcic.c node = prom_getchild (prom_root_node); node 303 arch/sparc/kernel/pcic.c node = prom_searchsiblings (node, "pci"); node 304 arch/sparc/kernel/pcic.c if (node == 0) node 309 arch/sparc/kernel/pcic.c err = prom_getproperty(node, "reg", (char*)regs, sizeof(regs)); node 353 arch/sparc/kernel/pcic.c pbm->prom_node = node; node 354 arch/sparc/kernel/pcic.c prom_getstring(node, "name", namebuf, 63); namebuf[63] = 0; node 451 arch/sparc/kernel/pcic.c phandle node = prom_getchild(pbm->prom_node); node 453 arch/sparc/kernel/pcic.c while(node) { node 454 arch/sparc/kernel/pcic.c err = prom_getproperty(node, "reg", node 459 arch/sparc/kernel/pcic.c return node; node 461 arch/sparc/kernel/pcic.c node = prom_getsibling(node); node 472 arch/sparc/kernel/pcic.c struct pci_dev *dev, int node) node 479 arch/sparc/kernel/pcic.c if (node == 0 || node == -1) { node 482 arch/sparc/kernel/pcic.c prom_getstring(node, "name", namebuf, 63); namebuf[63] = 0; node 531 arch/sparc/kernel/pcic.c pcic_fill_irq(struct linux_pcic *pcic, struct pci_dev *dev, int node) node 538 arch/sparc/kernel/pcic.c if (node == 0 || node == -1) { node 541 arch/sparc/kernel/pcic.c prom_getstring(node, "name", namebuf, sizeof(namebuf)); node 607 arch/sparc/kernel/pcic.c int node; node 626 arch/sparc/kernel/pcic.c node = pdev_to_pnode(&pcic->pbm, dev); node 627 arch/sparc/kernel/pcic.c if(node == 0) node 628 arch/sparc/kernel/pcic.c node = -1; node 633 arch/sparc/kernel/pcic.c pcp->prom_node = of_find_node_by_phandle(node); node 638 arch/sparc/kernel/pcic.c pcic_map_pci_device(pcic, dev, node); node 640 arch/sparc/kernel/pcic.c pcic_fill_irq(pcic, dev, node); node 222 arch/sparc/kernel/prom_32.c phandle node; node 273 arch/sparc/kernel/prom_32.c node = (*romvec->pv_v2devops.v2_inst2pkg)(fd); node 277 arch/sparc/kernel/prom_32.c if (!node) { node 282 arch/sparc/kernel/prom_32.c dp = of_find_node_by_phandle(node); node 601 arch/sparc/kernel/prom_64.c phandle node; node 615 arch/sparc/kernel/prom_64.c node = prom_inst2pkg(prom_stdout); node 616 arch/sparc/kernel/prom_64.c if (!node) { node 622 arch/sparc/kernel/prom_64.c dp = of_find_node_by_phandle(node); node 132 arch/sparc/kernel/prom_common.c static int __init prom_common_nextprop(phandle node, char *prev, char *buf) node 137 arch/sparc/kernel/prom_common.c name = prom_nextprop(node, prev, buf); node 620 arch/sparc/kernel/setup_64.c unsigned int i, node; node 623 arch/sparc/kernel/setup_64.c node = cpu_to_node(i); node 626 arch/sparc/kernel/setup_64.c THREAD_SIZE, node); node 629 arch/sparc/kernel/setup_64.c __func__, THREAD_SIZE, THREAD_SIZE, node); node 631 arch/sparc/kernel/setup_64.c THREAD_SIZE, node); node 634 arch/sparc/kernel/setup_64.c __func__, THREAD_SIZE, THREAD_SIZE, node); node 1587 arch/sparc/kernel/smp_64.c int node = cpu_to_node(cpu); node 1590 arch/sparc/kernel/smp_64.c if (!node_online(node) || !NODE_DATA(node)) { node 1593 arch/sparc/kernel/smp_64.c cpu, node); node 1598 arch/sparc/kernel/smp_64.c MEMBLOCK_ALLOC_ACCESSIBLE, node); node 1600 arch/sparc/kernel/smp_64.c "%016lx\n", cpu, size, node, __pa(ptr)); node 166 arch/sparc/kernel/time_64.c static unsigned long cpuid_to_freq(phandle node, int cpuid) node 172 arch/sparc/kernel/time_64.c if (!node) node 175 arch/sparc/kernel/time_64.c if (prom_getproperty(node, "device_type", type, sizeof(type)) != -1) node 179 arch/sparc/kernel/time_64.c if (is_cpu_node && (prom_getint(node, "upa-portid") == cpuid || node 180 arch/sparc/kernel/time_64.c prom_getint(node, "cpuid") == cpuid)) node 181 arch/sparc/kernel/time_64.c freq = prom_getintdefault(node, "clock-frequency", 0); node 183 arch/sparc/kernel/time_64.c freq = cpuid_to_freq(prom_getchild(node), cpuid); node 185 arch/sparc/kernel/time_64.c freq = cpuid_to_freq(prom_getsibling(node), cpuid); node 207 arch/sparc/kernel/vio.c static const u64 *vio_cfg_handle(struct mdesc_handle *hp, u64 node) node 212 arch/sparc/kernel/vio.c mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) { node 240 arch/sparc/kernel/vio.c u64 node; node 245 arch/sparc/kernel/vio.c node = mdesc_get_node(hp, (const char *)vdev->node_name, node 248 arch/sparc/kernel/vio.c return node; node 413 arch/sparc/kernel/vio.c static void vio_add(struct mdesc_handle *hp, u64 node, node 416 arch/sparc/kernel/vio.c (void) vio_create_one(hp, node, node_name, &root_vdev->dev); node 421 arch/sparc/kernel/vio.c u64 node; node 428 arch/sparc/kernel/vio.c u64 node; node 432 arch/sparc/kernel/vio.c node = vio_vdev_node(node_data->hp, vdev); node 434 arch/sparc/kernel/vio.c if (node == node_data->node) node 440 arch/sparc/kernel/vio.c static void vio_remove(struct mdesc_handle *hp, u64 node, const char *node_name) node 446 arch/sparc/kernel/vio.c node_data.node = node; node 471 arch/sparc/kernel/vio.c static void vio_add_ds(struct mdesc_handle *hp, u64 node, node 478 arch/sparc/kernel/vio.c mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) { node 489 arch/sparc/kernel/vio.c (void) vio_create_one(hp, node, node_name, &root_vdev->dev); node 113 arch/sparc/mm/init_64.c phandle node = prom_finddevice("/memory"); node 114 arch/sparc/mm/init_64.c int prop_size = prom_getproplen(node, property); node 125 arch/sparc/mm/init_64.c ret = prom_getproperty(node, property, (char *) regs, prop_size); node 614 arch/sparc/mm/init_64.c int n, node, ents, first, last, i; node 616 arch/sparc/mm/init_64.c node = prom_finddevice("/virtual-memory"); node 617 arch/sparc/mm/init_64.c n = prom_getproplen(node, "translations"); node 627 arch/sparc/mm/init_64.c if ((n = prom_getproperty(node, "translations", node 938 arch/sparc/mm/init_64.c u64 node; node 1259 arch/sparc/mm/init_64.c u64 node; node 1261 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "memory-latency-group") node 1275 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "memory-latency-group") { node 1279 arch/sparc/mm/init_64.c m->node = node; node 1281 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, "latency", NULL); node 1283 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, "address-match", NULL); node 1285 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, "address-mask", NULL); node 1290 arch/sparc/mm/init_64.c count - 1, m->node, m->latency, m->match, m->mask); node 1300 arch/sparc/mm/init_64.c u64 node; node 1302 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "mblock") node 1316 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "mblock") { node 1320 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, "base", NULL); node 1322 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, "size", NULL); node 1324 arch/sparc/mm/init_64.c val = mdesc_get_property(md, node, node 1362 arch/sparc/mm/init_64.c static struct mdesc_mlgroup * __init find_mlgroup(u64 node) node 1368 arch/sparc/mm/init_64.c if (m->node == node) node 1481 arch/sparc/mm/init_64.c u64 node; node 1483 arch/sparc/mm/init_64.c node = mdesc_node_by_name(md, MDESC_NODE_NULL, "latency-groups"); node 1484 arch/sparc/mm/init_64.c if (node == MDESC_NODE_NULL) { node 1498 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "group") { node 1499 arch/sparc/mm/init_64.c err = numa_parse_mdesc_group(md, node, count); node 1506 arch/sparc/mm/init_64.c mdesc_for_each_node_by_name(md, node, "group") { node 1507 arch/sparc/mm/init_64.c find_numa_latencies_for_group(md, node, count); node 2597 arch/sparc/mm/init_64.c int node, struct vmem_altmap *altmap) node 2613 arch/sparc/mm/init_64.c pgd_t *pgd = vmemmap_pgd_populate(vstart, node); node 2621 arch/sparc/mm/init_64.c pud = vmemmap_pud_populate(pgd, vstart, node); node 2628 arch/sparc/mm/init_64.c void *block = vmemmap_alloc_block(PMD_SIZE, node); node 34 arch/sparc/prom/init_64.c phandle node; node 44 arch/sparc/prom/init_64.c node = prom_finddevice("/openprom"); node 45 arch/sparc/prom/init_64.c if (!node || (s32)node == -1) node 48 arch/sparc/prom/init_64.c prom_getstring(node, "version", prom_version, sizeof(prom_version)); node 35 arch/sparc/prom/memory.c phandle node; node 38 arch/sparc/prom/memory.c node = prom_searchsiblings(prom_getchild(prom_root_node), "memory"); node 39 arch/sparc/prom/memory.c size = prom_getproperty(node, "available", (char *) reg, sizeof(reg)); node 167 arch/sparc/prom/misc_64.c phandle node; node 173 arch/sparc/prom/misc_64.c node = prom_finddevice(prom_chosen_path); node 174 arch/sparc/prom/misc_64.c ret = prom_getint(node, prom_mmu_name); node 186 arch/sparc/prom/misc_64.c phandle node; node 192 arch/sparc/prom/misc_64.c node = prom_finddevice("/chosen"); node 193 arch/sparc/prom/misc_64.c ret = prom_getint(node, "memory"); node 67 arch/sparc/prom/ranges.c phandle node, obio_node; node 73 arch/sparc/prom/ranges.c node = prom_getchild(prom_root_node); node 74 arch/sparc/prom/ranges.c obio_node = prom_searchsiblings(node, "obio"); node 88 arch/sparc/prom/ranges.c void prom_apply_generic_ranges(phandle node, phandle parent, node 95 arch/sparc/prom/ranges.c success = prom_getproperty(node, "ranges", node 24 arch/sparc/prom/tree_32.c static phandle __prom_getchild(phandle node) node 30 arch/sparc/prom/tree_32.c cnode = prom_nodeops->no_child(node); node 40 arch/sparc/prom/tree_32.c phandle prom_getchild(phandle node) node 44 arch/sparc/prom/tree_32.c if ((s32)node == -1) node 47 arch/sparc/prom/tree_32.c cnode = __prom_getchild(node); node 56 arch/sparc/prom/tree_32.c static phandle __prom_getsibling(phandle node) node 62 arch/sparc/prom/tree_32.c cnode = prom_nodeops->no_nextnode(node); node 72 arch/sparc/prom/tree_32.c phandle prom_getsibling(phandle node) node 76 arch/sparc/prom/tree_32.c if ((s32)node == -1) node 79 arch/sparc/prom/tree_32.c sibnode = __prom_getsibling(node); node 90 arch/sparc/prom/tree_32.c int prom_getproplen(phandle node, const char *prop) node 95 arch/sparc/prom/tree_32.c if((!node) || (!prop)) node 99 arch/sparc/prom/tree_32.c ret = prom_nodeops->no_proplen(node, prop); node 110 arch/sparc/prom/tree_32.c int prom_getproperty(phandle node, const char *prop, char *buffer, int bufsize) node 115 arch/sparc/prom/tree_32.c plen = prom_getproplen(node, prop); node 120 arch/sparc/prom/tree_32.c ret = prom_nodeops->no_getprop(node, prop, buffer); node 130 arch/sparc/prom/tree_32.c int prom_getint(phandle node, char *prop) node 134 arch/sparc/prom/tree_32.c if(prom_getproperty(node, prop, (char *) &intprop, sizeof(int)) != -1) node 144 arch/sparc/prom/tree_32.c int prom_getintdefault(phandle node, char *property, int deflt) node 148 arch/sparc/prom/tree_32.c retval = prom_getint(node, property); node 156 arch/sparc/prom/tree_32.c int prom_getbool(phandle node, char *prop) node 160 arch/sparc/prom/tree_32.c retval = prom_getproplen(node, prop); node 170 arch/sparc/prom/tree_32.c void prom_getstring(phandle node, char *prop, char *user_buf, int ubuf_size) node 174 arch/sparc/prom/tree_32.c len = prom_getproperty(node, prop, user_buf, ubuf_size); node 204 arch/sparc/prom/tree_32.c static char *__prom_nextprop(phandle node, char * oprop) node 210 arch/sparc/prom/tree_32.c prop = prom_nodeops->no_nextprop(node, oprop); node 221 arch/sparc/prom/tree_32.c char *prom_nextprop(phandle node, char *oprop, char *buffer) node 223 arch/sparc/prom/tree_32.c if (node == 0 || (s32)node == -1) node 226 arch/sparc/prom/tree_32.c return __prom_nextprop(node, oprop); node 234 arch/sparc/prom/tree_32.c phandle node = prom_root_node, node2; node 239 arch/sparc/prom/tree_32.c if (!*s) return node; /* path '.../' is legal */ node 240 arch/sparc/prom/tree_32.c node = prom_getchild(node); node 246 arch/sparc/prom/tree_32.c node = prom_searchsiblings(node, nbuf); node 247 arch/sparc/prom/tree_32.c if (!node) node 256 arch/sparc/prom/tree_32.c node2 = node; node 260 arch/sparc/prom/tree_32.c node = node2; node 274 arch/sparc/prom/tree_32.c return node; node 281 arch/sparc/prom/tree_32.c int prom_setprop(phandle node, const char *pname, char *value, int size) node 291 arch/sparc/prom/tree_32.c ret = prom_nodeops->no_setprop(node, pname, value, size); node 300 arch/sparc/prom/tree_32.c phandle node; node 304 arch/sparc/prom/tree_32.c node = (*romvec->pv_v2devops.v2_inst2pkg)(inst); node 307 arch/sparc/prom/tree_32.c if ((s32)node == -1) node 309 arch/sparc/prom/tree_32.c return node; node 20 arch/sparc/prom/tree_64.c static phandle prom_node_to_node(const char *type, phandle node) node 27 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 38 arch/sparc/prom/tree_64.c inline phandle __prom_getchild(phandle node) node 40 arch/sparc/prom/tree_64.c return prom_node_to_node("child", node); node 43 arch/sparc/prom/tree_64.c phandle prom_getchild(phandle node) node 47 arch/sparc/prom/tree_64.c if ((s32)node == -1) node 49 arch/sparc/prom/tree_64.c cnode = __prom_getchild(node); node 56 arch/sparc/prom/tree_64.c inline phandle prom_getparent(phandle node) node 60 arch/sparc/prom/tree_64.c if ((s32)node == -1) node 62 arch/sparc/prom/tree_64.c cnode = prom_node_to_node("parent", node); node 71 arch/sparc/prom/tree_64.c inline phandle __prom_getsibling(phandle node) node 73 arch/sparc/prom/tree_64.c return prom_node_to_node(prom_peer_name, node); node 76 arch/sparc/prom/tree_64.c phandle prom_getsibling(phandle node) node 80 arch/sparc/prom/tree_64.c if ((s32)node == -1) node 82 arch/sparc/prom/tree_64.c sibnode = __prom_getsibling(node); node 93 arch/sparc/prom/tree_64.c int prom_getproplen(phandle node, const char *prop) node 97 arch/sparc/prom/tree_64.c if (!node || !prop) node 103 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 117 arch/sparc/prom/tree_64.c int prom_getproperty(phandle node, const char *prop, node 123 arch/sparc/prom/tree_64.c plen = prom_getproplen(node, prop); node 130 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 145 arch/sparc/prom/tree_64.c int prom_getint(phandle node, const char *prop) node 149 arch/sparc/prom/tree_64.c if (prom_getproperty(node, prop, (char *) &intprop, sizeof(int)) != -1) node 160 arch/sparc/prom/tree_64.c int prom_getintdefault(phandle node, const char *property, int deflt) node 164 arch/sparc/prom/tree_64.c retval = prom_getint(node, property); node 173 arch/sparc/prom/tree_64.c int prom_getbool(phandle node, const char *prop) node 177 arch/sparc/prom/tree_64.c retval = prom_getproplen(node, prop); node 188 arch/sparc/prom/tree_64.c void prom_getstring(phandle node, const char *prop, char *user_buf, node 193 arch/sparc/prom/tree_64.c len = prom_getproperty(node, prop, user_buf, ubuf_size); node 203 arch/sparc/prom/tree_64.c int prom_nodematch(phandle node, const char *name) node 206 arch/sparc/prom/tree_64.c prom_getproperty(node, "name", namebuf, sizeof(namebuf)); node 239 arch/sparc/prom/tree_64.c char *prom_firstprop(phandle node, char *buffer) node 244 arch/sparc/prom/tree_64.c if ((s32)node == -1) node 250 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 265 arch/sparc/prom/tree_64.c char *prom_nextprop(phandle node, const char *oprop, char *buffer) node 270 arch/sparc/prom/tree_64.c if ((s32)node == -1) { node 282 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 311 arch/sparc/prom/tree_64.c int prom_node_has_property(phandle node, const char *prop) node 317 arch/sparc/prom/tree_64.c prom_nextprop(node, buf, buf); node 329 arch/sparc/prom/tree_64.c prom_setprop(phandle node, const char *pname, char *value, int size) node 347 arch/sparc/prom/tree_64.c args[3] = (unsigned int) node; node 362 arch/sparc/prom/tree_64.c phandle node; node 372 arch/sparc/prom/tree_64.c node = (int) args[4]; node 373 arch/sparc/prom/tree_64.c if ((s32)node == -1) node 375 arch/sparc/prom/tree_64.c return node; node 28 arch/unicore32/kernel/clock.c struct list_head node; node 62 arch/unicore32/kernel/clock.c list_for_each_entry(p, &clocks, node) { node 217 arch/unicore32/kernel/clock.c list_add(&clk->node, &clocks); node 228 arch/unicore32/kernel/clock.c list_del(&clk->node); node 14 arch/x86/entry/vdso/vgetcpu.c __vdso_getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *unused) node 16 arch/x86/entry/vdso/vgetcpu.c vdso_read_cpunode(cpu, node); node 21 arch/x86/entry/vdso/vgetcpu.c long getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache) node 50 arch/x86/events/amd/uncore.c struct hlist_node node; node 384 arch/x86/events/amd/uncore.c hlist_add_head(&this->node, &uncore_unused_list); node 424 arch/x86/events/amd/uncore.c hlist_for_each_entry_safe(uncore, n, &uncore_unused_list, node) { node 425 arch/x86/events/amd/uncore.c hlist_del(&uncore->node); node 86 arch/x86/events/intel/bts.c int node = (cpu == -1) ? cpu : cpu_to_node(cpu); node 104 arch/x86/events/intel/bts.c buf = kzalloc_node(offsetof(struct bts_buffer, buf[nbuf]), GFP_KERNEL, node); node 321 arch/x86/events/intel/ds.c int node = cpu_to_node(cpu); node 324 arch/x86/events/intel/ds.c page = __alloc_pages_node(node, flags | __GFP_ZERO, order); node 339 arch/x86/events/intel/ds.c int max, node = cpu_to_node(cpu); node 354 arch/x86/events/intel/ds.c insn_buff = kzalloc_node(PEBS_FIXUP_SIZE, GFP_KERNEL, node); node 614 arch/x86/events/intel/pt.c int node = cpu_to_node(cpu); node 618 arch/x86/events/intel/pt.c p = alloc_pages_node(node, gfp | __GFP_ZERO, 0); node 1196 arch/x86/events/intel/pt.c int node, ret, cpu = event->cpu; node 1203 arch/x86/events/intel/pt.c node = cpu_to_node(cpu); node 1205 arch/x86/events/intel/pt.c buf = kzalloc_node(sizeof(struct pt_buffer), GFP_KERNEL, node); node 1240 arch/x86/events/intel/pt.c int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu); node 1245 arch/x86/events/intel/pt.c filters = kzalloc_node(sizeof(struct pt_filters), GFP_KERNEL, node); node 314 arch/x86/events/intel/uncore.c int node) node 321 arch/x86/events/intel/uncore.c box = kzalloc_node(size, GFP_KERNEL, node); node 26 arch/x86/include/asm/amd_nb.h extern int amd_smn_read(u16 node, u32 address, u32 *value); node 27 arch/x86/include/asm/amd_nb.h extern int amd_smn_write(u16 node, u32 address, u32 value); node 28 arch/x86/include/asm/amd_nb.h extern int amd_df_indirect_read(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo); node 85 arch/x86/include/asm/amd_nb.h struct amd_northbridge *node_to_amd_nb(int node); node 16 arch/x86/include/asm/device.h struct list_head node; node 178 arch/x86/include/asm/io_apic.h int node, int trigger, int polarity); node 489 arch/x86/include/asm/kvm_host.h struct list_head node; node 23 arch/x86/include/asm/kvm_page_track.h struct hlist_node node; node 36 arch/x86/include/asm/kvm_page_track.h int bytes, struct kvm_page_track_notifier_node *node); node 46 arch/x86/include/asm/kvm_page_track.h struct kvm_page_track_notifier_node *node); node 37 arch/x86/include/asm/numa.h static inline void set_apicid_to_node(int apicid, s16 node) node 39 arch/x86/include/asm/numa.h __apicid_to_node[apicid] = node; node 45 arch/x86/include/asm/numa.h static inline void set_apicid_to_node(int apicid, s16 node) node 60 arch/x86/include/asm/numa.h extern void numa_set_node(int cpu, int node); node 66 arch/x86/include/asm/numa.h static inline void numa_set_node(int cpu, int node) { } node 74 arch/x86/include/asm/numa.h void debug_cpumask_set_cpu(int cpu, int node, bool enable); node 19 arch/x86/include/asm/pci.h int node; /* NUMA node */ node 136 arch/x86/include/asm/pci.h return sd->node; node 142 arch/x86/include/asm/pci.h int node; node 144 arch/x86/include/asm/pci.h node = __pcibus_to_node(bus); node 145 arch/x86/include/asm/pci.h return (node == NUMA_NO_NODE) ? cpu_online_mask : node 146 arch/x86/include/asm/pci.h cpumask_of_node(node); node 249 arch/x86/include/asm/segment.h static inline unsigned long vdso_encode_cpunode(int cpu, unsigned long node) node 251 arch/x86/include/asm/segment.h return (node << VDSO_CPUNODE_BITS) | cpu; node 254 arch/x86/include/asm/segment.h static inline void vdso_read_cpunode(unsigned *cpu, unsigned *node) node 273 arch/x86/include/asm/segment.h if (node) node 274 arch/x86/include/asm/segment.h *node = (p >> VDSO_CPUNODE_BITS); node 67 arch/x86/include/asm/topology.h extern const struct cpumask *cpumask_of_node(int node); node 70 arch/x86/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node) node 72 arch/x86/include/asm/topology.h return node_to_cpumask_map[node]; node 205 arch/x86/include/asm/uv/uv_hub.h static inline struct uv_hub_info_s *uv_hub_info_list(int node) node 207 arch/x86/include/asm/uv/uv_hub.h return (struct uv_hub_info_s *)__uv_hub_info_list[node]; node 591 arch/x86/include/asm/uv/uv_hub.h unsigned short sockid, node, *p2s; node 598 arch/x86/include/asm/uv/uv_hub.h node = uv_socket_to_node(sockid); node 601 arch/x86/include/asm/uv/uv_hub.h if (!node) node 604 arch/x86/include/asm/uv/uv_hub.h base = (unsigned long)(uv_hub_info->gr_table[node - 1].limit); node 174 arch/x86/include/asm/x86_init.h void (*fixup_cpu_id)(struct cpuinfo_x86 *c, int node); node 659 arch/x86/kernel/acpi/boot.c int node; node 662 arch/x86/kernel/acpi/boot.c node = dev ? dev_to_node(dev) : NUMA_NO_NODE; node 665 arch/x86/kernel/acpi/boot.c ioapic_set_alloc_attr(&info, node, trigger, polarity); node 109 arch/x86/kernel/amd_nb.c struct amd_northbridge *node_to_amd_nb(int node) node 111 arch/x86/kernel/amd_nb.c return (node < amd_northbridges.num) ? &amd_northbridges.nb[node] : NULL; node 126 arch/x86/kernel/amd_nb.c static int __amd_smn_rw(u16 node, u32 address, u32 *value, bool write) node 131 arch/x86/kernel/amd_nb.c if (node >= amd_northbridges.num) node 134 arch/x86/kernel/amd_nb.c root = node_to_amd_nb(node)->root; node 159 arch/x86/kernel/amd_nb.c int amd_smn_read(u16 node, u32 address, u32 *value) node 161 arch/x86/kernel/amd_nb.c return __amd_smn_rw(node, address, value, false); node 165 arch/x86/kernel/amd_nb.c int amd_smn_write(u16 node, u32 address, u32 value) node 167 arch/x86/kernel/amd_nb.c return __amd_smn_rw(node, address, &value, true); node 181 arch/x86/kernel/amd_nb.c int amd_df_indirect_read(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo) node 187 arch/x86/kernel/amd_nb.c if (node >= amd_northbridges.num) node 190 arch/x86/kernel/amd_nb.c F4 = node_to_amd_nb(node)->link; node 391 arch/x86/kernel/aperture_64.c int i, node; node 406 arch/x86/kernel/aperture_64.c node = 0; node 442 arch/x86/kernel/aperture_64.c node, aper_base, aper_base + aper_size - 1, node 444 arch/x86/kernel/aperture_64.c node++; node 170 arch/x86/kernel/apic/apic_numachip.c static void fixup_cpu_id(struct cpuinfo_x86 *c, int node) node 175 arch/x86/kernel/apic/apic_numachip.c this_cpu_write(cpu_llc_id, node); node 183 arch/x86/kernel/apic/apic_numachip.c c->phys_proc_id = node / nodes; node 375 arch/x86/kernel/apic/io_apic.c int node, int apic, int pin) node 384 arch/x86/kernel/apic/io_apic.c entry = kzalloc_node(sizeof(struct irq_pin_list), GFP_ATOMIC, node); node 387 arch/x86/kernel/apic/io_apic.c node, apic, pin); node 410 arch/x86/kernel/apic/io_apic.c int node, int apic, int pin) node 412 arch/x86/kernel/apic/io_apic.c if (__add_pin_to_irq_node(data, node, apic, pin)) node 419 arch/x86/kernel/apic/io_apic.c static void __init replace_pin_at_irq_node(struct mp_chip_data *data, int node, node 435 arch/x86/kernel/apic/io_apic.c add_pin_to_irq_node(data, node, newapic, newpin); node 871 arch/x86/kernel/apic/io_apic.c void ioapic_set_alloc_attr(struct irq_alloc_info *info, int node, node 876 arch/x86/kernel/apic/io_apic.c info->ioapic_node = node; node 1007 arch/x86/kernel/apic/io_apic.c int node = ioapic_alloc_attr_node(info); node 1017 arch/x86/kernel/apic/io_apic.c if (__add_pin_to_irq_node(irq_data->chip_data, node, ioapic, node 1022 arch/x86/kernel/apic/io_apic.c irq = __irq_domain_alloc_irqs(domain, irq, 1, node, info, true, node 2131 arch/x86/kernel/apic/io_apic.c int node = cpu_to_node(0); node 2220 arch/x86/kernel/apic/io_apic.c replace_pin_at_irq_node(data, node, apic1, pin1, apic2, pin2); node 373 arch/x86/kernel/apic/msi.c int dmar_alloc_hwirq(int id, int node, void *arg) node 386 arch/x86/kernel/apic/msi.c return irq_domain_alloc_irqs(domain, 1, node, &info); node 102 arch/x86/kernel/apic/vector.c static struct apic_chip_data *alloc_apic_chip_data(int node) node 106 arch/x86/kernel/apic/vector.c apicd = kzalloc_node(sizeof(*apicd), GFP_KERNEL, node); node 273 arch/x86/kernel/apic/vector.c int node = irq_data_get_node(irqd); node 275 arch/x86/kernel/apic/vector.c if (node == NUMA_NO_NODE) node 278 arch/x86/kernel/apic/vector.c cpumask_and(vector_searchmask, cpumask_of_node(node), affmsk); node 282 arch/x86/kernel/apic/vector.c if (!assign_vector_locked(irqd, cpumask_of_node(node))) node 534 arch/x86/kernel/apic/vector.c int i, err, node; node 546 arch/x86/kernel/apic/vector.c node = irq_data_get_node(irqd); node 548 arch/x86/kernel/apic/vector.c apicd = alloc_apic_chip_data(node); node 14 arch/x86/kernel/apic/x2apic_cluster.c int node; node 124 arch/x86/kernel/apic/x2apic_cluster.c static int alloc_clustermask(unsigned int cpu, int node) node 133 arch/x86/kernel/apic/x2apic_cluster.c if (cluster_hotplug_mask->node == node) node 139 arch/x86/kernel/apic/x2apic_cluster.c GFP_KERNEL, node); node 142 arch/x86/kernel/apic/x2apic_cluster.c cluster_hotplug_mask->node = node; node 291 arch/x86/kernel/cpu/amd.c int i, node; node 294 arch/x86/kernel/cpu/amd.c node = __apicid_to_node[i]; node 295 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node)) node 296 arch/x86/kernel/cpu/amd.c return node; node 299 arch/x86/kernel/cpu/amd.c node = __apicid_to_node[i]; node 300 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node)) node 301 arch/x86/kernel/cpu/amd.c return node; node 420 arch/x86/kernel/cpu/amd.c int node; node 423 arch/x86/kernel/cpu/amd.c node = numa_cpu_node(cpu); node 424 arch/x86/kernel/cpu/amd.c if (node == NUMA_NO_NODE) node 425 arch/x86/kernel/cpu/amd.c node = per_cpu(cpu_llc_id, cpu); node 433 arch/x86/kernel/cpu/amd.c x86_cpuinit.fixup_cpu_id(c, node); node 435 arch/x86/kernel/cpu/amd.c if (!node_online(node)) { node 458 arch/x86/kernel/cpu/amd.c node = __apicid_to_node[ht_nodeid]; node 460 arch/x86/kernel/cpu/amd.c if (!node_online(node)) node 461 arch/x86/kernel/cpu/amd.c node = nearby_node(apicid); node 463 arch/x86/kernel/cpu/amd.c numa_set_node(cpu, node); node 577 arch/x86/kernel/cpu/cacheinfo.c int node; node 583 arch/x86/kernel/cpu/cacheinfo.c node = amd_get_nb_id(smp_processor_id()); node 584 arch/x86/kernel/cpu/cacheinfo.c this_leaf->nb = node_to_amd_nb(node); node 37 arch/x86/kernel/cpu/hygon.c int i, node; node 40 arch/x86/kernel/cpu/hygon.c node = __apicid_to_node[i]; node 41 arch/x86/kernel/cpu/hygon.c if (node != NUMA_NO_NODE && node_online(node)) node 42 arch/x86/kernel/cpu/hygon.c return node; node 45 arch/x86/kernel/cpu/hygon.c node = __apicid_to_node[i]; node 46 arch/x86/kernel/cpu/hygon.c if (node != NUMA_NO_NODE && node_online(node)) node 47 arch/x86/kernel/cpu/hygon.c return node; node 132 arch/x86/kernel/cpu/hygon.c int node; node 135 arch/x86/kernel/cpu/hygon.c node = numa_cpu_node(cpu); node 136 arch/x86/kernel/cpu/hygon.c if (node == NUMA_NO_NODE) node 137 arch/x86/kernel/cpu/hygon.c node = per_cpu(cpu_llc_id, cpu); node 145 arch/x86/kernel/cpu/hygon.c x86_cpuinit.fixup_cpu_id(c, node); node 147 arch/x86/kernel/cpu/hygon.c if (!node_online(node)) { node 169 arch/x86/kernel/cpu/hygon.c node = __apicid_to_node[ht_nodeid]; node 171 arch/x86/kernel/cpu/hygon.c if (!node_online(node)) node 172 arch/x86/kernel/cpu/hygon.c node = nearby_node(apicid); node 174 arch/x86/kernel/cpu/hygon.c numa_set_node(cpu, node); node 483 arch/x86/kernel/cpu/intel.c unsigned node; node 488 arch/x86/kernel/cpu/intel.c node = numa_cpu_node(cpu); node 489 arch/x86/kernel/cpu/intel.c if (node == NUMA_NO_NODE || !node_online(node)) { node 491 arch/x86/kernel/cpu/intel.c node = cpu_to_node(cpu); node 493 arch/x86/kernel/cpu/intel.c numa_set_node(cpu, node); node 34 arch/x86/kernel/cpu/mce/genpool.c struct mce_evt_llist *node; node 39 arch/x86/kernel/cpu/mce/genpool.c llist_for_each_entry(node, &l->llnode, llnode) { node 40 arch/x86/kernel/cpu/mce/genpool.c m2 = &node->mce; node 59 arch/x86/kernel/cpu/mce/genpool.c struct mce_evt_llist *node, *t; node 66 arch/x86/kernel/cpu/mce/genpool.c llist_for_each_entry_safe(node, t, head, llnode) { node 67 arch/x86/kernel/cpu/mce/genpool.c if (!is_duplicate_mce_record(node, t)) node 68 arch/x86/kernel/cpu/mce/genpool.c llist_add(&node->llnode, &new_head); node 77 arch/x86/kernel/cpu/mce/genpool.c struct mce_evt_llist *node, *tmp; node 85 arch/x86/kernel/cpu/mce/genpool.c llist_for_each_entry_safe(node, tmp, head, llnode) { node 86 arch/x86/kernel/cpu/mce/genpool.c mce = &node->mce; node 88 arch/x86/kernel/cpu/mce/genpool.c gen_pool_free(mce_evt_pool, (unsigned long)node, sizeof(*node)); node 99 arch/x86/kernel/cpu/mce/genpool.c struct mce_evt_llist *node; node 107 arch/x86/kernel/cpu/mce/genpool.c node = (void *)gen_pool_alloc(mce_evt_pool, sizeof(*node)); node 108 arch/x86/kernel/cpu/mce/genpool.c if (!node) { node 113 arch/x86/kernel/cpu/mce/genpool.c memcpy(&node->mce, mce, sizeof(*mce)); node 114 arch/x86/kernel/cpu/mce/genpool.c llist_add(&node->llnode, &mce_event_llist); node 33 arch/x86/kernel/devicetree.c void __init early_init_dt_scan_chosen_arch(unsigned long node) node 138 arch/x86/kernel/espfix_64.c int n, node; node 161 arch/x86/kernel/espfix_64.c node = cpu_to_node(cpu); node 167 arch/x86/kernel/espfix_64.c struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); node 179 arch/x86/kernel/espfix_64.c struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0); node 189 arch/x86/kernel/espfix_64.c stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0)); node 114 arch/x86/kernel/irq_32.c int node = cpu_to_node(cpu); node 120 arch/x86/kernel/irq_32.c ph = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER); node 123 arch/x86/kernel/irq_32.c ps = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER); node 32 arch/x86/kernel/kdebugfs.c struct setup_data_node *node = file->private_data; node 41 arch/x86/kernel/kdebugfs.c if (pos >= node->len) node 44 arch/x86/kernel/kdebugfs.c if (count > node->len - pos) node 45 arch/x86/kernel/kdebugfs.c count = node->len - pos; node 47 arch/x86/kernel/kdebugfs.c pa = node->paddr + sizeof(struct setup_data) + pos; node 72 arch/x86/kernel/kdebugfs.c struct setup_data_node *node) node 80 arch/x86/kernel/kdebugfs.c debugfs_create_x32("type", S_IRUGO, d, &node->type); node 81 arch/x86/kernel/kdebugfs.c debugfs_create_file("data", S_IRUGO, d, node, &fops_setup_data); node 86 arch/x86/kernel/kdebugfs.c struct setup_data_node *node; node 98 arch/x86/kernel/kdebugfs.c node = kmalloc(sizeof(*node), GFP_KERNEL); node 99 arch/x86/kernel/kdebugfs.c if (!node) { node 106 arch/x86/kernel/kdebugfs.c kfree(node); node 111 arch/x86/kernel/kdebugfs.c node->paddr = pa_data; node 112 arch/x86/kernel/kdebugfs.c node->type = data->type; node 113 arch/x86/kernel/kdebugfs.c node->len = data->len; node 114 arch/x86/kernel/kdebugfs.c create_setup_data_node(d, no, node); node 540 arch/x86/kernel/quirks.c u32 node; node 549 arch/x86/kernel/quirks.c node = pcibus_to_node(dev->bus) | (val & 7); node 554 arch/x86/kernel/quirks.c if (node_online(node)) node 555 arch/x86/kernel/quirks.c set_dev_node(&dev->dev, node); node 74 arch/x86/kernel/setup_percpu.c int node = early_cpu_to_node(cpu); node 76 arch/x86/kernel/setup_percpu.c if (node_online(node) && NODE_DATA(node) && node 77 arch/x86/kernel/setup_percpu.c last && last != NODE_DATA(node)) node 80 arch/x86/kernel/setup_percpu.c last = NODE_DATA(node); node 105 arch/x86/kernel/setup_percpu.c int node = early_cpu_to_node(cpu); node 108 arch/x86/kernel/setup_percpu.c if (!node_online(node) || !NODE_DATA(node)) { node 111 arch/x86/kernel/setup_percpu.c cpu, node); node 117 arch/x86/kernel/setup_percpu.c node); node 120 arch/x86/kernel/setup_percpu.c cpu, size, node, __pa(ptr)); node 905 arch/x86/kernel/smpboot.c int node = early_cpu_to_node(cpu); node 918 arch/x86/kernel/smpboot.c if (node != current_node) { node 921 arch/x86/kernel/smpboot.c current_node = node; node 924 arch/x86/kernel/smpboot.c node_width - num_digits(node), " ", node); node 935 arch/x86/kernel/smpboot.c node, cpu, apicid); node 5433 arch/x86/kvm/mmu.c struct kvm_page_track_notifier_node *node) node 5831 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node; node 5835 arch/x86/kvm/mmu.c list_for_each_entry_safe_reverse(sp, node, node 5926 arch/x86/kvm/mmu.c struct kvm_page_track_notifier_node *node) node 5933 arch/x86/kvm/mmu.c struct kvm_page_track_notifier_node *node = &kvm->arch.mmu_sp_tracker; node 5935 arch/x86/kvm/mmu.c node->track_write = kvm_mmu_pte_write; node 5936 arch/x86/kvm/mmu.c node->track_flush_slot = kvm_mmu_invalidate_zap_pages_in_memslot; node 5937 arch/x86/kvm/mmu.c kvm_page_track_register_notifier(kvm, node); node 5942 arch/x86/kvm/mmu.c struct kvm_page_track_notifier_node *node = &kvm->arch.mmu_sp_tracker; node 5944 arch/x86/kvm/mmu.c kvm_page_track_unregister_notifier(kvm, node); node 6123 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node; node 6129 arch/x86/kvm/mmu.c list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link) { node 356 arch/x86/kvm/mtrr.c list_del(&mtrr_state->var_ranges[index].node); node 369 arch/x86/kvm/mtrr.c list_for_each_entry(tmp, &mtrr_state->head, node) node 372 arch/x86/kvm/mtrr.c list_add_tail(&cur->node, &tmp->node); node 522 arch/x86/kvm/mtrr.c list_for_each_entry_continue(iter->range, &mtrr_state->head, node) node 537 arch/x86/kvm/mtrr.c iter->range = list_prepare_entry(iter->range, &mtrr_state->head, node); node 192 arch/x86/kvm/page_track.c hlist_add_head_rcu(&n->node, &head->track_notifier_list); node 210 arch/x86/kvm/page_track.c hlist_del_rcu(&n->node); node 236 arch/x86/kvm/page_track.c hlist_for_each_entry_rcu(n, &head->track_notifier_list, node) node 261 arch/x86/kvm/page_track.c hlist_for_each_entry_rcu(n, &head->track_notifier_list, node) node 255 arch/x86/kvm/svm.c struct list_head node; /* Used by SVM for per-vcpu ir_list */ node 2066 arch/x86/kvm/svm.c list_for_each_entry(ir, &svm->ir_list, node) { node 5221 arch/x86/kvm/svm.c list_for_each_entry(cur, &svm->ir_list, node) { node 5224 arch/x86/kvm/svm.c list_del(&cur->node); node 5269 arch/x86/kvm/svm.c list_add(&ir->node, &svm->ir_list); node 2504 arch/x86/kvm/vmx/vmx.c int node = cpu_to_node(cpu); node 2508 arch/x86/kvm/vmx/vmx.c pages = __alloc_pages_node(node, flags, vmcs_config.order); node 1450 arch/x86/mm/init_64.c unsigned long end, int node, struct vmem_altmap *altmap) node 1462 arch/x86/mm/init_64.c pgd = vmemmap_pgd_populate(addr, node); node 1466 arch/x86/mm/init_64.c p4d = vmemmap_p4d_populate(pgd, addr, node); node 1470 arch/x86/mm/init_64.c pud = vmemmap_pud_populate(p4d, addr, node); node 1481 arch/x86/mm/init_64.c p = vmemmap_alloc_block_buf(PMD_SIZE, node); node 1490 arch/x86/mm/init_64.c if (p_end != p || node_start != node) { node 1495 arch/x86/mm/init_64.c node_start = node; node 1505 arch/x86/mm/init_64.c vmemmap_verify((pte_t *)pmd, node, addr, next); node 1508 arch/x86/mm/init_64.c if (vmemmap_populate_basepages(addr, next, node)) node 1514 arch/x86/mm/init_64.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, node 1520 arch/x86/mm/init_64.c err = vmemmap_populate_basepages(start, end, node); node 1522 arch/x86/mm/init_64.c err = vmemmap_populate_hugepages(start, end, node, altmap); node 1528 arch/x86/mm/init_64.c err = vmemmap_populate_basepages(start, end, node); node 80 arch/x86/mm/numa.c void numa_set_node(int cpu, int node) node 86 arch/x86/mm/numa.c cpu_to_node_map[cpu] = node; node 97 arch/x86/mm/numa.c per_cpu(x86_cpu_to_node_map, cpu) = node; node 99 arch/x86/mm/numa.c set_cpu_numa_node(cpu, node); node 116 arch/x86/mm/numa.c unsigned int node; node 123 arch/x86/mm/numa.c for (node = 0; node < nr_node_ids; node++) node 124 arch/x86/mm/numa.c alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]); node 758 arch/x86/mm/numa.c int node = numa_cpu_node(cpu); node 760 arch/x86/mm/numa.c if (node == NUMA_NO_NODE) node 763 arch/x86/mm/numa.c if (!node_online(node)) node 764 arch/x86/mm/numa.c init_memory_less_node(node); node 766 arch/x86/mm/numa.c numa_set_node(cpu, node); node 816 arch/x86/mm/numa.c void debug_cpumask_set_cpu(int cpu, int node, bool enable) node 820 arch/x86/mm/numa.c if (node == NUMA_NO_NODE) { node 824 arch/x86/mm/numa.c mask = node_to_cpumask_map[node]; node 826 arch/x86/mm/numa.c pr_err("node_to_cpumask_map[%i] NULL\n", node); node 838 arch/x86/mm/numa.c cpu, node, cpumask_pr_args(mask)); node 862 arch/x86/mm/numa.c const struct cpumask *cpumask_of_node(int node) node 864 arch/x86/mm/numa.c if ((unsigned)node >= nr_node_ids) { node 867 arch/x86/mm/numa.c node, nr_node_ids); node 871 arch/x86/mm/numa.c if (node_to_cpumask_map[node] == NULL) { node 874 arch/x86/mm/numa.c node); node 878 arch/x86/mm/numa.c return node_to_cpumask_map[node]; node 39 arch/x86/mm/pat_rbtree.c static int is_node_overlap(struct memtype *node, u64 start, u64 end) node 41 arch/x86/mm/pat_rbtree.c if (node->start >= end || node->end <= start) node 47 arch/x86/mm/pat_rbtree.c static u64 get_subtree_max_end(struct rb_node *node) node 50 arch/x86/mm/pat_rbtree.c if (node) { node 51 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(node, struct memtype, rb); node 57 arch/x86/mm/pat_rbtree.c #define NODE_END(node) ((node)->end) node 66 arch/x86/mm/pat_rbtree.c struct rb_node *node = root->rb_node; node 69 arch/x86/mm/pat_rbtree.c while (node) { node 70 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(node, struct memtype, rb); node 72 arch/x86/mm/pat_rbtree.c if (get_subtree_max_end(node->rb_left) > start) { node 74 arch/x86/mm/pat_rbtree.c node = node->rb_left; node 80 arch/x86/mm/pat_rbtree.c node = node->rb_right; node 100 arch/x86/mm/pat_rbtree.c struct rb_node *node; node 110 arch/x86/mm/pat_rbtree.c node = rb_next(&match->rb); node 111 arch/x86/mm/pat_rbtree.c if (node) node 112 arch/x86/mm/pat_rbtree.c match = rb_entry(node, struct memtype, rb); node 125 arch/x86/mm/pat_rbtree.c struct rb_node *node; node 139 arch/x86/mm/pat_rbtree.c node = rb_next(&match->rb); node 140 arch/x86/mm/pat_rbtree.c while (node) { node 141 arch/x86/mm/pat_rbtree.c match = rb_entry(node, struct memtype, rb); node 151 arch/x86/mm/pat_rbtree.c node = rb_next(&match->rb); node 168 arch/x86/mm/pat_rbtree.c struct rb_node **node = &(root->rb_node); node 171 arch/x86/mm/pat_rbtree.c while (*node) { node 172 arch/x86/mm/pat_rbtree.c struct memtype *data = rb_entry(*node, struct memtype, rb); node 174 arch/x86/mm/pat_rbtree.c parent = *node; node 178 arch/x86/mm/pat_rbtree.c node = &((*node)->rb_left); node 180 arch/x86/mm/pat_rbtree.c node = &((*node)->rb_right); node 184 arch/x86/mm/pat_rbtree.c rb_link_node(&newdata->rb, parent, node); node 251 arch/x86/mm/pat_rbtree.c struct rb_node *node; node 254 arch/x86/mm/pat_rbtree.c node = rb_first(&memtype_rbroot); node 255 arch/x86/mm/pat_rbtree.c while (node && pos != i) { node 256 arch/x86/mm/pat_rbtree.c node = rb_next(node); node 260 arch/x86/mm/pat_rbtree.c if (node) { /* pos == i */ node 261 arch/x86/mm/pat_rbtree.c struct memtype *this = rb_entry(node, struct memtype, rb); node 30 arch/x86/mm/srat.c int pxm, node; node 48 arch/x86/mm/srat.c node = acpi_map_pxm_to_node(pxm); node 49 arch/x86/mm/srat.c if (node < 0) { node 56 arch/x86/mm/srat.c printk(KERN_INFO "SRAT: PXM %u -> APIC 0x%04x -> Node %u skipped apicid that is too big\n", pxm, apic_id, node); node 59 arch/x86/mm/srat.c set_apicid_to_node(apic_id, node); node 60 arch/x86/mm/srat.c node_set(node, numa_nodes_parsed); node 62 arch/x86/mm/srat.c pxm, apic_id, node); node 69 arch/x86/mm/srat.c int pxm, node; node 83 arch/x86/mm/srat.c node = acpi_map_pxm_to_node(pxm); node 84 arch/x86/mm/srat.c if (node < 0) { node 96 arch/x86/mm/srat.c printk(KERN_INFO "SRAT: PXM %u -> APIC 0x%02x -> Node %u skipped apicid that is too big\n", pxm, apic_id, node); node 100 arch/x86/mm/srat.c set_apicid_to_node(apic_id, node); node 101 arch/x86/mm/srat.c node_set(node, numa_nodes_parsed); node 103 arch/x86/mm/srat.c pxm, apic_id, node); node 249 arch/x86/pci/acpi.c int node = acpi_get_node(device->handle); node 251 arch/x86/pci/acpi.c if (node == NUMA_NO_NODE) { node 252 arch/x86/pci/acpi.c node = x86_pci_root_bus_node(busnum); node 253 arch/x86/pci/acpi.c if (node != 0 && node != NUMA_NO_NODE) node 255 arch/x86/pci/acpi.c node); node 257 arch/x86/pci/acpi.c if (node != NUMA_NO_NODE && !node_online(node)) node 258 arch/x86/pci/acpi.c node = NUMA_NO_NODE; node 260 arch/x86/pci/acpi.c return node; node 330 arch/x86/pci/acpi.c int node = pci_acpi_root_get_node(root); node 351 arch/x86/pci/acpi.c .node = node, node 366 arch/x86/pci/acpi.c info->sd.node = node; node 378 arch/x86/pci/acpi.c list_for_each_entry(child, &bus->children, node) node 42 arch/x86/pci/amd_bus.c static struct pci_root_info __init *find_pci_root_info(int node, int link) node 48 arch/x86/pci/amd_bus.c if (info->node == node && info->link == link) node 65 arch/x86/pci/amd_bus.c int node; node 126 arch/x86/pci/amd_bus.c node = (reg >> 4) & 0x07; node 129 arch/x86/pci/amd_bus.c info = alloc_pci_root_info(min_bus, max_bus, node, link); node 159 arch/x86/pci/amd_bus.c node = reg & 0x07; node 163 arch/x86/pci/amd_bus.c info = find_pci_root_info(node, link); node 168 arch/x86/pci/amd_bus.c node, link, start, end); node 226 arch/x86/pci/amd_bus.c node = reg & 0x07; node 232 arch/x86/pci/amd_bus.c info = find_pci_root_info(node, link); node 238 arch/x86/pci/amd_bus.c node, link, start, end); node 321 arch/x86/pci/amd_bus.c &info->busn, info->node, info->link); node 28 arch/x86/pci/bus_numa.c return info->node; node 71 arch/x86/pci/bus_numa.c int node, int link) node 87 arch/x86/pci/bus_numa.c info->node = node; node 18 arch/x86/pci/bus_numa.h int node; node 24 arch/x86/pci/bus_numa.h int node, int link); node 468 arch/x86/pci/common.c sd->node = x86_pci_root_bus_node(busnum); node 635 arch/x86/pci/common.c list_add(&domain->node, &dma_domain_list); node 643 arch/x86/pci/common.c list_del(&domain->node); node 653 arch/x86/pci/common.c list_for_each_entry(domain, &dma_domain_list, node) { node 240 arch/x86/pci/i386.c list_for_each_entry(child, &bus->children, node) node 360 arch/x86/pci/i386.c list_for_each_entry(bus, &pci_root_buses, node) node 394 arch/x86/pci/i386.c list_for_each_entry(bus, &pci_root_buses, node) node 397 arch/x86/pci/i386.c list_for_each_entry(bus, &pci_root_buses, node) node 399 arch/x86/pci/i386.c list_for_each_entry(bus, &pci_root_buses, node) node 59 arch/x86/platform/olpc/olpc-xo1-rtc.c struct device_node *node; node 61 arch/x86/platform/olpc/olpc-xo1-rtc.c node = of_find_compatible_node(NULL, NULL, "olpc,xo1-rtc"); node 62 arch/x86/platform/olpc/olpc-xo1-rtc.c if (!node) node 64 arch/x86/platform/olpc/olpc-xo1-rtc.c of_node_put(node); node 22 arch/x86/platform/olpc/olpc_dt.c static phandle __init olpc_dt_getsibling(phandle node) node 24 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node }; node 25 arch/x86/platform/olpc/olpc_dt.c void *res[] = { &node }; node 27 arch/x86/platform/olpc/olpc_dt.c if ((s32)node == -1) node 30 arch/x86/platform/olpc/olpc_dt.c if (olpc_ofw("peer", args, res) || (s32)node == -1) node 33 arch/x86/platform/olpc/olpc_dt.c return node; node 36 arch/x86/platform/olpc/olpc_dt.c static phandle __init olpc_dt_getchild(phandle node) node 38 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node }; node 39 arch/x86/platform/olpc/olpc_dt.c void *res[] = { &node }; node 41 arch/x86/platform/olpc/olpc_dt.c if ((s32)node == -1) node 44 arch/x86/platform/olpc/olpc_dt.c if (olpc_ofw("child", args, res) || (s32)node == -1) { node 49 arch/x86/platform/olpc/olpc_dt.c return node; node 52 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_getproplen(phandle node, const char *prop) node 54 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node, prop }; node 58 arch/x86/platform/olpc/olpc_dt.c if ((s32)node == -1) node 69 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_getproperty(phandle node, const char *prop, node 74 arch/x86/platform/olpc/olpc_dt.c plen = olpc_dt_getproplen(node, prop); node 78 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node, prop, buf, (void *)plen }; node 90 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_nextprop(phandle node, char *prev, char *buf) node 92 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node, prev, buf }; node 98 arch/x86/platform/olpc/olpc_dt.c if ((s32)node == -1) node 107 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_pkg2path(phandle node, char *buf, node 110 arch/x86/platform/olpc/olpc_dt.c const void *args[] = { (void *)node, buf, (void *)buflen }; node 113 arch/x86/platform/olpc/olpc_dt.c if ((s32)node == -1) node 168 arch/x86/platform/olpc/olpc_dt.c phandle node; node 170 arch/x86/platform/olpc/olpc_dt.c void *res[] = { &node }; node 177 arch/x86/platform/olpc/olpc_dt.c if ((s32) node == -1) node 180 arch/x86/platform/olpc/olpc_dt.c return node; node 203 arch/x86/platform/olpc/olpc_dt.c phandle node; node 207 arch/x86/platform/olpc/olpc_dt.c node = olpc_dt_finddevice("/"); node 208 arch/x86/platform/olpc/olpc_dt.c if (!node) node 211 arch/x86/platform/olpc/olpc_dt.c r = olpc_dt_getproperty(node, "board-revision-int", node 219 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_compatible_match(phandle node, const char *compat) node 224 arch/x86/platform/olpc/olpc_dt.c plen = olpc_dt_getproperty(node, "compatible", buf, sizeof(buf)); node 239 arch/x86/platform/olpc/olpc_dt.c phandle node; node 242 arch/x86/platform/olpc/olpc_dt.c node = olpc_dt_finddevice("/battery@0"); node 243 arch/x86/platform/olpc/olpc_dt.c if (!node) node 253 arch/x86/platform/olpc/olpc_dt.c if (olpc_dt_compatible_match(node, "olpc,xo1.5-battery")) node 261 arch/x86/platform/olpc/olpc_dt.c if (olpc_dt_compatible_match(node, "olpc,xo1-battery")) { node 280 arch/x86/platform/olpc/olpc_dt.c if (olpc_dt_compatible_match(node, "olpc,xo1-battery")) { node 176 arch/x86/platform/uv/tlb_uv.c int node, b; node 178 arch/x86/platform/uv/tlb_uv.c for_each_online_node(node) { node 179 arch/x86/platform/uv/tlb_uv.c b = uv_node_to_blade_id(node); node 181 arch/x86/platform/uv/tlb_uv.c return node; node 1710 arch/x86/platform/uv/tlb_uv.c static void activation_descriptor_init(int node, int pnode, int base_pnode) node 1730 arch/x86/platform/uv/tlb_uv.c bau_desc = kmalloc_node(dsize, GFP_KERNEL, node); node 1795 arch/x86/platform/uv/tlb_uv.c static void pq_init(int node, int pnode) node 1806 arch/x86/platform/uv/tlb_uv.c vp = kmalloc_node(plsize, GFP_KERNEL, node); node 1850 arch/x86/platform/uv/tlb_uv.c int node; node 1854 arch/x86/platform/uv/tlb_uv.c node = uvhub_to_first_node(uvhub); node 1857 arch/x86/platform/uv/tlb_uv.c activation_descriptor_init(node, pnode, base_pnode); node 1859 arch/x86/platform/uv/tlb_uv.c pq_init(node, pnode); node 53 arch/x86/um/vdso/um_vdso.c __vdso_getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *unused) node 61 arch/x86/um/vdso/um_vdso.c if (node) node 62 arch/x86/um/vdso/um_vdso.c *node = 0; node 67 arch/x86/um/vdso/um_vdso.c long getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache) node 193 arch/xtensa/kernel/setup.c static int __init xtensa_dt_io_area(unsigned long node, const char *uname, node 202 arch/xtensa/kernel/setup.c if (!of_flat_dt_is_compatible(node, "simple-bus")) node 205 arch/xtensa/kernel/setup.c ranges = of_get_flat_dt_prop(node, "ranges", &len); node 220 arch/xtensa/kernel/setup.c static int __init xtensa_dt_io_area(unsigned long node, const char *uname, node 114 arch/xtensa/platforms/xtfpga/setup.c static void __init update_local_mac(struct device_node *node) node 120 arch/xtensa/platforms/xtfpga/setup.c macaddr = of_get_property(node, "local-mac-address", &prop_len); node 138 arch/xtensa/platforms/xtfpga/setup.c of_update_property(node, newmac); node 509 block/bfq-cgroup.c bfqg = kzalloc_node(sizeof(*bfqg), gfp, q->node); node 1211 block/bfq-cgroup.c struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node) node 1413 block/bfq-cgroup.c struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node) node 1418 block/bfq-cgroup.c bfqg = kmalloc_node(sizeof(*bfqg), GFP_KERNEL | __GFP_ZERO, node); node 2415 block/bfq-iosched.c struct rb_node *parent, *node; node 2439 block/bfq-iosched.c node = rb_next(&__bfqq->pos_node); node 2441 block/bfq-iosched.c node = rb_prev(&__bfqq->pos_node); node 2442 block/bfq-iosched.c if (!node) node 2445 block/bfq-iosched.c __bfqq = rb_entry(node, struct bfq_queue, pos_node); node 5152 block/bfq-iosched.c bfqd->queue->node); node 6432 block/bfq-iosched.c bfqd = kzalloc_node(sizeof(*bfqd), GFP_KERNEL, q->node); node 6540 block/bfq-iosched.c bfqd->root_group = bfq_create_group_hierarchy(bfqd, q->node); node 982 block/bfq-iosched.h struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node); node 1027 block/bfq-iosched.h struct bfq_entity *bfq_entity_of(struct rb_node *node); node 25 block/bfq-wf2q.c struct rb_node *node = tree->rb_node; node 27 block/bfq-wf2q.c return rb_entry(node, struct bfq_entity, rb_node); node 318 block/bfq-wf2q.c struct bfq_entity *bfq_entity_of(struct rb_node *node) node 322 block/bfq-wf2q.c if (node) node 323 block/bfq-wf2q.c entity = rb_entry(node, struct bfq_entity, rb_node); node 377 block/bfq-wf2q.c struct rb_node **node = &root->rb_node; node 380 block/bfq-wf2q.c while (*node) { node 381 block/bfq-wf2q.c parent = *node; node 385 block/bfq-wf2q.c node = &parent->rb_left; node 387 block/bfq-wf2q.c node = &parent->rb_right; node 390 block/bfq-wf2q.c rb_link_node(&entity->rb_node, parent, node); node 406 block/bfq-wf2q.c static void bfq_update_min(struct bfq_entity *entity, struct rb_node *node) node 410 block/bfq-wf2q.c if (node) { node 411 block/bfq-wf2q.c child = rb_entry(node, struct bfq_entity, rb_node); node 425 block/bfq-wf2q.c static void bfq_update_active_node(struct rb_node *node) node 427 block/bfq-wf2q.c struct bfq_entity *entity = rb_entry(node, struct bfq_entity, rb_node); node 430 block/bfq-wf2q.c bfq_update_min(entity, node->rb_right); node 431 block/bfq-wf2q.c bfq_update_min(entity, node->rb_left); node 444 block/bfq-wf2q.c static void bfq_update_active_tree(struct rb_node *node) node 449 block/bfq-wf2q.c bfq_update_active_node(node); node 451 block/bfq-wf2q.c parent = rb_parent(node); node 455 block/bfq-wf2q.c if (node == parent->rb_left && parent->rb_right) node 460 block/bfq-wf2q.c node = parent; node 479 block/bfq-wf2q.c struct rb_node *node = &entity->rb_node; node 488 block/bfq-wf2q.c if (node->rb_left) node 489 block/bfq-wf2q.c node = node->rb_left; node 490 block/bfq-wf2q.c else if (node->rb_right) node 491 block/bfq-wf2q.c node = node->rb_right; node 493 block/bfq-wf2q.c bfq_update_active_tree(node); node 553 block/bfq-wf2q.c static struct rb_node *bfq_find_deepest(struct rb_node *node) node 557 block/bfq-wf2q.c if (!node->rb_right && !node->rb_left) node 558 block/bfq-wf2q.c deepest = rb_parent(node); node 559 block/bfq-wf2q.c else if (!node->rb_right) node 560 block/bfq-wf2q.c deepest = node->rb_left; node 561 block/bfq-wf2q.c else if (!node->rb_left) node 562 block/bfq-wf2q.c deepest = node->rb_right; node 564 block/bfq-wf2q.c deepest = rb_next(node); node 567 block/bfq-wf2q.c else if (rb_parent(deepest) != node) node 583 block/bfq-wf2q.c struct rb_node *node; node 590 block/bfq-wf2q.c node = bfq_find_deepest(&entity->rb_node); node 593 block/bfq-wf2q.c if (node) node 594 block/bfq-wf2q.c bfq_update_active_tree(node); node 1362 block/bfq-wf2q.c struct rb_node *node = st->active.rb_node; node 1364 block/bfq-wf2q.c while (node) { node 1365 block/bfq-wf2q.c entry = rb_entry(node, struct bfq_entity, rb_node); node 1370 block/bfq-wf2q.c if (node->rb_left) { node 1371 block/bfq-wf2q.c entry = rb_entry(node->rb_left, node 1374 block/bfq-wf2q.c node = node->rb_left; node 1380 block/bfq-wf2q.c node = node->rb_right; node 152 block/blk-cgroup.c blkg = kzalloc_node(sizeof(*blkg), gfp_mask, q->node); node 507 block/blk-core.c q->node = node_id; node 960 block/blk-core.c create_io_context(GFP_ATOMIC, q->node); node 476 block/blk-flush.c int node, int cmd_size, gfp_t flags) node 481 block/blk-flush.c fq = kzalloc_node(sizeof(*fq), flags, node); node 488 block/blk-flush.c fq->flush_rq = kzalloc_node(rq_sz, flags, node); node 250 block/blk-ioc.c int create_task_io_context(struct task_struct *task, gfp_t gfp_flags, int node) node 256 block/blk-ioc.c node); node 304 block/blk-ioc.c gfp_t gfp_flags, int node) node 319 block/blk-ioc.c } while (!create_task_io_context(task, gfp_flags, node)); node 380 block/blk-ioc.c q->node); node 2009 block/blk-iocost.c gfp, q->node); node 943 block/blk-iolatency.c iolat = kzalloc_node(sizeof(*iolat), gfp, q->node); node 437 block/blk-mq-tag.c bool round_robin, int node) node 440 block/blk-mq-tag.c node); node 444 block/blk-mq-tag.c int node, int alloc_policy) node 449 block/blk-mq-tag.c if (bt_alloc(&tags->bitmap_tags, depth, round_robin, node)) node 452 block/blk-mq-tag.c node)) node 465 block/blk-mq-tag.c int node, int alloc_policy) node 474 block/blk-mq-tag.c tags = kzalloc_node(sizeof(*tags), GFP_KERNEL, node); node 481 block/blk-mq-tag.c return blk_mq_init_bitmap_tags(tags, node, alloc_policy); node 25 block/blk-mq-tag.h extern struct blk_mq_tags *blk_mq_init_tags(unsigned int nr_tags, unsigned int reserved_tags, int node, int alloc_policy); node 2106 block/blk-mq.c int node; node 2108 block/blk-mq.c node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], hctx_idx); node 2109 block/blk-mq.c if (node == NUMA_NO_NODE) node 2110 block/blk-mq.c node = set->numa_node; node 2112 block/blk-mq.c tags = blk_mq_init_tags(nr_tags, reserved_tags, node, node 2119 block/blk-mq.c node); node 2127 block/blk-mq.c node); node 2143 block/blk-mq.c unsigned int hctx_idx, int node) node 2148 block/blk-mq.c ret = set->ops->init_request(set, rq, hctx_idx, node); node 2162 block/blk-mq.c int node; node 2164 block/blk-mq.c node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], hctx_idx); node 2165 block/blk-mq.c if (node == NUMA_NO_NODE) node 2166 block/blk-mq.c node = set->numa_node; node 2188 block/blk-mq.c page = alloc_pages_node(node, node 2218 block/blk-mq.c if (blk_mq_init_request(set, rq, hctx_idx, node)) { node 2239 block/blk-mq.c static int blk_mq_hctx_notify_dead(unsigned int cpu, struct hlist_node *node) node 2246 block/blk-mq.c hctx = hlist_entry_safe(node, struct blk_mq_hw_ctx, cpuhp_dead); node 2352 block/blk-mq.c int node) node 2357 block/blk-mq.c hctx = kzalloc_node(blk_mq_hw_ctx_size(set), gfp, node); node 2361 block/blk-mq.c if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node)) node 2365 block/blk-mq.c if (node == NUMA_NO_NODE) node 2366 block/blk-mq.c node = set->numa_node; node 2367 block/blk-mq.c hctx->numa_node = node; node 2382 block/blk-mq.c gfp, node); node 2387 block/blk-mq.c gfp, node)) node 2760 block/blk-mq.c int hctx_idx, int node) node 2767 block/blk-mq.c if (tmp->numa_node == node) { node 2777 block/blk-mq.c hctx = blk_mq_alloc_hctx(q, set, node); node 2801 block/blk-mq.c int node; node 2804 block/blk-mq.c node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], i); node 2810 block/blk-mq.c if (hctxs[i] && (hctxs[i]->numa_node == node)) node 2813 block/blk-mq.c hctx = blk_mq_alloc_and_init_hctx(set, q, i, node); node 2822 block/blk-mq.c node, hctxs[i]->numa_node); node 3207 block/blk-mq.c struct list_head node; node 3228 block/blk-mq.c INIT_LIST_HEAD(&qe->node); node 3231 block/blk-mq.c list_add(&qe->node, head); node 3254 block/blk-mq.c list_for_each_entry(qe, head, node) node 3263 block/blk-mq.c list_del(&qe->node); node 68 block/blk-throttle.c struct list_head node; /* service_queue->queued[] */ node 98 block/blk-throttle.c #define rb_entry_tg(node) rb_entry((node), struct throtl_grp, rb_node) node 391 block/blk-throttle.c INIT_LIST_HEAD(&qn->node); node 410 block/blk-throttle.c if (list_empty(&qn->node)) { node 411 block/blk-throttle.c list_add_tail(&qn->node, queued); node 422 block/blk-throttle.c struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); node 450 block/blk-throttle.c struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); node 460 block/blk-throttle.c list_del_init(&qn->node); node 466 block/blk-throttle.c list_move_tail(&qn->node, queued); node 488 block/blk-throttle.c tg = kzalloc_node(sizeof(*tg), gfp, q->node); node 654 block/blk-throttle.c struct rb_node **node = &parent_sq->pending_tree.rb_root.rb_node; node 660 block/blk-throttle.c while (*node != NULL) { node 661 block/blk-throttle.c parent = *node; node 665 block/blk-throttle.c node = &parent->rb_left; node 667 block/blk-throttle.c node = &parent->rb_right; node 672 block/blk-throttle.c rb_link_node(&tg->rb_node, parent, node); node 2376 block/blk-throttle.c td = kzalloc_node(sizeof(*td), GFP_KERNEL, q->node); node 404 block/blk-zoned.c static inline unsigned long *blk_alloc_zone_bitmap(int node, node 408 block/blk-zoned.c GFP_NOIO, node); node 486 block/blk-zoned.c seq_zones_wlock = blk_alloc_zone_bitmap(q->node, nr_zones); node 489 block/blk-zoned.c seq_zones_bitmap = blk_alloc_zone_bitmap(q->node, nr_zones); node 59 block/blk.h int node, int cmd_size, gfp_t flags); node 286 block/blk.h int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node); node 300 block/blk.h static inline struct io_context *create_io_context(gfp_t gfp_mask, int node) node 304 block/blk.h create_task_io_context(current, gfp_mask, node); node 169 block/elevator.c eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node); node 1622 block/genhd.c struct list_head node; /* all disk_event's */ node 1995 block/genhd.c list_for_each_entry(ev, &disk_events, node) node 2030 block/genhd.c INIT_LIST_HEAD(&ev->node); node 2052 block/genhd.c list_add_tail(&disk->ev->node, &disk_events); node 2068 block/genhd.c list_del_init(&disk->ev->node); node 372 block/kyber-iosched.c kqd = kzalloc_node(sizeof(*kqd), GFP_KERNEL, q->node); node 390 block/kyber-iosched.c GFP_KERNEL, q->node); node 79 block/mq-deadline.c struct rb_node *node = rb_next(&rq->rb_node); node 81 block/mq-deadline.c if (node) node 82 block/mq-deadline.c return rb_entry_rq(node); node 415 block/mq-deadline.c dd = kzalloc_node(sizeof(*dd), GFP_KERNEL, q->node); node 218 block/sed-opal.c struct list_head node; node 1073 block/sed-opal.c list_for_each_entry(iter, &dev->unlk_lst, node) { node 1075 block/sed-opal.c list_del(&iter->node); node 1080 block/sed-opal.c list_add_tail(&sus->node, &dev->unlk_lst); node 2009 block/sed-opal.c list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) { node 2010 block/sed-opal.c list_del(&suspend->node); node 2422 block/sed-opal.c list_for_each_entry(suspend, &dev->unlk_lst, node) { node 42 crypto/af_alg.c struct alg_type_list *node; node 45 crypto/af_alg.c list_for_each_entry(node, &alg_types, list) { node 46 crypto/af_alg.c if (strcmp(node->type->name, name)) node 49 crypto/af_alg.c if (try_module_get(node->type->owner)) node 50 crypto/af_alg.c type = node->type; node 60 crypto/af_alg.c struct alg_type_list *node; node 64 crypto/af_alg.c list_for_each_entry(node, &alg_types, list) { node 65 crypto/af_alg.c if (!strcmp(node->type->name, type->name)) node 69 crypto/af_alg.c node = kmalloc(sizeof(*node), GFP_KERNEL); node 71 crypto/af_alg.c if (!node) node 77 crypto/af_alg.c node->type = type; node 78 crypto/af_alg.c list_add(&node->list, &alg_types); node 90 crypto/af_alg.c struct alg_type_list *node; node 94 crypto/af_alg.c list_for_each_entry(node, &alg_types, list) { node 95 crypto/af_alg.c if (strcmp(node->type->name, type->name)) node 98 crypto/af_alg.c list_del(&node->list); node 99 crypto/af_alg.c kfree(node); node 74 drivers/acpi/acpi_amba.c list_for_each_entry(rentry, &resource_list, node) { node 101 drivers/acpi/acpi_apd.c list_for_each_entry(rentry, &resource_list, node) { node 657 drivers/acpi/acpi_lpss.c list_for_each_entry(rentry, &resource_list, node) node 187 drivers/acpi/acpi_memhotplug.c int node; node 189 drivers/acpi/acpi_memhotplug.c node = acpi_get_node(handle); node 207 drivers/acpi/acpi_memhotplug.c if (node < 0) node 208 drivers/acpi/acpi_memhotplug.c node = memory_add_physaddr_to_nid(info->start_addr); node 210 drivers/acpi/acpi_memhotplug.c result = __add_memory(node, info->start_addr, info->length); node 129 drivers/acpi/acpi_platform.c list_for_each_entry(rentry, &resource_list, node) node 1535 drivers/acpi/acpi_video.c list_for_each_entry(dev, &device->children, node) { node 142 drivers/acpi/acpica/acdispat.h union acpi_operand_object ***node); node 163 drivers/acpi/acpica/acdispat.h struct acpi_namespace_node **node); node 171 drivers/acpi/acpica/acdispat.h acpi_ds_auto_serialize_method(struct acpi_namespace_node *node, node 229 drivers/acpi/acpica/acdispat.h struct acpi_namespace_node *node, node 277 drivers/acpi/acpica/acdispat.h acpi_ds_scope_stack_push(struct acpi_namespace_node *node, node 39 drivers/acpi/acpica/acevents.h u8 acpi_ev_is_notify_object(struct acpi_namespace_node *node); node 44 drivers/acpi/acpica/acevents.h acpi_ev_queue_notify_request(struct acpi_namespace_node *node, node 159 drivers/acpi/acpica/acevents.h acpi_ev_has_default_handler(struct acpi_namespace_node *node, node 165 drivers/acpi/acpica/acevents.h acpi_ev_install_space_handler(struct acpi_namespace_node *node, node 191 drivers/acpi/acpica/acevents.h acpi_ev_execute_reg_methods(struct acpi_namespace_node *node, node 233 drivers/acpi/acpica/acevents.h u8 acpi_ev_is_pci_root_bridge(struct acpi_namespace_node *node); node 393 drivers/acpi/acpica/acinterp.h void acpi_ex_dump_namespace_node(struct acpi_namespace_node *node, u32 flags); node 413 drivers/acpi/acpica/acinterp.h struct acpi_namespace_node *node, node 455 drivers/acpi/acpica/acinterp.h struct acpi_namespace_node *node); node 207 drivers/acpi/acpica/aclocal.h struct acpi_namespace_node *node; node 474 drivers/acpi/acpica/aclocal.h struct acpi_namespace_node *node; node 598 drivers/acpi/acpica/aclocal.h ACPI_STATE_COMMON struct acpi_namespace_node *node; node 649 drivers/acpi/acpica/aclocal.h struct acpi_namespace_node *node; node 737 drivers/acpi/acpica/aclocal.h struct acpi_namespace_node *node; /* For use by interpreter */\ node 71 drivers/acpi/acpica/acnamesp.h acpi_ns_load_table(u32 table_index, struct acpi_namespace_node *node); node 131 drivers/acpi/acpica/acnamesp.h void acpi_ns_delete_node(struct acpi_namespace_node *node); node 133 drivers/acpi/acpica/acnamesp.h void acpi_ns_remove_node(struct acpi_namespace_node *node); node 140 drivers/acpi/acpica/acnamesp.h void acpi_ns_detach_object(struct acpi_namespace_node *node); node 215 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, node 221 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, node 230 drivers/acpi/acpica/acnamesp.h acpi_ns_check_return_value(struct acpi_namespace_node *node, node 253 drivers/acpi/acpica/acnamesp.h char *acpi_ns_get_external_pathname(struct acpi_namespace_node *node); node 256 drivers/acpi/acpica/acnamesp.h acpi_ns_build_normalized_path(struct acpi_namespace_node *node, node 261 drivers/acpi/acpica/acnamesp.h char *acpi_ns_get_normalized_pathname(struct acpi_namespace_node *node, node 289 drivers/acpi/acpica/acnamesp.h acpi_size acpi_ns_get_pathname_length(struct acpi_namespace_node *node); node 295 drivers/acpi/acpica/acnamesp.h acpi_ns_attach_object(struct acpi_namespace_node *node, node 300 drivers/acpi/acpica/acnamesp.h *node); node 307 drivers/acpi/acpica/acnamesp.h acpi_ns_attach_data(struct acpi_namespace_node *node, node 311 drivers/acpi/acpica/acnamesp.h acpi_ns_detach_data(struct acpi_namespace_node *node, node 315 drivers/acpi/acpica/acnamesp.h acpi_ns_get_attached_data(struct acpi_namespace_node *node, node 350 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, node 360 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, node 367 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, node 374 drivers/acpi/acpica/acnamesp.h struct acpi_namespace_node *node, acpi_object_type type); node 379 drivers/acpi/acpica/acnamesp.h acpi_object_type acpi_ns_get_type(struct acpi_namespace_node *node); node 384 drivers/acpi/acpica/acnamesp.h acpi_ns_print_node_pathname(struct acpi_namespace_node *node, const char *msg); node 101 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */ node 105 drivers/acpi/acpica/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Link back to parent node */ node 130 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */ node 136 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */ node 148 drivers/acpi/acpica/acobject.h union acpi_operand_object *node; node 223 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */\ node 273 drivers/acpi/acpica/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Parent device */ node 284 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* Parent device */ node 311 drivers/acpi/acpica/acobject.h struct acpi_namespace_node *node; /* ref_of or Namepath */ node 398 drivers/acpi/acpica/acobject.h struct acpi_namespace_node node; node 434 drivers/acpi/acpica/acobject.h struct acpi_namespace_node node; node 168 drivers/acpi/acpica/acresrc.h acpi_rs_get_prt_method_data(struct acpi_namespace_node *node, node 172 drivers/acpi/acpica/acresrc.h acpi_rs_get_crs_method_data(struct acpi_namespace_node *node, node 176 drivers/acpi/acpica/acresrc.h acpi_rs_get_prs_method_data(struct acpi_namespace_node *node, node 184 drivers/acpi/acpica/acresrc.h acpi_rs_set_srs_method_data(struct acpi_namespace_node *node, node 188 drivers/acpi/acpica/acresrc.h acpi_rs_get_aei_method_data(struct acpi_namespace_node *node, node 159 drivers/acpi/acpica/acstruct.h struct acpi_namespace_node *node; /* Resolved node (prefix_node:relative_pathname) */ node 721 drivers/acpi/acpica/acutils.h struct acpi_namespace_node *node, node 27 drivers/acpi/acpica/dbcmds.c acpi_dm_test_resource_conversion(struct acpi_namespace_node *node, char *name); node 55 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 63 drivers/acpi/acpica/dbcmds.c node = ACPI_TO_POINTER(address); node 64 drivers/acpi/acpica/dbcmds.c if (!acpi_os_readable(node, sizeof(struct acpi_namespace_node))) { node 65 drivers/acpi/acpica/dbcmds.c acpi_os_printf("Address %p is invalid", node); node 71 drivers/acpi/acpica/dbcmds.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { node 74 drivers/acpi/acpica/dbcmds.c node, acpi_ut_get_descriptor_name(node)); node 82 drivers/acpi/acpica/dbcmds.c node = acpi_db_local_ns_lookup(in_string); node 83 drivers/acpi/acpica/dbcmds.c if (!node) { node 87 drivers/acpi/acpica/dbcmds.c node = acpi_gbl_root_node; node 91 drivers/acpi/acpica/dbcmds.c return (node); node 337 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 342 drivers/acpi/acpica/dbcmds.c node = acpi_db_convert_to_node(object_name); node 343 drivers/acpi/acpica/dbcmds.c if (!node) { node 347 drivers/acpi/acpica/dbcmds.c status = acpi_unload_parent_table(ACPI_CAST_PTR(acpi_handle, node)); node 350 drivers/acpi/acpica/dbcmds.c object_name, node); node 373 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 378 drivers/acpi/acpica/dbcmds.c node = acpi_db_convert_to_node(name); node 379 drivers/acpi/acpica/dbcmds.c if (!node) { node 385 drivers/acpi/acpica/dbcmds.c if (acpi_ev_is_notify_object(node)) { node 386 drivers/acpi/acpica/dbcmds.c status = acpi_ev_queue_notify_request(node, value); node 393 drivers/acpi/acpica/dbcmds.c acpi_ut_get_node_name(node), node 394 drivers/acpi/acpica/dbcmds.c acpi_ut_get_type_name(node->type)); node 493 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 499 drivers/acpi/acpica/dbcmds.c node = acpi_db_convert_to_node(buffer_arg); node 500 drivers/acpi/acpica/dbcmds.c if (!node || (node == acpi_gbl_root_node)) { node 507 drivers/acpi/acpica/dbcmds.c if (node->type != ACPI_TYPE_BUFFER) { node 519 drivers/acpi/acpica/dbcmds.c status = acpi_rs_create_resource_list(node->object, &return_buffer); node 538 drivers/acpi/acpica/dbcmds.c acpi_ut_debug_dump_buffer((u8 *)node->object->buffer.pointer, node 539 drivers/acpi/acpica/dbcmds.c node->object->buffer.length, node 659 drivers/acpi/acpica/dbcmds.c acpi_dm_test_resource_conversion(struct acpi_namespace_node *node, char *name) node 675 drivers/acpi/acpica/dbcmds.c status = acpi_evaluate_object(node, name, NULL, &return_buffer); node 684 drivers/acpi/acpica/dbcmds.c status = acpi_get_current_resources(node, &resource_buffer); node 755 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 764 drivers/acpi/acpica/dbcmds.c node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle); node 765 drivers/acpi/acpica/dbcmds.c parent_path = acpi_ns_get_normalized_pathname(node, TRUE); node 772 drivers/acpi/acpica/dbcmds.c (void)acpi_get_handle(node, METHOD_NAME__PRT, node 774 drivers/acpi/acpica/dbcmds.c (void)acpi_get_handle(node, METHOD_NAME__CRS, node 776 drivers/acpi/acpica/dbcmds.c (void)acpi_get_handle(node, METHOD_NAME__PRS, node 778 drivers/acpi/acpica/dbcmds.c (void)acpi_get_handle(node, METHOD_NAME__AEI, node 808 drivers/acpi/acpica/dbcmds.c status = acpi_get_irq_routing_table(node, &return_buffer); node 837 drivers/acpi/acpica/dbcmds.c status = acpi_walk_resources(node, METHOD_NAME__CRS, node 850 drivers/acpi/acpica/dbcmds.c status = acpi_get_current_resources(node, &return_buffer); node 879 drivers/acpi/acpica/dbcmds.c (void)acpi_dm_test_resource_conversion(node, METHOD_NAME__CRS); node 885 drivers/acpi/acpica/dbcmds.c status = acpi_set_current_resources(node, &return_buffer); node 916 drivers/acpi/acpica/dbcmds.c status = acpi_get_possible_resources(node, &return_buffer); node 948 drivers/acpi/acpica/dbcmds.c status = acpi_get_event_resources(node, &return_buffer); node 981 drivers/acpi/acpica/dbcmds.c struct acpi_namespace_node *node; node 996 drivers/acpi/acpica/dbcmds.c node = acpi_db_convert_to_node(object_arg); node 997 drivers/acpi/acpica/dbcmds.c if (node) { node 998 drivers/acpi/acpica/dbcmds.c if (node->type != ACPI_TYPE_DEVICE) { node 1001 drivers/acpi/acpica/dbcmds.c node->name.ascii, node 1002 drivers/acpi/acpica/dbcmds.c acpi_ut_get_type_name(node->type)); node 1004 drivers/acpi/acpica/dbcmds.c (void)acpi_db_device_resources(node, 0, NULL, node 142 drivers/acpi/acpica/dbdisply.c struct acpi_namespace_node *node; node 196 drivers/acpi/acpica/dbdisply.c node = obj_ptr; node 263 drivers/acpi/acpica/dbdisply.c node = acpi_db_local_ns_lookup(target); node 264 drivers/acpi/acpica/dbdisply.c if (!node) { node 271 drivers/acpi/acpica/dbdisply.c status = acpi_get_name(node, ACPI_FULL_PATHNAME_NO_TRAILING, &ret_buf); node 278 drivers/acpi/acpica/dbdisply.c node, (char *)ret_buf.pointer); node 281 drivers/acpi/acpica/dbdisply.c if (!acpi_os_readable(node, sizeof(struct acpi_namespace_node))) { node 282 drivers/acpi/acpica/dbdisply.c acpi_os_printf("Invalid Named object at address %p\n", node); node 286 drivers/acpi/acpica/dbdisply.c acpi_ut_debug_dump_buffer((void *)node, node 289 drivers/acpi/acpica/dbdisply.c acpi_ex_dump_namespace_node(node, 1); node 291 drivers/acpi/acpica/dbdisply.c obj_desc = acpi_ns_get_attached_object(node); node 352 drivers/acpi/acpica/dbdisply.c struct acpi_namespace_node *node; node 371 drivers/acpi/acpica/dbdisply.c node = walk_state->method_node; node 374 drivers/acpi/acpica/dbdisply.c acpi_ut_get_node_name(node)); node 506 drivers/acpi/acpica/dbdisply.c struct acpi_namespace_node *node; node 517 drivers/acpi/acpica/dbdisply.c node = walk_state->method_node; node 524 drivers/acpi/acpica/dbdisply.c acpi_ut_get_node_name(node), result_count); node 560 drivers/acpi/acpica/dbdisply.c struct acpi_namespace_node *node; node 568 drivers/acpi/acpica/dbdisply.c node = walk_state->method_node; node 572 drivers/acpi/acpica/dbdisply.c node = walk_state->method_node; node 573 drivers/acpi/acpica/dbdisply.c acpi_os_printf(" [%4.4s]\n", acpi_ut_get_node_name(node)); node 751 drivers/acpi/acpica/dbdisply.c status = acpi_get_name(gpe_block->node, node 759 drivers/acpi/acpica/dbdisply.c if (gpe_block->node == acpi_gbl_fadt_gpe_device) { node 767 drivers/acpi/acpica/dbdisply.c block, gpe_block, gpe_block->node, buffer, node 1059 drivers/acpi/acpica/dbdisply.c struct acpi_namespace_node *node = node 1065 drivers/acpi/acpica/dbdisply.c obj_desc = acpi_ns_get_attached_object(node); node 1070 drivers/acpi/acpica/dbdisply.c pathname = acpi_ns_get_normalized_pathname(node, TRUE); node 1089 drivers/acpi/acpica/dbdisply.c acpi_os_printf(" Device Name: %s (%p)\n", pathname, node); node 295 drivers/acpi/acpica/dbexec.c struct acpi_namespace_node *node = node 300 drivers/acpi/acpica/dbexec.c obj_desc = acpi_ns_get_attached_object(node); node 308 drivers/acpi/acpica/dbexec.c acpi_ns_print_node_pathname(node, "Evaluating"); node 315 drivers/acpi/acpica/dbexec.c status = acpi_evaluate_object(node, NULL, NULL, &return_obj); node 318 drivers/acpi/acpica/dbexec.c acpi_ut_get_node_name(node), node 115 drivers/acpi/acpica/dbmethod.c struct acpi_namespace_node *node; node 129 drivers/acpi/acpica/dbmethod.c node = acpi_db_convert_to_node(index_arg); node 130 drivers/acpi/acpica/dbmethod.c if (!node) { node 134 drivers/acpi/acpica/dbmethod.c if (node->type != ACPI_TYPE_INTEGER) { node 138 drivers/acpi/acpica/dbmethod.c obj_desc = node->object; node 360 drivers/acpi/acpica/dbmethod.c struct acpi_namespace_node *node = node 373 drivers/acpi/acpica/dbmethod.c predefined = acpi_ut_match_predefined_method(node->name.ascii); node 378 drivers/acpi/acpica/dbmethod.c if (node->type == ACPI_TYPE_LOCAL_SCOPE) { node 382 drivers/acpi/acpica/dbmethod.c pathname = acpi_ns_get_external_pathname(node); node 419 drivers/acpi/acpica/dbmethod.c status = acpi_evaluate_object(node, NULL, ¶m_objects, &return_obj); node 103 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node; node 117 drivers/acpi/acpica/dbnames.c ACPI_NS_NO_UPSEARCH, &node); node 127 drivers/acpi/acpica/dbnames.c ACPI_NS_NO_UPSEARCH, &node); node 147 drivers/acpi/acpica/dbnames.c acpi_gbl_db_scope_node = node; node 399 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node = node 407 drivers/acpi/acpica/dbnames.c predefined = acpi_ut_match_predefined_method(node->name.ascii); node 412 drivers/acpi/acpica/dbnames.c pathname = acpi_ns_get_normalized_pathname(node, TRUE); node 441 drivers/acpi/acpica/dbnames.c acpi_ns_check_acpi_compliance(pathname, node, predefined); node 492 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node = node 495 drivers/acpi/acpica/dbnames.c if (node->type > ACPI_TYPE_NS_NODE_MAX) { node 497 drivers/acpi/acpica/dbnames.c node->name.ascii, node->type); node 499 drivers/acpi/acpica/dbnames.c info->types[node->type]++; node 649 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node = node 659 drivers/acpi/acpica/dbnames.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { node 662 drivers/acpi/acpica/dbnames.c "is %2.2X should be %2.2X\n", node, node 663 drivers/acpi/acpica/dbnames.c acpi_ut_get_descriptor_name(node), node 664 drivers/acpi/acpica/dbnames.c ACPI_GET_DESCRIPTOR_TYPE(node), node 669 drivers/acpi/acpica/dbnames.c if ((node->type == ACPI_TYPE_LOCAL_ALIAS) || node 670 drivers/acpi/acpica/dbnames.c (node->type == ACPI_TYPE_LOCAL_METHOD_ALIAS)) { node 671 drivers/acpi/acpica/dbnames.c node = (struct acpi_namespace_node *)node->object; node 677 drivers/acpi/acpica/dbnames.c if (node->type > ACPI_TYPE_LOCAL_MAX) { node 679 drivers/acpi/acpica/dbnames.c node, node->type); node 683 drivers/acpi/acpica/dbnames.c if (!acpi_ut_valid_nameseg(node->name.ascii)) { node 684 drivers/acpi/acpica/dbnames.c acpi_os_printf("Invalid AcpiName for Node %p\n", node); node 688 drivers/acpi/acpica/dbnames.c object = acpi_ns_get_attached_object(node); node 749 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node = node 754 drivers/acpi/acpica/dbnames.c if (node == (void *)obj_desc) { node 756 drivers/acpi/acpica/dbnames.c acpi_ut_get_node_name(node)); node 761 drivers/acpi/acpica/dbnames.c if (acpi_ns_get_attached_object(node) == obj_desc) { node 763 drivers/acpi/acpica/dbnames.c node, acpi_ut_get_node_name(node)); node 815 drivers/acpi/acpica/dbnames.c struct acpi_namespace_node *node = node 823 drivers/acpi/acpica/dbnames.c if ((node->type != ACPI_TYPE_DEVICE) && node 824 drivers/acpi/acpica/dbnames.c (node->type != ACPI_TYPE_PROCESSOR)) { node 830 drivers/acpi/acpica/dbnames.c status = acpi_get_handle(node, METHOD_NAME__PRT, node 853 drivers/acpi/acpica/dbnames.c acpi_os_printf("%-32s Type %X", (char *)buffer.pointer, node->type); node 17 drivers/acpi/acpica/dbobject.c static void acpi_db_decode_node(struct acpi_namespace_node *node); node 38 drivers/acpi/acpica/dbobject.c struct acpi_namespace_node *node; node 40 drivers/acpi/acpica/dbobject.c node = walk_state->method_node; node 44 drivers/acpi/acpica/dbobject.c if (node == acpi_gbl_root_node) { node 156 drivers/acpi/acpica/dbobject.c static void acpi_db_decode_node(struct acpi_namespace_node *node) node 160 drivers/acpi/acpica/dbobject.c acpi_ut_get_node_name(node)); node 162 drivers/acpi/acpica/dbobject.c if (node->flags & ANOBJ_METHOD_ARG) { node 165 drivers/acpi/acpica/dbobject.c if (node->flags & ANOBJ_METHOD_LOCAL) { node 169 drivers/acpi/acpica/dbobject.c switch (node->type) { node 186 drivers/acpi/acpica/dbobject.c (node)); node 342 drivers/acpi/acpica/dbobject.c acpi_db_decode_node(obj_desc->reference.node); node 393 drivers/acpi/acpica/dbobject.c struct acpi_namespace_node *node; node 396 drivers/acpi/acpica/dbobject.c node = walk_state->method_node; node 400 drivers/acpi/acpica/dbobject.c if (node == acpi_gbl_root_node) { node 404 drivers/acpi/acpica/dbobject.c if (!node) { node 410 drivers/acpi/acpica/dbobject.c if (node->type != ACPI_TYPE_METHOD) { node 430 drivers/acpi/acpica/dbobject.c acpi_ut_get_node_name(node)); node 443 drivers/acpi/acpica/dbobject.c acpi_ut_get_node_name(node)); node 463 drivers/acpi/acpica/dbobject.c struct acpi_namespace_node *node; node 466 drivers/acpi/acpica/dbobject.c node = walk_state->method_node; node 471 drivers/acpi/acpica/dbobject.c if (node == acpi_gbl_root_node) { node 475 drivers/acpi/acpica/dbobject.c if (!node) { node 481 drivers/acpi/acpica/dbobject.c if (node->type != ACPI_TYPE_METHOD) { node 501 drivers/acpi/acpica/dbobject.c acpi_ut_get_node_name(node), node 502 drivers/acpi/acpica/dbobject.c node->object->method.param_count); node 515 drivers/acpi/acpica/dbobject.c acpi_ut_get_node_name(node)); node 218 drivers/acpi/acpica/dbstats.c struct acpi_namespace_node *node; node 224 drivers/acpi/acpica/dbstats.c node = (struct acpi_namespace_node *)obj_handle; node 225 drivers/acpi/acpica/dbstats.c obj_desc = acpi_ns_get_attached_object(node); node 229 drivers/acpi/acpica/dbstats.c type = node->type; node 26 drivers/acpi/acpica/dbtest.c acpi_db_test_integer_type(struct acpi_namespace_node *node, u32 bit_length); node 29 drivers/acpi/acpica/dbtest.c acpi_db_test_buffer_type(struct acpi_namespace_node *node, u32 bit_length); node 32 drivers/acpi/acpica/dbtest.c acpi_db_test_string_type(struct acpi_namespace_node *node, u32 byte_length); node 34 drivers/acpi/acpica/dbtest.c static acpi_status acpi_db_test_package_type(struct acpi_namespace_node *node); node 40 drivers/acpi/acpica/dbtest.c acpi_db_read_from_object(struct acpi_namespace_node *node, node 45 drivers/acpi/acpica/dbtest.c acpi_db_write_to_object(struct acpi_namespace_node *node, node 243 drivers/acpi/acpica/dbtest.c struct acpi_namespace_node *node; node 250 drivers/acpi/acpica/dbtest.c node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle); node 251 drivers/acpi/acpica/dbtest.c obj_desc = node->object; node 257 drivers/acpi/acpica/dbtest.c switch (node->type) { node 318 drivers/acpi/acpica/dbtest.c acpi_ut_get_type_name(node->type), node->name.ascii); node 330 drivers/acpi/acpica/dbtest.c status = acpi_db_test_integer_type(node, bit_length); node 335 drivers/acpi/acpica/dbtest.c status = acpi_db_test_string_type(node, byte_length); node 340 drivers/acpi/acpica/dbtest.c status = acpi_db_test_buffer_type(node, bit_length); node 345 drivers/acpi/acpica/dbtest.c status = acpi_db_test_package_type(node); node 388 drivers/acpi/acpica/dbtest.c acpi_db_test_integer_type(struct acpi_namespace_node *node, u32 bit_length) node 405 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp1); node 422 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 429 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp2); node 443 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 450 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp3); node 490 drivers/acpi/acpica/dbtest.c acpi_db_test_buffer_type(struct acpi_namespace_node *node, u32 bit_length) node 517 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp1); node 549 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 556 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp2); node 570 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 577 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp3); node 616 drivers/acpi/acpica/dbtest.c acpi_db_test_string_type(struct acpi_namespace_node *node, u32 byte_length) node 627 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp1); node 642 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 649 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp2); node 664 drivers/acpi/acpica/dbtest.c status = acpi_db_write_to_object(node, &write_value); node 671 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp3); node 706 drivers/acpi/acpica/dbtest.c static acpi_status acpi_db_test_package_type(struct acpi_namespace_node *node) node 713 drivers/acpi/acpica/dbtest.c status = acpi_db_read_from_object(node, ACPI_TYPE_PACKAGE, &temp1); node 785 drivers/acpi/acpica/dbtest.c region_obj->region.node->name.ascii); node 808 drivers/acpi/acpica/dbtest.c acpi_db_read_from_object(struct acpi_namespace_node *node, node 819 drivers/acpi/acpica/dbtest.c params[0].reference.actual_type = node->type; node 820 drivers/acpi/acpica/dbtest.c params[0].reference.handle = ACPI_CAST_PTR(acpi_handle, node); node 892 drivers/acpi/acpica/dbtest.c acpi_db_write_to_object(struct acpi_namespace_node *node, node 900 drivers/acpi/acpica/dbtest.c params[0].reference.actual_type = node->type; node 901 drivers/acpi/acpica/dbtest.c params[0].reference.handle = ACPI_CAST_PTR(acpi_handle, node); node 975 drivers/acpi/acpica/dbtest.c struct acpi_namespace_node *node = node 994 drivers/acpi/acpica/dbtest.c predefined = acpi_ut_match_predefined_method(node->name.ascii); node 999 drivers/acpi/acpica/dbtest.c if (node->type == ACPI_TYPE_LOCAL_SCOPE) { node 1003 drivers/acpi/acpica/dbtest.c pathname = acpi_ns_get_normalized_pathname(node, TRUE); node 1085 drivers/acpi/acpica/dbtest.c status = acpi_evaluate_object(node, NULL, ¶m_objects, &return_obj); node 248 drivers/acpi/acpica/dbutils.c struct acpi_namespace_node *node = NULL; node 267 drivers/acpi/acpica/dbutils.c NULL, &node); node 274 drivers/acpi/acpica/dbutils.c return (node); node 23 drivers/acpi/acpica/dsargs.c acpi_ds_execute_arguments(struct acpi_namespace_node *node, node 43 drivers/acpi/acpica/dsargs.c acpi_ds_execute_arguments(struct acpi_namespace_node *node, node 62 drivers/acpi/acpica/dsargs.c op->common.node = scope_node; node 82 drivers/acpi/acpica/dsargs.c walk_state->deferred_node = node; node 93 drivers/acpi/acpica/dsargs.c op->common.node = node; node 103 drivers/acpi/acpica/dsargs.c op->common.node = scope_node; node 124 drivers/acpi/acpica/dsargs.c walk_state->deferred_node = node; node 149 drivers/acpi/acpica/dsargs.c struct acpi_namespace_node *node; node 161 drivers/acpi/acpica/dsargs.c node = obj_desc->buffer_field.node; node 164 drivers/acpi/acpica/dsargs.c (ACPI_TYPE_BUFFER_FIELD, node, NULL)); node 167 drivers/acpi/acpica/dsargs.c acpi_ut_get_node_name(node))); node 171 drivers/acpi/acpica/dsargs.c status = acpi_ds_execute_arguments(node, node->parent, node 194 drivers/acpi/acpica/dsargs.c struct acpi_namespace_node *node; node 206 drivers/acpi/acpica/dsargs.c node = obj_desc->bank_field.node; node 209 drivers/acpi/acpica/dsargs.c (ACPI_TYPE_LOCAL_BANK_FIELD, node, NULL)); node 212 drivers/acpi/acpica/dsargs.c acpi_ut_get_node_name(node))); node 216 drivers/acpi/acpica/dsargs.c status = acpi_ds_execute_arguments(node, node->parent, node 225 drivers/acpi/acpica/dsargs.c obj_desc->region.length, node); node 244 drivers/acpi/acpica/dsargs.c struct acpi_namespace_node *node; node 255 drivers/acpi/acpica/dsargs.c node = obj_desc->buffer.node; node 256 drivers/acpi/acpica/dsargs.c if (!node) { node 267 drivers/acpi/acpica/dsargs.c status = acpi_ds_execute_arguments(node, node, node 288 drivers/acpi/acpica/dsargs.c struct acpi_namespace_node *node; node 299 drivers/acpi/acpica/dsargs.c node = obj_desc->package.node; node 300 drivers/acpi/acpica/dsargs.c if (!node) { node 312 drivers/acpi/acpica/dsargs.c status = acpi_ds_execute_arguments(node, node, node 334 drivers/acpi/acpica/dsargs.c struct acpi_namespace_node *node; node 351 drivers/acpi/acpica/dsargs.c node = obj_desc->region.node; node 354 drivers/acpi/acpica/dsargs.c (ACPI_TYPE_REGION, node, NULL)); node 358 drivers/acpi/acpica/dsargs.c acpi_ut_get_node_name(node), node 363 drivers/acpi/acpica/dsargs.c status = acpi_ds_execute_arguments(node, extra_desc->extra.scope_node, node 372 drivers/acpi/acpica/dsargs.c obj_desc->region.length, node); node 25 drivers/acpi/acpica/dsdebug.c acpi_ds_print_node_pathname(struct acpi_namespace_node *node, node 41 drivers/acpi/acpica/dsdebug.c acpi_ds_print_node_pathname(struct acpi_namespace_node *node, node 49 drivers/acpi/acpica/dsdebug.c if (!node) { node 58 drivers/acpi/acpica/dsdebug.c status = acpi_ns_handle_to_pathname(node, &buffer, TRUE); node 66 drivers/acpi/acpica/dsdebug.c (char *)buffer.pointer, node)); node 145 drivers/acpi/acpica/dsdebug.c method_desc->method.node, node 33 drivers/acpi/acpica/dsfield.c struct acpi_namespace_node **node); node 64 drivers/acpi/acpica/dsfield.c struct acpi_namespace_node **node) node 82 drivers/acpi/acpica/dsfield.c walk_state, node); node 94 drivers/acpi/acpica/dsfield.c obj_desc->region.node = *node; node 95 drivers/acpi/acpica/dsfield.c status = acpi_ns_attach_object(*node, obj_desc, ACPI_TYPE_REGION); node 124 drivers/acpi/acpica/dsfield.c struct acpi_namespace_node *node; node 151 drivers/acpi/acpica/dsfield.c node = walk_state->deferred_node; node 180 drivers/acpi/acpica/dsfield.c walk_state, &node); node 193 drivers/acpi/acpica/dsfield.c op->common.node = node; node 200 drivers/acpi/acpica/dsfield.c obj_desc = acpi_ns_get_attached_object(node); node 225 drivers/acpi/acpica/dsfield.c obj_desc->buffer_field.node = node; node 229 drivers/acpi/acpica/dsfield.c status = acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_BUFFER_FIELD); node 390 drivers/acpi/acpica/dsfield.c arg->common.node = info->field_node; node 552 drivers/acpi/acpica/dsfield.c struct acpi_namespace_node *node; node 629 drivers/acpi/acpica/dsfield.c walk_state, &node); node 643 drivers/acpi/acpica/dsfield.c arg->common.node = node; node 51 drivers/acpi/acpica/dsinit.c struct acpi_namespace_node *node = node 62 drivers/acpi/acpica/dsinit.c if (node->owner_id != info->owner_id) { node 94 drivers/acpi/acpica/dsinit.c obj_desc = acpi_ns_get_attached_object(node); node 110 drivers/acpi/acpica/dsinit.c acpi_ds_auto_serialize_method(node, obj_desc); node 55 drivers/acpi/acpica/dsmethod.c acpi_ds_auto_serialize_method(struct acpi_namespace_node *node, node 62 drivers/acpi/acpica/dsmethod.c ACPI_FUNCTION_TRACE_PTR(ds_auto_serialize_method, node); node 66 drivers/acpi/acpica/dsmethod.c acpi_ut_get_node_name(node), node)); node 75 drivers/acpi/acpica/dsmethod.c acpi_ps_set_name(op, node->name.integer); node 76 drivers/acpi/acpica/dsmethod.c op->common.node = node; node 81 drivers/acpi/acpica/dsmethod.c acpi_ds_create_walk_state(node->owner_id, NULL, NULL, NULL); node 87 drivers/acpi/acpica/dsmethod.c status = acpi_ds_init_aml_walk(walk_state, op, node, node 839 drivers/acpi/acpica/dsmethod.c method.node, method_desc, walk_state); node 216 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node **node) node 235 drivers/acpi/acpica/dsmthdat.c *node = &walk_state->local_variables[index]; node 249 drivers/acpi/acpica/dsmthdat.c *node = &walk_state->arguments[index]; node 285 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node *node; node 296 drivers/acpi/acpica/dsmthdat.c status = acpi_ds_method_data_get_node(type, index, walk_state, &node); node 311 drivers/acpi/acpica/dsmthdat.c node->object = object; node 339 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node *node; node 353 drivers/acpi/acpica/dsmthdat.c status = acpi_ds_method_data_get_node(type, index, walk_state, &node); node 360 drivers/acpi/acpica/dsmthdat.c object = node->object; node 381 drivers/acpi/acpica/dsmthdat.c node->object = object; node 392 drivers/acpi/acpica/dsmthdat.c index, node)); node 443 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node *node; node 450 drivers/acpi/acpica/dsmthdat.c status = acpi_ds_method_data_get_node(type, index, walk_state, &node); node 457 drivers/acpi/acpica/dsmthdat.c object = acpi_ns_get_attached_object(node); node 464 drivers/acpi/acpica/dsmthdat.c node->object = NULL; node 504 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node *node; node 520 drivers/acpi/acpica/dsmthdat.c status = acpi_ds_method_data_get_node(type, index, walk_state, &node); node 525 drivers/acpi/acpica/dsmthdat.c current_obj_desc = acpi_ns_get_attached_object(node); node 655 drivers/acpi/acpica/dsmthdat.c struct acpi_namespace_node *node; node 662 drivers/acpi/acpica/dsmthdat.c status = acpi_ds_method_data_get_node(opcode, index, walk_state, &node); node 669 drivers/acpi/acpica/dsmthdat.c object = acpi_ns_get_attached_object(node); node 52 drivers/acpi/acpica/dsobject.c if (!op->common.node) { node 78 drivers/acpi/acpica/dsobject.c &(op->common.node))); node 118 drivers/acpi/acpica/dsobject.c !obj_desc->reference.node) { node 123 drivers/acpi/acpica/dsobject.c obj_desc->reference.node = node 124 drivers/acpi/acpica/dsobject.c walk_state->scope_info->scope.node; node 234 drivers/acpi/acpica/dsobject.c op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc); node 254 drivers/acpi/acpica/dsobject.c struct acpi_namespace_node *node, node 267 drivers/acpi/acpica/dsobject.c if (acpi_ns_get_attached_object(node)) { node 289 drivers/acpi/acpica/dsobject.c node->type = obj_desc->common.type; node 293 drivers/acpi/acpica/dsobject.c status = acpi_ns_attach_object(node, obj_desc, node->type); node 346 drivers/acpi/acpica/dsobject.c obj_desc->buffer.node = node 361 drivers/acpi/acpica/dsobject.c obj_desc->package.node = node 511 drivers/acpi/acpica/dsobject.c obj_desc->reference.node = op->common.node; node 513 drivers/acpi/acpica/dsobject.c if (op->common.node) { node 515 drivers/acpi/acpica/dsobject.c op->common.node->object; node 271 drivers/acpi/acpica/dsopcode.c struct acpi_namespace_node *node; node 280 drivers/acpi/acpica/dsopcode.c node = op->common.node; node 293 drivers/acpi/acpica/dsopcode.c obj_desc = acpi_ns_get_attached_object(node); node 357 drivers/acpi/acpica/dsopcode.c struct acpi_namespace_node *node; node 367 drivers/acpi/acpica/dsopcode.c node = op->common.node; node 394 drivers/acpi/acpica/dsopcode.c obj_desc = acpi_ns_get_attached_object(node); node 414 drivers/acpi/acpica/dsopcode.c node->name.ascii, space_id)); node 434 drivers/acpi/acpica/dsopcode.c obj_desc->region.length, node); node 464 drivers/acpi/acpica/dsopcode.c struct acpi_namespace_node *node; node 475 drivers/acpi/acpica/dsopcode.c node = op->common.node; node 524 drivers/acpi/acpica/dsopcode.c obj_desc = acpi_ns_get_attached_object(node); node 684 drivers/acpi/acpica/dsopcode.c struct acpi_namespace_node *node; node 741 drivers/acpi/acpica/dsopcode.c node = arg->common.node; node 743 drivers/acpi/acpica/dsopcode.c obj_desc = acpi_ns_get_attached_object(node); node 98 drivers/acpi/acpica/dspkginit.c obj_desc->package.node = parent->common.node; node 155 drivers/acpi/acpica/dspkginit.c if (!arg->common.node) { node 181 drivers/acpi/acpica/dspkginit.c if (arg->common.node->type == ACPI_TYPE_METHOD) { node 198 drivers/acpi/acpica/dspkginit.c arg->common.node); node 269 drivers/acpi/acpica/dspkginit.c if (arg->common.node) { node 273 drivers/acpi/acpica/dspkginit.c arg->common.node)); node 274 drivers/acpi/acpica/dspkginit.c arg->common.node = NULL; node 307 drivers/acpi/acpica/dspkginit.c op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc); node 406 drivers/acpi/acpica/dspkginit.c scope_info.scope.node = element->reference.node; /* Prefix node */ node 457 drivers/acpi/acpica/dspkginit.c scope_info.scope.node->name.ascii)); node 474 drivers/acpi/acpica/dspkginit.c element->reference.node = resolved_node; node 475 drivers/acpi/acpica/dspkginit.c type = element->reference.node->type; node 284 drivers/acpi/acpica/dswexec.c acpi_ds_scope_stack_push(op->named.node, node 285 drivers/acpi/acpica/dswexec.c op->named.node-> node 478 drivers/acpi/acpica/dswexec.c op->common.node = (struct acpi_namespace_node *) node 479 drivers/acpi/acpica/dswexec.c op->asl.value.arg->asl.node; node 481 drivers/acpi/acpica/dswexec.c node->object); node 556 drivers/acpi/acpica/dswexec.c acpi_ns_get_external_pathname(op->common.node); node 562 drivers/acpi/acpica/dswexec.c node->object, node 590 drivers/acpi/acpica/dswexec.c op->common.parent->common.node; node 595 drivers/acpi/acpica/dswexec.c common.node, node 611 drivers/acpi/acpica/dswexec.c node)); node 107 drivers/acpi/acpica/dswload.c struct acpi_namespace_node *node; node 129 drivers/acpi/acpica/dswload.c if (op->common.node) { node 155 drivers/acpi/acpica/dswload.c walk_state, &(node)); node 169 drivers/acpi/acpica/dswload.c &node); node 182 drivers/acpi/acpica/dswload.c switch (node->type) { node 209 drivers/acpi/acpica/dswload.c acpi_ut_get_node_name(node), node 210 drivers/acpi/acpica/dswload.c acpi_ut_get_type_name(node->type))); node 212 drivers/acpi/acpica/dswload.c node->type = ACPI_TYPE_ANY; node 221 drivers/acpi/acpica/dswload.c if ((node == acpi_gbl_root_node) && node 236 drivers/acpi/acpica/dswload.c acpi_ut_get_type_name(node->type), node 237 drivers/acpi/acpica/dswload.c acpi_ut_get_node_name(node))); node 264 drivers/acpi/acpica/dswload.c node = walk_state->deferred_node; node 274 drivers/acpi/acpica/dswload.c node = NULL; node 310 drivers/acpi/acpica/dswload.c &node); node 316 drivers/acpi/acpica/dswload.c if (node->flags & ANOBJ_IS_EXTERNAL) { node 321 drivers/acpi/acpica/dswload.c node->flags &= ~ANOBJ_IS_EXTERNAL; node 322 drivers/acpi/acpica/dswload.c node->type = (u8) object_type; node 329 drivers/acpi/acpica/dswload.c (node, object_type, node 368 drivers/acpi/acpica/dswload.c if (node) { node 373 drivers/acpi/acpica/dswload.c op->common.node = node; node 374 drivers/acpi/acpica/dswload.c op->named.name = node->name.integer; node 499 drivers/acpi/acpica/dswload.c if (op->common.node) { node 500 drivers/acpi/acpica/dswload.c op->common.node->type = (u8) object_type; node 510 drivers/acpi/acpica/dswload.c op->common.node && op->common.aml_opcode == AML_EXTERNAL_OP) { node 519 drivers/acpi/acpica/dswload.c op->common.node->flags |= ANOBJ_IS_EXTERNAL; node 520 drivers/acpi/acpica/dswload.c op->common.node->type = (u8)object_type; node 523 drivers/acpi/acpica/dswload.c &op->common.node, node 553 drivers/acpi/acpica/dswload.c walk_state, op, op->named.node)); node 555 drivers/acpi/acpica/dswload.c if (!acpi_ns_get_attached_object(op->named.node)) { node 557 drivers/acpi/acpica/dswload.c ACPI_CAST_PTR(void, op->named.node); node 42 drivers/acpi/acpica/dswload2.c struct acpi_namespace_node *node; node 111 drivers/acpi/acpica/dswload2.c node = NULL; node 124 drivers/acpi/acpica/dswload2.c ACPI_NS_SEARCH_PARENT, walk_state, &(node)); node 131 drivers/acpi/acpica/dswload2.c if (op && (op->named.node == acpi_gbl_root_node)) { node 132 drivers/acpi/acpica/dswload2.c node = op->named.node; node 135 drivers/acpi/acpica/dswload2.c acpi_ds_scope_stack_push(node, object_type, node 150 drivers/acpi/acpica/dswload2.c &(node)); node 173 drivers/acpi/acpica/dswload2.c switch (node->type) { node 198 drivers/acpi/acpica/dswload2.c acpi_ut_get_node_name(node), node 199 drivers/acpi/acpica/dswload2.c acpi_ut_get_type_name(node->type))); node 201 drivers/acpi/acpica/dswload2.c node->type = ACPI_TYPE_ANY; node 211 drivers/acpi/acpica/dswload2.c if ((node == acpi_gbl_root_node) && node 226 drivers/acpi/acpica/dswload2.c acpi_ut_get_type_name(node->type), node 227 drivers/acpi/acpica/dswload2.c acpi_ut_get_node_name(node))); node 237 drivers/acpi/acpica/dswload2.c if (op && op->common.node) { node 241 drivers/acpi/acpica/dswload2.c node = op->common.node; node 245 drivers/acpi/acpica/dswload2.c acpi_ds_scope_stack_push(node, object_type, node 267 drivers/acpi/acpica/dswload2.c node = walk_state->deferred_node; node 315 drivers/acpi/acpica/dswload2.c walk_state, &node); node 320 drivers/acpi/acpica/dswload2.c acpi_ut_get_node_name(node), node)); node 342 drivers/acpi/acpica/dswload2.c if (node) { node 343 drivers/acpi/acpica/dswload2.c op->named.name = node->name.integer; node 352 drivers/acpi/acpica/dswload2.c op->common.node = node; node 374 drivers/acpi/acpica/dswload2.c struct acpi_namespace_node *node; node 408 drivers/acpi/acpica/dswload2.c node = op->common.node; node 414 drivers/acpi/acpica/dswload2.c walk_state->operands[0] = (void *)node; node 462 drivers/acpi/acpica/dswload2.c walk_state, op, node)); node 497 drivers/acpi/acpica/dswload2.c node, walk_state); node 503 drivers/acpi/acpica/dswload2.c acpi_ds_create_bank_field(op, arg->common.node, node 510 drivers/acpi/acpica/dswload2.c acpi_ds_create_field(op, arg->common.node, node 613 drivers/acpi/acpica/dswload2.c (acpi_ns_get_attached_object(node)); node 618 drivers/acpi/acpica/dswload2.c status = acpi_ds_create_node(walk_state, node, op); node 627 drivers/acpi/acpica/dswload2.c namepath = acpi_ns_get_external_pathname(node); node 633 drivers/acpi/acpica/dswload2.c if (node->object) { node 634 drivers/acpi/acpica/dswload2.c acpi_ns_detach_object(node); node 636 drivers/acpi/acpica/dswload2.c acpi_ns_attach_object(node, obj_desc, node 655 drivers/acpi/acpica/dswload2.c walk_state, op, op->named.node)); node 657 drivers/acpi/acpica/dswload2.c if (!acpi_ns_get_attached_object(op->named.node)) { node 659 drivers/acpi/acpica/dswload2.c ACPI_CAST_PTR(void, op->named.node); node 700 drivers/acpi/acpica/dswload2.c walk_state, op, node)); node 726 drivers/acpi/acpica/dswload2.c op->common.node = new_node; node 67 drivers/acpi/acpica/dswscope.c acpi_ds_scope_stack_push(struct acpi_namespace_node *node, node 76 drivers/acpi/acpica/dswscope.c if (!node) { node 100 drivers/acpi/acpica/dswscope.c scope_info->scope.node = node; node 114 drivers/acpi/acpica/dswscope.c scope.node), node 123 drivers/acpi/acpica/dswscope.c acpi_ut_get_node_name(scope_info->scope.node), node 164 drivers/acpi/acpica/dswscope.c acpi_ut_get_node_name(scope_info->scope.node), node 171 drivers/acpi/acpica/dswscope.c scope.node), node 630 drivers/acpi/acpica/dswstate.c while (extra_op && !extra_op->common.node) { node 637 drivers/acpi/acpica/dswstate.c parser_state->start_node = extra_op->common.node; node 378 drivers/acpi/acpica/evgpe.c gpe_device = gpe_block->node; node 328 drivers/acpi/acpica/evgpeblk.c gpe_block->node = gpe_device; node 230 drivers/acpi/acpica/evgpeinit.c walk_info.gpe_device = gpe_block->node; node 103 drivers/acpi/acpica/evgpeutil.c if ((gpe_block->node)->type == ACPI_TYPE_DEVICE) { node 104 drivers/acpi/acpica/evgpeutil.c info->gpe_device = gpe_block->node; node 118 drivers/acpi/acpica/evhandler.c acpi_ev_has_default_handler(struct acpi_namespace_node *node, node 126 drivers/acpi/acpica/evhandler.c obj_desc = acpi_ns_get_attached_object(node); node 171 drivers/acpi/acpica/evhandler.c struct acpi_namespace_node *node; node 186 drivers/acpi/acpica/evhandler.c node = acpi_ns_validate_handle(obj_handle); node 187 drivers/acpi/acpica/evhandler.c if (!node) { node 195 drivers/acpi/acpica/evhandler.c if ((node->type != ACPI_TYPE_DEVICE) && node 196 drivers/acpi/acpica/evhandler.c (node->type != ACPI_TYPE_REGION) && (node != acpi_gbl_root_node)) { node 202 drivers/acpi/acpica/evhandler.c obj_desc = acpi_ns_get_attached_object(node); node 328 drivers/acpi/acpica/evhandler.c acpi_ev_install_space_handler(struct acpi_namespace_node *node, node 345 drivers/acpi/acpica/evhandler.c if ((node->type != ACPI_TYPE_DEVICE) && node 346 drivers/acpi/acpica/evhandler.c (node->type != ACPI_TYPE_PROCESSOR) && node 347 drivers/acpi/acpica/evhandler.c (node->type != ACPI_TYPE_THERMAL) && (node != acpi_gbl_root_node)) { node 407 drivers/acpi/acpica/evhandler.c obj_desc = acpi_ns_get_attached_object(node); node 438 drivers/acpi/acpica/evhandler.c node)); node 442 drivers/acpi/acpica/evhandler.c if (node->type == ACPI_TYPE_ANY) { node 445 drivers/acpi/acpica/evhandler.c type = node->type; node 460 drivers/acpi/acpica/evhandler.c status = acpi_ns_attach_object(node, obj_desc, type); node 475 drivers/acpi/acpica/evhandler.c acpi_ut_get_node_name(node), node, obj_desc)); node 495 drivers/acpi/acpica/evhandler.c handler_obj->address_space.node = node; node 520 drivers/acpi/acpica/evhandler.c status = acpi_ns_walk_namespace(ACPI_TYPE_ANY, node, node 35 drivers/acpi/acpica/evmisc.c u8 acpi_ev_is_notify_object(struct acpi_namespace_node *node) node 38 drivers/acpi/acpica/evmisc.c switch (node->type) { node 68 drivers/acpi/acpica/evmisc.c acpi_ev_queue_notify_request(struct acpi_namespace_node *node, u32 notify_value) node 80 drivers/acpi/acpica/evmisc.c if (!acpi_ev_is_notify_object(node)) { node 94 drivers/acpi/acpica/evmisc.c obj_desc = acpi_ns_get_attached_object(node); node 111 drivers/acpi/acpica/evmisc.c acpi_ut_get_node_name(node), notify_value, node 112 drivers/acpi/acpica/evmisc.c node)); node 126 drivers/acpi/acpica/evmisc.c info->notify.node = node; node 134 drivers/acpi/acpica/evmisc.c acpi_ut_get_node_name(node), node 135 drivers/acpi/acpica/evmisc.c acpi_ut_get_type_name(node->type), notify_value, node 137 drivers/acpi/acpica/evmisc.c node)); node 171 drivers/acpi/acpica/evmisc.c info->notify.global->handler(info->notify.node, node 180 drivers/acpi/acpica/evmisc.c handler_obj->notify.handler(info->notify.node, node 129 drivers/acpi/acpica/evregion.c acpi_ut_get_node_name(region_obj->region.node), node 488 drivers/acpi/acpica/evregion.c acpi_ut_get_node_name(region_obj->region.node), node 525 drivers/acpi/acpica/evregion.c struct acpi_namespace_node *node; node 545 drivers/acpi/acpica/evregion.c node = region_obj->region.node->parent; node 547 drivers/acpi/acpica/evregion.c acpi_ns_search_one_scope(*reg_name_ptr, node, ACPI_TYPE_METHOD, node 649 drivers/acpi/acpica/evregion.c acpi_ev_execute_reg_methods(struct acpi_namespace_node *node, node 683 drivers/acpi/acpica/evregion.c (void)acpi_ns_walk_namespace(ACPI_TYPE_ANY, node, ACPI_UINT32_MAX, node 690 drivers/acpi/acpica/evregion.c acpi_ev_orphan_ec_reg_method(node); node 716 drivers/acpi/acpica/evregion.c struct acpi_namespace_node *node; node 724 drivers/acpi/acpica/evregion.c node = acpi_ns_validate_handle(obj_handle); node 725 drivers/acpi/acpica/evregion.c if (!node) { node 733 drivers/acpi/acpica/evregion.c if ((node->type != ACPI_TYPE_REGION) && (node != acpi_gbl_root_node)) { node 739 drivers/acpi/acpica/evregion.c obj_desc = acpi_ns_get_attached_object(node); node 165 drivers/acpi/acpica/evrgnini.c parent_node = region_obj->region.node->parent; node 180 drivers/acpi/acpica/evrgnini.c if (handler_obj->address_space.node == acpi_gbl_root_node) { node 218 drivers/acpi/acpica/evrgnini.c pci_root_node = handler_obj->address_space.node; node 243 drivers/acpi/acpica/evrgnini.c pci_device_node = region_obj->region.node; node 289 drivers/acpi/acpica/evrgnini.c region_obj->region.node); node 312 drivers/acpi/acpica/evrgnini.c u8 acpi_ev_is_pci_root_bridge(struct acpi_namespace_node *node) node 322 drivers/acpi/acpica/evrgnini.c status = acpi_ut_execute_HID(node, &hid); node 336 drivers/acpi/acpica/evrgnini.c status = acpi_ut_execute_CID(node, &cid); node 480 drivers/acpi/acpica/evrgnini.c struct acpi_namespace_node *node; node 494 drivers/acpi/acpica/evrgnini.c node = region_obj->region.node->parent; node 501 drivers/acpi/acpica/evrgnini.c while (node) { node 506 drivers/acpi/acpica/evrgnini.c obj_desc = acpi_ns_get_attached_object(node); node 511 drivers/acpi/acpica/evrgnini.c switch (node->type) { node 554 drivers/acpi/acpica/evrgnini.c node = node->parent; node 61 drivers/acpi/acpica/evxface.c struct acpi_namespace_node *node = node 114 drivers/acpi/acpica/evxface.c if (!acpi_ev_is_notify_object(node)) { node 121 drivers/acpi/acpica/evxface.c obj_desc = acpi_ns_get_attached_object(node); node 126 drivers/acpi/acpica/evxface.c obj_desc = acpi_ut_create_internal_object(node->type); node 134 drivers/acpi/acpica/evxface.c status = acpi_ns_attach_object(device, obj_desc, node->type); node 165 drivers/acpi/acpica/evxface.c handler_obj->notify.node = node; node 214 drivers/acpi/acpica/evxface.c struct acpi_namespace_node *node = node 268 drivers/acpi/acpica/evxface.c if (!acpi_ev_is_notify_object(node)) { node 274 drivers/acpi/acpica/evxface.c obj_desc = acpi_ns_get_attached_object(node); node 859 drivers/acpi/acpica/evxfgpe.c struct acpi_namespace_node *node; node 873 drivers/acpi/acpica/evxfgpe.c node = acpi_ns_validate_handle(gpe_device); node 874 drivers/acpi/acpica/evxfgpe.c if (!node) { node 881 drivers/acpi/acpica/evxfgpe.c if (node->type != ACPI_TYPE_DEVICE) { node 886 drivers/acpi/acpica/evxfgpe.c if (node->object) { node 895 drivers/acpi/acpica/evxfgpe.c status = acpi_ev_create_gpe_block(node, gpe_block_address->address, node 905 drivers/acpi/acpica/evxfgpe.c obj_desc = acpi_ns_get_attached_object(node); node 919 drivers/acpi/acpica/evxfgpe.c acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_DEVICE); node 956 drivers/acpi/acpica/evxfgpe.c struct acpi_namespace_node *node; node 969 drivers/acpi/acpica/evxfgpe.c node = acpi_ns_validate_handle(gpe_device); node 970 drivers/acpi/acpica/evxfgpe.c if (!node) { node 977 drivers/acpi/acpica/evxfgpe.c if (node->type != ACPI_TYPE_DEVICE) { node 984 drivers/acpi/acpica/evxfgpe.c obj_desc = acpi_ns_get_attached_object(node); node 48 drivers/acpi/acpica/evxfregn.c struct acpi_namespace_node *node; node 66 drivers/acpi/acpica/evxfregn.c node = acpi_ns_validate_handle(device); node 67 drivers/acpi/acpica/evxfregn.c if (!node) { node 75 drivers/acpi/acpica/evxfregn.c acpi_ev_install_space_handler(node, space_id, handler, setup, node 83 drivers/acpi/acpica/evxfregn.c acpi_ev_execute_reg_methods(node, space_id, ACPI_REG_CONNECT); node 114 drivers/acpi/acpica/evxfregn.c struct acpi_namespace_node *node; node 132 drivers/acpi/acpica/evxfregn.c node = acpi_ns_validate_handle(device); node 133 drivers/acpi/acpica/evxfregn.c if (!node || node 134 drivers/acpi/acpica/evxfregn.c ((node->type != ACPI_TYPE_DEVICE) && node 135 drivers/acpi/acpica/evxfregn.c (node->type != ACPI_TYPE_PROCESSOR) && node 136 drivers/acpi/acpica/evxfregn.c (node->type != ACPI_TYPE_THERMAL) && node 137 drivers/acpi/acpica/evxfregn.c (node != acpi_gbl_root_node))) { node 144 drivers/acpi/acpica/evxfregn.c obj_desc = acpi_ns_get_attached_object(node); node 174 drivers/acpi/acpica/evxfregn.c node, obj_desc)); node 219 drivers/acpi/acpica/evxfregn.c node, obj_desc)); node 120 drivers/acpi/acpica/exconfig.c start_node = walk_state->scope_info->scope.node; node 181 drivers/acpi/acpica/excreate.c obj_desc->mutex.node = node 185 drivers/acpi/acpica/excreate.c acpi_ns_attach_object(obj_desc->mutex.node, obj_desc, node 219 drivers/acpi/acpica/excreate.c struct acpi_namespace_node *node; node 226 drivers/acpi/acpica/excreate.c node = walk_state->op->common.node; node 232 drivers/acpi/acpica/excreate.c if (acpi_ns_get_attached_object(node)) { node 272 drivers/acpi/acpica/excreate.c walk_state->scope_info->scope.node; node 274 drivers/acpi/acpica/excreate.c region_obj2->extra.scope_node = node; node 282 drivers/acpi/acpica/excreate.c obj_desc->region.node = node; node 290 drivers/acpi/acpica/excreate.c status = acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_REGION); node 430 drivers/acpi/acpica/excreate.c obj_desc->method.node = operand[0]; node 205 drivers/acpi/acpica/exdebug.c if (source_desc->reference.node) { node 207 drivers/acpi/acpica/exdebug.c (source_desc->reference.node) != node 211 drivers/acpi/acpica/exdebug.c source_desc->reference.node); node 214 drivers/acpi/acpica/exdebug.c source_desc->reference.node, node 215 drivers/acpi/acpica/exdebug.c (source_desc->reference.node)-> node 218 drivers/acpi/acpica/exdebug.c switch ((source_desc->reference.node)->type) { node 234 drivers/acpi/acpica/exdebug.c node)->object, node 63 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(buffer.node), "Parent Node"}, node 69 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(package.node), "Parent Node"}, node 118 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(region.node), "Parent Node"}, node 199 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(reference.node), "Node"}, node 213 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(address_space.node), "Node"}, node 219 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(notify.node), "Node"}, node 267 drivers/acpi/acpica/exdump.c {ACPI_EXD_NODE, ACPI_EXD_OFFSET(common_field.node), "Parent Node"} node 339 drivers/acpi/acpica/exdump.c struct acpi_namespace_node *node; node 545 drivers/acpi/acpica/exdump.c node = node 549 drivers/acpi/acpica/exdump.c acpi_os_printf("%20s : %p", name, node); node 550 drivers/acpi/acpica/exdump.c if (node) { node 551 drivers/acpi/acpica/exdump.c acpi_os_printf(" [%4.4s]", node->name.ascii); node 668 drivers/acpi/acpica/exdump.c acpi_ut_repair_name(obj_desc->reference.node->name. node 671 drivers/acpi/acpica/exdump.c obj_desc->reference.node->name.ascii, node 672 drivers/acpi/acpica/exdump.c obj_desc->reference.node); node 935 drivers/acpi/acpica/exdump.c void acpi_ex_dump_namespace_node(struct acpi_namespace_node *node, u32 flags) node 949 drivers/acpi/acpica/exdump.c acpi_os_printf("%20s : %4.4s\n", "Name", acpi_ut_get_node_name(node)); node 951 drivers/acpi/acpica/exdump.c node->type, acpi_ut_get_type_name(node->type)); node 953 drivers/acpi/acpica/exdump.c acpi_ex_dump_object(ACPI_CAST_PTR(union acpi_operand_object, node), node 975 drivers/acpi/acpica/exdump.c acpi_os_printf(" %p ", obj_desc->reference.node); node 977 drivers/acpi/acpica/exdump.c status = acpi_ns_handle_to_pathname(obj_desc->reference.node, node 986 drivers/acpi/acpica/exdump.c reference.node-> node 152 drivers/acpi/acpica/exfldio.c common_field.node), node 155 drivers/acpi/acpica/exfldio.c node), node 166 drivers/acpi/acpica/exfldio.c acpi_ut_get_node_name(obj_desc->common_field.node), node 170 drivers/acpi/acpica/exfldio.c acpi_ut_get_node_name(rgn_desc->region.node), node 208 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node))); node 220 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node), node 359 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node))); node 368 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node))); node 381 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node), node 397 drivers/acpi/acpica/exmutex.c acpi_ut_get_node_name(obj_desc->mutex.node), node 474 drivers/acpi/acpica/exmutex.c obj_desc->mutex.node->name.ascii, node 861 drivers/acpi/acpica/exoparg1.c node, node 58 drivers/acpi/acpica/exoparg2.c struct acpi_namespace_node *node; node 72 drivers/acpi/acpica/exoparg2.c node = (struct acpi_namespace_node *)operand[0]; node 80 drivers/acpi/acpica/exoparg2.c if (!acpi_ev_is_notify_object(node)) { node 83 drivers/acpi/acpica/exoparg2.c acpi_ut_get_type_name(node->type))); node 96 drivers/acpi/acpica/exoparg2.c status = acpi_ev_queue_notify_request(node, value); node 415 drivers/acpi/acpica/exprep.c obj_desc->common_field.node = info->field_node; node 51 drivers/acpi/acpica/exresnte.c struct acpi_namespace_node *node; node 60 drivers/acpi/acpica/exresnte.c node = *object_ptr; node 61 drivers/acpi/acpica/exresnte.c source_desc = acpi_ns_get_attached_object(node); node 62 drivers/acpi/acpica/exresnte.c entry_type = acpi_ns_get_type((acpi_handle)node); node 65 drivers/acpi/acpica/exresnte.c node, source_desc, node 73 drivers/acpi/acpica/exresnte.c node = ACPI_CAST_PTR(struct acpi_namespace_node, node->object); node 74 drivers/acpi/acpica/exresnte.c source_desc = acpi_ns_get_attached_object(node); node 75 drivers/acpi/acpica/exresnte.c entry_type = acpi_ns_get_type((acpi_handle)node); node 76 drivers/acpi/acpica/exresnte.c *object_ptr = node; node 88 drivers/acpi/acpica/exresnte.c (node->flags & (ANOBJ_METHOD_ARG | ANOBJ_METHOD_LOCAL))) { node 94 drivers/acpi/acpica/exresnte.c node->name.ascii, node)); node 174 drivers/acpi/acpica/exresnte.c node, source_desc, entry_type)); node 200 drivers/acpi/acpica/exresnte.c "Untyped entry %p, no attached object!", node)); node 235 drivers/acpi/acpica/exresnte.c node, entry_type)); node 219 drivers/acpi/acpica/exresolv.c if ((stack_desc->reference.node->type == node 221 drivers/acpi/acpica/exresolv.c || (stack_desc->reference.node->type == node 226 drivers/acpi/acpica/exresolv.c *stack_ptr = (void *)stack_desc->reference.node; node 231 drivers/acpi/acpica/exresolv.c (stack_desc->reference.node)->object; node 308 drivers/acpi/acpica/exresolv.c struct acpi_namespace_node *node = node 326 drivers/acpi/acpica/exresolv.c obj_desc = acpi_ns_get_attached_object(node); node 351 drivers/acpi/acpica/exresolv.c acpi_ut_get_node_name(node))); node 382 drivers/acpi/acpica/exresolv.c node = obj_desc->reference.object; node 385 drivers/acpi/acpica/exresolv.c node = obj_desc->reference.node; node 390 drivers/acpi/acpica/exresolv.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != node 394 drivers/acpi/acpica/exresolv.c node, node 395 drivers/acpi/acpica/exresolv.c acpi_ut_get_descriptor_name(node))); node 401 drivers/acpi/acpica/exresolv.c obj_desc = acpi_ns_get_attached_object(node); node 406 drivers/acpi/acpica/exresolv.c type = acpi_ns_get_type(node); node 474 drivers/acpi/acpica/exresolv.c &node); node 479 drivers/acpi/acpica/exresolv.c obj_desc = acpi_ns_get_attached_object(node); node 28 drivers/acpi/acpica/exstore.c struct acpi_namespace_node *node, node 362 drivers/acpi/acpica/exstore.c struct acpi_namespace_node *node, node 375 drivers/acpi/acpica/exstore.c target_type = acpi_ns_get_type(node); node 376 drivers/acpi/acpica/exstore.c target_desc = acpi_ns_get_attached_object(node); node 380 drivers/acpi/acpica/exstore.c acpi_ut_get_object_type_name(source_desc), node, node 438 drivers/acpi/acpica/exstore.c acpi_ut_get_type_name(node->type), node 439 drivers/acpi/acpica/exstore.c node->name.ascii)); node 476 drivers/acpi/acpica/exstore.c acpi_ex_store_direct_to_node(source_desc, node, node 501 drivers/acpi/acpica/exstore.c acpi_ns_attach_object(node, new_desc, node 537 drivers/acpi/acpica/exstore.c acpi_ex_store_direct_to_node(source_desc, node, walk_state); node 561 drivers/acpi/acpica/exstore.c struct acpi_namespace_node *node, node 573 drivers/acpi/acpica/exstore.c source_desc, acpi_ut_get_type_name(node->type), node 574 drivers/acpi/acpica/exstore.c node)); node 586 drivers/acpi/acpica/exstore.c status = acpi_ns_attach_object(node, new_desc, new_desc->common.type); node 191 drivers/acpi/acpica/nsaccess.c obj_desc->mutex.node = new_node; node 322 drivers/acpi/acpica/nsaccess.c if ((!scope_info) || (!scope_info->scope.node)) { node 329 drivers/acpi/acpica/nsaccess.c prefix_node = scope_info->scope.node; node 28 drivers/acpi/acpica/nsalloc.c struct acpi_namespace_node *node; node 35 drivers/acpi/acpica/nsalloc.c node = acpi_os_acquire_object(acpi_gbl_namespace_cache); node 36 drivers/acpi/acpica/nsalloc.c if (!node) { node 50 drivers/acpi/acpica/nsalloc.c node->name.integer = name; node 51 drivers/acpi/acpica/nsalloc.c ACPI_SET_DESCRIPTOR_TYPE(node, ACPI_DESC_TYPE_NAMED); node 52 drivers/acpi/acpica/nsalloc.c return_PTR(node); node 70 drivers/acpi/acpica/nsalloc.c void acpi_ns_delete_node(struct acpi_namespace_node *node) node 77 drivers/acpi/acpica/nsalloc.c if (!node) { node 83 drivers/acpi/acpica/nsalloc.c acpi_ns_detach_object(node); node 91 drivers/acpi/acpica/nsalloc.c obj_desc = node->object; node 97 drivers/acpi/acpica/nsalloc.c obj_desc->data.handler(node, obj_desc->data.pointer); node 107 drivers/acpi/acpica/nsalloc.c if (node == acpi_gbl_root_node) { node 113 drivers/acpi/acpica/nsalloc.c (void)acpi_os_release_object(acpi_gbl_namespace_cache, node); node 117 drivers/acpi/acpica/nsalloc.c node, acpi_gbl_current_node_count)); node 132 drivers/acpi/acpica/nsalloc.c void acpi_ns_remove_node(struct acpi_namespace_node *node) node 138 drivers/acpi/acpica/nsalloc.c ACPI_FUNCTION_TRACE_PTR(ns_remove_node, node); node 140 drivers/acpi/acpica/nsalloc.c parent_node = node->parent; node 147 drivers/acpi/acpica/nsalloc.c while (next_node != node) { node 156 drivers/acpi/acpica/nsalloc.c prev_node->peer = node->peer; node 162 drivers/acpi/acpica/nsalloc.c parent_node->child = node->peer; node 167 drivers/acpi/acpica/nsalloc.c acpi_ns_delete_node(node); node 192 drivers/acpi/acpica/nsalloc.c struct acpi_namespace_node *node, /* New Child */ node 222 drivers/acpi/acpica/nsalloc.c node->peer = NULL; node 223 drivers/acpi/acpica/nsalloc.c node->parent = parent_node; node 227 drivers/acpi/acpica/nsalloc.c parent_node->child = node; node 235 drivers/acpi/acpica/nsalloc.c child_node->peer = node; node 240 drivers/acpi/acpica/nsalloc.c node->owner_id = owner_id; node 241 drivers/acpi/acpica/nsalloc.c node->type = (u8) type; node 245 drivers/acpi/acpica/nsalloc.c acpi_ut_get_node_name(node), node 246 drivers/acpi/acpica/nsalloc.c acpi_ut_get_type_name(node->type), node, owner_id, node 45 drivers/acpi/acpica/nsarguments.c if (!info->predefined || (info->node->flags & ANOBJ_EVALUATED)) { node 70 drivers/acpi/acpica/nsarguments.c info->node->flags |= ANOBJ_EVALUATED; node 93 drivers/acpi/acpica/nsarguments.c struct acpi_namespace_node *node, node 99 drivers/acpi/acpica/nsarguments.c if (!predefined || (node->flags & ANOBJ_EVALUATED)) { node 112 drivers/acpi/acpica/nsarguments.c if (node->type != ACPI_TYPE_METHOD) { node 120 drivers/acpi/acpica/nsarguments.c acpi_ut_get_type_name(node-> node 132 drivers/acpi/acpica/nsarguments.c acpi_ut_get_type_name(node-> node 149 drivers/acpi/acpica/nsarguments.c aml_param_count = node->object->method.param_count; node 186 drivers/acpi/acpica/nsarguments.c struct acpi_namespace_node *node, node 193 drivers/acpi/acpica/nsarguments.c if (node->flags & ANOBJ_EVALUATED) { node 202 drivers/acpi/acpica/nsarguments.c if (node->type != ACPI_TYPE_METHOD) { node 209 drivers/acpi/acpica/nsarguments.c (node->type))); node 227 drivers/acpi/acpica/nsarguments.c aml_param_count = node->object->method.param_count; node 438 drivers/acpi/acpica/nsconvert.c struct acpi_namespace_node *node; node 454 drivers/acpi/acpica/nsconvert.c scope_info.scope.node = node 459 drivers/acpi/acpica/nsconvert.c NULL, &node); node 476 drivers/acpi/acpica/nsconvert.c new_object->reference.node = node; node 477 drivers/acpi/acpica/nsconvert.c new_object->reference.object = node->object; node 484 drivers/acpi/acpica/nsconvert.c acpi_ut_add_reference(node->object); node 338 drivers/acpi/acpica/nsdump.c obj_desc->buffer_field.buffer_obj->buffer.node) { node 344 drivers/acpi/acpica/nsdump.c node)); node 354 drivers/acpi/acpica/nsdump.c node)); node 363 drivers/acpi/acpica/nsdump.c node), node 368 drivers/acpi/acpica/nsdump.c node)); node 377 drivers/acpi/acpica/nsdump.c common_field.node), node 382 drivers/acpi/acpica/nsdump.c node)); node 682 drivers/acpi/acpica/nsdump.c struct acpi_namespace_node *node; node 689 drivers/acpi/acpica/nsdump.c node = acpi_ns_validate_handle(obj_handle); node 690 drivers/acpi/acpica/nsdump.c if (!node) { node 697 drivers/acpi/acpica/nsdump.c pathname = acpi_ns_get_normalized_pathname(node, TRUE); node 705 drivers/acpi/acpica/nsdump.c level, level, " ", acpi_ut_get_type_name(node->type), node 52 drivers/acpi/acpica/nseval.c if (!info->node) { node 63 drivers/acpi/acpica/nseval.c ACPI_NS_NO_UPSEARCH, &info->node); node 73 drivers/acpi/acpica/nseval.c if (acpi_ns_get_type(info->node) == ACPI_TYPE_LOCAL_METHOD_ALIAS) { node 74 drivers/acpi/acpica/nseval.c info->node = node 76 drivers/acpi/acpica/nseval.c info->node->object); node 82 drivers/acpi/acpica/nseval.c info->node_flags = info->node->flags; node 83 drivers/acpi/acpica/nseval.c info->obj_desc = acpi_ns_get_attached_object(info->node); node 86 drivers/acpi/acpica/nseval.c info->relative_pathname, info->node, node 87 drivers/acpi/acpica/nseval.c acpi_ns_get_attached_object(info->node))); node 92 drivers/acpi/acpica/nseval.c acpi_ut_match_predefined_method(info->node->name.ascii); node 96 drivers/acpi/acpica/nseval.c info->full_pathname = acpi_ns_get_normalized_pathname(info->node, TRUE); node 106 drivers/acpi/acpica/nseval.c acpi_ut_get_type_name(info->node->type))); node 133 drivers/acpi/acpica/nseval.c acpi_ns_check_acpi_compliance(info->full_pathname, info->node, node 140 drivers/acpi/acpica/nseval.c acpi_ns_check_argument_count(info->full_pathname, info->node, node 154 drivers/acpi/acpica/nseval.c switch (acpi_ns_get_type(info->node)) { node 170 drivers/acpi/acpica/nseval.c acpi_ut_get_type_name(info->node->type))); node 235 drivers/acpi/acpica/nseval.c ACPI_CAST_PTR(union acpi_operand_object, info->node); node 261 drivers/acpi/acpica/nseval.c (void)acpi_ns_check_return_value(info->node, info, info->param_count, node 265 drivers/acpi/acpica/nsinit.c struct acpi_namespace_node *node = node 268 drivers/acpi/acpica/nsinit.c obj_desc = acpi_ns_get_attached_object(node); node 323 drivers/acpi/acpica/nsinit.c struct acpi_namespace_node *node = node 334 drivers/acpi/acpica/nsinit.c obj_desc = acpi_ns_get_attached_object(node); node 413 drivers/acpi/acpica/nsinit.c acpi_ut_get_node_name(node), node 421 drivers/acpi/acpica/nsinit.c acpi_ut_get_node_name(node), node 455 drivers/acpi/acpica/nsinit.c struct acpi_namespace_node *node; node 460 drivers/acpi/acpica/nsinit.c node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle); node 461 drivers/acpi/acpica/nsinit.c if ((node->type == ACPI_TYPE_DEVICE) || node 462 drivers/acpi/acpica/nsinit.c (node->type == ACPI_TYPE_PROCESSOR) || node 463 drivers/acpi/acpica/nsinit.c (node->type == ACPI_TYPE_THERMAL)) { node 470 drivers/acpi/acpica/nsinit.c if (!ACPI_COMPARE_NAMESEG(node->name.ascii, METHOD_NAME__INI)) { node 478 drivers/acpi/acpica/nsinit.c parent_node = node->parent; node 41 drivers/acpi/acpica/nsload.c acpi_ns_load_table(u32 table_index, struct acpi_namespace_node *node) node 71 drivers/acpi/acpica/nsload.c status = acpi_ns_parse_table(table_index, node); node 106 drivers/acpi/acpica/nsload.c status = acpi_ds_initialize_objects(table_index, node); node 30 drivers/acpi/acpica/nsnames.c char *acpi_ns_get_external_pathname(struct acpi_namespace_node *node) node 34 drivers/acpi/acpica/nsnames.c ACPI_FUNCTION_TRACE_PTR(ns_get_external_pathname, node); node 36 drivers/acpi/acpica/nsnames.c name_buffer = acpi_ns_get_normalized_pathname(node, FALSE); node 52 drivers/acpi/acpica/nsnames.c acpi_size acpi_ns_get_pathname_length(struct acpi_namespace_node *node) node 58 drivers/acpi/acpica/nsnames.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { node 61 drivers/acpi/acpica/nsnames.c node, ACPI_GET_DESCRIPTOR_TYPE(node))); node 65 drivers/acpi/acpica/nsnames.c size = acpi_ns_build_normalized_path(node, NULL, 0, FALSE); node 87 drivers/acpi/acpica/nsnames.c struct acpi_namespace_node *node; node 92 drivers/acpi/acpica/nsnames.c node = acpi_ns_validate_handle(target_handle); node 93 drivers/acpi/acpica/nsnames.c if (!node) { node 106 drivers/acpi/acpica/nsnames.c node_name = acpi_ut_get_node_name(node); node 135 drivers/acpi/acpica/nsnames.c struct acpi_namespace_node *node; node 140 drivers/acpi/acpica/nsnames.c node = acpi_ns_validate_handle(target_handle); node 141 drivers/acpi/acpica/nsnames.c if (!node) { node 148 drivers/acpi/acpica/nsnames.c acpi_ns_build_normalized_path(node, NULL, 0, no_trailing); node 162 drivers/acpi/acpica/nsnames.c (void)acpi_ns_build_normalized_path(node, buffer->pointer, node 193 drivers/acpi/acpica/nsnames.c acpi_ns_build_normalized_path(struct acpi_namespace_node *node, node 202 drivers/acpi/acpica/nsnames.c ACPI_FUNCTION_TRACE_PTR(ns_build_normalized_path, node); node 221 drivers/acpi/acpica/nsnames.c if (!node) { node 225 drivers/acpi/acpica/nsnames.c next_node = node; node 227 drivers/acpi/acpica/nsnames.c if (next_node != node) { node 289 drivers/acpi/acpica/nsnames.c char *acpi_ns_get_normalized_pathname(struct acpi_namespace_node *node, node 295 drivers/acpi/acpica/nsnames.c ACPI_FUNCTION_TRACE_PTR(ns_get_normalized_pathname, node); node 299 drivers/acpi/acpica/nsnames.c size = acpi_ns_build_normalized_path(node, NULL, 0, no_trailing); node 314 drivers/acpi/acpica/nsnames.c (void)acpi_ns_build_normalized_path(node, name_buffer, size, node 349 drivers/acpi/acpica/nsnames.c if (prefix_scope && prefix_scope->scope.node) { node 351 drivers/acpi/acpica/nsnames.c acpi_ns_get_normalized_pathname(prefix_scope->scope.node, node 37 drivers/acpi/acpica/nsobject.c acpi_ns_attach_object(struct acpi_namespace_node *node, node 49 drivers/acpi/acpica/nsobject.c if (!node) { node 66 drivers/acpi/acpica/nsobject.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { node 71 drivers/acpi/acpica/nsobject.c node, acpi_ut_get_descriptor_name(node))); node 77 drivers/acpi/acpica/nsobject.c if (node->object == object) { node 80 drivers/acpi/acpica/nsobject.c object, node)); node 119 drivers/acpi/acpica/nsobject.c obj_desc, node, acpi_ut_get_node_name(node))); node 123 drivers/acpi/acpica/nsobject.c if (node->object) { node 124 drivers/acpi/acpica/nsobject.c acpi_ns_detach_object(node); node 145 drivers/acpi/acpica/nsobject.c last_obj_desc->common.next_object = node->object; node 148 drivers/acpi/acpica/nsobject.c node->type = (u8) object_type; node 149 drivers/acpi/acpica/nsobject.c node->object = obj_desc; node 168 drivers/acpi/acpica/nsobject.c void acpi_ns_detach_object(struct acpi_namespace_node *node) node 174 drivers/acpi/acpica/nsobject.c obj_desc = node->object; node 180 drivers/acpi/acpica/nsobject.c if (node->flags & ANOBJ_ALLOCATED_BUFFER) { node 190 drivers/acpi/acpica/nsobject.c acpi_ut_remove_address_range(obj_desc->region.space_id, node); node 195 drivers/acpi/acpica/nsobject.c node->object = NULL; node 200 drivers/acpi/acpica/nsobject.c node->object = obj_desc->common.next_object; node 204 drivers/acpi/acpica/nsobject.c if (node->object && node 205 drivers/acpi/acpica/nsobject.c (node->object->common.type != ACPI_TYPE_LOCAL_DATA)) { node 206 drivers/acpi/acpica/nsobject.c node->object = node->object->common.next_object; node 222 drivers/acpi/acpica/nsobject.c node->type = ACPI_TYPE_ANY; node 225 drivers/acpi/acpica/nsobject.c node, acpi_ut_get_node_name(node), obj_desc)); node 248 drivers/acpi/acpica/nsobject.c *node) node 250 drivers/acpi/acpica/nsobject.c ACPI_FUNCTION_TRACE_PTR(ns_get_attached_object, node); node 252 drivers/acpi/acpica/nsobject.c if (!node) { node 257 drivers/acpi/acpica/nsobject.c if (!node->object || node 258 drivers/acpi/acpica/nsobject.c ((ACPI_GET_DESCRIPTOR_TYPE(node->object) != ACPI_DESC_TYPE_OPERAND) node 259 drivers/acpi/acpica/nsobject.c && (ACPI_GET_DESCRIPTOR_TYPE(node->object) != node 261 drivers/acpi/acpica/nsobject.c || ((node->object)->common.type == ACPI_TYPE_LOCAL_DATA)) { node 265 drivers/acpi/acpica/nsobject.c return_PTR(node->object); node 313 drivers/acpi/acpica/nsobject.c acpi_ns_attach_data(struct acpi_namespace_node *node, node 323 drivers/acpi/acpica/nsobject.c obj_desc = node->object; node 349 drivers/acpi/acpica/nsobject.c node->object = data_desc; node 370 drivers/acpi/acpica/nsobject.c acpi_ns_detach_data(struct acpi_namespace_node *node, node 377 drivers/acpi/acpica/nsobject.c obj_desc = node->object; node 385 drivers/acpi/acpica/nsobject.c node->object = obj_desc->common.next_object; node 415 drivers/acpi/acpica/nsobject.c acpi_ns_get_attached_data(struct acpi_namespace_node *node, node 420 drivers/acpi/acpica/nsobject.c obj_desc = node->object; node 101 drivers/acpi/acpica/nsparse.c info->node = start_node; node 103 drivers/acpi/acpica/nsparse.c info->node_flags = info->node->flags; node 104 drivers/acpi/acpica/nsparse.c info->full_pathname = acpi_ns_get_normalized_pathname(info->node, TRUE); node 65 drivers/acpi/acpica/nspredef.c acpi_ns_check_return_value(struct acpi_namespace_node *node, node 154 drivers/acpi/acpica/nspredef.c status = acpi_ns_complex_repairs(info, node, status, return_object_ptr); node 163 drivers/acpi/acpica/nspredef.c node->flags |= ANOBJ_EVALUATED; node 204 drivers/acpi/acpica/nspredef.c return_object->node.name.ascii, node 205 drivers/acpi/acpica/nspredef.c acpi_ut_get_type_name(return_object->node. node 55 drivers/acpi/acpica/nsrepair.c *node, node 136 drivers/acpi/acpica/nsrepair.c predefined = acpi_ns_match_simple_repair(info->node, node 146 drivers/acpi/acpica/nsrepair.c status = predefined->object_converter(info->node, return_object, node 307 drivers/acpi/acpica/nsrepair.c *node, node 319 drivers/acpi/acpica/nsrepair.c if (ACPI_COMPARE_NAMESEG(node->name.ascii, this_name->name)) { node 37 drivers/acpi/acpica/nsrepair2.c *node); node 151 drivers/acpi/acpica/nsrepair2.c struct acpi_namespace_node *node, node 160 drivers/acpi/acpica/nsrepair2.c predefined = acpi_ns_match_complex_repair(node); node 183 drivers/acpi/acpica/nsrepair2.c *node) node 191 drivers/acpi/acpica/nsrepair2.c if (ACPI_COMPARE_NAMESEG(node->name.ascii, this_name->name)) { node 715 drivers/acpi/acpica/nsrepair2.c struct acpi_namespace_node *node; node 725 drivers/acpi/acpica/nsrepair2.c status = acpi_ns_get_node(info->node, "^_PSS", node 726 drivers/acpi/acpica/nsrepair2.c ACPI_NS_NO_UPSEARCH, &node); node 22 drivers/acpi/acpica/nssearch.c struct acpi_namespace_node *node, node 64 drivers/acpi/acpica/nssearch.c struct acpi_namespace_node *node; node 89 drivers/acpi/acpica/nssearch.c node = parent_node->child; node 90 drivers/acpi/acpica/nssearch.c while (node) { node 94 drivers/acpi/acpica/nssearch.c if (node->name.integer == target_name) { node 98 drivers/acpi/acpica/nssearch.c if (acpi_ns_get_type(node) == node 100 drivers/acpi/acpica/nssearch.c node = node 102 drivers/acpi/acpica/nssearch.c node->object); node 110 drivers/acpi/acpica/nssearch.c acpi_ut_get_type_name(node->type), node 111 drivers/acpi/acpica/nssearch.c node, node 115 drivers/acpi/acpica/nssearch.c *return_node = node; node 121 drivers/acpi/acpica/nssearch.c node = node->peer; node 164 drivers/acpi/acpica/nssearch.c struct acpi_namespace_node *node, node 173 drivers/acpi/acpica/nssearch.c parent_node = node->parent; node 253 drivers/acpi/acpica/nssearch.c struct acpi_namespace_node *node, node 265 drivers/acpi/acpica/nssearch.c if (!node || !target_name || !return_node) { node 268 drivers/acpi/acpica/nssearch.c node, target_name, return_node)); node 286 drivers/acpi/acpica/nssearch.c status = acpi_ns_search_one_scope(target_name, node, type, return_node); node 354 drivers/acpi/acpica/nssearch.c acpi_ns_search_parent_tree(target_name, node, type, node 366 drivers/acpi/acpica/nssearch.c ACPI_CAST_PTR(char, &target_name), node)); node 393 drivers/acpi/acpica/nssearch.c acpi_ns_install_node(walk_state, node, new_node, type); node 37 drivers/acpi/acpica/nsutils.c acpi_ns_print_node_pathname(struct acpi_namespace_node *node, node 43 drivers/acpi/acpica/nsutils.c if (!node) { node 52 drivers/acpi/acpica/nsutils.c status = acpi_ns_handle_to_pathname(node, &buffer, TRUE); node 75 drivers/acpi/acpica/nsutils.c acpi_object_type acpi_ns_get_type(struct acpi_namespace_node * node) node 79 drivers/acpi/acpica/nsutils.c if (!node) { node 84 drivers/acpi/acpica/nsutils.c return_UINT8(node->type); node 673 drivers/acpi/acpica/nsutils.c scope_info.scope.node = prefix_node; node 277 drivers/acpi/acpica/nsxfeval.c switch (acpi_ns_get_type(info->node)) { node 475 drivers/acpi/acpica/nsxfeval.c struct acpi_namespace_node *node; node 498 drivers/acpi/acpica/nsxfeval.c node = info->return_object->reference.object; node 499 drivers/acpi/acpica/nsxfeval.c if (node) { node 500 drivers/acpi/acpica/nsxfeval.c obj_desc = node->object; node 641 drivers/acpi/acpica/nsxfeval.c struct acpi_namespace_node *node; node 654 drivers/acpi/acpica/nsxfeval.c node = acpi_ns_validate_handle(obj_handle); node 660 drivers/acpi/acpica/nsxfeval.c if (!node) { node 679 drivers/acpi/acpica/nsxfeval.c status = acpi_ut_execute_HID(node, &hid); node 694 drivers/acpi/acpica/nsxfeval.c status = acpi_ut_execute_CID(node, &cid); node 723 drivers/acpi/acpica/nsxfeval.c status = acpi_ut_execute_STA(node, &flags); node 833 drivers/acpi/acpica/nsxfeval.c struct acpi_namespace_node *node; node 849 drivers/acpi/acpica/nsxfeval.c node = acpi_ns_validate_handle(obj_handle); node 850 drivers/acpi/acpica/nsxfeval.c if (!node) { node 855 drivers/acpi/acpica/nsxfeval.c status = acpi_ns_attach_data(node, handler, data); node 879 drivers/acpi/acpica/nsxfeval.c struct acpi_namespace_node *node; node 895 drivers/acpi/acpica/nsxfeval.c node = acpi_ns_validate_handle(obj_handle); node 896 drivers/acpi/acpica/nsxfeval.c if (!node) { node 901 drivers/acpi/acpica/nsxfeval.c status = acpi_ns_detach_data(node, handler); node 929 drivers/acpi/acpica/nsxfeval.c struct acpi_namespace_node *node; node 945 drivers/acpi/acpica/nsxfeval.c node = acpi_ns_validate_handle(obj_handle); node 946 drivers/acpi/acpica/nsxfeval.c if (!node) { node 951 drivers/acpi/acpica/nsxfeval.c status = acpi_ns_get_attached_data(node, handler, data); node 50 drivers/acpi/acpica/nsxfname.c struct acpi_namespace_node *node = NULL; node 98 drivers/acpi/acpica/nsxfname.c acpi_ns_get_node(prefix_node, pathname, ACPI_NS_NO_UPSEARCH, &node); node 100 drivers/acpi/acpica/nsxfname.c *ret_handle = ACPI_CAST_PTR(acpi_handle, node); node 229 drivers/acpi/acpica/nsxfname.c struct acpi_namespace_node *node; node 255 drivers/acpi/acpica/nsxfname.c node = acpi_ns_validate_handle(handle); node 256 drivers/acpi/acpica/nsxfname.c if (!node) { node 264 drivers/acpi/acpica/nsxfname.c type = node->type; node 265 drivers/acpi/acpica/nsxfname.c name = node->name.integer; node 267 drivers/acpi/acpica/nsxfname.c if (node->type == ACPI_TYPE_METHOD) { node 268 drivers/acpi/acpica/nsxfname.c param_count = node->object->method.param_count; node 288 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_HID(node, &hid); node 296 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_UID(node, &uid); node 304 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_CID(node, &cid_list); node 317 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_CLS(node, &cls); node 348 drivers/acpi/acpica/nsxfname.c status = acpi_ut_evaluate_numeric_object(METHOD_NAME__ADR, node, node 356 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_power_methods(node, node 366 drivers/acpi/acpica/nsxfname.c status = acpi_ut_execute_power_methods(node, node 482 drivers/acpi/acpica/nsxfname.c struct acpi_namespace_node *node; node 548 drivers/acpi/acpica/nsxfname.c NULL, &node); node 559 drivers/acpi/acpica/nsxfname.c if (node->type != ACPI_TYPE_METHOD) { node 588 drivers/acpi/acpica/nsxfname.c status = acpi_ns_attach_object(node, method_obj, ACPI_TYPE_METHOD); node 594 drivers/acpi/acpica/nsxfname.c node->flags |= ANOBJ_ALLOCATED_BUFFER; node 32 drivers/acpi/acpica/nsxfobj.c struct acpi_namespace_node *node; node 55 drivers/acpi/acpica/nsxfobj.c node = acpi_ns_validate_handle(handle); node 56 drivers/acpi/acpica/nsxfobj.c if (!node) { node 61 drivers/acpi/acpica/nsxfobj.c *ret_type = node->type; node 84 drivers/acpi/acpica/nsxfobj.c struct acpi_namespace_node *node; node 105 drivers/acpi/acpica/nsxfobj.c node = acpi_ns_validate_handle(handle); node 106 drivers/acpi/acpica/nsxfobj.c if (!node) { node 113 drivers/acpi/acpica/nsxfobj.c parent_node = node->parent; node 153 drivers/acpi/acpica/nsxfobj.c struct acpi_namespace_node *node; node 192 drivers/acpi/acpica/nsxfobj.c node = acpi_ns_get_next_node_typed(type, parent_node, child_node); node 193 drivers/acpi/acpica/nsxfobj.c if (!node) { node 199 drivers/acpi/acpica/nsxfobj.c *ret_handle = ACPI_CAST_PTR(acpi_handle, node); node 205 drivers/acpi/acpica/psargs.c struct acpi_namespace_node *node; node 231 drivers/acpi/acpica/psargs.c NULL, &node); node 238 drivers/acpi/acpica/psargs.c possible_method_call && (node->type == ACPI_TYPE_METHOD)) { node 256 drivers/acpi/acpica/psargs.c method_desc = acpi_ns_get_attached_object(node); node 259 drivers/acpi/acpica/psargs.c node->name.ascii, node, method_desc, path)); node 273 drivers/acpi/acpica/psargs.c name_op->common.node = node; node 279 drivers/acpi/acpica/psargs.c node)); node 285 drivers/acpi/acpica/psargs.c node, method_desc->method.param_count)); node 562 drivers/acpi/acpica/psobject.c node); node 563 drivers/acpi/acpica/psobject.c acpi_ns_remove_node((*op)->common.node); node 564 drivers/acpi/acpica/psobject.c (*op)->common.node = NULL; node 232 drivers/acpi/acpica/psparse.c replacement_op->common.node = op->common.node; node 256 drivers/acpi/acpica/psparse.c replacement_op->common.node = node 257 drivers/acpi/acpica/psparse.c op->common.node; node 375 drivers/acpi/acpica/psparse.c (op->common.value.arg)->common.node; node 98 drivers/acpi/acpica/psxface.c if (!info || !info->node) { node 105 drivers/acpi/acpica/psxface.c acpi_ds_begin_method_execution(info->node, info->obj_desc, NULL); node 120 drivers/acpi/acpica/psxface.c info->node->name.ascii, info->node, info->obj_desc)); node 141 drivers/acpi/acpica/psxface.c status = acpi_ds_init_aml_walk(walk_state, op, info->node, node 265 drivers/acpi/acpica/psxface.c status = acpi_ds_init_aml_walk(walk_state, op, info->node, node 282 drivers/acpi/acpica/psxface.c if (info->node && info->node != acpi_gbl_root_node) { node 284 drivers/acpi/acpica/psxface.c acpi_ds_scope_stack_push(info->node, ACPI_TYPE_METHOD, node 798 drivers/acpi/acpica/rscalc.c temp_size_needed += acpi_ns_get_pathname_length((*sub_object_list)->reference.node); node 190 drivers/acpi/acpica/rscreate.c struct acpi_namespace_node *node; node 305 drivers/acpi/acpica/rscreate.c node = obj_desc->reference.node; node 314 drivers/acpi/acpica/rscreate.c status = acpi_ns_handle_to_pathname((acpi_handle)node, &path_buffer, FALSE); node 433 drivers/acpi/acpica/rsutils.c acpi_rs_get_prt_method_data(struct acpi_namespace_node *node, node 446 drivers/acpi/acpica/rsutils.c acpi_ut_evaluate_object(node, METHOD_NAME__PRT, ACPI_BTYPE_PACKAGE, node 483 drivers/acpi/acpica/rsutils.c acpi_rs_get_crs_method_data(struct acpi_namespace_node *node, node 496 drivers/acpi/acpica/rsutils.c acpi_ut_evaluate_object(node, METHOD_NAME__CRS, ACPI_BTYPE_BUFFER, node 534 drivers/acpi/acpica/rsutils.c acpi_rs_get_prs_method_data(struct acpi_namespace_node *node, node 547 drivers/acpi/acpica/rsutils.c acpi_ut_evaluate_object(node, METHOD_NAME__PRS, ACPI_BTYPE_BUFFER, node 585 drivers/acpi/acpica/rsutils.c acpi_rs_get_aei_method_data(struct acpi_namespace_node *node, node 598 drivers/acpi/acpica/rsutils.c acpi_ut_evaluate_object(node, METHOD_NAME__AEI, ACPI_BTYPE_BUFFER, node 691 drivers/acpi/acpica/rsutils.c acpi_rs_set_srs_method_data(struct acpi_namespace_node *node, node 708 drivers/acpi/acpica/rsutils.c info->prefix_node = node; node 62 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 73 drivers/acpi/acpica/rsxface.c node = acpi_ns_validate_handle(device_handle); node 74 drivers/acpi/acpica/rsxface.c if (!node) { node 78 drivers/acpi/acpica/rsxface.c if (node->type != ACPI_TYPE_DEVICE) { node 94 drivers/acpi/acpica/rsxface.c *return_node = node; node 126 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 132 drivers/acpi/acpica/rsxface.c status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node); node 137 drivers/acpi/acpica/rsxface.c status = acpi_rs_get_prt_method_data(node, ret_buffer); node 171 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 177 drivers/acpi/acpica/rsxface.c status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node); node 182 drivers/acpi/acpica/rsxface.c status = acpi_rs_get_crs_method_data(node, ret_buffer); node 213 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 219 drivers/acpi/acpica/rsxface.c status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node); node 224 drivers/acpi/acpica/rsxface.c status = acpi_rs_get_prs_method_data(node, ret_buffer); node 252 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 264 drivers/acpi/acpica/rsxface.c status = acpi_rs_validate_parameters(device_handle, in_buffer, &node); node 269 drivers/acpi/acpica/rsxface.c status = acpi_rs_set_srs_method_data(node, in_buffer); node 298 drivers/acpi/acpica/rsxface.c struct acpi_namespace_node *node; node 304 drivers/acpi/acpica/rsxface.c status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node); node 309 drivers/acpi/acpica/rsxface.c status = acpi_rs_get_aei_method_data(node, ret_buffer); node 329 drivers/acpi/acpica/tbxfload.c struct acpi_namespace_node *node = node 347 drivers/acpi/acpica/tbxfload.c owner_id = node->owner_id; node 152 drivers/acpi/acpica/utcopy.c internal_object->reference.node; node 154 drivers/acpi/acpica/utcopy.c acpi_ns_get_type(internal_object->reference.node); node 240 drivers/acpi/acpica/utdecode.c struct acpi_namespace_node *node = (struct acpi_namespace_node *)object; node 256 drivers/acpi/acpica/utdecode.c if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { node 264 drivers/acpi/acpica/utdecode.c acpi_ut_repair_name(node->name.ascii); node 268 drivers/acpi/acpica/utdecode.c return (node->name.ascii); node 177 drivers/acpi/acpica/utdelete.c if (object->method.node) { node 178 drivers/acpi/acpica/utdelete.c object->method.node = NULL; node 191 drivers/acpi/acpica/utdelete.c if (!(object->region.node->flags & ANOBJ_TEMPORARY)) { node 193 drivers/acpi/acpica/utdelete.c object->region.node); node 306 drivers/acpi/acpica/uterror.c struct acpi_namespace_node *node = prefix_node; node 313 drivers/acpi/acpica/uterror.c ACPI_NS_NO_UPSEARCH, &node); node 319 drivers/acpi/acpica/uterror.c acpi_ns_print_node_pathname(node, message); node 497 drivers/acpi/acpica/utobject.c reference.node); node 673 drivers/acpi/acpica/uttrack.c node)); node 80 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 87 drivers/acpi/arm64/iort.c if (curr->iort_node == node) { node 102 drivers/acpi/arm64/iort.c static inline void iort_delete_fwnode(struct acpi_iort_node *node) node 108 drivers/acpi/arm64/iort.c if (curr->iort_node == node) { node 143 drivers/acpi/arm64/iort.c (struct acpi_iort_node *node, void *context); node 258 drivers/acpi/arm64/iort.c static acpi_status iort_match_node_callback(struct acpi_iort_node *node, node 264 drivers/acpi/arm64/iort.c if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT) { node 278 drivers/acpi/arm64/iort.c ncomp = (struct acpi_iort_named_component *)node->node_data; node 282 drivers/acpi/arm64/iort.c } else if (node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX) { node 287 drivers/acpi/arm64/iort.c pci_rc = (struct acpi_iort_root_complex *)node->node_data; node 325 drivers/acpi/arm64/iort.c static struct acpi_iort_node *iort_node_get_id(struct acpi_iort_node *node, node 331 drivers/acpi/arm64/iort.c if (!node->mapping_offset || !node->mapping_count || node 332 drivers/acpi/arm64/iort.c index >= node->mapping_count) node 335 drivers/acpi/arm64/iort.c map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node, node 336 drivers/acpi/arm64/iort.c node->mapping_offset + index * sizeof(*map)); node 341 drivers/acpi/arm64/iort.c node, node->type); node 349 drivers/acpi/arm64/iort.c if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT || node 350 drivers/acpi/arm64/iort.c node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX || node 351 drivers/acpi/arm64/iort.c node->type == ACPI_IORT_NODE_SMMU_V3 || node 352 drivers/acpi/arm64/iort.c node->type == ACPI_IORT_NODE_PMCG) { node 361 drivers/acpi/arm64/iort.c static int iort_get_id_mapping_index(struct acpi_iort_node *node) node 365 drivers/acpi/arm64/iort.c switch (node->type) { node 371 drivers/acpi/arm64/iort.c if (node->revision < 1) node 374 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 383 drivers/acpi/arm64/iort.c if (smmu->id_mapping_index >= node->mapping_count) { node 385 drivers/acpi/arm64/iort.c node, node->type); node 397 drivers/acpi/arm64/iort.c static struct acpi_iort_node *iort_node_map_id(struct acpi_iort_node *node, node 404 drivers/acpi/arm64/iort.c while (node) { node 408 drivers/acpi/arm64/iort.c if (IORT_TYPE_MASK(node->type) & type_mask) { node 411 drivers/acpi/arm64/iort.c return node; node 414 drivers/acpi/arm64/iort.c if (!node->mapping_offset || !node->mapping_count) node 417 drivers/acpi/arm64/iort.c map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node, node 418 drivers/acpi/arm64/iort.c node->mapping_offset); node 423 drivers/acpi/arm64/iort.c node, node->type); node 432 drivers/acpi/arm64/iort.c index = iort_get_id_mapping_index(node); node 435 drivers/acpi/arm64/iort.c for (i = 0; i < node->mapping_count; i++, map++) { node 440 drivers/acpi/arm64/iort.c if (!iort_id_map(map, node->type, id, &id)) node 444 drivers/acpi/arm64/iort.c if (i == node->mapping_count) node 447 drivers/acpi/arm64/iort.c node = ACPI_ADD_PTR(struct acpi_iort_node, iort_table, node 460 drivers/acpi/arm64/iort.c struct acpi_iort_node *node, u32 *id_out, u8 type_mask, node 467 drivers/acpi/arm64/iort.c parent = iort_node_get_id(node, &id, index); node 491 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 498 drivers/acpi/arm64/iort.c node = iort_get_iort_node(dev->fwnode); node 499 drivers/acpi/arm64/iort.c if (node) node 500 drivers/acpi/arm64/iort.c return node; node 528 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 531 drivers/acpi/arm64/iort.c node = iort_find_dev_node(dev); node 532 drivers/acpi/arm64/iort.c if (!node) node 535 drivers/acpi/arm64/iort.c iort_node_map_id(node, req_id, &dev_id, IORT_MSI_TYPE); node 549 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 551 drivers/acpi/arm64/iort.c node = iort_find_dev_node(dev); node 552 drivers/acpi/arm64/iort.c if (!node) node 555 drivers/acpi/arm64/iort.c index = iort_get_id_mapping_index(node); node 558 drivers/acpi/arm64/iort.c if (iort_node_get_id(node, dev_id, index)) node 561 drivers/acpi/arm64/iort.c for (i = 0; i < node->mapping_count; i++) { node 562 drivers/acpi/arm64/iort.c if (iort_node_map_platform_id(node, dev_id, node 602 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 604 drivers/acpi/arm64/iort.c node = iort_find_dev_node(dev); node 605 drivers/acpi/arm64/iort.c if (!node) node 608 drivers/acpi/arm64/iort.c node = iort_node_map_id(node, req_id, NULL, IORT_MSI_TYPE); node 609 drivers/acpi/arm64/iort.c if (!node) node 613 drivers/acpi/arm64/iort.c its = (struct acpi_iort_its_group *)node->node_data; node 647 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 656 drivers/acpi/arm64/iort.c index = iort_get_id_mapping_index(node); node 660 drivers/acpi/arm64/iort.c map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node, node 661 drivers/acpi/arm64/iort.c node->mapping_offset + index * sizeof(*map)); node 667 drivers/acpi/arm64/iort.c node, node->type); node 698 drivers/acpi/arm64/iort.c struct acpi_iort_node *node, *msi_parent = NULL; node 704 drivers/acpi/arm64/iort.c node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT, node 706 drivers/acpi/arm64/iort.c if (!node) node 710 drivers/acpi/arm64/iort.c for (i = 0; i < node->mapping_count; i++) { node 711 drivers/acpi/arm64/iort.c msi_parent = iort_node_map_platform_id(node, NULL, node 874 drivers/acpi/arm64/iort.c static bool iort_pci_rc_supports_ats(struct acpi_iort_node *node) node 878 drivers/acpi/arm64/iort.c pci_rc = (struct acpi_iort_root_complex *)node->node_data; node 882 drivers/acpi/arm64/iort.c static int iort_iommu_xlate(struct device *dev, struct acpi_iort_node *node, node 888 drivers/acpi/arm64/iort.c if (!node) node 891 drivers/acpi/arm64/iort.c iort_fwnode = iort_get_fwnode(node); node 905 drivers/acpi/arm64/iort.c return iort_iommu_driver_enabled(node->type) ? node 913 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 922 drivers/acpi/arm64/iort.c parent = iort_node_map_id(info->node, alias, &streamid, node 937 drivers/acpi/arm64/iort.c struct acpi_iort_node *node, *parent; node 954 drivers/acpi/arm64/iort.c node = iort_scan_node(ACPI_IORT_NODE_PCI_ROOT_COMPLEX, node 956 drivers/acpi/arm64/iort.c if (!node) node 959 drivers/acpi/arm64/iort.c info.node = node; node 963 drivers/acpi/arm64/iort.c if (!err && iort_pci_rc_supports_ats(node)) node 968 drivers/acpi/arm64/iort.c node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT, node 970 drivers/acpi/arm64/iort.c if (!node) node 974 drivers/acpi/arm64/iort.c parent = iort_node_map_platform_id(node, &streamid, node 1016 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 1019 drivers/acpi/arm64/iort.c node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT, node 1021 drivers/acpi/arm64/iort.c if (!node) node 1024 drivers/acpi/arm64/iort.c ncomp = (struct acpi_iort_named_component *)node->node_data; node 1034 drivers/acpi/arm64/iort.c struct acpi_iort_node *node; node 1038 drivers/acpi/arm64/iort.c node = iort_scan_node(ACPI_IORT_NODE_PCI_ROOT_COMPLEX, node 1040 drivers/acpi/arm64/iort.c if (!node || node->revision < 1) node 1043 drivers/acpi/arm64/iort.c rc = (struct acpi_iort_root_complex *)node->node_data; node 1131 drivers/acpi/arm64/iort.c static int __init arm_smmu_v3_count_resources(struct acpi_iort_node *node) node 1138 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 1186 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1192 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 1230 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1236 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 1253 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1257 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 1275 drivers/acpi/arm64/iort.c static int __init arm_smmu_count_resources(struct acpi_iort_node *node) node 1280 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu *)node->node_data; node 1294 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1301 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu *)node->node_data; node 1308 drivers/acpi/arm64/iort.c glb_irq = ACPI_ADD_PTR(u64, node, smmu->global_interrupt_offset); node 1317 drivers/acpi/arm64/iort.c ctx_irq = ACPI_ADD_PTR(u64, node, smmu->context_interrupt_offset); node 1328 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1334 drivers/acpi/arm64/iort.c smmu = (struct acpi_iort_smmu *)node->node_data; node 1346 drivers/acpi/arm64/iort.c static int __init arm_smmu_v3_pmcg_count_resources(struct acpi_iort_node *node) node 1351 drivers/acpi/arm64/iort.c pmcg = (struct acpi_iort_pmcg *)node->node_data; node 1361 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1366 drivers/acpi/arm64/iort.c pmcg = (struct acpi_iort_pmcg *)node->node_data; node 1403 drivers/acpi/arm64/iort.c int (*dev_init)(struct acpi_iort_node *node); node 1405 drivers/acpi/arm64/iort.c struct acpi_iort_node *node); node 1406 drivers/acpi/arm64/iort.c int (*dev_count_resources)(struct acpi_iort_node *node); node 1408 drivers/acpi/arm64/iort.c struct acpi_iort_node *node); node 1410 drivers/acpi/arm64/iort.c struct acpi_iort_node *node); node 1437 drivers/acpi/arm64/iort.c struct acpi_iort_node *node) node 1439 drivers/acpi/arm64/iort.c switch (node->type) { node 1457 drivers/acpi/arm64/iort.c static int __init iort_add_platform_device(struct acpi_iort_node *node, node 1470 drivers/acpi/arm64/iort.c ret = ops->dev_set_proximity(&pdev->dev, node); node 1475 drivers/acpi/arm64/iort.c count = ops->dev_count_resources(node); node 1483 drivers/acpi/arm64/iort.c ops->dev_init_resources(r, node); node 1504 drivers/acpi/arm64/iort.c ret = platform_device_add_data(pdev, &node, sizeof(node)); node 1509 drivers/acpi/arm64/iort.c fwnode = iort_get_fwnode(node); node 1519 drivers/acpi/arm64/iort.c ops->dev_dma_configure(&pdev->dev, node); node 1521 drivers/acpi/arm64/iort.c iort_set_device_domain(&pdev->dev, node); node 544 drivers/acpi/bus.c const struct acpi_device_physical_node *node; node 546 drivers/acpi/bus.c node = list_first_entry(&adev->physical_node_list, node 547 drivers/acpi/bus.c struct acpi_device_physical_node, node); node 549 drivers/acpi/bus.c phys_dev = node->dev; node 35 drivers/acpi/container.c list_for_each_entry(child, &adev->children, node) node 146 drivers/acpi/ec.c struct list_head node; node 1061 drivers/acpi/ec.c list_for_each_entry(handler, &ec->list, node) { node 1100 drivers/acpi/ec.c list_add(&handler->node, &ec->list); node 1113 drivers/acpi/ec.c list_for_each_entry_safe(handler, tmp, &ec->list, node) { node 1115 drivers/acpi/ec.c list_del_init(&handler->node); node 1116 drivers/acpi/ec.c list_add(&handler->node, &free_list); node 1120 drivers/acpi/ec.c list_for_each_entry_safe(handler, tmp, &free_list, node) node 49 drivers/acpi/evged.c struct list_head node; node 137 drivers/acpi/evged.c list_add_tail(&event->node, &geddev->event_list); node 168 drivers/acpi/evged.c list_for_each_entry_safe(event, next, &geddev->event_list, node) { node 170 drivers/acpi/evged.c list_del(&event->node); node 123 drivers/acpi/glue.c list_for_each_entry(adev, &parent->children, node) { node 209 drivers/acpi/glue.c list_for_each_entry(pn, &acpi_dev->physical_node_list, node) { node 224 drivers/acpi/glue.c physnode_list = &pn->node; node 231 drivers/acpi/glue.c list_add(&physical_node->node, physnode_list); node 275 drivers/acpi/glue.c list_for_each_entry(entry, &acpi_dev->physical_node_list, node) node 279 drivers/acpi/glue.c list_del(&entry->node); node 44 drivers/acpi/hmat/hmat.c struct list_head node; node 49 drivers/acpi/hmat/hmat.c struct list_head node; node 59 drivers/acpi/hmat/hmat.c struct list_head node; node 64 drivers/acpi/hmat/hmat.c struct list_head node; node 72 drivers/acpi/hmat/hmat.c list_for_each_entry(initiator, &initiators, node) node 82 drivers/acpi/hmat/hmat.c list_for_each_entry(target, &targets, node) node 104 drivers/acpi/hmat/hmat.c list_add_tail(&initiator->node, &initiators); node 121 drivers/acpi/hmat/hmat.c list_add_tail(&target->node, &targets); node 236 drivers/acpi/hmat/hmat.c list_add_tail(&loc->node, &localities); node 381 drivers/acpi/hmat/hmat.c list_add_tail(&tcache->node, &target->caches); node 532 drivers/acpi/hmat/hmat.c ia = list_entry(a, struct memory_initiator, node); node 533 drivers/acpi/hmat/hmat.c ib = list_entry(b, struct memory_initiator, node); node 579 drivers/acpi/hmat/hmat.c list_for_each_entry(initiator, &initiators, node) { node 606 drivers/acpi/hmat/hmat.c list_for_each_entry(tcache, &target->caches, node) node 644 drivers/acpi/hmat/hmat.c list_for_each_entry(target, &targets, node) node 679 drivers/acpi/hmat/hmat.c list_for_each_entry_safe(target, tnext, &targets, node) { node 680 drivers/acpi/hmat/hmat.c list_for_each_entry_safe(tcache, cnext, &target->caches, node) { node 681 drivers/acpi/hmat/hmat.c list_del(&tcache->node); node 684 drivers/acpi/hmat/hmat.c list_del(&target->node); node 688 drivers/acpi/hmat/hmat.c list_for_each_entry_safe(initiator, inext, &initiators, node) { node 689 drivers/acpi/hmat/hmat.c list_del(&initiator->node); node 693 drivers/acpi/hmat/hmat.c list_for_each_entry_safe(loc, lnext, &localities, node) { node 694 drivers/acpi/hmat/hmat.c list_del(&loc->node); node 102 drivers/acpi/internal.h struct list_head node; node 39 drivers/acpi/numa.c int node_to_pxm(int node) node 41 drivers/acpi/numa.c if (node < 0) node 43 drivers/acpi/numa.c return node_to_pxm_map[node]; node 46 drivers/acpi/numa.c static void __acpi_map_pxm_to_node(int pxm, int node) node 48 drivers/acpi/numa.c if (pxm_to_node_map[pxm] == NUMA_NO_NODE || node < pxm_to_node_map[pxm]) node 49 drivers/acpi/numa.c pxm_to_node_map[pxm] = node; node 50 drivers/acpi/numa.c if (node_to_pxm_map[node] == PXM_INVAL || pxm < node_to_pxm_map[node]) node 51 drivers/acpi/numa.c node_to_pxm_map[node] = pxm; node 56 drivers/acpi/numa.c int node; node 61 drivers/acpi/numa.c node = pxm_to_node_map[pxm]; node 63 drivers/acpi/numa.c if (node == NUMA_NO_NODE) { node 66 drivers/acpi/numa.c node = first_unset_node(nodes_found_map); node 67 drivers/acpi/numa.c __acpi_map_pxm_to_node(pxm, node); node 68 drivers/acpi/numa.c node_set(node, nodes_found_map); node 71 drivers/acpi/numa.c return node; node 92 drivers/acpi/numa.c int node, min_node; node 94 drivers/acpi/numa.c node = acpi_map_pxm_to_node(pxm); node 96 drivers/acpi/numa.c if (node == NUMA_NO_NODE) node 97 drivers/acpi/numa.c node = 0; node 99 drivers/acpi/numa.c min_node = node; node 100 drivers/acpi/numa.c if (!node_online(node)) { node 104 drivers/acpi/numa.c dist = node_distance(node, n); node 251 drivers/acpi/numa.c int node, pxm; node 272 drivers/acpi/numa.c node = acpi_map_pxm_to_node(pxm); node 273 drivers/acpi/numa.c if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) { node 278 drivers/acpi/numa.c if (numa_add_memblk(node, start, end) < 0) { node 280 drivers/acpi/numa.c node, (unsigned long long) start, node 285 drivers/acpi/numa.c node_set(node, numa_nodes_parsed); node 288 drivers/acpi/numa.c node, pxm, node 22 drivers/acpi/nvs.c struct list_head node; node 45 drivers/acpi/nvs.c list_add_tail(®ion->node, &nvs_region_list); node 56 drivers/acpi/nvs.c list_for_each_entry(region, &nvs_region_list, node) { node 79 drivers/acpi/nvs.c struct list_head node; node 107 drivers/acpi/nvs.c list_add_tail(&entry->node, &nvs_list); node 118 drivers/acpi/nvs.c list_for_each_entry_safe(entry, next, &nvs_list, node) { node 119 drivers/acpi/nvs.c list_del(&entry->node); node 132 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) node 156 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) { node 175 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) node 207 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) node 81 drivers/acpi/pci_mcfg.c #define THUNDER_PEM_RES(addr, node) \ node 82 drivers/acpi/pci_mcfg.c DEFINE_RES_MEM((addr) + ((u64) (node) << 44), 0x39 * SZ_16M) node 84 drivers/acpi/pci_mcfg.c #define THUNDER_PEM_QUIRK(rev, node) \ node 85 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 4 + (10 * (node)), MCFG_BUS_ANY, \ node 86 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x88001f000000UL, node) }, \ node 87 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 5 + (10 * (node)), MCFG_BUS_ANY, \ node 88 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x884057000000UL, node) }, \ node 89 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 6 + (10 * (node)), MCFG_BUS_ANY, \ node 90 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x88808f000000UL, node) }, \ node 91 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 7 + (10 * (node)), MCFG_BUS_ANY, \ node 92 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x89001f000000UL, node) }, \ node 93 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 8 + (10 * (node)), MCFG_BUS_ANY, \ node 94 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x894057000000UL, node) }, \ node 95 drivers/acpi/pci_mcfg.c { "CAVIUM", "THUNDERX", rev, 9 + (10 * (node)), MCFG_BUS_ANY, \ node 96 drivers/acpi/pci_mcfg.c &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x89808f000000UL, node) } node 250 drivers/acpi/pci_root.c struct list_head node; node 274 drivers/acpi/pci_root.c struct acpi_handle_node *node, *tmp; node 283 drivers/acpi/pci_root.c node = kzalloc(sizeof(struct acpi_handle_node), GFP_KERNEL); node 284 drivers/acpi/pci_root.c if (!node) node 287 drivers/acpi/pci_root.c INIT_LIST_HEAD(&node->node); node 288 drivers/acpi/pci_root.c node->handle = phandle; node 289 drivers/acpi/pci_root.c list_add(&node->node, &device_list); node 307 drivers/acpi/pci_root.c list_for_each_entry(node, &device_list, node) { node 308 drivers/acpi/pci_root.c acpi_handle hnd = node->handle; node 334 drivers/acpi/pci_root.c list_for_each_entry_safe(node, tmp, &device_list, node) node 335 drivers/acpi/pci_root.c kfree(node); node 880 drivers/acpi/pci_root.c int node = acpi_get_node(device->handle); node 934 drivers/acpi/pci_root.c if (node != NUMA_NO_NODE) node 935 drivers/acpi/pci_root.c dev_printk(KERN_DEBUG, &bus->dev, "on NUMA node %d\n", node); node 47 drivers/acpi/power.c struct list_head node; node 63 drivers/acpi/power.c struct list_head node; node 107 drivers/acpi/power.c list_for_each_entry(e, list, node) node 109 drivers/acpi/power.c list_add_tail(&entry->node, &e->node); node 113 drivers/acpi/power.c list_add_tail(&entry->node, list); node 121 drivers/acpi/power.c list_for_each_entry_safe(entry, e, list, node) { node 122 drivers/acpi/power.c list_del(&entry->node); node 219 drivers/acpi/power.c list_for_each_entry(entry, list, node) { node 249 drivers/acpi/power.c list_for_each_entry(dep, &resource->dependents, node) { node 262 drivers/acpi/power.c list_add_tail(&dep->node, &resource->dependents); node 277 drivers/acpi/power.c list_for_each_entry(dep, &resource->dependents, node) { node 279 drivers/acpi/power.c list_del(&dep->node); node 315 drivers/acpi/power.c list_for_each_entry(entry, resources, node) { node 324 drivers/acpi/power.c list_for_each_entry(entry, resources, node) node 349 drivers/acpi/power.c list_for_each_entry_reverse(entry, resources, node) node 374 drivers/acpi/power.c list_for_each_entry(dep, &resource->dependents, node) { node 461 drivers/acpi/power.c list_for_each_entry_reverse(entry, list, node) { node 469 drivers/acpi/power.c list_for_each_entry_continue(entry, list, node) node 480 drivers/acpi/power.c list_for_each_entry(entry, list, node) { node 488 drivers/acpi/power.c list_for_each_entry_continue_reverse(entry, list, node) node 531 drivers/acpi/power.c list_for_each_entry_reverse(entry, resources, node) { node 555 drivers/acpi/power.c list_for_each_entry(entry, resources, node) { node 602 drivers/acpi/power.c list_for_each_entry(entry, list, node) { node 712 drivers/acpi/power.c list_for_each_entry(entry, &dev->wakeup.resources, node) { node 775 drivers/acpi/power.c list_for_each_entry(entry, &dev->wakeup.resources, node) { node 60 drivers/acpi/pptt.c struct acpi_pptt_processor *node, node 65 drivers/acpi/pptt.c if (resource >= node->number_of_priv_resources) node 68 drivers/acpi/pptt.c ref = ACPI_ADD_PTR(u32, node, sizeof(struct acpi_pptt_processor)); node 204 drivers/acpi/pptt.c struct acpi_pptt_processor *node) node 213 drivers/acpi/pptt.c return (node->flags & ACPI_PPTT_ACPI_LEAF_NODE); node 216 drivers/acpi/pptt.c node_entry = ACPI_PTR_DIFF(node, table_hdr); node 322 drivers/acpi/pptt.c struct acpi_pptt_processor **node) node 337 drivers/acpi/pptt.c *node = cpu_node; node 25 drivers/acpi/proc.c struct list_head *node, *next; node 30 drivers/acpi/proc.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { node 32 drivers/acpi/proc.c container_of(node, struct acpi_device, wakeup_list); node 52 drivers/acpi/proc.c node) { node 57 drivers/acpi/proc.c if (&entry->node != node 85 drivers/acpi/proc.c &adev->physical_node_list, node) node 99 drivers/acpi/proc.c struct list_head *node, *next; node 112 drivers/acpi/proc.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { node 114 drivers/acpi/proc.c container_of(node, struct acpi_device, wakeup_list); node 1034 drivers/acpi/property.c next = adev->node.next; node 1039 drivers/acpi/property.c child_adev = list_entry(next, struct acpi_device, node); node 1042 drivers/acpi/property.c node); node 732 drivers/acpi/resource.c list_for_each_entry(rentry, &resource_list, node) { node 54 drivers/acpi/scan.c struct list_head node; node 128 drivers/acpi/scan.c list_for_each_entry(pn, &adev->physical_node_list, node) node 159 drivers/acpi/scan.c list_for_each_entry(pn, &device->physical_node_list, node) { node 197 drivers/acpi/scan.c list_for_each_entry(pn, &device->physical_node_list, node) node 362 drivers/acpi/scan.c list_for_each_entry(child, &adev->children, node) { node 481 drivers/acpi/scan.c list_del(&device->node); node 483 drivers/acpi/scan.c list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node) node 489 drivers/acpi/scan.c list_del(&acpi_device_bus_id->node); node 649 drivers/acpi/scan.c INIT_LIST_HEAD(&device->node); node 667 drivers/acpi/scan.c list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node) { node 680 drivers/acpi/scan.c list_add_tail(&acpi_device_bus_id->node, &acpi_bus_id_list); node 685 drivers/acpi/scan.c list_add_tail(&device->node, &device->parent->children); node 711 drivers/acpi/scan.c list_del(&device->node); node 1425 drivers/acpi/scan.c list_for_each_entry(rentry, &list, node) { node 1857 drivers/acpi/scan.c list_add_tail(&dep->node , &acpi_dep_list); node 2015 drivers/acpi/scan.c list_for_each_entry(child, &device->children, node) node 2028 drivers/acpi/scan.c list_for_each_entry_safe(dep, tmp, &acpi_dep_list, node) { node 2037 drivers/acpi/scan.c list_del(&dep->node); node 2086 drivers/acpi/scan.c list_for_each_entry_reverse(child, &adev->children, node) node 333 drivers/acpi/sysfs.c struct list_head node; node 373 drivers/acpi/sysfs.c list_for_each_entry(attr, &acpi_table_attr_list, node) { node 419 drivers/acpi/sysfs.c list_add_tail(&table_attr->node, &acpi_table_attr_list); node 539 drivers/acpi/sysfs.c list_add_tail(&table_attr->node, &acpi_table_attr_list); node 716 drivers/acpi/utils.c list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node) node 42 drivers/acpi/wakeup.c struct list_head *node, *next; node 44 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { node 46 drivers/acpi/wakeup.c container_of(node, struct acpi_device, wakeup_list); node 69 drivers/acpi/wakeup.c struct list_head *node, *next; node 71 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { node 73 drivers/acpi/wakeup.c container_of(node, struct acpi_device, wakeup_list); node 91 drivers/acpi/wakeup.c struct list_head *node, *next; node 94 drivers/acpi/wakeup.c list_for_each_safe(node, next, &acpi_wakeup_device_list) { node 95 drivers/acpi/wakeup.c struct acpi_device *dev = container_of(node, node 497 drivers/amba/bus.c struct list_head node; node 514 drivers/amba/bus.c list_for_each_entry_safe(ddev, tmp, &deferred_devices, node) { node 520 drivers/amba/bus.c list_del_init(&ddev->node); node 560 drivers/amba/bus.c list_add_tail(&ddev->node, &deferred_devices); node 400 drivers/android/binder.c struct binder_node *node; node 691 drivers/android/binder.c #define binder_node_lock(node) _binder_node_lock(node, __LINE__) node 693 drivers/android/binder.c _binder_node_lock(struct binder_node *node, int line) node 694 drivers/android/binder.c __acquires(&node->lock) node 698 drivers/android/binder.c spin_lock(&node->lock); node 707 drivers/android/binder.c #define binder_node_unlock(node) _binder_node_unlock(node, __LINE__) node 709 drivers/android/binder.c _binder_node_unlock(struct binder_node *node, int line) node 710 drivers/android/binder.c __releases(&node->lock) node 714 drivers/android/binder.c spin_unlock(&node->lock); node 724 drivers/android/binder.c #define binder_node_inner_lock(node) _binder_node_inner_lock(node, __LINE__) node 726 drivers/android/binder.c _binder_node_inner_lock(struct binder_node *node, int line) node 727 drivers/android/binder.c __acquires(&node->lock) __acquires(&node->proc->inner_lock) node 731 drivers/android/binder.c spin_lock(&node->lock); node 732 drivers/android/binder.c if (node->proc) node 733 drivers/android/binder.c binder_inner_proc_lock(node->proc); node 736 drivers/android/binder.c __acquire(&node->proc->inner_lock); node 745 drivers/android/binder.c #define binder_node_inner_unlock(node) _binder_node_inner_unlock(node, __LINE__) node 747 drivers/android/binder.c _binder_node_inner_unlock(struct binder_node *node, int line) node 748 drivers/android/binder.c __releases(&node->lock) __releases(&node->proc->inner_lock) node 750 drivers/android/binder.c struct binder_proc *proc = node->proc; node 758 drivers/android/binder.c __release(&node->proc->inner_lock); node 759 drivers/android/binder.c spin_unlock(&node->lock); node 917 drivers/android/binder.c static void binder_inc_node_tmpref_ilocked(struct binder_node *node); node 1068 drivers/android/binder.c struct binder_node *node; node 1073 drivers/android/binder.c node = rb_entry(n, struct binder_node, rb_node); node 1075 drivers/android/binder.c if (ptr < node->ptr) node 1077 drivers/android/binder.c else if (ptr > node->ptr) node 1085 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 1086 drivers/android/binder.c return node; node 1095 drivers/android/binder.c struct binder_node *node; node 1098 drivers/android/binder.c node = binder_get_node_ilocked(proc, ptr); node 1100 drivers/android/binder.c return node; node 1110 drivers/android/binder.c struct binder_node *node; node 1120 drivers/android/binder.c node = rb_entry(parent, struct binder_node, rb_node); node 1122 drivers/android/binder.c if (ptr < node->ptr) node 1124 drivers/android/binder.c else if (ptr > node->ptr) node 1132 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 1133 drivers/android/binder.c return node; node 1136 drivers/android/binder.c node = new_node; node 1138 drivers/android/binder.c node->tmp_refs++; node 1139 drivers/android/binder.c rb_link_node(&node->rb_node, parent, p); node 1140 drivers/android/binder.c rb_insert_color(&node->rb_node, &proc->nodes); node 1141 drivers/android/binder.c node->debug_id = atomic_inc_return(&binder_last_id); node 1142 drivers/android/binder.c node->proc = proc; node 1143 drivers/android/binder.c node->ptr = ptr; node 1144 drivers/android/binder.c node->cookie = cookie; node 1145 drivers/android/binder.c node->work.type = BINDER_WORK_NODE; node 1146 drivers/android/binder.c node->min_priority = flags & FLAT_BINDER_FLAG_PRIORITY_MASK; node 1147 drivers/android/binder.c node->accept_fds = !!(flags & FLAT_BINDER_FLAG_ACCEPTS_FDS); node 1148 drivers/android/binder.c node->txn_security_ctx = !!(flags & FLAT_BINDER_FLAG_TXN_SECURITY_CTX); node 1149 drivers/android/binder.c spin_lock_init(&node->lock); node 1150 drivers/android/binder.c INIT_LIST_HEAD(&node->work.entry); node 1151 drivers/android/binder.c INIT_LIST_HEAD(&node->async_todo); node 1154 drivers/android/binder.c proc->pid, current->pid, node->debug_id, node 1155 drivers/android/binder.c (u64)node->ptr, (u64)node->cookie); node 1157 drivers/android/binder.c return node; node 1163 drivers/android/binder.c struct binder_node *node; node 1164 drivers/android/binder.c struct binder_node *new_node = kzalloc(sizeof(*node), GFP_KERNEL); node 1169 drivers/android/binder.c node = binder_init_node_ilocked(proc, new_node, fp); node 1171 drivers/android/binder.c if (node != new_node) node 1177 drivers/android/binder.c return node; node 1180 drivers/android/binder.c static void binder_free_node(struct binder_node *node) node 1182 drivers/android/binder.c kfree(node); node 1186 drivers/android/binder.c static int binder_inc_node_nilocked(struct binder_node *node, int strong, node 1190 drivers/android/binder.c struct binder_proc *proc = node->proc; node 1192 drivers/android/binder.c assert_spin_locked(&node->lock); node 1198 drivers/android/binder.c node->internal_strong_refs == 0 && node 1199 drivers/android/binder.c !(node->proc && node 1200 drivers/android/binder.c node == node->proc->context->binder_context_mgr_node && node 1201 drivers/android/binder.c node->has_strong_ref)) { node 1203 drivers/android/binder.c node->debug_id); node 1206 drivers/android/binder.c node->internal_strong_refs++; node 1208 drivers/android/binder.c node->local_strong_refs++; node 1209 drivers/android/binder.c if (!node->has_strong_ref && target_list) { node 1212 drivers/android/binder.c binder_dequeue_work_ilocked(&node->work); node 1215 drivers/android/binder.c &node->work); node 1219 drivers/android/binder.c node->local_weak_refs++; node 1220 drivers/android/binder.c if (!node->has_weak_ref && list_empty(&node->work.entry)) { node 1223 drivers/android/binder.c node->debug_id); node 1229 drivers/android/binder.c binder_enqueue_work_ilocked(&node->work, target_list); node 1235 drivers/android/binder.c static int binder_inc_node(struct binder_node *node, int strong, int internal, node 1240 drivers/android/binder.c binder_node_inner_lock(node); node 1241 drivers/android/binder.c ret = binder_inc_node_nilocked(node, strong, internal, target_list); node 1242 drivers/android/binder.c binder_node_inner_unlock(node); node 1247 drivers/android/binder.c static bool binder_dec_node_nilocked(struct binder_node *node, node 1250 drivers/android/binder.c struct binder_proc *proc = node->proc; node 1252 drivers/android/binder.c assert_spin_locked(&node->lock); node 1257 drivers/android/binder.c node->internal_strong_refs--; node 1259 drivers/android/binder.c node->local_strong_refs--; node 1260 drivers/android/binder.c if (node->local_strong_refs || node->internal_strong_refs) node 1264 drivers/android/binder.c node->local_weak_refs--; node 1265 drivers/android/binder.c if (node->local_weak_refs || node->tmp_refs || node 1266 drivers/android/binder.c !hlist_empty(&node->refs)) node 1270 drivers/android/binder.c if (proc && (node->has_strong_ref || node->has_weak_ref)) { node 1271 drivers/android/binder.c if (list_empty(&node->work.entry)) { node 1272 drivers/android/binder.c binder_enqueue_work_ilocked(&node->work, &proc->todo); node 1276 drivers/android/binder.c if (hlist_empty(&node->refs) && !node->local_strong_refs && node 1277 drivers/android/binder.c !node->local_weak_refs && !node->tmp_refs) { node 1279 drivers/android/binder.c binder_dequeue_work_ilocked(&node->work); node 1280 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); node 1283 drivers/android/binder.c node->debug_id); node 1285 drivers/android/binder.c BUG_ON(!list_empty(&node->work.entry)); node 1291 drivers/android/binder.c if (node->tmp_refs) { node 1295 drivers/android/binder.c hlist_del(&node->dead_node); node 1299 drivers/android/binder.c node->debug_id); node 1307 drivers/android/binder.c static void binder_dec_node(struct binder_node *node, int strong, int internal) node 1311 drivers/android/binder.c binder_node_inner_lock(node); node 1312 drivers/android/binder.c free_node = binder_dec_node_nilocked(node, strong, internal); node 1313 drivers/android/binder.c binder_node_inner_unlock(node); node 1315 drivers/android/binder.c binder_free_node(node); node 1318 drivers/android/binder.c static void binder_inc_node_tmpref_ilocked(struct binder_node *node) node 1325 drivers/android/binder.c node->tmp_refs++; node 1341 drivers/android/binder.c static void binder_inc_node_tmpref(struct binder_node *node) node 1343 drivers/android/binder.c binder_node_lock(node); node 1344 drivers/android/binder.c if (node->proc) node 1345 drivers/android/binder.c binder_inner_proc_lock(node->proc); node 1348 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 1349 drivers/android/binder.c if (node->proc) node 1350 drivers/android/binder.c binder_inner_proc_unlock(node->proc); node 1353 drivers/android/binder.c binder_node_unlock(node); node 1362 drivers/android/binder.c static void binder_dec_node_tmpref(struct binder_node *node) node 1366 drivers/android/binder.c binder_node_inner_lock(node); node 1367 drivers/android/binder.c if (!node->proc) node 1371 drivers/android/binder.c node->tmp_refs--; node 1372 drivers/android/binder.c BUG_ON(node->tmp_refs < 0); node 1373 drivers/android/binder.c if (!node->proc) node 1383 drivers/android/binder.c free_node = binder_dec_node_nilocked(node, 0, 1); node 1384 drivers/android/binder.c binder_node_inner_unlock(node); node 1386 drivers/android/binder.c binder_free_node(node); node 1389 drivers/android/binder.c static void binder_put_node(struct binder_node *node) node 1391 drivers/android/binder.c binder_dec_node_tmpref(node); node 1437 drivers/android/binder.c struct binder_node *node, node 1450 drivers/android/binder.c if (node < ref->node) node 1452 drivers/android/binder.c else if (node > ref->node) node 1463 drivers/android/binder.c new_ref->node = node; node 1467 drivers/android/binder.c new_ref->data.desc = (node == context->binder_context_mgr_node) ? 0 : 1; node 1490 drivers/android/binder.c binder_node_lock(node); node 1491 drivers/android/binder.c hlist_add_head(&new_ref->node_entry, &node->refs); node 1496 drivers/android/binder.c node->debug_id); node 1497 drivers/android/binder.c binder_node_unlock(node); node 1508 drivers/android/binder.c ref->node->debug_id); node 1513 drivers/android/binder.c binder_node_inner_lock(ref->node); node 1515 drivers/android/binder.c binder_dec_node_nilocked(ref->node, 1, 1); node 1518 drivers/android/binder.c delete_node = binder_dec_node_nilocked(ref->node, 0, 1); node 1519 drivers/android/binder.c binder_node_inner_unlock(ref->node); node 1529 drivers/android/binder.c ref->node = NULL; node 1560 drivers/android/binder.c ret = binder_inc_node(ref->node, 1, 1, target_list); node 1567 drivers/android/binder.c ret = binder_inc_node(ref->node, 0, 1, target_list); node 1597 drivers/android/binder.c binder_dec_node(ref->node, strong, 1); node 1631 drivers/android/binder.c struct binder_node *node; node 1638 drivers/android/binder.c node = ref->node; node 1643 drivers/android/binder.c binder_inc_node_tmpref(node); node 1648 drivers/android/binder.c return node; node 1664 drivers/android/binder.c if (ref->node) node 1665 drivers/android/binder.c binder_free_node(ref->node); node 1747 drivers/android/binder.c struct binder_node *node, node 1757 drivers/android/binder.c ref = binder_get_ref_for_node_olocked(proc, node, NULL); node 1764 drivers/android/binder.c ref = binder_get_ref_for_node_olocked(proc, node, new_ref); node 2304 drivers/android/binder.c struct binder_node *node; node 2307 drivers/android/binder.c node = binder_get_node(proc, fp->binder); node 2308 drivers/android/binder.c if (node == NULL) { node 2315 drivers/android/binder.c node->debug_id, (u64)node->ptr); node 2316 drivers/android/binder.c binder_dec_node(node, hdr->type == BINDER_TYPE_BINDER, node 2318 drivers/android/binder.c binder_put_node(node); node 2440 drivers/android/binder.c struct binder_node *node; node 2446 drivers/android/binder.c node = binder_get_node(proc, fp->binder); node 2447 drivers/android/binder.c if (!node) { node 2448 drivers/android/binder.c node = binder_new_node(proc, fp); node 2449 drivers/android/binder.c if (!node) node 2452 drivers/android/binder.c if (fp->cookie != node->cookie) { node 2455 drivers/android/binder.c node->debug_id, (u64)fp->cookie, node 2456 drivers/android/binder.c (u64)node->cookie); node 2465 drivers/android/binder.c ret = binder_inc_ref_for_node(target_proc, node, node 2479 drivers/android/binder.c trace_binder_transaction_node_to_ref(t, node, &rdata); node 2482 drivers/android/binder.c node->debug_id, (u64)node->ptr, node 2485 drivers/android/binder.c binder_put_node(node); node 2495 drivers/android/binder.c struct binder_node *node; node 2499 drivers/android/binder.c node = binder_get_node_from_ref(proc, fp->handle, node 2501 drivers/android/binder.c if (!node) { node 2511 drivers/android/binder.c binder_node_lock(node); node 2512 drivers/android/binder.c if (node->proc == target_proc) { node 2517 drivers/android/binder.c fp->binder = node->ptr; node 2518 drivers/android/binder.c fp->cookie = node->cookie; node 2519 drivers/android/binder.c if (node->proc) node 2520 drivers/android/binder.c binder_inner_proc_lock(node->proc); node 2522 drivers/android/binder.c __acquire(&node->proc->inner_lock); node 2523 drivers/android/binder.c binder_inc_node_nilocked(node, node 2526 drivers/android/binder.c if (node->proc) node 2527 drivers/android/binder.c binder_inner_proc_unlock(node->proc); node 2529 drivers/android/binder.c __release(&node->proc->inner_lock); node 2530 drivers/android/binder.c trace_binder_transaction_ref_to_node(t, node, &src_rdata); node 2533 drivers/android/binder.c src_rdata.debug_id, src_rdata.desc, node->debug_id, node 2534 drivers/android/binder.c (u64)node->ptr); node 2535 drivers/android/binder.c binder_node_unlock(node); node 2539 drivers/android/binder.c binder_node_unlock(node); node 2540 drivers/android/binder.c ret = binder_inc_ref_for_node(target_proc, node, node 2549 drivers/android/binder.c trace_binder_transaction_ref_to_ref(t, node, &src_rdata, node 2555 drivers/android/binder.c node->debug_id); node 2558 drivers/android/binder.c binder_put_node(node); node 2757 drivers/android/binder.c struct binder_node *node = t->buffer->target_node; node 2761 drivers/android/binder.c BUG_ON(!node); node 2762 drivers/android/binder.c binder_node_lock(node); node 2765 drivers/android/binder.c if (node->has_async_transaction) { node 2768 drivers/android/binder.c node->has_async_transaction = true; node 2776 drivers/android/binder.c binder_node_unlock(node); node 2788 drivers/android/binder.c binder_enqueue_work_ilocked(&t->work, &node->async_todo); node 2794 drivers/android/binder.c binder_node_unlock(node); node 2821 drivers/android/binder.c struct binder_node *node, node 2827 drivers/android/binder.c binder_node_inner_lock(node); node 2828 drivers/android/binder.c if (node->proc) { node 2829 drivers/android/binder.c target_node = node; node 2830 drivers/android/binder.c binder_inc_node_nilocked(node, 1, 0, NULL); node 2831 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 2832 drivers/android/binder.c node->proc->tmp_ref++; node 2833 drivers/android/binder.c *procp = node->proc; node 2836 drivers/android/binder.c binder_node_inner_unlock(node); node 2951 drivers/android/binder.c ref->node, &target_proc, node 3687 drivers/android/binder.c struct binder_node *node; node 3696 drivers/android/binder.c node = binder_get_node(proc, node_ptr); node 3697 drivers/android/binder.c if (node == NULL) { node 3706 drivers/android/binder.c if (cookie != node->cookie) { node 3711 drivers/android/binder.c (u64)node_ptr, node->debug_id, node 3712 drivers/android/binder.c (u64)cookie, (u64)node->cookie); node 3713 drivers/android/binder.c binder_put_node(node); node 3716 drivers/android/binder.c binder_node_inner_lock(node); node 3718 drivers/android/binder.c if (node->pending_strong_ref == 0) { node 3721 drivers/android/binder.c node->debug_id); node 3722 drivers/android/binder.c binder_node_inner_unlock(node); node 3723 drivers/android/binder.c binder_put_node(node); node 3726 drivers/android/binder.c node->pending_strong_ref = 0; node 3728 drivers/android/binder.c if (node->pending_weak_ref == 0) { node 3731 drivers/android/binder.c node->debug_id); node 3732 drivers/android/binder.c binder_node_inner_unlock(node); node 3733 drivers/android/binder.c binder_put_node(node); node 3736 drivers/android/binder.c node->pending_weak_ref = 0; node 3738 drivers/android/binder.c free_node = binder_dec_node_nilocked(node, node 3745 drivers/android/binder.c node->debug_id, node->local_strong_refs, node 3746 drivers/android/binder.c node->local_weak_refs, node->tmp_refs); node 3747 drivers/android/binder.c binder_node_inner_unlock(node); node 3748 drivers/android/binder.c binder_put_node(node); node 3907 drivers/android/binder.c ref->data.weak, ref->node->debug_id); node 3909 drivers/android/binder.c binder_node_lock(ref->node); node 3914 drivers/android/binder.c binder_node_unlock(ref->node); node 3923 drivers/android/binder.c if (ref->node->proc == NULL) { node 3936 drivers/android/binder.c binder_node_unlock(ref->node); node 3946 drivers/android/binder.c binder_node_unlock(ref->node); node 3973 drivers/android/binder.c binder_node_unlock(ref->node); node 4293 drivers/android/binder.c struct binder_node *node = container_of(w, struct binder_node, work); node 4295 drivers/android/binder.c binder_uintptr_t node_ptr = node->ptr; node 4296 drivers/android/binder.c binder_uintptr_t node_cookie = node->cookie; node 4297 drivers/android/binder.c int node_debug_id = node->debug_id; node 4302 drivers/android/binder.c BUG_ON(proc != node->proc); node 4303 drivers/android/binder.c strong = node->internal_strong_refs || node 4304 drivers/android/binder.c node->local_strong_refs; node 4305 drivers/android/binder.c weak = !hlist_empty(&node->refs) || node 4306 drivers/android/binder.c node->local_weak_refs || node 4307 drivers/android/binder.c node->tmp_refs || strong; node 4308 drivers/android/binder.c has_strong_ref = node->has_strong_ref; node 4309 drivers/android/binder.c has_weak_ref = node->has_weak_ref; node 4312 drivers/android/binder.c node->has_weak_ref = 1; node 4313 drivers/android/binder.c node->pending_weak_ref = 1; node 4314 drivers/android/binder.c node->local_weak_refs++; node 4317 drivers/android/binder.c node->has_strong_ref = 1; node 4318 drivers/android/binder.c node->pending_strong_ref = 1; node 4319 drivers/android/binder.c node->local_strong_refs++; node 4322 drivers/android/binder.c node->has_strong_ref = 0; node 4324 drivers/android/binder.c node->has_weak_ref = 0; node 4332 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); node 4334 drivers/android/binder.c binder_node_lock(node); node 4344 drivers/android/binder.c binder_node_unlock(node); node 4345 drivers/android/binder.c binder_free_node(node); node 4941 drivers/android/binder.c struct binder_node *node; node 4961 drivers/android/binder.c node = binder_get_node_from_ref(proc, handle, true, NULL); node 4962 drivers/android/binder.c if (!node) node 4965 drivers/android/binder.c info->strong_count = node->local_strong_refs + node 4966 drivers/android/binder.c node->internal_strong_refs; node 4967 drivers/android/binder.c info->weak_count = node->local_weak_refs; node 4969 drivers/android/binder.c binder_put_node(node); node 4984 drivers/android/binder.c struct binder_node *node = rb_entry(n, struct binder_node, node 4986 drivers/android/binder.c if (node->ptr > ptr) { node 4987 drivers/android/binder.c info->ptr = node->ptr; node 4988 drivers/android/binder.c info->cookie = node->cookie; node 4989 drivers/android/binder.c info->has_strong_ref = node->has_strong_ref; node 4990 drivers/android/binder.c info->has_weak_ref = node->has_weak_ref; node 5343 drivers/android/binder.c static int binder_node_release(struct binder_node *node, int refs) node 5347 drivers/android/binder.c struct binder_proc *proc = node->proc; node 5349 drivers/android/binder.c binder_release_work(proc, &node->async_todo); node 5351 drivers/android/binder.c binder_node_lock(node); node 5353 drivers/android/binder.c binder_dequeue_work_ilocked(&node->work); node 5357 drivers/android/binder.c BUG_ON(!node->tmp_refs); node 5358 drivers/android/binder.c if (hlist_empty(&node->refs) && node->tmp_refs == 1) { node 5360 drivers/android/binder.c binder_node_unlock(node); node 5361 drivers/android/binder.c binder_free_node(node); node 5366 drivers/android/binder.c node->proc = NULL; node 5367 drivers/android/binder.c node->local_strong_refs = 0; node 5368 drivers/android/binder.c node->local_weak_refs = 0; node 5372 drivers/android/binder.c hlist_add_head(&node->dead_node, &binder_dead_nodes); node 5375 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) { node 5401 drivers/android/binder.c node->debug_id, refs, death); node 5402 drivers/android/binder.c binder_node_unlock(node); node 5403 drivers/android/binder.c binder_put_node(node); node 5457 drivers/android/binder.c struct binder_node *node; node 5459 drivers/android/binder.c node = rb_entry(n, struct binder_node, rb_node); node 5466 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 5467 drivers/android/binder.c rb_erase(&node->rb_node, &proc->nodes); node 5469 drivers/android/binder.c incoming_refs = binder_node_release(node, incoming_refs); node 5587 drivers/android/binder.c struct binder_node *node; node 5607 drivers/android/binder.c node = container_of(w, struct binder_node, work); node 5609 drivers/android/binder.c prefix, node->debug_id, node 5610 drivers/android/binder.c (u64)node->ptr, (u64)node->cookie); node 5666 drivers/android/binder.c struct binder_node *node) node 5673 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) node 5677 drivers/android/binder.c node->debug_id, (u64)node->ptr, (u64)node->cookie, node 5678 drivers/android/binder.c node->has_strong_ref, node->has_weak_ref, node 5679 drivers/android/binder.c node->local_strong_refs, node->local_weak_refs, node 5680 drivers/android/binder.c node->internal_strong_refs, count, node->tmp_refs); node 5683 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) node 5687 drivers/android/binder.c if (node->proc) { node 5688 drivers/android/binder.c list_for_each_entry(w, &node->async_todo, entry) node 5689 drivers/android/binder.c print_binder_work_ilocked(m, node->proc, " ", node 5697 drivers/android/binder.c binder_node_lock(ref->node); node 5700 drivers/android/binder.c ref->node->proc ? "" : "dead ", node 5701 drivers/android/binder.c ref->node->debug_id, ref->data.strong, node 5703 drivers/android/binder.c binder_node_unlock(ref->node); node 5725 drivers/android/binder.c struct binder_node *node = rb_entry(n, struct binder_node, node 5727 drivers/android/binder.c if (!print_all && !node->has_async_transaction) node 5735 drivers/android/binder.c binder_inc_node_tmpref_ilocked(node); node 5740 drivers/android/binder.c binder_node_inner_lock(node); node 5741 drivers/android/binder.c print_binder_node_nilocked(m, node); node 5742 drivers/android/binder.c binder_node_inner_unlock(node); node 5743 drivers/android/binder.c last_node = node; node 5937 drivers/android/binder.c struct binder_node *node; node 5945 drivers/android/binder.c hlist_for_each_entry(node, &binder_dead_nodes, dead_node) { node 5951 drivers/android/binder.c node->tmp_refs++; node 5955 drivers/android/binder.c binder_node_lock(node); node 5956 drivers/android/binder.c print_binder_node_nilocked(m, node); node 5957 drivers/android/binder.c binder_node_unlock(node); node 5958 drivers/android/binder.c last_node = node; node 140 drivers/android/binder_trace.h TP_PROTO(struct binder_transaction *t, struct binder_node *node, node 142 drivers/android/binder_trace.h TP_ARGS(t, node, rdata), node 153 drivers/android/binder_trace.h __entry->node_debug_id = node->debug_id; node 154 drivers/android/binder_trace.h __entry->node_ptr = node->ptr; node 165 drivers/android/binder_trace.h TP_PROTO(struct binder_transaction *t, struct binder_node *node, node 167 drivers/android/binder_trace.h TP_ARGS(t, node, rdata), node 180 drivers/android/binder_trace.h __entry->node_debug_id = node->debug_id; node 181 drivers/android/binder_trace.h __entry->node_ptr = node->ptr; node 190 drivers/android/binder_trace.h TP_PROTO(struct binder_transaction *t, struct binder_node *node, node 193 drivers/android/binder_trace.h TP_ARGS(t, node, src_ref, dest_ref), node 205 drivers/android/binder_trace.h __entry->node_debug_id = node->debug_id; node 34 drivers/ata/ahci_octeon.c struct device_node *node = dev->of_node; node 62 drivers/ata/ahci_octeon.c if (!node) { node 67 drivers/ata/ahci_octeon.c ret = of_platform_populate(node, NULL, NULL, dev); node 310 drivers/ata/libahci_platform.c struct device *dev, struct device_node *node) node 314 drivers/ata/libahci_platform.c hpriv->phys[port] = devm_of_phy_get(dev, node, NULL); node 323 drivers/ata/libahci_platform.c if (of_find_property(node, "phys", NULL)) { node 326 drivers/ata/libahci_platform.c node); node 342 drivers/ata/libahci_platform.c node, rc); node 231 drivers/ata/pata_macio.c struct device_node *node; node 482 drivers/ata/pata_macio.c const char* cable = of_get_property(priv->node, "cable-type", node 505 drivers/ata/pata_macio.c if (of_device_is_compatible(priv->node, "K2-UATA") || node 506 drivers/ata/pata_macio.c of_device_is_compatible(priv->node, "shasta-ata")) node 753 drivers/ata/pata_macio.c ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, priv->node, 0, 1); node 759 drivers/ata/pata_macio.c priv->node, priv->aapl_bus_id, 1); node 761 drivers/ata/pata_macio.c priv->node, priv->aapl_bus_id, 1); node 766 drivers/ata/pata_macio.c priv->node, priv->aapl_bus_id, 0); node 890 drivers/ata/pata_macio.c ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, priv->node, node 948 drivers/ata/pata_macio.c if (of_device_is_compatible(priv->node, "shasta-ata")) { node 951 drivers/ata/pata_macio.c } else if (of_device_is_compatible(priv->node, "kauai-ata")) { node 954 drivers/ata/pata_macio.c } else if (of_device_is_compatible(priv->node, "K2-UATA")) { node 957 drivers/ata/pata_macio.c } else if (of_device_is_compatible(priv->node, "keylargo-ata")) { node 958 drivers/ata/pata_macio.c if (of_node_name_eq(priv->node, "ata-4")) { node 965 drivers/ata/pata_macio.c } else if (of_device_is_compatible(priv->node, "heathrow-ata")) { node 976 drivers/ata/pata_macio.c bidp = of_get_property(priv->node, "AAPL,bus-id", NULL); node 1139 drivers/ata/pata_macio.c priv->node = of_node_get(mdev->ofdev.dev.of_node); node 1283 drivers/ata/pata_macio.c priv->node = of_node_get(np); node 842 drivers/ata/pata_octeon_cf.c struct device_node *node; node 854 drivers/ata/pata_octeon_cf.c node = pdev->dev.of_node; node 855 drivers/ata/pata_octeon_cf.c if (node == NULL) node 862 drivers/ata/pata_octeon_cf.c cf_port->is_true_ide = of_property_read_bool(node, "cavium,true-ide"); node 864 drivers/ata/pata_octeon_cf.c if (of_property_read_u32(node, "cavium,bus-width", &bus_width) == 0) node 869 drivers/ata/pata_octeon_cf.c n_addr = of_n_addr_cells(node); node 870 drivers/ata/pata_octeon_cf.c n_size = of_n_size_cells(node); node 872 drivers/ata/pata_octeon_cf.c reg_prop = of_find_property(node, "reg", ®_len); node 881 drivers/ata/pata_octeon_cf.c dma_node = of_parse_phandle(node, node 320 drivers/auxdisplay/ht16k33.c struct device_node *node = client->dev.of_node; node 339 drivers/auxdisplay/ht16k33.c if (!of_get_property(node, "linux,no-autorepeat", NULL)) node 342 drivers/auxdisplay/ht16k33.c err = of_property_read_u32(node, "debounce-delay-ms", node 398 drivers/auxdisplay/ht16k33.c struct device_node *node = client->dev.of_node; node 440 drivers/auxdisplay/ht16k33.c err = of_property_read_u32(node, "refresh-rate-hz", node 480 drivers/auxdisplay/ht16k33.c err = of_property_read_u32(node, "default-brightness-level", node 251 drivers/base/arch_topology.c static int __init get_cpu_for_node(struct device_node *node) node 256 drivers/base/arch_topology.c cpu_node = of_parse_phandle(node, "cpu", 0); node 26 drivers/base/attribute_container.c struct klist_node node; node 34 drivers/base/attribute_container.c container_of(n, struct internal_container, node); node 41 drivers/base/attribute_container.c container_of(n, struct internal_container, node); node 75 drivers/base/attribute_container.c INIT_LIST_HEAD(&cont->node); node 80 drivers/base/attribute_container.c list_add_tail(&cont->node, &attribute_container_list); node 102 drivers/base/attribute_container.c list_del(&cont->node); node 149 drivers/base/attribute_container.c list_for_each_entry(cont, &attribute_container_list, node) { node 174 drivers/base/attribute_container.c klist_add_tail(&ic->node, &cont->containers); node 214 drivers/base/attribute_container.c list_for_each_entry(cont, &attribute_container_list, node) { node 224 drivers/base/attribute_container.c klist_for_each_entry(ic, &cont->containers, node, &iter) { node 227 drivers/base/attribute_container.c klist_del(&ic->node); node 258 drivers/base/attribute_container.c list_for_each_entry(cont, &attribute_container_list, node) { node 270 drivers/base/attribute_container.c klist_for_each_entry(ic, &cont->containers, node, &iter) { node 298 drivers/base/attribute_container.c list_for_each_entry(cont, &attribute_container_list, node) { node 430 drivers/base/attribute_container.c klist_for_each_entry(ic, &cont->containers, node, &iter) { node 493 drivers/base/bus.c list_for_each_entry(sif, &bus->p->interfaces, node) node 518 drivers/base/bus.c list_for_each_entry(sif, &bus->p->interfaces, node) node 1053 drivers/base/bus.c list_add_tail(&sif->node, &subsys->p->interfaces); node 1078 drivers/base/bus.c list_del_init(&sif->node); node 440 drivers/base/class.c list_add_tail(&class_intf->node, &parent->p->interfaces); node 462 drivers/base/class.c list_del_init(&class_intf->node); node 63 drivers/base/component.c struct list_head node; node 73 drivers/base/component.c struct list_head node; node 155 drivers/base/component.c list_for_each_entry(m, &masters, node) node 167 drivers/base/component.c list_for_each_entry(c, &component_list, node) { node 274 drivers/base/component.c list_for_each_entry(m, &masters, node) { node 443 drivers/base/component.c list_del(&master->node); node 491 drivers/base/component.c list_add(&master->node, &masters); node 683 drivers/base/component.c list_add_tail(&component->node, &component_list); node 689 drivers/base/component.c list_del(&component->node); node 762 drivers/base/component.c list_for_each_entry(c, &component_list, node) node 764 drivers/base/component.c list_del(&c->node); node 2215 drivers/base/core.c &dev->class->p->interfaces, node) node 2367 drivers/base/core.c &dev->class->p->interfaces, node) node 28 drivers/base/devres.c struct devres_node node; node 40 drivers/base/devres.c struct devres_node node[2]; node 50 drivers/base/devres.c static void set_node_dbginfo(struct devres_node *node, const char *name, node 53 drivers/base/devres.c node->name = name; node 54 drivers/base/devres.c node->size = size; node 57 drivers/base/devres.c static void devres_log(struct device *dev, struct devres_node *node, node 62 drivers/base/devres.c op, node, node->name, (unsigned long)node->size); node 65 drivers/base/devres.c #define set_node_dbginfo(node, n, s) do {} while (0) node 66 drivers/base/devres.c #define devres_log(dev, node, op) do {} while (0) node 83 drivers/base/devres.c static struct devres_group * node_to_group(struct devres_node *node) node 85 drivers/base/devres.c if (node->release == &group_open_release) node 86 drivers/base/devres.c return container_of(node, struct devres_group, node[0]); node 87 drivers/base/devres.c if (node->release == &group_close_release) node 88 drivers/base/devres.c return container_of(node, struct devres_group, node[1]); node 109 drivers/base/devres.c INIT_LIST_HEAD(&dr->node.entry); node 110 drivers/base/devres.c dr->node.release = release; node 114 drivers/base/devres.c static void add_dr(struct device *dev, struct devres_node *node) node 116 drivers/base/devres.c devres_log(dev, node, "ADD"); node 117 drivers/base/devres.c BUG_ON(!list_empty(&node->entry)); node 118 drivers/base/devres.c list_add_tail(&node->entry, &dev->devres_head); node 130 drivers/base/devres.c set_node_dbginfo(&dr->node, name, size); node 181 drivers/base/devres.c struct devres_node *node; node 189 drivers/base/devres.c list_for_each_entry_safe_reverse(node, tmp, node 191 drivers/base/devres.c struct devres *dr = container_of(node, struct devres, node); node 193 drivers/base/devres.c if (node->release != release) node 214 drivers/base/devres.c BUG_ON(!list_empty(&dr->node.entry)); node 235 drivers/base/devres.c add_dr(dev, &dr->node); node 243 drivers/base/devres.c struct devres_node *node; node 245 drivers/base/devres.c list_for_each_entry_reverse(node, &dev->devres_head, entry) { node 246 drivers/base/devres.c struct devres *dr = container_of(node, struct devres, node); node 248 drivers/base/devres.c if (node->release != release) node 310 drivers/base/devres.c dr = find_dr(dev, new_dr->node.release, match, match_data); node 312 drivers/base/devres.c add_dr(dev, &new_dr->node); node 347 drivers/base/devres.c list_del_init(&dr->node.entry); node 348 drivers/base/devres.c devres_log(dev, &dr->node, "REM"); node 433 drivers/base/devres.c struct devres_node *node; node 436 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); node 439 drivers/base/devres.c grp = node_to_group(node); node 446 drivers/base/devres.c if (&node->entry == first) node 448 drivers/base/devres.c list_move_tail(&node->entry, todo); node 464 drivers/base/devres.c struct devres_node *node; node 467 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); node 470 drivers/base/devres.c grp = node_to_group(node); node 471 drivers/base/devres.c BUG_ON(!grp || list_empty(&grp->node[0].entry)); node 474 drivers/base/devres.c if (list_empty(&grp->node[1].entry)) node 482 drivers/base/devres.c list_move_tail(&grp->node[0].entry, todo); node 483 drivers/base/devres.c list_del_init(&grp->node[1].entry); node 505 drivers/base/devres.c list_for_each_entry_safe_reverse(dr, tmp, &todo, node.entry) { node 506 drivers/base/devres.c devres_log(dev, &dr->node, "REL"); node 507 drivers/base/devres.c dr->node.release(dev, dr->data); node 555 drivers/base/devres.c grp->node[0].release = &group_open_release; node 556 drivers/base/devres.c grp->node[1].release = &group_close_release; node 557 drivers/base/devres.c INIT_LIST_HEAD(&grp->node[0].entry); node 558 drivers/base/devres.c INIT_LIST_HEAD(&grp->node[1].entry); node 559 drivers/base/devres.c set_node_dbginfo(&grp->node[0], "grp<", 0); node 560 drivers/base/devres.c set_node_dbginfo(&grp->node[1], "grp>", 0); node 566 drivers/base/devres.c add_dr(dev, &grp->node[0]); node 575 drivers/base/devres.c struct devres_node *node; node 577 drivers/base/devres.c list_for_each_entry_reverse(node, &dev->devres_head, entry) { node 580 drivers/base/devres.c if (node->release != &group_open_release) node 583 drivers/base/devres.c grp = container_of(node, struct devres_group, node[0]); node 588 drivers/base/devres.c } else if (list_empty(&grp->node[1].entry)) node 612 drivers/base/devres.c add_dr(dev, &grp->node[1]); node 638 drivers/base/devres.c list_del_init(&grp->node[0].entry); node 639 drivers/base/devres.c list_del_init(&grp->node[1].entry); node 640 drivers/base/devres.c devres_log(dev, &grp->node[0], "REM"); node 672 drivers/base/devres.c struct list_head *first = &grp->node[0].entry; node 675 drivers/base/devres.c if (!list_empty(&grp->node[1].entry)) node 676 drivers/base/devres.c end = grp->node[1].entry.next; node 823 drivers/base/devres.c set_node_dbginfo(&dr->node, "devm_kzalloc_release", size); node 34 drivers/base/node.c struct node *node_dev = to_node(dev); node 105 drivers/base/node.c static void node_remove_accesses(struct node *node) node 109 drivers/base/node.c list_for_each_entry_safe(c, cnext, &node->access_list, list_node) { node 120 drivers/base/node.c static struct node_access_nodes *node_init_node_access(struct node *node, node 126 drivers/base/node.c list_for_each_entry(access_node, &node->access_list, list_node) node 136 drivers/base/node.c dev->parent = &node->dev; node 146 drivers/base/node.c list_add_tail(&access_node->list_node, &node->access_list); node 188 drivers/base/node.c struct node *node; node 194 drivers/base/node.c node = node_devices[nid]; node 195 drivers/base/node.c c = node_init_node_access(node, access); node 218 drivers/base/node.c struct list_head node; node 257 drivers/base/node.c static void node_init_cache_dev(struct node *node) node 265 drivers/base/node.c dev->parent = &node->dev; node 274 drivers/base/node.c node->cache_dev = dev; node 291 drivers/base/node.c struct node *node; node 296 drivers/base/node.c node = node_devices[nid]; node 297 drivers/base/node.c list_for_each_entry(info, &node->cache_attrs, node) { node 299 drivers/base/node.c dev_warn(&node->dev, node 306 drivers/base/node.c if (!node->cache_dev) node 307 drivers/base/node.c node_init_cache_dev(node); node 308 drivers/base/node.c if (!node->cache_dev) node 316 drivers/base/node.c dev->parent = node->cache_dev; node 324 drivers/base/node.c dev_warn(&node->dev, "failed to add cache level:%d\n", node 329 drivers/base/node.c list_add_tail(&info->node, &node->cache_attrs); node 337 drivers/base/node.c static void node_remove_caches(struct node *node) node 341 drivers/base/node.c if (!node->cache_dev) node 344 drivers/base/node.c list_for_each_entry_safe(info, next, &node->cache_attrs, node) { node 345 drivers/base/node.c list_del(&info->node); node 348 drivers/base/node.c device_unregister(node->cache_dev); node 357 drivers/base/node.c static void node_remove_caches(struct node *node) { } node 564 drivers/base/node.c static inline bool hugetlb_register_node(struct node *node) node 567 drivers/base/node.c node_state(node->dev.id, N_MEMORY)) { node 568 drivers/base/node.c __hugetlb_register_node(node); node 574 drivers/base/node.c static inline void hugetlb_unregister_node(struct node *node) node 577 drivers/base/node.c __hugetlb_unregister_node(node); node 587 drivers/base/node.c static inline void hugetlb_register_node(struct node *node) {} node 589 drivers/base/node.c static inline void hugetlb_unregister_node(struct node *node) {} node 594 drivers/base/node.c struct node *node = to_node(dev); node 606 drivers/base/node.c flush_work(&node->node_work); node 608 drivers/base/node.c kfree(node); node 617 drivers/base/node.c static int register_node(struct node *node, int num) node 621 drivers/base/node.c node->dev.id = num; node 622 drivers/base/node.c node->dev.bus = &node_subsys; node 623 drivers/base/node.c node->dev.release = node_device_release; node 624 drivers/base/node.c node->dev.groups = node_dev_groups; node 625 drivers/base/node.c error = device_register(&node->dev); node 628 drivers/base/node.c put_device(&node->dev); node 630 drivers/base/node.c hugetlb_register_node(node); node 632 drivers/base/node.c compaction_register_node(node); node 644 drivers/base/node.c void unregister_node(struct node *node) node 646 drivers/base/node.c hugetlb_unregister_node(node); /* no-op, if memoryless node */ node 647 drivers/base/node.c node_remove_accesses(node); node 648 drivers/base/node.c node_remove_caches(node); node 649 drivers/base/node.c device_unregister(&node->dev); node 652 drivers/base/node.c struct node *node_devices[MAX_NUMNODES]; node 697 drivers/base/node.c struct node *init_node, *targ_node; node 846 drivers/base/node.c struct node *node = container_of(work, struct node, node_work); node 856 drivers/base/node.c if (!hugetlb_register_node(node)) node 857 drivers/base/node.c hugetlb_unregister_node(node); node 912 drivers/base/node.c node_devices[nid] = kzalloc(sizeof(struct node), GFP_KERNEL); node 31 drivers/base/power/clock_ops.c struct list_head node; node 110 drivers/base/power/clock_ops.c list_add_tail(&ce->node, &psd->clock_list); node 274 drivers/base/power/clock_ops.c list_for_each_entry(ce, &psd->clock_list, node) { node 287 drivers/base/power/clock_ops.c list_del(&ce->node); node 312 drivers/base/power/clock_ops.c list_for_each_entry(ce, &psd->clock_list, node) { node 321 drivers/base/power/clock_ops.c list_del(&ce->node); node 377 drivers/base/power/clock_ops.c list_for_each_entry_safe_reverse(ce, c, &psd->clock_list, node) node 378 drivers/base/power/clock_ops.c list_move(&ce->node, &list); node 384 drivers/base/power/clock_ops.c list_for_each_entry_safe_reverse(ce, c, &list, node) { node 385 drivers/base/power/clock_ops.c list_del(&ce->node); node 408 drivers/base/power/clock_ops.c list_for_each_entry_reverse(ce, &psd->clock_list, node) { node 439 drivers/base/power/clock_ops.c list_for_each_entry(ce, &psd->clock_list, node) node 1939 drivers/base/power/domain.c struct device_node *node; node 2011 drivers/base/power/domain.c cp->node = of_node_get(np); node 2173 drivers/base/power/domain.c if (cp->node == np) { node 2192 drivers/base/power/domain.c of_node_put(cp->node); node 2226 drivers/base/power/domain.c if (provider->node == genpdspec->np) node 2622 drivers/base/power/domain.c np = it.node; node 273 drivers/base/power/qos.c list_for_each_entry_safe(req, tmp, &f->list, data.flr.node) { node 192 drivers/base/regmap/internal.h struct rb_node node; node 31 drivers/base/regmap/regcache-rbtree.c struct rb_node node; node 66 drivers/base/regmap/regcache-rbtree.c struct rb_node *node; node 78 drivers/base/regmap/regcache-rbtree.c node = rbtree_ctx->root.rb_node; node 79 drivers/base/regmap/regcache-rbtree.c while (node) { node 80 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); node 87 drivers/base/regmap/regcache-rbtree.c node = node->rb_right; node 89 drivers/base/regmap/regcache-rbtree.c node = node->rb_left; node 107 drivers/base/regmap/regcache-rbtree.c rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node); node 125 drivers/base/regmap/regcache-rbtree.c rb_link_node(&rbnode->node, parent, new); node 126 drivers/base/regmap/regcache-rbtree.c rb_insert_color(&rbnode->node, root); node 137 drivers/base/regmap/regcache-rbtree.c struct rb_node *node; node 148 drivers/base/regmap/regcache-rbtree.c for (node = rb_first(&rbtree_ctx->root); node != NULL; node 149 drivers/base/regmap/regcache-rbtree.c node = rb_next(node)) { node 150 drivers/base/regmap/regcache-rbtree.c n = rb_entry(node, struct regcache_rbtree_node, node); node 227 drivers/base/regmap/regcache-rbtree.c rbtree_node = rb_entry(next, struct regcache_rbtree_node, node); node 228 drivers/base/regmap/regcache-rbtree.c next = rb_next(&rbtree_node->node); node 229 drivers/base/regmap/regcache-rbtree.c rb_erase(&rbtree_node->node, &rbtree_ctx->root); node 374 drivers/base/regmap/regcache-rbtree.c struct rb_node *node; node 403 drivers/base/regmap/regcache-rbtree.c node = rbtree_ctx->root.rb_node; node 404 drivers/base/regmap/regcache-rbtree.c while (node) { node 405 drivers/base/regmap/regcache-rbtree.c rbnode_tmp = rb_entry(node, struct regcache_rbtree_node, node 406 drivers/base/regmap/regcache-rbtree.c node); node 432 drivers/base/regmap/regcache-rbtree.c node = node->rb_left; node 434 drivers/base/regmap/regcache-rbtree.c node = node->rb_right; node 469 drivers/base/regmap/regcache-rbtree.c struct rb_node *node; node 476 drivers/base/regmap/regcache-rbtree.c for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { node 477 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); node 511 drivers/base/regmap/regcache-rbtree.c struct rb_node *node; node 516 drivers/base/regmap/regcache-rbtree.c for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { node 517 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); node 555 drivers/base/regmap/regmap-debugfs.c struct regmap_debugfs_node *node; node 556 drivers/base/regmap/regmap-debugfs.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 557 drivers/base/regmap/regmap-debugfs.c if (!node) node 559 drivers/base/regmap/regmap-debugfs.c node->map = map; node 560 drivers/base/regmap/regmap-debugfs.c node->name = name; node 562 drivers/base/regmap/regmap-debugfs.c list_add(&node->link, ®map_debugfs_early_list); node 625 drivers/base/regmap/regmap-debugfs.c range_node = rb_entry(next, struct regmap_range_node, node); node 632 drivers/base/regmap/regmap-debugfs.c next = rb_next(&range_node->node); node 648 drivers/base/regmap/regmap-debugfs.c struct regmap_debugfs_node *node, *tmp; node 651 drivers/base/regmap/regmap-debugfs.c list_for_each_entry_safe(node, tmp, ®map_debugfs_early_list, node 653 drivers/base/regmap/regmap-debugfs.c if (node->map == map) { node 654 drivers/base/regmap/regmap-debugfs.c list_del(&node->link); node 655 drivers/base/regmap/regmap-debugfs.c kfree(node); node 664 drivers/base/regmap/regmap-debugfs.c struct regmap_debugfs_node *node, *tmp; node 669 drivers/base/regmap/regmap-debugfs.c list_for_each_entry_safe(node, tmp, ®map_debugfs_early_list, link) { node 670 drivers/base/regmap/regmap-debugfs.c regmap_debugfs_init(node->map, node->name); node 671 drivers/base/regmap/regmap-debugfs.c list_del(&node->link); node 672 drivers/base/regmap/regmap-debugfs.c kfree(node); node 539 drivers/base/regmap/regmap.c rb_entry(*new, struct regmap_range_node, node); node 550 drivers/base/regmap/regmap.c rb_link_node(&data->node, parent, new); node 551 drivers/base/regmap/regmap.c rb_insert_color(&data->node, root); node 559 drivers/base/regmap/regmap.c struct rb_node *node = map->range_tree.rb_node; node 561 drivers/base/regmap/regmap.c while (node) { node 563 drivers/base/regmap/regmap.c rb_entry(node, struct regmap_range_node, node); node 566 drivers/base/regmap/regmap.c node = node->rb_left; node 568 drivers/base/regmap/regmap.c node = node->rb_right; node 583 drivers/base/regmap/regmap.c range_node = rb_entry(next, struct regmap_range_node, node); node 584 drivers/base/regmap/regmap.c next = rb_next(&range_node->node); node 585 drivers/base/regmap/regmap.c rb_erase(&range_node->node, &map->range_tree); node 18 drivers/base/swnode.c const struct software_node *node; node 52 drivers/base/swnode.c software_node_to_swnode(const struct software_node *node) node 57 drivers/base/swnode.c if (!node) node 64 drivers/base/swnode.c if (swnode->node == node) node 78 drivers/base/swnode.c return swnode ? swnode->node : NULL; node 82 drivers/base/swnode.c struct fwnode_handle *software_node_fwnode(const struct software_node *node) node 84 drivers/base/swnode.c struct swnode *swnode = software_node_to_swnode(node); node 494 drivers/base/swnode.c return !!property_entry_get(swnode->node->properties, propname); node 504 drivers/base/swnode.c return property_entry_read_int_array(swnode->node->properties, propname, node 514 drivers/base/swnode.c return property_entry_read_string_array(swnode->node->properties, node 578 drivers/base/swnode.c if (!swnode || !swnode->node->references) node 581 drivers/base/swnode.c for (ref = swnode->node->references; ref->name; ref++) node 588 drivers/base/swnode.c refnode = software_node_fwnode(ref->refs[index].node); node 593 drivers/base/swnode.c prop = property_entry_get(swnode->node->properties, nargs_prop); node 649 drivers/base/swnode.c if (parent == swnode->node->parent && swnode->node->name && node 650 drivers/base/swnode.c !strcmp(name, swnode->node->name)) { node 659 drivers/base/swnode.c return swnode ? swnode->node : NULL; node 664 drivers/base/swnode.c software_node_register_properties(struct software_node *node, node 673 drivers/base/swnode.c node->properties = props; node 690 drivers/base/swnode.c property_entries_free(swnode->node->properties); node 691 drivers/base/swnode.c kfree(swnode->node); node 703 drivers/base/swnode.c swnode_register(const struct software_node *node, struct swnode *parent, node 723 drivers/base/swnode.c swnode->node = node; node 733 drivers/base/swnode.c if (node->name) node 736 drivers/base/swnode.c "%s", node->name); node 754 drivers/base/swnode.c property_entries_free(node->properties); node 804 drivers/base/swnode.c int software_node_register(const struct software_node *node) node 806 drivers/base/swnode.c struct swnode *parent = software_node_to_swnode(node->parent); node 808 drivers/base/swnode.c if (software_node_to_swnode(node)) node 811 drivers/base/swnode.c return PTR_ERR_OR_ZERO(swnode_register(node, parent, 0)); node 819 drivers/base/swnode.c struct software_node *node; node 831 drivers/base/swnode.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 832 drivers/base/swnode.c if (!node) node 835 drivers/base/swnode.c ret = software_node_register_properties(node, properties); node 837 drivers/base/swnode.c kfree(node); node 841 drivers/base/swnode.c node->parent = p ? p->node : NULL; node 843 drivers/base/swnode.c return swnode_register(node, p, 1); node 24 drivers/base/syscore.c list_add_tail(&ops->node, &syscore_ops_list); node 36 drivers/base/syscore.c list_del(&ops->node); node 62 drivers/base/syscore.c list_for_each_entry_reverse(ops, &syscore_ops_list, node) node 79 drivers/base/syscore.c list_for_each_entry_continue(ops, &syscore_ops_list, node) node 100 drivers/base/syscore.c list_for_each_entry(ops, &syscore_ops_list, node) node 122 drivers/base/syscore.c list_for_each_entry_reverse(ops, &syscore_ops_list, node) node 142 drivers/bcma/main.c struct device_node *node; node 149 drivers/bcma/main.c for_each_child_of_node(parent->of_node, node) { node 150 drivers/bcma/main.c reg = of_get_address(node, 0, &size, NULL); node 153 drivers/bcma/main.c if (of_translate_address(node, reg) == core->addr) node 154 drivers/bcma/main.c return node; node 202 drivers/bcma/main.c struct device_node *node; node 204 drivers/bcma/main.c node = bcma_of_find_child_device(parent, core); node 205 drivers/bcma/main.c if (node) node 206 drivers/bcma/main.c core->dev.of_node = node; node 210 drivers/bcma/main.c of_dma_configure(&core->dev, node, false); node 10 drivers/block/drbd/drbd_interval.c sector_t interval_end(struct rb_node *node) node 12 drivers/block/drbd/drbd_interval.c struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb); node 16 drivers/block/drbd/drbd_interval.c #define NODE_END(node) ((node)->sector + ((node)->size >> 9)) node 71 drivers/block/drbd/drbd_interval.c struct rb_node *node = root->rb_node; node 73 drivers/block/drbd/drbd_interval.c while (node) { node 75 drivers/block/drbd/drbd_interval.c rb_entry(node, struct drbd_interval, rb); node 78 drivers/block/drbd/drbd_interval.c node = node->rb_left; node 80 drivers/block/drbd/drbd_interval.c node = node->rb_right; node 82 drivers/block/drbd/drbd_interval.c node = node->rb_left; node 84 drivers/block/drbd/drbd_interval.c node = node->rb_right; node 114 drivers/block/drbd/drbd_interval.c struct rb_node *node = root->rb_node; node 120 drivers/block/drbd/drbd_interval.c while (node) { node 122 drivers/block/drbd/drbd_interval.c rb_entry(node, struct drbd_interval, rb); node 124 drivers/block/drbd/drbd_interval.c if (node->rb_left && node 125 drivers/block/drbd/drbd_interval.c sector < interval_end(node->rb_left)) { node 127 drivers/block/drbd/drbd_interval.c node = node->rb_left; node 134 drivers/block/drbd/drbd_interval.c node = node->rb_right; node 145 drivers/block/drbd/drbd_interval.c struct rb_node *node; node 148 drivers/block/drbd/drbd_interval.c node = rb_next(&i->rb); node 149 drivers/block/drbd/drbd_interval.c if (!node) node 151 drivers/block/drbd/drbd_interval.c i = rb_entry(node, struct drbd_interval, rb); node 3891 drivers/block/mtip32xx/mtip32xx.c static int get_least_used_cpu_on_node(int node) node 3896 drivers/block/mtip32xx/mtip32xx.c node_mask = cpumask_of_node(node); node 623 drivers/block/pktcdvd.c static inline struct pkt_rb_node *pkt_rbtree_next(struct pkt_rb_node *node) node 625 drivers/block/pktcdvd.c struct rb_node *n = rb_next(&node->rb_node); node 631 drivers/block/pktcdvd.c static void pkt_rbtree_erase(struct pktcdvd_device *pd, struct pkt_rb_node *node) node 633 drivers/block/pktcdvd.c rb_erase(&node->rb_node, &pd->bio_queue); node 634 drivers/block/pktcdvd.c mempool_free(node, &pd->rb_pool); node 676 drivers/block/pktcdvd.c static void pkt_rbtree_insert(struct pktcdvd_device *pd, struct pkt_rb_node *node) node 680 drivers/block/pktcdvd.c sector_t s = node->bio->bi_iter.bi_sector; node 691 drivers/block/pktcdvd.c rb_link_node(&node->rb_node, parent, p); node 692 drivers/block/pktcdvd.c rb_insert_color(&node->rb_node, &pd->bio_queue); node 1167 drivers/block/pktcdvd.c struct pkt_rb_node *node, *first_node; node 1188 drivers/block/pktcdvd.c node = first_node; node 1189 drivers/block/pktcdvd.c while (node) { node 1190 drivers/block/pktcdvd.c bio = node->bio; node 1200 drivers/block/pktcdvd.c node = pkt_rbtree_next(node); node 1201 drivers/block/pktcdvd.c if (!node) { node 1204 drivers/block/pktcdvd.c node = rb_entry(n, struct pkt_rb_node, rb_node); node 1206 drivers/block/pktcdvd.c if (node == first_node) node 1207 drivers/block/pktcdvd.c node = NULL; node 1228 drivers/block/pktcdvd.c while ((node = pkt_rbtree_find(pd, zone)) != NULL) { node 1229 drivers/block/pktcdvd.c bio = node->bio; node 1234 drivers/block/pktcdvd.c pkt_rbtree_erase(pd, node); node 2353 drivers/block/pktcdvd.c struct pkt_rb_node *node; node 2407 drivers/block/pktcdvd.c node = mempool_alloc(&pd->rb_pool, GFP_NOIO); node 2408 drivers/block/pktcdvd.c node->bio = bio; node 2412 drivers/block/pktcdvd.c pkt_rbtree_insert(pd, node); node 207 drivers/block/rbd.c struct list_head node; node 452 drivers/block/rbd.c struct list_head node; node 756 drivers/block/rbd.c INIT_LIST_HEAD(&rbdc->node); node 768 drivers/block/rbd.c list_add_tail(&rbdc->node, &rbd_client_list); node 806 drivers/block/rbd.c list_for_each_entry(client_node, &rbd_client_list, node) { node 982 drivers/block/rbd.c list_del(&rbdc->node); node 5544 drivers/block/rbd.c INIT_LIST_HEAD(&rbd_dev->node); node 7133 drivers/block/rbd.c list_add_tail(&rbd_dev->node, &rbd_dev_list); node 7230 drivers/block/rbd.c rbd_dev = list_entry(tmp, struct rbd_device, node); node 7260 drivers/block/rbd.c list_del_init(&rbd_dev->node); node 212 drivers/block/xen-blkback/blkback.c #define foreach_grant_safe(pos, n, rbtree, node) \ node 213 drivers/block/xen-blkback/blkback.c for ((pos) = container_of(rb_first((rbtree)), typeof(*(pos)), node), \ node 214 drivers/block/xen-blkback/blkback.c (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL; \ node 215 drivers/block/xen-blkback/blkback.c &(pos)->node != NULL; \ node 216 drivers/block/xen-blkback/blkback.c (pos) = container_of(n, typeof(*(pos)), node), \ node 217 drivers/block/xen-blkback/blkback.c (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL) node 245 drivers/block/xen-blkback/blkback.c this = container_of(*new, struct persistent_gnt, node); node 260 drivers/block/xen-blkback/blkback.c rb_link_node(&(persistent_gnt->node), parent, new); node 261 drivers/block/xen-blkback/blkback.c rb_insert_color(&(persistent_gnt->node), &ring->persistent_gnts); node 271 drivers/block/xen-blkback/blkback.c struct rb_node *node = NULL; node 273 drivers/block/xen-blkback/blkback.c node = ring->persistent_gnts.rb_node; node 274 drivers/block/xen-blkback/blkback.c while (node) { node 275 drivers/block/xen-blkback/blkback.c data = container_of(node, struct persistent_gnt, node); node 278 drivers/block/xen-blkback/blkback.c node = node->rb_left; node 280 drivers/block/xen-blkback/blkback.c node = node->rb_right; node 318 drivers/block/xen-blkback/blkback.c foreach_grant_safe(persistent_gnt, n, root, node) { node 330 drivers/block/xen-blkback/blkback.c !rb_next(&persistent_gnt->node)) { node 339 drivers/block/xen-blkback/blkback.c rb_erase(&persistent_gnt->node, root); node 427 drivers/block/xen-blkback/blkback.c foreach_grant_safe(persistent_gnt, n, root, node) { node 438 drivers/block/xen-blkback/blkback.c rb_erase(&persistent_gnt->node, root); node 245 drivers/block/xen-blkback/common.h struct rb_node node; node 511 drivers/block/xen-blkback/xenbus.c if (be->backend_watch.node) { node 513 drivers/block/xen-blkback/xenbus.c kfree(be->backend_watch.node); node 514 drivers/block/xen-blkback/xenbus.c be->backend_watch.node = NULL; node 88 drivers/block/xen-blkfront.c struct list_head node; node 311 drivers/block/xen-blkfront.c list_add(&gnt_list_entry->node, &rinfo->grants); node 319 drivers/block/xen-blkfront.c &rinfo->grants, node) { node 320 drivers/block/xen-blkfront.c list_del(&gnt_list_entry->node); node 336 drivers/block/xen-blkfront.c node); node 337 drivers/block/xen-blkfront.c list_del(&gnt_list_entry->node); node 1266 drivers/block/xen-blkfront.c &rinfo->grants, node) { node 1267 drivers/block/xen-blkfront.c list_del(&persistent_gnt->node); node 1505 drivers/block/xen-blkfront.c list_add(&s->grants_used[i]->node, &rinfo->grants); node 1516 drivers/block/xen-blkfront.c list_add_tail(&s->grants_used[i]->node, &rinfo->grants); node 1525 drivers/block/xen-blkfront.c list_add(&s->indirect_grants[i]->node, &rinfo->grants); node 1540 drivers/block/xen-blkfront.c list_add_tail(&s->indirect_grants[i]->node, &rinfo->grants); node 2676 drivers/block/xen-blkfront.c node) { node 2681 drivers/block/xen-blkfront.c list_del(&gnt_list_entry->node); node 2685 drivers/block/xen-blkfront.c list_add_tail(&gnt_list_entry->node, &rinfo->grants); node 158 drivers/block/zram/zcomp.c int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) node 160 drivers/block/zram/zcomp.c struct zcomp *comp = hlist_entry(node, struct zcomp, node); node 175 drivers/block/zram/zcomp.c int zcomp_cpu_dead(unsigned int cpu, struct hlist_node *node) node 177 drivers/block/zram/zcomp.c struct zcomp *comp = hlist_entry(node, struct zcomp, node); node 195 drivers/block/zram/zcomp.c ret = cpuhp_state_add_instance(CPUHP_ZCOMP_PREPARE, &comp->node); node 207 drivers/block/zram/zcomp.c cpuhp_state_remove_instance(CPUHP_ZCOMP_PREPARE, &comp->node); node 19 drivers/block/zram/zcomp.h struct hlist_node node; node 22 drivers/block/zram/zcomp.h int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node); node 23 drivers/block/zram/zcomp.h int zcomp_cpu_dead(unsigned int cpu, struct hlist_node *node); node 901 drivers/bluetooth/btmtkuart.c struct device_node *node = serdev->dev.of_node; node 906 drivers/bluetooth/btmtkuart.c of_property_read_u32(node, "current-speed", &speed); node 76 drivers/bus/fsl-mc/fsl-mc-allocator.c INIT_LIST_HEAD(&resource->node); node 77 drivers/bus/fsl-mc/fsl-mc-allocator.c list_add_tail(&resource->node, &res_pool->free_list); node 131 drivers/bus/fsl-mc/fsl-mc-allocator.c if (list_empty(&resource->node)) { node 139 drivers/bus/fsl-mc/fsl-mc-allocator.c list_del_init(&resource->node); node 197 drivers/bus/fsl-mc/fsl-mc-allocator.c struct fsl_mc_resource, node); node 215 drivers/bus/fsl-mc/fsl-mc-allocator.c list_del_init(&resource->node); node 240 drivers/bus/fsl-mc/fsl-mc-allocator.c if (!list_empty(&resource->node)) node 243 drivers/bus/fsl-mc/fsl-mc-allocator.c list_add_tail(&resource->node, &res_pool->free_list); node 385 drivers/bus/fsl-mc/fsl-mc-allocator.c INIT_LIST_HEAD(&mc_dev_irq->resource.node); node 386 drivers/bus/fsl-mc/fsl-mc-allocator.c list_add_tail(&mc_dev_irq->resource.node, &res_pool->free_list); node 561 drivers/bus/fsl-mc/fsl-mc-allocator.c list_for_each_entry_safe(resource, next, &res_pool->free_list, node) { node 422 drivers/bus/hisi_lpc.c list_for_each_entry(rentry, &resource_list, node) node 466 drivers/bus/hisi_lpc.c list_for_each_entry(child, &adev->children, node) node 487 drivers/bus/hisi_lpc.c list_for_each_entry(child, &adev->children, node) { node 1203 drivers/bus/mvebu-mbus.c mbus_parse_ranges(struct device_node *node, node 1212 drivers/bus/mvebu-mbus.c *ranges_start = of_get_property(node, "ranges", &ranges_len); node 1220 drivers/bus/mvebu-mbus.c *addr_cells = of_n_addr_cells(node); node 1222 drivers/bus/mvebu-mbus.c prop = of_get_property(node, "#address-cells", NULL); node 1225 drivers/bus/mvebu-mbus.c prop = of_get_property(node, "#size-cells", NULL); node 1232 drivers/bus/mvebu-mbus.c pr_warn("malformed ranges entry '%pOFn'\n", node); node 199 drivers/bus/sunxi-rsb.c struct device_node *node, u16 hwaddr, u8 rtaddr) node 213 drivers/bus/sunxi-rsb.c rdev->dev.of_node = node; node 46 drivers/bus/vexpress-config.c static void vexpress_config_find_prop(struct device_node *node, node 52 drivers/bus/vexpress-config.c of_node_get(node); node 53 drivers/bus/vexpress-config.c while (node) { node 54 drivers/bus/vexpress-config.c if (of_property_read_u32(node, name, val) == 0) { node 55 drivers/bus/vexpress-config.c of_node_put(node); node 58 drivers/bus/vexpress-config.c node = of_get_next_parent(node); node 62 drivers/bus/vexpress-config.c int vexpress_config_get_topo(struct device_node *node, u32 *site, node 65 drivers/bus/vexpress-config.c vexpress_config_find_prop(node, "arm,vexpress,site", site); node 70 drivers/bus/vexpress-config.c vexpress_config_find_prop(node, "arm,vexpress,position", position); node 71 drivers/bus/vexpress-config.c vexpress_config_find_prop(node, "arm,vexpress,dcc", dcc); node 155 drivers/bus/vexpress-config.c const struct device_node *node = data; node 158 drivers/bus/vexpress-config.c dev->parent->of_node, node); node 160 drivers/bus/vexpress-config.c return dev->parent->of_node == node; node 163 drivers/bus/vexpress-config.c static int vexpress_config_populate(struct device_node *node) node 169 drivers/bus/vexpress-config.c bridge = of_parse_phandle(node, "arm,vexpress,config-bridge", 0); node 179 drivers/bus/vexpress-config.c ret = of_platform_populate(node, NULL, NULL, parent); node 189 drivers/bus/vexpress-config.c struct device_node *node; node 192 drivers/bus/vexpress-config.c for_each_compatible_node(node, NULL, "arm,vexpress,config-bus") { node 193 drivers/bus/vexpress-config.c err = vexpress_config_populate(node); node 195 drivers/bus/vexpress-config.c of_node_put(node); node 169 drivers/clk/axs10x/i2s_pll_clock.c struct device_node *node = dev->of_node; node 187 drivers/clk/axs10x/i2s_pll_clock.c clk_name = node->name; node 190 drivers/clk/axs10x/i2s_pll_clock.c parent_name = of_clk_get_parent_name(node, 0); node 203 drivers/clk/axs10x/i2s_pll_clock.c return of_clk_add_provider(node, of_clk_src_simple_get, clk); node 272 drivers/clk/axs10x/pll_clock.c static void __init of_axs10x_pll_clk_setup(struct device_node *node) node 283 drivers/clk/axs10x/pll_clock.c pll_clk->base = of_iomap(node, 0); node 289 drivers/clk/axs10x/pll_clock.c pll_clk->lock = of_iomap(node, 1); node 295 drivers/clk/axs10x/pll_clock.c init.name = node->name; node 297 drivers/clk/axs10x/pll_clock.c parent_name = of_clk_get_parent_name(node, 0); node 305 drivers/clk/axs10x/pll_clock.c pr_err("failed to register %pOFn clock\n", node); node 309 drivers/clk/axs10x/pll_clock.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clk->hw); node 311 drivers/clk/axs10x/pll_clock.c pr_err("failed to add hw provider for %pOFn clock\n", node); node 263 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_root_ccu_setup(struct device_node *node) node 265 drivers/clk/bcm/clk-bcm21664.c kona_dt_ccu_setup(&root_ccu_data, node); node 268 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_aon_ccu_setup(struct device_node *node) node 270 drivers/clk/bcm/clk-bcm21664.c kona_dt_ccu_setup(&aon_ccu_data, node); node 273 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_master_ccu_setup(struct device_node *node) node 275 drivers/clk/bcm/clk-bcm21664.c kona_dt_ccu_setup(&master_ccu_data, node); node 278 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_slave_ccu_setup(struct device_node *node) node 280 drivers/clk/bcm/clk-bcm21664.c kona_dt_ccu_setup(&slave_ccu_data, node); node 341 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_root_ccu_setup(struct device_node *node) node 343 drivers/clk/bcm/clk-bcm281xx.c kona_dt_ccu_setup(&root_ccu_data, node); node 346 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_aon_ccu_setup(struct device_node *node) node 348 drivers/clk/bcm/clk-bcm281xx.c kona_dt_ccu_setup(&aon_ccu_data, node); node 351 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_hub_ccu_setup(struct device_node *node) node 353 drivers/clk/bcm/clk-bcm281xx.c kona_dt_ccu_setup(&hub_ccu_data, node); node 356 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_master_ccu_setup(struct device_node *node) node 358 drivers/clk/bcm/clk-bcm281xx.c kona_dt_ccu_setup(&master_ccu_data, node); node 361 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_slave_ccu_setup(struct device_node *node) node 363 drivers/clk/bcm/clk-bcm281xx.c kona_dt_ccu_setup(&slave_ccu_data, node); node 18 drivers/clk/bcm/clk-bcm63xx.c static void __init bcm63138_armpll_init(struct device_node *node) node 20 drivers/clk/bcm/clk-bcm63xx.c iproc_armpll_setup(node); node 51 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_armpll_init(struct device_node *node) node 53 drivers/clk/bcm/clk-cygnus.c iproc_armpll_setup(node); node 110 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_genpll_clk_init(struct device_node *node) node 112 drivers/clk/bcm/clk-cygnus.c iproc_pll_clk_setup(node, &genpll, NULL, 0, genpll_clk, node 168 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_lcpll0_clk_init(struct device_node *node) node 170 drivers/clk/bcm/clk-cygnus.c iproc_pll_clk_setup(node, &lcpll0, NULL, 0, lcpll0_clk, node 246 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_mipipll_clk_init(struct device_node *node) node 248 drivers/clk/bcm/clk-cygnus.c iproc_pll_clk_setup(node, &mipipll, mipipll_vco_params, node 266 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_asiu_init(struct device_node *node) node 268 drivers/clk/bcm/clk-cygnus.c iproc_asiu_setup(node, asiu_div, asiu_gate, ARRAY_SIZE(asiu_div)); node 308 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_audiopll_clk_init(struct device_node *node) node 310 drivers/clk/bcm/clk-cygnus.c iproc_pll_clk_setup(node, &audiopll, NULL, 0, node 23 drivers/clk/bcm/clk-hr2.c static void __init hr2_armpll_init(struct device_node *node) node 25 drivers/clk/bcm/clk-hr2.c iproc_armpll_setup(node); node 244 drivers/clk/bcm/clk-iproc-armpll.c void __init iproc_armpll_setup(struct device_node *node) node 255 drivers/clk/bcm/clk-iproc-armpll.c pll->base = of_iomap(node, 0); node 259 drivers/clk/bcm/clk-iproc-armpll.c init.name = node->name; node 262 drivers/clk/bcm/clk-iproc-armpll.c parent_name = of_clk_get_parent_name(node, 0); node 271 drivers/clk/bcm/clk-iproc-armpll.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll->hw); node 185 drivers/clk/bcm/clk-iproc-asiu.c void __init iproc_asiu_setup(struct device_node *node, node 210 drivers/clk/bcm/clk-iproc-asiu.c asiu->div_base = of_iomap(node, 0); node 214 drivers/clk/bcm/clk-iproc-asiu.c asiu->gate_base = of_iomap(node, 1); node 224 drivers/clk/bcm/clk-iproc-asiu.c ret = of_property_read_string_index(node, "clock-output-names", node 237 drivers/clk/bcm/clk-iproc-asiu.c parent_name = of_clk_get_parent_name(node, 0); node 248 drivers/clk/bcm/clk-iproc-asiu.c ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, node 725 drivers/clk/bcm/clk-iproc-pll.c void iproc_pll_clk_setup(struct device_node *node, node 756 drivers/clk/bcm/clk-iproc-pll.c pll->control_base = of_iomap(node, 0); node 761 drivers/clk/bcm/clk-iproc-pll.c pll->pwr_base = of_iomap(node, 1); node 765 drivers/clk/bcm/clk-iproc-pll.c pll->asiu_base = of_iomap(node, 2); node 774 drivers/clk/bcm/clk-iproc-pll.c pll->status_base = of_iomap(node, 2); node 786 drivers/clk/bcm/clk-iproc-pll.c init.name = node->name; node 789 drivers/clk/bcm/clk-iproc-pll.c parent_name = of_clk_get_parent_name(node, 0); node 812 drivers/clk/bcm/clk-iproc-pll.c parent_name = node->name; node 814 drivers/clk/bcm/clk-iproc-pll.c ret = of_property_read_string_index(node, "clock-output-names", node 837 drivers/clk/bcm/clk-iproc-pll.c ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, clk_data); node 212 drivers/clk/bcm/clk-iproc.h void iproc_armpll_setup(struct device_node *node); node 213 drivers/clk/bcm/clk-iproc.h void iproc_pll_clk_setup(struct device_node *node, node 219 drivers/clk/bcm/clk-iproc.h void iproc_asiu_setup(struct device_node *node, node 760 drivers/clk/bcm/clk-kona-setup.c of_clk_del_provider(ccu->node); /* safe if never added */ node 762 drivers/clk/bcm/clk-kona-setup.c of_node_put(ccu->node); node 763 drivers/clk/bcm/clk-kona-setup.c ccu->node = NULL; node 802 drivers/clk/bcm/clk-kona-setup.c struct device_node *node) node 809 drivers/clk/bcm/clk-kona-setup.c ret = of_address_to_resource(node, 0, &res); node 812 drivers/clk/bcm/clk-kona-setup.c node); node 819 drivers/clk/bcm/clk-kona-setup.c node); node 826 drivers/clk/bcm/clk-kona-setup.c pr_err("%s: ccu data not valid for %pOFn\n", __func__, node); node 833 drivers/clk/bcm/clk-kona-setup.c node); node 836 drivers/clk/bcm/clk-kona-setup.c ccu->node = of_node_get(node); node 849 drivers/clk/bcm/clk-kona-setup.c ret = of_clk_add_hw_provider(node, of_clk_kona_onecell_get, ccu); node 852 drivers/clk/bcm/clk-kona-setup.c node, ret); node 857 drivers/clk/bcm/clk-kona-setup.c pr_err("Broadcom %pOFn initialization had errors\n", node); node 862 drivers/clk/bcm/clk-kona-setup.c pr_err("Broadcom %pOFn setup aborted\n", node); node 483 drivers/clk/bcm/clk-kona.h struct device_node *node; node 507 drivers/clk/bcm/clk-kona.h struct device_node *node); node 96 drivers/clk/bcm/clk-ns2.c static void __init ns2_genpll_scr_clk_init(struct device_node *node) node 98 drivers/clk/bcm/clk-ns2.c iproc_pll_clk_setup(node, &genpll_scr, NULL, 0, genpll_scr_clk, node 158 drivers/clk/bcm/clk-ns2.c static void __init ns2_genpll_sw_clk_init(struct device_node *node) node 160 drivers/clk/bcm/clk-ns2.c iproc_pll_clk_setup(node, &genpll_sw, NULL, 0, genpll_sw_clk, node 220 drivers/clk/bcm/clk-ns2.c static void __init ns2_lcpll_ddr_clk_init(struct device_node *node) node 222 drivers/clk/bcm/clk-ns2.c iproc_pll_clk_setup(node, &lcpll_ddr, NULL, 0, lcpll_ddr_clk, node 282 drivers/clk/bcm/clk-ns2.c static void __init ns2_lcpll_ports_clk_init(struct device_node *node) node 284 drivers/clk/bcm/clk-ns2.c iproc_pll_clk_setup(node, &lcpll_ports, NULL, 0, lcpll_ports_clk, node 39 drivers/clk/bcm/clk-nsp.c static void __init nsp_armpll_init(struct device_node *node) node 41 drivers/clk/bcm/clk-nsp.c iproc_armpll_setup(node); node 95 drivers/clk/bcm/clk-nsp.c static void __init nsp_genpll_clk_init(struct device_node *node) node 97 drivers/clk/bcm/clk-nsp.c iproc_pll_clk_setup(node, &genpll, NULL, 0, genpll_clk, node 134 drivers/clk/bcm/clk-nsp.c static void __init nsp_lcpll0_clk_init(struct device_node *node) node 136 drivers/clk/bcm/clk-nsp.c iproc_pll_clk_setup(node, &lcpll0, NULL, 0, lcpll0_clk, node 181 drivers/clk/bcm/clk-sr.c static void sr_genpll3_clk_init(struct device_node *node) node 183 drivers/clk/bcm/clk-sr.c iproc_pll_clk_setup(node, &sr_genpll3, NULL, 0, sr_genpll3_clk, node 641 drivers/clk/clk-cdce925.c struct device_node *node = client->dev.of_node; node 683 drivers/clk/clk-cdce925.c parent_name = of_clk_get_parent_name(node, 0); node 690 drivers/clk/clk-cdce925.c if (of_property_read_u32(node, "xtal-load-pf", &value) == 0) node 718 drivers/clk/clk-cdce925.c np_output = of_get_child_by_name(node, child_name); node 14 drivers/clk/clk-conf.c static int __set_clk_parents(struct device_node *node, bool clk_supplier) node 20 drivers/clk/clk-conf.c num_parents = of_count_phandle_with_args(node, "assigned-clock-parents", node 24 drivers/clk/clk-conf.c node); node 27 drivers/clk/clk-conf.c rc = of_parse_phandle_with_args(node, "assigned-clock-parents", node 36 drivers/clk/clk-conf.c if (clkspec.np == node && !clk_supplier) node 42 drivers/clk/clk-conf.c index, node); node 46 drivers/clk/clk-conf.c rc = of_parse_phandle_with_args(node, "assigned-clocks", node 50 drivers/clk/clk-conf.c if (clkspec.np == node && !clk_supplier) { node 58 drivers/clk/clk-conf.c index, node); node 76 drivers/clk/clk-conf.c static int __set_clk_rates(struct device_node *node, bool clk_supplier) node 85 drivers/clk/clk-conf.c of_property_for_each_u32(node, "assigned-clock-rates", prop, cur, rate) { node 87 drivers/clk/clk-conf.c rc = of_parse_phandle_with_args(node, "assigned-clocks", node 96 drivers/clk/clk-conf.c if (clkspec.np == node && !clk_supplier) node 103 drivers/clk/clk-conf.c index, node); node 131 drivers/clk/clk-conf.c int of_clk_set_defaults(struct device_node *node, bool clk_supplier) node 135 drivers/clk/clk-conf.c if (!node) node 138 drivers/clk/clk-conf.c rc = __set_clk_parents(node, clk_supplier); node 142 drivers/clk/clk-conf.c return __set_clk_rates(node, clk_supplier); node 162 drivers/clk/clk-fixed-factor.c static struct clk_hw *_of_fixed_factor_clk_setup(struct device_node *node) node 165 drivers/clk/clk-fixed-factor.c const char *clk_name = node->name; node 170 drivers/clk/clk-fixed-factor.c if (of_property_read_u32(node, "clock-div", &div)) { node 172 drivers/clk/clk-fixed-factor.c __func__, node); node 176 drivers/clk/clk-fixed-factor.c if (of_property_read_u32(node, "clock-mult", &mult)) { node 178 drivers/clk/clk-fixed-factor.c __func__, node); node 182 drivers/clk/clk-fixed-factor.c of_property_read_string(node, "clock-output-names", &clk_name); node 184 drivers/clk/clk-fixed-factor.c if (of_match_node(set_rate_parent_matches, node)) node 187 drivers/clk/clk-fixed-factor.c hw = __clk_hw_register_fixed_factor(NULL, node, clk_name, NULL, 0, node 194 drivers/clk/clk-fixed-factor.c of_node_clear_flag(node, OF_POPULATED); node 198 drivers/clk/clk-fixed-factor.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 210 drivers/clk/clk-fixed-factor.c void __init of_fixed_factor_clk_setup(struct device_node *node) node 212 drivers/clk/clk-fixed-factor.c _of_fixed_factor_clk_setup(node); node 18 drivers/clk/clk-fixed-mmio.c static struct clk_hw *fixed_mmio_clk_setup(struct device_node *node) node 21 drivers/clk/clk-fixed-mmio.c const char *clk_name = node->name; node 26 drivers/clk/clk-fixed-mmio.c base = of_iomap(node, 0); node 28 drivers/clk/clk-fixed-mmio.c pr_err("%pOFn: failed to map address\n", node); node 34 drivers/clk/clk-fixed-mmio.c of_property_read_string(node, "clock-output-names", &clk_name); node 38 drivers/clk/clk-fixed-mmio.c pr_err("%pOFn: failed to register fixed rate clock\n", node); node 42 drivers/clk/clk-fixed-mmio.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, clk); node 44 drivers/clk/clk-fixed-mmio.c pr_err("%pOFn: failed to add clock provider\n", node); node 52 drivers/clk/clk-fixed-mmio.c static void __init of_fixed_mmio_clk_setup(struct device_node *node) node 54 drivers/clk/clk-fixed-mmio.c fixed_mmio_clk_setup(node); node 158 drivers/clk/clk-fixed-rate.c static struct clk *_of_fixed_clk_setup(struct device_node *node) node 161 drivers/clk/clk-fixed-rate.c const char *clk_name = node->name; node 166 drivers/clk/clk-fixed-rate.c if (of_property_read_u32(node, "clock-frequency", &rate)) node 169 drivers/clk/clk-fixed-rate.c of_property_read_u32(node, "clock-accuracy", &accuracy); node 171 drivers/clk/clk-fixed-rate.c of_property_read_string(node, "clock-output-names", &clk_name); node 178 drivers/clk/clk-fixed-rate.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 190 drivers/clk/clk-fixed-rate.c void __init of_fixed_clk_setup(struct device_node *node) node 192 drivers/clk/clk-fixed-rate.c _of_fixed_clk_setup(node); node 241 drivers/clk/clk-gpio.c struct device_node *node = pdev->dev.of_node; node 249 drivers/clk/clk-gpio.c num_parents = of_clk_get_parent_count(node); node 256 drivers/clk/clk-gpio.c of_clk_parent_fill(node, parent_names, num_parents); node 261 drivers/clk/clk-gpio.c is_mux = of_device_is_compatible(node, "gpio-mux-clock"); node 269 drivers/clk/clk-gpio.c node, __func__); node 272 drivers/clk/clk-gpio.c node, __func__, node 278 drivers/clk/clk-gpio.c clk = clk_register_gpio_mux(&pdev->dev, node->name, node 281 drivers/clk/clk-gpio.c clk = clk_register_gpio_gate(&pdev->dev, node->name, node 287 drivers/clk/clk-gpio.c return of_clk_add_provider(node, of_clk_src_simple_get, clk); node 263 drivers/clk/clk-highbank.c static void __init hb_clk_init(struct device_node *node, const struct clk_ops *ops, unsigned long clkflags) node 267 drivers/clk/clk-highbank.c const char *clk_name = node->name; node 273 drivers/clk/clk-highbank.c rc = of_property_read_u32(node, "reg", ®); node 288 drivers/clk/clk-highbank.c of_property_read_string(node, "clock-output-names", &clk_name); node 293 drivers/clk/clk-highbank.c parent_name = of_clk_get_parent_name(node, 0); node 304 drivers/clk/clk-highbank.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, &hb_clk->hw); node 307 drivers/clk/clk-highbank.c static void __init hb_pll_init(struct device_node *node) node 309 drivers/clk/clk-highbank.c hb_clk_init(node, &clk_pll_ops, 0); node 313 drivers/clk/clk-highbank.c static void __init hb_a9periph_init(struct device_node *node) node 315 drivers/clk/clk-highbank.c hb_clk_init(node, &a9periphclk_ops, 0); node 319 drivers/clk/clk-highbank.c static void __init hb_a9bus_init(struct device_node *node) node 321 drivers/clk/clk-highbank.c hb_clk_init(node, &a9bclk_ops, CLK_IS_CRITICAL); node 325 drivers/clk/clk-highbank.c static void __init hb_emmc_init(struct device_node *node) node 327 drivers/clk/clk-highbank.c hb_clk_init(node, &periclk_ops, 0); node 354 drivers/clk/clk-hsdk-pll.c static void __init of_hsdk_pll_clk_setup(struct device_node *node) node 366 drivers/clk/clk-hsdk-pll.c pll_clk->regs = of_iomap(node, 0); node 372 drivers/clk/clk-hsdk-pll.c pll_clk->spec_regs = of_iomap(node, 1); node 378 drivers/clk/clk-hsdk-pll.c init.name = node->name; node 380 drivers/clk/clk-hsdk-pll.c parent_name = of_clk_get_parent_name(node, 0); node 382 drivers/clk/clk-hsdk-pll.c num_parents = of_clk_get_parent_count(node); node 394 drivers/clk/clk-hsdk-pll.c pr_err("failed to register %pOFn clock\n", node); node 398 drivers/clk/clk-hsdk-pll.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clk->hw); node 400 drivers/clk/clk-hsdk-pll.c pr_err("failed to add hw provider for %pOFn clock\n", node); node 19 drivers/clk/clk-moxart.c static void __init moxart_of_pll_clk_init(struct device_node *node) node 25 drivers/clk/clk-moxart.c const char *name = node->name; node 28 drivers/clk/clk-moxart.c of_property_read_string(node, "clock-output-names", &name); node 29 drivers/clk/clk-moxart.c parent_name = of_clk_get_parent_name(node, 0); node 31 drivers/clk/clk-moxart.c base = of_iomap(node, 0); node 33 drivers/clk/clk-moxart.c pr_err("%pOF: of_iomap failed\n", node); node 40 drivers/clk/clk-moxart.c ref_clk = of_clk_get(node, 0); node 42 drivers/clk/clk-moxart.c pr_err("%pOF: of_clk_get failed\n", node); node 48 drivers/clk/clk-moxart.c pr_err("%pOF: failed to register clock\n", node); node 53 drivers/clk/clk-moxart.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 58 drivers/clk/clk-moxart.c static void __init moxart_of_apb_clk_init(struct device_node *node) node 65 drivers/clk/clk-moxart.c const char *name = node->name; node 68 drivers/clk/clk-moxart.c of_property_read_string(node, "clock-output-names", &name); node 69 drivers/clk/clk-moxart.c parent_name = of_clk_get_parent_name(node, 0); node 71 drivers/clk/clk-moxart.c base = of_iomap(node, 0); node 73 drivers/clk/clk-moxart.c pr_err("%pOF: of_iomap failed\n", node); node 84 drivers/clk/clk-moxart.c pll_clk = of_clk_get(node, 0); node 86 drivers/clk/clk-moxart.c pr_err("%pOF: of_clk_get failed\n", node); node 92 drivers/clk/clk-moxart.c pr_err("%pOF: failed to register clock\n", node); node 97 drivers/clk/clk-moxart.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 63 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup(struct device_node *node, node 69 drivers/clk/clk-nspire.c const char *clk_name = node->name; node 73 drivers/clk/clk-nspire.c io = of_iomap(node, 0); node 81 drivers/clk/clk-nspire.c of_property_read_string(node, "clock-output-names", &clk_name); node 82 drivers/clk/clk-nspire.c parent_name = of_clk_get_parent_name(node, 0); node 87 drivers/clk/clk-nspire.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 90 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup_cx(struct device_node *node) node 92 drivers/clk/clk-nspire.c nspire_ahbdiv_setup(node, nspire_clkinfo_cx); node 95 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup_classic(struct device_node *node) node 97 drivers/clk/clk-nspire.c nspire_ahbdiv_setup(node, nspire_clkinfo_classic); node 105 drivers/clk/clk-nspire.c static void __init nspire_clk_setup(struct device_node *node, node 111 drivers/clk/clk-nspire.c const char *clk_name = node->name; node 114 drivers/clk/clk-nspire.c io = of_iomap(node, 0); node 122 drivers/clk/clk-nspire.c of_property_read_string(node, "clock-output-names", &clk_name); node 127 drivers/clk/clk-nspire.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 137 drivers/clk/clk-nspire.c static void __init nspire_clk_setup_cx(struct device_node *node) node 139 drivers/clk/clk-nspire.c nspire_clk_setup(node, nspire_clkinfo_cx); node 142 drivers/clk/clk-nspire.c static void __init nspire_clk_setup_classic(struct device_node *node) node 144 drivers/clk/clk-nspire.c nspire_clk_setup(node, nspire_clkinfo_classic); node 178 drivers/clk/clk-palmas.c struct device_node *node = pdev->dev.of_node; node 182 drivers/clk/clk-palmas.c ret = of_property_read_u32(node, "ti,external-sleep-control", node 199 drivers/clk/clk-palmas.c node, prop); node 242 drivers/clk/clk-palmas.c struct device_node *node = pdev->dev.of_node; node 276 drivers/clk/clk-palmas.c ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &cinfo->hw); node 69 drivers/clk/clk-pwm.c struct device_node *node = pdev->dev.of_node; node 91 drivers/clk/clk-pwm.c if (of_property_read_u32(node, "clock-frequency", &clk_pwm->fixed_rate)) node 110 drivers/clk/clk-pwm.c clk_name = node->name; node 111 drivers/clk/clk-pwm.c of_property_read_string(node, "clock-output-names", &clk_name); node 124 drivers/clk/clk-pwm.c return of_clk_add_hw_provider(node, of_clk_hw_simple_get, &clk_pwm->hw); node 86 drivers/clk/clk-qoriq.c struct device_node *node; node 1003 drivers/clk/clk-qoriq.c if (!clockgen.node) node 1031 drivers/clk/clk-qoriq.c *sysclk_from_fixed(struct device_node *node, const char *name) node 1035 drivers/clk/clk-qoriq.c if (of_property_read_u32(node, "clock-frequency", &rate)) node 1061 drivers/clk/clk-qoriq.c clk = of_clk_get_by_name(clockgen.node, dtname); node 1072 drivers/clk/clk-qoriq.c clk = of_clk_get(clockgen.node, 0); node 1084 drivers/clk/clk-qoriq.c clk = sysclk_from_fixed(clockgen.node, name); node 1096 drivers/clk/clk-qoriq.c sysclk = of_get_child_by_name(clockgen.node, "sysclk"); node 1128 drivers/clk/clk-qoriq.c static void __init sysclk_init(struct device_node *node) node 1132 drivers/clk/clk-qoriq.c legacy_init_clockgen(node); node 1136 drivers/clk/clk-qoriq.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 1427 drivers/clk/clk-qoriq.c if (clockgen.node) node 1430 drivers/clk/clk-qoriq.c clockgen.node = np; node 157 drivers/clk/clk-rk808.c struct device_node *node = client->dev.of_node; node 176 drivers/clk/clk-rk808.c of_property_read_string_index(node, "clock-output-names", node 188 drivers/clk/clk-rk808.c of_property_read_string_index(node, "clock-output-names", node 223 drivers/clk/clk-vt8500.c static __init void vtwm_device_clk_init(struct device_node *node) node 228 drivers/clk/clk-vt8500.c const char *clk_name = node->name; node 243 drivers/clk/clk-vt8500.c rc = of_property_read_u32(node, "enable-reg", &en_reg); node 246 drivers/clk/clk-vt8500.c rc = of_property_read_u32(node, "enable-bit", &dev_clk->en_bit); node 255 drivers/clk/clk-vt8500.c rc = of_property_read_u32(node, "divisor-reg", &div_reg); node 264 drivers/clk/clk-vt8500.c of_property_read_u32(node, "divisor-mask", &dev_clk->div_mask); node 268 drivers/clk/clk-vt8500.c of_property_read_string(node, "clock-output-names", &clk_name); node 289 drivers/clk/clk-vt8500.c parent_name = of_clk_get_parent_name(node, 0); node 301 drivers/clk/clk-vt8500.c rc = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 673 drivers/clk/clk-vt8500.c static __init void vtwm_pll_clk_init(struct device_node *node, int pll_type) node 678 drivers/clk/clk-vt8500.c const char *clk_name = node->name; node 686 drivers/clk/clk-vt8500.c rc = of_property_read_u32(node, "reg", ®); node 698 drivers/clk/clk-vt8500.c of_property_read_string(node, "clock-output-names", &clk_name); node 703 drivers/clk/clk-vt8500.c parent_name = of_clk_get_parent_name(node, 0); node 715 drivers/clk/clk-vt8500.c rc = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 722 drivers/clk/clk-vt8500.c static void __init vt8500_pll_init(struct device_node *node) node 724 drivers/clk/clk-vt8500.c vtwm_pll_clk_init(node, PLL_TYPE_VT8500); node 728 drivers/clk/clk-vt8500.c static void __init wm8650_pll_init(struct device_node *node) node 730 drivers/clk/clk-vt8500.c vtwm_pll_clk_init(node, PLL_TYPE_WM8650); node 734 drivers/clk/clk-vt8500.c static void __init wm8750_pll_init(struct device_node *node) node 736 drivers/clk/clk-vt8500.c vtwm_pll_clk_init(node, PLL_TYPE_WM8750); node 740 drivers/clk/clk-vt8500.c static void __init wm8850_pll_init(struct device_node *node) node 742 drivers/clk/clk-vt8500.c vtwm_pll_clk_init(node, PLL_TYPE_WM8850); node 1482 drivers/clk/clk.c list_for_each_entry(cn, &clk_notifier_list, node) { node 3813 drivers/clk/clk.c int of_clk_hw_register(struct device_node *node, struct clk_hw *hw) node 3815 drivers/clk/clk.c return PTR_ERR_OR_ZERO(__clk_register(NULL, node, hw)); node 4152 drivers/clk/clk.c list_for_each_entry(cn, &clk_notifier_list, node) node 4165 drivers/clk/clk.c list_add(&cn->node, &clk_notifier_list); node 4200 drivers/clk/clk.c list_for_each_entry(cn, &clk_notifier_list, node) node 4212 drivers/clk/clk.c list_del(&cn->node); node 4245 drivers/clk/clk.c struct device_node *node; node 4320 drivers/clk/clk.c cp->node = of_node_get(np); node 4357 drivers/clk/clk.c cp->node = of_node_get(np); node 4450 drivers/clk/clk.c if (cp->node == np) { node 4452 drivers/clk/clk.c of_node_put(cp->node); node 4587 drivers/clk/clk.c if (provider->node == clkspec->np) { node 4766 drivers/clk/clk.c struct list_head node; node 4869 drivers/clk/clk.c &clk_provider_list, node) { node 4870 drivers/clk/clk.c list_del(&clk_provider->node); node 4880 drivers/clk/clk.c list_add_tail(&parent->node, &clk_provider_list); node 4886 drivers/clk/clk.c &clk_provider_list, node) { node 4896 drivers/clk/clk.c list_del(&clk_provider->node); node 48 drivers/clk/clkdev.c list_for_each_entry(p, &clocks, node) { node 124 drivers/clk/clkdev.c list_add_tail(&cl->node, &clocks); node 141 drivers/clk/clkdev.c list_add_tail(&cl->node, &clocks); node 290 drivers/clk/clkdev.c list_del(&cl->node); node 142 drivers/clk/davinci/pll-da850.c void of_da850_pll0_init(struct device_node *node) node 147 drivers/clk/davinci/pll-da850.c base = of_iomap(node, 0); node 155 drivers/clk/davinci/pll-da850.c of_davinci_pll_init(NULL, node, &da850_pll0_info, node 745 drivers/clk/davinci/pll.c int of_davinci_pll_init(struct device *dev, struct device_node *node, node 758 drivers/clk/davinci/pll.c parent_name = of_clk_get_parent_name(node, 0); node 768 drivers/clk/davinci/pll.c child = of_get_child_by_name(node, "pllout"); node 773 drivers/clk/davinci/pll.c child = of_get_child_by_name(node, "sysclk"); node 811 drivers/clk/davinci/pll.c child = of_get_child_by_name(node, "auxclk"); node 826 drivers/clk/davinci/pll.c child = of_get_child_by_name(node, "obsclk"); node 115 drivers/clk/davinci/pll.h int of_davinci_pll_init(struct device *dev, struct device_node *node, node 127 drivers/clk/davinci/pll.h void of_da850_pll0_init(struct device_node *node); node 483 drivers/clk/davinci/psc.c struct device_node *node = dev->of_node; node 490 drivers/clk/davinci/psc.c of_genpd_add_provider_onecell(node, &psc->pm_data); node 492 drivers/clk/davinci/psc.c of_clk_add_provider(node, of_clk_src_onecell_get, &psc->clk_data); node 16 drivers/clk/h8300/clk-div.c static void __init h8300_div_clk_setup(struct device_node *node) node 20 drivers/clk/h8300/clk-div.c const char *clk_name = node->name; node 26 drivers/clk/h8300/clk-div.c num_parents = of_clk_get_parent_count(node); node 32 drivers/clk/h8300/clk-div.c divcr = of_iomap(node, 0); node 41 drivers/clk/h8300/clk-div.c parent_name = of_clk_get_parent_name(node, 0); node 42 drivers/clk/h8300/clk-div.c of_property_read_u32(node, "renesas,width", &width); node 47 drivers/clk/h8300/clk-div.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw); node 86 drivers/clk/h8300/clk-h8s2678.c static void __init h8s2678_pll_clk_setup(struct device_node *node) node 89 drivers/clk/h8300/clk-h8s2678.c const char *clk_name = node->name; node 95 drivers/clk/h8300/clk-h8s2678.c num_parents = of_clk_get_parent_count(node); node 106 drivers/clk/h8300/clk-h8s2678.c pll_clock->sckcr = of_iomap(node, 0); node 112 drivers/clk/h8300/clk-h8s2678.c pll_clock->pllcr = of_iomap(node, 1); node 118 drivers/clk/h8300/clk-h8s2678.c parent_name = of_clk_get_parent_name(node, 0); node 133 drivers/clk/h8300/clk-h8s2678.c of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clock->hw); node 448 drivers/clk/hisilicon/clk-hi3620.c static void __init hi3620_mmc_clk_init(struct device_node *node) node 454 drivers/clk/hisilicon/clk-hi3620.c if (!node) { node 459 drivers/clk/hisilicon/clk-hi3620.c base = of_iomap(node, 0); node 476 drivers/clk/hisilicon/clk-hi3620.c hisi_register_clk_mmc(mmc_clk, base, node); node 480 drivers/clk/hisilicon/clk-hi3620.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 170 drivers/clk/imx/clk-imx6q.c static void of_assigned_ldb_sels(struct device_node *node, node 178 drivers/clk/imx/clk-imx6q.c num_parents = of_count_phandle_with_args(node, "assigned-clock-parents", node 181 drivers/clk/imx/clk-imx6q.c rc = of_parse_phandle_with_args(node, "assigned-clock-parents", node 190 drivers/clk/imx/clk-imx6q.c if (clkspec.np != node || clkspec.args[0] >= IMX6QDL_CLK_END) { node 196 drivers/clk/imx/clk-imx6q.c rc = of_parse_phandle_with_args(node, "assigned-clocks", node 200 drivers/clk/imx/clk-imx6q.c if (clkspec.np != node || clkspec.args[0] >= IMX6QDL_CLK_END) { node 224 drivers/clk/imx/clk-imx6q.c static bool pll6_bypassed(struct device_node *node) node 229 drivers/clk/imx/clk-imx6q.c num_clocks = of_count_phandle_with_args(node, "assigned-clocks", node 235 drivers/clk/imx/clk-imx6q.c ret = of_parse_phandle_with_args(node, "assigned-clocks", node 241 drivers/clk/imx/clk-imx6q.c if (clkspec.np == node && node 250 drivers/clk/imx/clk-imx6q.c ret = of_parse_phandle_with_args(node, "assigned-clock-parents", node 197 drivers/clk/keystone/gate.c static void __init of_psc_clk_init(struct device_node *node, spinlock_t *lock) node 199 drivers/clk/keystone/gate.c const char *clk_name = node->name; node 211 drivers/clk/keystone/gate.c i = of_property_match_string(node, "reg-names", "control"); node 212 drivers/clk/keystone/gate.c data->control_base = of_iomap(node, i); node 218 drivers/clk/keystone/gate.c i = of_property_match_string(node, "reg-names", "domain"); node 219 drivers/clk/keystone/gate.c data->domain_base = of_iomap(node, i); node 225 drivers/clk/keystone/gate.c of_property_read_u32(node, "domain-id", &data->domain_id); node 231 drivers/clk/keystone/gate.c of_property_read_string(node, "clock-output-names", &clk_name); node 232 drivers/clk/keystone/gate.c parent_name = of_clk_get_parent_name(node, 0); node 240 drivers/clk/keystone/gate.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 244 drivers/clk/keystone/gate.c pr_err("%s: error registering clk %pOFn\n", __func__, node); node 259 drivers/clk/keystone/gate.c static void __init of_keystone_psc_clk_init(struct device_node *node) node 261 drivers/clk/keystone/gate.c of_psc_clk_init(node, &psc_lock); node 158 drivers/clk/keystone/pll.c static void __init _of_pll_clk_init(struct device_node *node, bool pllctrl) node 171 drivers/clk/keystone/pll.c parent_name = of_clk_get_parent_name(node, 0); node 172 drivers/clk/keystone/pll.c if (of_property_read_u32(node, "fixed-postdiv", &pll_data->postdiv)) { node 181 drivers/clk/keystone/pll.c i = of_property_match_string(node, "reg-names", node 183 drivers/clk/keystone/pll.c pll_data->pllod = of_iomap(node, i); node 186 drivers/clk/keystone/pll.c i = of_property_match_string(node, "reg-names", "control"); node 187 drivers/clk/keystone/pll.c pll_data->pll_ctl0 = of_iomap(node, i); node 202 drivers/clk/keystone/pll.c i = of_property_match_string(node, "reg-names", "multiplier"); node 203 drivers/clk/keystone/pll.c pll_data->pllm = of_iomap(node, i); node 211 drivers/clk/keystone/pll.c clk = clk_register_pll(NULL, node->name, parent_name, pll_data); node 213 drivers/clk/keystone/pll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 218 drivers/clk/keystone/pll.c pr_err("%s: error initializing pll %pOFn\n", __func__, node); node 226 drivers/clk/keystone/pll.c static void __init of_keystone_pll_clk_init(struct device_node *node) node 228 drivers/clk/keystone/pll.c _of_pll_clk_init(node, false); node 237 drivers/clk/keystone/pll.c static void __init of_keystone_main_pll_clk_init(struct device_node *node) node 239 drivers/clk/keystone/pll.c _of_pll_clk_init(node, true); node 248 drivers/clk/keystone/pll.c static void __init of_pll_div_clk_init(struct device_node *node) node 254 drivers/clk/keystone/pll.c const char *clk_name = node->name; node 256 drivers/clk/keystone/pll.c of_property_read_string(node, "clock-output-names", &clk_name); node 257 drivers/clk/keystone/pll.c reg = of_iomap(node, 0); node 263 drivers/clk/keystone/pll.c parent_name = of_clk_get_parent_name(node, 0); node 270 drivers/clk/keystone/pll.c if (of_property_read_u32(node, "bit-shift", &shift)) { node 276 drivers/clk/keystone/pll.c if (of_property_read_u32(node, "bit-mask", &mask)) { node 285 drivers/clk/keystone/pll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 297 drivers/clk/keystone/pll.c static void __init of_pll_mux_clk_init(struct device_node *node) node 303 drivers/clk/keystone/pll.c const char *clk_name = node->name; node 305 drivers/clk/keystone/pll.c of_property_read_string(node, "clock-output-names", &clk_name); node 306 drivers/clk/keystone/pll.c reg = of_iomap(node, 0); node 312 drivers/clk/keystone/pll.c of_clk_parent_fill(node, parents, 2); node 318 drivers/clk/keystone/pll.c if (of_property_read_u32(node, "bit-shift", &shift)) { node 323 drivers/clk/keystone/pll.c if (of_property_read_u32(node, "bit-mask", &mask)) { node 332 drivers/clk/keystone/pll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 65 drivers/clk/keystone/sci-clk.c struct list_head node; node 494 drivers/clk/keystone/sci-clk.c struct sci_clk *ca = container_of(a, struct sci_clk, node); node 495 drivers/clk/keystone/sci-clk.c struct sci_clk *cb = container_of(b, struct sci_clk, node); node 553 drivers/clk/keystone/sci-clk.c list_add_tail(&sci_clk->node, &clks); node 589 drivers/clk/keystone/sci-clk.c list_add_tail(&sci_clk->node, &clks); node 609 drivers/clk/keystone/sci-clk.c list_for_each_entry(sci_clk, &clks, node) { node 79 drivers/clk/mediatek/clk-cpumux.c int mtk_clk_register_cpumuxes(struct device_node *node, node 87 drivers/clk/mediatek/clk-cpumux.c regmap = syscon_node_to_regmap(node); node 89 drivers/clk/mediatek/clk-cpumux.c pr_err("Cannot find regmap for %pOF: %ld\n", node, node 18 drivers/clk/mediatek/clk-cpumux.h int mtk_clk_register_cpumuxes(struct device_node *node, node 149 drivers/clk/mediatek/clk-mt2701-aud.c struct device_node *node = pdev->dev.of_node; node 154 drivers/clk/mediatek/clk-mt2701-aud.c mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks), node 157 drivers/clk/mediatek/clk-mt2701-aud.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 173 drivers/clk/mediatek/clk-mt2701-aud.c of_clk_del_provider(node); node 106 drivers/clk/mediatek/clk-mt2701-bdp.c struct device_node *node = pdev->dev.of_node; node 110 drivers/clk/mediatek/clk-mt2701-bdp.c mtk_clk_register_gates(node, bdp_clks, ARRAY_SIZE(bdp_clks), node 113 drivers/clk/mediatek/clk-mt2701-bdp.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 48 drivers/clk/mediatek/clk-mt2701-eth.c struct device_node *node = pdev->dev.of_node; node 52 drivers/clk/mediatek/clk-mt2701-eth.c mtk_clk_register_gates(node, eth_clks, ARRAY_SIZE(eth_clks), node 55 drivers/clk/mediatek/clk-mt2701-eth.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 61 drivers/clk/mediatek/clk-mt2701-eth.c mtk_register_reset_controller(node, 1, 0x34); node 41 drivers/clk/mediatek/clk-mt2701-g3d.c struct device_node *node = pdev->dev.of_node; node 46 drivers/clk/mediatek/clk-mt2701-g3d.c mtk_clk_register_gates(node, g3d_clks, ARRAY_SIZE(g3d_clks), node 49 drivers/clk/mediatek/clk-mt2701-g3d.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 55 drivers/clk/mediatek/clk-mt2701-g3d.c mtk_register_reset_controller(node, 1, 0xc); node 45 drivers/clk/mediatek/clk-mt2701-hif.c struct device_node *node = pdev->dev.of_node; node 49 drivers/clk/mediatek/clk-mt2701-hif.c mtk_clk_register_gates(node, hif_clks, ARRAY_SIZE(hif_clks), node 52 drivers/clk/mediatek/clk-mt2701-hif.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 60 drivers/clk/mediatek/clk-mt2701-hif.c mtk_register_reset_controller(node, 1, 0x34); node 48 drivers/clk/mediatek/clk-mt2701-img.c struct device_node *node = pdev->dev.of_node; node 52 drivers/clk/mediatek/clk-mt2701-img.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 55 drivers/clk/mediatek/clk-mt2701-img.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 91 drivers/clk/mediatek/clk-mt2701-mm.c struct device_node *node = pdev->dev.of_node; node 95 drivers/clk/mediatek/clk-mt2701-mm.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 98 drivers/clk/mediatek/clk-mt2701-mm.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 59 drivers/clk/mediatek/clk-mt2701-vdec.c struct device_node *node = pdev->dev.of_node; node 63 drivers/clk/mediatek/clk-mt2701-vdec.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 66 drivers/clk/mediatek/clk-mt2701-vdec.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 670 drivers/clk/mediatek/clk-mt2701.c struct device_node *node = pdev->dev.of_node; node 691 drivers/clk/mediatek/clk-mt2701.c mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), node 694 drivers/clk/mediatek/clk-mt2701.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 739 drivers/clk/mediatek/clk-mt2701.c static void __init mtk_infrasys_init_early(struct device_node *node) node 753 drivers/clk/mediatek/clk-mt2701.c mtk_clk_register_cpumuxes(node, cpu_muxes, ARRAY_SIZE(cpu_muxes), node 756 drivers/clk/mediatek/clk-mt2701.c r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data); node 767 drivers/clk/mediatek/clk-mt2701.c struct device_node *node = pdev->dev.of_node; node 778 drivers/clk/mediatek/clk-mt2701.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 783 drivers/clk/mediatek/clk-mt2701.c r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data); node 787 drivers/clk/mediatek/clk-mt2701.c mtk_register_reset_controller(node, 2, 0x30); node 891 drivers/clk/mediatek/clk-mt2701.c struct device_node *node = pdev->dev.of_node; node 900 drivers/clk/mediatek/clk-mt2701.c mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks), node 906 drivers/clk/mediatek/clk-mt2701.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 910 drivers/clk/mediatek/clk-mt2701.c mtk_register_reset_controller(node, 2, 0x0); node 972 drivers/clk/mediatek/clk-mt2701.c struct device_node *node = pdev->dev.of_node; node 978 drivers/clk/mediatek/clk-mt2701.c mtk_clk_register_plls(node, apmixed_plls, ARRAY_SIZE(apmixed_plls), node 983 drivers/clk/mediatek/clk-mt2701.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 65 drivers/clk/mediatek/clk-mt2712-bdp.c struct device_node *node = pdev->dev.of_node; node 69 drivers/clk/mediatek/clk-mt2712-bdp.c mtk_clk_register_gates(node, bdp_clks, ARRAY_SIZE(bdp_clks), node 72 drivers/clk/mediatek/clk-mt2712-bdp.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 43 drivers/clk/mediatek/clk-mt2712-img.c struct device_node *node = pdev->dev.of_node; node 47 drivers/clk/mediatek/clk-mt2712-img.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 50 drivers/clk/mediatek/clk-mt2712-img.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 39 drivers/clk/mediatek/clk-mt2712-jpgdec.c struct device_node *node = pdev->dev.of_node; node 43 drivers/clk/mediatek/clk-mt2712-jpgdec.c mtk_clk_register_gates(node, jpgdec_clks, ARRAY_SIZE(jpgdec_clks), node 46 drivers/clk/mediatek/clk-mt2712-jpgdec.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 38 drivers/clk/mediatek/clk-mt2712-mfg.c struct device_node *node = pdev->dev.of_node; node 42 drivers/clk/mediatek/clk-mt2712-mfg.c mtk_clk_register_gates(node, mfg_clks, ARRAY_SIZE(mfg_clks), node 45 drivers/clk/mediatek/clk-mt2712-mfg.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 133 drivers/clk/mediatek/clk-mt2712-mm.c struct device_node *node = pdev->dev.of_node; node 137 drivers/clk/mediatek/clk-mt2712-mm.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 140 drivers/clk/mediatek/clk-mt2712-mm.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 57 drivers/clk/mediatek/clk-mt2712-vdec.c struct device_node *node = pdev->dev.of_node; node 61 drivers/clk/mediatek/clk-mt2712-vdec.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 64 drivers/clk/mediatek/clk-mt2712-vdec.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 40 drivers/clk/mediatek/clk-mt2712-venc.c struct device_node *node = pdev->dev.of_node; node 44 drivers/clk/mediatek/clk-mt2712-venc.c mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks), node 47 drivers/clk/mediatek/clk-mt2712-venc.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1264 drivers/clk/mediatek/clk-mt2712.c struct device_node *node = pdev->dev.of_node; node 1268 drivers/clk/mediatek/clk-mt2712.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 1270 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1281 drivers/clk/mediatek/clk-mt2712.c static void clk_mt2712_top_init_early(struct device_node *node) node 1295 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data); node 1307 drivers/clk/mediatek/clk-mt2712.c struct device_node *node = pdev->dev.of_node; node 1335 drivers/clk/mediatek/clk-mt2712.c mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), node 1338 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data); node 1351 drivers/clk/mediatek/clk-mt2712.c struct device_node *node = pdev->dev.of_node; node 1355 drivers/clk/mediatek/clk-mt2712.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 1358 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1364 drivers/clk/mediatek/clk-mt2712.c mtk_register_reset_controller(node, 2, 0x30); node 1373 drivers/clk/mediatek/clk-mt2712.c struct device_node *node = pdev->dev.of_node; node 1377 drivers/clk/mediatek/clk-mt2712.c mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks), node 1380 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1386 drivers/clk/mediatek/clk-mt2712.c mtk_register_reset_controller(node, 2, 0); node 1395 drivers/clk/mediatek/clk-mt2712.c struct device_node *node = pdev->dev.of_node; node 1410 drivers/clk/mediatek/clk-mt2712.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 99 drivers/clk/mediatek/clk-mt6779-aud.c struct device_node *node = pdev->dev.of_node; node 103 drivers/clk/mediatek/clk-mt6779-aud.c mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks), node 106 drivers/clk/mediatek/clk-mt6779-aud.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 48 drivers/clk/mediatek/clk-mt6779-cam.c struct device_node *node = pdev->dev.of_node; node 52 drivers/clk/mediatek/clk-mt6779-cam.c mtk_clk_register_gates(node, cam_clks, ARRAY_SIZE(cam_clks), node 55 drivers/clk/mediatek/clk-mt6779-cam.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 40 drivers/clk/mediatek/clk-mt6779-img.c struct device_node *node = pdev->dev.of_node; node 44 drivers/clk/mediatek/clk-mt6779-img.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 47 drivers/clk/mediatek/clk-mt6779-img.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 42 drivers/clk/mediatek/clk-mt6779-ipe.c struct device_node *node = pdev->dev.of_node; node 46 drivers/clk/mediatek/clk-mt6779-ipe.c mtk_clk_register_gates(node, ipe_clks, ARRAY_SIZE(ipe_clks), node 49 drivers/clk/mediatek/clk-mt6779-ipe.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 32 drivers/clk/mediatek/clk-mt6779-mfg.c struct device_node *node = pdev->dev.of_node; node 36 drivers/clk/mediatek/clk-mt6779-mfg.c mtk_clk_register_gates(node, mfg_clks, ARRAY_SIZE(mfg_clks), node 39 drivers/clk/mediatek/clk-mt6779-mfg.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 95 drivers/clk/mediatek/clk-mt6779-mm.c struct device_node *node = pdev->dev.of_node; node 99 drivers/clk/mediatek/clk-mt6779-mm.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 102 drivers/clk/mediatek/clk-mt6779-mm.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 49 drivers/clk/mediatek/clk-mt6779-vdec.c struct device_node *node = pdev->dev.of_node; node 53 drivers/clk/mediatek/clk-mt6779-vdec.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 56 drivers/clk/mediatek/clk-mt6779-vdec.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 40 drivers/clk/mediatek/clk-mt6779-venc.c struct device_node *node = pdev->dev.of_node; node 44 drivers/clk/mediatek/clk-mt6779-venc.c mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks), node 47 drivers/clk/mediatek/clk-mt6779-venc.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1214 drivers/clk/mediatek/clk-mt6779.c struct device_node *node = pdev->dev.of_node; node 1218 drivers/clk/mediatek/clk-mt6779.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 1220 drivers/clk/mediatek/clk-mt6779.c mtk_clk_register_gates(node, apmixed_clks, node 1223 drivers/clk/mediatek/clk-mt6779.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1231 drivers/clk/mediatek/clk-mt6779.c struct device_node *node = pdev->dev.of_node; node 1245 drivers/clk/mediatek/clk-mt6779.c node, &mt6779_clk_lock, clk_data); node 1253 drivers/clk/mediatek/clk-mt6779.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1259 drivers/clk/mediatek/clk-mt6779.c struct device_node *node = pdev->dev.of_node; node 1263 drivers/clk/mediatek/clk-mt6779.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 1266 drivers/clk/mediatek/clk-mt6779.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 44 drivers/clk/mediatek/clk-mt6797-img.c struct device_node *node = pdev->dev.of_node; node 48 drivers/clk/mediatek/clk-mt6797-img.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 51 drivers/clk/mediatek/clk-mt6797-img.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 104 drivers/clk/mediatek/clk-mt6797-mm.c struct device_node *node = pdev->dev.of_node; node 108 drivers/clk/mediatek/clk-mt6797-mm.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 111 drivers/clk/mediatek/clk-mt6797-mm.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 61 drivers/clk/mediatek/clk-mt6797-vdec.c struct device_node *node = pdev->dev.of_node; node 65 drivers/clk/mediatek/clk-mt6797-vdec.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 68 drivers/clk/mediatek/clk-mt6797-vdec.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 46 drivers/clk/mediatek/clk-mt6797-venc.c struct device_node *node = pdev->dev.of_node; node 50 drivers/clk/mediatek/clk-mt6797-venc.c mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks), node 53 drivers/clk/mediatek/clk-mt6797-venc.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 387 drivers/clk/mediatek/clk-mt6797.c struct device_node *node = pdev->dev.of_node; node 402 drivers/clk/mediatek/clk-mt6797.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 561 drivers/clk/mediatek/clk-mt6797.c static void mtk_infrasys_init_early(struct device_node *node) node 575 drivers/clk/mediatek/clk-mt6797.c r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data); node 587 drivers/clk/mediatek/clk-mt6797.c struct device_node *node = pdev->dev.of_node; node 598 drivers/clk/mediatek/clk-mt6797.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 603 drivers/clk/mediatek/clk-mt6797.c r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data); node 667 drivers/clk/mediatek/clk-mt6797.c struct device_node *node = pdev->dev.of_node; node 673 drivers/clk/mediatek/clk-mt6797.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 675 drivers/clk/mediatek/clk-mt6797.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 136 drivers/clk/mediatek/clk-mt7622-aud.c struct device_node *node = pdev->dev.of_node; node 141 drivers/clk/mediatek/clk-mt7622-aud.c mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks), node 144 drivers/clk/mediatek/clk-mt7622-aud.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 160 drivers/clk/mediatek/clk-mt7622-aud.c of_clk_del_provider(node); node 71 drivers/clk/mediatek/clk-mt7622-eth.c struct device_node *node = pdev->dev.of_node; node 76 drivers/clk/mediatek/clk-mt7622-eth.c mtk_clk_register_gates(node, eth_clks, ARRAY_SIZE(eth_clks), node 79 drivers/clk/mediatek/clk-mt7622-eth.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 85 drivers/clk/mediatek/clk-mt7622-eth.c mtk_register_reset_controller(node, 1, 0x34); node 93 drivers/clk/mediatek/clk-mt7622-eth.c struct device_node *node = pdev->dev.of_node; node 98 drivers/clk/mediatek/clk-mt7622-eth.c mtk_clk_register_gates(node, sgmii_clks, ARRAY_SIZE(sgmii_clks), node 101 drivers/clk/mediatek/clk-mt7622-eth.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 82 drivers/clk/mediatek/clk-mt7622-hif.c struct device_node *node = pdev->dev.of_node; node 87 drivers/clk/mediatek/clk-mt7622-hif.c mtk_clk_register_gates(node, ssusb_clks, ARRAY_SIZE(ssusb_clks), node 90 drivers/clk/mediatek/clk-mt7622-hif.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 96 drivers/clk/mediatek/clk-mt7622-hif.c mtk_register_reset_controller(node, 1, 0x34); node 104 drivers/clk/mediatek/clk-mt7622-hif.c struct device_node *node = pdev->dev.of_node; node 109 drivers/clk/mediatek/clk-mt7622-hif.c mtk_clk_register_gates(node, pcie_clks, ARRAY_SIZE(pcie_clks), node 112 drivers/clk/mediatek/clk-mt7622-hif.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 118 drivers/clk/mediatek/clk-mt7622-hif.c mtk_register_reset_controller(node, 1, 0x34); node 616 drivers/clk/mediatek/clk-mt7622.c struct device_node *node = pdev->dev.of_node; node 637 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), node 644 drivers/clk/mediatek/clk-mt7622.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 649 drivers/clk/mediatek/clk-mt7622.c struct device_node *node = pdev->dev.of_node; node 655 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 658 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_cpumuxes(node, infra_muxes, ARRAY_SIZE(infra_muxes), node 661 drivers/clk/mediatek/clk-mt7622.c r = of_clk_add_provider(node, of_clk_src_onecell_get, node 666 drivers/clk/mediatek/clk-mt7622.c mtk_register_reset_controller(node, 1, 0x30); node 674 drivers/clk/mediatek/clk-mt7622.c struct device_node *node = pdev->dev.of_node; node 680 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), node 683 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_gates(node, apmixed_clks, node 689 drivers/clk/mediatek/clk-mt7622.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 697 drivers/clk/mediatek/clk-mt7622.c struct device_node *node = pdev->dev.of_node; node 706 drivers/clk/mediatek/clk-mt7622.c mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks), node 712 drivers/clk/mediatek/clk-mt7622.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 718 drivers/clk/mediatek/clk-mt7622.c mtk_register_reset_controller(node, 2, 0x0); node 82 drivers/clk/mediatek/clk-mt7629-eth.c struct device_node *node = pdev->dev.of_node; node 87 drivers/clk/mediatek/clk-mt7629-eth.c mtk_clk_register_gates(node, eth_clks, CLK_ETH_NR_CLK, clk_data); node 89 drivers/clk/mediatek/clk-mt7629-eth.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 95 drivers/clk/mediatek/clk-mt7629-eth.c mtk_register_reset_controller(node, 1, 0x34); node 103 drivers/clk/mediatek/clk-mt7629-eth.c struct device_node *node = pdev->dev.of_node; node 109 drivers/clk/mediatek/clk-mt7629-eth.c mtk_clk_register_gates(node, sgmii_clks[id++], CLK_SGMII_NR_CLK, node 112 drivers/clk/mediatek/clk-mt7629-eth.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 77 drivers/clk/mediatek/clk-mt7629-hif.c struct device_node *node = pdev->dev.of_node; node 82 drivers/clk/mediatek/clk-mt7629-hif.c mtk_clk_register_gates(node, ssusb_clks, ARRAY_SIZE(ssusb_clks), node 85 drivers/clk/mediatek/clk-mt7629-hif.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 91 drivers/clk/mediatek/clk-mt7629-hif.c mtk_register_reset_controller(node, 1, 0x34); node 99 drivers/clk/mediatek/clk-mt7629-hif.c struct device_node *node = pdev->dev.of_node; node 104 drivers/clk/mediatek/clk-mt7629-hif.c mtk_clk_register_gates(node, pcie_clks, ARRAY_SIZE(pcie_clks), node 107 drivers/clk/mediatek/clk-mt7629-hif.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 113 drivers/clk/mediatek/clk-mt7629-hif.c mtk_register_reset_controller(node, 1, 0x34); node 576 drivers/clk/mediatek/clk-mt7629.c struct device_node *node = pdev->dev.of_node; node 598 drivers/clk/mediatek/clk-mt7629.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 603 drivers/clk/mediatek/clk-mt7629.c struct device_node *node = pdev->dev.of_node; node 609 drivers/clk/mediatek/clk-mt7629.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 612 drivers/clk/mediatek/clk-mt7629.c mtk_clk_register_cpumuxes(node, infra_muxes, ARRAY_SIZE(infra_muxes), node 615 drivers/clk/mediatek/clk-mt7629.c r = of_clk_add_provider(node, of_clk_src_onecell_get, node 628 drivers/clk/mediatek/clk-mt7629.c struct device_node *node = pdev->dev.of_node; node 637 drivers/clk/mediatek/clk-mt7629.c mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks), node 643 drivers/clk/mediatek/clk-mt7629.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 655 drivers/clk/mediatek/clk-mt7629.c struct device_node *node = pdev->dev.of_node; node 661 drivers/clk/mediatek/clk-mt7629.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), node 664 drivers/clk/mediatek/clk-mt7629.c mtk_clk_register_gates(node, apmixed_clks, node 670 drivers/clk/mediatek/clk-mt7629.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 516 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_topckgen_init(struct device_node *node) node 522 drivers/clk/mediatek/clk-mt8135.c base = of_iomap(node, 0); node 537 drivers/clk/mediatek/clk-mt8135.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 544 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_infrasys_init(struct device_node *node) node 551 drivers/clk/mediatek/clk-mt8135.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 556 drivers/clk/mediatek/clk-mt8135.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 561 drivers/clk/mediatek/clk-mt8135.c mtk_register_reset_controller(node, 2, 0x30); node 565 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_pericfg_init(struct device_node *node) node 571 drivers/clk/mediatek/clk-mt8135.c base = of_iomap(node, 0); node 579 drivers/clk/mediatek/clk-mt8135.c mtk_clk_register_gates(node, peri_gates, ARRAY_SIZE(peri_gates), node 584 drivers/clk/mediatek/clk-mt8135.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 589 drivers/clk/mediatek/clk-mt8135.c mtk_register_reset_controller(node, 2, 0); node 626 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_apmixedsys_init(struct device_node *node) node 634 drivers/clk/mediatek/clk-mt8135.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 924 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_topckgen_init(struct device_node *node) node 930 drivers/clk/mediatek/clk-mt8173.c base = of_iomap(node, 0); node 943 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 952 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_infrasys_init(struct device_node *node) node 959 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 963 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_cpumuxes(node, cpu_muxes, ARRAY_SIZE(cpu_muxes), node 966 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 971 drivers/clk/mediatek/clk-mt8173.c mtk_register_reset_controller(node, 2, 0x30); node 975 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_pericfg_init(struct device_node *node) node 981 drivers/clk/mediatek/clk-mt8173.c base = of_iomap(node, 0); node 989 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, peri_gates, ARRAY_SIZE(peri_gates), node 994 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 999 drivers/clk/mediatek/clk-mt8173.c mtk_register_reset_controller(node, 2, 0); node 1078 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_apmixedsys_init(struct device_node *node) node 1085 drivers/clk/mediatek/clk-mt8173.c base = of_iomap(node, 0); node 1097 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 1119 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1129 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_imgsys_init(struct device_node *node) node 1136 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 1139 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1147 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_mmsys_init(struct device_node *node) node 1154 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 1157 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1164 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vdecsys_init(struct device_node *node) node 1171 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 1174 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1181 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vencsys_init(struct device_node *node) node 1188 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks), node 1191 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1198 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vencltsys_init(struct device_node *node) node 1205 drivers/clk/mediatek/clk-mt8173.c mtk_clk_register_gates(node, venclt_clks, ARRAY_SIZE(venclt_clks), node 1208 drivers/clk/mediatek/clk-mt8173.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 74 drivers/clk/mediatek/clk-mt8183-audio.c struct device_node *node = pdev->dev.of_node; node 78 drivers/clk/mediatek/clk-mt8183-audio.c mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks), node 81 drivers/clk/mediatek/clk-mt8183-audio.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 87 drivers/clk/mediatek/clk-mt8183-audio.c of_clk_del_provider(node); node 40 drivers/clk/mediatek/clk-mt8183-cam.c struct device_node *node = pdev->dev.of_node; node 44 drivers/clk/mediatek/clk-mt8183-cam.c mtk_clk_register_gates(node, cam_clks, ARRAY_SIZE(cam_clks), node 47 drivers/clk/mediatek/clk-mt8183-cam.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 40 drivers/clk/mediatek/clk-mt8183-img.c struct device_node *node = pdev->dev.of_node; node 44 drivers/clk/mediatek/clk-mt8183-img.c mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks), node 47 drivers/clk/mediatek/clk-mt8183-img.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 33 drivers/clk/mediatek/clk-mt8183-ipu0.c struct device_node *node = pdev->dev.of_node; node 37 drivers/clk/mediatek/clk-mt8183-ipu0.c mtk_clk_register_gates(node, ipu_core0_clks, ARRAY_SIZE(ipu_core0_clks), node 40 drivers/clk/mediatek/clk-mt8183-ipu0.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 33 drivers/clk/mediatek/clk-mt8183-ipu1.c struct device_node *node = pdev->dev.of_node; node 37 drivers/clk/mediatek/clk-mt8183-ipu1.c mtk_clk_register_gates(node, ipu_core1_clks, ARRAY_SIZE(ipu_core1_clks), node 40 drivers/clk/mediatek/clk-mt8183-ipu1.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 31 drivers/clk/mediatek/clk-mt8183-ipu_adl.c struct device_node *node = pdev->dev.of_node; node 35 drivers/clk/mediatek/clk-mt8183-ipu_adl.c mtk_clk_register_gates(node, ipu_adl_clks, ARRAY_SIZE(ipu_adl_clks), node 38 drivers/clk/mediatek/clk-mt8183-ipu_adl.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 100 drivers/clk/mediatek/clk-mt8183-ipu_conn.c struct device_node *node = pdev->dev.of_node; node 104 drivers/clk/mediatek/clk-mt8183-ipu_conn.c mtk_clk_register_gates(node, ipu_conn_clks, ARRAY_SIZE(ipu_conn_clks), node 107 drivers/clk/mediatek/clk-mt8183-ipu_conn.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 32 drivers/clk/mediatek/clk-mt8183-mfgcfg.c struct device_node *node = pdev->dev.of_node; node 38 drivers/clk/mediatek/clk-mt8183-mfgcfg.c mtk_clk_register_gates_with_dev(node, mfg_clks, ARRAY_SIZE(mfg_clks), node 41 drivers/clk/mediatek/clk-mt8183-mfgcfg.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 88 drivers/clk/mediatek/clk-mt8183-mm.c struct device_node *node = pdev->dev.of_node; node 92 drivers/clk/mediatek/clk-mt8183-mm.c mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks), node 95 drivers/clk/mediatek/clk-mt8183-mm.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 44 drivers/clk/mediatek/clk-mt8183-vdec.c struct device_node *node = pdev->dev.of_node; node 48 drivers/clk/mediatek/clk-mt8183-vdec.c mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks), node 51 drivers/clk/mediatek/clk-mt8183-vdec.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 36 drivers/clk/mediatek/clk-mt8183-venc.c struct device_node *node = pdev->dev.of_node; node 40 drivers/clk/mediatek/clk-mt8183-venc.c mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks), node 43 drivers/clk/mediatek/clk-mt8183-venc.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1158 drivers/clk/mediatek/clk-mt8183.c struct device_node *node = pdev->dev.of_node; node 1162 drivers/clk/mediatek/clk-mt8183.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 1164 drivers/clk/mediatek/clk-mt8183.c mtk_clk_register_gates(node, apmixed_clks, ARRAY_SIZE(apmixed_clks), node 1167 drivers/clk/mediatek/clk-mt8183.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1172 drivers/clk/mediatek/clk-mt8183.c static void clk_mt8183_top_init_early(struct device_node *node) node 1184 drivers/clk/mediatek/clk-mt8183.c of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data); node 1194 drivers/clk/mediatek/clk-mt8183.c struct device_node *node = pdev->dev.of_node; node 1209 drivers/clk/mediatek/clk-mt8183.c node, &mt8183_clk_lock, top_clk_data); node 1217 drivers/clk/mediatek/clk-mt8183.c mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), node 1220 drivers/clk/mediatek/clk-mt8183.c return of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data); node 1226 drivers/clk/mediatek/clk-mt8183.c struct device_node *node = pdev->dev.of_node; node 1231 drivers/clk/mediatek/clk-mt8183.c mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks), node 1234 drivers/clk/mediatek/clk-mt8183.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1242 drivers/clk/mediatek/clk-mt8183.c mtk_register_reset_controller_set_clr(node, 4, INFRA_RST0_SET_OFFSET); node 1250 drivers/clk/mediatek/clk-mt8183.c struct device_node *node = pdev->dev.of_node; node 1254 drivers/clk/mediatek/clk-mt8183.c mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks), node 1257 drivers/clk/mediatek/clk-mt8183.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 1263 drivers/clk/mediatek/clk-mt8183.c struct device_node *node = pdev->dev.of_node; node 1276 drivers/clk/mediatek/clk-mt8183.c return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 50 drivers/clk/mediatek/clk-mt8516-aud.c static void __init mtk_audsys_init(struct device_node *node) node 57 drivers/clk/mediatek/clk-mt8516-aud.c mtk_clk_register_gates(node, aud_clks, ARRAY_SIZE(aud_clks), clk_data); node 59 drivers/clk/mediatek/clk-mt8516-aud.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 677 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_topckgen_init(struct device_node *node) node 683 drivers/clk/mediatek/clk-mt8516.c base = of_iomap(node, 0); node 693 drivers/clk/mediatek/clk-mt8516.c mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), clk_data); node 701 drivers/clk/mediatek/clk-mt8516.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 708 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_infracfg_init(struct device_node *node) node 714 drivers/clk/mediatek/clk-mt8516.c base = of_iomap(node, 0); node 725 drivers/clk/mediatek/clk-mt8516.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 787 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_apmixedsys_init(struct device_node *node) node 793 drivers/clk/mediatek/clk-mt8516.c base = of_iomap(node, 0); node 801 drivers/clk/mediatek/clk-mt8516.c mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data); node 803 drivers/clk/mediatek/clk-mt8516.c r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 97 drivers/clk/mediatek/clk-mtk.c int mtk_clk_register_gates_with_dev(struct device_node *node, node 109 drivers/clk/mediatek/clk-mtk.c regmap = syscon_node_to_regmap(node); node 111 drivers/clk/mediatek/clk-mtk.c pr_err("Cannot find regmap for %pOF: %ld\n", node, node 141 drivers/clk/mediatek/clk-mtk.c int mtk_clk_register_gates(struct device_node *node, node 145 drivers/clk/mediatek/clk-mtk.c return mtk_clk_register_gates_with_dev(node, node 168 drivers/clk/mediatek/clk-mtk.h int mtk_clk_register_gates(struct device_node *node, node 172 drivers/clk/mediatek/clk-mtk.h int mtk_clk_register_gates_with_dev(struct device_node *node, node 238 drivers/clk/mediatek/clk-mtk.h void mtk_clk_register_plls(struct device_node *node, node 191 drivers/clk/mediatek/clk-mux.c int num, struct device_node *node, node 199 drivers/clk/mediatek/clk-mux.c regmap = syscon_node_to_regmap(node); node 201 drivers/clk/mediatek/clk-mux.c pr_err("Cannot find regmap for %pOF: %ld\n", node, node 85 drivers/clk/mediatek/clk-mux.h int num, struct device_node *node, node 344 drivers/clk/mediatek/clk-pll.c void mtk_clk_register_plls(struct device_node *node, node 351 drivers/clk/mediatek/clk-pll.c base = of_iomap(node, 0); node 251 drivers/clk/mvebu/clk-corediv.c mvebu_corediv_clk_init(struct device_node *node, node 262 drivers/clk/mvebu/clk-corediv.c base = of_iomap(node, 0); node 266 drivers/clk/mvebu/clk-corediv.c parent_name = of_clk_get_parent_name(node, 0); node 284 drivers/clk/mvebu/clk-corediv.c of_property_read_string_index(node, "clock-output-names", node 302 drivers/clk/mvebu/clk-corediv.c of_clk_add_provider(node, of_clk_src_onecell_get, &clk_data); node 311 drivers/clk/mvebu/clk-corediv.c static void __init armada370_corediv_clk_init(struct device_node *node) node 313 drivers/clk/mvebu/clk-corediv.c return mvebu_corediv_clk_init(node, &armada370_corediv_soc); node 318 drivers/clk/mvebu/clk-corediv.c static void __init armada375_corediv_clk_init(struct device_node *node) node 320 drivers/clk/mvebu/clk-corediv.c return mvebu_corediv_clk_init(node, &armada375_corediv_soc); node 325 drivers/clk/mvebu/clk-corediv.c static void __init armada380_corediv_clk_init(struct device_node *node) node 327 drivers/clk/mvebu/clk-corediv.c return mvebu_corediv_clk_init(node, &armada380_corediv_soc); node 332 drivers/clk/mvebu/clk-corediv.c static void __init mv98dx3236_corediv_clk_init(struct device_node *node) node 334 drivers/clk/mvebu/clk-corediv.c return mvebu_corediv_clk_init(node, &mv98dx3236_corediv_soc); node 166 drivers/clk/mvebu/clk-cpu.c static void __init of_cpu_clk_setup(struct device_node *node) node 169 drivers/clk/mvebu/clk-cpu.c void __iomem *clock_complex_base = of_iomap(node, 0); node 170 drivers/clk/mvebu/clk-cpu.c void __iomem *pmu_dfs_base = of_iomap(node, 1); node 210 drivers/clk/mvebu/clk-cpu.c cpuclk[cpu].parent_name = of_clk_get_parent_name(node, 0); node 231 drivers/clk/mvebu/clk-cpu.c of_clk_add_provider(node, of_clk_src_onecell_get, &clk_data); node 247 drivers/clk/mvebu/clk-cpu.c static void __init of_mv98dx3236_cpu_clk_setup(struct device_node *node) node 249 drivers/clk/mvebu/clk-cpu.c of_clk_add_provider(node, of_clk_src_simple_get, NULL); node 16 drivers/clk/pistachio/clk.c pistachio_clk_alloc_provider(struct device_node *node, unsigned int num_clks) node 28 drivers/clk/pistachio/clk.c p->node = node; node 29 drivers/clk/pistachio/clk.c p->base = of_iomap(node, 0); node 54 drivers/clk/pistachio/clk.c of_clk_add_provider(p->node, of_clk_src_onecell_get, &p->clk_data); node 142 drivers/clk/pistachio/clk.h struct device_node *node; node 148 drivers/clk/pistachio/clk.h pistachio_clk_alloc_provider(struct device_node *node, unsigned int num_clks); node 123 drivers/clk/qcom/common.c struct device_node *node = NULL; node 132 drivers/clk/qcom/common.c node = of_get_child_by_name(clocks_node, path); node 136 drivers/clk/qcom/common.c if (!node) { node 151 drivers/clk/qcom/common.c of_node_put(node); node 54 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos_clkout_init(struct device_node *node, u32 mux_mask) node 74 drivers/clk/samsung/clk-exynos-clkout.c parents[i] = of_clk_get_by_name(node, name); node 87 drivers/clk/samsung/clk-exynos-clkout.c clkout->reg = of_iomap(node, 0); node 110 drivers/clk/samsung/clk-exynos-clkout.c ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, &clkout->data); node 138 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos4_clkout_init(struct device_node *node) node 140 drivers/clk/samsung/clk-exynos-clkout.c exynos_clkout_init(node, EXYNOS4_CLKOUT_MUX_MASK); node 149 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos5_clkout_init(struct device_node *node) node 151 drivers/clk/samsung/clk-exynos-clkout.c exynos_clkout_init(node, EXYNOS5_CLKOUT_MUX_MASK); node 291 drivers/clk/samsung/clk.c list_for_each_entry(reg_cache, &clock_reg_cache_list, node) { node 304 drivers/clk/samsung/clk.c list_for_each_entry(reg_cache, &clock_reg_cache_list, node) node 338 drivers/clk/samsung/clk.c list_add_tail(®_cache->node, &clock_reg_cache_list); node 275 drivers/clk/samsung/clk.h struct list_head node; node 94 drivers/clk/socfpga/clk-gate-a10.c static void __init __socfpga_gate_init(struct device_node *node, node 103 drivers/clk/socfpga/clk-gate-a10.c const char *clk_name = node->name; node 112 drivers/clk/socfpga/clk-gate-a10.c rc = of_property_read_u32_array(node, "clk-gate", clk_gate, 2); node 124 drivers/clk/socfpga/clk-gate-a10.c rc = of_property_read_u32(node, "fixed-divider", &fixed_div); node 130 drivers/clk/socfpga/clk-gate-a10.c rc = of_property_read_u32_array(node, "div-reg", div_reg, 3); node 139 drivers/clk/socfpga/clk-gate-a10.c rc = of_property_read_u32_array(node, "clk-phase", clk_phase, 2); node 153 drivers/clk/socfpga/clk-gate-a10.c of_property_read_string(node, "clock-output-names", &clk_name); node 159 drivers/clk/socfpga/clk-gate-a10.c init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS); node 168 drivers/clk/socfpga/clk-gate-a10.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 173 drivers/clk/socfpga/clk-gate-a10.c void __init socfpga_a10_gate_init(struct device_node *node) node 175 drivers/clk/socfpga/clk-gate-a10.c __socfpga_gate_init(node, &gateclk_ops); node 171 drivers/clk/socfpga/clk-gate.c void __init socfpga_gate_init(struct device_node *node) node 179 drivers/clk/socfpga/clk-gate.c const char *clk_name = node->name; node 193 drivers/clk/socfpga/clk-gate.c rc = of_property_read_u32_array(node, "clk-gate", clk_gate, 2); node 205 drivers/clk/socfpga/clk-gate.c rc = of_property_read_u32(node, "fixed-divider", &fixed_div); node 211 drivers/clk/socfpga/clk-gate.c rc = of_property_read_u32_array(node, "div-reg", div_reg, 3); node 220 drivers/clk/socfpga/clk-gate.c rc = of_property_read_u32_array(node, "clk-phase", clk_phase, 2); node 226 drivers/clk/socfpga/clk-gate.c of_property_read_string(node, "clock-output-names", &clk_name); node 232 drivers/clk/socfpga/clk-gate.c init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS); node 246 drivers/clk/socfpga/clk-gate.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 60 drivers/clk/socfpga/clk-periph-a10.c static __init void __socfpga_periph_init(struct device_node *node, node 66 drivers/clk/socfpga/clk-periph-a10.c const char *clk_name = node->name; node 73 drivers/clk/socfpga/clk-periph-a10.c of_property_read_u32(node, "reg", ®); node 81 drivers/clk/socfpga/clk-periph-a10.c rc = of_property_read_u32_array(node, "div-reg", div_reg, 3); node 90 drivers/clk/socfpga/clk-periph-a10.c rc = of_property_read_u32(node, "fixed-divider", &fixed_div); node 96 drivers/clk/socfpga/clk-periph-a10.c of_property_read_string(node, "clock-output-names", &clk_name); node 102 drivers/clk/socfpga/clk-periph-a10.c init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS); node 112 drivers/clk/socfpga/clk-periph-a10.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 125 drivers/clk/socfpga/clk-periph-a10.c void __init socfpga_a10_periph_init(struct device_node *node) node 127 drivers/clk/socfpga/clk-periph-a10.c __socfpga_periph_init(node, &periclk_ops); node 50 drivers/clk/socfpga/clk-periph.c static __init void __socfpga_periph_init(struct device_node *node, node 56 drivers/clk/socfpga/clk-periph.c const char *clk_name = node->name; node 63 drivers/clk/socfpga/clk-periph.c of_property_read_u32(node, "reg", ®); node 71 drivers/clk/socfpga/clk-periph.c rc = of_property_read_u32_array(node, "div-reg", div_reg, 3); node 80 drivers/clk/socfpga/clk-periph.c rc = of_property_read_u32(node, "fixed-divider", &fixed_div); node 86 drivers/clk/socfpga/clk-periph.c of_property_read_string(node, "clock-output-names", &clk_name); node 92 drivers/clk/socfpga/clk-periph.c init.num_parents = of_clk_parent_fill(node, parent_name, node 103 drivers/clk/socfpga/clk-periph.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 106 drivers/clk/socfpga/clk-periph.c void __init socfpga_periph_init(struct device_node *node) node 108 drivers/clk/socfpga/clk-periph.c __socfpga_periph_init(node, &periclk_ops); node 66 drivers/clk/socfpga/clk-pll-a10.c static struct clk * __init __socfpga_pll_init(struct device_node *node, node 72 drivers/clk/socfpga/clk-pll-a10.c const char *clk_name = node->name; node 79 drivers/clk/socfpga/clk-pll-a10.c of_property_read_u32(node, "reg", ®); node 91 drivers/clk/socfpga/clk-pll-a10.c of_property_read_string(node, "clock-output-names", &clk_name); node 98 drivers/clk/socfpga/clk-pll-a10.c of_clk_get_parent_name(node, i)) != NULL) node 113 drivers/clk/socfpga/clk-pll-a10.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 117 drivers/clk/socfpga/clk-pll-a10.c void __init socfpga_a10_pll_init(struct device_node *node) node 119 drivers/clk/socfpga/clk-pll-a10.c __socfpga_pll_init(node, &clk_pll_ops); node 73 drivers/clk/socfpga/clk-pll.c static __init struct clk *__socfpga_pll_init(struct device_node *node, node 79 drivers/clk/socfpga/clk-pll.c const char *clk_name = node->name; node 85 drivers/clk/socfpga/clk-pll.c of_property_read_u32(node, "reg", ®); node 97 drivers/clk/socfpga/clk-pll.c of_property_read_string(node, "clock-output-names", &clk_name); node 103 drivers/clk/socfpga/clk-pll.c init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS); node 116 drivers/clk/socfpga/clk-pll.c rc = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 120 drivers/clk/socfpga/clk-pll.c void __init socfpga_pll_init(struct device_node *node) node 122 drivers/clk/socfpga/clk-pll.c __socfpga_pll_init(node, &clk_pll_ops); node 32 drivers/clk/socfpga/clk.h void __init socfpga_pll_init(struct device_node *node); node 33 drivers/clk/socfpga/clk.h void __init socfpga_periph_init(struct device_node *node); node 34 drivers/clk/socfpga/clk.h void __init socfpga_gate_init(struct device_node *node); node 35 drivers/clk/socfpga/clk.h void socfpga_a10_pll_init(struct device_node *node); node 36 drivers/clk/socfpga/clk.h void socfpga_a10_periph_init(struct device_node *node); node 37 drivers/clk/socfpga/clk.h void socfpga_a10_gate_init(struct device_node *node); node 43 drivers/clk/sprd/common.c struct device_node *node = pdev->dev.of_node; node 47 drivers/clk/sprd/common.c if (of_find_property(node, "sprd,syscon", NULL)) { node 48 drivers/clk/sprd/common.c regmap = syscon_regmap_lookup_by_phandle(node, "sprd,syscon"); node 1428 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun4i_ccu_init(struct device_node *node, node 1434 drivers/clk/sunxi-ng/ccu-sun4i-a10.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 1437 drivers/clk/sunxi-ng/ccu-sun4i-a10.c of_node_full_name(node)); node 1467 drivers/clk/sunxi-ng/ccu-sun4i-a10.c sunxi_ccu_probe(node, reg, desc); node 1470 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun4i_a10_ccu_setup(struct device_node *node) node 1472 drivers/clk/sunxi-ng/ccu-sun4i-a10.c sun4i_ccu_init(node, &sun4i_a10_ccu_desc); node 1477 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun7i_a20_ccu_setup(struct device_node *node) node 1479 drivers/clk/sunxi-ng/ccu-sun4i-a10.c sun4i_ccu_init(node, &sun7i_a20_ccu_desc); node 188 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c static void __init sunxi_r_ccu_init(struct device_node *node, node 193 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 195 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c pr_err("%pOF: Could not map the clock registers\n", node); node 199 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c sunxi_ccu_probe(node, reg, desc); node 202 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c static void __init sun50i_h6_r_ccu_setup(struct device_node *node) node 204 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c sunxi_r_ccu_init(node, &sun50i_h6_r_ccu_desc); node 987 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_ccu_init(struct device_node *node, node 993 drivers/clk/sunxi-ng/ccu-sun5i.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 995 drivers/clk/sunxi-ng/ccu-sun5i.c pr_err("%pOF: Could not map the clock registers\n", node); node 1015 drivers/clk/sunxi-ng/ccu-sun5i.c sunxi_ccu_probe(node, reg, desc); node 1018 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_a10s_ccu_setup(struct device_node *node) node 1020 drivers/clk/sunxi-ng/ccu-sun5i.c sun5i_ccu_init(node, &sun5i_a10s_ccu_desc); node 1025 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_a13_ccu_setup(struct device_node *node) node 1027 drivers/clk/sunxi-ng/ccu-sun5i.c sun5i_ccu_init(node, &sun5i_a13_ccu_desc); node 1032 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_gr8_ccu_setup(struct device_node *node) node 1034 drivers/clk/sunxi-ng/ccu-sun5i.c sun5i_ccu_init(node, &sun5i_gr8_ccu_desc); node 1229 drivers/clk/sunxi-ng/ccu-sun6i-a31.c static void __init sun6i_a31_ccu_setup(struct device_node *node) node 1234 drivers/clk/sunxi-ng/ccu-sun6i-a31.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 1236 drivers/clk/sunxi-ng/ccu-sun6i-a31.c pr_err("%pOF: Could not map the clock registers\n", node); node 1260 drivers/clk/sunxi-ng/ccu-sun6i-a31.c sunxi_ccu_probe(node, reg, &sun6i_a31_ccu_desc); node 727 drivers/clk/sunxi-ng/ccu-sun8i-a23.c static void __init sun8i_a23_ccu_setup(struct device_node *node) node 732 drivers/clk/sunxi-ng/ccu-sun8i-a23.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 734 drivers/clk/sunxi-ng/ccu-sun8i-a23.c pr_err("%pOF: Could not map the clock registers\n", node); node 748 drivers/clk/sunxi-ng/ccu-sun8i-a23.c sunxi_ccu_probe(node, reg, &sun8i_a23_ccu_desc); node 787 drivers/clk/sunxi-ng/ccu-sun8i-a33.c static void __init sun8i_a33_ccu_setup(struct device_node *node) node 792 drivers/clk/sunxi-ng/ccu-sun8i-a33.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 794 drivers/clk/sunxi-ng/ccu-sun8i-a33.c pr_err("%pOF: Could not map the clock registers\n", node); node 808 drivers/clk/sunxi-ng/ccu-sun8i-a33.c sunxi_ccu_probe(node, reg, &sun8i_a33_ccu_desc); node 1139 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sunxi_h3_h5_ccu_init(struct device_node *node, node 1145 drivers/clk/sunxi-ng/ccu-sun8i-h3.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 1147 drivers/clk/sunxi-ng/ccu-sun8i-h3.c pr_err("%pOF: Could not map the clock registers\n", node); node 1156 drivers/clk/sunxi-ng/ccu-sun8i-h3.c sunxi_ccu_probe(node, reg, desc); node 1166 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sun8i_h3_ccu_setup(struct device_node *node) node 1168 drivers/clk/sunxi-ng/ccu-sun8i-h3.c sunxi_h3_h5_ccu_init(node, &sun8i_h3_ccu_desc); node 1173 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sun50i_h5_ccu_setup(struct device_node *node) node 1175 drivers/clk/sunxi-ng/ccu-sun8i-h3.c sunxi_h3_h5_ccu_init(node, &sun50i_h5_ccu_desc); node 257 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sunxi_r_ccu_init(struct device_node *node, node 262 drivers/clk/sunxi-ng/ccu-sun8i-r.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 264 drivers/clk/sunxi-ng/ccu-sun8i-r.c pr_err("%pOF: Could not map the clock registers\n", node); node 268 drivers/clk/sunxi-ng/ccu-sun8i-r.c sunxi_ccu_probe(node, reg, desc); node 271 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun8i_a83t_r_ccu_setup(struct device_node *node) node 273 drivers/clk/sunxi-ng/ccu-sun8i-r.c sunxi_r_ccu_init(node, &sun8i_a83t_r_ccu_desc); node 278 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun8i_h3_r_ccu_setup(struct device_node *node) node 280 drivers/clk/sunxi-ng/ccu-sun8i-r.c sunxi_r_ccu_init(node, &sun8i_h3_r_ccu_desc); node 285 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun50i_a64_r_ccu_setup(struct device_node *node) node 287 drivers/clk/sunxi-ng/ccu-sun8i-r.c sunxi_r_ccu_init(node, &sun50i_a64_r_ccu_desc); node 797 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3_v3s_ccu_init(struct device_node *node, node 803 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 805 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c pr_err("%pOF: Could not map the clock registers\n", node); node 814 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c sunxi_ccu_probe(node, reg, ccu_desc); node 817 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3s_ccu_setup(struct device_node *node) node 819 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c sun8i_v3_v3s_ccu_init(node, &sun8i_v3s_ccu_desc); node 822 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3_ccu_setup(struct device_node *node) node 824 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c sun8i_v3_v3s_ccu_init(node, &sun8i_v3_ccu_desc); node 525 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c static void __init suniv_f1c100s_ccu_setup(struct device_node *node) node 530 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 532 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c pr_err("%pOF: Could not map the clock registers\n", node); node 541 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c sunxi_ccu_probe(node, reg, &suniv_ccu_desc); node 82 drivers/clk/sunxi-ng/ccu_common.c int sunxi_ccu_probe(struct device_node *node, void __iomem *reg, node 106 drivers/clk/sunxi-ng/ccu_common.c ret = of_clk_hw_register(node, hw); node 113 drivers/clk/sunxi-ng/ccu_common.c ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, node 124 drivers/clk/sunxi-ng/ccu_common.c reset->rcdev.of_node = node; node 141 drivers/clk/sunxi-ng/ccu_common.c of_clk_del_provider(node); node 66 drivers/clk/sunxi-ng/ccu_common.h int sunxi_ccu_probe(struct device_node *node, void __iomem *reg, node 14 drivers/clk/sunxi/clk-a10-codec.c static void __init sun4i_codec_clk_setup(struct device_node *node) node 17 drivers/clk/sunxi/clk-a10-codec.c const char *clk_name = node->name, *parent_name; node 20 drivers/clk/sunxi/clk-a10-codec.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 24 drivers/clk/sunxi/clk-a10-codec.c of_property_read_string(node, "clock-output-names", &clk_name); node 25 drivers/clk/sunxi/clk-a10-codec.c parent_name = of_clk_get_parent_name(node, 0); node 32 drivers/clk/sunxi/clk-a10-codec.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 17 drivers/clk/sunxi/clk-a10-hosc.c static void __init sun4i_osc_clk_setup(struct device_node *node) node 22 drivers/clk/sunxi/clk-a10-hosc.c const char *clk_name = node->name; node 25 drivers/clk/sunxi/clk-a10-hosc.c if (of_property_read_u32(node, "clock-frequency", &rate)) node 36 drivers/clk/sunxi/clk-a10-hosc.c of_property_read_string(node, "clock-output-names", &clk_name); node 39 drivers/clk/sunxi/clk-a10-hosc.c gate->reg = of_iomap(node, 0); node 53 drivers/clk/sunxi/clk-a10-hosc.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 21 drivers/clk/sunxi/clk-a10-mod1.c static void __init sun4i_mod1_clk_setup(struct device_node *node) node 27 drivers/clk/sunxi/clk-a10-mod1.c const char *clk_name = node->name; node 31 drivers/clk/sunxi/clk-a10-mod1.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 43 drivers/clk/sunxi/clk-a10-mod1.c of_property_read_string(node, "clock-output-names", &clk_name); node 44 drivers/clk/sunxi/clk-a10-mod1.c i = of_clk_parent_fill(node, parents, SUN4I_MOD1_MAX_PARENTS); node 61 drivers/clk/sunxi/clk-a10-mod1.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 38 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun4i_pll2_setup(struct device_node *node, node 41 drivers/clk/sunxi/clk-a10-pll2.c const char *clk_name = node->name, *parent; node 49 drivers/clk/sunxi/clk-a10-pll2.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 61 drivers/clk/sunxi/clk-a10-pll2.c parent = of_clk_get_parent_name(node, 0); node 120 drivers/clk/sunxi/clk-a10-pll2.c of_property_read_string_index(node, "clock-output-names", node 135 drivers/clk/sunxi/clk-a10-pll2.c of_property_read_string_index(node, "clock-output-names", node 144 drivers/clk/sunxi/clk-a10-pll2.c of_property_read_string_index(node, "clock-output-names", node 153 drivers/clk/sunxi/clk-a10-pll2.c of_property_read_string_index(node, "clock-output-names", node 163 drivers/clk/sunxi/clk-a10-pll2.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 181 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun4i_a10_pll2_setup(struct device_node *node) node 183 drivers/clk/sunxi/clk-a10-pll2.c sun4i_pll2_setup(node, 0); node 189 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun5i_a13_pll2_setup(struct device_node *node) node 191 drivers/clk/sunxi/clk-a10-pll2.c sun4i_pll2_setup(node, 1); node 85 drivers/clk/sunxi/clk-a10-ve.c static void __init sun4i_ve_clk_setup(struct device_node *node) node 92 drivers/clk/sunxi/clk-a10-ve.c const char *clk_name = node->name; node 96 drivers/clk/sunxi/clk-a10-ve.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 108 drivers/clk/sunxi/clk-a10-ve.c of_property_read_string(node, "clock-output-names", &clk_name); node 109 drivers/clk/sunxi/clk-a10-ve.c parent = of_clk_get_parent_name(node, 0); node 128 drivers/clk/sunxi/clk-a10-ve.c err = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 140 drivers/clk/sunxi/clk-a10-ve.c reset_data->rcdev.of_node = node; node 152 drivers/clk/sunxi/clk-a10-ve.c of_clk_del_provider(node); node 52 drivers/clk/sunxi/clk-a20-gmac.c static void __init sun7i_a20_gmac_clk_setup(struct device_node *node) node 57 drivers/clk/sunxi/clk-a20-gmac.c const char *clk_name = node->name; node 61 drivers/clk/sunxi/clk-a20-gmac.c if (of_property_read_string(node, "clock-output-names", &clk_name)) node 74 drivers/clk/sunxi/clk-a20-gmac.c if (of_clk_parent_fill(node, parents, 2) != 2) node 77 drivers/clk/sunxi/clk-a20-gmac.c reg = of_iomap(node, 0); node 100 drivers/clk/sunxi/clk-a20-gmac.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 176 drivers/clk/sunxi/clk-factors.c static struct clk *__sunxi_factors_register(struct device_node *node, node 187 drivers/clk/sunxi/clk-factors.c const char *clk_name = node->name; node 192 drivers/clk/sunxi/clk-factors.c i = of_clk_parent_fill(node, parents, FACTORS_MAX_PARENTS); node 201 drivers/clk/sunxi/clk-factors.c of_property_read_string(node, "clock-output-names", &clk_name); node 253 drivers/clk/sunxi/clk-factors.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 272 drivers/clk/sunxi/clk-factors.c struct clk *sunxi_factors_register(struct device_node *node, node 277 drivers/clk/sunxi/clk-factors.c return __sunxi_factors_register(node, data, lock, reg, 0); node 280 drivers/clk/sunxi/clk-factors.c struct clk *sunxi_factors_register_critical(struct device_node *node, node 285 drivers/clk/sunxi/clk-factors.c return __sunxi_factors_register(node, data, lock, reg, CLK_IS_CRITICAL); node 288 drivers/clk/sunxi/clk-factors.c void sunxi_factors_unregister(struct device_node *node, struct clk *clk) node 298 drivers/clk/sunxi/clk-factors.c of_clk_del_provider(node); node 54 drivers/clk/sunxi/clk-factors.h struct clk *sunxi_factors_register(struct device_node *node, node 58 drivers/clk/sunxi/clk-factors.h struct clk *sunxi_factors_register_critical(struct device_node *node, node 63 drivers/clk/sunxi/clk-factors.h void sunxi_factors_unregister(struct device_node *node, struct clk *clk); node 68 drivers/clk/sunxi/clk-mod0.c static void __init sun4i_a10_mod0_setup(struct device_node *node) node 72 drivers/clk/sunxi/clk-mod0.c reg = of_iomap(node, 0); node 82 drivers/clk/sunxi/clk-mod0.c sunxi_factors_register(node, &sun4i_a10_mod0_data, node 129 drivers/clk/sunxi/clk-mod0.c static void __init sun9i_a80_mod0_setup(struct device_node *node) node 133 drivers/clk/sunxi/clk-mod0.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 136 drivers/clk/sunxi/clk-mod0.c node); node 140 drivers/clk/sunxi/clk-mod0.c sunxi_factors_register(node, &sun9i_a80_mod0_data, node 147 drivers/clk/sunxi/clk-mod0.c static void __init sun5i_a13_mbus_setup(struct device_node *node) node 151 drivers/clk/sunxi/clk-mod0.c reg = of_iomap(node, 0); node 158 drivers/clk/sunxi/clk-mod0.c sunxi_factors_register_critical(node, &sun4i_a10_mod0_data, node 290 drivers/clk/sunxi/clk-mod0.c static void __init sunxi_mmc_setup(struct device_node *node, node 299 drivers/clk/sunxi/clk-mod0.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 301 drivers/clk/sunxi/clk-mod0.c pr_err("Couldn't map the %pOFn clock registers\n", node); node 314 drivers/clk/sunxi/clk-mod0.c clk_data->clks[0] = sunxi_factors_register(node, data, lock, reg); node 341 drivers/clk/sunxi/clk-mod0.c if (of_property_read_string_index(node, "clock-output-names", node 343 drivers/clk/sunxi/clk-mod0.c init.name = node->name; node 352 drivers/clk/sunxi/clk-mod0.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 364 drivers/clk/sunxi/clk-mod0.c static void __init sun4i_a10_mmc_setup(struct device_node *node) node 366 drivers/clk/sunxi/clk-mod0.c sunxi_mmc_setup(node, &sun4i_a10_mod0_data, &sun4i_a10_mmc_lock); node 372 drivers/clk/sunxi/clk-mod0.c static void __init sun9i_a80_mmc_setup(struct device_node *node) node 374 drivers/clk/sunxi/clk-mod0.c sunxi_mmc_setup(node, &sun9i_a80_mod0_data, &sun9i_a80_mmc_lock); node 18 drivers/clk/sunxi/clk-simple-gates.c static void __init sunxi_simple_gates_setup(struct device_node *node, node 33 drivers/clk/sunxi/clk-simple-gates.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 37 drivers/clk/sunxi/clk-simple-gates.c clk_parent = of_clk_get_parent_name(node, 0); node 43 drivers/clk/sunxi/clk-simple-gates.c number = of_property_count_u32_elems(node, "clock-indices"); node 44 drivers/clk/sunxi/clk-simple-gates.c of_property_read_u32_index(node, "clock-indices", number - 1, &number); node 50 drivers/clk/sunxi/clk-simple-gates.c of_property_for_each_u32(node, "clock-indices", prop, p, index) { node 51 drivers/clk/sunxi/clk-simple-gates.c of_property_read_string_index(node, "clock-output-names", node 76 drivers/clk/sunxi/clk-simple-gates.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 84 drivers/clk/sunxi/clk-simple-gates.c of_address_to_resource(node, 0, &res); node 88 drivers/clk/sunxi/clk-simple-gates.c static void __init sunxi_simple_gates_init(struct device_node *node) node 90 drivers/clk/sunxi/clk-simple-gates.c sunxi_simple_gates_setup(node, NULL, 0); node 146 drivers/clk/sunxi/clk-simple-gates.c static void __init sun4i_a10_ahb_init(struct device_node *node) node 148 drivers/clk/sunxi/clk-simple-gates.c sunxi_simple_gates_setup(node, sun4i_a10_ahb_critical_clocks, node 164 drivers/clk/sunxi/clk-simple-gates.c static void __init sun4i_a10_dram_init(struct device_node *node) node 166 drivers/clk/sunxi/clk-simple-gates.c sunxi_simple_gates_setup(node, sun4i_a10_dram_critical_clocks, node 101 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_display_init(struct device_node *node, node 105 drivers/clk/sunxi/clk-sun4i-display.c const char *clk_name = node->name; node 115 drivers/clk/sunxi/clk-sun4i-display.c of_property_read_string(node, "clock-output-names", &clk_name); node 117 drivers/clk/sunxi/clk-sun4i-display.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 123 drivers/clk/sunxi/clk-sun4i-display.c ret = of_clk_parent_fill(node, parents, data->parents); node 169 drivers/clk/sunxi/clk-sun4i-display.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 187 drivers/clk/sunxi/clk-sun4i-display.c reset_data->rcdev.of_node = node; node 207 drivers/clk/sunxi/clk-sun4i-display.c of_clk_del_provider(node); node 218 drivers/clk/sunxi/clk-sun4i-display.c of_address_to_resource(node, 0, &res); node 232 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_tcon_ch0_setup(struct device_node *node) node 234 drivers/clk/sunxi/clk-sun4i-display.c sun4i_a10_display_init(node, &sun4i_a10_tcon_ch0_data); node 251 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_display_setup(struct device_node *node) node 253 drivers/clk/sunxi/clk-sun4i-display.c sun4i_a10_display_init(node, &sun4i_a10_display_data); node 21 drivers/clk/sunxi/clk-sun4i-pll3.c static void __init sun4i_a10_pll3_setup(struct device_node *node) node 23 drivers/clk/sunxi/clk-sun4i-pll3.c const char *clk_name = node->name, *parent; node 31 drivers/clk/sunxi/clk-sun4i-pll3.c of_property_read_string(node, "clock-output-names", &clk_name); node 32 drivers/clk/sunxi/clk-sun4i-pll3.c parent = of_clk_get_parent_name(node, 0); node 34 drivers/clk/sunxi/clk-sun4i-pll3.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 68 drivers/clk/sunxi/clk-sun4i-pll3.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 85 drivers/clk/sunxi/clk-sun4i-pll3.c of_address_to_resource(node, 0, &res); node 224 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c static void __init tcon_ch1_setup(struct device_node *node) node 227 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c const char *clk_name = node->name; node 235 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c of_property_read_string(node, "clock-output-names", &clk_name); node 237 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 243 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c ret = of_clk_parent_fill(node, parents, TCON_CH1_SCLK2_PARENTS); node 269 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 283 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c of_address_to_resource(node, 0, &res); node 22 drivers/clk/sunxi/clk-sun8i-apb0.c static struct clk *sun8i_a23_apb0_register(struct device_node *node, node 25 drivers/clk/sunxi/clk-sun8i-apb0.c const char *clk_name = node->name; node 30 drivers/clk/sunxi/clk-sun8i-apb0.c clk_parent = of_clk_get_parent_name(node, 0); node 34 drivers/clk/sunxi/clk-sun8i-apb0.c of_property_read_string(node, "clock-output-names", &clk_name); node 42 drivers/clk/sunxi/clk-sun8i-apb0.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 54 drivers/clk/sunxi/clk-sun8i-apb0.c static void sun8i_a23_apb0_setup(struct device_node *node) node 60 drivers/clk/sunxi/clk-sun8i-apb0.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 73 drivers/clk/sunxi/clk-sun8i-apb0.c clk = sun8i_a23_apb0_register(node, reg); node 81 drivers/clk/sunxi/clk-sun8i-apb0.c of_address_to_resource(node, 0, &res); node 20 drivers/clk/sunxi/clk-sun8i-bus-gates.c static void __init sun8i_h3_bus_gates_init(struct device_node *node) node 36 drivers/clk/sunxi/clk-sun8i-bus-gates.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 41 drivers/clk/sunxi/clk-sun8i-bus-gates.c int idx = of_property_match_string(node, "clock-names", node 46 drivers/clk/sunxi/clk-sun8i-bus-gates.c parents[i] = of_clk_get_parent_name(node, idx); node 53 drivers/clk/sunxi/clk-sun8i-bus-gates.c number = of_property_count_u32_elems(node, "clock-indices"); node 54 drivers/clk/sunxi/clk-sun8i-bus-gates.c of_property_read_u32_index(node, "clock-indices", number - 1, &number); node 61 drivers/clk/sunxi/clk-sun8i-bus-gates.c of_property_for_each_u32(node, "clock-indices", prop, p, index) { node 62 drivers/clk/sunxi/clk-sun8i-bus-gates.c of_property_read_string_index(node, "clock-output-names", node 94 drivers/clk/sunxi/clk-sun8i-bus-gates.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 102 drivers/clk/sunxi/clk-sun8i-bus-gates.c of_address_to_resource(node, 0, &res); node 24 drivers/clk/sunxi/clk-sun8i-mbus.c static void __init sun8i_a23_mbus_setup(struct device_node *node) node 26 drivers/clk/sunxi/clk-sun8i-mbus.c int num_parents = of_clk_get_parent_count(node); node 28 drivers/clk/sunxi/clk-sun8i-mbus.c const char *clk_name = node->name; node 41 drivers/clk/sunxi/clk-sun8i-mbus.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 59 drivers/clk/sunxi/clk-sun8i-mbus.c of_property_read_string(node, "clock-output-names", &clk_name); node 60 drivers/clk/sunxi/clk-sun8i-mbus.c of_clk_parent_fill(node, parents, num_parents); node 85 drivers/clk/sunxi/clk-sun8i-mbus.c err = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 104 drivers/clk/sunxi/clk-sun8i-mbus.c of_address_to_resource(node, 0, &res); node 76 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_pll4_setup(struct device_node *node) node 80 drivers/clk/sunxi/clk-sun9i-core.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 83 drivers/clk/sunxi/clk-sun9i-core.c node); node 87 drivers/clk/sunxi/clk-sun9i-core.c sunxi_factors_register(node, &sun9i_a80_pll4_data, node 130 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_gt_setup(struct device_node *node) node 134 drivers/clk/sunxi/clk-sun9i-core.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 137 drivers/clk/sunxi/clk-sun9i-core.c node); node 142 drivers/clk/sunxi/clk-sun9i-core.c sunxi_factors_register_critical(node, &sun9i_a80_gt_data, node 185 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_ahb_setup(struct device_node *node) node 189 drivers/clk/sunxi/clk-sun9i-core.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 192 drivers/clk/sunxi/clk-sun9i-core.c node); node 196 drivers/clk/sunxi/clk-sun9i-core.c sunxi_factors_register(node, &sun9i_a80_ahb_data, node 211 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_apb0_setup(struct device_node *node) node 215 drivers/clk/sunxi/clk-sun9i-core.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 218 drivers/clk/sunxi/clk-sun9i-core.c node); node 222 drivers/clk/sunxi/clk-sun9i-core.c sunxi_factors_register(node, &sun9i_a80_apb0_data, node 268 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_apb1_setup(struct device_node *node) node 272 drivers/clk/sunxi/clk-sun9i-core.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 275 drivers/clk/sunxi/clk-sun9i-core.c node); node 279 drivers/clk/sunxi/clk-sun9i-core.c sunxi_factors_register(node, &sun9i_a80_apb1_data, node 183 drivers/clk/sunxi/clk-sun9i-cpus.c static void sun9i_a80_cpus_setup(struct device_node *node) node 185 drivers/clk/sunxi/clk-sun9i-cpus.c const char *clk_name = node->name; node 197 drivers/clk/sunxi/clk-sun9i-cpus.c cpus->reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 201 drivers/clk/sunxi/clk-sun9i-cpus.c of_property_read_string(node, "clock-output-names", &clk_name); node 204 drivers/clk/sunxi/clk-sun9i-cpus.c ret = of_clk_parent_fill(node, parents, SUN9I_CPUS_MAX_PARENTS); node 224 drivers/clk/sunxi/clk-sun9i-cpus.c ret = of_clk_add_provider(node, of_clk_src_simple_get, clk); node 236 drivers/clk/sunxi/clk-sun9i-cpus.c of_address_to_resource(node, 0, &res); node 556 drivers/clk/sunxi/clk-sunxi.c static struct clk * __init sunxi_factors_clk_setup(struct device_node *node, node 561 drivers/clk/sunxi/clk-sunxi.c reg = of_iomap(node, 0); node 564 drivers/clk/sunxi/clk-sunxi.c node); node 568 drivers/clk/sunxi/clk-sunxi.c return sunxi_factors_register(node, data, &clk_lock, reg); node 571 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll1_clk_setup(struct device_node *node) node 573 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun4i_pll1_data); node 578 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_pll1_clk_setup(struct device_node *node) node 580 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun6i_a31_pll1_data); node 585 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_pll1_clk_setup(struct device_node *node) node 587 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun8i_a23_pll1_data); node 592 drivers/clk/sunxi/clk-sunxi.c static void __init sun7i_pll4_clk_setup(struct device_node *node) node 594 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun7i_a20_pll4_data); node 599 drivers/clk/sunxi/clk-sunxi.c static void __init sun5i_ahb_clk_setup(struct device_node *node) node 601 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun5i_a13_ahb_data); node 606 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_ahb1_clk_setup(struct device_node *node) node 608 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun6i_ahb1_data); node 613 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_apb1_clk_setup(struct device_node *node) node 615 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun4i_apb1_data); node 620 drivers/clk/sunxi/clk-sunxi.c static void __init sun7i_out_clk_setup(struct device_node *node) node 622 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun7i_a20_out_data); node 650 drivers/clk/sunxi/clk-sunxi.c static struct clk * __init sunxi_mux_clk_setup(struct device_node *node, node 655 drivers/clk/sunxi/clk-sunxi.c const char *clk_name = node->name; node 660 drivers/clk/sunxi/clk-sunxi.c reg = of_iomap(node, 0); node 662 drivers/clk/sunxi/clk-sunxi.c pr_err("Could not map registers for mux-clk: %pOF\n", node); node 666 drivers/clk/sunxi/clk-sunxi.c i = of_clk_parent_fill(node, parents, SUNXI_MAX_PARENTS); node 667 drivers/clk/sunxi/clk-sunxi.c if (of_property_read_string(node, "clock-output-names", &clk_name)) { node 669 drivers/clk/sunxi/clk-sunxi.c __func__, node); node 684 drivers/clk/sunxi/clk-sunxi.c if (of_clk_add_provider(node, of_clk_src_simple_get, clk)) { node 697 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_cpu_clk_setup(struct device_node *node) node 700 drivers/clk/sunxi/clk-sunxi.c sunxi_mux_clk_setup(node, &sun4i_cpu_mux_data, CLK_IS_CRITICAL); node 705 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_ahb1_mux_clk_setup(struct device_node *node) node 707 drivers/clk/sunxi/clk-sunxi.c sunxi_mux_clk_setup(node, &sun6i_a31_ahb1_mux_data, 0); node 712 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_ahb2_clk_setup(struct device_node *node) node 714 drivers/clk/sunxi/clk-sunxi.c sunxi_mux_clk_setup(node, &sun8i_h3_ahb2_mux_data, 0); node 775 drivers/clk/sunxi/clk-sunxi.c static void __init sunxi_divider_clk_setup(struct device_node *node, node 779 drivers/clk/sunxi/clk-sunxi.c const char *clk_name = node->name; node 783 drivers/clk/sunxi/clk-sunxi.c reg = of_iomap(node, 0); node 785 drivers/clk/sunxi/clk-sunxi.c pr_err("Could not map registers for mux-clk: %pOF\n", node); node 789 drivers/clk/sunxi/clk-sunxi.c clk_parent = of_clk_get_parent_name(node, 0); node 791 drivers/clk/sunxi/clk-sunxi.c if (of_property_read_string(node, "clock-output-names", &clk_name)) { node 793 drivers/clk/sunxi/clk-sunxi.c __func__, node); node 807 drivers/clk/sunxi/clk-sunxi.c if (of_clk_add_provider(node, of_clk_src_simple_get, clk)) { node 814 drivers/clk/sunxi/clk-sunxi.c of_clk_del_provider(node); node 826 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_ahb_clk_setup(struct device_node *node) node 828 drivers/clk/sunxi/clk-sunxi.c sunxi_divider_clk_setup(node, &sun4i_ahb_data); node 833 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_apb0_clk_setup(struct device_node *node) node 835 drivers/clk/sunxi/clk-sunxi.c sunxi_divider_clk_setup(node, &sun4i_apb0_data); node 840 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_axi_clk_setup(struct device_node *node) node 842 drivers/clk/sunxi/clk-sunxi.c sunxi_divider_clk_setup(node, &sun4i_axi_data); node 847 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_axi_clk_setup(struct device_node *node) node 849 drivers/clk/sunxi/clk-sunxi.c sunxi_divider_clk_setup(node, &sun8i_a23_axi_data); node 943 drivers/clk/sunxi/clk-sunxi.c static struct clk ** __init sunxi_divs_clk_setup(struct device_node *node, node 968 drivers/clk/sunxi/clk-sunxi.c of_property_read_string_index(node, "clock-output-names", node 977 drivers/clk/sunxi/clk-sunxi.c of_property_read_string_index(node, "clock-output-names", node 992 drivers/clk/sunxi/clk-sunxi.c pclk = sunxi_factors_clk_setup(node, &factors); node 999 drivers/clk/sunxi/clk-sunxi.c reg = of_iomap(node, 0); node 1001 drivers/clk/sunxi/clk-sunxi.c pr_err("Could not map registers for divs-clk: %pOF\n", node); node 1020 drivers/clk/sunxi/clk-sunxi.c if (of_property_read_string_index(node, "clock-output-names", node 1092 drivers/clk/sunxi/clk-sunxi.c if (of_clk_add_provider(node, of_clk_src_onecell_get, clk_data)) { node 1110 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll5_clk_setup(struct device_node *node) node 1112 drivers/clk/sunxi/clk-sunxi.c sunxi_divs_clk_setup(node, &pll5_divs_data); node 1117 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll6_clk_setup(struct device_node *node) node 1119 drivers/clk/sunxi/clk-sunxi.c sunxi_divs_clk_setup(node, &pll6_divs_data); node 1124 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_pll6_clk_setup(struct device_node *node) node 1126 drivers/clk/sunxi/clk-sunxi.c sunxi_divs_clk_setup(node, &sun6i_a31_pll6_divs_data); node 1162 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_display_setup(struct device_node *node) node 1164 drivers/clk/sunxi/clk-sunxi.c sunxi_factors_clk_setup(node, &sun6i_display_data); node 88 drivers/clk/sunxi/clk-usb.c static void __init sunxi_usb_clk_setup(struct device_node *node, node 101 drivers/clk/sunxi/clk-usb.c reg = of_io_request_and_map(node, 0, of_node_full_name(node)); node 105 drivers/clk/sunxi/clk-usb.c clk_parent = of_clk_get_parent_name(node, 0); node 125 drivers/clk/sunxi/clk-usb.c of_property_read_string_index(node, "clock-output-names", node 138 drivers/clk/sunxi/clk-usb.c of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 149 drivers/clk/sunxi/clk-usb.c reset_data->clk = of_clk_get(node, 0); node 161 drivers/clk/sunxi/clk-usb.c reset_data->rcdev.of_node = node; node 172 drivers/clk/sunxi/clk-usb.c static void __init sun4i_a10_usb_setup(struct device_node *node) node 174 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun4i_a10_usb_clk_data, &sun4i_a10_usb_lock); node 183 drivers/clk/sunxi/clk-usb.c static void __init sun5i_a13_usb_setup(struct device_node *node) node 185 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun5i_a13_usb_clk_data, &sun4i_a10_usb_lock); node 194 drivers/clk/sunxi/clk-usb.c static void __init sun6i_a31_usb_setup(struct device_node *node) node 196 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun6i_a31_usb_clk_data, &sun4i_a10_usb_lock); node 205 drivers/clk/sunxi/clk-usb.c static void __init sun8i_a23_usb_setup(struct device_node *node) node 207 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun8i_a23_usb_clk_data, &sun4i_a10_usb_lock); node 217 drivers/clk/sunxi/clk-usb.c static void __init sun8i_h3_usb_setup(struct device_node *node) node 219 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun8i_h3_usb_clk_data, &sun4i_a10_usb_lock); node 231 drivers/clk/sunxi/clk-usb.c static void __init sun9i_a80_usb_mod_setup(struct device_node *node) node 233 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun9i_a80_usb_mod_data, &a80_usb_mod_lock); node 245 drivers/clk/sunxi/clk-usb.c static void __init sun9i_a80_usb_phy_setup(struct device_node *node) node 247 drivers/clk/sunxi/clk-usb.c sunxi_usb_clk_setup(node, &sun9i_a80_usb_phy_data, &a80_usb_phy_lock); node 378 drivers/clk/tegra/clk-emc.c struct device_node *node) node 383 drivers/clk/tegra/clk-emc.c err = of_property_read_u32(node, "clock-frequency", &tmp); node 385 drivers/clk/tegra/clk-emc.c pr_err("timing %pOF: failed to read rate\n", node); node 391 drivers/clk/tegra/clk-emc.c err = of_property_read_u32(node, "nvidia,parent-clock-frequency", &tmp); node 393 drivers/clk/tegra/clk-emc.c pr_err("timing %pOF: failed to read parent rate\n", node); node 399 drivers/clk/tegra/clk-emc.c timing->parent = of_clk_get_by_name(node, "emc-parent"); node 401 drivers/clk/tegra/clk-emc.c pr_err("timing %pOF: failed to get parent clock\n", node); node 415 drivers/clk/tegra/clk-emc.c node, __clk_get_name(timing->parent)); node 437 drivers/clk/tegra/clk-emc.c struct device_node *node, node 442 drivers/clk/tegra/clk-emc.c int child_count = of_get_child_count(node); node 455 drivers/clk/tegra/clk-emc.c for_each_child_of_node(node, child) { node 485 drivers/clk/tegra/clk-emc.c struct device_node *node; node 499 drivers/clk/tegra/clk-emc.c for_each_child_of_node(np, node) { node 500 drivers/clk/tegra/clk-emc.c err = of_property_read_u32(node, "nvidia,ram-code", node 509 drivers/clk/tegra/clk-emc.c err = load_timings_from_dt(tegra, node, node_ram_code); node 511 drivers/clk/tegra/clk-emc.c of_node_put(node); node 1324 drivers/clk/tegra/clk-tegra114.c struct device_node *node; node 1332 drivers/clk/tegra/clk-tegra114.c node = of_find_matching_node(NULL, pmc_match); node 1333 drivers/clk/tegra/clk-tegra114.c if (!node) { node 1339 drivers/clk/tegra/clk-tegra114.c pmc_base = of_iomap(node, 0); node 1415 drivers/clk/tegra/clk-tegra124.c struct device_node *node; node 1424 drivers/clk/tegra/clk-tegra124.c node = of_find_matching_node(NULL, pmc_match); node 1425 drivers/clk/tegra/clk-tegra124.c if (!node) { node 1431 drivers/clk/tegra/clk-tegra124.c pmc_base = of_iomap(node, 0); node 1139 drivers/clk/tegra/clk-tegra20.c struct device_node *node; node 1147 drivers/clk/tegra/clk-tegra20.c node = of_find_matching_node(NULL, pmc_match); node 1148 drivers/clk/tegra/clk-tegra20.c if (!node) { node 1153 drivers/clk/tegra/clk-tegra20.c pmc_base = of_iomap(node, 0); node 3509 drivers/clk/tegra/clk-tegra210.c struct device_node *node; node 3518 drivers/clk/tegra/clk-tegra210.c node = of_find_matching_node(NULL, pmc_match); node 3519 drivers/clk/tegra/clk-tegra210.c if (!node) { node 3525 drivers/clk/tegra/clk-tegra210.c pmc_base = of_iomap(node, 0); node 1307 drivers/clk/tegra/clk-tegra30.c struct device_node *node; node 1315 drivers/clk/tegra/clk-tegra30.c node = of_find_matching_node(NULL, pmc_match); node 1316 drivers/clk/tegra/clk-tegra30.c if (!node) { node 1321 drivers/clk/tegra/clk-tegra30.c pmc_base = of_iomap(node, 0); node 877 drivers/clk/ti/adpll.c struct device_node *node = pdev->dev.of_node; node 895 drivers/clk/ti/adpll.c d->np = node; node 137 drivers/clk/ti/apll.c struct device_node *node) node 145 drivers/clk/ti/apll.c clk = of_clk_get(node, 0); node 148 drivers/clk/ti/apll.c node); node 149 drivers/clk/ti/apll.c if (!ti_clk_retry_init(node, hw, omap_clk_register_apll)) node 157 drivers/clk/ti/apll.c clk = of_clk_get(node, 1); node 160 drivers/clk/ti/apll.c node); node 161 drivers/clk/ti/apll.c if (!ti_clk_retry_init(node, hw, omap_clk_register_apll)) node 169 drivers/clk/ti/apll.c clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name); node 171 drivers/clk/ti/apll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 184 drivers/clk/ti/apll.c static void __init of_dra7_apll_setup(struct device_node *node) node 201 drivers/clk/ti/apll.c init->name = node->name; node 204 drivers/clk/ti/apll.c init->num_parents = of_clk_get_parent_count(node); node 206 drivers/clk/ti/apll.c pr_err("dra7 apll %pOFn must have parent(s)\n", node); node 214 drivers/clk/ti/apll.c of_clk_parent_fill(node, parent_names, init->num_parents); node 218 drivers/clk/ti/apll.c ret = ti_clk_get_reg_addr(node, 0, &ad->control_reg); node 219 drivers/clk/ti/apll.c ret |= ti_clk_get_reg_addr(node, 1, &ad->idlest_reg); node 227 drivers/clk/ti/apll.c omap_clk_register_apll(&clk_hw->hw, node); node 345 drivers/clk/ti/apll.c static void __init of_omap2_apll_setup(struct device_node *node) node 365 drivers/clk/ti/apll.c init->name = node->name; node 368 drivers/clk/ti/apll.c init->num_parents = of_clk_get_parent_count(node); node 370 drivers/clk/ti/apll.c pr_err("%pOFn must have one parent\n", node); node 374 drivers/clk/ti/apll.c parent_name = of_clk_get_parent_name(node, 0); node 377 drivers/clk/ti/apll.c if (of_property_read_u32(node, "ti,clock-frequency", &val)) { node 378 drivers/clk/ti/apll.c pr_err("%pOFn missing clock-frequency\n", node); node 383 drivers/clk/ti/apll.c if (of_property_read_u32(node, "ti,bit-shift", &val)) { node 384 drivers/clk/ti/apll.c pr_err("%pOFn missing bit-shift\n", node); node 392 drivers/clk/ti/apll.c if (of_property_read_u32(node, "ti,idlest-shift", &val)) { node 393 drivers/clk/ti/apll.c pr_err("%pOFn missing idlest-shift\n", node); node 399 drivers/clk/ti/apll.c ret = ti_clk_get_reg_addr(node, 0, &ad->control_reg); node 400 drivers/clk/ti/apll.c ret |= ti_clk_get_reg_addr(node, 1, &ad->autoidle_reg); node 401 drivers/clk/ti/apll.c ret |= ti_clk_get_reg_addr(node, 2, &ad->idlest_reg); node 406 drivers/clk/ti/apll.c clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name); node 408 drivers/clk/ti/apll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 32 drivers/clk/ti/autoidle.c struct list_head node; node 153 drivers/clk/ti/autoidle.c list_for_each_entry(c, &autoidle_clks, node) node 167 drivers/clk/ti/autoidle.c list_for_each_entry(c, &autoidle_clks, node) node 182 drivers/clk/ti/autoidle.c int __init of_ti_clk_autoidle_setup(struct device_node *node) node 189 drivers/clk/ti/autoidle.c if (of_property_read_u32(node, "ti,autoidle-shift", &shift)) node 198 drivers/clk/ti/autoidle.c clk->name = node->name; node 199 drivers/clk/ti/autoidle.c ret = ti_clk_get_reg_addr(node, 0, &clk->reg); node 205 drivers/clk/ti/autoidle.c if (of_property_read_bool(node, "ti,invert-autoidle-bit")) node 208 drivers/clk/ti/autoidle.c list_add(&clk->node, &autoidle_clks); node 171 drivers/clk/ti/clk-dra7-atl.c static void __init of_dra7_atl_clock_setup(struct device_node *node) node 186 drivers/clk/ti/clk-dra7-atl.c init.name = node->name; node 189 drivers/clk/ti/clk-dra7-atl.c init.num_parents = of_clk_get_parent_count(node); node 193 drivers/clk/ti/clk-dra7-atl.c node); node 202 drivers/clk/ti/clk-dra7-atl.c parent_names[0] = of_clk_get_parent_name(node, 0); node 206 drivers/clk/ti/clk-dra7-atl.c clk = ti_clk_register(NULL, &clk_hw->hw, node->name); node 209 drivers/clk/ti/clk-dra7-atl.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 221 drivers/clk/ti/clk-dra7-atl.c struct device_node *node = pdev->dev.of_node; node 226 drivers/clk/ti/clk-dra7-atl.c if (!node) node 233 drivers/clk/ti/clk-dra7-atl.c cinfo->iobase = of_iomap(node, 0); node 249 drivers/clk/ti/clk-dra7-atl.c rc = of_parse_phandle_with_args(node, "ti,provided-clocks", node 271 drivers/clk/ti/clk-dra7-atl.c cfg_node = of_get_child_by_name(node, prop); node 134 drivers/clk/ti/clk.c struct device_node *node, *parent; node 171 drivers/clk/ti/clk.c node = of_find_node_by_name(NULL, buf); node 173 drivers/clk/ti/clk.c parent = node; node 174 drivers/clk/ti/clk.c node = of_get_child_by_name(parent, "clk"); node 178 drivers/clk/ti/clk.c clkspec.np = node; node 185 drivers/clk/ti/clk.c of_node_put(node); node 190 drivers/clk/ti/clk.c of_node_put(node); node 218 drivers/clk/ti/clk.c struct device_node *node; node 235 drivers/clk/ti/clk.c int __init ti_clk_retry_init(struct device_node *node, void *user, node 240 drivers/clk/ti/clk.c pr_debug("%pOFn: adding to retry list...\n", node); node 245 drivers/clk/ti/clk.c retry->node = node; node 263 drivers/clk/ti/clk.c int ti_clk_get_reg_addr(struct device_node *node, int index, node 270 drivers/clk/ti/clk.c if (clocks_node_ptr[i] == node->parent) node 275 drivers/clk/ti/clk.c pr_err("clk-provider not found for %pOFn!\n", node); node 281 drivers/clk/ti/clk.c if (of_property_read_u32_index(node, "reg", index, &val)) { node 282 drivers/clk/ti/clk.c pr_err("%pOFn must have reg[%d]!\n", node, index); node 385 drivers/clk/ti/clk.c pr_debug("retry-init: %pOFn\n", retry->node); node 386 drivers/clk/ti/clk.c retry->func(retry->user, retry->node); node 553 drivers/clk/ti/clk.c list_add(&oclk->node, &clk_hw_omap_clocks); node 573 drivers/clk/ti/clk.c list_for_each_entry(hw, &clk_hw_omap_clocks, node) { node 593 drivers/clk/ti/clk.c list_for_each_entry(oclk, &clk_hw_omap_clocks, node) { node 57 drivers/clk/ti/clkctrl.c struct list_head node; node 241 drivers/clk/ti/clkctrl.c list_for_each_entry(entry, &provider->clocks, node) { node 257 drivers/clk/ti/clkctrl.c struct device_node *node, struct clk_hw *clk_hw, node 268 drivers/clk/ti/clkctrl.c node->parent, node, offset, node 271 drivers/clk/ti/clkctrl.c init.name = kasprintf(GFP_KERNEL, "%pOFn:%04x:%d", node, node 295 drivers/clk/ti/clkctrl.c list_add(&clkctrl_clk->node, &provider->clocks); node 307 drivers/clk/ti/clkctrl.c struct device_node *node, u16 offset, node 320 drivers/clk/ti/clkctrl.c if (_ti_clkctrl_clk_register(provider, node, &clk_hw->hw, offset, node 328 drivers/clk/ti/clkctrl.c struct device_node *node, u16 offset, node 355 drivers/clk/ti/clkctrl.c if (_ti_clkctrl_clk_register(provider, node, &mux->hw, offset, node 363 drivers/clk/ti/clkctrl.c struct device_node *node, u16 offset, node 386 drivers/clk/ti/clkctrl.c node, offset, data->bit); node 391 drivers/clk/ti/clkctrl.c if (_ti_clkctrl_clk_register(provider, node, &div->hw, offset, node 399 drivers/clk/ti/clkctrl.c struct device_node *node, node 411 drivers/clk/ti/clkctrl.c _ti_clkctrl_setup_gate(provider, node, data->offset, node 416 drivers/clk/ti/clkctrl.c _ti_clkctrl_setup_div(provider, node, data->offset, node 421 drivers/clk/ti/clkctrl.c _ti_clkctrl_setup_mux(provider, node, data->offset, node 440 drivers/clk/ti/clkctrl.c static void __init _ti_omap4_clkctrl_setup(struct device_node *node) node 456 drivers/clk/ti/clkctrl.c of_node_name_eq(node, "clk")) node 459 drivers/clk/ti/clkctrl.c addrp = of_get_address(node, 0, NULL, NULL); node 460 drivers/clk/ti/clkctrl.c addr = (u32)of_translate_address(node, addrp); node 527 drivers/clk/ti/clkctrl.c pr_err("%pOF not found from clkctrl data.\n", node); node 535 drivers/clk/ti/clkctrl.c provider->base = of_iomap(node, 0); node 538 drivers/clk/ti/clkctrl.c provider->clkdm_name = kasprintf(GFP_KERNEL, "%pOFnxxx", node->parent); node 550 drivers/clk/ti/clkctrl.c provider->clkdm_name = kasprintf(GFP_KERNEL, "%pOFn", node); node 592 drivers/clk/ti/clkctrl.c _ti_clkctrl_setup_subclks(provider, node, reg_data, node 614 drivers/clk/ti/clkctrl.c node->parent, node, node 618 drivers/clk/ti/clkctrl.c node, reg_data->offset, 0); node 633 drivers/clk/ti/clkctrl.c list_add(&clkctrl_clk->node, &provider->clocks); node 638 drivers/clk/ti/clkctrl.c ret = of_clk_add_hw_provider(node, _ti_omap4_clkctrl_xlate, provider); node 640 drivers/clk/ti/clkctrl.c ti_clk_retry_init(node, provider, _clkctrl_add_provider); node 226 drivers/clk/ti/clock.h int ti_clk_get_reg_addr(struct device_node *node, int index, node 229 drivers/clk/ti/clock.h int ti_clk_retry_init(struct device_node *node, void *user, node 231 drivers/clk/ti/clock.h int ti_clk_add_component(struct device_node *node, struct clk_hw *hw, int type); node 233 drivers/clk/ti/clock.h int of_ti_clk_autoidle_setup(struct device_node *node); node 128 drivers/clk/ti/clockdomain.c static void __init of_ti_clockdomain_setup(struct device_node *node) node 132 drivers/clk/ti/clockdomain.c const char *clkdm_name = node->name; node 136 drivers/clk/ti/clockdomain.c num_clks = of_clk_get_parent_count(node); node 139 drivers/clk/ti/clockdomain.c clk = of_clk_get(node, i); node 142 drivers/clk/ti/clockdomain.c __func__, node, i, PTR_ERR(clk)); node 64 drivers/clk/ti/composite.c struct device_node *node; node 76 drivers/clk/ti/composite.c static struct device_node *_get_component_node(struct device_node *node, int i) node 81 drivers/clk/ti/composite.c rc = of_parse_phandle_with_args(node, "clocks", "#clock-cells", i, node 89 drivers/clk/ti/composite.c static struct component_clk *_lookup_component(struct device_node *node) node 94 drivers/clk/ti/composite.c if (comp->node == node) node 120 drivers/clk/ti/composite.c struct device_node *node) node 139 drivers/clk/ti/composite.c cclk->comp_nodes[i]->name, node); node 140 drivers/clk/ti/composite.c if (!ti_clk_retry_init(node, hw, node 148 drivers/clk/ti/composite.c node, component_clk_types[comp->type]); node 171 drivers/clk/ti/composite.c pr_err("%s: no parents found for %pOFn!\n", __func__, node); node 175 drivers/clk/ti/composite.c clk = clk_register_composite(NULL, node->name, node 185 drivers/clk/ti/composite.c ret = ti_clk_add_alias(NULL, clk, node->name); node 190 drivers/clk/ti/composite.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 205 drivers/clk/ti/composite.c static void __init of_ti_composite_clk_setup(struct device_node *node) node 212 drivers/clk/ti/composite.c num_clks = of_clk_get_parent_count(node); node 215 drivers/clk/ti/composite.c pr_err("composite clk %pOFn must have component(s)\n", node); node 225 drivers/clk/ti/composite.c cclk->comp_nodes[i] = _get_component_node(node, i); node 227 drivers/clk/ti/composite.c _register_composite(&cclk->hw, node); node 241 drivers/clk/ti/composite.c int __init ti_clk_add_component(struct device_node *node, struct clk_hw *hw, node 248 drivers/clk/ti/composite.c num_parents = of_clk_get_parent_count(node); node 251 drivers/clk/ti/composite.c pr_err("component-clock %pOFn must have parent(s)\n", node); node 259 drivers/clk/ti/composite.c of_clk_parent_fill(node, parent_names, num_parents); node 270 drivers/clk/ti/composite.c clk->node = node; node 429 drivers/clk/ti/divider.c __init ti_clk_get_div_table(struct device_node *node) node 438 drivers/clk/ti/divider.c divspec = of_get_property(node, "ti,dividers", &num_div); node 449 drivers/clk/ti/divider.c of_property_read_u32_index(node, "ti,dividers", i, &val); node 455 drivers/clk/ti/divider.c pr_err("no valid dividers for %pOFn table\n", node); node 467 drivers/clk/ti/divider.c of_property_read_u32_index(node, "ti,dividers", i, &val); node 478 drivers/clk/ti/divider.c static int _get_divider_width(struct device_node *node, node 489 drivers/clk/ti/divider.c if (of_property_read_u32(node, "ti,min-div", &min_div)) node 492 drivers/clk/ti/divider.c if (of_property_read_u32(node, "ti,max-div", &max_div)) { node 493 drivers/clk/ti/divider.c pr_err("no max-div for %pOFn!\n", node); node 522 drivers/clk/ti/divider.c static int __init ti_clk_divider_populate(struct device_node *node, node 529 drivers/clk/ti/divider.c ret = ti_clk_get_reg_addr(node, 0, reg); node 533 drivers/clk/ti/divider.c if (!of_property_read_u32(node, "ti,bit-shift", &val)) node 539 drivers/clk/ti/divider.c if (!of_property_read_u32(node, "ti,latch-bit", &val)) node 548 drivers/clk/ti/divider.c if (of_property_read_bool(node, "ti,index-starts-at-one")) node 551 drivers/clk/ti/divider.c if (of_property_read_bool(node, "ti,index-power-of-two")) node 554 drivers/clk/ti/divider.c if (of_property_read_bool(node, "ti,set-rate-parent")) node 557 drivers/clk/ti/divider.c *table = ti_clk_get_div_table(node); node 562 drivers/clk/ti/divider.c *width = _get_divider_width(node, *table, *div_flags); node 573 drivers/clk/ti/divider.c static void __init of_ti_divider_clk_setup(struct device_node *node) node 585 drivers/clk/ti/divider.c parent_name = of_clk_get_parent_name(node, 0); node 587 drivers/clk/ti/divider.c if (ti_clk_divider_populate(node, ®, &table, &flags, node 591 drivers/clk/ti/divider.c clk = _register_divider(NULL, node->name, parent_name, flags, ®, node 595 drivers/clk/ti/divider.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 596 drivers/clk/ti/divider.c of_ti_clk_autoidle_setup(node); node 605 drivers/clk/ti/divider.c static void __init of_ti_composite_divider_clk_setup(struct device_node *node) node 614 drivers/clk/ti/divider.c if (ti_clk_divider_populate(node, &div->reg, &div->table, &val, node 619 drivers/clk/ti/divider.c if (!ti_clk_add_component(node, &div->hw, CLK_COMPONENT_TYPE_DIVIDER)) node 162 drivers/clk/ti/dpll.c struct device_node *node) node 170 drivers/clk/ti/dpll.c clk = of_clk_get(node, 0); node 173 drivers/clk/ti/dpll.c node); node 174 drivers/clk/ti/dpll.c if (!ti_clk_retry_init(node, hw, _register_dpll)) node 182 drivers/clk/ti/dpll.c clk = of_clk_get(node, 1); node 186 drivers/clk/ti/dpll.c node); node 187 drivers/clk/ti/dpll.c if (!ti_clk_retry_init(node, hw, _register_dpll)) node 196 drivers/clk/ti/dpll.c clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name); node 199 drivers/clk/ti/dpll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 223 drivers/clk/ti/dpll.c static void _register_dpll_x2(struct device_node *node, node 230 drivers/clk/ti/dpll.c const char *name = node->name; node 233 drivers/clk/ti/dpll.c parent_name = of_clk_get_parent_name(node, 0); node 235 drivers/clk/ti/dpll.c pr_err("%pOFn must have parent\n", node); node 257 drivers/clk/ti/dpll.c ret = of_property_count_elems_of_size(node, "reg", 1); node 260 drivers/clk/ti/dpll.c } else if (ti_clk_get_reg_addr(node, 0, &clk_hw->clksel_reg)) { node 273 drivers/clk/ti/dpll.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 285 drivers/clk/ti/dpll.c static void __init of_ti_dpll_setup(struct device_node *node, node 305 drivers/clk/ti/dpll.c init->name = node->name; node 308 drivers/clk/ti/dpll.c init->num_parents = of_clk_get_parent_count(node); node 310 drivers/clk/ti/dpll.c pr_err("%pOFn must have parent(s)\n", node); node 318 drivers/clk/ti/dpll.c of_clk_parent_fill(node, parent_names, init->num_parents); node 322 drivers/clk/ti/dpll.c if (ti_clk_get_reg_addr(node, 0, &dd->control_reg)) node 331 drivers/clk/ti/dpll.c if (ti_clk_get_reg_addr(node, 1, &dd->mult_div1_reg)) node 338 drivers/clk/ti/dpll.c if (ti_clk_get_reg_addr(node, 1, &dd->idlest_reg)) node 341 drivers/clk/ti/dpll.c if (ti_clk_get_reg_addr(node, 2, &dd->mult_div1_reg)) node 346 drivers/clk/ti/dpll.c if (ti_clk_get_reg_addr(node, 3, &dd->autoidle_reg)) node 350 drivers/clk/ti/dpll.c if (of_property_read_bool(node, "ti,low-power-stop")) node 353 drivers/clk/ti/dpll.c if (of_property_read_bool(node, "ti,low-power-bypass")) node 356 drivers/clk/ti/dpll.c if (of_property_read_bool(node, "ti,lock")) node 362 drivers/clk/ti/dpll.c _register_dpll(&clk_hw->hw, node); node 374 drivers/clk/ti/dpll.c static void __init of_ti_omap4_dpll_x2_setup(struct device_node *node) node 376 drivers/clk/ti/dpll.c _register_dpll_x2(node, &dpll_x2_ck_ops, &clkhwops_omap4_dpllmx); node 383 drivers/clk/ti/dpll.c static void __init of_ti_am3_dpll_x2_setup(struct device_node *node) node 385 drivers/clk/ti/dpll.c _register_dpll_x2(node, &dpll_x2_ck_ops, NULL); node 392 drivers/clk/ti/dpll.c static void __init of_ti_omap3_dpll_setup(struct device_node *node) node 409 drivers/clk/ti/dpll.c of_node_name_eq(node, "dpll5_ck")) node 410 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap3_dpll5_ck_ops, &dd); node 412 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap3_dpll_ck_ops, &dd); node 417 drivers/clk/ti/dpll.c static void __init of_ti_omap3_core_dpll_setup(struct device_node *node) node 431 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap3_dpll_core_ck_ops, &dd); node 436 drivers/clk/ti/dpll.c static void __init of_ti_omap3_per_dpll_setup(struct device_node *node) node 451 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd); node 456 drivers/clk/ti/dpll.c static void __init of_ti_omap3_per_jtype_dpll_setup(struct device_node *node) node 473 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd); node 479 drivers/clk/ti/dpll.c static void __init of_ti_omap4_dpll_setup(struct device_node *node) node 493 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_ck_ops, &dd); node 498 drivers/clk/ti/dpll.c static void __init of_ti_omap5_mpu_dpll_setup(struct device_node *node) node 514 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_ck_ops, &dd); node 519 drivers/clk/ti/dpll.c static void __init of_ti_omap4_core_dpll_setup(struct device_node *node) node 533 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd); node 540 drivers/clk/ti/dpll.c static void __init of_ti_omap4_m4xen_dpll_setup(struct device_node *node) node 556 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd); node 561 drivers/clk/ti/dpll.c static void __init of_ti_omap4_jtype_dpll_setup(struct device_node *node) node 577 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd); node 583 drivers/clk/ti/dpll.c static void __init of_ti_am3_no_gate_dpll_setup(struct device_node *node) node 597 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd); node 602 drivers/clk/ti/dpll.c static void __init of_ti_am3_jtype_dpll_setup(struct device_node *node) node 617 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_ck_ops, &dd); node 622 drivers/clk/ti/dpll.c static void __init of_ti_am3_no_gate_jtype_dpll_setup(struct device_node *node) node 637 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd); node 643 drivers/clk/ti/dpll.c static void __init of_ti_am3_dpll_setup(struct device_node *node) node 657 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_ck_ops, &dd); node 661 drivers/clk/ti/dpll.c static void __init of_ti_am3_core_dpll_setup(struct device_node *node) node 675 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd); node 680 drivers/clk/ti/dpll.c static void __init of_ti_omap2_core_dpll_setup(struct device_node *node) node 690 drivers/clk/ti/dpll.c of_ti_dpll_setup(node, &omap2_dpll_core_ck_ops, &dd); node 532 drivers/clk/ti/fapll.c static void __init ti_fapll_setup(struct device_node *node) node 555 drivers/clk/ti/fapll.c init->name = node->name; node 557 drivers/clk/ti/fapll.c init->num_parents = of_clk_get_parent_count(node); node 559 drivers/clk/ti/fapll.c pr_err("%pOFn must have two parents\n", node); node 563 drivers/clk/ti/fapll.c of_clk_parent_fill(node, parent_name, 2); node 566 drivers/clk/ti/fapll.c fd->clk_ref = of_clk_get(node, 0); node 568 drivers/clk/ti/fapll.c pr_err("%pOFn could not get clk_ref\n", node); node 572 drivers/clk/ti/fapll.c fd->clk_bypass = of_clk_get(node, 1); node 574 drivers/clk/ti/fapll.c pr_err("%pOFn could not get clk_bypass\n", node); node 578 drivers/clk/ti/fapll.c fd->base = of_iomap(node, 0); node 580 drivers/clk/ti/fapll.c pr_err("%pOFn could not get IO base\n", node); node 587 drivers/clk/ti/fapll.c fd->name = node->name; node 612 drivers/clk/ti/fapll.c if (of_property_read_string_index(node, "clock-output-names", node 616 drivers/clk/ti/fapll.c if (of_property_read_u32_index(node, "clock-indices", i, node 634 drivers/clk/ti/fapll.c output_name, node->name, node 646 drivers/clk/ti/fapll.c of_clk_add_provider(node, of_clk_src_onecell_get, &fd->outputs); node 36 drivers/clk/ti/fixed-factor.c static void __init of_ti_fixed_factor_clk_setup(struct device_node *node) node 39 drivers/clk/ti/fixed-factor.c const char *clk_name = node->name; node 44 drivers/clk/ti/fixed-factor.c if (of_property_read_u32(node, "ti,clock-div", &div)) { node 45 drivers/clk/ti/fixed-factor.c pr_err("%pOFn must have a clock-div property\n", node); node 49 drivers/clk/ti/fixed-factor.c if (of_property_read_u32(node, "ti,clock-mult", &mult)) { node 50 drivers/clk/ti/fixed-factor.c pr_err("%pOFn must have a clock-mult property\n", node); node 54 drivers/clk/ti/fixed-factor.c if (of_property_read_bool(node, "ti,set-rate-parent")) node 57 drivers/clk/ti/fixed-factor.c parent_name = of_clk_get_parent_name(node, 0); node 63 drivers/clk/ti/fixed-factor.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 64 drivers/clk/ti/fixed-factor.c of_ti_clk_autoidle_setup(node); node 134 drivers/clk/ti/gate.c static void __init _of_ti_gate_clk_setup(struct device_node *node, node 147 drivers/clk/ti/gate.c if (ti_clk_get_reg_addr(node, 0, ®)) node 150 drivers/clk/ti/gate.c if (!of_property_read_u32(node, "ti,bit-shift", &val)) node 154 drivers/clk/ti/gate.c if (of_clk_get_parent_count(node) != 1) { node 155 drivers/clk/ti/gate.c pr_err("%pOFn must have 1 parent\n", node); node 159 drivers/clk/ti/gate.c parent_name = of_clk_get_parent_name(node, 0); node 161 drivers/clk/ti/gate.c if (of_property_read_bool(node, "ti,set-rate-parent")) node 164 drivers/clk/ti/gate.c if (of_property_read_bool(node, "ti,set-bit-to-disable")) node 167 drivers/clk/ti/gate.c clk = _register_gate(NULL, node->name, parent_name, flags, ®, node 171 drivers/clk/ti/gate.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 175 drivers/clk/ti/gate.c _of_ti_composite_gate_clk_setup(struct device_node *node, node 185 drivers/clk/ti/gate.c if (ti_clk_get_reg_addr(node, 0, &gate->enable_reg)) node 188 drivers/clk/ti/gate.c of_property_read_u32(node, "ti,bit-shift", &val); node 193 drivers/clk/ti/gate.c if (!ti_clk_add_component(node, &gate->hw, CLK_COMPONENT_TYPE_GATE)) node 201 drivers/clk/ti/gate.c of_ti_composite_no_wait_gate_clk_setup(struct device_node *node) node 203 drivers/clk/ti/gate.c _of_ti_composite_gate_clk_setup(node, NULL); node 209 drivers/clk/ti/gate.c static void __init of_ti_composite_interface_clk_setup(struct device_node *node) node 211 drivers/clk/ti/gate.c _of_ti_composite_gate_clk_setup(node, &clkhwops_iclk_wait); node 217 drivers/clk/ti/gate.c static void __init of_ti_composite_gate_clk_setup(struct device_node *node) node 219 drivers/clk/ti/gate.c _of_ti_composite_gate_clk_setup(node, &clkhwops_wait); node 225 drivers/clk/ti/gate.c static void __init of_ti_clkdm_gate_clk_setup(struct device_node *node) node 227 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clkdm_clk_ops, NULL); node 232 drivers/clk/ti/gate.c static void __init of_ti_hsdiv_gate_clk_setup(struct device_node *node) node 234 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clk_hsdiv_restore_ops, node 240 drivers/clk/ti/gate.c static void __init of_ti_gate_clk_setup(struct device_node *node) node 242 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clk_ops, NULL); node 246 drivers/clk/ti/gate.c static void __init of_ti_wait_gate_clk_setup(struct device_node *node) node 248 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clk_ops, &clkhwops_wait); node 254 drivers/clk/ti/gate.c static void __init of_ti_am35xx_gate_clk_setup(struct device_node *node) node 256 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clk_ops, node 262 drivers/clk/ti/gate.c static void __init of_ti_dss_gate_clk_setup(struct device_node *node) node 264 drivers/clk/ti/gate.c _of_ti_gate_clk_setup(node, &omap_gate_clk_ops, node 68 drivers/clk/ti/interface.c static void __init _of_ti_interface_clk_setup(struct device_node *node, node 77 drivers/clk/ti/interface.c if (ti_clk_get_reg_addr(node, 0, ®)) node 80 drivers/clk/ti/interface.c if (!of_property_read_u32(node, "ti,bit-shift", &val)) node 83 drivers/clk/ti/interface.c parent_name = of_clk_get_parent_name(node, 0); node 85 drivers/clk/ti/interface.c pr_err("%pOFn must have a parent\n", node); node 89 drivers/clk/ti/interface.c clk = _register_interface(NULL, node->name, parent_name, ®, node 93 drivers/clk/ti/interface.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 96 drivers/clk/ti/interface.c static void __init of_ti_interface_clk_setup(struct device_node *node) node 98 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, &clkhwops_iclk_wait); node 103 drivers/clk/ti/interface.c static void __init of_ti_no_wait_interface_clk_setup(struct device_node *node) node 105 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, &clkhwops_iclk); node 111 drivers/clk/ti/interface.c static void __init of_ti_hsotgusb_interface_clk_setup(struct device_node *node) node 113 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, node 119 drivers/clk/ti/interface.c static void __init of_ti_dss_interface_clk_setup(struct device_node *node) node 121 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, node 127 drivers/clk/ti/interface.c static void __init of_ti_ssi_interface_clk_setup(struct device_node *node) node 129 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, &clkhwops_omap3430es2_iclk_ssi_wait); node 134 drivers/clk/ti/interface.c static void __init of_ti_am35xx_interface_clk_setup(struct device_node *node) node 136 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, &clkhwops_am35xx_ipss_wait); node 143 drivers/clk/ti/interface.c static void __init of_ti_omap2430_interface_clk_setup(struct device_node *node) node 145 drivers/clk/ti/interface.c _of_ti_interface_clk_setup(node, &clkhwops_omap2430_i2chs_wait); node 173 drivers/clk/ti/mux.c static void of_mux_clk_setup(struct device_node *node) node 185 drivers/clk/ti/mux.c num_parents = of_clk_get_parent_count(node); node 187 drivers/clk/ti/mux.c pr_err("mux-clock %pOFn must have parents\n", node); node 194 drivers/clk/ti/mux.c of_clk_parent_fill(node, parent_names, num_parents); node 196 drivers/clk/ti/mux.c if (ti_clk_get_reg_addr(node, 0, ®)) node 199 drivers/clk/ti/mux.c of_property_read_u32(node, "ti,bit-shift", &shift); node 201 drivers/clk/ti/mux.c of_property_read_u32(node, "ti,latch-bit", &latch); node 203 drivers/clk/ti/mux.c if (of_property_read_bool(node, "ti,index-starts-at-one")) node 206 drivers/clk/ti/mux.c if (of_property_read_bool(node, "ti,set-rate-parent")) node 216 drivers/clk/ti/mux.c clk = _register_mux(NULL, node->name, parent_names, num_parents, node 221 drivers/clk/ti/mux.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 257 drivers/clk/ti/mux.c static void __init of_ti_composite_mux_clk_setup(struct device_node *node) node 267 drivers/clk/ti/mux.c if (ti_clk_get_reg_addr(node, 0, &mux->reg)) node 270 drivers/clk/ti/mux.c if (!of_property_read_u32(node, "ti,bit-shift", &val)) node 273 drivers/clk/ti/mux.c if (of_property_read_bool(node, "ti,index-starts-at-one")) node 276 drivers/clk/ti/mux.c num_parents = of_clk_get_parent_count(node); node 279 drivers/clk/ti/mux.c pr_err("%pOFn must have parents\n", node); node 286 drivers/clk/ti/mux.c if (!ti_clk_add_component(node, &mux->hw, CLK_COMPONENT_TYPE_MUX)) node 29 drivers/clk/versatile/clk-sp810.c struct device_node *node; node 82 drivers/clk/versatile/clk-sp810.c static void __init clk_sp810_of_setup(struct device_node *node) node 96 drivers/clk/versatile/clk-sp810.c if (of_clk_parent_fill(node, parent_names, num) != num) { node 102 drivers/clk/versatile/clk-sp810.c sp810->node = node; node 103 drivers/clk/versatile/clk-sp810.c sp810->base = of_iomap(node, 0); node 112 drivers/clk/versatile/clk-sp810.c deprecated = !of_find_property(node, "assigned-clock-parents", NULL); node 135 drivers/clk/versatile/clk-sp810.c of_clk_add_provider(node, clk_sp810_timerclken_of_get, sp810); node 74 drivers/clk/zynqmp/clkc.c struct clock_topology node[MAX_NODES]; node 522 drivers/clk/zynqmp/clkc.c clk_nodes = clock[clk_id].node; node 565 drivers/clk/zynqmp/clkc.c nodes = clock[clk_id].node; node 690 drivers/clk/zynqmp/clkc.c ret = zynqmp_clock_get_topology(i, clock[i].node, node 32 drivers/clocksource/arc_timer.c static int noinline arc_get_timer_clk(struct device_node *node) node 37 drivers/clocksource/arc_timer.c clk = of_clk_get(node, 0); node 103 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_gfrc(struct device_node *node) node 114 drivers/clocksource/arc_timer.c ret = arc_get_timer_clk(node); node 161 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_rtc(struct device_node *node) node 178 drivers/clocksource/arc_timer.c ret = arc_get_timer_clk(node); node 214 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_timer1(struct device_node *node) node 222 drivers/clocksource/arc_timer.c ret = arc_get_timer_clk(node); node 325 drivers/clocksource/arc_timer.c static int __init arc_clockevent_setup(struct device_node *node) node 330 drivers/clocksource/arc_timer.c arc_timer_irq = irq_of_parse_and_map(node, 0); node 336 drivers/clocksource/arc_timer.c ret = arc_get_timer_clk(node); node 70 drivers/clocksource/bcm2835_timer.c static int __init bcm2835_timer_init(struct device_node *node) node 77 drivers/clocksource/bcm2835_timer.c base = of_iomap(node, 0); node 83 drivers/clocksource/bcm2835_timer.c ret = of_property_read_u32(node, "clock-frequency", &freq); node 92 drivers/clocksource/bcm2835_timer.c clocksource_mmio_init(base + REG_COUNTER_LO, node->name, node 95 drivers/clocksource/bcm2835_timer.c irq = irq_of_parse_and_map(node, DEFAULT_TIMER); node 111 drivers/clocksource/bcm2835_timer.c timer->evt.name = node->name; node 116 drivers/clocksource/bcm2835_timer.c timer->act.name = node->name; node 169 drivers/clocksource/bcm_kona_timer.c static int __init kona_timer_init(struct device_node *node) node 174 drivers/clocksource/bcm_kona_timer.c external_clk = of_clk_get_by_name(node, NULL); node 179 drivers/clocksource/bcm_kona_timer.c } else if (!of_property_read_u32(node, "clock-frequency", &freq)) { node 187 drivers/clocksource/bcm_kona_timer.c timers.tmr_irq = irq_of_parse_and_map(node, 0); node 190 drivers/clocksource/bcm_kona_timer.c timers.tmr_regs = of_iomap(node, 0); node 52 drivers/clocksource/clksrc-dbx500-prcmu.c static int __init clksrc_dbx500_prcmu_init(struct device_node *node) node 54 drivers/clocksource/clksrc-dbx500-prcmu.c clksrc_dbx500_timer_base = of_iomap(node, 0); node 130 drivers/clocksource/h8300_timer16.c static int __init h8300_16timer_init(struct device_node *node) node 137 drivers/clocksource/h8300_timer16.c clk = of_clk_get(node, 0); node 144 drivers/clocksource/h8300_timer16.c base[REG_CH] = of_iomap(node, 0); node 150 drivers/clocksource/h8300_timer16.c base[REG_COMM] = of_iomap(node, 1); node 157 drivers/clocksource/h8300_timer16.c irq = irq_of_parse_and_map(node, 0); node 163 drivers/clocksource/h8300_timer16.c of_property_read_u32(node, "renesas,channel", &ch); node 160 drivers/clocksource/h8300_timer8.c static int __init h8300_8timer_init(struct device_node *node) node 166 drivers/clocksource/h8300_timer8.c clk = of_clk_get(node, 0); node 173 drivers/clocksource/h8300_timer8.c base = of_iomap(node, 0); node 180 drivers/clocksource/h8300_timer8.c irq = irq_of_parse_and_map(node, 0); node 123 drivers/clocksource/h8300_tpu.c static int __init h8300_tpu_init(struct device_node *node) node 129 drivers/clocksource/h8300_tpu.c clk = of_clk_get(node, 0); node 135 drivers/clocksource/h8300_tpu.c base[CH_L] = of_iomap(node, CH_L); node 140 drivers/clocksource/h8300_tpu.c base[CH_H] = of_iomap(node, CH_H); node 136 drivers/clocksource/jcore-pit.c static int __init jcore_pit_init(struct device_node *node) node 143 drivers/clocksource/jcore-pit.c jcore_pit_base = of_iomap(node, 0); node 149 drivers/clocksource/jcore-pit.c pit_irq = irq_of_parse_and_map(node, 0); node 221 drivers/clocksource/jcore-pit.c pit->base = of_iomap(node, cpu); node 186 drivers/clocksource/mips-gic-timer.c static int __init gic_clocksource_of_init(struct device_node *node) node 191 drivers/clocksource/mips-gic-timer.c if (!mips_gic_present() || !node->parent || node 192 drivers/clocksource/mips-gic-timer.c !of_device_is_compatible(node->parent, "mti,gic")) { node 197 drivers/clocksource/mips-gic-timer.c clk = of_clk_get(node, 0); node 207 drivers/clocksource/mips-gic-timer.c } else if (of_property_read_u32(node, "clock-frequency", node 212 drivers/clocksource/mips-gic-timer.c gic_timer_irq = irq_of_parse_and_map(node, 0); node 247 drivers/clocksource/nomadik-mtu.c static int __init nmdk_timer_of_init(struct device_node *node) node 254 drivers/clocksource/nomadik-mtu.c base = of_iomap(node, 0); node 260 drivers/clocksource/nomadik-mtu.c pclk = of_clk_get_by_name(node, "apb_pclk"); node 266 drivers/clocksource/nomadik-mtu.c clk = of_clk_get_by_name(node, "timclk"); node 272 drivers/clocksource/nomadik-mtu.c irq = irq_of_parse_and_map(node, 0); node 211 drivers/clocksource/timer-atcpit100.c static void fill_vdso_need_info(struct device_node *node) node 214 drivers/clocksource/timer-atcpit100.c of_address_to_resource(node, 0, &timer_res); node 221 drivers/clocksource/timer-atcpit100.c static int __init atcpit100_timer_init(struct device_node *node) node 227 drivers/clocksource/timer-atcpit100.c ret = timer_of_init(node, &to); node 237 drivers/clocksource/timer-atcpit100.c node->name, timer_of_rate(&to), 300, 32, node 260 drivers/clocksource/timer-atcpit100.c fill_vdso_need_info(node); node 166 drivers/clocksource/timer-atmel-pit.c static int __init at91sam926x_pit_dt_init(struct device_node *node) node 177 drivers/clocksource/timer-atmel-pit.c data->base = of_iomap(node, 0); node 184 drivers/clocksource/timer-atmel-pit.c data->mck = of_clk_get(node, 0); node 198 drivers/clocksource/timer-atmel-pit.c data->irq = irq_of_parse_and_map(node, 0); node 184 drivers/clocksource/timer-atmel-st.c static int __init atmel_st_timer_init(struct device_node *node) node 190 drivers/clocksource/timer-atmel-st.c regmap_st = syscon_node_to_regmap(node); node 202 drivers/clocksource/timer-atmel-st.c irq = irq_of_parse_and_map(node, 0); node 217 drivers/clocksource/timer-atmel-st.c sclk = of_clk_get(node, 0); node 357 drivers/clocksource/timer-atmel-tcb.c static int __init tcb_clksrc_init(struct device_node *node) node 374 drivers/clocksource/timer-atmel-tcb.c tc.regs = of_iomap(node->parent, 0); node 378 drivers/clocksource/timer-atmel-tcb.c t0_clk = of_clk_get_by_name(node->parent, "t0_clk"); node 382 drivers/clocksource/timer-atmel-tcb.c tc.slow_clk = of_clk_get_by_name(node->parent, "slow_clk"); node 387 drivers/clocksource/timer-atmel-tcb.c tc.clk[1] = of_clk_get_by_name(node->parent, "t1_clk"); node 390 drivers/clocksource/timer-atmel-tcb.c tc.clk[2] = of_clk_get_by_name(node->parent, "t2_clk"); node 394 drivers/clocksource/timer-atmel-tcb.c tc.irq[2] = of_irq_get(node->parent, 2); node 396 drivers/clocksource/timer-atmel-tcb.c tc.irq[2] = of_irq_get(node->parent, 0); node 401 drivers/clocksource/timer-atmel-tcb.c match = of_match_node(atmel_tcb_of_match, node->parent); node 435 drivers/clocksource/timer-atmel-tcb.c clksrc.name = kbasename(node->parent->full_name); node 436 drivers/clocksource/timer-atmel-tcb.c clkevt.clkevt.name = kbasename(node->parent->full_name); node 152 drivers/clocksource/timer-digicolor.c static int __init digicolor_timer_init(struct device_node *node) node 162 drivers/clocksource/timer-digicolor.c dc_timer_dev.base = of_iomap(node, 0); node 168 drivers/clocksource/timer-digicolor.c irq = irq_of_parse_and_map(node, dc_timer_dev.timer_id); node 174 drivers/clocksource/timer-digicolor.c clk = of_clk_get(node, 0); node 188 drivers/clocksource/timer-digicolor.c clocksource_mmio_init(dc_timer_dev.base + COUNT(TIMER_B), node->name, node 163 drivers/clocksource/timer-integrator-ap.c static int __init integrator_ap_timer_init_of(struct device_node *node) node 173 drivers/clocksource/timer-integrator-ap.c base = of_io_request_and_map(node, 0, "integrator-timer"); node 177 drivers/clocksource/timer-integrator-ap.c clk = of_clk_get(node, 0); node 179 drivers/clocksource/timer-integrator-ap.c pr_err("No clock for %pOFn\n", node); node 202 drivers/clocksource/timer-integrator-ap.c if (node == alias_node) node 217 drivers/clocksource/timer-integrator-ap.c if (node == alias_node) { node 219 drivers/clocksource/timer-integrator-ap.c irq = irq_of_parse_and_map(node, 0); node 258 drivers/clocksource/timer-mediatek.c static int __init mtk_syst_init(struct device_node *node) node 269 drivers/clocksource/timer-mediatek.c ret = timer_of_init(node, &to); node 279 drivers/clocksource/timer-mediatek.c static int __init mtk_gpt_init(struct device_node *node) node 291 drivers/clocksource/timer-mediatek.c ret = timer_of_init(node, &to); node 298 drivers/clocksource/timer-mediatek.c node->name, timer_of_rate(&to), 300, 32, node 160 drivers/clocksource/timer-meson6.c static int __init meson6_timer_init(struct device_node *node) node 165 drivers/clocksource/timer-meson6.c timer_base = of_io_request_and_map(node, 0, "meson6-timer"); node 171 drivers/clocksource/timer-meson6.c irq = irq_of_parse_and_map(node, 0); node 185 drivers/clocksource/timer-meson6.c clocksource_mmio_init(timer_base + MESON_ISA_TIMERE, node->name, node 168 drivers/clocksource/timer-milbeaut.c static int __init mlb_timer_init(struct device_node *node) node 173 drivers/clocksource/timer-milbeaut.c ret = timer_of_init(node, &to); node 180 drivers/clocksource/timer-milbeaut.c node->name, rate, MLB_TIMER_RATING, 32, node 49 drivers/clocksource/timer-nps.c static int __init nps_get_timer_clk(struct device_node *node, node 55 drivers/clocksource/timer-nps.c *clk = of_clk_get(node, 0); node 87 drivers/clocksource/timer-nps.c static int __init nps_setup_clocksource(struct device_node *node) node 99 drivers/clocksource/timer-nps.c ret = nps_get_timer_clk(node, &nps_timer1_freq, &clk); node 243 drivers/clocksource/timer-nps.c static int __init nps_setup_clockevent(struct device_node *node) node 248 drivers/clocksource/timer-nps.c nps_timer0_irq = irq_of_parse_and_map(node, 0); node 254 drivers/clocksource/timer-nps.c ret = nps_get_timer_clk(node, &nps_timer0_freq, &clk); node 116 drivers/clocksource/timer-owl.c static int __init owl_timer_init(struct device_node *node) node 122 drivers/clocksource/timer-owl.c owl_timer_base = of_io_request_and_map(node, 0, "owl-timer"); node 131 drivers/clocksource/timer-owl.c timer1_irq = of_irq_get_byname(node, "timer1"); node 137 drivers/clocksource/timer-owl.c clk = of_clk_get(node, 0); node 147 drivers/clocksource/timer-owl.c clocksource_mmio_init(owl_clksrc_base + OWL_Tx_VAL, node->name, node 151 drivers/clocksource/timer-pistachio.c static int __init pistachio_clksrc_of_init(struct device_node *node) node 158 drivers/clocksource/timer-pistachio.c pcs_gpt.base = of_iomap(node, 0); node 164 drivers/clocksource/timer-pistachio.c periph_regs = syscon_regmap_lookup_by_phandle(node, "img,cr-periph"); node 177 drivers/clocksource/timer-pistachio.c sys_clk = of_clk_get_by_name(node, "sys"); node 183 drivers/clocksource/timer-pistachio.c fast_clk = of_clk_get_by_name(node, "fast"); node 288 drivers/clocksource/timer-stm32.c static int __init stm32_timer_init(struct device_node *node) node 301 drivers/clocksource/timer-stm32.c ret = timer_of_init(node, to); node 312 drivers/clocksource/timer-stm32.c rstc = of_reset_control_get(node, NULL); node 169 drivers/clocksource/timer-sun4i.c static int __init sun4i_timer_init(struct device_node *node) node 174 drivers/clocksource/timer-sun4i.c ret = timer_of_init(node, &to); node 195 drivers/clocksource/timer-sun4i.c node->name, timer_of_rate(&to), 350, 32, node 186 drivers/clocksource/timer-sun5i.c static int __init sun5i_setup_clocksource(struct device_node *node, node 226 drivers/clocksource/timer-sun5i.c cs->clksrc.name = node->name; node 264 drivers/clocksource/timer-sun5i.c static int __init sun5i_setup_clockevent(struct device_node *node, void __iomem *base, node 301 drivers/clocksource/timer-sun5i.c ce->clkevt.name = node->name; node 337 drivers/clocksource/timer-sun5i.c static int __init sun5i_timer_init(struct device_node *node) node 344 drivers/clocksource/timer-sun5i.c timer_base = of_io_request_and_map(node, 0, of_node_full_name(node)); node 350 drivers/clocksource/timer-sun5i.c irq = irq_of_parse_and_map(node, 0); node 356 drivers/clocksource/timer-sun5i.c clk = of_clk_get(node, 0); node 362 drivers/clocksource/timer-sun5i.c rstc = of_reset_control_get(node, NULL); node 366 drivers/clocksource/timer-sun5i.c ret = sun5i_setup_clocksource(node, timer_base, clk, irq); node 370 drivers/clocksource/timer-sun5i.c return sun5i_setup_clockevent(node, timer_base, clk, irq); node 325 drivers/clocksource/timer-ti-dm.c list_for_each_entry(t, &omap_timer_list, node) { node 474 drivers/clocksource/timer-ti-dm.c list_for_each_entry(timer, &omap_timer_list, node) { node 758 drivers/clocksource/timer-ti-dm.c list_for_each_entry(timer, &omap_timer_list, node) { node 857 drivers/clocksource/timer-ti-dm.c list_add_tail(&timer->node, &omap_timer_list); node 885 drivers/clocksource/timer-ti-dm.c list_for_each_entry(timer, &omap_timer_list, node) node 888 drivers/clocksource/timer-ti-dm.c list_del(&timer->node); node 21 drivers/clocksource/timer-versatile.c static int __init versatile_sched_clock_init(struct device_node *node) node 23 drivers/clocksource/timer-versatile.c void __iomem *base = of_iomap(node, 0); node 117 drivers/clocksource/timer-zevio.c static int __init zevio_timer_add(struct device_node *node) node 127 drivers/clocksource/timer-zevio.c timer->base = of_iomap(node, 0); node 135 drivers/clocksource/timer-zevio.c timer->clk = of_clk_get(node, 0); node 142 drivers/clocksource/timer-zevio.c timer->interrupt_regs = of_iomap(node, 1); node 143 drivers/clocksource/timer-zevio.c irqnr = irq_of_parse_and_map(node, 0); node 145 drivers/clocksource/timer-zevio.c of_address_to_resource(node, 0, &res); node 148 drivers/clocksource/timer-zevio.c (unsigned long long)res.start, node); node 152 drivers/clocksource/timer-zevio.c (unsigned long long)res.start, node); node 209 drivers/clocksource/timer-zevio.c static int __init zevio_timer_init(struct device_node *node) node 211 drivers/clocksource/timer-zevio.c return zevio_timer_add(node); node 285 drivers/counter/ftm-quaddec.c struct device_node *node = pdev->dev.of_node; node 302 drivers/counter/ftm-quaddec.c ftm->big_endian = of_property_read_bool(node, "big-endian"); node 127 drivers/cpufreq/armada-8k-cpufreq.c struct device_node *node; node 130 drivers/cpufreq/armada-8k-cpufreq.c node = of_find_compatible_node(NULL, NULL, "marvell,ap806-cpu-clock"); node 131 drivers/cpufreq/armada-8k-cpufreq.c if (!node || !of_device_is_available(node)) { node 132 drivers/cpufreq/armada-8k-cpufreq.c of_node_put(node); node 135 drivers/cpufreq/armada-8k-cpufreq.c of_node_put(node); node 69 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c u8 node, slow_mode; node 74 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c node = pmi_msg.data1; node 77 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c cpu = cbe_node_to_cpu(node); node 79 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c pr_debug("cbe_handle_pmi: node: %d max_freq: %d\n", node, slow_mode); node 94 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c pr_debug("limiting node %d to slow mode %d\n", node, slow_mode); node 99 drivers/crypto/cavium/cpt/cptvf.h u8 node; /* Operating node: Bits (46:44) in BAR0 address */ node 176 drivers/crypto/cavium/cpt/cptvf_main.c struct hlist_node *node; node 184 drivers/crypto/cavium/cpt/cptvf_main.c hlist_for_each_entry_safe(chunk, node, &cqinfo->queue[i].chead, node 627 drivers/crypto/cavium/cpt/cptvf_main.c cpumask_set_cpu(cpumask_local_spread(cpu, cptvf->node), node 710 drivers/crypto/cavium/cpt/cptvf_main.c cptvf->node = dev_to_node(&pdev->dev); node 247 drivers/crypto/cavium/cpt/cptvf_reqmanager.c struct hlist_node *node; node 249 drivers/crypto/cavium/cpt/cptvf_reqmanager.c hlist_for_each(node, &queue->chead) { node 250 drivers/crypto/cavium/cpt/cptvf_reqmanager.c chunk = hlist_entry(node, struct command_chunk, node 233 drivers/crypto/cavium/nitrox/nitrox_dev.h int node; node 106 drivers/crypto/cavium/nitrox/nitrox_lib.c cmdq = kzalloc_node(sizeof(*cmdq), GFP_KERNEL, ndev->node); node 157 drivers/crypto/cavium/nitrox/nitrox_lib.c GFP_KERNEL, ndev->node); node 475 drivers/crypto/cavium/nitrox/nitrox_main.c ndev->node = dev_to_node(&pdev->dev); node 476 drivers/crypto/cavium/nitrox/nitrox_main.c if (ndev->node == NUMA_NO_NODE) node 477 drivers/crypto/cavium/nitrox/nitrox_main.c ndev->node = 0; node 100 drivers/crypto/cavium/zip/zip_main.c struct zip_device *zip_get_device(int node) node 102 drivers/crypto/cavium/zip/zip_main.c if ((node < MAX_ZIP_DEVICES) && (node >= 0)) node 103 drivers/crypto/cavium/zip/zip_main.c return zip_dev[node]; node 105 drivers/crypto/cavium/zip/zip_main.c zip_err("ZIP device not found for node id %d\n", node); node 391 drivers/crypto/hisilicon/sec/sec_drv.h u8 node[16]; node 72 drivers/crypto/hisilicon/zip/zip.h struct hisi_zip *find_zip_device(int node); node 91 drivers/crypto/hisilicon/zip/zip_main.c static struct hisi_zip *find_zip_device_numa(int node) node 100 drivers/crypto/hisilicon/zip/zip_main.c if (node_distance(dev->numa_node, node) < min_distance) { node 102 drivers/crypto/hisilicon/zip/zip_main.c min_distance = node_distance(dev->numa_node, node); node 110 drivers/crypto/hisilicon/zip/zip_main.c struct hisi_zip *find_zip_device(int node) node 116 drivers/crypto/hisilicon/zip/zip_main.c zip = find_zip_device_numa(node); node 1725 drivers/crypto/n2_core.c u64 node, struct spu_queue *p, node 1730 drivers/crypto/n2_core.c mdesc_for_each_arc(arc, mdesc, node, MDESC_ARC_TYPE_BACK) { node 1752 drivers/crypto/n2_core.c u64 node, const char *iname, unsigned long q_type, node 1771 drivers/crypto/n2_core.c err = spu_mdesc_walk_arcs(mdesc, dev, node, p, table); node 1788 drivers/crypto/n2_core.c u64 node; node 1790 drivers/crypto/n2_core.c mdesc_for_each_node_by_name(mdesc, node, "exec-unit") { node 1793 drivers/crypto/n2_core.c type = mdesc_get_property(mdesc, node, "type", NULL); node 1797 drivers/crypto/n2_core.c err = handle_exec_unit(ip, list, dev, mdesc, node, node 1808 drivers/crypto/n2_core.c static int get_irq_props(struct mdesc_handle *mdesc, u64 node, node 1815 drivers/crypto/n2_core.c ino = mdesc_get_property(mdesc, node, "ino", &ino_len); node 1843 drivers/crypto/n2_core.c u64 node; node 1849 drivers/crypto/n2_core.c mdesc_for_each_node_by_name(mdesc, node, "virtual-device") { node 1853 drivers/crypto/n2_core.c name = mdesc_get_property(mdesc, node, "name", NULL); node 1856 drivers/crypto/n2_core.c chdl = mdesc_get_property(mdesc, node, "cfg-handle", NULL); node 1860 drivers/crypto/n2_core.c return get_irq_props(mdesc, node, ip); node 828 drivers/crypto/nx/nx-842-pseries.c struct device_node *node = NULL; node 833 drivers/crypto/nx/nx-842-pseries.c node = local_devdata->dev->of_node; node 837 drivers/crypto/nx/nx-842-pseries.c !strcmp(upd->dn->name, node->name)) { node 984 drivers/crypto/omap-aes.c struct device_node *node = dev->of_node; node 994 drivers/crypto/omap-aes.c err = of_address_to_resource(node, 0, res); node 1934 drivers/crypto/omap-sham.c struct device_node *node = dev->of_node; node 1944 drivers/crypto/omap-sham.c err = of_address_to_resource(node, 0, res); node 1951 drivers/crypto/omap-sham.c dd->irq = irq_of_parse_and_map(node, 0); node 164 drivers/crypto/qat/qat_common/adf_common_drv.h struct qat_crypto_instance *qat_crypto_get_instance_node(int node); node 617 drivers/crypto/qat/qat_common/qat_algs.c int node = get_current_node(); node 621 drivers/crypto/qat/qat_common/qat_algs.c inst = qat_crypto_get_instance_node(node); node 970 drivers/crypto/qat/qat_common/qat_algs.c int node = get_current_node(); node 973 drivers/crypto/qat/qat_common/qat_algs.c inst = qat_crypto_get_instance_node(node); node 95 drivers/crypto/qat/qat_common/qat_crypto.c struct qat_crypto_instance *qat_crypto_get_instance_node(int node) node 104 drivers/crypto/qat/qat_common/qat_crypto.c if ((node == dev_to_node(&GET_DEV(tmp_dev)) || node 117 drivers/crypto/qat/qat_common/qat_crypto.c pr_info("QAT: Could not find a device on node %d\n", node); node 314 drivers/crypto/virtio/virtio_crypto_algs.c int node = virtio_crypto_get_current_node(); node 316 drivers/crypto/virtio/virtio_crypto_algs.c virtcrypto_get_dev_node(node, node 110 drivers/crypto/virtio/virtio_crypto_common.h struct virtio_crypto *virtcrypto_get_dev_node(int node, node 123 drivers/crypto/virtio/virtio_crypto_common.h int cpu, node; node 126 drivers/crypto/virtio/virtio_crypto_common.h node = topology_physical_package_id(cpu); node 129 drivers/crypto/virtio/virtio_crypto_common.h return node; node 184 drivers/crypto/virtio/virtio_crypto_mgr.c struct virtio_crypto *virtcrypto_get_dev_node(int node, uint32_t service, node 194 drivers/crypto/virtio/virtio_crypto_mgr.c if ((node == dev_to_node(&tmp_dev->vdev->dev) || node 208 drivers/crypto/virtio/virtio_crypto_mgr.c node); node 58 drivers/dca/dca-core.c list_del(&domain->node); node 97 drivers/dca/dca-core.c domain = list_first_entry(&dca_domains, struct dca_domain, node); node 99 drivers/dca/dca-core.c list_for_each_entry_safe(dca, _dca, &domain->dca_providers, node) node 100 drivers/dca/dca-core.c list_move(&dca->node, &unregistered_providers); node 106 drivers/dca/dca-core.c list_for_each_entry_safe(dca, _dca, &unregistered_providers, node) { node 108 drivers/dca/dca-core.c list_del(&dca->node); node 116 drivers/dca/dca-core.c list_for_each_entry(domain, &dca_domains, node) node 154 drivers/dca/dca-core.c node); node 159 drivers/dca/dca-core.c list_for_each_entry(dca, &domain->dca_providers, node) node 197 drivers/dca/dca-core.c list_for_each_entry(dca, &domain->dca_providers, node) { node 380 drivers/dca/dca-core.c list_add(&domain->node, &dca_domains); node 383 drivers/dca/dca-core.c list_add(&dca->node, &domain->dca_providers); node 413 drivers/dca/dca-core.c list_del(&dca->node); node 223 drivers/devfreq/devfreq-event.c struct device_node *node; node 229 drivers/devfreq/devfreq-event.c node = of_parse_phandle(dev->of_node, "devfreq-events", index); node 230 drivers/devfreq/devfreq-event.c if (!node) node 234 drivers/devfreq/devfreq-event.c list_for_each_entry(edev, &devfreq_event_list, node) { node 235 drivers/devfreq/devfreq-event.c if (edev->dev.parent && edev->dev.parent->of_node == node) node 239 drivers/devfreq/devfreq-event.c list_for_each_entry(edev, &devfreq_event_list, node) { node 240 drivers/devfreq/devfreq-event.c if (of_node_name_eq(node, edev->desc->name)) node 248 drivers/devfreq/devfreq-event.c of_node_put(node); node 252 drivers/devfreq/devfreq-event.c of_node_put(node); node 337 drivers/devfreq/devfreq-event.c INIT_LIST_HEAD(&edev->node); node 340 drivers/devfreq/devfreq-event.c list_add(&edev->node, &devfreq_event_list); node 361 drivers/devfreq/devfreq-event.c list_del(&edev->node); node 65 drivers/devfreq/devfreq.c list_for_each_entry(tmp_devfreq, &devfreq_list, node) { node 215 drivers/devfreq/devfreq.c list_for_each_entry(tmp_governor, &devfreq_governor_list, node) { node 591 drivers/devfreq/devfreq.c list_del(&devfreq->node); node 644 drivers/devfreq/devfreq.c INIT_LIST_HEAD(&devfreq->node); node 733 drivers/devfreq/devfreq.c list_add(&devfreq->node, &devfreq_list); node 832 drivers/devfreq/devfreq.c struct device_node *node; node 841 drivers/devfreq/devfreq.c node = of_parse_phandle(dev->of_node, "devfreq", index); node 842 drivers/devfreq/devfreq.c if (!node) node 846 drivers/devfreq/devfreq.c list_for_each_entry(devfreq, &devfreq_list, node) { node 848 drivers/devfreq/devfreq.c && devfreq->dev.parent->of_node == node) { node 850 drivers/devfreq/devfreq.c of_node_put(node); node 855 drivers/devfreq/devfreq.c of_node_put(node); node 967 drivers/devfreq/devfreq.c list_for_each_entry(devfreq, &devfreq_list, node) { node 988 drivers/devfreq/devfreq.c list_for_each_entry(devfreq, &devfreq_list, node) { node 1021 drivers/devfreq/devfreq.c list_add(&governor->node, &devfreq_governor_list); node 1023 drivers/devfreq/devfreq.c list_for_each_entry(devfreq, &devfreq_list, node) { node 1085 drivers/devfreq/devfreq.c list_for_each_entry(devfreq, &devfreq_list, node) { node 1109 drivers/devfreq/devfreq.c list_del(&governor->node); node 1219 drivers/devfreq/devfreq.c list_for_each_entry(governor, &devfreq_governor_list, node) { node 503 drivers/devfreq/event/exynos-ppmu.c struct device_node *events_np, *node; node 528 drivers/devfreq/event/exynos-ppmu.c for_each_child_of_node(events_np, node) { node 533 drivers/devfreq/event/exynos-ppmu.c if (of_node_name_eq(node, ppmu_events[i].name)) node 540 drivers/devfreq/event/exynos-ppmu.c node); node 555 drivers/devfreq/event/exynos-ppmu.c of_property_read_string(node, "event-name", &desc[j].name); node 556 drivers/devfreq/event/exynos-ppmu.c ret = of_property_read_u32(node, "event-data-type", node 182 drivers/devfreq/event/rockchip-dfi.c struct device_node *np = pdev->dev.of_node, *node; node 200 drivers/devfreq/event/rockchip-dfi.c node = of_parse_phandle(np, "rockchip,pmu", 0); node 201 drivers/devfreq/event/rockchip-dfi.c if (node) { node 202 drivers/devfreq/event/rockchip-dfi.c data->regmap_pmu = syscon_node_to_regmap(node); node 293 drivers/devfreq/exynos-bus.c struct device_node *np = dev->of_node, *node; node 319 drivers/devfreq/exynos-bus.c node = of_parse_phandle(dev->of_node, "devfreq", 0); node 320 drivers/devfreq/exynos-bus.c if (node) { node 321 drivers/devfreq/exynos-bus.c of_node_put(node); node 48 drivers/devfreq/governor.h struct list_head node; node 310 drivers/devfreq/rk3399_dmc.c struct device_node *np = pdev->dev.of_node, *node; node 372 drivers/devfreq/rk3399_dmc.c node = of_parse_phandle(np, "rockchip,pmu", 0); node 373 drivers/devfreq/rk3399_dmc.c if (node) { node 374 drivers/devfreq/rk3399_dmc.c data->regmap_pmu = syscon_node_to_regmap(node); node 690 drivers/dma-buf/dma-buf.c list_add(&attach->node, &dmabuf->attachments); node 720 drivers/dma-buf/dma-buf.c list_del(&attach->node); node 1224 drivers/dma-buf/dma-buf.c list_for_each_entry(attach_obj, &buf_obj->attachments, node) { node 147 drivers/dma-buf/dma-fence.c list_for_each_entry_safe(cur, tmp, &cb_list, node) { node 148 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cur->node); node 340 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cb->node); node 362 drivers/dma-buf/dma-fence.c list_add_tail(&cb->node, &fence->cb_list); node 364 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cb->node); node 422 drivers/dma-buf/dma-fence.c ret = !list_empty(&cb->node); node 424 drivers/dma-buf/dma-fence.c list_del_init(&cb->node); node 498 drivers/dma-buf/dma-fence.c list_add(&cb.base.node, &fence->cb_list); node 514 drivers/dma-buf/dma-fence.c if (!list_empty(&cb.base.node)) node 515 drivers/dma-buf/dma-fence.c list_del(&cb.base.node); node 140 drivers/dma-buf/sw_sync.c rb_erase(&pt->node, &parent->pt_tree); node 207 drivers/dma-buf/sw_sync.c rb_erase(&pt->node, &obj->pt_tree); node 257 drivers/dma-buf/sw_sync.c other = rb_entry(parent, typeof(*pt), node); node 273 drivers/dma-buf/sw_sync.c rb_link_node(&pt->node, parent, p); node 274 drivers/dma-buf/sw_sync.c rb_insert_color(&pt->node, &obj->pt_tree); node 276 drivers/dma-buf/sw_sync.c parent = rb_next(&pt->node); node 278 drivers/dma-buf/sw_sync.c parent ? &rb_entry(parent, typeof(*pt), node)->link : &obj->pt_list); node 62 drivers/dma-buf/sync_debug.h struct rb_node node; node 37 drivers/dma-buf/sync_file.c INIT_LIST_HEAD(&sync_file->cb.node); node 313 drivers/dma-buf/sync_file.c if (list_empty(&sync_file->cb.node) && node 58 drivers/dma/acpi-dma.c list_for_each_entry(rentry, &resource_list, node) { node 161 drivers/dma/altera-msgdma.c struct list_head node; node 214 drivers/dma/altera-msgdma.c desc = list_first_entry(&mdev->free_list, struct msgdma_sw_desc, node); node 215 drivers/dma/altera-msgdma.c list_del(&desc->node); node 234 drivers/dma/altera-msgdma.c list_add_tail(&desc->node, &mdev->free_list); node 235 drivers/dma/altera-msgdma.c list_for_each_entry_safe(child, next, &desc->tx_list, node) { node 237 drivers/dma/altera-msgdma.c list_move_tail(&child->node, &mdev->free_list); node 251 drivers/dma/altera-msgdma.c list_for_each_entry_safe(desc, next, list, node) node 312 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &mdev->pending_list); node 364 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &first->tx_list); node 438 drivers/dma/altera-msgdma.c list_add_tail(&new->node, &first->tx_list); node 538 drivers/dma/altera-msgdma.c list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) node 554 drivers/dma/altera-msgdma.c struct msgdma_sw_desc, node); node 584 drivers/dma/altera-msgdma.c list_for_each_entry_safe(desc, next, &mdev->done_list, node) { node 588 drivers/dma/altera-msgdma.c list_del(&desc->node); node 612 drivers/dma/altera-msgdma.c struct msgdma_sw_desc, node); node 615 drivers/dma/altera-msgdma.c list_del(&desc->node); node 617 drivers/dma/altera-msgdma.c list_add_tail(&desc->node, &mdev->done_list); node 671 drivers/dma/altera-msgdma.c list_add_tail(&desc->node, &mdev->free_list); node 183 drivers/dma/amba-pl08x.c struct list_head node; node 535 drivers/dma/amba-pl08x.c list_del(&txd->vd.node); node 1284 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) { node 1500 drivers/dma/amba-pl08x.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { node 1501 drivers/dma/amba-pl08x.c list_del(&dsg->node); node 1583 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) node 1911 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); node 2035 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); node 112 drivers/dma/bcm-sba-raid.c struct list_head node; node 210 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_free_list, node) { node 212 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_alloc_list); node 248 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_pending_list); node 264 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_active_list); node 277 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_aborted_list); node 289 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_free_list); node 332 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_alloc_list, node) node 336 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_active_list, node) node 355 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_pending_list, node) node 398 drivers/dma/bcm-sba-raid.c struct sba_request, node); node 466 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_free_list, node) node 470 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_alloc_list, node) node 473 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_pending_list, node) node 476 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_active_list, node) node 479 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_aborted_list, node) node 1506 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&req->node); node 1524 drivers/dma/bcm-sba-raid.c list_add_tail(&req->node, &sba->reqs_free_list); node 450 drivers/dma/bcm2835-dma.c list_del(&vd->node); node 1270 drivers/dma/coh901318.c struct list_head node; node 1525 drivers/dma/coh901318.c INIT_LIST_HEAD(&desc->node); node 1531 drivers/dma/coh901318.c node); node 1532 drivers/dma/coh901318.c list_del(&desc->node); node 1547 drivers/dma/coh901318.c list_add_tail(&cohd->node, &cohc->free); node 1554 drivers/dma/coh901318.c list_add_tail(&desc->node, &cohc->active); node 1561 drivers/dma/coh901318.c node); node 1567 drivers/dma/coh901318.c list_del(&cohd->node); node 1573 drivers/dma/coh901318.c list_add_tail(&desc->node, &cohc->queue); node 1580 drivers/dma/coh901318.c node); node 1624 drivers/dma/coh901318.c cohd = list_entry(pos, struct coh901318_desc, node); node 1670 drivers/dma/coh901318.c cohd = list_entry(pos, struct coh901318_desc, node); node 213 drivers/dma/dma-axi-dmac.c list_move_tail(&vdesc->node, &chan->active_descs); node 269 drivers/dma/dma-axi-dmac.c struct axi_dmac_desc, vdesc.node); node 294 drivers/dma/dma-axi-dmac.c list_for_each_entry(desc, &chan->active_descs, vdesc.node) { node 388 drivers/dma/dma-axi-dmac.c list_del(&active->vdesc.node); node 490 drivers/dma/dma-jz4780.c list_del(&vdesc->node); node 377 drivers/dma/dmaengine.c int node = dev_to_node(chan->device->dev); node 378 drivers/dma/dmaengine.c return node == NUMA_NO_NODE || node 379 drivers/dma/dmaengine.c cpumask_test_cpu(cpu, cpumask_of_node(node)); node 213 drivers/dma/dmatest.c struct list_head node; node 227 drivers/dma/dmatest.c struct list_head node; node 239 drivers/dma/dmatest.c list_for_each_entry(dtc, &info->channels, node) { node 242 drivers/dma/dmatest.c list_for_each_entry(thread, &dtc->threads, node) { node 255 drivers/dma/dmatest.c list_for_each_entry(dtc, &info->channels, node) { node 258 drivers/dma/dmatest.c list_for_each_entry(thread, &dtc->threads, node) { node 919 drivers/dma/dmatest.c list_for_each_entry_safe(thread, _thread, &dtc->threads, node) { node 923 drivers/dma/dmatest.c list_del(&thread->node); node 978 drivers/dma/dmatest.c list_add_tail(&thread->node, &dtc->threads); node 1028 drivers/dma/dmatest.c list_add_tail(&dtc->node, &info->channels); node 1101 drivers/dma/dmatest.c list_for_each_entry(dtc, &info->channels, node) { node 1105 drivers/dma/dmatest.c list_for_each_entry(thread, &dtc->threads, node) { node 1119 drivers/dma/dmatest.c list_for_each_entry_safe(dtc, _dtc, &info->channels, node) { node 1120 drivers/dma/dmatest.c list_del(&dtc->node); node 1201 drivers/dma/dmatest.c list_for_each_entry(dtc, &info->channels, node) { node 1206 drivers/dma/dmatest.c node); node 1219 drivers/dma/dmatest.c dtc = list_last_entry(&info->channels, struct dmatest_chan, node); node 1274 drivers/dma/dmatest.c list_for_each_entry(dtc, &info->channels, node) { node 1278 drivers/dma/dmatest.c list_for_each_entry(thread, &dtc->threads, node) { node 555 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c list_del(&vd->node); node 586 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c list_del(&vd->node); node 489 drivers/dma/dw-edma/dw-edma-core.c list_del(&vd->node); node 496 drivers/dma/dw-edma/dw-edma-core.c list_del(&vd->node); node 524 drivers/dma/dw-edma/dw-edma-core.c list_del(&vd->node); node 129 drivers/dma/ep93xx_dma.c struct list_head node; node 243 drivers/dma/ep93xx_dma.c list_add_tail(&desc->node, &edmac->active); node 248 drivers/dma/ep93xx_dma.c struct ep93xx_dma_desc, node); node 259 drivers/dma/ep93xx_dma.c list_move_tail(&d->node, &edmac->active); node 268 drivers/dma/ep93xx_dma.c struct ep93xx_dma_desc, node); node 682 drivers/dma/ep93xx_dma.c list_for_each_entry_safe(desc, _desc, &edmac->free_list, node) { node 684 drivers/dma/ep93xx_dma.c list_del_init(&desc->node); node 711 drivers/dma/ep93xx_dma.c list_add(&desc->node, &edmac->free_list); node 736 drivers/dma/ep93xx_dma.c new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node); node 737 drivers/dma/ep93xx_dma.c list_del_init(&new->node); node 776 drivers/dma/ep93xx_dma.c list_for_each_entry_safe(desc, d, &list, node) { node 850 drivers/dma/ep93xx_dma.c list_add_tail(&desc->node, &edmac->queue); node 967 drivers/dma/ep93xx_dma.c list_for_each_entry_safe(desc, d, &list, node) node 1009 drivers/dma/ep93xx_dma.c list_add_tail(&desc->node, &first->tx_list); node 1084 drivers/dma/ep93xx_dma.c list_add_tail(&desc->node, &first->tx_list); node 1165 drivers/dma/ep93xx_dma.c list_add_tail(&desc->node, &first->tx_list); node 1224 drivers/dma/ep93xx_dma.c list_for_each_entry_safe(desc, _d, &list, node) node 49 drivers/dma/fsl-edma.c list_del(&fsl_chan->edesc->vdesc.node); node 951 drivers/dma/fsl-qdma.c list_del(&vdesc->node); node 98 drivers/dma/fsl_raid.c list_add_tail(&desc->node, &re_chan->submit_q); node 118 drivers/dma/fsl_raid.c list_for_each_entry_safe(desc, _desc, &re_chan->submit_q, node) { node 122 drivers/dma/fsl_raid.c list_move_tail(&desc->node, &re_chan->active_q); node 148 drivers/dma/fsl_raid.c list_for_each_entry_safe(desc, _desc, &re_chan->ack_q, node) { node 150 drivers/dma/fsl_raid.c list_move_tail(&desc->node, &re_chan->free_q); node 176 drivers/dma/fsl_raid.c node) { node 187 drivers/dma/fsl_raid.c list_move_tail(&desc->node, &re_chan->ack_q); node 258 drivers/dma/fsl_raid.c INIT_LIST_HEAD(&desc->node); node 286 drivers/dma/fsl_raid.c struct fsl_re_desc, node); node 287 drivers/dma/fsl_raid.c list_del(&desc->node); node 594 drivers/dma/fsl_raid.c INIT_LIST_HEAD(&desc->node); node 597 drivers/dma/fsl_raid.c list_add_tail(&desc->node, &re_chan->free_q); node 612 drivers/dma/fsl_raid.c node); node 614 drivers/dma/fsl_raid.c list_del(&desc->node); node 295 drivers/dma/fsl_raid.h struct list_head node; node 427 drivers/dma/fsldma.c list_for_each_entry(child, &desc->tx_list, node) { node 447 drivers/dma/fsldma.c list_del(&desc->node); node 492 drivers/dma/fsldma.c list_for_each_entry_safe(desc, _desc, &chan->ld_completed, node) node 541 drivers/dma/fsldma.c list_del(&desc->node); node 552 drivers/dma/fsldma.c list_add_tail(&desc->node, &chan->ld_completed); node 599 drivers/dma/fsldma.c desc = list_first_entry(&chan->ld_pending, struct fsl_desc_sw, node); node 645 drivers/dma/fsldma.c list_for_each_entry_safe(desc, _desc, &chan->ld_running, node) { node 725 drivers/dma/fsldma.c list_for_each_entry_safe(desc, _desc, list, node) node 734 drivers/dma/fsldma.c list_for_each_entry_safe_reverse(desc, _desc, list, node) node 804 drivers/dma/fsldma.c list_add_tail(&new->node, &first->tx_list); node 1106 drivers/dma/fsldma.c struct device_node *node, u32 feature, const char *compatible) node 1120 drivers/dma/fsldma.c chan->regs = of_iomap(node, 0); node 1127 drivers/dma/fsldma.c err = of_address_to_resource(node, 0, &res); node 1187 drivers/dma/fsldma.c chan->irq = irq_of_parse_and_map(node, 0); node 102 drivers/dma/fsldma.h struct list_head node; node 191 drivers/dma/fsldma.h #define to_fsl_desc(lh) container_of(lh, struct fsl_desc_sw, node) node 119 drivers/dma/hsu/hsu.c list_del(&vdesc->node); node 124 drivers/dma/idma64.c list_del(&vdesc->node); node 535 drivers/dma/img-mdc-dma.c list_del(&vd->node); node 121 drivers/dma/imx-dma.c struct list_head node; node 248 drivers/dma/imx-dma.c node); node 442 drivers/dma/imx-dma.c node); node 630 drivers/dma/imx-dma.c desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, node); node 651 drivers/dma/imx-dma.c struct imxdma_desc, node); node 780 drivers/dma/imx-dma.c list_add_tail(&desc->node, &imxdmac->ld_free); node 805 drivers/dma/imx-dma.c list_for_each_entry_safe(desc, _desc, &imxdmac->ld_free, node) { node 829 drivers/dma/imx-dma.c desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); node 884 drivers/dma/imx-dma.c desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); node 940 drivers/dma/imx-dma.c desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); node 977 drivers/dma/imx-dma.c desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); node 1010 drivers/dma/imx-dma.c struct imxdma_desc, node); node 768 drivers/dma/imx-sdma.c list_del(&vd->node); node 84 drivers/dma/k3dma.c struct list_head node; node 287 drivers/dma/k3dma.c list_del(&ds->vd.node); node 333 drivers/dma/k3dma.c struct k3_dma_chan, node); node 335 drivers/dma/k3dma.c list_del_init(&c->node); node 368 drivers/dma/k3dma.c list_del_init(&c->node); node 434 drivers/dma/k3dma.c if (list_empty(&c->node)) { node 436 drivers/dma/k3dma.c list_add_tail(&c->node, &d->chan_pending); node 733 drivers/dma/k3dma.c list_del_init(&c->node); node 776 drivers/dma/k3dma.c list_del_init(&c->node); node 799 drivers/dma/k3dma.c list_add_tail(&c->node, &d->chan_pending); node 945 drivers/dma/k3dma.c INIT_LIST_HEAD(&c->node); node 39 drivers/dma/mcf-edma.c list_del(&mcf_chan->edesc->vdesc.node); node 87 drivers/dma/mediatek/mtk-cqdma.c struct list_head node; node 276 drivers/dma/mediatek/mtk-cqdma.c list_for_each_entry_safe(vd, vd2, &cvc->vc.desc_issued, node) { node 284 drivers/dma/mediatek/mtk-cqdma.c list_add_tail(&cvd->node, &pc->queue); node 291 drivers/dma/mediatek/mtk-cqdma.c list_del(&vd->node); node 303 drivers/dma/mediatek/mtk-cqdma.c list_for_each_entry(cvd, &cvc->pc->queue, node) node 321 drivers/dma/mediatek/mtk-cqdma.c struct mtk_cqdma_vdesc, node); node 332 drivers/dma/mediatek/mtk-cqdma.c list_del(&cvd->node); node 352 drivers/dma/mediatek/mtk-cqdma.c struct mtk_cqdma_vdesc, node); node 427 drivers/dma/mediatek/mtk-cqdma.c list_for_each_entry(vd, &cvc->pc->queue, node) node 434 drivers/dma/mediatek/mtk-cqdma.c list_for_each_entry(vd, &cvc->vc.desc_issued, node) node 503 drivers/dma/mediatek/mtk-hsdma.c list_for_each_entry_safe(vd, vd2, &hvc->vc.desc_issued, node) { node 529 drivers/dma/mediatek/mtk-hsdma.c list_move_tail(&vd->node, &hvc->desc_hw_processing); node 589 drivers/dma/mediatek/mtk-hsdma.c list_del(&cb->vd->node); node 666 drivers/dma/mediatek/mtk-hsdma.c list_for_each_entry(vd, &hvc->desc_hw_processing, node) node 670 drivers/dma/mediatek/mtk-hsdma.c list_for_each_entry(vd, &hvc->vc.desc_issued, node) node 216 drivers/dma/mediatek/mtk-uart-apdma.c list_del(&d->vd.node); node 249 drivers/dma/mediatek/mtk-uart-apdma.c list_del(&d->vd.node); node 83 drivers/dma/mmp_pdma.c struct list_head node; node 135 drivers/dma/mmp_pdma.c container_of(lh, struct mmp_pdma_desc_sw, node) node 327 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw, node); node 351 drivers/dma/mmp_pdma.c list_for_each_entry(child, &desc->tx_list, node) { node 420 drivers/dma/mmp_pdma.c list_for_each_entry_safe(desc, _desc, list, node) { node 421 drivers/dma/mmp_pdma.c list_del(&desc->node); node 505 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); node 579 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); node 674 drivers/dma/mmp_pdma.c list_add_tail(&new->node, &first->tx_list); node 789 drivers/dma/mmp_pdma.c list_for_each_entry(sw, &chan->chain_running, node) { node 898 drivers/dma/mmp_pdma.c list_for_each_entry_safe(desc, _desc, &chan->chain_running, node) { node 903 drivers/dma/mmp_pdma.c list_move(&desc->node, &chain_cleanup); node 929 drivers/dma/mmp_pdma.c list_for_each_entry_safe(desc, _desc, &chain_cleanup, node) { node 933 drivers/dma/mmp_pdma.c list_del(&desc->node); node 431 drivers/dma/moxart-dma.c list_del(&vd->node); node 569 drivers/dma/moxart-dma.c struct device_node *node = dev->of_node; node 581 drivers/dma/moxart-dma.c irq = irq_of_parse_and_map(node, 0); node 627 drivers/dma/moxart-dma.c ret = of_dma_controller_register(node, moxart_of_xlate, mdc); node 189 drivers/dma/mpc512x_dma.c struct list_head node; node 264 drivers/dma/mpc512x_dma.c struct mpc_dma_desc, node); node 273 drivers/dma/mpc512x_dma.c list_move_tail(&mdesc->node, &mchan->active); node 276 drivers/dma/mpc512x_dma.c list_move_tail(&mdesc->node, &mchan->active); node 281 drivers/dma/mpc512x_dma.c list_for_each_entry(mdesc, &mchan->active, node) { node 336 drivers/dma/mpc512x_dma.c list_for_each_entry(mdesc, &mchan->active, node) node 399 drivers/dma/mpc512x_dma.c list_for_each_entry(mdesc, &list, node) { node 472 drivers/dma/mpc512x_dma.c list_move_tail(&mdesc->node, &mchan->queued); node 520 drivers/dma/mpc512x_dma.c list_add_tail(&mdesc->node, &descs); node 575 drivers/dma/mpc512x_dma.c list_for_each_entry_safe(mdesc, tmp, &descs, node) node 614 drivers/dma/mpc512x_dma.c node); node 615 drivers/dma/mpc512x_dma.c list_del(&mdesc->node); node 668 drivers/dma/mpc512x_dma.c list_add_tail(&mdesc->node, &mchan->prepared); node 710 drivers/dma/mpc512x_dma.c struct mpc_dma_desc, node); node 718 drivers/dma/mpc512x_dma.c list_del(&mdesc->node); node 795 drivers/dma/mpc512x_dma.c list_add_tail(&mdesc->node, &mchan->prepared); node 804 drivers/dma/mpc512x_dma.c list_add_tail(&mdesc->node, &mchan->free); node 221 drivers/dma/mv_xor.c node) { node 224 drivers/dma/mv_xor.c list_move_tail(&iter->node, &mv_chan->free_slots); node 246 drivers/dma/mv_xor.c list_move_tail(&desc->node, &mv_chan->completed_slots); node 252 drivers/dma/mv_xor.c list_move_tail(&desc->node, &mv_chan->free_slots); node 281 drivers/dma/mv_xor.c node) { node 313 drivers/dma/mv_xor.c node); node 316 drivers/dma/mv_xor.c if (!list_is_last(&iter->node, &mv_chan->chain)) { node 321 drivers/dma/mv_xor.c iter = list_entry(iter->node.next, node 323 drivers/dma/mv_xor.c node); node 358 drivers/dma/mv_xor.c node); node 360 drivers/dma/mv_xor.c list_move_tail(&iter->node, &mv_chan->allocated_slots); node 398 drivers/dma/mv_xor.c list_move_tail(&sw_desc->node, &mv_chan->chain); node 404 drivers/dma/mv_xor.c node); node 405 drivers/dma/mv_xor.c list_move_tail(&sw_desc->node, &mv_chan->chain); node 458 drivers/dma/mv_xor.c INIT_LIST_HEAD(&slot->node); node 466 drivers/dma/mv_xor.c list_add_tail(&slot->node, &mv_chan->free_slots); node 638 drivers/dma/mv_xor.c node) { node 640 drivers/dma/mv_xor.c list_move_tail(&iter->node, &mv_chan->free_slots); node 643 drivers/dma/mv_xor.c node) { node 645 drivers/dma/mv_xor.c list_move_tail(&iter->node, &mv_chan->free_slots); node 648 drivers/dma/mv_xor.c node) { node 650 drivers/dma/mv_xor.c list_move_tail(&iter->node, &mv_chan->free_slots); node 653 drivers/dma/mv_xor.c iter, _iter, &mv_chan->free_slots, node) { node 654 drivers/dma/mv_xor.c list_del(&iter->node); node 142 drivers/dma/mv_xor.h struct list_head node; node 139 drivers/dma/nbpfaxi.c struct list_head node; node 156 drivers/dma/nbpfaxi.c struct list_head node; node 168 drivers/dma/nbpfaxi.c struct list_head node; node 384 drivers/dma/nbpfaxi.c struct nbpf_link_desc *ldesc = list_first_entry(&desc->sg, struct nbpf_link_desc, node); node 610 drivers/dma/nbpfaxi.c struct nbpf_desc, node); node 640 drivers/dma/nbpfaxi.c list_for_each_entry(desc, &chan->active, node) node 647 drivers/dma/nbpfaxi.c list_for_each_entry(desc, &chan->queued, node) node 675 drivers/dma/nbpfaxi.c list_add_tail(&desc->node, &chan->queued); node 705 drivers/dma/nbpfaxi.c list_add_tail(&ldesc->node, &lhead); node 720 drivers/dma/nbpfaxi.c list_add_tail(&desc->node, &head); node 730 drivers/dma/nbpfaxi.c list_add(&dpage->node, &chan->desc_page); node 743 drivers/dma/nbpfaxi.c list_for_each_entry_safe(ldesc, tmp, &desc->sg, node) node 744 drivers/dma/nbpfaxi.c list_move(&ldesc->node, &chan->free_links); node 746 drivers/dma/nbpfaxi.c list_add(&desc->node, &chan->free); node 757 drivers/dma/nbpfaxi.c list_for_each_entry_safe(desc, tmp, &chan->done, node) node 759 drivers/dma/nbpfaxi.c list_move(&desc->node, &head); node 764 drivers/dma/nbpfaxi.c list_for_each_entry_safe(desc, tmp, &head, node) { node 765 drivers/dma/nbpfaxi.c list_del(&desc->node); node 797 drivers/dma/nbpfaxi.c desc = list_first_entry(&chan->free, struct nbpf_desc, node); node 798 drivers/dma/nbpfaxi.c list_del(&desc->node); node 814 drivers/dma/nbpfaxi.c struct nbpf_link_desc, node); node 820 drivers/dma/nbpfaxi.c list_move_tail(&ldesc->node, &desc->sg); node 849 drivers/dma/nbpfaxi.c list_for_each_entry_safe(desc, tmp, &head, node) { node 852 drivers/dma/nbpfaxi.c list_del(&desc->node); node 957 drivers/dma/nbpfaxi.c list_for_each_entry(ldesc, &desc->sg, node) { node 1068 drivers/dma/nbpfaxi.c list_for_each_entry_safe(dpage, tmp, &chan->desc_page, node) { node 1071 drivers/dma/nbpfaxi.c list_del(&dpage->node); node 1120 drivers/dma/nbpfaxi.c list_for_each_entry_safe(desc, tmp, &chan->done, node) { node 1130 drivers/dma/nbpfaxi.c list_del(&desc->node); node 1154 drivers/dma/nbpfaxi.c list_del(&desc->node); node 1198 drivers/dma/nbpfaxi.c list_move_tail(&desc->node, &chan->done); node 1203 drivers/dma/nbpfaxi.c struct nbpf_desc, node); node 158 drivers/dma/owl-dma.c struct list_head node; node 324 drivers/dma/owl-dma.c list_del(&lli->node); node 337 drivers/dma/owl-dma.c INIT_LIST_HEAD(&lli->node); node 349 drivers/dma/owl-dma.c list_add_tail(&next->node, &txd->lli_list); node 505 drivers/dma/owl-dma.c list_del(&vd->node); node 514 drivers/dma/owl-dma.c struct owl_dma_lli, node); node 643 drivers/dma/owl-dma.c list_for_each_entry_safe(lli, _lli, &txd->lli_list, node) node 750 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) { node 753 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) node 784 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) node 517 drivers/dma/pl330.c struct list_head node; node 2035 drivers/dma/pl330.c list_for_each_entry(desc, &pch->work_list, node) { node 2067 drivers/dma/pl330.c list_for_each_entry_safe(desc, _dt, &pch->work_list, node) node 2071 drivers/dma/pl330.c list_move_tail(&desc->node, &pch->completed_list); node 2094 drivers/dma/pl330.c struct dma_pl330_desc, node); node 2100 drivers/dma/pl330.c list_move_tail(&desc->node, &pch->work_list); node 2110 drivers/dma/pl330.c list_move_tail(&desc->node, &pch->dmac->desc_pool); node 2292 drivers/dma/pl330.c list_for_each_entry(desc, &pch->submitted_list, node) { node 2297 drivers/dma/pl330.c list_for_each_entry(desc, &pch->work_list , node) { node 2418 drivers/dma/pl330.c list_for_each_entry(desc, &pch->work_list, node) { node 2499 drivers/dma/pl330.c while (!list_empty(&last->node)) { node 2500 drivers/dma/pl330.c desc = list_entry(last->node.next, struct dma_pl330_desc, node); node 2509 drivers/dma/pl330.c list_move_tail(&desc->node, &pch->submitted_list); node 2514 drivers/dma/pl330.c list_add_tail(&last->node, &pch->submitted_list); node 2527 drivers/dma/pl330.c INIT_LIST_HEAD(&desc->node); node 2546 drivers/dma/pl330.c list_add_tail(&desc[i].node, pool); node 2564 drivers/dma/pl330.c struct dma_pl330_desc, node); node 2566 drivers/dma/pl330.c list_del_init(&desc->node); node 2701 drivers/dma/pl330.c while (!list_empty(&first->node)) { node 2702 drivers/dma/pl330.c desc = list_entry(first->node.next, node 2703 drivers/dma/pl330.c struct dma_pl330_desc, node); node 2704 drivers/dma/pl330.c list_move_tail(&desc->node, &pl330->desc_pool); node 2707 drivers/dma/pl330.c list_move_tail(&first->node, &pl330->desc_pool); node 2740 drivers/dma/pl330.c list_add_tail(&desc->node, &first->node); node 2817 drivers/dma/pl330.c while (!list_empty(&first->node)) { node 2818 drivers/dma/pl330.c desc = list_entry(first->node.next, node 2819 drivers/dma/pl330.c struct dma_pl330_desc, node); node 2820 drivers/dma/pl330.c list_move_tail(&desc->node, &pl330->desc_pool); node 2823 drivers/dma/pl330.c list_move_tail(&first->node, &pl330->desc_pool); node 2865 drivers/dma/pl330.c list_add_tail(&desc->node, &first->node); node 68 drivers/dma/ppc4xx/adma.c struct list_head node; node 1401 drivers/dma/ppc4xx/adma.c list_for_each_entry(ref, &ppc440spe_adma_chan_list, node) { node 4171 drivers/dma/ppc4xx/adma.c INIT_LIST_HEAD(&ref->node); node 4172 drivers/dma/ppc4xx/adma.c list_add_tail(&ref->node, &ppc440spe_adma_chan_list); node 4197 drivers/dma/ppc4xx/adma.c list_for_each_entry_safe(ref, _ref, &ppc440spe_adma_chan_list, node) { node 4199 drivers/dma/ppc4xx/adma.c list_del(&ref->node); node 4265 drivers/dma/ppc4xx/adma.c node) { node 4268 drivers/dma/ppc4xx/adma.c list_del(&ref->node); node 575 drivers/dma/pxa_dma.c struct virt_dma_desc, node); node 620 drivers/dma/pxa_dma.c list_for_each_entry_safe(vd, tmp, &chan->vc.desc_issued, node) { node 632 drivers/dma/pxa_dma.c list_del(&vd->node); node 657 drivers/dma/pxa_dma.c struct virt_dma_desc, node); node 791 drivers/dma/pxa_dma.c list_move_tail(&vd->node, &vc->desc_issued); node 803 drivers/dma/pxa_dma.c struct virt_dma_desc, node); node 818 drivers/dma/pxa_dma.c list_move_tail(&vd->node, &vc->desc_submitted); node 837 drivers/dma/pxa_dma.c struct virt_dma_desc, node); node 855 drivers/dma/pxa_dma.c INIT_LIST_HEAD(&vd->node); node 1100 drivers/dma/pxa_dma.c list_for_each_entry(vd, &head, node) { node 372 drivers/dma/qcom/bam_dma.c struct list_head node; node 718 drivers/dma/qcom/bam_dma.c list_add(&async_desc->vd.node, &bchan->vc.desc_issued); node 846 drivers/dma/qcom/bam_dma.c list_add(&async_desc->vd.node, node 1000 drivers/dma/qcom/bam_dma.c list_del(&vd->node); node 133 drivers/dma/qcom/hidma.c list_for_each_entry_safe(mdesc, next, &list, node) { node 159 drivers/dma/qcom/hidma.c list_move(&mdesc->node, &mchan->free); node 181 drivers/dma/qcom/hidma.c if (mdesc->node.next) { node 183 drivers/dma/qcom/hidma.c list_move_tail(&mdesc->node, &mchan->completed); node 188 drivers/dma/qcom/hidma.c struct hidma_desc, node); node 244 drivers/dma/qcom/hidma.c list_for_each_entry_safe(qdesc, next, &mchan->queued, node) { node 246 drivers/dma/qcom/hidma.c list_move_tail(&qdesc->node, &mchan->active); node 252 drivers/dma/qcom/hidma.c node); node 338 drivers/dma/qcom/hidma.c list_move_tail(&mdesc->node, &mchan->queued); node 380 drivers/dma/qcom/hidma.c list_add_tail(&mdesc->node, &descs); node 385 drivers/dma/qcom/hidma.c list_for_each_entry_safe(mdesc, tmp, &descs, node) { node 411 drivers/dma/qcom/hidma.c mdesc = list_first_entry(&mchan->free, struct hidma_desc, node); node 412 drivers/dma/qcom/hidma.c list_del(&mdesc->node); node 426 drivers/dma/qcom/hidma.c list_add_tail(&mdesc->node, &mchan->prepared); node 444 drivers/dma/qcom/hidma.c mdesc = list_first_entry(&mchan->free, struct hidma_desc, node); node 445 drivers/dma/qcom/hidma.c list_del(&mdesc->node); node 459 drivers/dma/qcom/hidma.c list_add_tail(&mdesc->node, &mchan->prepared); node 494 drivers/dma/qcom/hidma.c list_for_each_entry_safe(mdesc, tmp, &list, node) { node 502 drivers/dma/qcom/hidma.c list_move(&mdesc->node, &mchan->free); node 547 drivers/dma/qcom/hidma.c list_for_each_entry_safe(mdesc, tmp, &descs, node) { node 549 drivers/dma/qcom/hidma.c list_del(&mdesc->node); node 80 drivers/dma/qcom/hidma.h struct list_head node; node 94 drivers/dma/qcom/hidma_dbg.c list_for_each_entry(mdesc, &mchan->prepared, node) node 98 drivers/dma/qcom/hidma_dbg.c list_for_each_entry(mdesc, &mchan->active, node) node 102 drivers/dma/qcom/hidma_dbg.c list_for_each_entry(mdesc, &mchan->completed, node) node 152 drivers/dma/s3c24xx-dma.c struct list_head node; node 431 drivers/dma/s3c24xx-dma.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { node 432 drivers/dma/s3c24xx-dma.c list_del(&dsg->node); node 445 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg = list_entry(txd->at, struct s3c24xx_sg, node); node 509 drivers/dma/s3c24xx-dma.c list_del(&txd->vd.node); node 785 drivers/dma/s3c24xx-dma.c list_for_each_entry(dsg, &txd->dsg_list, node) node 794 drivers/dma/s3c24xx-dma.c dsg = list_entry(txd->at, struct s3c24xx_sg, node); node 795 drivers/dma/s3c24xx-dma.c list_for_each_entry_from(dsg, &txd->dsg_list, node) node 842 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); node 956 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); node 1048 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); node 94 drivers/dma/sa11x0-dma.c struct list_head node; node 150 drivers/dma/sa11x0-dma.c list_del(&txd->vd.node); node 358 drivers/dma/sa11x0-dma.c struct sa11x0_dma_chan, node); node 359 drivers/dma/sa11x0-dma.c list_del_init(&c->node); node 395 drivers/dma/sa11x0-dma.c list_del_init(&c->node); node 510 drivers/dma/sa11x0-dma.c if (list_empty(&c->node)) { node 511 drivers/dma/sa11x0-dma.c list_add_tail(&c->node, &d->chan_pending); node 717 drivers/dma/sa11x0-dma.c list_del_init(&c->node); node 743 drivers/dma/sa11x0-dma.c list_add_tail(&c->node, &d->chan_pending); node 776 drivers/dma/sa11x0-dma.c list_add_tail(&p->txd_load->vd.node, &head); node 780 drivers/dma/sa11x0-dma.c list_add_tail(&p->txd_done->vd.node, &head); node 862 drivers/dma/sa11x0-dma.c INIT_LIST_HEAD(&c->node); node 35 drivers/dma/sh/rcar-dmac.c struct list_head node; node 78 drivers/dma/sh/rcar-dmac.c struct list_head node; node 103 drivers/dma/sh/rcar-dmac.c struct list_head node; node 352 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_xfer_chunk, node); node 474 drivers/dma/sh/rcar-dmac.c list_add_tail(&desc->node, &chan->desc.pending); node 476 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_xfer_chunk, node); node 510 drivers/dma/sh/rcar-dmac.c list_add_tail(&desc->node, &list); node 515 drivers/dma/sh/rcar-dmac.c list_add_tail(&page->node, &chan->desc.pages); node 540 drivers/dma/sh/rcar-dmac.c list_add(&desc->node, &chan->desc.free); node 560 drivers/dma/sh/rcar-dmac.c list_for_each_entry_safe(desc, _desc, &list, node) { node 562 drivers/dma/sh/rcar-dmac.c list_del(&desc->node); node 610 drivers/dma/sh/rcar-dmac.c desc = list_first_entry(&chan->desc.free, struct rcar_dmac_desc, node); node 611 drivers/dma/sh/rcar-dmac.c list_del(&desc->node); node 637 drivers/dma/sh/rcar-dmac.c list_add_tail(&chunk->node, &list); node 642 drivers/dma/sh/rcar-dmac.c list_add_tail(&page->node, &chan->desc.pages); node 681 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_xfer_chunk, node); node 682 drivers/dma/sh/rcar-dmac.c list_del(&chunk->node); node 733 drivers/dma/sh/rcar-dmac.c list_for_each_entry(chunk, &desc->chunks, node) { node 805 drivers/dma/sh/rcar-dmac.c list_for_each_entry_safe(desc, _desc, &descs, node) { node 806 drivers/dma/sh/rcar-dmac.c list_del(&desc->node); node 991 drivers/dma/sh/rcar-dmac.c list_add_tail(&chunk->node, &desc->chunks); node 1074 drivers/dma/sh/rcar-dmac.c list_for_each_entry(desc, &list, node) node 1077 drivers/dma/sh/rcar-dmac.c list_for_each_entry_safe(page, _page, &rchan->desc.pages, node) { node 1078 drivers/dma/sh/rcar-dmac.c list_del(&page->node); node 1318 drivers/dma/sh/rcar-dmac.c list_for_each_entry(desc, &chan->desc.done, node) { node 1322 drivers/dma/sh/rcar-dmac.c list_for_each_entry(desc, &chan->desc.pending, node) { node 1326 drivers/dma/sh/rcar-dmac.c list_for_each_entry(desc, &chan->desc.active, node) { node 1375 drivers/dma/sh/rcar-dmac.c list_for_each_entry_reverse(chunk, &desc->chunks, node) { node 1437 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_desc, node); node 1505 drivers/dma/sh/rcar-dmac.c if (!list_is_last(&desc->running->node, &desc->chunks)) { node 1506 drivers/dma/sh/rcar-dmac.c desc->running = list_next_entry(desc->running, node); node 1520 drivers/dma/sh/rcar-dmac.c node); node 1526 drivers/dma/sh/rcar-dmac.c list_move_tail(&desc->node, &chan->desc.done); node 1532 drivers/dma/sh/rcar-dmac.c node); node 1619 drivers/dma/sh/rcar-dmac.c node); node 1621 drivers/dma/sh/rcar-dmac.c list_del(&desc->node); node 1635 drivers/dma/sh/rcar-dmac.c list_add_tail(&desc->node, &chan->desc.wait); node 63 drivers/dma/sh/shdma-base.c list_for_each_entry(sdesc, &schan->ld_queue, node) node 86 drivers/dma/sh/shdma-base.c list_for_each_entry_safe(chunk, c, desc->node.prev, node) { node 94 drivers/dma/sh/shdma-base.c &chunk->node == &schan->ld_free)) node 105 drivers/dma/sh/shdma-base.c list_move_tail(&chunk->node, &schan->ld_queue); node 161 drivers/dma/sh/shdma-base.c list_for_each_entry(sdesc, &schan->ld_free, node) node 164 drivers/dma/sh/shdma-base.c list_del(&sdesc->node); node 242 drivers/dma/sh/shdma-base.c list_add(&desc->node, &schan->ld_free); node 336 drivers/dma/sh/shdma-base.c list_for_each_entry_safe(desc, _desc, &schan->ld_queue, node) { node 403 drivers/dma/sh/shdma-base.c list_move(&desc->node, &schan->ld_free); node 407 drivers/dma/sh/shdma-base.c list_move_tail(&desc->node, &cyclic_list); node 615 drivers/dma/sh/shdma-base.c list_add_tail(&new->node, &tx_list); node 630 drivers/dma/sh/shdma-base.c list_for_each_entry(new, &tx_list, node) node 766 drivers/dma/sh/shdma-base.c struct shdma_desc, node); node 839 drivers/dma/sh/shdma-base.c list_for_each_entry(sdesc, &schan->ld_queue, node) node 883 drivers/dma/sh/shdma-base.c list_for_each_entry(sdesc, &dl, node) { node 925 drivers/dma/sh/shdma-base.c list_for_each_entry(sdesc, &schan->ld_queue, node) { node 37 drivers/dma/sh/shdma.h struct list_head node; node 56 drivers/dma/sh/shdma.h #define to_sh_desc(lh) container_of(lh, struct sh_desc, node) node 495 drivers/dma/sh/shdmac.c list_for_each_entry_rcu(shdev, &sh_dmae_devices, node) { node 772 drivers/dma/sh/shdmac.c list_add_tail_rcu(&shdev->node, &sh_dmae_devices); node 874 drivers/dma/sh/shdmac.c list_del_rcu(&shdev->node); node 895 drivers/dma/sh/shdmac.c list_del_rcu(&shdev->node); node 58 drivers/dma/sh/usb-dmac.c struct list_head node; node 237 drivers/dma/sh/usb-dmac.c list_del(&vd->node); node 274 drivers/dma/sh/usb-dmac.c INIT_LIST_HEAD(&desc->node); node 277 drivers/dma/sh/usb-dmac.c list_add_tail(&desc->node, &chan->desc_freed); node 291 drivers/dma/sh/usb-dmac.c list_for_each_entry_safe(desc, _desc, &list, node) { node 292 drivers/dma/sh/usb-dmac.c list_del(&desc->node); node 306 drivers/dma/sh/usb-dmac.c list_for_each_entry(desc, &chan->desc_freed, node) { node 308 drivers/dma/sh/usb-dmac.c list_move_tail(&desc->node, &chan->desc_got); node 320 drivers/dma/sh/usb-dmac.c node); node 321 drivers/dma/sh/usb-dmac.c list_move_tail(&desc->node, &chan->desc_got); node 335 drivers/dma/sh/usb-dmac.c list_move_tail(&desc->node, &chan->desc_freed); node 459 drivers/dma/sh/usb-dmac.c list_for_each_entry_safe(desc, _desc, &list, node) node 460 drivers/dma/sh/usb-dmac.c list_move_tail(&desc->node, &uchan->desc_freed); node 493 drivers/dma/sh/usb-dmac.c list_for_each_entry_reverse(desc, &chan->desc_freed, node) { node 83 drivers/dma/sirf-dma.c struct list_head node; node 249 drivers/dma/sirf-dma.c node); node 251 drivers/dma/sirf-dma.c list_move_tail(&sdesc->node, &schan->active); node 285 drivers/dma/sirf-dma.c struct sirfsoc_dma_desc, node); node 307 drivers/dma/sirf-dma.c struct sirfsoc_dma_desc, node); node 359 drivers/dma/sirf-dma.c list_for_each_entry(sdesc, &list, node) { node 380 drivers/dma/sirf-dma.c struct sirfsoc_dma_desc, node); node 416 drivers/dma/sirf-dma.c list_move_tail(&sdesc->node, &schan->queued); node 580 drivers/dma/sirf-dma.c list_add_tail(&sdesc->node, &descs); node 618 drivers/dma/sirf-dma.c list_for_each_entry_safe(sdesc, tmp, &descs, node) node 661 drivers/dma/sirf-dma.c sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, node); node 707 drivers/dma/sirf-dma.c node); node 708 drivers/dma/sirf-dma.c list_del(&sdesc->node); node 740 drivers/dma/sirf-dma.c list_add_tail(&sdesc->node, &schan->prepared); node 783 drivers/dma/sirf-dma.c node); node 784 drivers/dma/sirf-dma.c list_del(&sdesc->node); node 798 drivers/dma/sirf-dma.c list_add_tail(&sdesc->node, &schan->prepared); node 1090 drivers/dma/sirf-dma.c node); node 516 drivers/dma/sprd-dma.c list_del(&vd->node); node 83 drivers/dma/st_fdma.c nbytes = fchan->fdesc->node[0].desc->nbytes; node 85 drivers/dma/st_fdma.c ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START; node 143 drivers/dma/st_fdma.c list_del(&fchan->fdesc->vdesc.node); node 231 drivers/dma/st_fdma.c dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc, node 232 drivers/dma/st_fdma.c fdesc->node[i].pdesc); node 242 drivers/dma/st_fdma.c fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT); node 249 drivers/dma/st_fdma.c fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool, node 250 drivers/dma/st_fdma.c GFP_NOWAIT, &fdesc->node[i].pdesc); node 251 drivers/dma/st_fdma.c if (!fdesc->node[i].desc) node 258 drivers/dma/st_fdma.c dma_pool_free(fchan->node_pool, fdesc->node[i].desc, node 259 drivers/dma/st_fdma.c fdesc->node[i].pdesc); node 329 drivers/dma/st_fdma.c hw_node = fdesc->node[0].desc; node 483 drivers/dma/st_fdma.c struct st_fdma_hw_node *hw_node = fdesc->node[i].desc; node 485 drivers/dma/st_fdma.c hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; node 532 drivers/dma/st_fdma.c hw_node = fdesc->node[i].desc; node 534 drivers/dma/st_fdma.c hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; node 569 drivers/dma/st_fdma.c if (cur_addr == fdesc->node[i].pdesc) { node 573 drivers/dma/st_fdma.c residue += fdesc->node[i].desc->nbytes; node 100 drivers/dma/st_fdma.h struct st_fdma_sw_node node[]; node 374 drivers/dma/ste_dma40.c struct list_head node; node 763 drivers/dma/ste_dma40.c list_del(&d40d->node); node 774 drivers/dma/ste_dma40.c list_for_each_entry_safe(d, _d, &d40c->client, node) { node 788 drivers/dma/ste_dma40.c INIT_LIST_HEAD(&desc->node); node 803 drivers/dma/ste_dma40.c list_add_tail(&desc->node, &d40c->active); node 825 drivers/dma/ste_dma40.c list_add_tail(&desc->node, &d40c->done); node 951 drivers/dma/ste_dma40.c return list_first_entry_or_null(&d40c->active, struct d40_desc, node); node 959 drivers/dma/ste_dma40.c list_add_tail(&desc->node, &d40c->pending_queue); node 965 drivers/dma/ste_dma40.c node); node 970 drivers/dma/ste_dma40.c return list_first_entry_or_null(&d40c->queue, struct d40_desc, node); node 975 drivers/dma/ste_dma40.c return list_first_entry_or_null(&d40c->done, struct d40_desc, node); node 1139 drivers/dma/ste_dma40.c list_for_each_entry_safe(d40d, _d, &d40c->client, node) { node 1147 drivers/dma/ste_dma40.c &d40c->prepare_queue, node) { node 1616 drivers/dma/ste_dma40.c list_add_tail(&d40d->node, &d40c->client); node 2260 drivers/dma/ste_dma40.c list_add_tail(&desc->node, &chan->prepare_queue); node 625 drivers/dma/stm32-dma.c list_del(&chan->desc->vdesc.node); node 176 drivers/dma/stm32-dmamux.c struct device_node *node = pdev->dev.of_node; node 185 drivers/dma/stm32-dmamux.c if (!node) node 201 drivers/dma/stm32-dmamux.c dma_node = of_parse_phandle(node, "dma-masters", i - 1); node 290 drivers/dma/stm32-dmamux.c return of_dma_router_register(node, stm32_dmamux_route_allocate, node 255 drivers/dma/stm32-mdma.c struct stm32_mdma_desc_node node[]; node 343 drivers/dma/stm32-mdma.c desc = kzalloc(offsetof(typeof(*desc), node[count]), GFP_NOWAIT); node 348 drivers/dma/stm32-mdma.c desc->node[i].hwdesc = node 350 drivers/dma/stm32-mdma.c &desc->node[i].hwdesc_phys); node 351 drivers/dma/stm32-mdma.c if (!desc->node[i].hwdesc) node 362 drivers/dma/stm32-mdma.c dma_pool_free(chan->desc_pool, desc->node[i].hwdesc, node 363 drivers/dma/stm32-mdma.c desc->node[i].hwdesc_phys); node 375 drivers/dma/stm32-mdma.c dma_pool_free(chan->desc_pool, desc->node[i].hwdesc, node 376 drivers/dma/stm32-mdma.c desc->node[i].hwdesc_phys); node 675 drivers/dma/stm32-mdma.c struct stm32_mdma_desc_node *node) node 677 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "hwdesc: %pad\n", &node->hwdesc_phys); node 678 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CTCR: 0x%08x\n", node->hwdesc->ctcr); node 679 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CBNDTR: 0x%08x\n", node->hwdesc->cbndtr); node 680 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CSAR: 0x%08x\n", node->hwdesc->csar); node 681 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CDAR: 0x%08x\n", node->hwdesc->cdar); node 682 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CBRUR: 0x%08x\n", node->hwdesc->cbrur); node 683 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CLAR: 0x%08x\n", node->hwdesc->clar); node 684 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CTBR: 0x%08x\n", node->hwdesc->ctbr); node 685 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CMAR: 0x%08x\n", node->hwdesc->cmar); node 686 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "CMDR: 0x%08x\n\n", node->hwdesc->cmdr); node 700 drivers/dma/stm32-mdma.c hwdesc = desc->node[count].hwdesc; node 716 drivers/dma/stm32-mdma.c hwdesc->clar = desc->node[0].hwdesc_phys; node 720 drivers/dma/stm32-mdma.c hwdesc->clar = desc->node[next].hwdesc_phys; node 723 drivers/dma/stm32-mdma.c stm32_mdma_dump_hwdesc(chan, &desc->node[count]); node 814 drivers/dma/stm32-mdma.c dma_pool_free(chan->desc_pool, desc->node[i].hwdesc, node 815 drivers/dma/stm32-mdma.c desc->node[i].hwdesc_phys); node 905 drivers/dma/stm32-mdma.c dma_pool_free(chan->desc_pool, desc->node[i].hwdesc, node 906 drivers/dma/stm32-mdma.c desc->node[i].hwdesc_phys); node 1020 drivers/dma/stm32-mdma.c hwdesc = desc->node[0].hwdesc; node 1031 drivers/dma/stm32-mdma.c stm32_mdma_dump_hwdesc(chan, &desc->node[0]); node 1131 drivers/dma/stm32-mdma.c hwdesc = chan->desc->node[0].hwdesc; node 1209 drivers/dma/stm32-mdma.c hwdesc = chan->desc->node[chan->curr_hwdesc].hwdesc; node 1279 drivers/dma/stm32-mdma.c struct stm32_mdma_hwdesc *hwdesc = desc->node[0].hwdesc; node 1285 drivers/dma/stm32-mdma.c hwdesc = desc->node[i].hwdesc; node 1334 drivers/dma/stm32-mdma.c list_del(&chan->desc->vdesc.node); node 359 drivers/dma/sun4i-dma.c list_del(&contract->vd.node); node 177 drivers/dma/sun6i-dma.c struct list_head node; node 443 drivers/dma/sun6i-dma.c list_del(&desc->node); node 507 drivers/dma/sun6i-dma.c struct sun6i_vchan, node); node 510 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); node 852 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); node 875 drivers/dma/sun6i-dma.c list_add_tail(&vchan->node, &sdev->pending); node 893 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); node 904 drivers/dma/sun6i-dma.c list_add_tail(&vd->node, &vc->desc_completed); node 976 drivers/dma/sun6i-dma.c if (!vchan->phy && list_empty(&vchan->node)) { node 977 drivers/dma/sun6i-dma.c list_add_tail(&vchan->node, &sdev->pending); node 999 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); node 1358 drivers/dma/sun6i-dma.c INIT_LIST_HEAD(&vchan->node); node 153 drivers/dma/tegra20-apb-dma.c struct list_head node; node 168 drivers/dma/tegra20-apb-dma.c struct list_head node; node 283 drivers/dma/tegra20-apb-dma.c list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) { node 285 drivers/dma/tegra20-apb-dma.c list_del(&dma_desc->node); node 313 drivers/dma/tegra20-apb-dma.c list_add_tail(&dma_desc->node, &tdc->free_dma_desc); node 326 drivers/dma/tegra20-apb-dma.c typeof(*sg_req), node); node 327 drivers/dma/tegra20-apb-dma.c list_del(&sg_req->node); node 513 drivers/dma/tegra20-apb-dma.c typeof(*sg_req), node); node 528 drivers/dma/tegra20-apb-dma.c hsgreq = list_first_entry(&tdc->pending_sg_req, typeof(*hsgreq), node); node 529 drivers/dma/tegra20-apb-dma.c if (!list_is_last(&hsgreq->node, &tdc->pending_sg_req)) { node 530 drivers/dma/tegra20-apb-dma.c hnsgreq = list_first_entry(&hsgreq->node, node 531 drivers/dma/tegra20-apb-dma.c typeof(*hnsgreq), node); node 549 drivers/dma/tegra20-apb-dma.c typeof(*sgreq), node); node 550 drivers/dma/tegra20-apb-dma.c list_move_tail(&sgreq->node, &tdc->free_sg_req); node 554 drivers/dma/tegra20-apb-dma.c list_add_tail(&dma_desc->node, &tdc->free_dma_desc); node 582 drivers/dma/tegra20-apb-dma.c hsgreq = list_first_entry(&tdc->pending_sg_req, typeof(*hsgreq), node); node 603 drivers/dma/tegra20-apb-dma.c sgreq = list_first_entry(&tdc->pending_sg_req, typeof(*sgreq), node); node 607 drivers/dma/tegra20-apb-dma.c list_del(&sgreq->node); node 614 drivers/dma/tegra20-apb-dma.c list_add_tail(&dma_desc->node, &tdc->free_dma_desc); node 616 drivers/dma/tegra20-apb-dma.c list_add_tail(&sgreq->node, &tdc->free_sg_req); node 632 drivers/dma/tegra20-apb-dma.c sgreq = list_first_entry(&tdc->pending_sg_req, typeof(*sgreq), node); node 647 drivers/dma/tegra20-apb-dma.c if (!list_is_last(&sgreq->node, &tdc->pending_sg_req)) { node 648 drivers/dma/tegra20-apb-dma.c list_move_tail(&sgreq->node, &tdc->pending_sg_req); node 782 drivers/dma/tegra20-apb-dma.c typeof(*sgreq), node); node 806 drivers/dma/tegra20-apb-dma.c if (!list_is_first(&sg_req->node, &tdc->pending_sg_req)) node 878 drivers/dma/tegra20-apb-dma.c list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) { node 886 drivers/dma/tegra20-apb-dma.c list_for_each_entry(sg_req, &tdc->pending_sg_req, node) { node 1099 drivers/dma/tegra20-apb-dma.c list_add_tail(&sg_req->node, &dma_desc->tx_list); node 1233 drivers/dma/tegra20-apb-dma.c list_add_tail(&sg_req->node, &dma_desc->tx_list); node 1304 drivers/dma/tegra20-apb-dma.c typeof(*dma_desc), node); node 1305 drivers/dma/tegra20-apb-dma.c list_del(&dma_desc->node); node 1310 drivers/dma/tegra20-apb-dma.c sg_req = list_first_entry(&sg_req_list, typeof(*sg_req), node); node 1311 drivers/dma/tegra20-apb-dma.c list_del(&sg_req->node); node 360 drivers/dma/tegra210-adma.c list_del(&vd->node); node 104 drivers/dma/ti/cppi41.c struct list_head node; /* Node for pending list */ node 484 drivers/dma/ti/cppi41.c list_for_each_entry_safe(c, _c, &cdd->pending, node) { node 486 drivers/dma/ti/cppi41.c list_del(&c->node); node 507 drivers/dma/ti/cppi41.c list_add_tail(&c->node, &cdd->pending); node 754 drivers/dma/ti/cppi41.c list_for_each_entry_safe(cc, _ct, &cdd->pending, node) { node 757 drivers/dma/ti/cppi41.c list_del(&cc->node); node 132 drivers/dma/ti/dma-crossbar.c struct device_node *node = pdev->dev.of_node; node 140 drivers/dma/ti/dma-crossbar.c if (!node) node 147 drivers/dma/ti/dma-crossbar.c dma_node = of_parse_phandle(node, "dma-masters", 0); node 169 drivers/dma/ti/dma-crossbar.c if (of_property_read_u32(node, "dma-requests", &xbar->xbar_events)) { node 192 drivers/dma/ti/dma-crossbar.c ret = of_dma_router_register(node, ti_am335x_xbar_route_allocate, node 321 drivers/dma/ti/dma-crossbar.c struct device_node *node = pdev->dev.of_node; node 332 drivers/dma/ti/dma-crossbar.c if (!node) node 339 drivers/dma/ti/dma-crossbar.c dma_node = of_parse_phandle(node, "dma-masters", 0); node 367 drivers/dma/ti/dma-crossbar.c if (of_property_read_u32(node, "dma-requests", &xbar->xbar_requests)) { node 374 drivers/dma/ti/dma-crossbar.c if (!of_property_read_u32(node, "ti,dma-safe-map", &safe_val)) node 378 drivers/dma/ti/dma-crossbar.c prop = of_find_property(node, "ti,reserved-dma-request-ranges", &sz); node 392 drivers/dma/ti/dma-crossbar.c ret = of_property_read_u32_array(node, pname, (u32 *)rsv_events, node 426 drivers/dma/ti/dma-crossbar.c ret = of_dma_router_register(node, ti_dra7_xbar_route_allocate, node 180 drivers/dma/ti/edma.c struct list_head node; node 220 drivers/dma/ti/edma.c struct device_node *node; node 226 drivers/dma/ti/edma.c struct list_head node; node 775 drivers/dma/ti/edma.c list_del(&vdesc->node); node 1978 drivers/dma/ti/edma.c INIT_LIST_HEAD(&echan->node); node 2258 drivers/dma/ti/edma.c struct device_node *node = pdev->dev.of_node; node 2264 drivers/dma/ti/edma.c if (node) { node 2267 drivers/dma/ti/edma.c match = of_match_node(edma_of_ids, node); node 2361 drivers/dma/ti/edma.c if (irq < 0 && node) node 2362 drivers/dma/ti/edma.c irq = irq_of_parse_and_map(node, 0); node 2377 drivers/dma/ti/edma.c if (irq < 0 && node) node 2378 drivers/dma/ti/edma.c irq = irq_of_parse_and_map(node, 2); node 2412 drivers/dma/ti/edma.c ret = of_parse_phandle_with_fixed_args(node, "ti,tptcs", node 2417 drivers/dma/ti/edma.c ecc->tc_list[i].node = tc_args.np; node 2469 drivers/dma/ti/edma.c if (node) node 2470 drivers/dma/ti/edma.c of_dma_controller_register(node, of_edma_xlate, ecc); node 540 drivers/dma/ti/omap-dma.c list_del(&vd->node); node 98 drivers/dma/uniphier-mdmac.c list_del(&vd->node); node 29 drivers/dma/virt-dma.c list_move_tail(&vd->node, &vc->desc_submitted); node 56 drivers/dma/virt-dma.c list_del(&vd->node); node 71 drivers/dma/virt-dma.c list_for_each_entry(vd, &vc->desc_issued, node) node 103 drivers/dma/virt-dma.c list_for_each_entry_safe(vd, _vd, &head, node) { node 106 drivers/dma/virt-dma.c list_del(&vd->node); node 116 drivers/dma/virt-dma.c list_for_each_entry_safe(vd, _vd, head, node) { node 118 drivers/dma/virt-dma.c list_move_tail(&vd->node, &vc->desc_allocated); node 121 drivers/dma/virt-dma.c list_del(&vd->node); node 19 drivers/dma/virt-dma.h struct list_head node; node 70 drivers/dma/virt-dma.h list_add_tail(&vd->node, &vc->desc_allocated); node 103 drivers/dma/virt-dma.h list_add_tail(&vd->node, &vc->desc_completed); node 117 drivers/dma/virt-dma.h list_add(&vd->node, &vc->desc_allocated); node 162 drivers/dma/virt-dma.h struct virt_dma_desc, node); node 192 drivers/dma/virt-dma.h list_for_each_entry(vd, &head, node) node 239 drivers/dma/xgene-dma.c struct list_head node; node 481 drivers/dma/xgene-dma.c list_del(&desc->node); node 520 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc, _desc, &chan->ld_completed, node) { node 572 drivers/dma/xgene-dma.c list_del(&desc->node); node 583 drivers/dma/xgene-dma.c list_add_tail(&desc->node, &chan->ld_completed); node 655 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_pending, node) { node 671 drivers/dma/xgene-dma.c list_move_tail(&desc_sw->node, &chan->ld_running); node 700 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_running, node) { node 753 drivers/dma/xgene-dma.c list_move_tail(&desc_sw->node, &ld_completed); node 766 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &ld_completed, node) { node 805 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc, _desc, list, node) node 866 drivers/dma/xgene-dma.c list_add_tail(&new->node, &first->tx_list); node 924 drivers/dma/xgene-dma.c list_add_tail(&new->node, &first->tx_list); node 272 drivers/dma/xilinx/xilinx_dma.c struct list_head node; node 284 drivers/dma/xilinx/xilinx_dma.c struct list_head node; node 296 drivers/dma/xilinx/xilinx_dma.c struct list_head node; node 310 drivers/dma/xilinx/xilinx_dma.c struct list_head node; node 611 drivers/dma/xilinx/xilinx_dma.c node); node 612 drivers/dma/xilinx/xilinx_dma.c list_del(&segment->node); node 640 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &chan->free_seg_list); node 702 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(segment, next, &desc->segments, node) { node 703 drivers/dma/xilinx/xilinx_dma.c list_del(&segment->node); node 708 drivers/dma/xilinx/xilinx_dma.c &desc->segments, node) { node 709 drivers/dma/xilinx/xilinx_dma.c list_del(&cdma_segment->node); node 714 drivers/dma/xilinx/xilinx_dma.c &desc->segments, node) { node 715 drivers/dma/xilinx/xilinx_dma.c list_del(&axidma_segment->node); node 735 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, list, node) { node 736 drivers/dma/xilinx/xilinx_dma.c list_del(&desc->node); node 825 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, &chan->done_list, node) { node 834 drivers/dma/xilinx/xilinx_dma.c list_del(&desc->node); node 922 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&chan->seg_v[i].node, node 1020 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1022 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry(segment, &desc->segments, node) { node 1117 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1164 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry(segment, &desc->segments, node) { node 1189 drivers/dma/xilinx/xilinx_dma.c list_del(&desc->node); node 1190 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&desc->node, &chan->active_list); node 1217 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1219 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1221 drivers/dma/xilinx/xilinx_dma.c struct xilinx_cdma_tx_segment, node); node 1250 drivers/dma/xilinx/xilinx_dma.c node); node 1289 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1291 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1293 drivers/dma/xilinx/xilinx_dma.c struct xilinx_axidma_tx_segment, node); node 1357 drivers/dma/xilinx/xilinx_dma.c node); node 1401 drivers/dma/xilinx/xilinx_dma.c list_for_each_entry_safe(desc, next, &chan->active_list, node) { node 1402 drivers/dma/xilinx/xilinx_dma.c list_del(&desc->node); node 1405 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&desc->node, &chan->done_list); node 1549 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor, node); node 1553 drivers/dma/xilinx/xilinx_dma.c node); node 1558 drivers/dma/xilinx/xilinx_dma.c node); node 1563 drivers/dma/xilinx/xilinx_dma.c node); node 1572 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&desc->node, &chan->pending_list); node 1694 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &desc->segments); node 1698 drivers/dma/xilinx/xilinx_dma.c struct xilinx_vdma_tx_segment, node); node 1752 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &desc->segments); node 1839 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &desc->segments); node 1844 drivers/dma/xilinx/xilinx_dma.c struct xilinx_axidma_tx_segment, node); node 1852 drivers/dma/xilinx/xilinx_dma.c node); node 1939 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &desc->segments); node 1944 drivers/dma/xilinx/xilinx_dma.c struct xilinx_axidma_tx_segment, node); node 1954 drivers/dma/xilinx/xilinx_dma.c node); node 2031 drivers/dma/xilinx/xilinx_dma.c list_add_tail(&segment->node, &desc->segments); node 2035 drivers/dma/xilinx/xilinx_dma.c struct xilinx_axidma_tx_segment, node); node 2043 drivers/dma/xilinx/xilinx_dma.c node); node 2388 drivers/dma/xilinx/xilinx_dma.c struct device_node *node, int chan_id) node 2418 drivers/dma/xilinx/xilinx_dma.c has_dre = of_property_read_bool(node, "xlnx,include-dre"); node 2420 drivers/dma/xilinx/xilinx_dma.c chan->genlock = of_property_read_bool(node, "xlnx,genlock-mode"); node 2422 drivers/dma/xilinx/xilinx_dma.c err = of_property_read_u32(node, "xlnx,datawidth", &value); node 2436 drivers/dma/xilinx/xilinx_dma.c if (of_device_is_compatible(node, "xlnx,axi-vdma-mm2s-channel") || node 2437 drivers/dma/xilinx/xilinx_dma.c of_device_is_compatible(node, "xlnx,axi-dma-mm2s-channel") || node 2438 drivers/dma/xilinx/xilinx_dma.c of_device_is_compatible(node, "xlnx,axi-cdma-channel")) { node 2452 drivers/dma/xilinx/xilinx_dma.c } else if (of_device_is_compatible(node, node 2454 drivers/dma/xilinx/xilinx_dma.c of_device_is_compatible(node, node 2459 drivers/dma/xilinx/xilinx_dma.c chan->has_vflip = of_property_read_bool(node, node 2482 drivers/dma/xilinx/xilinx_dma.c chan->irq = irq_of_parse_and_map(node, 0); node 2544 drivers/dma/xilinx/xilinx_dma.c struct device_node *node) node 2548 drivers/dma/xilinx/xilinx_dma.c ret = of_property_read_u32(node, "dma-channels", &nr_channels); node 2553 drivers/dma/xilinx/xilinx_dma.c xilinx_dma_chan_probe(xdev, node, xdev->chan_id++); node 2613 drivers/dma/xilinx/xilinx_dma.c struct device_node *node = pdev->dev.of_node; node 2651 drivers/dma/xilinx/xilinx_dma.c xdev->mcdma = of_property_read_bool(node, "xlnx,mcdma"); node 2652 drivers/dma/xilinx/xilinx_dma.c if (!of_property_read_u32(node, "xlnx,sg-length-width", node 2668 drivers/dma/xilinx/xilinx_dma.c err = of_property_read_u32(node, "xlnx,num-fstores", node 2676 drivers/dma/xilinx/xilinx_dma.c err = of_property_read_u32(node, "xlnx,flush-fsync", node 2683 drivers/dma/xilinx/xilinx_dma.c err = of_property_read_u32(node, "xlnx,addrwidth", &addr_width); node 2732 drivers/dma/xilinx/xilinx_dma.c for_each_child_of_node(node, child) { node 2747 drivers/dma/xilinx/xilinx_dma.c err = of_dma_controller_register(node, of_dma_xilinx_xlate, node 181 drivers/dma/xilinx/zynqmp_dma.c struct list_head node; node 382 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw, node); node 385 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw, node); node 392 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&new->node, &chan->pending_list); node 412 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw, node); node 413 drivers/dma/xilinx/zynqmp_dma.c list_del(&desc->node); node 435 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&sdesc->node, &chan->free_list); node 436 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(child, next, &sdesc->tx_list, node) { node 438 drivers/dma/xilinx/zynqmp_dma.c list_move_tail(&child->node, &chan->free_list); node 452 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(desc, next, list, node) node 486 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&desc->node, &chan->free_list); node 583 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw, node); node 601 drivers/dma/xilinx/zynqmp_dma.c list_for_each_entry_safe(desc, next, &chan->done_list, node) { node 605 drivers/dma/xilinx/zynqmp_dma.c list_del(&desc->node); node 629 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_desc_sw, node); node 632 drivers/dma/xilinx/zynqmp_dma.c list_del(&desc->node); node 634 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&desc->node, &chan->done_list); node 839 drivers/dma/xilinx/zynqmp_dma.c list_add_tail(&new->node, &first->tx_list); node 875 drivers/dma/xilinx/zynqmp_dma.c struct device_node *node = pdev->dev.of_node; node 892 drivers/dma/xilinx/zynqmp_dma.c err = of_property_read_u32(node, "xlnx,bus-width", &chan->bus_width); node 904 drivers/dma/xilinx/zynqmp_dma.c chan->is_dmacoherent = of_property_read_bool(node, "dma-coherent"); node 104 drivers/dma/zx_dma.c struct list_head node; node 205 drivers/dma/zx_dma.c list_del(&ds->vd.node); node 243 drivers/dma/zx_dma.c struct zx_dma_chan, node); node 247 drivers/dma/zx_dma.c list_del_init(&c->node); node 324 drivers/dma/zx_dma.c list_del_init(&c->node); node 390 drivers/dma/zx_dma.c if (!c->phy && list_empty(&c->node)) { node 392 drivers/dma/zx_dma.c list_add_tail(&c->node, &d->chan_pending); node 671 drivers/dma/zx_dma.c list_del_init(&c->node); node 843 drivers/dma/zx_dma.c INIT_LIST_HEAD(&c->node); node 27 drivers/edac/cell_edac.c int node; node 41 drivers/edac/cell_edac.c priv->node, chan, ar); node 64 drivers/edac/cell_edac.c priv->node, chan, ar); node 146 drivers/edac/cell_edac.c if (of_node_to_nid(np) != priv->node) node 161 drivers/edac/cell_edac.c priv->node, priv->chanmask, node 214 drivers/edac/cell_edac.c priv->node = pdev->id; node 335 drivers/edac/ghes_edac.c p += sprintf(p, "node:%d ", mem_err->node); node 193 drivers/edac/thunderx_edac.c u64 node; node 418 drivers/edac/thunderx_edac.c lmc->mem = alloc_pages_node(lmc->node, GFP_KERNEL, 0); node 496 drivers/edac/thunderx_edac.c addr |= lmc->node << 40; node 520 drivers/edac/thunderx_edac.c static unsigned int thunderx_get_num_lmcs(unsigned int node) node 531 drivers/edac/thunderx_edac.c if (pdev->dev.numa_node == node) node 653 drivers/edac/thunderx_edac.c int node = dev_to_node(&pdev->dev); node 656 drivers/edac/thunderx_edac.c ret += max(node, 0) << 3; node 743 drivers/edac/thunderx_edac.c lmc->node = FIELD_GET(THUNDERX_NODE, pci_resource_start(pdev, 0)); node 745 drivers/edac/thunderx_edac.c lmc->xbits = thunderx_get_num_lmcs(lmc->node) >> 1; node 760 drivers/edac/thunderx_edac.c l2c_ioaddr = ioremap(L2C_CTL | FIELD_PREP(THUNDERX_NODE, lmc->node), PAGE_SIZE); node 201 drivers/edac/ti_edac.c static int _emif_get_id(struct device_node *node) node 208 drivers/edac/ti_edac.c addrp = of_get_address(node, 0, NULL, NULL); node 209 drivers/edac/ti_edac.c my_addr = (u32)of_translate_address(node, addrp); node 212 drivers/edac/ti_edac.c if (np == node) node 1818 drivers/edac/xgene_edac.c struct xgene_edac_dev_ctx *node; node 1838 drivers/edac/xgene_edac.c list_for_each_entry(node, &ctx->l3s, next) node 1839 drivers/edac/xgene_edac.c xgene_edac_l3_check(node->edac_dev); node 1841 drivers/edac/xgene_edac.c list_for_each_entry(node, &ctx->socs, next) node 1842 drivers/edac/xgene_edac.c xgene_edac_soc_check(node->edac_dev); node 1973 drivers/edac/xgene_edac.c struct xgene_edac_dev_ctx *node; node 1982 drivers/edac/xgene_edac.c list_for_each_entry_safe(node, temp_node, &edac->l3s, next) node 1983 drivers/edac/xgene_edac.c xgene_edac_l3_remove(node); node 1985 drivers/edac/xgene_edac.c list_for_each_entry_safe(node, temp_node, &edac->socs, next) node 1986 drivers/edac/xgene_edac.c xgene_edac_soc_remove(node); node 178 drivers/extcon/extcon-palmas.c struct device_node *node = pdev->dev.of_node; node 191 drivers/extcon/extcon-palmas.c if (node && !pdata) { node 192 drivers/extcon/extcon-palmas.c palmas_usb->wakeup = of_property_read_bool(node, "ti,wakeup"); node 193 drivers/extcon/extcon-palmas.c palmas_usb->enable_id_detection = of_property_read_bool(node, node 195 drivers/extcon/extcon-palmas.c palmas_usb->enable_vbus_detection = of_property_read_bool(node, node 233 drivers/extcon/extcon-palmas.c if (of_property_read_u32(node, "debounce-delay-ms", &debounce)) node 1336 drivers/extcon/extcon.c struct extcon_dev *extcon_find_edev_by_node(struct device_node *node) node 1342 drivers/extcon/extcon.c if (edev->dev.parent && edev->dev.parent->of_node == node) node 1360 drivers/extcon/extcon.c struct device_node *node; node 1371 drivers/extcon/extcon.c node = of_parse_phandle(dev->of_node, "extcon", index); node 1372 drivers/extcon/extcon.c if (!node) { node 1378 drivers/extcon/extcon.c edev = extcon_find_edev_by_node(node); node 1379 drivers/extcon/extcon.c of_node_put(node); node 1386 drivers/extcon/extcon.c struct extcon_dev *extcon_find_edev_by_node(struct device_node *node) node 544 drivers/firewire/core-device.c device->max_speed = device->node->max_speed; node 830 drivers/firewire/core-device.c device->node->data = NULL; node 833 drivers/firewire/core-device.c fw_node_put(device->node); node 894 drivers/firewire/core-device.c struct fw_node *current_node = new->node; node 895 drivers/firewire/core-device.c struct fw_node *obsolete_node = old->node; node 897 drivers/firewire/core-device.c new->node = obsolete_node; node 898 drivers/firewire/core-device.c new->node->data = new; node 899 drivers/firewire/core-device.c old->node = current_node; node 900 drivers/firewire/core-device.c old->node->data = old; node 1005 drivers/firewire/core-device.c if (device->node->link_on) node 1009 drivers/firewire/core-device.c if (device->node == card->root_node) node 1088 drivers/firewire/core-device.c if (device->node == card->root_node) node 1203 drivers/firewire/core-device.c void fw_node_event(struct fw_card *card, struct fw_node *node, int event) node 1232 drivers/firewire/core-device.c device->node = fw_node_get(node); node 1233 drivers/firewire/core-device.c device->node_id = node->node_id; node 1235 drivers/firewire/core-device.c device->is_local = node == card->local_node; node 1244 drivers/firewire/core-device.c node->data = device; node 1259 drivers/firewire/core-device.c device = node->data; node 1263 drivers/firewire/core-device.c device->node_id = node->node_id; node 1276 drivers/firewire/core-device.c device = node->data; node 1280 drivers/firewire/core-device.c device->node_id = node->node_id; node 1291 drivers/firewire/core-device.c if (!node->data) node 1306 drivers/firewire/core-device.c device = node->data; node 101 drivers/firewire/core-topology.c struct fw_node *node; node 103 drivers/firewire/core-topology.c node = kzalloc(struct_size(node, ports, port_count), GFP_ATOMIC); node 104 drivers/firewire/core-topology.c if (node == NULL) node 107 drivers/firewire/core-topology.c node->color = color; node 108 drivers/firewire/core-topology.c node->node_id = LOCAL_BUS | SELF_ID_PHY_ID(sid); node 109 drivers/firewire/core-topology.c node->link_on = SELF_ID_LINK_ON(sid); node 110 drivers/firewire/core-topology.c node->phy_speed = SELF_ID_PHY_SPEED(sid); node 111 drivers/firewire/core-topology.c node->initiated_reset = SELF_ID_PHY_INITIATOR(sid); node 112 drivers/firewire/core-topology.c node->port_count = port_count; node 114 drivers/firewire/core-topology.c refcount_set(&node->ref_count, 1); node 115 drivers/firewire/core-topology.c INIT_LIST_HEAD(&node->link); node 117 drivers/firewire/core-topology.c return node; node 134 drivers/firewire/core-topology.c static void update_hop_count(struct fw_node *node) node 140 drivers/firewire/core-topology.c for (i = 0; i < node->port_count; i++) { node 141 drivers/firewire/core-topology.c if (node->ports[i] == NULL) node 144 drivers/firewire/core-topology.c if (node->ports[i]->max_hops > max_child_hops) node 145 drivers/firewire/core-topology.c max_child_hops = node->ports[i]->max_hops; node 147 drivers/firewire/core-topology.c if (node->ports[i]->max_depth > depths[0]) { node 149 drivers/firewire/core-topology.c depths[0] = node->ports[i]->max_depth; node 150 drivers/firewire/core-topology.c } else if (node->ports[i]->max_depth > depths[1]) node 151 drivers/firewire/core-topology.c depths[1] = node->ports[i]->max_depth; node 154 drivers/firewire/core-topology.c node->max_depth = depths[0] + 1; node 155 drivers/firewire/core-topology.c node->max_hops = max(max_child_hops, depths[0] + depths[1] + 2); node 173 drivers/firewire/core-topology.c struct fw_node *node, *child, *local_node, *irm_node; node 181 drivers/firewire/core-topology.c node = NULL; node 222 drivers/firewire/core-topology.c node = fw_node_create(q, port_count, card->color); node 223 drivers/firewire/core-topology.c if (node == NULL) { node 229 drivers/firewire/core-topology.c local_node = node; node 232 drivers/firewire/core-topology.c irm_node = node; node 250 drivers/firewire/core-topology.c node->color = i; node 254 drivers/firewire/core-topology.c node->ports[i] = child; node 259 drivers/firewire/core-topology.c child->ports[child->color] = node; node 280 drivers/firewire/core-topology.c list_add_tail(&node->link, &stack); node 283 drivers/firewire/core-topology.c if (node->phy_speed == SCODE_BETA && node 294 drivers/firewire/core-topology.c update_hop_count(node); node 300 drivers/firewire/core-topology.c card->root_node = node; node 309 drivers/firewire/core-topology.c struct fw_node * node, node 316 drivers/firewire/core-topology.c struct fw_node *node, *next, *child, *parent; node 324 drivers/firewire/core-topology.c list_for_each_entry(node, &list, link) { node 325 drivers/firewire/core-topology.c node->color = card->color; node 327 drivers/firewire/core-topology.c for (i = 0; i < node->port_count; i++) { node 328 drivers/firewire/core-topology.c child = node->ports[i]; node 339 drivers/firewire/core-topology.c callback(card, node, parent); node 342 drivers/firewire/core-topology.c list_for_each_entry_safe(node, next, &list, link) node 343 drivers/firewire/core-topology.c fw_node_put(node); node 347 drivers/firewire/core-topology.c struct fw_node *node, struct fw_node *parent) node 349 drivers/firewire/core-topology.c fw_node_event(card, node, FW_NODE_DESTROYED); node 350 drivers/firewire/core-topology.c fw_node_put(node); node 357 drivers/firewire/core-topology.c struct fw_node *node, struct fw_node *parent) node 359 drivers/firewire/core-topology.c int b_path = (node->phy_speed == SCODE_BETA); node 363 drivers/firewire/core-topology.c node->max_speed = parent->max_speed < node->phy_speed ? node 364 drivers/firewire/core-topology.c parent->max_speed : node->phy_speed; node 365 drivers/firewire/core-topology.c node->b_path = parent->b_path && b_path; node 367 drivers/firewire/core-topology.c node->max_speed = node->phy_speed; node 368 drivers/firewire/core-topology.c node->b_path = b_path; node 371 drivers/firewire/core-topology.c fw_node_event(card, node, FW_NODE_CREATED); node 153 drivers/firewire/core.h void fw_node_event(struct fw_card *card, struct fw_node *node, int event); node 199 drivers/firewire/core.h static inline struct fw_node *fw_node_get(struct fw_node *node) node 201 drivers/firewire/core.h refcount_inc(&node->ref_count); node 203 drivers/firewire/core.h return node; node 206 drivers/firewire/core.h static inline void fw_node_put(struct fw_node *node) node 208 drivers/firewire/core.h if (refcount_dec_and_test(&node->ref_count)) node 209 drivers/firewire/core.h kfree(node); node 137 drivers/firmware/arm_scmi/driver.c struct list_head node; node 635 drivers/firmware/arm_scmi/driver.c info = list_entry(p, struct scmi_info, node); node 852 drivers/firmware/arm_scmi/driver.c INIT_LIST_HEAD(&info->node); node 877 drivers/firmware/arm_scmi/driver.c list_add_tail(&info->node, &scmi_list); node 926 drivers/firmware/arm_scmi/driver.c list_del(&info->node); node 228 drivers/firmware/arm_scpi.c struct list_head node; node 370 drivers/firmware/arm_scpi.c node); node 371 drivers/firmware/arm_scpi.c list_del(&match->node); node 373 drivers/firmware/arm_scpi.c list_for_each_entry(t, &ch->rx_pending, node) node 375 drivers/firmware/arm_scpi.c list_del(&t->node); node 440 drivers/firmware/arm_scpi.c list_add_tail(&t->node, &ch->rx_pending); node 457 drivers/firmware/arm_scpi.c t = list_first_entry(&ch->xfers_list, struct scpi_xfer, node); node 458 drivers/firmware/arm_scpi.c list_del(&t->node); node 466 drivers/firmware/arm_scpi.c list_add_tail(&t->node, &ch->xfers_list); node 889 drivers/firmware/arm_scpi.c list_add_tail(&xfers->node, &ch->xfers_list); node 226 drivers/firmware/efi/cper.c n += scnprintf(msg + n, len - n, "node: %d ", mem->node); node 284 drivers/firmware/efi/cper.c cmem->node = mem->node; node 34 drivers/firmware/efi/dev-path-parser.c static long __init parse_acpi_path(struct efi_dev_path *node, node 40 drivers/firmware/efi/dev-path-parser.c if (node->length != 12) node 44 drivers/firmware/efi/dev-path-parser.c 'A' + ((node->acpi.hid >> 10) & 0x1f) - 1, node 45 drivers/firmware/efi/dev-path-parser.c 'A' + ((node->acpi.hid >> 5) & 0x1f) - 1, node 46 drivers/firmware/efi/dev-path-parser.c 'A' + ((node->acpi.hid >> 0) & 0x1f) - 1, node 47 drivers/firmware/efi/dev-path-parser.c node->acpi.hid >> 16); node 48 drivers/firmware/efi/dev-path-parser.c sprintf(hid_uid.uid, "%u", node->acpi.uid); node 72 drivers/firmware/efi/dev-path-parser.c static long __init parse_pci_path(struct efi_dev_path *node, node 77 drivers/firmware/efi/dev-path-parser.c if (node->length != 6) node 82 drivers/firmware/efi/dev-path-parser.c devfn = PCI_DEVFN(node->pci.dev, node->pci.fn); node 108 drivers/firmware/efi/dev-path-parser.c static long __init parse_end_path(struct efi_dev_path *node, node 111 drivers/firmware/efi/dev-path-parser.c if (node->length != 4) node 113 drivers/firmware/efi/dev-path-parser.c if (node->sub_type != EFI_DEV_END_INSTANCE && node 114 drivers/firmware/efi/dev-path-parser.c node->sub_type != EFI_DEV_END_ENTIRE) node 120 drivers/firmware/efi/dev-path-parser.c return node->sub_type; node 159 drivers/firmware/efi/dev-path-parser.c struct device * __init efi_get_device_by_path(struct efi_dev_path **node, node 169 drivers/firmware/efi/dev-path-parser.c if (*len < 4 || *len < (*node)->length) node 171 drivers/firmware/efi/dev-path-parser.c else if ((*node)->type == EFI_DEV_ACPI && node 172 drivers/firmware/efi/dev-path-parser.c (*node)->sub_type == EFI_DEV_BASIC_ACPI) node 173 drivers/firmware/efi/dev-path-parser.c ret = parse_acpi_path(*node, parent, &child); node 174 drivers/firmware/efi/dev-path-parser.c else if ((*node)->type == EFI_DEV_HW && node 175 drivers/firmware/efi/dev-path-parser.c (*node)->sub_type == EFI_DEV_PCI) node 176 drivers/firmware/efi/dev-path-parser.c ret = parse_pci_path(*node, parent, &child); node 177 drivers/firmware/efi/dev-path-parser.c else if (((*node)->type == EFI_DEV_END_PATH || node 178 drivers/firmware/efi/dev-path-parser.c (*node)->type == EFI_DEV_END_PATH2)) node 179 drivers/firmware/efi/dev-path-parser.c ret = parse_end_path(*node, parent, &child); node 188 drivers/firmware/efi/dev-path-parser.c *node = (void *)*node + (*node)->length; node 189 drivers/firmware/efi/dev-path-parser.c *len -= (*node)->length; node 719 drivers/firmware/efi/efi.c static int __init __find_uefi_params(unsigned long node, node 729 drivers/firmware/efi/efi.c prop = of_get_flat_dt_prop(node, params[i].propname, &len); node 753 drivers/firmware/efi/efi.c static int __init fdt_find_uefi_params(unsigned long node, const char *uname, node 768 drivers/firmware/efi/efi.c int err = of_get_flat_dt_subnode_by_name(node, subnode); node 773 drivers/firmware/efi/efi.c node = err; node 776 drivers/firmware/efi/efi.c return __find_uefi_params(node, info, dt_params[i].params); node 35 drivers/firmware/efi/libstub/fdt.c int node, num_rsv; node 80 drivers/firmware/efi/libstub/fdt.c node = fdt_subnode_offset(fdt, 0, "chosen"); node 81 drivers/firmware/efi/libstub/fdt.c if (node < 0) { node 82 drivers/firmware/efi/libstub/fdt.c node = fdt_add_subnode(fdt, 0, "chosen"); node 83 drivers/firmware/efi/libstub/fdt.c if (node < 0) { node 85 drivers/firmware/efi/libstub/fdt.c status = node; node 91 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop(fdt, node, "bootargs", cmdline_ptr, node 102 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,initrd-start", initrd_image_start); node 107 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,initrd-end", initrd_image_end); node 113 drivers/firmware/efi/libstub/fdt.c node = fdt_subnode_offset(fdt, 0, "chosen"); node 116 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,uefi-system-table", fdt_val64); node 122 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-start", fdt_val64); node 128 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-size", fdt_val32); node 132 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-desc-size", fdt_val32); node 136 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-desc-ver", fdt_val32); node 146 drivers/firmware/efi/libstub/fdt.c status = fdt_setprop_var(fdt, node, "kaslr-seed", fdt_val64); node 168 drivers/firmware/efi/libstub/fdt.c int node = fdt_path_offset(fdt, "/chosen"); node 173 drivers/firmware/efi/libstub/fdt.c if (node < 0) node 178 drivers/firmware/efi/libstub/fdt.c err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-start", fdt_val64); node 184 drivers/firmware/efi/libstub/fdt.c err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-size", fdt_val32); node 190 drivers/firmware/efi/libstub/fdt.c err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-desc-size", fdt_val32); node 196 drivers/firmware/efi/libstub/fdt.c err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-desc-ver", fdt_val32); node 43 drivers/firmware/efi/vars.c struct efi_generic_dev_path *node; node 46 drivers/firmware/efi/vars.c node = (struct efi_generic_dev_path *)buffer; node 48 drivers/firmware/efi/vars.c if (len < sizeof(*node)) node 51 drivers/firmware/efi/vars.c while (offset <= len - sizeof(*node) && node 52 drivers/firmware/efi/vars.c node->length >= sizeof(*node) && node 53 drivers/firmware/efi/vars.c node->length <= len - offset) { node 54 drivers/firmware/efi/vars.c offset += node->length; node 56 drivers/firmware/efi/vars.c if ((node->type == EFI_DEV_END_PATH || node 57 drivers/firmware/efi/vars.c node->type == EFI_DEV_END_PATH2) && node 58 drivers/firmware/efi/vars.c node->sub_type == EFI_DEV_END_ENTIRE) node 61 drivers/firmware/efi/vars.c node = (struct efi_generic_dev_path *)(buffer + offset); node 94 drivers/firmware/stratix10-svc.c struct list_head node; node 139 drivers/firmware/stratix10-svc.c struct list_head node; node 180 drivers/firmware/stratix10-svc.c list_for_each_entry(pmem, &svc_data_mem, node) node 739 drivers/firmware/stratix10-svc.c struct stratix10_svc_controller, node); node 838 drivers/firmware/stratix10-svc.c list_for_each_entry(p_mem, &svc_data_mem, node) node 919 drivers/firmware/stratix10-svc.c list_add_tail(&pmem->node, &svc_data_mem); node 939 drivers/firmware/stratix10-svc.c list_for_each_entry(pmem, &svc_data_mem, node) node 947 drivers/firmware/stratix10-svc.c list_del(&pmem->node); node 1024 drivers/firmware/stratix10-svc.c list_add_tail(&controller->node, &svc_ctrl); node 1064 drivers/firmware/stratix10-svc.c list_del(&ctrl->node); node 131 drivers/firmware/ti_sci.c struct list_head node; node 3023 drivers/firmware/ti_sci.c info = list_entry(p, struct ti_sci_info, node); node 3146 drivers/firmware/ti_sci.c info = list_entry(p, struct ti_sci_info, node); node 3428 drivers/firmware/ti_sci.c INIT_LIST_HEAD(&info->node); node 3517 drivers/firmware/ti_sci.c list_add_tail(&info->node, &ti_sci_list); node 3547 drivers/firmware/ti_sci.c list_del(&info->node); node 145 drivers/firmware/trusted_foundations.c struct device_node *node; node 149 drivers/firmware/trusted_foundations.c node = of_find_compatible_node(NULL, NULL, "tlm,trusted-foundations"); node 150 drivers/firmware/trusted_foundations.c if (!node) node 153 drivers/firmware/trusted_foundations.c err = of_property_read_u32(node, "tlm,version-major", node 157 drivers/firmware/trusted_foundations.c err = of_property_read_u32(node, "tlm,version-minor", node 624 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_request_node(const u32 node, const u32 capabilities, node 628 drivers/firmware/xilinx/zynqmp.c return zynqmp_pm_invoke_fn(PM_REQUEST_NODE, node, capabilities, node 642 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_release_node(const u32 node) node 644 drivers/firmware/xilinx/zynqmp.c return zynqmp_pm_invoke_fn(PM_RELEASE_NODE, node, 0, 0, 0, NULL); node 659 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_set_requirement(const u32 node, const u32 capabilities, node 663 drivers/firmware/xilinx/zynqmp.c return zynqmp_pm_invoke_fn(PM_SET_REQUIREMENT, node, capabilities, node 166 drivers/fpga/dfl-afu-dma-region.c this = container_of(*new, struct dfl_afu_dma_region, node); node 181 drivers/fpga/dfl-afu-dma-region.c rb_link_node(®ion->node, parent, new); node 182 drivers/fpga/dfl-afu-dma-region.c rb_insert_color(®ion->node, &afu->dma_regions); node 203 drivers/fpga/dfl-afu-dma-region.c rb_erase(®ion->node, &afu->dma_regions); node 215 drivers/fpga/dfl-afu-dma-region.c struct rb_node *node = rb_first(&afu->dma_regions); node 218 drivers/fpga/dfl-afu-dma-region.c while (node) { node 219 drivers/fpga/dfl-afu-dma-region.c region = container_of(node, struct dfl_afu_dma_region, node); node 224 drivers/fpga/dfl-afu-dma-region.c rb_erase(node, &afu->dma_regions); node 234 drivers/fpga/dfl-afu-dma-region.c node = rb_next(node); node 257 drivers/fpga/dfl-afu-dma-region.c struct rb_node *node = afu->dma_regions.rb_node; node 260 drivers/fpga/dfl-afu-dma-region.c while (node) { node 263 drivers/fpga/dfl-afu-dma-region.c region = container_of(node, struct dfl_afu_dma_region, node); node 272 drivers/fpga/dfl-afu-dma-region.c node = node->rb_left; node 274 drivers/fpga/dfl-afu-dma-region.c node = node->rb_right; node 25 drivers/fpga/dfl-afu-region.c list_for_each_entry((region), &(afu)->regions, node) node 78 drivers/fpga/dfl-afu-region.c list_add(®ion->node, &afu->regions); node 100 drivers/fpga/dfl-afu-region.c list_for_each_entry_safe(region, tmp, &afu->regions, node) node 40 drivers/fpga/dfl-afu.h struct list_head node; node 58 drivers/fpga/dfl-afu.h struct rb_node node; node 37 drivers/fpga/dfl-fme-pr.c list_for_each_entry(fme_region, &fme->region_list, node) node 288 drivers/fpga/dfl-fme-pr.c list_for_each_entry_safe(fbridge, tmp, &priv->bridge_list, node) { node 289 drivers/fpga/dfl-fme-pr.c list_del(&fbridge->node); node 367 drivers/fpga/dfl-fme-pr.c list_for_each_entry_safe(fme_region, tmp, &priv->region_list, node) { node 368 drivers/fpga/dfl-fme-pr.c list_del(&fme_region->node); node 418 drivers/fpga/dfl-fme-pr.c list_add(&fme_br->node, &priv->bridge_list); node 428 drivers/fpga/dfl-fme-pr.c list_add(&fme_region->node, &priv->region_list); node 32 drivers/fpga/dfl-fme-pr.h struct list_head node; node 57 drivers/fpga/dfl-fme-pr.h struct list_head node; node 174 drivers/fpga/dfl.c list_for_each_entry(ops, &dfl_port_ops_list, node) { node 208 drivers/fpga/dfl.c list_add_tail(&ops->node, &dfl_port_ops_list); node 220 drivers/fpga/dfl.c list_del(&ops->node); node 450 drivers/fpga/dfl.c struct list_head node; node 459 drivers/fpga/dfl.c list_add(&pdata->node, &cdev->port_dev_list); node 521 drivers/fpga/dfl.c list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) { node 530 drivers/fpga/dfl.c list_del(&finfo->node); node 605 drivers/fpga/dfl.c list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) { node 606 drivers/fpga/dfl.c list_del(&finfo->node); node 672 drivers/fpga/dfl.c list_add_tail(&finfo->node, &binfo->sub_features); node 851 drivers/fpga/dfl.c list_for_each_entry_safe(dfl, tmp, &info->dfls, node) { node 852 drivers/fpga/dfl.c list_del(&dfl->node); node 889 drivers/fpga/dfl.c list_add_tail(&dfl->node, &info->dfls); node 966 drivers/fpga/dfl.c list_for_each_entry(dfl, &info->dfls, node) { node 1002 drivers/fpga/dfl.c list_for_each_entry_safe(pdata, ptmp, &cdev->port_dev_list, node) { node 1012 drivers/fpga/dfl.c list_del(&pdata->node); node 1045 drivers/fpga/dfl.c list_for_each_entry(pdata, &cdev->port_dev_list, node) { node 1184 drivers/fpga/dfl.c list_for_each_entry(pdata, &cdev->port_dev_list, node) { node 1221 drivers/fpga/dfl.c list_for_each_entry(pdata, &cdev->port_dev_list, node) { node 160 drivers/fpga/dfl.h struct list_head node; node 228 drivers/fpga/dfl.h struct list_head node; node 392 drivers/fpga/dfl.h struct list_head node; node 167 drivers/fpga/fpga-bridge.c list_for_each_entry(bridge, bridge_list, node) { node 191 drivers/fpga/fpga-bridge.c list_for_each_entry(bridge, bridge_list, node) { node 214 drivers/fpga/fpga-bridge.c list_for_each_entry_safe(bridge, next, bridge_list, node) { node 218 drivers/fpga/fpga-bridge.c list_del(&bridge->node); node 247 drivers/fpga/fpga-bridge.c list_add(&bridge->node, bridge_list); node 277 drivers/fpga/fpga-bridge.c list_add(&bridge->node, bridge_list); node 349 drivers/fpga/fpga-bridge.c INIT_LIST_HEAD(&bridge->node); node 492 drivers/fpga/fpga-mgr.c struct fpga_manager *of_fpga_mgr_get(struct device_node *node) node 496 drivers/fpga/fpga-mgr.c dev = class_find_device_by_of_node(fpga_mgr_class, node); node 113 drivers/gnss/serial.c struct device_node *node = serdev->dev.of_node; node 116 drivers/gnss/serial.c of_property_read_u32(node, "current-speed", &speed); node 385 drivers/gnss/sirf.c struct device_node *node = serdev->dev.of_node; node 388 drivers/gnss/sirf.c of_property_read_u32(node, "current-speed", &speed); node 238 drivers/gpio/gpio-altera.c struct device_node *node = pdev->dev.of_node; node 249 drivers/gpio/gpio-altera.c if (of_property_read_u32(node, "altr,ngpio", ®)) node 274 drivers/gpio/gpio-altera.c if (of_property_read_u32(node, "altr,interrupt-type", ®)) { node 304 drivers/gpio/gpio-altera.c ret = of_mm_gpiochip_add_data(node, &altera_gc->mmchip, altera_gc); node 47 drivers/gpio/gpio-brcmstb.c struct list_head node; node 304 drivers/gpio/gpio-brcmstb.c list_for_each_entry(bank, &priv->bank_list, node) node 316 drivers/gpio/gpio-brcmstb.c list_for_each_entry_reverse(bank, &priv->bank_list, node) { node 412 drivers/gpio/gpio-brcmstb.c list_for_each_entry(bank, &priv->bank_list, node) node 535 drivers/gpio/gpio-brcmstb.c list_for_each_entry(bank, &priv->bank_list, node) { node 581 drivers/gpio/gpio-brcmstb.c list_for_each_entry(bank, &priv->bank_list, node) { node 741 drivers/gpio/gpio-brcmstb.c list_add(&bank->node, &priv->bank_list); node 620 drivers/gpio/gpio-max732x.c struct device_node *node; node 627 drivers/gpio/gpio-max732x.c node = client->dev.of_node; node 629 drivers/gpio/gpio-max732x.c if (!pdata && node) node 211 drivers/gpio/gpio-mt7621.c struct device_node *node, int bank) node 222 drivers/gpio/gpio-mt7621.c rg->chip.of_node = node; node 59 drivers/gpio/gpio-mxc.c struct list_head node; node 301 drivers/gpio/gpio-mxc.c list_for_each_entry(port, &mxc_gpio_ports, node) { node 519 drivers/gpio/gpio-mxc.c list_add_tail(&port->node, &mxc_gpio_ports); node 564 drivers/gpio/gpio-mxc.c list_for_each_entry(port, &mxc_gpio_ports, node) { node 578 drivers/gpio/gpio-mxc.c list_for_each_entry(port, &mxc_gpio_ports, node) { node 1346 drivers/gpio/gpio-omap.c struct device_node *node = dev->of_node; node 1399 drivers/gpio/gpio-omap.c bank->chip.of_node = of_node_get(node); node 1402 drivers/gpio/gpio-omap.c if (node) { node 1403 drivers/gpio/gpio-omap.c if (!of_property_read_bool(node, "ti,gpio-always-on")) node 92 drivers/gpio/gpio-sodaville.c static int sdv_xlate(struct irq_domain *h, struct device_node *node, node 98 drivers/gpio/gpio-sodaville.c if (node != irq_domain_get_of_node(h)) node 24 drivers/gpio/gpio-ts4800.c struct device_node *node; node 38 drivers/gpio/gpio-ts4800.c node = pdev->dev.of_node; node 39 drivers/gpio/gpio-ts4800.c if (!node) node 42 drivers/gpio/gpio-ts4800.c retval = of_property_read_u32(node, "ngpios", &ngpios); node 500 drivers/gpio/gpio-twl4030.c struct device_node *node = pdev->dev.of_node; node 522 drivers/gpio/gpio-twl4030.c irq_domain_add_legacy(node, TWL4030_GPIO_MAX, irq_base, 0, node 539 drivers/gpio/gpio-twl4030.c if (node) node 54 drivers/gpio/gpiolib-acpi.c struct list_head node; node 66 drivers/gpio/gpiolib-acpi.c struct list_head node; node 204 drivers/gpio/gpiolib-acpi.c list_for_each_entry(event, &acpi_gpio->events, node) node 343 drivers/gpio/gpiolib-acpi.c list_add_tail(&event->node, &acpi_gpio->events); node 430 drivers/gpio/gpiolib-acpi.c list_for_each_entry_safe_reverse(event, ep, &acpi_gpio->events, node) { node 440 drivers/gpio/gpiolib-acpi.c list_del(&event->node); node 1011 drivers/gpio/gpiolib-acpi.c list_for_each_entry(conn, &achip->conns, node) { node 1028 drivers/gpio/gpiolib-acpi.c list_for_each_entry(event, &achip->events, node) { node 1060 drivers/gpio/gpiolib-acpi.c list_add_tail(&conn->node, &achip->conns); node 1108 drivers/gpio/gpiolib-acpi.c list_for_each_entry_safe_reverse(conn, tmp, &achip->conns, node) { node 1110 drivers/gpio/gpiolib-acpi.c list_del(&conn->node); node 148 drivers/gpio/gpiolib-devres.c struct device_node *node, node 156 drivers/gpio/gpiolib-devres.c desc = gpiod_get_from_of_node(node, propname, index, dflags, label); node 316 drivers/gpio/gpiolib-of.c struct gpio_desc *gpiod_get_from_of_node(struct device_node *node, node 330 drivers/gpio/gpiolib-of.c desc = of_get_named_gpiod_flags(node, propname, node 2581 drivers/gpio/gpiolib.c list_add_tail(&pin_range->node, &gdev->pin_ranges); node 2638 drivers/gpio/gpiolib.c list_add_tail(&pin_range->node, &gdev->pin_ranges); node 2653 drivers/gpio/gpiolib.c list_for_each_entry_safe(pin_range, tmp, &gdev->pin_ranges, node) { node 2654 drivers/gpio/gpiolib.c list_del(&pin_range->node); node 857 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 858 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_device *dev = node->minor->dev; node 897 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 898 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_device *dev = node->minor->dev; node 907 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 908 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_device *dev = node->minor->dev; node 917 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 918 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct drm_device *dev = node->minor->dev; node 983 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c list_for_each_entry(s_job, &sched->ring_mirror_list, node) { node 1011 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c list_for_each_entry(s_job, &sched->ring_mirror_list, node) { node 699 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 700 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c struct drm_device *dev = node->minor->dev; node 748 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 749 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c struct drm_device *dev = node->minor->dev; node 849 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 850 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct drm_device *dev = node->minor->dev; node 34 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct drm_mm_node node; node 153 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct amdgpu_gtt_node *node = mem->mm_node; node 155 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c return (node->node.start != AMDGPU_BO_INVALID_OFFSET); node 175 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct amdgpu_gtt_node *node = mem->mm_node; node 198 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c r = drm_mm_insert_node_in_range(&mgr->mm, &node->node, mem->num_pages, node 204 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c mem->start = node->node.start; node 225 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct amdgpu_gtt_node *node; node 237 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 238 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c if (!node) { node 243 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node->node.start = AMDGPU_BO_INVALID_OFFSET; node 244 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node->node.size = mem->num_pages; node 245 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node->tbo = tbo; node 246 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c mem->mm_node = node; node 251 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c kfree(node); node 257 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c mem->start = node->node.start; node 281 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct amdgpu_gtt_node *node = mem->mm_node; node 283 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c if (!node) node 287 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c if (node->node.start != AMDGPU_BO_INVALID_OFFSET) node 288 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c drm_mm_remove_node(&node->node); node 292 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c kfree(node); node 314 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct amdgpu_gtt_node *node; node 320 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node = container_of(mm_node, struct amdgpu_gtt_node, node); node 321 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c r = amdgpu_ttm_recover_gart(node->tbo); node 408 drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 409 drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c struct drm_device *dev = node->minor->dev; node 1226 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 1227 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c struct drm_device *dev = node->minor->dev; node 77 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_mn_node *node, *next_node; node 82 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c hash_del(&amn->node); node 83 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c rbtree_postorder_for_each_entry_safe(node, next_node, node 85 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_for_each_entry_safe(bo, next_bo, &node->bos, mn_list) { node 89 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c kfree(node); node 170 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c static void amdgpu_mn_invalidate_node(struct amdgpu_mn_node *node, node 177 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_for_each_entry(bo, &node->bos, mn_list) { node 219 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_mn_node *node; node 226 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node = container_of(it, struct amdgpu_mn_node, it); node 229 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c amdgpu_mn_invalidate_node(node, start, end); node 265 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_mn_node *node; node 273 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node = container_of(it, struct amdgpu_mn_node, it); node 276 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_for_each_entry(bo, &node->bos, mn_list) { node 329 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c hash_for_each_possible(adev->mn_hash, amn, node, key) node 350 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c hash_add(adev->mn_hash, &amn->node, AMDGPU_MN_KEY(mm, type)); node 382 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_mn_node *node = NULL, *new_node; node 399 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c kfree(node); node 400 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node = container_of(it, struct amdgpu_mn_node, it); node 401 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c interval_tree_remove(&node->it, &amn->objects); node 404 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_splice(&node->bos, &bos); node 407 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c if (!node) node 408 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node = new_node; node 414 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node->it.start = addr; node 415 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node->it.last = end; node 416 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c INIT_LIST_HEAD(&node->bos); node 417 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_splice(&bos, &node->bos); node 418 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c list_add(&bo->mn_list, &node->bos); node 420 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c interval_tree_insert(&node->it, &amn->objects); node 457 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_mn_node *node; node 459 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c node = container_of(head, struct amdgpu_mn_node, bos); node 460 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c interval_tree_remove(&node->it, &amn->objects); node 461 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c kfree(node); node 62 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.h struct hlist_node node; node 204 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct drm_mm_node *node = bo->tbo.mem.mm_node; node 211 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h pages_left -= node->size, node++) node 212 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h if (node->start < fpfn) node 3145 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 3146 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c struct drm_device *dev = node->minor->dev; node 318 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_del(&obj->node); node 344 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_add(&obj->node, &con->head); node 538 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry_safe(obj, tmp, &con->head, node) { node 695 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry(obj, &con->head, node) { node 914 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry_safe(obj, tmp, &con->head, node) { node 973 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry_safe(obj, tmp, &con->head, node) { node 1143 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry_safe(obj, tmp, &con->head, node) { node 1505 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c list_for_each_entry_safe(obj, tmp, &con->head, node) { node 391 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h struct list_head node; node 36 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c struct hlist_node node; node 136 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_possible(sync->fences, e, node, f->context) { node 177 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_add(sync->fences, &e->node, f->context); node 265 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { node 270 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_del(&e->node); node 307 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { node 313 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_del(&e->node); node 340 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(source->fences, i, tmp, e, node) { node 347 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_del(&e->node); node 365 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { node 370 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_del(&e->node); node 391 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { node 392 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_del(&e->node); node 1478 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct drm_mm_node *node = bo->mem.mm_node; node 1511 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (place->fpfn < (node->start + node->size) && node 1512 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c !(place->lpfn && place->lpfn <= node->start)) node 1515 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c num_pages -= node->size; node 1516 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c ++node; node 2133 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 2134 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c unsigned ttm_pl = (uintptr_t)node->info_ent->data; node 2135 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct drm_device *dev = node->minor->dev; node 60 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define START(node) ((node)->start) node 61 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define LAST(node) ((node)->last) node 195 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct drm_mm_node *node) node 197 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c uint64_t start = node->start << PAGE_SHIFT; node 198 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c uint64_t end = (node->size + node->start) << PAGE_SHIFT; node 245 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct drm_mm_node *node) node 249 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c start = node->start + node->size; node 170 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c char node[10] = { 0 }; node 195 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c sprintf(node, "node%d", hive->number_devices); node 197 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c ret = sysfs_create_link(hive->kobj, &adev->dev->kobj, node); node 1029 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 1030 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_device *dev = node->minor->dev; node 1046 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 1047 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_device *dev = node->minor->dev; node 1058 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 1059 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c struct drm_device *dev = node->minor->dev; node 115 drivers/gpu/drm/arc/arcpgu_drv.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 116 drivers/gpu/drm/arc/arcpgu_drv.c struct drm_device *drm = node->minor->dev; node 105 drivers/gpu/drm/arm/display/komeda/komeda_kms.c struct komeda_plane_state *node, *last; node 120 drivers/gpu/drm/arm/display/komeda/komeda_kms.c list_for_each_entry(node, zorder_list, zlist_node) { node 121 drivers/gpu/drm/arm/display/komeda/komeda_kms.c if (new->base.zpos < node->base.zpos) { node 122 drivers/gpu/drm/arm/display/komeda/komeda_kms.c list_add_tail(&new->zlist_node, &node->zlist_node); node 124 drivers/gpu/drm/arm/display/komeda/komeda_kms.c } else if (node->base.zpos == new->base.zpos) { node 125 drivers/gpu/drm/arm/display/komeda/komeda_kms.c struct drm_plane *a = node->base.plane; node 132 drivers/gpu/drm/arm/display/komeda/komeda_kms.c a->name, b->name, node->base.zpos); node 198 drivers/gpu/drm/arm/hdlcd_drv.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 199 drivers/gpu/drm/arm/hdlcd_drv.c struct drm_device *drm = node->minor->dev; node 211 drivers/gpu/drm/arm/hdlcd_drv.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 212 drivers/gpu/drm/arm/hdlcd_drv.c struct drm_device *drm = node->minor->dev; node 20 drivers/gpu/drm/armada/armada_debugfs.c struct drm_info_node *node = m->private; node 21 drivers/gpu/drm/armada/armada_debugfs.c struct drm_device *dev = node->minor->dev; node 132 drivers/gpu/drm/armada/armada_gem.c struct drm_mm_node *node; node 137 drivers/gpu/drm/armada/armada_gem.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 138 drivers/gpu/drm/armada/armada_gem.c if (!node) node 142 drivers/gpu/drm/armada/armada_gem.c ret = drm_mm_insert_node_generic(&priv->linear, node, node 146 drivers/gpu/drm/armada/armada_gem.c kfree(node); node 150 drivers/gpu/drm/armada/armada_gem.c obj->linear = node; node 147 drivers/gpu/drm/bridge/adv7511/adv7533.c .node = NULL, node 39 drivers/gpu/drm/bridge/analogix/analogix_dp_core.c struct device_node *node; node 111 drivers/gpu/drm/bridge/sil-sii8620.c struct list_head node; node 270 drivers/gpu/drm/bridge/sil-sii8620.c list_for_each_entry_safe(msg, n, &ctx->mt_queue, node) { node 271 drivers/gpu/drm/bridge/sil-sii8620.c list_del(&msg->node); node 289 drivers/gpu/drm/bridge/sil-sii8620.c node); node 290 drivers/gpu/drm/bridge/sil-sii8620.c list_del(&msg->node); node 302 drivers/gpu/drm/bridge/sil-sii8620.c msg = list_first_entry(&ctx->mt_queue, struct sii8620_mt_msg, node); node 391 drivers/gpu/drm/bridge/sil-sii8620.c list_add_tail(&msg->node, &ctx->mt_queue); node 407 drivers/gpu/drm/bridge/sil-sii8620.c msg = list_last_entry(&ctx->mt_queue, struct sii8620_mt_msg, node); node 1813 drivers/gpu/drm/bridge/sil-sii8620.c return list_first_entry(&ctx->mt_queue, struct sii8620_mt_msg, node); node 274 drivers/gpu/drm/bridge/ti-sn65dsi86.c .node = NULL, node 1511 drivers/gpu/drm/drm_atomic.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 1512 drivers/gpu/drm/drm_atomic.c struct drm_device *dev = node->minor->dev; node 443 drivers/gpu/drm/drm_client.c struct drm_info_node *node = m->private; node 444 drivers/gpu/drm/drm_client.c struct drm_device *dev = node->minor->dev; node 53 drivers/gpu/drm/drm_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 54 drivers/gpu/drm/drm_debugfs.c struct drm_minor *minor = node->minor; node 75 drivers/gpu/drm/drm_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 76 drivers/gpu/drm/drm_debugfs.c struct drm_device *dev = node->minor->dev; node 126 drivers/gpu/drm/drm_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 127 drivers/gpu/drm/drm_debugfs.c struct drm_device *dev = node->minor->dev; node 148 drivers/gpu/drm/drm_debugfs.c struct drm_info_node *node = inode->i_private; node 150 drivers/gpu/drm/drm_debugfs.c return single_open(file, node->info_ent->show, node); node 287 drivers/gpu/drm/drm_debugfs.c struct drm_info_node *node, *tmp; node 290 drivers/gpu/drm/drm_debugfs.c list_for_each_entry_safe(node, tmp, &minor->debugfs_list, list) { node 291 drivers/gpu/drm/drm_debugfs.c debugfs_remove(node->dent); node 292 drivers/gpu/drm/drm_debugfs.c list_del(&node->list); node 293 drivers/gpu/drm/drm_debugfs.c kfree(node); node 1842 drivers/gpu/drm/drm_fb_helper.c info->node, info->fix.id); node 345 drivers/gpu/drm/drm_file.c struct pci_bus, node); node 63 drivers/gpu/drm/drm_flip_work.c list_add_tail(&task->node, &work->queued); node 132 drivers/gpu/drm/drm_flip_work.c list_for_each_entry_safe(task, tmp, &tasks, node) { node 1069 drivers/gpu/drm/drm_framebuffer.c struct drm_info_node *node = m->private; node 1070 drivers/gpu/drm/drm_framebuffer.c struct drm_device *dev = node->minor->dev; node 1151 drivers/gpu/drm/drm_gem.c struct drm_vma_offset_node *node; node 1158 drivers/gpu/drm/drm_gem.c node = drm_vma_offset_exact_lookup_locked(dev->vma_offset_manager, node 1161 drivers/gpu/drm/drm_gem.c if (likely(node)) { node 1162 drivers/gpu/drm/drm_gem.c obj = container_of(node, struct drm_gem_object, vma_node); node 1181 drivers/gpu/drm/drm_gem.c if (!drm_vma_node_is_allowed(node, priv)) { node 1186 drivers/gpu/drm/drm_gem.c if (node->readonly) { node 1195 drivers/gpu/drm/drm_gem.c ret = drm_gem_mmap_obj(obj, drm_vma_node_size(node) << PAGE_SHIFT, node 352 drivers/gpu/drm/drm_gem_cma_helper.c struct drm_vma_offset_node *node; node 358 drivers/gpu/drm/drm_gem_cma_helper.c node = drm_vma_offset_exact_lookup_locked(dev->vma_offset_manager, node 361 drivers/gpu/drm/drm_gem_cma_helper.c if (likely(node)) { node 362 drivers/gpu/drm/drm_gem_cma_helper.c obj = container_of(node, struct drm_gem_object, vma_node); node 382 drivers/gpu/drm/drm_gem_cma_helper.c if (!drm_vma_node_is_allowed(node, priv)) { node 155 drivers/gpu/drm/drm_mipi_dsi.c of_mipi_dsi_device_add(struct mipi_dsi_host *host, struct device_node *node) node 162 drivers/gpu/drm/drm_mipi_dsi.c if (of_modalias_node(node, info.type, sizeof(info.type)) < 0) { node 163 drivers/gpu/drm/drm_mipi_dsi.c dev_err(dev, "modalias failure on %pOF\n", node); node 167 drivers/gpu/drm/drm_mipi_dsi.c ret = of_property_read_u32(node, "reg", ®); node 170 drivers/gpu/drm/drm_mipi_dsi.c node, ret); node 175 drivers/gpu/drm/drm_mipi_dsi.c info.node = of_node_get(node); node 181 drivers/gpu/drm/drm_mipi_dsi.c of_mipi_dsi_device_add(struct mipi_dsi_host *host, struct device_node *node) node 224 drivers/gpu/drm/drm_mipi_dsi.c dsi->dev.of_node = info->node; node 261 drivers/gpu/drm/drm_mipi_dsi.c struct mipi_dsi_host *of_find_mipi_dsi_host_by_node(struct device_node *node) node 268 drivers/gpu/drm/drm_mipi_dsi.c if (host->dev->of_node == node) { node 282 drivers/gpu/drm/drm_mipi_dsi.c struct device_node *node; node 284 drivers/gpu/drm/drm_mipi_dsi.c for_each_available_child_of_node(host->dev->of_node, node) { node 286 drivers/gpu/drm/drm_mipi_dsi.c if (!of_find_property(node, "reg", NULL)) node 288 drivers/gpu/drm/drm_mipi_dsi.c of_mipi_dsi_device_add(host, node); node 107 drivers/gpu/drm/drm_mm.c static noinline void save_stack(struct drm_mm_node *node) node 115 drivers/gpu/drm/drm_mm.c node->stack = stack_depot_save(entries, n, GFP_NOWAIT); node 120 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node; node 129 drivers/gpu/drm/drm_mm.c list_for_each_entry(node, drm_mm_nodes(mm), node_list) { node 130 drivers/gpu/drm/drm_mm.c if (!node->stack) { node 132 drivers/gpu/drm/drm_mm.c node->start, node->size); node 136 drivers/gpu/drm/drm_mm.c nr_entries = stack_depot_fetch(node->stack, &entries); node 139 drivers/gpu/drm/drm_mm.c node->start, node->size, buf); node 148 drivers/gpu/drm/drm_mm.c static void save_stack(struct drm_mm_node *node) { } node 152 drivers/gpu/drm/drm_mm.c #define START(node) ((node)->start) node 153 drivers/gpu/drm/drm_mm.c #define LAST(node) ((node)->start + (node)->size - 1) node 168 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node) node 175 drivers/gpu/drm/drm_mm.c node->__subtree_last = LAST(node); node 181 drivers/gpu/drm/drm_mm.c if (parent->__subtree_last >= node->__subtree_last) node 184 drivers/gpu/drm/drm_mm.c parent->__subtree_last = node->__subtree_last; node 200 drivers/gpu/drm/drm_mm.c if (parent->__subtree_last < node->__subtree_last) node 201 drivers/gpu/drm/drm_mm.c parent->__subtree_last = node->__subtree_last; node 202 drivers/gpu/drm/drm_mm.c if (node->start < parent->start) { node 210 drivers/gpu/drm/drm_mm.c rb_link_node(&node->rb, rb, link); node 211 drivers/gpu/drm/drm_mm.c rb_insert_augmented_cached(&node->rb, &mm->interval_tree, leftmost, node 217 drivers/gpu/drm/drm_mm.c u64 x = expr(node); \ node 225 drivers/gpu/drm/drm_mm.c rb_link_node(&node->member, rb, link); \ node 226 drivers/gpu/drm/drm_mm.c rb_insert_color(&node->member, &root); \ node 238 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node) node 241 drivers/gpu/drm/drm_mm.c u64 x = node->hole_size; node 254 drivers/gpu/drm/drm_mm.c rb_link_node(&node->rb_hole_size, rb, link); node 255 drivers/gpu/drm/drm_mm.c rb_insert_color_cached(&node->rb_hole_size, root, first); node 258 drivers/gpu/drm/drm_mm.c static void add_hole(struct drm_mm_node *node) node 260 drivers/gpu/drm/drm_mm.c struct drm_mm *mm = node->mm; node 262 drivers/gpu/drm/drm_mm.c node->hole_size = node 263 drivers/gpu/drm/drm_mm.c __drm_mm_hole_node_end(node) - __drm_mm_hole_node_start(node); node 264 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!drm_mm_hole_follows(node)); node 266 drivers/gpu/drm/drm_mm.c insert_hole_size(&mm->holes_size, node); node 269 drivers/gpu/drm/drm_mm.c list_add(&node->hole_stack, &mm->hole_stack); node 272 drivers/gpu/drm/drm_mm.c static void rm_hole(struct drm_mm_node *node) node 274 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!drm_mm_hole_follows(node)); node 276 drivers/gpu/drm/drm_mm.c list_del(&node->hole_stack); node 277 drivers/gpu/drm/drm_mm.c rb_erase_cached(&node->rb_hole_size, &node->mm->holes_size); node 278 drivers/gpu/drm/drm_mm.c rb_erase(&node->rb_hole_addr, &node->mm->holes_addr); node 279 drivers/gpu/drm/drm_mm.c node->hole_size = 0; node 281 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(drm_mm_hole_follows(node)); node 305 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node = node 308 drivers/gpu/drm/drm_mm.c if (size <= node->hole_size) { node 309 drivers/gpu/drm/drm_mm.c best = node; node 322 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node = NULL; node 327 drivers/gpu/drm/drm_mm.c node = rb_hole_addr_to_node(rb); node 328 drivers/gpu/drm/drm_mm.c hole_start = __drm_mm_hole_node_start(node); node 331 drivers/gpu/drm/drm_mm.c rb = node->rb_hole_addr.rb_left; node 332 drivers/gpu/drm/drm_mm.c else if (addr > hole_start + node->hole_size) node 333 drivers/gpu/drm/drm_mm.c rb = node->rb_hole_addr.rb_right; node 338 drivers/gpu/drm/drm_mm.c return node; node 366 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node, node 372 drivers/gpu/drm/drm_mm.c return rb_hole_size_to_node(rb_prev(&node->rb_hole_size)); node 375 drivers/gpu/drm/drm_mm.c return rb_hole_addr_to_node(rb_next(&node->rb_hole_addr)); node 378 drivers/gpu/drm/drm_mm.c return rb_hole_addr_to_node(rb_prev(&node->rb_hole_addr)); node 381 drivers/gpu/drm/drm_mm.c node = list_next_entry(node, hole_stack); node 382 drivers/gpu/drm/drm_mm.c return &node->hole_stack == &mm->hole_stack ? NULL : node; node 400 drivers/gpu/drm/drm_mm.c int drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node) node 402 drivers/gpu/drm/drm_mm.c u64 end = node->start + node->size; node 407 drivers/gpu/drm/drm_mm.c end = node->start + node->size; node 408 drivers/gpu/drm/drm_mm.c if (unlikely(end <= node->start)) node 412 drivers/gpu/drm/drm_mm.c hole = find_hole(mm, node->start); node 420 drivers/gpu/drm/drm_mm.c mm->color_adjust(hole, node->color, &adj_start, &adj_end); node 422 drivers/gpu/drm/drm_mm.c if (adj_start > node->start || adj_end < end) node 425 drivers/gpu/drm/drm_mm.c node->mm = mm; node 427 drivers/gpu/drm/drm_mm.c list_add(&node->node_list, &hole->node_list); node 428 drivers/gpu/drm/drm_mm.c drm_mm_interval_tree_add_node(hole, node); node 429 drivers/gpu/drm/drm_mm.c node->allocated = true; node 430 drivers/gpu/drm/drm_mm.c node->hole_size = 0; node 433 drivers/gpu/drm/drm_mm.c if (node->start > hole_start) node 436 drivers/gpu/drm/drm_mm.c add_hole(node); node 438 drivers/gpu/drm/drm_mm.c save_stack(node); node 465 drivers/gpu/drm/drm_mm.c struct drm_mm_node * const node, node 540 drivers/gpu/drm/drm_mm.c node->mm = mm; node 541 drivers/gpu/drm/drm_mm.c node->size = size; node 542 drivers/gpu/drm/drm_mm.c node->start = adj_start; node 543 drivers/gpu/drm/drm_mm.c node->color = color; node 544 drivers/gpu/drm/drm_mm.c node->hole_size = 0; node 546 drivers/gpu/drm/drm_mm.c list_add(&node->node_list, &hole->node_list); node 547 drivers/gpu/drm/drm_mm.c drm_mm_interval_tree_add_node(hole, node); node 548 drivers/gpu/drm/drm_mm.c node->allocated = true; node 554 drivers/gpu/drm/drm_mm.c add_hole(node); node 556 drivers/gpu/drm/drm_mm.c save_stack(node); node 572 drivers/gpu/drm/drm_mm.c void drm_mm_remove_node(struct drm_mm_node *node) node 574 drivers/gpu/drm/drm_mm.c struct drm_mm *mm = node->mm; node 577 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!node->allocated); node 578 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(node->scanned_block); node 580 drivers/gpu/drm/drm_mm.c prev_node = list_prev_entry(node, node_list); node 582 drivers/gpu/drm/drm_mm.c if (drm_mm_hole_follows(node)) node 583 drivers/gpu/drm/drm_mm.c rm_hole(node); node 585 drivers/gpu/drm/drm_mm.c drm_mm_interval_tree_remove(node, &mm->interval_tree); node 586 drivers/gpu/drm/drm_mm.c list_del(&node->node_list); node 587 drivers/gpu/drm/drm_mm.c node->allocated = false; node 725 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node) node 733 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(node->mm != mm); node 734 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!node->allocated); node 735 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(node->scanned_block); node 736 drivers/gpu/drm/drm_mm.c node->scanned_block = true; node 744 drivers/gpu/drm/drm_mm.c hole = list_prev_entry(node, node_list); node 745 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(list_next_entry(hole, node_list) != node); node 746 drivers/gpu/drm/drm_mm.c __list_del_entry(&node->node_list); node 816 drivers/gpu/drm/drm_mm.c struct drm_mm_node *node) node 820 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(node->mm != scan->mm); node 821 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!node->scanned_block); node 822 drivers/gpu/drm/drm_mm.c node->scanned_block = false; node 824 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(!node->mm->scan_active); node 825 drivers/gpu/drm/drm_mm.c node->mm->scan_active--; node 835 drivers/gpu/drm/drm_mm.c prev_node = list_prev_entry(node, node_list); node 837 drivers/gpu/drm/drm_mm.c list_next_entry(node, node_list)); node 838 drivers/gpu/drm/drm_mm.c list_add(&node->node_list, &prev_node->node_list); node 840 drivers/gpu/drm/drm_mm.c return (node->start + node->size > scan->hit_start && node 841 drivers/gpu/drm/drm_mm.c node->start < scan->hit_end); node 94 drivers/gpu/drm/drm_of.c struct device_node *node) node 96 drivers/gpu/drm/drm_of.c of_node_get(node); node 98 drivers/gpu/drm/drm_of.c compare, node); node 196 drivers/gpu/drm/drm_of.c int drm_of_encoder_active_endpoint(struct device_node *node, node 205 drivers/gpu/drm/drm_of.c if (!node || !crtc) node 208 drivers/gpu/drm/drm_of.c for_each_endpoint_of_node(node, ep) { node 142 drivers/gpu/drm/drm_syncobj.c struct list_head node; node 194 drivers/gpu/drm/drm_syncobj.c list_add_tail(&wait->node, &syncobj->cb_list); node 206 drivers/gpu/drm/drm_syncobj.c if (!wait->node.next) node 210 drivers/gpu/drm/drm_syncobj.c list_del_init(&wait->node); node 242 drivers/gpu/drm/drm_syncobj.c list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) node 275 drivers/gpu/drm/drm_syncobj.c list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) node 370 drivers/gpu/drm/drm_syncobj.c if (wait.node.next) node 866 drivers/gpu/drm/drm_syncobj.c list_del_init(&wait->node); node 144 drivers/gpu/drm/drm_vma_manager.c struct drm_mm_node *node, *best; node 152 drivers/gpu/drm/drm_vma_manager.c node = rb_entry(iter, struct drm_mm_node, rb); node 153 drivers/gpu/drm/drm_vma_manager.c offset = node->start; node 156 drivers/gpu/drm/drm_vma_manager.c best = node; node 202 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_node *node, unsigned long pages) node 208 drivers/gpu/drm/drm_vma_manager.c if (!drm_mm_node_allocated(&node->vm_node)) node 210 drivers/gpu/drm/drm_vma_manager.c &node->vm_node, pages); node 230 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_node *node) node 234 drivers/gpu/drm/drm_vma_manager.c if (drm_mm_node_allocated(&node->vm_node)) { node 235 drivers/gpu/drm/drm_vma_manager.c drm_mm_remove_node(&node->vm_node); node 236 drivers/gpu/drm/drm_vma_manager.c memset(&node->vm_node, 0, sizeof(node->vm_node)); node 263 drivers/gpu/drm/drm_vma_manager.c int drm_vma_node_allow(struct drm_vma_offset_node *node, struct drm_file *tag) node 276 drivers/gpu/drm/drm_vma_manager.c write_lock(&node->vm_lock); node 278 drivers/gpu/drm/drm_vma_manager.c iter = &node->vm_files.rb_node; node 302 drivers/gpu/drm/drm_vma_manager.c rb_insert_color(&new->vm_rb, &node->vm_files); node 306 drivers/gpu/drm/drm_vma_manager.c write_unlock(&node->vm_lock); node 325 drivers/gpu/drm/drm_vma_manager.c void drm_vma_node_revoke(struct drm_vma_offset_node *node, node 331 drivers/gpu/drm/drm_vma_manager.c write_lock(&node->vm_lock); node 333 drivers/gpu/drm/drm_vma_manager.c iter = node->vm_files.rb_node; node 338 drivers/gpu/drm/drm_vma_manager.c rb_erase(&entry->vm_rb, &node->vm_files); node 349 drivers/gpu/drm/drm_vma_manager.c write_unlock(&node->vm_lock); node 366 drivers/gpu/drm/drm_vma_manager.c bool drm_vma_node_is_allowed(struct drm_vma_offset_node *node, node 372 drivers/gpu/drm/drm_vma_manager.c read_lock(&node->vm_lock); node 374 drivers/gpu/drm/drm_vma_manager.c iter = node->vm_files.rb_node; node 385 drivers/gpu/drm/drm_vma_manager.c read_unlock(&node->vm_lock); node 193 drivers/gpu/drm/etnaviv/etnaviv_drv.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 194 drivers/gpu/drm/etnaviv/etnaviv_drv.c struct drm_device *dev = node->minor->dev; node 196 drivers/gpu/drm/etnaviv/etnaviv_drv.c node->info_ent->data; node 203 drivers/gpu/drm/etnaviv/etnaviv_drv.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 204 drivers/gpu/drm/etnaviv/etnaviv_drv.c struct drm_device *dev = node->minor->dev; node 208 drivers/gpu/drm/etnaviv/etnaviv_drv.c node->info_ent->data; node 100 drivers/gpu/drm/etnaviv/etnaviv_gem.h struct list_head node; /* GPU active submit list */ node 140 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct drm_mm_node *node, size_t size) node 154 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = drm_mm_insert_node_in_range(&context->mm, node, node 224 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct drm_mm_node *node, size_t size, u64 va) node 226 drivers/gpu/drm/etnaviv/etnaviv_mmu.c return drm_mm_insert_node_in_range(&context->mm, node, size, 0, 0, va, node 235 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct drm_mm_node *node; node 256 drivers/gpu/drm/etnaviv/etnaviv_mmu.c node = &mapping->vram_node; node 259 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_iommu_insert_exact(context, node, node 262 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_iommu_find_iova(context, node, node 267 drivers/gpu/drm/etnaviv/etnaviv_mmu.c mapping->iova = node->start; node 268 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size, node 272 drivers/gpu/drm/etnaviv/etnaviv_mmu.c drm_mm_remove_node(node); node 378 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct drm_mm_node *node = &mapping->vram_node; node 381 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_iommu_find_iova(context, node, size); node 387 drivers/gpu/drm/etnaviv/etnaviv_mmu.c mapping->iova = node->start; node 388 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_context_map(context, node->start, paddr, size, node 391 drivers/gpu/drm/etnaviv/etnaviv_mmu.c drm_mm_remove_node(node); node 410 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct drm_mm_node *node = &mapping->vram_node; node 420 drivers/gpu/drm/etnaviv/etnaviv_mmu.c etnaviv_context_unmap(context, node->start, node->size); node 421 drivers/gpu/drm/etnaviv/etnaviv_mmu.c drm_mm_remove_node(node); node 1654 drivers/gpu/drm/exynos/exynos_drm_dsi.c struct device_node *node = dev->of_node; node 1657 drivers/gpu/drm/exynos/exynos_drm_dsi.c ret = exynos_dsi_of_read_u32(node, "samsung,pll-clock-frequency", node 1662 drivers/gpu/drm/exynos/exynos_drm_dsi.c ret = exynos_dsi_of_read_u32(node, "samsung,burst-clock-frequency", node 1667 drivers/gpu/drm/exynos/exynos_drm_dsi.c ret = exynos_dsi_of_read_u32(node, "samsung,esc-clock-frequency", node 1672 drivers/gpu/drm/exynos/exynos_drm_dsi.c dsi->in_bridge_node = of_graph_get_remote_node(node, DSI_PORT_IN, 0); node 271 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node; node 287 drivers/gpu/drm/exynos/exynos_drm_g2d.c node = kcalloc(G2D_CMDLIST_NUM, sizeof(*node), GFP_KERNEL); node 288 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (!node) { node 296 drivers/gpu/drm/exynos/exynos_drm_g2d.c node[nr].cmdlist = node 298 drivers/gpu/drm/exynos/exynos_drm_g2d.c node[nr].dma_addr = node 301 drivers/gpu/drm/exynos/exynos_drm_g2d.c buf_info = &node[nr].buf_info; node 305 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_add_tail(&node[nr].list, &g2d->free_cmdlist); node 332 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node; node 341 drivers/gpu/drm/exynos/exynos_drm_g2d.c node = list_first_entry(&g2d->free_cmdlist, struct g2d_cmdlist_node, node 343 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_del_init(&node->list); node 346 drivers/gpu/drm/exynos/exynos_drm_g2d.c return node; node 349 drivers/gpu/drm/exynos/exynos_drm_g2d.c static void g2d_put_cmdlist(struct g2d_data *g2d, struct g2d_cmdlist_node *node) node 352 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_move_tail(&node->list, &g2d->free_cmdlist); node 357 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node) node 367 drivers/gpu/drm/exynos/exynos_drm_g2d.c lnode->cmdlist->data[lnode->cmdlist->last] = node->dma_addr; node 370 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_add_tail(&node->list, &file_priv->inuse_cmdlist); node 372 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (node->event) node 373 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_add_tail(&node->event->base.link, &file_priv->event_list); node 687 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node, node 691 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist *cmdlist = node->cmdlist; node 692 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_buf_info *buf_info = &node->buf_info; node 774 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node, node 777 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_buf_info *buf_info = &node->buf_info; node 807 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node = node 812 drivers/gpu/drm/exynos/exynos_drm_g2d.c writel_relaxed(node->dma_addr, g2d->regs + G2D_DMA_SFR_BASE_ADDR); node 832 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node; node 839 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_for_each_entry(node, &runqueue_node->run_cmdlist, list) node 840 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_unmap_cmdlist_gem(g2d, node, runqueue_node->filp); node 856 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_runqueue_node *node, *n; node 861 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_for_each_entry_safe(node, n, &g2d->runqueue, list) { node 862 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (file && node->filp != file) node 865 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_del_init(&node->list); node 866 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_free_runqueue_node(g2d, node); node 1021 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node, node 1024 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist *cmdlist = node->cmdlist; node 1030 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_buf_info *buf_info = &node->buf_info; node 1145 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node; node 1150 drivers/gpu/drm/exynos/exynos_drm_g2d.c node = g2d_get_cmdlist(g2d); node 1151 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (!node) node 1165 drivers/gpu/drm/exynos/exynos_drm_g2d.c node->event = NULL; node 1168 drivers/gpu/drm/exynos/exynos_drm_g2d.c e = kzalloc(sizeof(*node->event), GFP_KERNEL); node 1184 drivers/gpu/drm/exynos/exynos_drm_g2d.c node->event = e; node 1187 drivers/gpu/drm/exynos/exynos_drm_g2d.c cmdlist = node->cmdlist; node 1212 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (node->event) { node 1244 drivers/gpu/drm/exynos/exynos_drm_g2d.c ret = g2d_check_reg_offset(g2d, node, req->cmd_nr, false); node 1248 drivers/gpu/drm/exynos/exynos_drm_g2d.c node->buf_info.map_nr = req->cmd_buf_nr; node 1263 drivers/gpu/drm/exynos/exynos_drm_g2d.c ret = g2d_check_reg_offset(g2d, node, req->cmd_buf_nr, true); node 1267 drivers/gpu/drm/exynos/exynos_drm_g2d.c ret = g2d_map_cmdlist_gem(g2d, node, drm_dev, file); node 1281 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_add_cmdlist_to_inuse(file_priv, node); node 1286 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_unmap_cmdlist_gem(g2d, node, file); node 1288 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (node->event) node 1289 drivers/gpu/drm/exynos/exynos_drm_g2d.c drm_event_cancel_free(drm_dev, &node->event->base); node 1291 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_put_cmdlist(g2d, node); node 1361 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct g2d_cmdlist_node *node, *n; node 1387 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_for_each_entry_safe(node, n, &file_priv->inuse_cmdlist, list) { node 1388 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_unmap_cmdlist_gem(g2d, node, file); node 1389 drivers/gpu/drm/exynos/exynos_drm_g2d.c list_move_tail(&node->list, &g2d->free_cmdlist); node 442 drivers/gpu/drm/i915/display/intel_fbc.c struct drm_mm_node *node, node 467 drivers/gpu/drm/i915/display/intel_fbc.c ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size <<= 1, node 478 drivers/gpu/drm/i915/display/intel_fbc.c ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size >>= 1, node 240 drivers/gpu/drm/i915/display/intel_fbdev.c (unsigned long)(ggtt->gmadr.start + vma->node.start); node 241 drivers/gpu/drm/i915/display/intel_fbdev.c info->fix.smem_len = vma->node.size; node 250 drivers/gpu/drm/i915/display/intel_fbdev.c info->screen_size = vma->node.size; node 219 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c batch->node.start, batch->node.size, node 193 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (!drm_mm_node_allocated(&vma->node)) node 288 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (!drm_mm_node_allocated(&vma->node)) node 298 drivers/gpu/drm/i915/gem/i915_gem_domain.c vma->node.color = cache_level; node 491 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (!drm_mm_node_allocated(&vma->node)) node 246 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_mm_node node; /** temporary GTT binding */ node 355 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (vma->node.size < entry->pad_to_size) node 358 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->alignment && !IS_ALIGNED(vma->node.start, entry->alignment)) node 362 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vma->node.start != entry->offset) node 366 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vma->node.start < BATCH_OFFSET_BIAS) node 370 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c (vma->node.start + vma->node.size - 1) >> 32) node 388 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (vma->node.size) node 389 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c pin_flags = vma->node.start; node 547 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->offset != vma->node.start) { node 548 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset = vma->node.start | UPDATE; node 555 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (drm_mm_node_allocated(&vma->node)) node 615 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->offset != vma->node.start) { node 616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset = vma->node.start | UPDATE; node 812 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c GEM_BUG_ON(drm_mm_node_allocated(&vma->node) && node 894 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return gen8_canonical_addr((int)reloc->delta + target->node.start); node 908 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.allocated = false; node 964 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c i915_gem_object_finish_access((struct drm_i915_gem_object *)cache->node.mm); node 971 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (cache->node.allocated) { node 973 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.start, node 974 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.size); node 975 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c drm_mm_remove_node(&cache->node); node 977 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c i915_vma_unpin((struct i915_vma *)cache->node.mm); node 1005 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.mm = (void *)obj; node 1049 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c memset(&cache->node, 0, sizeof(cache->node)); node 1051 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c (&ggtt->vm.mm, &cache->node, node 1058 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.start = vma->node.start; node 1059 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c cache->node.mm = (void *)vma; node 1063 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c offset = cache->node.start; node 1064 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (cache->node.allocated) { node 1191 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c batch->node.start, PAGE_SIZE, node 1288 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c addr = gen8_canonical_addr(vma->node.start + offset); node 1345 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return target->node.start | UPDATE; node 1408 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c gen8_canonical_addr(target->node.start) == reloc->presumed_offset) node 2006 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c batch_start = gen8_canonical_addr(eb->batch->node.start) + node 2009 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c shadow_batch_start = gen8_canonical_addr(vma->node.start); node 2098 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c eb->batch->node.start + node 301 drivers/gpu/drm/i915/gem/i915_gem_mman.c (ggtt->gmadr.start + vma->node.start) >> PAGE_SHIFT, node 48 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c offset = vma->node.start; node 184 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c batch->node.start, batch->node.size, node 232 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c src_offset = src->node.start; node 233 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c dst_offset = dst->node.start; node 381 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c batch->node.start, batch->node.size, node 16 drivers/gpu/drm/i915/gem/i915_gem_pm.c struct llist_node *node, *next; node 18 drivers/gpu/drm/i915/gem/i915_gem_pm.c llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) { node 20 drivers/gpu/drm/i915/gem/i915_gem_pm.c container_of((struct list_head *)node, node 444 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c unsigned long count = vma->node.size >> PAGE_SHIFT; node 29 drivers/gpu/drm/i915/gem/i915_gem_stolen.c struct drm_mm_node *node, u64 size, node 42 drivers/gpu/drm/i915/gem/i915_gem_stolen.c ret = drm_mm_insert_node_in_range(&dev_priv->mm.stolen, node, node 51 drivers/gpu/drm/i915/gem/i915_gem_stolen.c struct drm_mm_node *node, u64 size, node 54 drivers/gpu/drm/i915/gem/i915_gem_stolen.c return i915_gem_stolen_insert_node_in_range(dev_priv, node, size, node 59 drivers/gpu/drm/i915/gem/i915_gem_stolen.c struct drm_mm_node *node) node 62 drivers/gpu/drm/i915/gem/i915_gem_stolen.c drm_mm_remove_node(node); node 677 drivers/gpu/drm/i915/gem/i915_gem_stolen.c ret = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, node 685 drivers/gpu/drm/i915/gem/i915_gem_stolen.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 16 drivers/gpu/drm/i915/gem/i915_gem_stolen.h struct drm_mm_node *node, u64 size, node 19 drivers/gpu/drm/i915/gem/i915_gem_stolen.h struct drm_mm_node *node, u64 size, node 23 drivers/gpu/drm/i915/gem/i915_gem_stolen.h struct drm_mm_node *node); node 169 drivers/gpu/drm/i915/gem/i915_gem_tiling.c if (vma->node.size < size) node 173 drivers/gpu/drm/i915/gem/i915_gem_tiling.c if (!IS_ALIGNED(vma->node.start, alignment)) node 24 drivers/gpu/drm/i915/gem/i915_gem_userptr.c struct hlist_node node; node 34 drivers/gpu/drm/i915/gem/i915_gem_userptr.c struct hlist_node node; node 335 drivers/gpu/drm/i915/gem/i915_gem_userptr.c hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real) node 378 drivers/gpu/drm/i915/gem/i915_gem_userptr.c &mm->node, (unsigned long)mm->mm); node 403 drivers/gpu/drm/i915/gem/i915_gem_userptr.c hash_del(&mm->node); node 678 drivers/gpu/drm/i915/gem/selftests/huge_pages.c if (!IS_ALIGNED(vma->node.start, node 681 drivers/gpu/drm/i915/gem/selftests/huge_pages.c vma->node.start); node 686 drivers/gpu/drm/i915/gem/selftests/huge_pages.c if (!IS_ALIGNED(vma->node.size, node 689 drivers/gpu/drm/i915/gem/selftests/huge_pages.c vma->node.size); node 834 drivers/gpu/drm/i915/gem/selftests/huge_pages.c if (!IS_ALIGNED(vma->node.start, node 837 drivers/gpu/drm/i915/gem/selftests/huge_pages.c vma->node.start); node 842 drivers/gpu/drm/i915/gem/selftests/huge_pages.c if (!IS_ALIGNED(vma->node.size, node 845 drivers/gpu/drm/i915/gem/selftests/huge_pages.c vma->node.size); node 600 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c *cmd++ = lower_32_bits(vma->node.start); node 601 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c *cmd++ = upper_32_bits(vma->node.start); node 663 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c batch->node.start, batch->node.size, node 1151 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c struct drm_mm_node *node = node 1154 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c if (!node || node->start > offset) node 1157 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c GEM_BUG_ON(offset >= node->start + node->size); node 1220 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c err = engine->emit_bb_start(rq, vma->node.start, vma->node.size, 0); node 1319 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c err = engine->emit_bb_start(rq, vma->node.start, vma->node.size, 0); node 62 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c GEM_BUG_ON(offset + (count - 1) * PAGE_SIZE > vma->node.size); node 63 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c offset += vma->node.start; node 136 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c batch->node.start, batch->node.size, node 125 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_entry_safe(cur, tmp, list, node) { node 126 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c INIT_LIST_HEAD(&cur->node); node 1329 drivers/gpu/drm/i915/gt/intel_engine_cs.c rq->batch ? upper_32_bits(rq->batch->node.start) : ~0u, node 1330 drivers/gpu/drm/i915/gt/intel_engine_cs.c rq->batch ? lower_32_bits(rq->batch->node.start) : ~0u); node 35 drivers/gpu/drm/i915/gt/intel_engine_pool.c static void node_free(struct intel_engine_pool_node *node) node 37 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_gem_object_put(node->obj); node 38 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_active_fini(&node->active); node 39 drivers/gpu/drm/i915/gt/intel_engine_pool.c kfree(node); node 44 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool_node *node = node 45 drivers/gpu/drm/i915/gt/intel_engine_pool.c container_of(ref, typeof(*node), active); node 46 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct dma_resv *resv = node->obj->base.resv; node 54 drivers/gpu/drm/i915/gt/intel_engine_pool.c err = i915_gem_object_pin_pages(node->obj); node 59 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_gem_object_make_unshrinkable(node->obj); node 66 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool_node *node = node 67 drivers/gpu/drm/i915/gt/intel_engine_pool.c container_of(ref, typeof(*node), active); node 68 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool *pool = node->pool; node 69 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct list_head *list = bucket_for_size(pool, node->obj->base.size); node 74 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_gem_object_unpin_pages(node->obj); node 77 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_gem_object_make_purgeable(node->obj); node 80 drivers/gpu/drm/i915/gt/intel_engine_pool.c list_add(&node->link, list); node 88 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool_node *node; node 91 drivers/gpu/drm/i915/gt/intel_engine_pool.c node = kmalloc(sizeof(*node), node 93 drivers/gpu/drm/i915/gt/intel_engine_pool.c if (!node) node 96 drivers/gpu/drm/i915/gt/intel_engine_pool.c node->pool = pool; node 97 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_active_init(engine->i915, &node->active, pool_active, pool_retire); node 101 drivers/gpu/drm/i915/gt/intel_engine_pool.c i915_active_fini(&node->active); node 102 drivers/gpu/drm/i915/gt/intel_engine_pool.c kfree(node); node 108 drivers/gpu/drm/i915/gt/intel_engine_pool.c node->obj = obj; node 109 drivers/gpu/drm/i915/gt/intel_engine_pool.c return node; node 115 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool_node *node; node 126 drivers/gpu/drm/i915/gt/intel_engine_pool.c list_for_each_entry(node, list, link) { node 127 drivers/gpu/drm/i915/gt/intel_engine_pool.c if (node->obj->base.size < size) node 129 drivers/gpu/drm/i915/gt/intel_engine_pool.c list_del(&node->link); node 134 drivers/gpu/drm/i915/gt/intel_engine_pool.c if (&node->link == list) { node 135 drivers/gpu/drm/i915/gt/intel_engine_pool.c node = node_create(pool, size); node 136 drivers/gpu/drm/i915/gt/intel_engine_pool.c if (IS_ERR(node)) node 137 drivers/gpu/drm/i915/gt/intel_engine_pool.c return node; node 140 drivers/gpu/drm/i915/gt/intel_engine_pool.c ret = i915_active_acquire(&node->active); node 142 drivers/gpu/drm/i915/gt/intel_engine_pool.c node_free(node); node 146 drivers/gpu/drm/i915/gt/intel_engine_pool.c return node; node 164 drivers/gpu/drm/i915/gt/intel_engine_pool.c struct intel_engine_pool_node *node, *nn; node 166 drivers/gpu/drm/i915/gt/intel_engine_pool.c list_for_each_entry_safe(node, nn, list, link) node 167 drivers/gpu/drm/i915/gt/intel_engine_pool.c node_free(node); node 18 drivers/gpu/drm/i915/gt/intel_engine_pool.h intel_engine_pool_mark_active(struct intel_engine_pool_node *node, node 21 drivers/gpu/drm/i915/gt/intel_engine_pool.h return i915_active_ref(&node->active, rq->timeline, rq); node 25 drivers/gpu/drm/i915/gt/intel_engine_pool.h intel_engine_pool_put(struct intel_engine_pool_node *node) node 27 drivers/gpu/drm/i915/gt/intel_engine_pool.h i915_active_release(&node->active); node 266 drivers/gpu/drm/i915/gt/intel_lrc.c return rb_entry(rb, struct i915_priolist, node); node 1334 drivers/gpu/drm/i915/gt/intel_lrc.c rb_erase_cached(&p->node, &execlists->queue); node 1655 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_sched_node *node, node 1658 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(!list_empty(&node->link)); node 1659 drivers/gpu/drm/i915/gt/intel_lrc.c list_add_tail(&node->link, i915_sched_lookup_priolist(engine, prio)); node 2608 drivers/gpu/drm/i915/gt/intel_lrc.c rb_erase_cached(&p->node, &execlists->queue); node 3430 drivers/gpu/drm/i915/gt/intel_lrc.c struct rb_node *node = &ve->nodes[sibling->id].rb; node 3432 drivers/gpu/drm/i915/gt/intel_lrc.c if (RB_EMPTY_NODE(node)) node 3438 drivers/gpu/drm/i915/gt/intel_lrc.c if (!RB_EMPTY_NODE(node)) node 3439 drivers/gpu/drm/i915/gt/intel_lrc.c rb_erase_cached(node, &sibling->execlists.virtual); node 3566 drivers/gpu/drm/i915/gt/intel_lrc.c struct ve_node * const node = &ve->nodes[sibling->id]; node 3571 drivers/gpu/drm/i915/gt/intel_lrc.c if (!RB_EMPTY_NODE(&node->rb)) { node 3573 drivers/gpu/drm/i915/gt/intel_lrc.c rb_erase_cached(&node->rb, node 3575 drivers/gpu/drm/i915/gt/intel_lrc.c RB_CLEAR_NODE(&node->rb); node 3583 drivers/gpu/drm/i915/gt/intel_lrc.c if (!RB_EMPTY_NODE(&node->rb)) { node 3589 drivers/gpu/drm/i915/gt/intel_lrc.c &node->rb; node 3590 drivers/gpu/drm/i915/gt/intel_lrc.c if (prio == node->prio || (prio > node->prio && first)) node 3593 drivers/gpu/drm/i915/gt/intel_lrc.c rb_erase_cached(&node->rb, &sibling->execlists.virtual); node 3612 drivers/gpu/drm/i915/gt/intel_lrc.c rb_link_node(&node->rb, rb, parent); node 3613 drivers/gpu/drm/i915/gt/intel_lrc.c rb_insert_color_cached(&node->rb, node 3618 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(RB_EMPTY_NODE(&node->rb)); node 3619 drivers/gpu/drm/i915/gt/intel_lrc.c node->prio = prio; node 3952 drivers/gpu/drm/i915/gt/intel_lrc.c struct i915_priolist *p = rb_entry(rb, typeof(*p), node); node 96 drivers/gpu/drm/i915/gt/intel_renderstate.c u64 r = s + so->vma->node.start; node 628 drivers/gpu/drm/i915/gt/intel_reset.c struct drm_vma_offset_node *node; node 640 drivers/gpu/drm/i915/gt/intel_reset.c node = &vma->obj->base.vma_node; node 643 drivers/gpu/drm/i915/gt/intel_reset.c drm_vma_node_offset_addr(node) + vma_offset, node 111 drivers/gpu/drm/i915/gt/selftest_hangcheck.c return hws->node.start + offset_in_page(sizeof(u32)*rq->fence.context); node 202 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = lower_32_bits(vma->node.start); node 203 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = upper_32_bits(vma->node.start); node 216 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = lower_32_bits(vma->node.start); node 229 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = lower_32_bits(vma->node.start); node 241 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = lower_32_bits(vma->node.start); node 256 drivers/gpu/drm/i915/gt/selftest_hangcheck.c err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags); node 1131 drivers/gpu/drm/i915/gt/selftest_hangcheck.c struct drm_mm_node evict = arg->vma->node; node 1467 drivers/gpu/drm/i915/gt/selftest_lrc.c vma->node.start, node 483 drivers/gpu/drm/i915/gt/selftest_workarounds.c u64 addr = scratch->node.start; node 569 drivers/gpu/drm/i915/gt/selftest_workarounds.c batch->node.start, PAGE_SIZE, node 786 drivers/gpu/drm/i915/gt/selftest_workarounds.c u64 offset = results->node.start + sizeof(u32) * i; node 854 drivers/gpu/drm/i915/gt/selftest_workarounds.c err = engine->emit_bb_start(rq, batch->node.start, 0, 0); node 77 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c return rb_entry(rb, struct i915_priolist, node); node 583 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c rb_erase_cached(&p->node, &execlists->queue); node 727 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c rb_erase_cached(&p->node, &execlists->queue); node 401 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c struct drm_mm_node *node = &ggtt->uc_fw; node 403 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c GEM_BUG_ON(!node->allocated); node 404 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c GEM_BUG_ON(upper_32_bits(node->start)); node 405 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c GEM_BUG_ON(upper_32_bits(node->start + node->size - 1)); node 407 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c return lower_32_bits(node->start); node 416 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c .node.start = uc_fw_ggtt_offset(uc_fw, ggtt), node 417 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c .node.size = obj->base.size, node 423 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c GEM_BUG_ON(dummy.node.size > ggtt->uc_fw.size); node 47 drivers/gpu/drm/i915/gvt/aperture_gm.c struct drm_mm_node *node; node 51 drivers/gpu/drm/i915/gvt/aperture_gm.c node = &vgpu->gm.high_gm_node; node 57 drivers/gpu/drm/i915/gvt/aperture_gm.c node = &vgpu->gm.low_gm_node; node 66 drivers/gpu/drm/i915/gvt/aperture_gm.c ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node, node 36 drivers/gpu/drm/i915/gvt/debugfs.c struct list_head node; node 49 drivers/gpu/drm/i915/gvt/debugfs.c ma = container_of(a, struct diff_mmio, node); node 50 drivers/gpu/drm/i915/gvt/debugfs.c mb = container_of(b, struct diff_mmio, node); node 63 drivers/gpu/drm/i915/gvt/debugfs.c struct diff_mmio *node; node 70 drivers/gpu/drm/i915/gvt/debugfs.c node = kmalloc(sizeof(*node), GFP_KERNEL); node 71 drivers/gpu/drm/i915/gvt/debugfs.c if (!node) node 74 drivers/gpu/drm/i915/gvt/debugfs.c node->offset = offset; node 75 drivers/gpu/drm/i915/gvt/debugfs.c node->preg = preg; node 76 drivers/gpu/drm/i915/gvt/debugfs.c node->vreg = vreg; node 77 drivers/gpu/drm/i915/gvt/debugfs.c list_add(&node->node, ¶m->diff_mmio_list); node 94 drivers/gpu/drm/i915/gvt/debugfs.c struct diff_mmio *node, *next; node 113 drivers/gpu/drm/i915/gvt/debugfs.c list_for_each_entry_safe(node, next, ¶m.diff_mmio_list, node) { node 114 drivers/gpu/drm/i915/gvt/debugfs.c u32 diff = node->preg ^ node->vreg; node 117 drivers/gpu/drm/i915/gvt/debugfs.c node->offset, node->preg, node->vreg, node 119 drivers/gpu/drm/i915/gvt/debugfs.c list_del(&node->node); node 120 drivers/gpu/drm/i915/gvt/debugfs.c kfree(node); node 89 drivers/gpu/drm/i915/gvt/handlers.c hash_for_each_possible(gvt->mmio.mmio_info_table, e, node, offset) { node 137 drivers/gpu/drm/i915/gvt/handlers.c INIT_HLIST_NODE(&info->node); node 138 drivers/gpu/drm/i915/gvt/handlers.c hash_add(gvt->mmio.mmio_info_table, &info->node, info->offset); node 3319 drivers/gpu/drm/i915/gvt/handlers.c hash_for_each_safe(gvt->mmio.mmio_info_table, i, tmp, e, node) node 3413 drivers/gpu/drm/i915/gvt/handlers.c hash_for_each(gvt->mmio.mmio_info_table, i, e, node) { node 222 drivers/gpu/drm/i915/gvt/kvmgt.c struct rb_node *node = vgpu->vdev.dma_addr_cache.rb_node; node 225 drivers/gpu/drm/i915/gvt/kvmgt.c while (node) { node 226 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, dma_addr_node); node 229 drivers/gpu/drm/i915/gvt/kvmgt.c node = node->rb_left; node 231 drivers/gpu/drm/i915/gvt/kvmgt.c node = node->rb_right; node 240 drivers/gpu/drm/i915/gvt/kvmgt.c struct rb_node *node = vgpu->vdev.gfn_cache.rb_node; node 243 drivers/gpu/drm/i915/gvt/kvmgt.c while (node) { node 244 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, gfn_node); node 247 drivers/gpu/drm/i915/gvt/kvmgt.c node = node->rb_left; node 249 drivers/gpu/drm/i915/gvt/kvmgt.c node = node->rb_right; node 317 drivers/gpu/drm/i915/gvt/kvmgt.c struct rb_node *node = NULL; node 321 drivers/gpu/drm/i915/gvt/kvmgt.c node = rb_first(&vgpu->vdev.gfn_cache); node 322 drivers/gpu/drm/i915/gvt/kvmgt.c if (!node) { node 326 drivers/gpu/drm/i915/gvt/kvmgt.c dma = rb_entry(node, struct gvt_dma, gfn_node); node 1704 drivers/gpu/drm/i915/gvt/kvmgt.c struct kvm_page_track_notifier_node *node) node 1706 drivers/gpu/drm/i915/gvt/kvmgt.c struct kvmgt_guest_info *info = container_of(node, node 1716 drivers/gpu/drm/i915/gvt/kvmgt.c struct kvm_page_track_notifier_node *node) node 1720 drivers/gpu/drm/i915/gvt/kvmgt.c struct kvmgt_guest_info *info = container_of(node, node 67 drivers/gpu/drm/i915/gvt/mmio.h struct hlist_node node; node 32 drivers/gpu/drm/i915/i915_active.c struct rb_node node; node 49 drivers/gpu/drm/i915/i915_active.c static inline struct llist_node *barrier_to_ll(struct active_node *node) node 51 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(!is_barrier(&node->base)); node 52 drivers/gpu/drm/i915/i915_active.c return (struct llist_node *)&node->base.link; node 56 drivers/gpu/drm/i915/i915_active.c __barrier_to_engine(struct active_node *node) node 58 drivers/gpu/drm/i915/i915_active.c return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev); node 62 drivers/gpu/drm/i915/i915_active.c barrier_to_engine(struct active_node *node) node 64 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(!is_barrier(&node->base)); node 65 drivers/gpu/drm/i915/i915_active.c return __barrier_to_engine(node); node 146 drivers/gpu/drm/i915/i915_active.c rbtree_postorder_for_each_entry_safe(it, n, &root, node) { node 184 drivers/gpu/drm/i915/i915_active.c struct active_node *node, *prealloc; node 195 drivers/gpu/drm/i915/i915_active.c node = READ_ONCE(ref->cache); node 196 drivers/gpu/drm/i915/i915_active.c if (node && node->timeline == idx) node 197 drivers/gpu/drm/i915/i915_active.c return &node->base; node 212 drivers/gpu/drm/i915/i915_active.c node = rb_entry(parent, struct active_node, node); node 213 drivers/gpu/drm/i915/i915_active.c if (node->timeline == idx) { node 218 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) node 224 drivers/gpu/drm/i915/i915_active.c node = prealloc; node 225 drivers/gpu/drm/i915/i915_active.c i915_active_request_init(&node->base, &tl->mutex, NULL, node_retire); node 226 drivers/gpu/drm/i915/i915_active.c node->ref = ref; node 227 drivers/gpu/drm/i915/i915_active.c node->timeline = idx; node 229 drivers/gpu/drm/i915/i915_active.c rb_link_node(&node->node, parent, p); node 230 drivers/gpu/drm/i915/i915_active.c rb_insert_color(&node->node, &ref->tree); node 233 drivers/gpu/drm/i915/i915_active.c ref->cache = node; node 236 drivers/gpu/drm/i915/i915_active.c BUILD_BUG_ON(offsetof(typeof(*node), base)); node 237 drivers/gpu/drm/i915/i915_active.c return &node->base; node 260 drivers/gpu/drm/i915/i915_active.c struct active_node *node, node 267 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(node->timeline != engine->kernel_context->timeline->fence_context); node 285 drivers/gpu/drm/i915/i915_active.c if (node == barrier_from_ll(pos)) { node 286 drivers/gpu/drm/i915/i915_active.c node = NULL; node 298 drivers/gpu/drm/i915/i915_active.c return !node; node 302 drivers/gpu/drm/i915/i915_active.c __active_del_barrier(struct i915_active *ref, struct active_node *node) node 304 drivers/gpu/drm/i915/i915_active.c return ____active_del_barrier(ref, node, barrier_to_engine(node)); node 426 drivers/gpu/drm/i915/i915_active.c rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) { node 474 drivers/gpu/drm/i915/i915_active.c rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) { node 495 drivers/gpu/drm/i915/i915_active.c static inline bool is_idle_barrier(struct active_node *node, u64 idx) node 497 drivers/gpu/drm/i915/i915_active.c return node->timeline == idx && !i915_active_request_isset(&node->base); node 518 drivers/gpu/drm/i915/i915_active.c p = &ref->cache->node; node 525 drivers/gpu/drm/i915/i915_active.c struct active_node *node = node 526 drivers/gpu/drm/i915/i915_active.c rb_entry(p, struct active_node, node); node 528 drivers/gpu/drm/i915/i915_active.c if (is_idle_barrier(node, idx)) node 532 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) node 545 drivers/gpu/drm/i915/i915_active.c struct active_node *node = node 546 drivers/gpu/drm/i915/i915_active.c rb_entry(p, struct active_node, node); node 549 drivers/gpu/drm/i915/i915_active.c if (node->timeline > idx) node 552 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) node 555 drivers/gpu/drm/i915/i915_active.c if (is_idle_barrier(node, idx)) node 565 drivers/gpu/drm/i915/i915_active.c engine = __barrier_to_engine(node); node 567 drivers/gpu/drm/i915/i915_active.c if (is_barrier(&node->base) && node 568 drivers/gpu/drm/i915/i915_active.c ____active_del_barrier(ref, node, engine)) node 578 drivers/gpu/drm/i915/i915_active.c if (p == &ref->cache->node) node 582 drivers/gpu/drm/i915/i915_active.c return rb_entry(p, struct active_node, node); node 603 drivers/gpu/drm/i915/i915_active.c struct active_node *node; node 605 drivers/gpu/drm/i915/i915_active.c node = reuse_idle_barrier(ref, idx); node 606 drivers/gpu/drm/i915/i915_active.c if (!node) { node 607 drivers/gpu/drm/i915/i915_active.c node = kmem_cache_alloc(global.slab_cache, GFP_KERNEL); node 608 drivers/gpu/drm/i915/i915_active.c if (!node) { node 614 drivers/gpu/drm/i915/i915_active.c node->base.lock = node 617 drivers/gpu/drm/i915/i915_active.c RCU_INIT_POINTER(node->base.request, NULL); node 618 drivers/gpu/drm/i915/i915_active.c node->base.retire = node_retire; node 619 drivers/gpu/drm/i915/i915_active.c node->timeline = idx; node 620 drivers/gpu/drm/i915/i915_active.c node->ref = ref; node 623 drivers/gpu/drm/i915/i915_active.c if (!i915_active_request_isset(&node->base)) { node 633 drivers/gpu/drm/i915/i915_active.c RCU_INIT_POINTER(node->base.request, ERR_PTR(-EAGAIN)); node 634 drivers/gpu/drm/i915/i915_active.c node->base.link.prev = (void *)engine; node 638 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(barrier_to_engine(node) != engine); node 639 drivers/gpu/drm/i915/i915_active.c llist_add(barrier_to_ll(node), &ref->preallocated_barriers); node 647 drivers/gpu/drm/i915/i915_active.c struct active_node *node = barrier_from_ll(pos); node 650 drivers/gpu/drm/i915/i915_active.c intel_engine_pm_put(barrier_to_engine(node)); node 652 drivers/gpu/drm/i915/i915_active.c kmem_cache_free(global.slab_cache, node); node 671 drivers/gpu/drm/i915/i915_active.c struct active_node *node = barrier_from_ll(pos); node 672 drivers/gpu/drm/i915/i915_active.c struct intel_engine_cs *engine = barrier_to_engine(node); node 682 drivers/gpu/drm/i915/i915_active.c it = rb_entry(parent, struct active_node, node); node 683 drivers/gpu/drm/i915/i915_active.c if (it->timeline < node->timeline) node 688 drivers/gpu/drm/i915/i915_active.c rb_link_node(&node->node, parent, p); node 689 drivers/gpu/drm/i915/i915_active.c rb_insert_color(&node->node, &ref->tree); node 691 drivers/gpu/drm/i915/i915_active.c llist_add(barrier_to_ll(node), &engine->barrier_tasks); node 700 drivers/gpu/drm/i915/i915_active.c struct llist_node *node, *next; node 710 drivers/gpu/drm/i915/i915_active.c llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) { node 711 drivers/gpu/drm/i915/i915_active.c RCU_INIT_POINTER(barrier_from_ll(node)->base.request, rq); node 713 drivers/gpu/drm/i915/i915_active.c list_add_tail((struct list_head *)node, &rq->active_list); node 861 drivers/gpu/drm/i915/i915_cmd_parser.c struct hlist_node node; node 908 drivers/gpu/drm/i915/i915_cmd_parser.c hash_add(engine->cmd_hash, &desc_node->node, node 922 drivers/gpu/drm/i915/i915_cmd_parser.c hash_for_each_safe(engine->cmd_hash, i, tmp, desc_node, node) { node 923 drivers/gpu/drm/i915/i915_cmd_parser.c hash_del(&desc_node->node); node 1053 drivers/gpu/drm/i915/i915_cmd_parser.c hash_for_each_possible(engine->cmd_hash, desc_node, node, node 54 drivers/gpu/drm/i915/i915_debugfs.c static inline struct drm_i915_private *node_to_i915(struct drm_info_node *node) node 56 drivers/gpu/drm/i915/i915_debugfs.c return to_i915(node->minor->dev); node 160 drivers/gpu/drm/i915/i915_debugfs.c if (!drm_mm_node_allocated(&vma->node)) node 170 drivers/gpu/drm/i915/i915_debugfs.c vma->node.start, vma->node.size, node 254 drivers/gpu/drm/i915/i915_debugfs.c if (!drm_mm_node_allocated(&vma->node)) node 258 drivers/gpu/drm/i915/i915_debugfs.c stats->active += vma->node.size; node 260 drivers/gpu/drm/i915/i915_debugfs.c stats->inactive += vma->node.size; node 263 drivers/gpu/drm/i915/i915_debugfs.c stats->closed += vma->node.size; node 274 drivers/gpu/drm/i915/i915_debugfs.c if (drm_mm_node_allocated(&vma->node)) { node 276 drivers/gpu/drm/i915/i915_debugfs.c stats->active += vma->node.size; node 278 drivers/gpu/drm/i915/i915_debugfs.c stats->inactive += vma->node.size; node 281 drivers/gpu/drm/i915/i915_debugfs.c stats->closed += vma->node.size; node 1967 drivers/gpu/drm/i915/i915_debugfs.c struct drm_info_node *node = m->private; node 1968 drivers/gpu/drm/i915/i915_debugfs.c struct drm_i915_private *dev_priv = node_to_i915(node); node 1969 drivers/gpu/drm/i915/i915_debugfs.c bool dump_load_err = !!node->info_ent->data; node 2386 drivers/gpu/drm/i915/i915_drv.h struct drm_mm_node *node, node 66 drivers/gpu/drm/i915/i915_gem.c struct drm_mm_node *node, u32 size) node 68 drivers/gpu/drm/i915/i915_gem.c memset(node, 0, sizeof(*node)); node 69 drivers/gpu/drm/i915/i915_gem.c return drm_mm_insert_node_in_range(&ggtt->vm.mm, node, node 76 drivers/gpu/drm/i915/i915_gem.c remove_mappable_node(struct drm_mm_node *node) node 78 drivers/gpu/drm/i915/i915_gem.c drm_mm_remove_node(node); node 95 drivers/gpu/drm/i915/i915_gem.c pinned += vma->node.size; node 334 drivers/gpu/drm/i915/i915_gem.c struct drm_mm_node node; node 353 drivers/gpu/drm/i915/i915_gem.c node.start = i915_ggtt_offset(vma); node 354 drivers/gpu/drm/i915/i915_gem.c node.allocated = false; node 356 drivers/gpu/drm/i915/i915_gem.c ret = insert_mappable_node(ggtt, &node, PAGE_SIZE); node 359 drivers/gpu/drm/i915/i915_gem.c GEM_BUG_ON(!node.allocated); node 392 drivers/gpu/drm/i915/i915_gem.c u32 page_base = node.start; node 396 drivers/gpu/drm/i915/i915_gem.c if (node.allocated) { node 399 drivers/gpu/drm/i915/i915_gem.c node.start, I915_CACHE_NONE, 0); node 418 drivers/gpu/drm/i915/i915_gem.c if (node.allocated) { node 419 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.clear_range(&ggtt->vm, node.start, node.size); node 420 drivers/gpu/drm/i915/i915_gem.c remove_mappable_node(&node); node 527 drivers/gpu/drm/i915/i915_gem.c struct drm_mm_node node; node 563 drivers/gpu/drm/i915/i915_gem.c node.start = i915_ggtt_offset(vma); node 564 drivers/gpu/drm/i915/i915_gem.c node.allocated = false; node 566 drivers/gpu/drm/i915/i915_gem.c ret = insert_mappable_node(ggtt, &node, PAGE_SIZE); node 569 drivers/gpu/drm/i915/i915_gem.c GEM_BUG_ON(!node.allocated); node 603 drivers/gpu/drm/i915/i915_gem.c u32 page_base = node.start; node 607 drivers/gpu/drm/i915/i915_gem.c if (node.allocated) { node 612 drivers/gpu/drm/i915/i915_gem.c node.start, I915_CACHE_NONE, 0); node 639 drivers/gpu/drm/i915/i915_gem.c if (node.allocated) { node 640 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.clear_range(&ggtt->vm, node.start, node.size); node 641 drivers/gpu/drm/i915/i915_gem.c remove_mappable_node(&node); node 65 drivers/gpu/drm/i915/i915_gem_evict.c return drm_mm_scan_add_block(scan, &vma->node); node 102 drivers/gpu/drm/i915/i915_gem_evict.c struct drm_mm_node *node; node 181 drivers/gpu/drm/i915/i915_gem_evict.c ret = drm_mm_scan_remove_block(&scan, &vma->node); node 227 drivers/gpu/drm/i915/i915_gem_evict.c if (drm_mm_scan_remove_block(&scan, &vma->node)) node 241 drivers/gpu/drm/i915/i915_gem_evict.c while (ret == 0 && (node = drm_mm_scan_color_evict(&scan))) { node 242 drivers/gpu/drm/i915/i915_gem_evict.c vma = container_of(node, struct i915_vma, node); node 265 drivers/gpu/drm/i915/i915_gem_evict.c struct drm_mm_node *node; node 297 drivers/gpu/drm/i915/i915_gem_evict.c drm_mm_for_each_node_in_range(node, &vm->mm, start, end) { node 299 drivers/gpu/drm/i915/i915_gem_evict.c if (node->color == I915_COLOR_UNEVICTABLE) { node 304 drivers/gpu/drm/i915/i915_gem_evict.c GEM_BUG_ON(!node->allocated); node 305 drivers/gpu/drm/i915/i915_gem_evict.c vma = container_of(node, typeof(*vma), node); node 314 drivers/gpu/drm/i915/i915_gem_evict.c if (node->start + node->size == target->start) { node 315 drivers/gpu/drm/i915/i915_gem_evict.c if (node->color == target->color) node 318 drivers/gpu/drm/i915/i915_gem_evict.c if (node->start == target->start + target->size) { node 319 drivers/gpu/drm/i915/i915_gem_evict.c if (node->color == target->color) node 85 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(!IS_ALIGNED(vma->node.start, I965_FENCE_PAGE)); node 89 drivers/gpu/drm/i915/i915_gem_fence_reg.c val = (vma->node.start + vma->fence_size - I965_FENCE_PAGE) << 32; node 90 drivers/gpu/drm/i915/i915_gem_fence_reg.c val |= vma->node.start; node 131 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(vma->node.start & ~I915_FENCE_START_MASK); node 133 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(!IS_ALIGNED(vma->node.start, vma->fence_size)); node 141 drivers/gpu/drm/i915/i915_gem_fence_reg.c val = vma->node.start; node 169 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(vma->node.start & ~I830_FENCE_START_MASK); node 172 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(!IS_ALIGNED(vma->node.start, vma->fence_size)); node 174 drivers/gpu/drm/i915/i915_gem_fence_reg.c val = vma->node.start; node 154 drivers/gpu/drm/i915/i915_gem_gtt.c vma->node.start, vma->size); node 171 drivers/gpu/drm/i915/i915_gem_gtt.c vma->vm->clear_range(vma->vm, vma->node.start, vma->size); node 1220 drivers/gpu/drm/i915/i915_gem_gtt.c u64 start = vma->node.start; node 1300 drivers/gpu/drm/i915/i915_gem_gtt.c !iter->sg && IS_ALIGNED(vma->node.start + node 1301 drivers/gpu/drm/i915/i915_gem_gtt.c vma->node.size, node 1345 drivers/gpu/drm/i915/i915_gem_gtt.c u64 idx = vma->node.start >> GEN8_PTE_SHIFT; node 1650 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned first_entry = vma->node.start / I915_GTT_PAGE_SIZE; node 2206 drivers/gpu/drm/i915/i915_gem_gtt.c gtt_entries += vma->node.start / I915_GTT_PAGE_SIZE; node 2245 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int i = vma->node.start / I915_GTT_PAGE_SIZE; node 2419 drivers/gpu/drm/i915/i915_gem_gtt.c intel_gtt_insert_sg_entries(vma->pages, vma->node.start >> PAGE_SHIFT, node 2464 drivers/gpu/drm/i915/i915_gem_gtt.c vma->vm->clear_range(vma->vm, vma->node.start, vma->size); node 2485 drivers/gpu/drm/i915/i915_gem_gtt.c vma->node.start, node 2516 drivers/gpu/drm/i915/i915_gem_gtt.c vm->clear_range(vm, vma->node.start, vma->size); node 2523 drivers/gpu/drm/i915/i915_gem_gtt.c vm->clear_range(vm, vma->node.start, vma->size); node 2560 drivers/gpu/drm/i915/i915_gem_gtt.c static void i915_gtt_color_adjust(const struct drm_mm_node *node, node 2565 drivers/gpu/drm/i915/i915_gem_gtt.c if (node->allocated && node->color != color) node 2573 drivers/gpu/drm/i915/i915_gem_gtt.c node = list_next_entry(node, node_list); node 2574 drivers/gpu/drm/i915/i915_gem_gtt.c if (node->color != color) node 3632 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_mm_node *node, node 3643 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(drm_mm_node_allocated(node)); node 3645 drivers/gpu/drm/i915/i915_gem_gtt.c node->size = size; node 3646 drivers/gpu/drm/i915/i915_gem_gtt.c node->start = offset; node 3647 drivers/gpu/drm/i915/i915_gem_gtt.c node->color = color; node 3649 drivers/gpu/drm/i915/i915_gem_gtt.c err = drm_mm_reserve_node(&vm->mm, node); node 3656 drivers/gpu/drm/i915/i915_gem_gtt.c err = i915_gem_evict_for_node(vm, node, flags); node 3658 drivers/gpu/drm/i915/i915_gem_gtt.c err = drm_mm_reserve_node(&vm->mm, node); node 3723 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_mm_node *node, node 3740 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(drm_mm_node_allocated(node)); node 3764 drivers/gpu/drm/i915/i915_gem_gtt.c err = drm_mm_insert_node_in_range(&vm->mm, node, node 3771 drivers/gpu/drm/i915/i915_gem_gtt.c err = drm_mm_insert_node_in_range(&vm->mm, node, node 3807 drivers/gpu/drm/i915/i915_gem_gtt.c err = i915_gem_gtt_reserve(vm, node, size, offset, color, flags); node 3820 drivers/gpu/drm/i915/i915_gem_gtt.c return drm_mm_insert_node_in_range(&vm->mm, node, node 593 drivers/gpu/drm/i915/i915_gem_gtt.h struct drm_mm_node *node, node 598 drivers/gpu/drm/i915/i915_gem_gtt.h struct drm_mm_node *node, node 985 drivers/gpu/drm/i915/i915_gpu_error.c dst->gtt_offset = vma->node.start; node 986 drivers/gpu/drm/i915/i915_gpu_error.c dst->gtt_size = vma->node.size; node 1350 drivers/gpu/drm/i915/i915_gpu_error.c .node = { .start = U64_MAX, .size = obj->base.size }, node 986 drivers/gpu/drm/i915/i915_pmu.c static int i915_pmu_cpu_online(unsigned int cpu, struct hlist_node *node) node 988 drivers/gpu/drm/i915/i915_pmu.c struct i915_pmu *pmu = hlist_entry_safe(node, typeof(*pmu), node); node 999 drivers/gpu/drm/i915/i915_pmu.c static int i915_pmu_cpu_offline(unsigned int cpu, struct hlist_node *node) node 1001 drivers/gpu/drm/i915/i915_pmu.c struct i915_pmu *pmu = hlist_entry_safe(node, typeof(*pmu), node); node 1033 drivers/gpu/drm/i915/i915_pmu.c ret = cpuhp_state_add_instance(slot, &pmu->node); node 1046 drivers/gpu/drm/i915/i915_pmu.c WARN_ON(cpuhp_state_remove_instance(cpuhp_slot, &pmu->node)); node 44 drivers/gpu/drm/i915/i915_pmu.h struct hlist_node node; node 47 drivers/gpu/drm/i915/i915_priolist_types.h struct rb_node node; node 23 drivers/gpu/drm/i915/i915_scheduler.c node_to_request(const struct i915_sched_node *node) node 25 drivers/gpu/drm/i915/i915_scheduler.c return container_of(node, const struct i915_request, sched); node 28 drivers/gpu/drm/i915/i915_scheduler.c static inline bool node_started(const struct i915_sched_node *node) node 30 drivers/gpu/drm/i915/i915_scheduler.c return i915_request_started(node_to_request(node)); node 33 drivers/gpu/drm/i915/i915_scheduler.c static inline bool node_signaled(const struct i915_sched_node *node) node 35 drivers/gpu/drm/i915/i915_scheduler.c return i915_request_completed(node_to_request(node)); node 40 drivers/gpu/drm/i915/i915_scheduler.c return rb_entry(rb, struct i915_priolist, node); node 130 drivers/gpu/drm/i915/i915_scheduler.c rb_link_node(&p->node, rb, parent); node 131 drivers/gpu/drm/i915/i915_scheduler.c rb_insert_color_cached(&p->node, &execlists->queue, first); node 149 drivers/gpu/drm/i915/i915_scheduler.c sched_lock_engine(const struct i915_sched_node *node, node 153 drivers/gpu/drm/i915/i915_scheduler.c const struct i915_request *rq = node_to_request(node); node 230 drivers/gpu/drm/i915/i915_scheduler.c static void __i915_schedule(struct i915_sched_node *node, node 244 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= READ_ONCE(node->attr.priority)) node 247 drivers/gpu/drm/i915/i915_scheduler.c if (node_signaled(node)) node 250 drivers/gpu/drm/i915/i915_scheduler.c stack.signaler = node; node 272 drivers/gpu/drm/i915/i915_scheduler.c struct i915_sched_node *node = dep->signaler; node 275 drivers/gpu/drm/i915/i915_scheduler.c if (node_started(node)) node 284 drivers/gpu/drm/i915/i915_scheduler.c list_for_each_entry(p, &node->signalers_list, signal_link) { node 301 drivers/gpu/drm/i915/i915_scheduler.c if (node->attr.priority == I915_PRIORITY_INVALID) { node 302 drivers/gpu/drm/i915/i915_scheduler.c GEM_BUG_ON(!list_empty(&node->link)); node 303 drivers/gpu/drm/i915/i915_scheduler.c node->attr = *attr; node 312 drivers/gpu/drm/i915/i915_scheduler.c engine = node_to_request(node)->engine; node 316 drivers/gpu/drm/i915/i915_scheduler.c engine = sched_lock_engine(node, engine, &cache); node 320 drivers/gpu/drm/i915/i915_scheduler.c node = dep->signaler; node 321 drivers/gpu/drm/i915/i915_scheduler.c engine = sched_lock_engine(node, engine, &cache); node 325 drivers/gpu/drm/i915/i915_scheduler.c if (prio <= node->attr.priority || node_signaled(node)) node 328 drivers/gpu/drm/i915/i915_scheduler.c GEM_BUG_ON(node_to_request(node)->engine != engine); node 330 drivers/gpu/drm/i915/i915_scheduler.c node->attr.priority = prio; node 332 drivers/gpu/drm/i915/i915_scheduler.c if (list_empty(&node->link)) { node 345 drivers/gpu/drm/i915/i915_scheduler.c !i915_request_is_active(node_to_request(node))) { node 350 drivers/gpu/drm/i915/i915_scheduler.c list_move_tail(&node->link, cache.priolist); node 354 drivers/gpu/drm/i915/i915_scheduler.c kick_submission(engine, node_to_request(node), prio); node 367 drivers/gpu/drm/i915/i915_scheduler.c static void __bump_priority(struct i915_sched_node *node, unsigned int bump) node 369 drivers/gpu/drm/i915/i915_scheduler.c struct i915_sched_attr attr = node->attr; node 372 drivers/gpu/drm/i915/i915_scheduler.c __i915_schedule(node, &attr); node 388 drivers/gpu/drm/i915/i915_scheduler.c void i915_sched_node_init(struct i915_sched_node *node) node 390 drivers/gpu/drm/i915/i915_scheduler.c INIT_LIST_HEAD(&node->signalers_list); node 391 drivers/gpu/drm/i915/i915_scheduler.c INIT_LIST_HEAD(&node->waiters_list); node 392 drivers/gpu/drm/i915/i915_scheduler.c INIT_LIST_HEAD(&node->link); node 393 drivers/gpu/drm/i915/i915_scheduler.c node->attr.priority = I915_PRIORITY_INVALID; node 394 drivers/gpu/drm/i915/i915_scheduler.c node->semaphores = 0; node 395 drivers/gpu/drm/i915/i915_scheduler.c node->flags = 0; node 410 drivers/gpu/drm/i915/i915_scheduler.c bool __i915_sched_node_add_dependency(struct i915_sched_node *node, node 422 drivers/gpu/drm/i915/i915_scheduler.c dep->waiter = node; node 428 drivers/gpu/drm/i915/i915_scheduler.c node->flags |= I915_SCHED_HAS_SEMAPHORE_CHAIN; node 431 drivers/gpu/drm/i915/i915_scheduler.c list_add(&dep->signal_link, &node->signalers_list); node 452 drivers/gpu/drm/i915/i915_scheduler.c int i915_sched_node_add_dependency(struct i915_sched_node *node, node 461 drivers/gpu/drm/i915/i915_scheduler.c if (!__i915_sched_node_add_dependency(node, signal, dep, node 469 drivers/gpu/drm/i915/i915_scheduler.c void i915_sched_node_fini(struct i915_sched_node *node) node 481 drivers/gpu/drm/i915/i915_scheduler.c list_for_each_entry_safe(dep, tmp, &node->signalers_list, signal_link) { node 491 drivers/gpu/drm/i915/i915_scheduler.c list_for_each_entry_safe(dep, tmp, &node->waiters_list, wait_link) { node 492 drivers/gpu/drm/i915/i915_scheduler.c GEM_BUG_ON(dep->signaler != node); node 28 drivers/gpu/drm/i915/i915_scheduler.h void i915_sched_node_init(struct i915_sched_node *node); node 30 drivers/gpu/drm/i915/i915_scheduler.h bool __i915_sched_node_add_dependency(struct i915_sched_node *node, node 35 drivers/gpu/drm/i915/i915_scheduler.h int i915_sched_node_add_dependency(struct i915_sched_node *node, node 38 drivers/gpu/drm/i915/i915_scheduler.h void i915_sched_node_fini(struct i915_sched_node *node); node 469 drivers/gpu/drm/i915/i915_trace.h __entry->offset = vma->node.start; node 470 drivers/gpu/drm/i915/i915_trace.h __entry->size = vma->node.size; node 494 drivers/gpu/drm/i915/i915_trace.h __entry->offset = vma->node.start; node 495 drivers/gpu/drm/i915/i915_trace.h __entry->size = vma->node.size; node 618 drivers/gpu/drm/i915/i915_trace.h TP_PROTO(struct i915_address_space *vm, struct drm_mm_node *node, unsigned int flags), node 619 drivers/gpu/drm/i915/i915_trace.h TP_ARGS(vm, node, flags), node 633 drivers/gpu/drm/i915/i915_trace.h __entry->start = node->start; node 634 drivers/gpu/drm/i915/i915_trace.h __entry->size = node->size; node 635 drivers/gpu/drm/i915/i915_trace.h __entry->color = node->color; node 121 drivers/gpu/drm/i915/i915_vgpu.c struct drm_mm_node *node) node 123 drivers/gpu/drm/i915/i915_vgpu.c if (!drm_mm_node_allocated(node)) node 127 drivers/gpu/drm/i915/i915_vgpu.c node->start, node 128 drivers/gpu/drm/i915/i915_vgpu.c node->start + node->size, node 129 drivers/gpu/drm/i915/i915_vgpu.c node->size / 1024); node 131 drivers/gpu/drm/i915/i915_vgpu.c ggtt->vm.reserved -= node->size; node 132 drivers/gpu/drm/i915/i915_vgpu.c drm_mm_remove_node(node); node 156 drivers/gpu/drm/i915/i915_vgpu.c struct drm_mm_node *node, node 167 drivers/gpu/drm/i915/i915_vgpu.c ret = i915_gem_gtt_reserve(&ggtt->vm, node, node 63 drivers/gpu/drm/i915/i915_vma.c if (!vma->node.stack) { node 65 drivers/gpu/drm/i915/i915_vma.c vma->node.start, vma->node.size, reason); node 69 drivers/gpu/drm/i915/i915_vma.c nr_entries = stack_depot_fetch(vma->node.stack, &entries); node 72 drivers/gpu/drm/i915/i915_vma.c vma->node.start, vma->node.size, reason, buf); node 311 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 312 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(vma->size > vma->node.size); node 314 drivers/gpu/drm/i915/i915_vma.c if (GEM_DEBUG_WARN_ON(range_overflows(vma->node.start, node 315 drivers/gpu/drm/i915/i915_vma.c vma->node.size, node 367 drivers/gpu/drm/i915/i915_vma.c vma->node.start, node 368 drivers/gpu/drm/i915/i915_vma.c vma->node.size); node 438 drivers/gpu/drm/i915/i915_vma.c if (!drm_mm_node_allocated(&vma->node)) node 441 drivers/gpu/drm/i915/i915_vma.c if (vma->node.size < size) node 445 drivers/gpu/drm/i915/i915_vma.c if (alignment && !IS_ALIGNED(vma->node.start, alignment)) node 452 drivers/gpu/drm/i915/i915_vma.c vma->node.start < (flags & PIN_OFFSET_MASK)) node 456 drivers/gpu/drm/i915/i915_vma.c vma->node.start != (flags & PIN_OFFSET_MASK)) node 469 drivers/gpu/drm/i915/i915_vma.c fenceable = (vma->node.size >= vma->fence_size && node 470 drivers/gpu/drm/i915/i915_vma.c IS_ALIGNED(vma->node.start, vma->fence_alignment)); node 472 drivers/gpu/drm/i915/i915_vma.c mappable = vma->node.start + vma->fence_size <= i915_vm_to_ggtt(vma->vm)->mappable_end; node 480 drivers/gpu/drm/i915/i915_vma.c static bool color_differs(struct drm_mm_node *node, unsigned long color) node 482 drivers/gpu/drm/i915/i915_vma.c return node->allocated && node->color != color; node 487 drivers/gpu/drm/i915/i915_vma.c struct drm_mm_node *node = &vma->node; node 501 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(node)); node 502 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(list_empty(&node->node_list)); node 504 drivers/gpu/drm/i915/i915_vma.c other = list_prev_entry(node, node_list); node 508 drivers/gpu/drm/i915/i915_vma.c other = list_next_entry(node, node_list); node 509 drivers/gpu/drm/i915/i915_vma.c if (color_differs(other, cache_level) && !drm_mm_hole_follows(node)) node 551 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(drm_mm_node_allocated(&vma->node)); node 610 drivers/gpu/drm/i915/i915_vma.c ret = i915_gem_gtt_reserve(vma->vm, &vma->node, node 649 drivers/gpu/drm/i915/i915_vma.c ret = i915_gem_gtt_insert(vma->vm, &vma->node, node 655 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(vma->node.start < start); node 656 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(vma->node.start + vma->node.size > end); node 658 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 683 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 689 drivers/gpu/drm/i915/i915_vma.c drm_mm_remove_node(&vma->node); node 732 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 800 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(vma->node.allocated); node 864 drivers/gpu/drm/i915/i915_vma.c struct drm_vma_offset_node *node = &vma->obj->base.vma_node; node 877 drivers/gpu/drm/i915/i915_vma.c drm_vma_node_offset_addr(node) + vma_offset, node 895 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 971 drivers/gpu/drm/i915/i915_vma.c if (!drm_mm_node_allocated(&vma->node)) node 53 drivers/gpu/drm/i915/i915_vma.h struct drm_mm_node node; node 217 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(!vma->node.allocated); node 218 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(upper_32_bits(vma->node.start)); node 219 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(upper_32_bits(vma->node.start + vma->node.size - 1)); node 220 drivers/gpu/drm/i915/i915_vma.h return lower_32_bits(vma->node.start); node 334 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 366 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 255 drivers/gpu/drm/i915/selftests/i915_gem_evict.c static void mock_color_adjust(const struct drm_mm_node *node, node 389 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct drm_mm_node node; node 434 drivers/gpu/drm/i915/selftests/i915_gem_evict.c if (i915_gem_gtt_insert(&i915->ggtt.vm, &r->node, node 518 drivers/gpu/drm/i915/selftests/i915_gem_evict.c drm_mm_remove_node(&reserved->node); node 293 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c mock_vma.node.size = BIT_ULL(size); node 294 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c mock_vma.node.start = addr; node 408 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 411 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node), node 438 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 441 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, node 450 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, node 481 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 484 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node), node 511 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 514 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node), node 523 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, p->name, vma->node.start, vma->node.size, node 599 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 614 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(drm_mm_node_allocated(&vma->node)); node 683 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 787 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 865 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node) || node 1084 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct drm_mm_node *node; node 1091 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) { node 1096 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt->vm.mm.color_adjust(node, 0, node 1333 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, node 1345 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1346 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (vma->node.start != total || node 1347 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.size != 2*I915_GTT_PAGE_SIZE) { node 1349 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.start, vma->node.size, node 1383 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, node 1395 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1396 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (vma->node.start != total || node 1397 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.size != 2*I915_GTT_PAGE_SIZE) { node 1399 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.start, vma->node.size, node 1427 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, node 1439 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1440 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (vma->node.start != offset || node 1441 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.size != 2*I915_GTT_PAGE_SIZE) { node 1443 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma->node.start, vma->node.size, node 1540 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, node 1557 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1569 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!drm_mm_node_allocated(&vma->node)) { node 1589 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1590 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c offset = vma->node.start; node 1598 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, node 1609 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 1610 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (vma->node.start != offset) { node 1612 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c offset, vma->node.start); node 1645 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, node 1656 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); node 669 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.start, node 670 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.size, node 793 drivers/gpu/drm/i915/selftests/i915_request.c *cmd++ = lower_32_bits(vma->node.start); node 794 drivers/gpu/drm/i915/selftests/i915_request.c *cmd++ = upper_32_bits(vma->node.start); node 797 drivers/gpu/drm/i915/selftests/i915_request.c *cmd++ = lower_32_bits(vma->node.start); node 800 drivers/gpu/drm/i915/selftests/i915_request.c *cmd++ = lower_32_bits(vma->node.start); node 872 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.start, node 873 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.size, node 990 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.start, node 991 drivers/gpu/drm/i915/selftests/i915_request.c batch->node.size, node 567 drivers/gpu/drm/i915/selftests/i915_vma.c if (vma->node.size < vma->size) { node 569 drivers/gpu/drm/i915/selftests/i915_vma.c vma->size, vma->node.size); node 661 drivers/gpu/drm/i915/selftests/i915_vma.c if (vma->node.size < vma->size) { node 663 drivers/gpu/drm/i915/selftests/i915_vma.c name, vma->size, vma->node.size); node 71 drivers/gpu/drm/i915/selftests/igt_spinner.c return hws->node.start + seqno_offset(rq->fence.context); node 143 drivers/gpu/drm/i915/selftests/igt_spinner.c *batch++ = lower_32_bits(vma->node.start); node 144 drivers/gpu/drm/i915/selftests/igt_spinner.c *batch++ = upper_32_bits(vma->node.start); node 156 drivers/gpu/drm/i915/selftests/igt_spinner.c err = engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, 0); node 16 drivers/gpu/drm/lima/lima_vm.c struct drm_mm_node node; node 131 drivers/gpu/drm/lima/lima_vm.c err = drm_mm_insert_node(&vm->mm, &bo_va->node, bo->gem.size); node 135 drivers/gpu/drm/lima/lima_vm.c err = lima_vm_map_page_table(vm, bo->pages_dma_addr, bo_va->node.start, node 136 drivers/gpu/drm/lima/lima_vm.c bo_va->node.start + bo_va->node.size - 1); node 148 drivers/gpu/drm/lima/lima_vm.c drm_mm_remove_node(&bo_va->node); node 171 drivers/gpu/drm/lima/lima_vm.c lima_vm_unmap_page_table(vm, bo_va->node.start, node 172 drivers/gpu/drm/lima/lima_vm.c bo_va->node.start + bo_va->node.size - 1); node 174 drivers/gpu/drm/lima/lima_vm.c drm_mm_remove_node(&bo_va->node); node 193 drivers/gpu/drm/lima/lima_vm.c ret = bo_va->node.start; node 555 drivers/gpu/drm/mediatek/mtk_drm_crtc.c struct device_node *node; node 557 drivers/gpu/drm/mediatek/mtk_drm_crtc.c node = priv->comp_node[comp_id]; node 558 drivers/gpu/drm/mediatek/mtk_drm_crtc.c if (!node) { node 588 drivers/gpu/drm/mediatek/mtk_drm_crtc.c struct device_node *node; node 590 drivers/gpu/drm/mediatek/mtk_drm_crtc.c node = priv->comp_node[comp_id]; node 593 drivers/gpu/drm/mediatek/mtk_drm_crtc.c dev_err(dev, "Component %pOF not initialized\n", node); node 240 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c int mtk_ddp_comp_get_id(struct device_node *node, node 243 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c int id = of_alias_get_id(node, mtk_ddp_comp_stem[comp_type]); node 255 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c int mtk_ddp_comp_init(struct device *dev, struct device_node *node, node 285 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c comp->regs = of_iomap(node, 0); node 286 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c comp->irq = of_irq_get(node, 0); node 287 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c comp->clk = of_clk_get(node, 0); node 298 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c larb_node = of_parse_phandle(node, "mediatek,larb", 0); node 301 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c "Missing mediadek,larb phandle in %pOF node\n", node); node 161 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h int mtk_ddp_comp_get_id(struct device_node *node, node 489 drivers/gpu/drm/mediatek/mtk_drm_drv.c struct device_node *node; node 512 drivers/gpu/drm/mediatek/mtk_drm_drv.c for_each_child_of_node(dev->of_node->parent, node) { node 517 drivers/gpu/drm/mediatek/mtk_drm_drv.c of_id = of_match_node(mtk_ddp_comp_dt_ids, node); node 521 drivers/gpu/drm/mediatek/mtk_drm_drv.c if (!of_device_is_available(node)) { node 523 drivers/gpu/drm/mediatek/mtk_drm_drv.c node); node 530 drivers/gpu/drm/mediatek/mtk_drm_drv.c private->mutex_node = of_node_get(node); node 534 drivers/gpu/drm/mediatek/mtk_drm_drv.c comp_id = mtk_ddp_comp_get_id(node, comp_type); node 537 drivers/gpu/drm/mediatek/mtk_drm_drv.c node); node 541 drivers/gpu/drm/mediatek/mtk_drm_drv.c private->comp_node[comp_id] = of_node_get(node); node 554 drivers/gpu/drm/mediatek/mtk_drm_drv.c node); node 556 drivers/gpu/drm/mediatek/mtk_drm_drv.c node); node 563 drivers/gpu/drm/mediatek/mtk_drm_drv.c of_node_put(node); node 567 drivers/gpu/drm/mediatek/mtk_drm_drv.c ret = mtk_ddp_comp_init(dev, node, comp, comp_id, NULL); node 569 drivers/gpu/drm/mediatek/mtk_drm_drv.c of_node_put(node); node 80 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 81 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c struct drm_device *dev = node->minor->dev; node 85 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c node->info_ent->data; node 1260 drivers/gpu/drm/msm/adreno/a6xx_gmu.c int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node) node 1263 drivers/gpu/drm/msm/adreno/a6xx_gmu.c struct platform_device *pdev = of_find_device_by_node(node); node 1271 drivers/gpu/drm/msm/adreno/a6xx_gmu.c of_dma_configure(gmu->dev, node, true); node 851 drivers/gpu/drm/msm/adreno/a6xx_gpu.c struct device_node *node; node 874 drivers/gpu/drm/msm/adreno/a6xx_gpu.c node = of_parse_phandle(pdev->dev.of_node, "qcom,gmu", 0); node 877 drivers/gpu/drm/msm/adreno/a6xx_gpu.c BUG_ON(!node); node 879 drivers/gpu/drm/msm/adreno/a6xx_gpu.c ret = a6xx_gmu_init(a6xx_gpu, node); node 56 drivers/gpu/drm/msm/adreno/a6xx_gpu.h int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node); node 79 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct list_head node; node 91 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c list_add_tail(&obj->node, &a6xx_state->objs); node 913 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c list_for_each_entry_safe(obj, tmp, &a6xx_state->objs, node) node 274 drivers/gpu/drm/msm/adreno/adreno_device.c struct device_node *node = dev->of_node; node 280 drivers/gpu/drm/msm/adreno/adreno_device.c ret = of_property_read_string_index(node, "compatible", 0, &compat); node 298 drivers/gpu/drm/msm/adreno/adreno_device.c ret = of_property_read_u32(node, "qcom,chipid", &chipid); node 823 drivers/gpu/drm/msm/adreno/adreno_gpu.c struct device_node *child, *node; node 826 drivers/gpu/drm/msm/adreno/adreno_gpu.c node = of_get_compatible_child(dev->of_node, "qcom,gpu-pwrlevels"); node 827 drivers/gpu/drm/msm/adreno/adreno_gpu.c if (!node) { node 832 drivers/gpu/drm/msm/adreno/adreno_gpu.c for_each_child_of_node(node, child) { node 847 drivers/gpu/drm/msm/adreno/adreno_gpu.c of_node_put(node); node 245 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 246 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c struct drm_device *dev = node->minor->dev; node 28 drivers/gpu/drm/msm/disp/mdp_kms.c list_for_each_entry(irq, &mdp_kms->irq_list, node) node 53 drivers/gpu/drm/msm/disp/mdp_kms.c list_for_each_entry_safe(handler, n, &mdp_kms->irq_list, node) { node 111 drivers/gpu/drm/msm/disp/mdp_kms.c list_add(&irq->node, &mdp_kms->irq_list); node 130 drivers/gpu/drm/msm/disp/mdp_kms.c list_del(&irq->node); node 58 drivers/gpu/drm/msm/disp/mdp_kms.h struct list_head node; node 162 drivers/gpu/drm/msm/msm_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 163 drivers/gpu/drm/msm/msm_debugfs.c struct drm_device *dev = node->minor->dev; node 165 drivers/gpu/drm/msm/msm_debugfs.c node->info_ent->data; node 312 drivers/gpu/drm/msm/msm_drv.c struct device_node *node; node 333 drivers/gpu/drm/msm/msm_drv.c node = of_parse_phandle(dev->dev->of_node, "memory-region", 0); node 334 drivers/gpu/drm/msm/msm_drv.c if (node) { node 336 drivers/gpu/drm/msm/msm_drv.c ret = of_address_to_resource(node, 0, &r); node 337 drivers/gpu/drm/msm/msm_drv.c of_node_put(node); node 1065 drivers/gpu/drm/msm/msm_gem.c to_msm_bo(obj)->vram_node = &vma->node; node 29 drivers/gpu/drm/msm/msm_gem.h struct drm_mm_node node; node 134 drivers/gpu/drm/msm/msm_gem.h struct list_head node; /* node in ring submit list */ node 55 drivers/gpu/drm/msm/msm_gem_submit.c INIT_LIST_HEAD(&submit->node); node 65 drivers/gpu/drm/msm/msm_gem_submit.c list_del(&submit->node); node 34 drivers/gpu/drm/msm/msm_gem_vma.c unsigned size = vma->node.size << PAGE_SHIFT; node 96 drivers/gpu/drm/msm/msm_gem_vma.c drm_mm_remove_node(&vma->node); node 114 drivers/gpu/drm/msm/msm_gem_vma.c ret = drm_mm_insert_node(&aspace->mm, &vma->node, npages); node 120 drivers/gpu/drm/msm/msm_gem_vma.c vma->iova = vma->node.start << PAGE_SHIFT; node 394 drivers/gpu/drm/msm/msm_gpu.c list_for_each_entry(submit, &ring->submits, node) { node 410 drivers/gpu/drm/msm/msm_gpu.c list_for_each_entry(submit, &ring->submits, node) node 501 drivers/gpu/drm/msm/msm_gpu.c list_for_each_entry(submit, &ring->submits, node) node 699 drivers/gpu/drm/msm/msm_gpu.c list_for_each_entry_safe(submit, tmp, &ring->submits, node) { node 745 drivers/gpu/drm/msm/msm_gpu.c list_add_tail(&submit->node, &ring->submits); node 176 drivers/gpu/drm/msm/msm_gpu.h struct list_head node; node 28 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry(entry, &ctx->submitqueues, node) { node 52 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry_safe(entry, tmp, &ctx->submitqueues, node) node 87 drivers/gpu/drm/msm/msm_submitqueue.c list_add_tail(&queue->node, &ctx->submitqueues); node 173 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry(entry, &ctx->submitqueues, node) { node 175 drivers/gpu/drm/msm/msm_submitqueue.c list_del(&entry->node); node 17 drivers/gpu/drm/nouveau/include/nvkm/core/gpuobj.h struct nvkm_mm_node *node; node 46 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h struct nvkm_mm_node *node; node 48 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h list_for_each_entry(node, &mm->nodes, nl_entry) { node 49 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h if (node->heap == heap) node 50 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h size += node->length; node 56 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_contiguous(struct nvkm_mm_node *node) node 58 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h return !node->next; node 62 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_addr(struct nvkm_mm_node *node) node 64 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h if (WARN_ON(!nvkm_mm_contiguous(node))) node 66 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h return node->offset; node 70 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_size(struct nvkm_mm_node *node) node 74 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h size += node->length; node 75 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h } while ((node = node->next)); node 20 drivers/gpu/drm/nouveau/include/nvkm/core/object.h struct rb_node node; node 118 drivers/gpu/drm/nouveau/nouveau_abi16.c nvkm_mm_free(&chan->heap, &ntfy->node); node 546 drivers/gpu/drm/nouveau/nouveau_abi16.c &ntfy->node); node 550 drivers/gpu/drm/nouveau/nouveau_abi16.c args.start = ntfy->node->offset; node 551 drivers/gpu/drm/nouveau/nouveau_abi16.c args.limit = ntfy->node->offset + ntfy->node->length - 1; node 580 drivers/gpu/drm/nouveau/nouveau_abi16.c info->offset = ntfy->node->offset; node 18 drivers/gpu/drm/nouveau/nouveau_abi16.h struct nvkm_mm_node *node; node 40 drivers/gpu/drm/nouveau/nouveau_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 41 drivers/gpu/drm/nouveau/nouveau_debugfs.c struct nouveau_drm *drm = nouveau_drm(node->minor->dev); node 52 drivers/gpu/drm/nouveau/nouveau_debugfs.c struct drm_info_node *node = m->private; node 53 drivers/gpu/drm/nouveau/nouveau_debugfs.c struct nouveau_drm *drm = nouveau_drm(node->minor->dev); node 112 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c return nvkm_memory_map(gpuobj->parent, gpuobj->node->offset + offset, node 119 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c return nvkm_ro32(gpuobj->parent, gpuobj->node->offset + offset); node 125 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c nvkm_wo32(gpuobj->parent, gpuobj->node->offset + offset, data); node 157 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c gpuobj->map = (u8 *)gpuobj->map + gpuobj->node->offset; node 181 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c max(align, 1), &gpuobj->node); node 184 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c -align, &gpuobj->node); node 191 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c gpuobj->addr = parent->addr + gpuobj->node->offset; node 192 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c gpuobj->size = gpuobj->node->length; node 220 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c nvkm_mm_free(&gpuobj->parent->heap, &gpuobj->node); node 32 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *node; node 36 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(node, &mm->nodes, nl_entry) { node 38 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->offset, node->length, node->type); node 41 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(node, &mm->free, fl_entry) { node 43 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->offset, node->length, node->type); node 53 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *prev = node(this, prev); node 54 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *next = node(this, next); node 129 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = node(this, prev); node 133 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = node(this, next); node 203 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = node(this, prev); node 207 drivers/gpu/drm/nouveau/nvkm/core/mm.c next = node(this, next); node 242 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *node, *prev; node 246 drivers/gpu/drm/nouveau/nvkm/core/mm.c prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); node 250 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (!(node = kzalloc(sizeof(*node), GFP_KERNEL))) node 252 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->type = NVKM_MM_TYPE_HOLE; node 253 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->offset = next; node 254 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->length = offset - next; node 255 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&node->nl_entry, &mm->nodes); node 265 drivers/gpu/drm/nouveau/nvkm/core/mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 266 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (!node) node 270 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->offset = roundup(offset, mm->block_size); node 271 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->length = rounddown(offset + length, mm->block_size); node 272 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->length -= node->offset; node 275 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&node->nl_entry, &mm->nodes); node 276 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&node->fl_entry, &mm->free); node 277 drivers/gpu/drm/nouveau/nvkm/core/mm.c node->heap = heap; node 285 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *node, *temp; node 291 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry(node, &mm->nodes, nl_entry) { node 292 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (node->type != NVKM_MM_TYPE_HOLE) { node 300 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) { node 301 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_del(&node->nl_entry); node 302 drivers/gpu/drm/nouveau/nvkm/core/mm.c kfree(node); node 35 drivers/gpu/drm/nouveau/nvkm/core/object.c struct rb_node *node = client->objroot.rb_node; node 36 drivers/gpu/drm/nouveau/nvkm/core/object.c while (node) { node 37 drivers/gpu/drm/nouveau/nvkm/core/object.c object = rb_entry(node, typeof(*object), node); node 39 drivers/gpu/drm/nouveau/nvkm/core/object.c node = node->rb_left; node 42 drivers/gpu/drm/nouveau/nvkm/core/object.c node = node->rb_right; node 60 drivers/gpu/drm/nouveau/nvkm/core/object.c if (!RB_EMPTY_NODE(&object->node)) node 61 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_erase(&object->node, &object->client->objroot); node 71 drivers/gpu/drm/nouveau/nvkm/core/object.c struct nvkm_object *this = rb_entry(*ptr, typeof(*this), node); node 82 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_link_node(&object->node, parent, ptr); node 83 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_insert_color(&object->node, &object->client->objroot); node 307 drivers/gpu/drm/nouveau/nvkm/core/object.c RB_CLEAR_NODE(&object->node); node 83 drivers/gpu/drm/nouveau/nvkm/core/ramht.c inst = data->inst->node->offset; node 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4); node 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c (chan->ramht->gpuobj->node->offset >> 4)); node 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4); node 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c (chan->ramht->gpuobj->node->offset >> 4)); node 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4); node 82 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c (chan->ramht->gpuobj->node->offset >> 4)); node 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4); node 82 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c (chan->ramht->gpuobj->node->offset >> 4)); node 69 drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c nvkm_wr32(device, 0x001708, 0x80000000 | bar->bar1->node->offset >> 4); node 91 drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c nvkm_wr32(device, 0x00170c, 0x80000000 | bar->bar2->node->offset >> 4); node 83 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c struct nvkm_mm_node *node; node 85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c while ((node = next)) { node 86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c next = node->next; node 87 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c nvkm_mm_free(&vram->ram->vram, &node); node 109 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c struct nvkm_mm_node **node, *r; node 130 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c node = &vram->mn; node 142 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c *node = r; node 143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c node = &r->next; node 176 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = gk20a_instobj(memory); node 177 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->imem; node 182 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c return node->vaddr; node 188 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory); node 189 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->base.imem; node 197 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (node->base.vaddr) { node 198 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (!node->use_cpt) { node 200 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c list_del(&node->vaddr_node); node 209 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP, node 211 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (!node->base.vaddr) { node 222 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->use_cpt++; node 225 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c return node->base.vaddr; node 231 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = gk20a_instobj(memory); node 232 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->imem; node 243 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory); node 244 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->base.imem; node 250 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (WARN_ON(node->use_cpt == 0)) node 254 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (--node->use_cpt == 0) node 255 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c list_add_tail(&node->vaddr_node, &imem->vaddr_lru); node 267 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = gk20a_instobj(memory); node 269 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c return node->vaddr[offset / 4]; node 275 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = gk20a_instobj(memory); node 277 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->vaddr[offset / 4] = data; node 284 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = gk20a_instobj(memory); node 286 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c .memory = &node->memory, node 288 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c .mem = node->mn, node 297 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_dma *node = gk20a_instobj_dma(memory); node 298 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->base.imem; node 301 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (unlikely(!node->base.vaddr)) node 304 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT, node 305 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.vaddr, node->handle, imem->attrs); node 308 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c return node; node 314 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory); node 315 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instmem *imem = node->base.imem; node 317 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct nvkm_mm_node *r = node->base.mn; node 326 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (node->base.vaddr) node 327 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c gk20a_instobj_iommu_recycle_vaddr(node); node 335 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c for (i = 0; i < node->base.mn->length; i++) { node 338 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c dma_unmap_page(dev, node->dma_addrs[i], PAGE_SIZE, node 340 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c __free_page(node->pages[i]); node 349 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c return node; node 386 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_dma *node; node 390 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (!(node = kzalloc(sizeof(*node), GFP_KERNEL))) node 392 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c *_node = &node->base; node 394 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c nvkm_memory_ctor(&gk20a_instobj_func_dma, &node->base.memory); node 395 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.memory.ptrs = &gk20a_instobj_ptrs; node 397 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.vaddr = dma_alloc_attrs(dev, npages << PAGE_SHIFT, node 398 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c &node->handle, GFP_KERNEL, node 400 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (!node->base.vaddr) { node 406 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (unlikely(node->handle & (align - 1))) node 409 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c &node->handle, align); node 412 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->r.type = 12; node 413 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->r.offset = node->handle >> 12; node 414 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->r.length = (npages << PAGE_SHIFT) >> 12; node 416 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.mn = &node->r; node 424 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj_iommu *node; node 435 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c if (!(node = kzalloc(sizeof(*node) + ((sizeof(node->pages[0]) + node 436 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c sizeof(*node->dma_addrs)) * npages), GFP_KERNEL))) node 438 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c *_node = &node->base; node 439 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->dma_addrs = (void *)(node->pages + npages); node 441 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c nvkm_memory_ctor(&gk20a_instobj_func_iommu, &node->base.memory); node 442 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.memory.ptrs = &gk20a_instobj_ptrs; node 453 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->pages[i] = p; node 460 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->dma_addrs[i] = dma_adr; node 477 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c ret = iommu_map(imem->domain, offset, node->dma_addrs[i], node 493 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->base.mn = r; node 502 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c for (i = 0; i < npages && node->pages[i] != NULL; i++) { node 503 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c dma_addr_t dma_addr = node->dma_addrs[i]; node 507 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c __free_page(node->pages[i]); node 519 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c struct gk20a_instobj *node = NULL; node 531 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c align, &node); node 534 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c align, &node); node 535 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c *pmemory = node ? &node->memory : NULL; node 539 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c node->imem = imem; node 542 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c size, align, (u64)node->mn->offset << 12); node 42 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c struct nvkm_mm_node *node; node 50 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c nvkm_wr32(device, 0x700000 + iobj->node->offset + offset, data); node 58 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c return nvkm_rd32(device, 0x700000 + iobj->node->offset + offset); node 77 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c return device->pri + 0x700000 + iobj->node->offset; node 83 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c return nv04_instobj(memory)->node->length; node 89 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c return nv04_instobj(memory)->node->offset; node 103 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c nvkm_mm_free(&iobj->imem->heap, &iobj->node); node 137 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c align ? align : 1, &iobj->node); node 44 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c struct nvkm_mm_node *node; node 51 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); node 58 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); node 77 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c return iobj->imem->iomem + iobj->node->offset; node 83 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c return nv40_instobj(memory)->node->length; node 89 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c return nv40_instobj(memory)->node->offset; node 103 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c nvkm_mm_free(&iobj->imem->heap, &iobj->node); node 137 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c align ? align : 1, &iobj->node); node 865 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct rb_node *node = vmm->root.rb_node; node 866 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while (node) { node 867 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); node 869 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c node = node->rb_left; node 872 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c node = node->rb_right; node 981 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct rb_node *node; node 986 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while ((node = rb_first(&vmm->root))) { node 987 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); node 1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->addr == addr && vma->part && (prev = node(vma, prev))) { node 1162 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->addr + vma->size == addr + size && (next = node(vma, next))) { node 1199 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((vma = node(vma, next)) && (start = vma->addr) < limit); node 1314 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma = node(vma, next); node 1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->part && (prev = node(vma, prev)) && prev->mapped) node 1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((next = node(vma, next)) && (!next->part || next->mapped)) node 1526 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((prev = node(vma, prev)) && !prev->used) { node 1532 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((next = node(vma, next)) && !next->used) { node 1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while ((next = node(next, next)) && next->part && node 1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((next = node(vma, next)) && next->part); node 1640 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct rb_node *node = NULL, *temp; node 1693 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c node = temp; node 1698 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (unlikely(!node)) node 1705 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *this = rb_entry(node, typeof(*this), tree); node 1706 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *prev = node(this, prev); node 1707 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vma *next = node(this, next); node 1724 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((node = rb_next(node))); node 404 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c list_for_each_entry(img, imgs, base.node) { node 424 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c list_for_each_entry(_img, imgs, node) { node 437 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c list_for_each_entry(_img, imgs, node) { node 512 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c list_add_tail(&img->node, &imgs); node 584 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c list_for_each_entry_safe(img, t, &imgs, node) { node 260 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c list_for_each_entry(img, imgs, base.node) { node 276 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c list_for_each_entry(_img, imgs, node) { node 289 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c list_for_each_entry(_img, imgs, node) { node 96 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/ls_ucode.h struct list_head node; node 1165 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c struct device_node *node = pdev->dev.of_node; node 1186 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c err = of_get_display_timing(node, "panel-timing", &timing); node 1198 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c of_property_read_u32(node, "width-mm", &ddata->width_mm); node 1201 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c of_property_read_u32(node, "height-mm", &ddata->height_mm); node 1219 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c backlight = of_parse_phandle(node, "backlight", 0); node 69 drivers/gpu/drm/omapdrm/dss/base.c static bool omapdss_device_is_registered(struct device_node *node) node 77 drivers/gpu/drm/omapdrm/dss/base.c if (dssdev->dev->of_node == node) { node 108 drivers/gpu/drm/omapdrm/dss/base.c struct omap_dss_device *omapdss_find_device_by_node(struct device_node *node) node 113 drivers/gpu/drm/omapdrm/dss/base.c if (dssdev->dev->of_node == node) node 296 drivers/gpu/drm/omapdrm/dss/base.c struct device_node *node; node 301 drivers/gpu/drm/omapdrm/dss/base.c static bool omapdss_list_contains(const struct device_node *node) node 306 drivers/gpu/drm/omapdrm/dss/base.c if (comp->node == node) node 313 drivers/gpu/drm/omapdrm/dss/base.c static void omapdss_walk_device(struct device *dev, struct device_node *node, node 321 drivers/gpu/drm/omapdrm/dss/base.c ret = of_property_read_string(node, "compatible", &compat); node 327 drivers/gpu/drm/omapdrm/dss/base.c comp->node = node; node 337 drivers/gpu/drm/omapdrm/dss/base.c n = of_get_child_by_name(node, "ports"); node 339 drivers/gpu/drm/omapdrm/dss/base.c n = of_get_child_by_name(node, "port"); node 346 drivers/gpu/drm/omapdrm/dss/base.c while ((n = of_graph_get_next_endpoint(node, n)) != NULL) { node 380 drivers/gpu/drm/omapdrm/dss/base.c if (omapdss_device_is_registered(comp->node)) node 5144 drivers/gpu/drm/omapdrm/dss/dsi.c struct device_node *node = dsi->dev->of_node; node 5152 drivers/gpu/drm/omapdrm/dss/dsi.c ep = of_graph_get_endpoint_by_regs(node, 0, 0); node 14 drivers/gpu/drm/omapdrm/dss/dss-of.c omapdss_of_find_connected_device(struct device_node *node, unsigned int port) node 19 drivers/gpu/drm/omapdrm/dss/dss-of.c remote_node = of_graph_get_remote_node(node, port, 0); node 699 drivers/gpu/drm/omapdrm/dss/hdmi4.c struct device_node *node = pdev->dev.of_node; node 703 drivers/gpu/drm/omapdrm/dss/hdmi4.c ep = of_graph_get_endpoint_by_regs(node, 0, 0); node 683 drivers/gpu/drm/omapdrm/dss/hdmi5.c struct device_node *node = pdev->dev.of_node; node 687 drivers/gpu/drm/omapdrm/dss/hdmi5.c ep = of_graph_get_endpoint_by_regs(node, 0, 0); node 28 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c struct device_node *node; node 44 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_update_prop(struct device_node *node, char *compat, node 57 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c of_update_property(node, prop); node 80 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_omapify_node(struct device_node *node) node 87 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c prop = of_find_property(node, "compatible", NULL); node 106 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c omapdss_update_prop(node, new_compat, new_len); node 109 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_add_to_list(struct device_node *node, bool root) node 113 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c n->node = node; node 119 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static bool __init omapdss_list_contains(const struct device_node *node) node 124 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c if (n->node == node) node 131 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_walk_device(struct device_node *node, bool root) node 135 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c omapdss_add_to_list(node, root); node 141 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c n = of_get_child_by_name(node, "ports"); node 143 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c n = of_get_child_by_name(node, "port"); node 150 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c while ((n = of_graph_get_next_endpoint(node, n)) != NULL) { node 212 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c if (of_match_node(omapdss_of_fixups_whitelist, n->node)) node 213 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c omapdss_omapify_node(n->node); node 216 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c of_node_put(n->node); node 472 drivers/gpu/drm/omapdrm/dss/omapdss.h struct omap_dss_device *omapdss_find_device_by_node(struct device_node *node); node 506 drivers/gpu/drm/omapdrm/dss/omapdss.h omapdss_of_find_connected_device(struct device_node *node, unsigned int port); node 777 drivers/gpu/drm/omapdrm/dss/venc.c struct device_node *node = venc->pdev->dev.of_node; node 782 drivers/gpu/drm/omapdrm/dss/venc.c ep = of_graph_get_endpoint_by_regs(node, 0, 0); node 21 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 22 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_device *dev = node->minor->dev; node 35 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 36 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_device *dev = node->minor->dev; node 47 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 48 drivers/gpu/drm/omapdrm/omap_debugfs.c struct drm_device *dev = node->minor->dev; node 208 drivers/gpu/drm/omapdrm/omap_drv.c struct device_node *node = NULL; node 214 drivers/gpu/drm/omapdrm/omap_drv.c node = display->dev->of_node; node 222 drivers/gpu/drm/omapdrm/omap_drv.c node = bridge->of_node; node 224 drivers/gpu/drm/omapdrm/omap_drv.c node = output->panel->dev->of_node; node 227 drivers/gpu/drm/omapdrm/omap_drv.c return node ? of_alias_get_id(node, "display") : -ENODEV; node 12 drivers/gpu/drm/omapdrm/omap_irq.c struct list_head node; node 27 drivers/gpu/drm/omapdrm/omap_irq.c list_for_each_entry(wait, &priv->wait_list, node) node 53 drivers/gpu/drm/omapdrm/omap_irq.c list_add(&wait->node, &priv->wait_list); node 70 drivers/gpu/drm/omapdrm/omap_irq.c list_del(&wait->node); node 242 drivers/gpu/drm/omapdrm/omap_irq.c list_for_each_entry_safe(wait, n, &priv->wait_list, node) { node 371 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c .node = NULL, node 416 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c info.node = of_graph_get_remote_port(endpoint); node 417 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c if (!info.node) node 567 drivers/gpu/drm/panel/panel-truly-nt35597.c .node = NULL, node 410 drivers/gpu/drm/panfrost/panfrost_drv.c node); node 457 drivers/gpu/drm/panfrost/panfrost_drv.c static void panfrost_drm_mm_color_adjust(const struct drm_mm_node *node, node 62 drivers/gpu/drm/panfrost/panfrost_gem.c list_for_each_entry(iter, &bo->mappings.list, node) { node 113 drivers/gpu/drm/panfrost/panfrost_gem.c list_for_each_entry(mapping, &bo->mappings.list, node) node 132 drivers/gpu/drm/panfrost/panfrost_gem.c INIT_LIST_HEAD(&mapping->node); node 164 drivers/gpu/drm/panfrost/panfrost_gem.c list_add_tail(&mapping->node, &bo->mappings.list); node 180 drivers/gpu/drm/panfrost/panfrost_gem.c list_for_each_entry(iter, &bo->mappings.list, node) { node 183 drivers/gpu/drm/panfrost/panfrost_gem.c list_del(&iter->node); node 44 drivers/gpu/drm/panfrost/panfrost_gem.h struct list_head node; node 59 drivers/gpu/drm/panfrost/panfrost_gem.h drm_mm_node_to_panfrost_mapping(struct drm_mm_node *node) node 61 drivers/gpu/drm/panfrost/panfrost_gem.h return container_of(node, struct panfrost_gem_mapping, mmnode); node 412 drivers/gpu/drm/panfrost/panfrost_mmu.c struct drm_mm_node *node; node 428 drivers/gpu/drm/panfrost/panfrost_mmu.c drm_mm_for_each_node(node, &priv->mm) { node 429 drivers/gpu/drm/panfrost/panfrost_mmu.c if (offset >= node->start && node 430 drivers/gpu/drm/panfrost/panfrost_mmu.c offset < (node->start + node->size)) { node 431 drivers/gpu/drm/panfrost/panfrost_mmu.c mapping = drm_mm_node_to_panfrost_mapping(node); node 36 drivers/gpu/drm/pl111/pl111_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 37 drivers/gpu/drm/pl111/pl111_debugfs.c struct drm_device *dev = node->minor->dev; node 41 drivers/gpu/drm/qxl/qxl_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 42 drivers/gpu/drm/qxl/qxl_debugfs.c struct qxl_device *qdev = node->minor->dev->dev_private; node 55 drivers/gpu/drm/qxl/qxl_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 56 drivers/gpu/drm/qxl/qxl_debugfs.c struct qxl_device *qdev = node->minor->dev->dev_private; node 369 drivers/gpu/drm/qxl/qxl_ttm.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 370 drivers/gpu/drm/qxl/qxl_ttm.c struct drm_mm *mm = (struct drm_mm *)node->info_ent->data; node 371 drivers/gpu/drm/qxl/qxl_ttm.c struct drm_device *dev = node->minor->dev; node 2925 drivers/gpu/drm/radeon/r100.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 2926 drivers/gpu/drm/radeon/r100.c struct drm_device *dev = node->minor->dev; node 2946 drivers/gpu/drm/radeon/r100.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 2947 drivers/gpu/drm/radeon/r100.c struct drm_device *dev = node->minor->dev; node 2974 drivers/gpu/drm/radeon/r100.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 2975 drivers/gpu/drm/radeon/r100.c struct drm_device *dev = node->minor->dev; node 3024 drivers/gpu/drm/radeon/r100.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 3025 drivers/gpu/drm/radeon/r100.c struct drm_device *dev = node->minor->dev; node 595 drivers/gpu/drm/radeon/r300.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 596 drivers/gpu/drm/radeon/r300.c struct drm_device *dev = node->minor->dev; node 485 drivers/gpu/drm/radeon/r420.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 486 drivers/gpu/drm/radeon/r420.c struct drm_device *dev = node->minor->dev; node 4351 drivers/gpu/drm/radeon/r600.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 4352 drivers/gpu/drm/radeon/r600.c struct drm_device *dev = node->minor->dev; node 751 drivers/gpu/drm/radeon/radeon_dp_mst.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 752 drivers/gpu/drm/radeon/radeon_dp_mst.c struct drm_device *dev = node->minor->dev; node 978 drivers/gpu/drm/radeon/radeon_fence.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 979 drivers/gpu/drm/radeon/radeon_fence.c struct drm_device *dev = node->minor->dev; node 1011 drivers/gpu/drm/radeon/radeon_fence.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 1012 drivers/gpu/drm/radeon/radeon_fence.c struct drm_device *dev = node->minor->dev; node 782 drivers/gpu/drm/radeon/radeon_gem.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 783 drivers/gpu/drm/radeon/radeon_gem.c struct drm_device *dev = node->minor->dev; node 299 drivers/gpu/drm/radeon/radeon_ib.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 300 drivers/gpu/drm/radeon/radeon_ib.c struct drm_device *dev = node->minor->dev; node 85 drivers/gpu/drm/radeon/radeon_mn.c struct radeon_mn_node *node; node 94 drivers/gpu/drm/radeon/radeon_mn.c node = container_of(it, struct radeon_mn_node, it); node 97 drivers/gpu/drm/radeon/radeon_mn.c list_for_each_entry(bo, &node->bos, mn_list) { node 180 drivers/gpu/drm/radeon/radeon_mn.c struct radeon_mn_node *node = NULL; node 194 drivers/gpu/drm/radeon/radeon_mn.c kfree(node); node 195 drivers/gpu/drm/radeon/radeon_mn.c node = container_of(it, struct radeon_mn_node, it); node 196 drivers/gpu/drm/radeon/radeon_mn.c interval_tree_remove(&node->it, &rmn->objects); node 199 drivers/gpu/drm/radeon/radeon_mn.c list_splice(&node->bos, &bos); node 202 drivers/gpu/drm/radeon/radeon_mn.c if (!node) { node 203 drivers/gpu/drm/radeon/radeon_mn.c node = kmalloc(sizeof(struct radeon_mn_node), GFP_KERNEL); node 204 drivers/gpu/drm/radeon/radeon_mn.c if (!node) { node 212 drivers/gpu/drm/radeon/radeon_mn.c node->it.start = addr; node 213 drivers/gpu/drm/radeon/radeon_mn.c node->it.last = end; node 214 drivers/gpu/drm/radeon/radeon_mn.c INIT_LIST_HEAD(&node->bos); node 215 drivers/gpu/drm/radeon/radeon_mn.c list_splice(&bos, &node->bos); node 216 drivers/gpu/drm/radeon/radeon_mn.c list_add(&bo->mn_list, &node->bos); node 218 drivers/gpu/drm/radeon/radeon_mn.c interval_tree_insert(&node->it, &rmn->objects); node 247 drivers/gpu/drm/radeon/radeon_mn.c struct radeon_mn_node *node; node 248 drivers/gpu/drm/radeon/radeon_mn.c node = container_of(head, struct radeon_mn_node, bos); node 249 drivers/gpu/drm/radeon/radeon_mn.c interval_tree_remove(&node->it, &rmn->objects); node 250 drivers/gpu/drm/radeon/radeon_mn.c kfree(node); node 1866 drivers/gpu/drm/radeon/radeon_pm.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 1867 drivers/gpu/drm/radeon/radeon_pm.c struct drm_device *dev = node->minor->dev; node 466 drivers/gpu/drm/radeon/radeon_ring.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 467 drivers/gpu/drm/radeon/radeon_ring.c struct drm_device *dev = node->minor->dev; node 469 drivers/gpu/drm/radeon/radeon_ring.c int ridx = *(int*)node->info_ent->data; node 929 drivers/gpu/drm/radeon/radeon_ttm.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 930 drivers/gpu/drm/radeon/radeon_ttm.c unsigned ttm_pl = *(int*)node->info_ent->data; node 931 drivers/gpu/drm/radeon/radeon_ttm.c struct drm_device *dev = node->minor->dev; node 312 drivers/gpu/drm/radeon/rs400.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 313 drivers/gpu/drm/radeon/rs400.c struct drm_device *dev = node->minor->dev; node 240 drivers/gpu/drm/radeon/rv515.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 241 drivers/gpu/drm/radeon/rv515.c struct drm_device *dev = node->minor->dev; node 258 drivers/gpu/drm/radeon/rv515.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 259 drivers/gpu/drm/radeon/rv515.c struct drm_device *dev = node->minor->dev; node 32 drivers/gpu/drm/rcar-du/rcar_du_encoder.c static unsigned int rcar_du_encoder_count_ports(struct device_node *node) node 38 drivers/gpu/drm/rcar-du/rcar_du_encoder.c ports = of_get_child_by_name(node, "ports"); node 40 drivers/gpu/drm/rcar-du/rcar_du_encoder.c ports = of_node_get(node); node 710 drivers/gpu/drm/rcar-du/rcar_lvds.c struct device_node *node; node 741 drivers/gpu/drm/rcar-du/rcar_lvds.c for_each_endpoint_of_node(remote, node) { node 742 drivers/gpu/drm/rcar-du/rcar_lvds.c if (node != remote_input) { node 748 drivers/gpu/drm/rcar-du/rcar_lvds.c of_node_put(node); node 694 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c struct device_node *node = NULL, *local; node 702 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c while ((node = of_find_compatible_node(node, NULL, node 707 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c if (node == dsi->dev->of_node) node 710 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c remote = of_graph_get_remote_node(node, 1, 0); node 719 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c pdev = of_find_device_by_node(node); node 727 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c of_node_put(node); node 274 drivers/gpu/drm/rockchip/rockchip_drm_drv.c struct device_node *node = of_graph_get_remote_port_parent(ep); node 279 drivers/gpu/drm/rockchip/rockchip_drm_drv.c if (!node) node 283 drivers/gpu/drm/rockchip/rockchip_drm_drv.c pdev = of_find_device_by_node(node); node 284 drivers/gpu/drm/rockchip/rockchip_drm_drv.c of_node_put(node); node 275 drivers/gpu/drm/scheduler/sched_main.c list_add_tail(&s_job->node, &sched->ring_mirror_list); node 288 drivers/gpu/drm/scheduler/sched_main.c struct drm_sched_job, node); node 378 drivers/gpu/drm/scheduler/sched_main.c list_for_each_entry_safe_reverse(s_job, tmp, &sched->ring_mirror_list, node) { node 389 drivers/gpu/drm/scheduler/sched_main.c list_del_init(&s_job->node); node 441 drivers/gpu/drm/scheduler/sched_main.c list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) { node 484 drivers/gpu/drm/scheduler/sched_main.c list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) { node 542 drivers/gpu/drm/scheduler/sched_main.c INIT_LIST_HEAD(&job->node); node 658 drivers/gpu/drm/scheduler/sched_main.c struct drm_sched_job, node); node 662 drivers/gpu/drm/scheduler/sched_main.c list_del_init(&job->node); node 108 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *check, *found; node 117 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node(node, mm) { node 118 drivers/gpu/drm/selftests/test-drm_mm.c if (node->start != addr) { node 120 drivers/gpu/drm/selftests/test-drm_mm.c n, addr, node->start); node 124 drivers/gpu/drm/selftests/test-drm_mm.c if (node->size != size) { node 126 drivers/gpu/drm/selftests/test-drm_mm.c n, size, node->size); node 130 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_hole_follows(node)) { node 137 drivers/gpu/drm/selftests/test-drm_mm.c if (node != check) { node 139 drivers/gpu/drm/selftests/test-drm_mm.c node->start, check->start); node 157 drivers/gpu/drm/selftests/test-drm_mm.c static u64 misalignment(struct drm_mm_node *node, u64 alignment) node 164 drivers/gpu/drm/selftests/test-drm_mm.c div64_u64_rem(node->start, alignment, &rem); node 168 drivers/gpu/drm/selftests/test-drm_mm.c static bool assert_node(struct drm_mm_node *node, struct drm_mm *mm, node 173 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_node_allocated(node) || node->mm != mm) { node 178 drivers/gpu/drm/selftests/test-drm_mm.c if (node->size != size) { node 180 drivers/gpu/drm/selftests/test-drm_mm.c node->size, size); node 184 drivers/gpu/drm/selftests/test-drm_mm.c if (misalignment(node, alignment)) { node 186 drivers/gpu/drm/selftests/test-drm_mm.c node->start, misalignment(node, alignment), alignment); node 190 drivers/gpu/drm/selftests/test-drm_mm.c if (node->color != color) { node 192 drivers/gpu/drm/selftests/test-drm_mm.c node->color, color); node 299 drivers/gpu/drm/selftests/test-drm_mm.c static struct drm_mm_node *set_node(struct drm_mm_node *node, node 302 drivers/gpu/drm/selftests/test-drm_mm.c node->start = start; node 303 drivers/gpu/drm/selftests/test-drm_mm.c node->size = size; node 304 drivers/gpu/drm/selftests/test-drm_mm.c return node; node 307 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_reserve_fail(struct drm_mm *mm, struct drm_mm_node *node) node 311 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(mm, node); node 317 drivers/gpu/drm/selftests/test-drm_mm.c node->start, node->size); node 318 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 321 drivers/gpu/drm/selftests/test-drm_mm.c err, -ENOSPC, node->start, node->size); node 375 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node tmp, *nodes, *node, *next; node 466 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 467 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 471 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 472 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, node); node 475 drivers/gpu/drm/selftests/test-drm_mm.c m, n, node->start); node 489 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 490 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 525 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_insert(struct drm_mm *mm, struct drm_mm_node *node, node 531 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node_generic(mm, node, node 540 drivers/gpu/drm/selftests/test-drm_mm.c if (!assert_node(node, mm, size, alignment, color)) { node 541 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 573 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; node 598 drivers/gpu/drm/selftests/test-drm_mm.c node = replace ? &tmp : &nodes[n]; node 599 drivers/gpu/drm/selftests/test-drm_mm.c memset(node, 0, sizeof(*node)); node 600 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, size, 0, n, mode)) { node 661 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 662 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 666 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 667 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, size, 0, n, mode)) { node 683 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 684 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 692 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 693 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 762 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_insert_in_range(struct drm_mm *mm, struct drm_mm_node *node, node 769 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node_in_range(mm, node, node 780 drivers/gpu/drm/selftests/test-drm_mm.c if (!assert_node(node, mm, size, alignment, color)) { node 781 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 820 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node; node 827 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node(node, mm) { node 828 drivers/gpu/drm/selftests/test-drm_mm.c if (node->start < start || node->start + node->size > end) { node 830 drivers/gpu/drm/selftests/test-drm_mm.c n, node->start, node->start + node->size, start, end); node 834 drivers/gpu/drm/selftests/test-drm_mm.c if (node->start != n * size) { node 836 drivers/gpu/drm/selftests/test-drm_mm.c n, n * size, node->start); node 840 drivers/gpu/drm/selftests/test-drm_mm.c if (node->size != size) { node 842 drivers/gpu/drm/selftests/test-drm_mm.c n, size, node->size); node 846 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_hole_follows(node) && node 847 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_hole_node_end(node) < end) { node 856 drivers/gpu/drm/selftests/test-drm_mm.c node = __drm_mm_interval_first(mm, 0, start - 1); node 857 drivers/gpu/drm/selftests/test-drm_mm.c if (node->allocated) { node 859 drivers/gpu/drm/selftests/test-drm_mm.c node->start, node->size, start); node 865 drivers/gpu/drm/selftests/test-drm_mm.c node = __drm_mm_interval_first(mm, end, U64_MAX); node 866 drivers/gpu/drm/selftests/test-drm_mm.c if (node->allocated) { node 868 drivers/gpu/drm/selftests/test-drm_mm.c node->start, node->size, end); node 880 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; node 947 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 948 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 956 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 957 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1041 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; node 1073 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1074 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1082 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1083 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1093 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; node 1104 drivers/gpu/drm/selftests/test-drm_mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 1105 drivers/gpu/drm/selftests/test-drm_mm.c if (!node) { node 1112 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, node 1125 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) { node 1126 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1127 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 1183 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node node; node 1200 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_scan_add_block(scan, &e->node)) node 1204 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_scan_remove_block(scan, &e->node)) node 1214 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&e->node); node 1217 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node; node 1219 drivers/gpu/drm/selftests/test-drm_mm.c while ((node = drm_mm_scan_color_evict(scan))) { node 1220 drivers/gpu/drm/selftests/test-drm_mm.c e = container_of(node, typeof(*e), node); node 1221 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&e->node); node 1241 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node; node 1248 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_scan_add_block(&scan, &e->node); node 1251 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_scan_remove_block(&scan, &e->node); node 1256 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_node_allocated(&e->node)) { node 1264 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node(node, mm) { node 1265 drivers/gpu/drm/selftests/test-drm_mm.c e = container_of(node, typeof(*e), node); node 1278 drivers/gpu/drm/selftests/test-drm_mm.c return assert_continuous(mm, nodes[0].node.size); node 1295 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_scan_add_block(&scan, &e->node)) node 1301 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_scan_remove_block(&scan, &e->node)) { node 1304 drivers/gpu/drm/selftests/test-drm_mm.c e->node.start); node 1313 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&e->node); node 1319 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(mm, &e->node); node 1322 drivers/gpu/drm/selftests/test-drm_mm.c e->node.start); node 1327 drivers/gpu/drm/selftests/test-drm_mm.c return assert_continuous(mm, nodes[0].node.size); node 1385 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(mm, &e->node); node 1388 drivers/gpu/drm/selftests/test-drm_mm.c e->node.start); node 1393 drivers/gpu/drm/selftests/test-drm_mm.c if (!assert_continuous(mm, nodes[0].node.size)) { node 1408 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; node 1431 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node(&mm, &nodes[n].node, 1); node 1501 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1502 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1521 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; node 1541 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node(&mm, &nodes[n].node, 1); node 1598 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1599 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1608 drivers/gpu/drm/selftests/test-drm_mm.c static unsigned int node_index(const struct drm_mm_node *node) node 1610 drivers/gpu/drm/selftests/test-drm_mm.c return div64_u64(node->start, node->size); node 1621 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; node 1670 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 1671 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1672 drivers/gpu/drm/selftests/test-drm_mm.c __set_bit(node_index(node), bitmap); node 1678 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 1679 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, node 1686 drivers/gpu/drm/selftests/test-drm_mm.c if (drm_mm_hole_follows(node)) { node 1688 drivers/gpu/drm/selftests/test-drm_mm.c m, n, node->start); node 1693 drivers/gpu/drm/selftests/test-drm_mm.c if (node_index(node) != last) { node 1695 drivers/gpu/drm/selftests/test-drm_mm.c m, n, size, last, node_index(node)); node 1707 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1708 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1715 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1716 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1735 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *nodes, *node, *next; node 1777 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 1778 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1779 drivers/gpu/drm/selftests/test-drm_mm.c __set_bit(node_index(node), bitmap); node 1785 drivers/gpu/drm/selftests/test-drm_mm.c node = &nodes[order[(o + m) % count]]; node 1786 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, node 1794 drivers/gpu/drm/selftests/test-drm_mm.c if (node_index(node) != first) { node 1796 drivers/gpu/drm/selftests/test-drm_mm.c m, n, first, node_index(node)); node 1807 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1808 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1815 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 1816 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1830 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node rsvd_lo, rsvd_hi, node; node 1859 drivers/gpu/drm/selftests/test-drm_mm.c memset(&node, 0, sizeof(node)); node 1860 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node_generic(&mm, &node, node 1865 drivers/gpu/drm/selftests/test-drm_mm.c node.start); node 1870 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_insert_node_generic(&mm, &node, 2, 0, 0, mode); node 1878 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(&node); node 1898 drivers/gpu/drm/selftests/test-drm_mm.c static void separate_adjacent_colors(const struct drm_mm_node *node, node 1903 drivers/gpu/drm/selftests/test-drm_mm.c if (node->allocated && node->color != color) node 1906 drivers/gpu/drm/selftests/test-drm_mm.c node = list_next_entry(node, node_list); node 1907 drivers/gpu/drm/selftests/test-drm_mm.c if (node->allocated && node->color != color) node 1911 drivers/gpu/drm/selftests/test-drm_mm.c static bool colors_abutt(const struct drm_mm_node *node) node 1913 drivers/gpu/drm/selftests/test-drm_mm.c if (!drm_mm_hole_follows(node) && node 1914 drivers/gpu/drm/selftests/test-drm_mm.c list_next_entry(node, node_list)->allocated) { node 1916 drivers/gpu/drm/selftests/test-drm_mm.c node->color, node->start, node->size, node 1917 drivers/gpu/drm/selftests/test-drm_mm.c list_next_entry(node, node_list)->color, node 1918 drivers/gpu/drm/selftests/test-drm_mm.c list_next_entry(node, node_list)->start, node 1919 drivers/gpu/drm/selftests/test-drm_mm.c list_next_entry(node, node_list)->size); node 1931 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *nn; node 1945 drivers/gpu/drm/selftests/test-drm_mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 1946 drivers/gpu/drm/selftests/test-drm_mm.c if (!node) { node 1951 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, node 1955 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 1960 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, nn, &mm) { node 1961 drivers/gpu/drm/selftests/test-drm_mm.c if (node->color != node->size) { node 1963 drivers/gpu/drm/selftests/test-drm_mm.c node->size, node->color); node 1968 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 1969 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 1977 drivers/gpu/drm/selftests/test-drm_mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 1978 drivers/gpu/drm/selftests/test-drm_mm.c if (!node) { node 1983 drivers/gpu/drm/selftests/test-drm_mm.c node->size = 1 + 2*count; node 1984 drivers/gpu/drm/selftests/test-drm_mm.c node->color = node->size; node 1986 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, node); node 1993 drivers/gpu/drm/selftests/test-drm_mm.c last = node->start + node->size; node 1998 drivers/gpu/drm/selftests/test-drm_mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 1999 drivers/gpu/drm/selftests/test-drm_mm.c if (!node) { node 2004 drivers/gpu/drm/selftests/test-drm_mm.c node->start = last; node 2005 drivers/gpu/drm/selftests/test-drm_mm.c node->size = n + count; node 2006 drivers/gpu/drm/selftests/test-drm_mm.c node->color = node->size; node 2008 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, node); node 2015 drivers/gpu/drm/selftests/test-drm_mm.c node->start += n + 1; node 2016 drivers/gpu/drm/selftests/test-drm_mm.c rem = misalignment(node, n + count); node 2017 drivers/gpu/drm/selftests/test-drm_mm.c node->start += n + count - rem; node 2019 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(&mm, node); node 2026 drivers/gpu/drm/selftests/test-drm_mm.c last = node->start + node->size; node 2030 drivers/gpu/drm/selftests/test-drm_mm.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 2031 drivers/gpu/drm/selftests/test-drm_mm.c if (!node) { node 2036 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, node, node 2041 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 2046 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, nn, &mm) { node 2049 drivers/gpu/drm/selftests/test-drm_mm.c if (node->color != node->size) { node 2051 drivers/gpu/drm/selftests/test-drm_mm.c mode->name, node->size, node->color); node 2056 drivers/gpu/drm/selftests/test-drm_mm.c if (colors_abutt(node)) node 2059 drivers/gpu/drm/selftests/test-drm_mm.c div64_u64_rem(node->start, node->size, &rem); node 2062 drivers/gpu/drm/selftests/test-drm_mm.c mode->name, node->start, node->size, rem); node 2066 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 2067 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 2075 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, nn, &mm) { node 2076 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 2077 drivers/gpu/drm/selftests/test-drm_mm.c kfree(node); node 2140 drivers/gpu/drm/selftests/test-drm_mm.c err = drm_mm_reserve_node(mm, &e->node); node 2143 drivers/gpu/drm/selftests/test-drm_mm.c e->node.start); node 2160 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; node 2183 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n].node, node 2242 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 2243 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 2263 drivers/gpu/drm/selftests/test-drm_mm.c struct drm_mm_node *node, *next; node 2284 drivers/gpu/drm/selftests/test-drm_mm.c if (!expect_insert(&mm, &nodes[n].node, node 2343 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_node_safe(node, next, &mm) node 2344 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_remove_node(node); node 108 drivers/gpu/drm/sti/sti_cursor.c struct drm_info_node *node = s->private; node 109 drivers/gpu/drm/sti/sti_cursor.c struct sti_cursor *cursor = (struct sti_cursor *)node->info_ent->data; node 77 drivers/gpu/drm/sti/sti_drv.c struct drm_info_node *node = s->private; node 78 drivers/gpu/drm/sti/sti_drv.c struct drm_device *dev = node->minor->dev; node 258 drivers/gpu/drm/sti/sti_drv.c struct device_node *node = dev->of_node; node 266 drivers/gpu/drm/sti/sti_drv.c child_np = of_get_next_available_child(node, NULL); node 271 drivers/gpu/drm/sti/sti_drv.c child_np = of_get_next_available_child(node, child_np); node 180 drivers/gpu/drm/sti/sti_dvo.c struct drm_info_node *node = s->private; node 181 drivers/gpu/drm/sti/sti_dvo.c struct sti_dvo *dvo = (struct sti_dvo *)node->info_ent->data; node 215 drivers/gpu/drm/sti/sti_gdp.c struct drm_info_node *node = s->private; node 216 drivers/gpu/drm/sti/sti_gdp.c struct sti_gdp *gdp = (struct sti_gdp *)node->info_ent->data; node 258 drivers/gpu/drm/sti/sti_gdp.c static void gdp_node_dump_node(struct seq_file *s, struct sti_gdp_node *node) node 260 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\t@:0x%p", node); node 261 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tCTL 0x%08X", node->gam_gdp_ctl); node 262 drivers/gpu/drm/sti/sti_gdp.c gdp_dbg_ctl(s, node->gam_gdp_ctl); node 263 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tAGC 0x%08X", node->gam_gdp_agc); node 264 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tVPO 0x%08X", node->gam_gdp_vpo); node 265 drivers/gpu/drm/sti/sti_gdp.c gdp_dbg_vpo(s, node->gam_gdp_vpo); node 266 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tVPS 0x%08X", node->gam_gdp_vps); node 267 drivers/gpu/drm/sti/sti_gdp.c gdp_dbg_vps(s, node->gam_gdp_vps); node 268 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tPML 0x%08X", node->gam_gdp_pml); node 269 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tPMP 0x%08X", node->gam_gdp_pmp); node 270 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tSIZE 0x%08X", node->gam_gdp_size); node 271 drivers/gpu/drm/sti/sti_gdp.c gdp_dbg_size(s, node->gam_gdp_size); node 272 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tNVN 0x%08X", node->gam_gdp_nvn); node 273 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tKEY1 0x%08X", node->gam_gdp_key1); node 274 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tKEY2 0x%08X", node->gam_gdp_key2); node 275 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tPPT 0x%08X", node->gam_gdp_ppt); node 276 drivers/gpu/drm/sti/sti_gdp.c gdp_dbg_ppt(s, node->gam_gdp_ppt); node 277 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n\tCML 0x%08X\n", node->gam_gdp_cml); node 282 drivers/gpu/drm/sti/sti_gdp.c struct drm_info_node *node = s->private; node 283 drivers/gpu/drm/sti/sti_gdp.c struct sti_gdp *gdp = (struct sti_gdp *)node->info_ent->data; node 346 drivers/gpu/drm/sti/sti_hda.c struct drm_info_node *node = s->private; node 347 drivers/gpu/drm/sti/sti_hda.c struct sti_hda *hda = (struct sti_hda *)node->info_ent->data; node 663 drivers/gpu/drm/sti/sti_hdmi.c struct drm_info_node *node = s->private; node 664 drivers/gpu/drm/sti/sti_hdmi.c struct sti_hdmi *hdmi = (struct sti_hdmi *)node->info_ent->data; node 565 drivers/gpu/drm/sti/sti_hqvdp.c struct drm_info_node *node = s->private; node 566 drivers/gpu/drm/sti/sti_hqvdp.c struct sti_hqvdp *hqvdp = (struct sti_hqvdp *)node->info_ent->data; node 150 drivers/gpu/drm/sti/sti_mixer.c struct drm_info_node *node = s->private; node 151 drivers/gpu/drm/sti/sti_mixer.c struct sti_mixer *mixer = (struct sti_mixer *)node->info_ent->data; node 499 drivers/gpu/drm/sti/sti_tvout.c struct drm_info_node *node = s->private; node 500 drivers/gpu/drm/sti/sti_tvout.c struct sti_tvout *tvout = (struct sti_tvout *)node->info_ent->data; node 842 drivers/gpu/drm/sti/sti_tvout.c struct device_node *node = dev->of_node; node 848 drivers/gpu/drm/sti/sti_tvout.c if (!node) node 95 drivers/gpu/drm/sti/sti_vid.c struct drm_info_node *node = s->private; node 96 drivers/gpu/drm/sti/sti_vid.c struct sti_vid *vid = (struct sti_vid *)node->info_ent->data; node 712 drivers/gpu/drm/sun4i/sun4i_backend.c static int sun4i_backend_of_get_id(struct device_node *node) node 718 drivers/gpu/drm/sun4i/sun4i_backend.c ep = of_graph_get_endpoint_by_regs(node, 0, -1); node 734 drivers/gpu/drm/sun4i/sun4i_backend.c struct device_node *node) node 739 drivers/gpu/drm/sun4i/sun4i_backend.c port = of_graph_get_port_by_id(node, 0); node 751 drivers/gpu/drm/sun4i/sun4i_backend.c if (remote == frontend->node) { node 818 drivers/gpu/drm/sun4i/sun4i_backend.c backend->engine.node = dev->of_node; node 148 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_connector(struct device_node *node) node 150 drivers/gpu/drm/sun4i/sun4i_drv.c return of_device_is_compatible(node, "hdmi-connector"); node 153 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_frontend(struct device_node *node) node 155 drivers/gpu/drm/sun4i/sun4i_drv.c return of_device_is_compatible(node, "allwinner,sun4i-a10-display-frontend") || node 156 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun5i-a13-display-frontend") || node 157 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun6i-a31-display-frontend") || node 158 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun7i-a20-display-frontend") || node 159 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun8i-a23-display-frontend") || node 160 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun8i-a33-display-frontend") || node 161 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_compatible(node, "allwinner,sun9i-a80-display-frontend"); node 164 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_deu(struct device_node *node) node 166 drivers/gpu/drm/sun4i/sun4i_drv.c return of_device_is_compatible(node, "allwinner,sun9i-a80-deu"); node 169 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_supported_frontend(struct device_node *node) node 172 drivers/gpu/drm/sun4i/sun4i_drv.c return !!of_match_node(sun4i_frontend_of_table, node); node 177 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon(struct device_node *node) node 179 drivers/gpu/drm/sun4i/sun4i_drv.c return !!of_match_node(sun4i_tcon_of_table, node); node 182 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon_with_ch0(struct device_node *node) node 186 drivers/gpu/drm/sun4i/sun4i_drv.c match = of_match_node(sun4i_tcon_of_table, node); node 198 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon_top(struct device_node *node) node 201 drivers/gpu/drm/sun4i/sun4i_drv.c !!of_match_node(sun8i_tcon_top_of_table, node); node 239 drivers/gpu/drm/sun4i/sun4i_drv.c struct device_node *node, node 244 drivers/gpu/drm/sun4i/sun4i_drv.c port = of_graph_get_port_by_id(node, port_id); node 257 drivers/gpu/drm/sun4i/sun4i_drv.c if (sun4i_drv_node_is_tcon(node)) { node 275 drivers/gpu/drm/sun4i/sun4i_drv.c if (sun4i_drv_node_is_tcon_with_ch0(node)) { node 299 drivers/gpu/drm/sun4i/sun4i_drv.c struct device_node *node) node 310 drivers/gpu/drm/sun4i/sun4i_drv.c if (!sun4i_drv_node_is_frontend(node) && node 311 drivers/gpu/drm/sun4i/sun4i_drv.c !of_device_is_available(node)) node 318 drivers/gpu/drm/sun4i/sun4i_drv.c if (sun4i_drv_node_is_connector(node)) node 326 drivers/gpu/drm/sun4i/sun4i_drv.c if (!(sun4i_drv_node_is_frontend(node) || node 327 drivers/gpu/drm/sun4i/sun4i_drv.c sun4i_drv_node_is_deu(node)) || node 328 drivers/gpu/drm/sun4i/sun4i_drv.c (sun4i_drv_node_is_supported_frontend(node) && node 329 drivers/gpu/drm/sun4i/sun4i_drv.c of_device_is_available(node))) { node 331 drivers/gpu/drm/sun4i/sun4i_drv.c DRM_DEBUG_DRIVER("Adding component %pOF\n", node); node 332 drivers/gpu/drm/sun4i/sun4i_drv.c drm_of_component_match_add(dev, match, compare_of, node); node 337 drivers/gpu/drm/sun4i/sun4i_drv.c sun4i_drv_traverse_endpoints(list, node, 1); node 340 drivers/gpu/drm/sun4i/sun4i_drv.c if (sun4i_drv_node_is_tcon_top(node)) { node 341 drivers/gpu/drm/sun4i/sun4i_drv.c sun4i_drv_traverse_endpoints(list, node, 3); node 342 drivers/gpu/drm/sun4i/sun4i_drv.c sun4i_drv_traverse_endpoints(list, node, 5); node 571 drivers/gpu/drm/sun4i/sun4i_frontend.c frontend->node = dev->of_node; node 128 drivers/gpu/drm/sun4i/sun4i_frontend.h struct device_node *node; node 859 drivers/gpu/drm/sun4i/sun4i_tcon.c struct device_node *node, node 866 drivers/gpu/drm/sun4i/sun4i_tcon.c port = of_graph_get_port_by_id(node, port_id); node 893 drivers/gpu/drm/sun4i/sun4i_tcon.c if (remote == engine->node) node 973 drivers/gpu/drm/sun4i/sun4i_tcon.c static bool sun4i_tcon_connected_to_tcon_top(struct device_node *node) node 978 drivers/gpu/drm/sun4i/sun4i_tcon.c remote = of_graph_get_remote_node(node, 0, -1); node 1038 drivers/gpu/drm/sun4i/sun4i_tcon.c struct device_node *node) node 1043 drivers/gpu/drm/sun4i/sun4i_tcon.c port = of_graph_get_port_by_id(node, 0); node 1066 drivers/gpu/drm/sun4i/sun4i_tcon.c if (sun4i_tcon_connected_to_tcon_top(node)) node 1080 drivers/gpu/drm/sun4i/sun4i_tcon.c return sun4i_tcon_find_engine_traverse(drv, node, 0); node 1293 drivers/gpu/drm/sun4i/sun4i_tcon.c struct device_node *node = pdev->dev.of_node; node 1303 drivers/gpu/drm/sun4i/sun4i_tcon.c ret = drm_of_find_panel_or_bridge(node, 1, 0, &panel, &bridge); node 269 drivers/gpu/drm/sun4i/sun4i_tcon.h struct drm_bridge *sun4i_tcon_find_bridge(struct device_node *node); node 270 drivers/gpu/drm/sun4i/sun4i_tcon.h struct drm_panel *sun4i_tcon_find_panel(struct device_node *node); node 62 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c static bool sun8i_dw_hdmi_node_is_tcon_top(struct device_node *node) node 65 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c !!of_match_node(sun8i_tcon_top_of_table, node); node 69 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c struct device_node *node) node 74 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c remote = of_graph_get_remote_node(node, 0, -1); node 91 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c crtcs = drm_of_find_possible_crtcs(drm, node); node 203 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.h int sun8i_hdmi_phy_probe(struct sun8i_dw_hdmi *hdmi, struct device_node *node); node 607 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c int sun8i_hdmi_phy_probe(struct sun8i_dw_hdmi *hdmi, struct device_node *node) node 616 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c match = of_match_node(sun8i_hdmi_phy_of_table, node); node 628 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c ret = of_address_to_resource(node, 0, &res); node 647 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c phy->clk_bus = of_clk_get_by_name(node, "bus"); node 653 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c phy->clk_mod = of_clk_get_by_name(node, "mod"); node 661 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c phy->clk_pll0 = of_clk_get_by_name(node, "pll-0"); node 669 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c phy->clk_pll1 = of_clk_get_by_name(node, "pll-1"); node 678 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c phy->rst_phy = of_reset_control_get_shared(node, "phy"); node 402 drivers/gpu/drm/sun4i/sun8i_mixer.c static int sun8i_mixer_of_get_id(struct device_node *node) node 408 drivers/gpu/drm/sun4i/sun8i_mixer.c ep = of_graph_get_endpoint_by_regs(node, 1, -1); node 453 drivers/gpu/drm/sun4i/sun8i_mixer.c mixer->engine.node = dev->of_node; node 23 drivers/gpu/drm/sun4i/sun8i_tcon_top.c static bool sun8i_tcon_top_node_is_tcon_top(struct device_node *node) node 25 drivers/gpu/drm/sun4i/sun8i_tcon_top.c return !!of_match_node(sun8i_tcon_top_of_table, node); node 123 drivers/gpu/drm/sun4i/sunxi_engine.h struct device_node *node; node 96 drivers/gpu/drm/tegra/dc.c if (it.node == dev->of_node) node 1417 drivers/gpu/drm/tegra/dc.c struct drm_info_node *node = s->private; node 1418 drivers/gpu/drm/tegra/dc.c struct tegra_dc *dc = node->info_ent->data; node 1443 drivers/gpu/drm/tegra/dc.c struct drm_info_node *node = s->private; node 1444 drivers/gpu/drm/tegra/dc.c struct tegra_dc *dc = node->info_ent->data; node 1474 drivers/gpu/drm/tegra/dc.c struct drm_info_node *node = s->private; node 1475 drivers/gpu/drm/tegra/dc.c struct tegra_dc *dc = node->info_ent->data; node 966 drivers/gpu/drm/tegra/drm.c struct drm_info_node *node = (struct drm_info_node *)s->private; node 967 drivers/gpu/drm/tegra/drm.c struct drm_device *drm = node->minor->dev; node 987 drivers/gpu/drm/tegra/drm.c struct drm_info_node *node = (struct drm_info_node *)s->private; node 988 drivers/gpu/drm/tegra/drm.c struct drm_device *drm = node->minor->dev; node 200 drivers/gpu/drm/tegra/dsi.c struct drm_info_node *node = s->private; node 201 drivers/gpu/drm/tegra/dsi.c struct tegra_dsi *dsi = node->info_ent->data; node 203 drivers/gpu/drm/tegra/dsi.c struct drm_device *drm = node->minor->dev; node 1030 drivers/gpu/drm/tegra/hdmi.c struct drm_info_node *node = s->private; node 1031 drivers/gpu/drm/tegra/hdmi.c struct tegra_hdmi *hdmi = node->info_ent->data; node 1033 drivers/gpu/drm/tegra/hdmi.c struct drm_device *drm = node->minor->dev; node 1265 drivers/gpu/drm/tegra/sor.c struct drm_info_node *node = s->private; node 1266 drivers/gpu/drm/tegra/sor.c struct tegra_sor *sor = node->info_ent->data; node 1268 drivers/gpu/drm/tegra/sor.c struct drm_device *drm = node->minor->dev; node 1427 drivers/gpu/drm/tegra/sor.c struct drm_info_node *node = s->private; node 1428 drivers/gpu/drm/tegra/sor.c struct tegra_sor *sor = node->info_ent->data; node 1430 drivers/gpu/drm/tegra/sor.c struct drm_device *drm = node->minor->dev; node 225 drivers/gpu/drm/tilcdc/tilcdc_drv.c struct device_node *node = dev->of_node; node 273 drivers/gpu/drm/tilcdc/tilcdc_drv.c if (of_property_read_u32(node, "max-bandwidth", &priv->max_bandwidth)) node 278 drivers/gpu/drm/tilcdc/tilcdc_drv.c if (of_property_read_u32(node, "max-width", &priv->max_width)) node 283 drivers/gpu/drm/tilcdc/tilcdc_drv.c if (of_property_read_u32(node, "max-pixelclock", node 319 drivers/gpu/drm/tilcdc/tilcdc_drv.c of_property_read_string(node, "blue-and-red-wiring", &str); node 455 drivers/gpu/drm/tilcdc/tilcdc_drv.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 456 drivers/gpu/drm/tilcdc/tilcdc_drv.c struct drm_device *dev = node->minor->dev; node 476 drivers/gpu/drm/tilcdc/tilcdc_drv.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 477 drivers/gpu/drm/tilcdc/tilcdc_drv.c struct drm_device *dev = node->minor->dev; node 169 drivers/gpu/drm/tilcdc/tilcdc_external.c struct device_node *node; node 171 drivers/gpu/drm/tilcdc/tilcdc_external.c node = of_graph_get_remote_node(dev->of_node, 0, 0); node 173 drivers/gpu/drm/tilcdc/tilcdc_external.c if (!of_device_is_compatible(node, "nxp,tda998x")) { node 174 drivers/gpu/drm/tilcdc/tilcdc_external.c of_node_put(node); node 179 drivers/gpu/drm/tilcdc/tilcdc_external.c drm_of_component_match_add(dev, match, dev_match_of, node); node 180 drivers/gpu/drm/tilcdc/tilcdc_external.c of_node_put(node); node 306 drivers/gpu/drm/tilcdc/tilcdc_panel.c struct device_node *bl_node, *node = pdev->dev.of_node; node 313 drivers/gpu/drm/tilcdc/tilcdc_panel.c if (!node) { node 322 drivers/gpu/drm/tilcdc/tilcdc_panel.c bl_node = of_parse_phandle(node, "backlight", 0); node 353 drivers/gpu/drm/tilcdc/tilcdc_panel.c panel_mod->timings = of_get_display_timings(node); node 360 drivers/gpu/drm/tilcdc/tilcdc_panel.c panel_mod->info = of_get_panel_info(node); node 274 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c struct device_node *node = pdev->dev.of_node; node 283 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c if (!node) { node 301 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c if (of_property_read_u32(node, "i2c", &i2c_phandle)) { node 321 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c tfp410_mod->gpio = of_get_named_gpio_flags(node, "powerdn-gpio", node 55 drivers/gpu/drm/ttm/ttm_agp_backend.c struct drm_mm_node *node = bo_mem->mm_node; node 78 drivers/gpu/drm/ttm/ttm_agp_backend.c ret = agp_bind_memory(mem, node->start); node 58 drivers/gpu/drm/ttm/ttm_bo_manager.c struct drm_mm_node *node; node 67 drivers/gpu/drm/ttm/ttm_bo_manager.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 68 drivers/gpu/drm/ttm/ttm_bo_manager.c if (!node) node 76 drivers/gpu/drm/ttm/ttm_bo_manager.c ret = drm_mm_insert_node_in_range(mm, node, node 83 drivers/gpu/drm/ttm/ttm_bo_manager.c kfree(node); node 85 drivers/gpu/drm/ttm/ttm_bo_manager.c mem->mm_node = node; node 86 drivers/gpu/drm/ttm/ttm_bo_manager.c mem->start = node->start; node 407 drivers/gpu/drm/ttm/ttm_bo_vm.c struct drm_vma_offset_node *node; node 412 drivers/gpu/drm/ttm/ttm_bo_vm.c node = drm_vma_offset_lookup_locked(&bdev->vma_manager, offset, pages); node 413 drivers/gpu/drm/ttm/ttm_bo_vm.c if (likely(node)) { node 414 drivers/gpu/drm/ttm/ttm_bo_vm.c bo = container_of(node, struct ttm_buffer_object, node 236 drivers/gpu/drm/udl/udl_fb.c info->node, user, info, ufbdev->fb_count); node 261 drivers/gpu/drm/udl/udl_fb.c info->node, user, ufbdev->fb_count); node 170 drivers/gpu/drm/udl/udl_main.c struct list_head *node; node 182 drivers/gpu/drm/udl/udl_main.c node = udl->urbs.list.next; /* have reserved one with sem */ node 183 drivers/gpu/drm/udl/udl_main.c list_del_init(node); node 187 drivers/gpu/drm/udl/udl_main.c unode = list_entry(node, struct urb_node, entry); node 194 drivers/gpu/drm/udl/udl_main.c kfree(node); node 44 drivers/gpu/drm/v3d/v3d_bo.c drm_mm_remove_node(&bo->node); node 107 drivers/gpu/drm/v3d/v3d_bo.c ret = drm_mm_insert_node_generic(&v3d->mm, &bo->node, node 185 drivers/gpu/drm/v3d/v3d_bo.c args->offset = bo->node.start << PAGE_SHIFT; node 230 drivers/gpu/drm/v3d/v3d_bo.c args->offset = bo->node.start << PAGE_SHIFT; node 82 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 83 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_device *dev = node->minor->dev; node 129 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 130 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_device *dev = node->minor->dev; node 198 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 199 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_device *dev = node->minor->dev; node 214 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 215 drivers/gpu/drm/v3d/v3d_debugfs.c struct drm_device *dev = node->minor->dev; node 143 drivers/gpu/drm/v3d/v3d_drv.h struct drm_mm_node node; node 71 drivers/gpu/drm/v3d/v3d_irq.c V3D_CORE_WRITE(0, V3D_PTB_BPOA, bo->node.start << PAGE_SHIFT); node 91 drivers/gpu/drm/v3d/v3d_mmu.c u32 page = bo->node.start; node 108 drivers/gpu/drm/v3d/v3d_mmu.c WARN_ON_ONCE(page - bo->node.start != node 121 drivers/gpu/drm/v3d/v3d_mmu.c for (page = bo->node.start; page < bo->node.start + npages; page++) node 68 drivers/gpu/drm/vc4/vc4_bo.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 69 drivers/gpu/drm/vc4/vc4_bo.c struct drm_device *dev = node->minor->dev; node 45 drivers/gpu/drm/vc4/vc4_debugfs.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 46 drivers/gpu/drm/vc4/vc4_debugfs.c struct debugfs_regset32 *regset = node->info_ent->data; node 257 drivers/gpu/drm/vc4/vc4_drv.c struct device_node *node; node 267 drivers/gpu/drm/vc4/vc4_drv.c node = of_find_matching_node_and_match(NULL, vc4_v3d_dt_match, NULL); node 268 drivers/gpu/drm/vc4/vc4_drv.c if (!node || !of_device_is_available(node)) node 270 drivers/gpu/drm/vc4/vc4_drv.c of_node_put(node); node 184 drivers/gpu/drm/vc4/vc4_hdmi.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 185 drivers/gpu/drm/vc4/vc4_hdmi.c struct drm_device *dev = node->minor->dev; node 85 drivers/gpu/drm/vc4/vc4_hvs.c struct drm_info_node *node = m->private; node 86 drivers/gpu/drm/vc4/vc4_hvs.c struct drm_device *dev = node->minor->dev; node 101 drivers/gpu/drm/vc4/vc4_v3d.c struct drm_info_node *node = (struct drm_info_node *)m->private; node 102 drivers/gpu/drm/vc4/vc4_v3d.c struct drm_device *dev = node->minor->dev; node 45 drivers/gpu/drm/virtio/virtgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 46 drivers/gpu/drm/virtio/virtgpu_debugfs.c struct virtio_gpu_device *vgdev = node->minor->dev->dev_private; node 58 drivers/gpu/drm/virtio/virtgpu_debugfs.c struct drm_info_node *node = (struct drm_info_node *) m->private; node 59 drivers/gpu/drm/virtio/virtgpu_debugfs.c struct virtio_gpu_device *vgdev = node->minor->dev->dev_private; node 104 drivers/gpu/drm/virtio/virtgpu_drv.h struct list_head node; node 98 drivers/gpu/drm/virtio/virtgpu_fence.c list_add_tail(&fence->node, &drv->fences); node 116 drivers/gpu/drm/virtio/virtgpu_fence.c list_for_each_entry_safe(fence, tmp, &drv->fences, node) { node 120 drivers/gpu/drm/virtio/virtgpu_fence.c list_del(&fence->node); node 88 drivers/gpu/drm/vkms/vkms_crtc.c *vblank_time = output->vblank_hrtimer.node.expires; node 158 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct drm_mm_node node; node 187 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct drm_mm_node *node; node 267 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c drm_mm_remove_node(&header->node); node 770 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c memset(info->node, 0, sizeof(*info->node)); node 772 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ret = drm_mm_insert_node(&man->mm, info->node, info->page_size); node 775 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ret = drm_mm_insert_node(&man->mm, info->node, info->page_size); node 797 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct drm_mm_node *node, node 804 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c info.node = node; node 872 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ret = vmw_cmdbuf_alloc_space(man, &header->node, size, interruptible); node 884 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c header->size = header->node.size << PAGE_SHIFT; node 886 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c offset = header->node.start << PAGE_SHIFT; node 900 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c drm_mm_remove_node(&header->node); node 211 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_validation_info *node) node 229 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->staged = vmw_binding_state_alloc(dev_priv); node 230 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (IS_ERR(node->staged)) { node 231 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c ret = PTR_ERR(node->staged); node 232 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->staged = NULL; node 236 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->staged = sw_context->staged_bindings; node 240 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->ctx = res; node 241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->cur = vmw_context_binding_state(res); node 242 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c list_add_tail(&node->head, &sw_context->ctx_list); node 899 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_validation_info *node; node 901 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node = vmw_execbuf_info_from_res(sw_context, ctx); node 902 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (!node) node 909 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_add(node->staged, &binding.bi, 0, binding.slot); node 1660 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_validation_info *node; node 1662 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node = vmw_execbuf_info_from_res(sw_context, ctx); node 1663 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (!node) node 1670 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_add(node->staged, &binding.bi, 0, node 198 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c list_add(&cb.base.node, &f->cb_list); node 234 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c if (!list_empty(&cb.base.node)) node 235 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c list_del(&cb.base.node); node 317 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_res_node *node; node 320 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node = vmw_validation_find_res_dup(ctx, res); node 321 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (node) { node 322 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->first_usage = 0; node 326 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node = vmw_validation_mem_alloc(ctx, sizeof(*node) + priv_size); node 327 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (!node) { node 333 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->hash.key = (unsigned long) res; node 334 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ret = drm_ht_insert_item(ctx->ht, &node->hash); node 341 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->res = vmw_resource_reference_unless_doomed(res); node 342 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (!node->res) node 345 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->first_usage = 1; node 347 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_add_tail(&node->head, &ctx->resource_list); node 352 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_add(&node->head, &ctx->resource_ctx_list); node 355 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_add_tail(&node->head, &ctx->resource_ctx_list); node 358 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_add_tail(&node->head, &ctx->resource_list); node 365 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->dirty_set = 1; node 367 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; node 370 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c *first_usage = node->first_usage; node 372 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c *p_node = &node->private; node 657 drivers/hid/hid-debug.c list_for_each_entry(list, &hdev->debug_list, node) node 1084 drivers/hid/hid-debug.c list_add_tail(&list->node, &list->hdev->debug_list); node 1169 drivers/hid/hid-debug.c list_del(&list->node); node 1016 drivers/hid/hid-quirks.c struct list_head node; node 1040 drivers/hid/hid-quirks.c list_for_each_entry(q, &dquirks_list, node) { node 1094 drivers/hid/hid-quirks.c list_for_each_entry(q, &dquirks_list, node) { node 1098 drivers/hid/hid-quirks.c list_replace(&q->node, &q_new->node); node 1108 drivers/hid/hid-quirks.c list_add_tail(&q_new->node, &dquirks_list); node 1131 drivers/hid/hid-quirks.c list_for_each_entry_safe(q, temp, &dquirks_list, node) { node 1133 drivers/hid/hid-quirks.c list_del(&q->node); node 61 drivers/hid/hid-roccat.c struct list_head node; node 191 drivers/hid/hid-roccat.c list_add_tail(&reader->node, &device->readers); node 219 drivers/hid/hid-roccat.c list_del(&reader->node); node 268 drivers/hid/hid-roccat.c list_for_each_entry(reader, &device->readers, node) { node 300 drivers/hid/hidraw.c list_add_tail(&list->node, &hidraw_table[minor]->list); node 353 drivers/hid/hidraw.c list_del(&list->node); node 486 drivers/hid/hidraw.c list_for_each_entry(list, &dev->list, node) { node 45 drivers/hid/usbhid/hiddev.c struct list_head node; node 148 drivers/hid/usbhid/hiddev.c list_for_each_entry(list, &hiddev->list, node) { node 222 drivers/hid/usbhid/hiddev.c list_del(&list->node); node 269 drivers/hid/usbhid/hiddev.c list_add_tail(&list->node, &hiddev->list); node 436 drivers/hsi/controllers/omap_ssi_port.c struct list_head *node, *tmp; node 439 drivers/hsi/controllers/omap_ssi_port.c list_for_each_safe(node, tmp, queue) { node 440 drivers/hsi/controllers/omap_ssi_port.c msg = list_entry(node, struct hsi_msg, link); node 443 drivers/hsi/controllers/omap_ssi_port.c list_del(node); node 40 drivers/hv/vmbus_drv.c struct list_head node; node 724 drivers/hv/vmbus_drv.c list_for_each_entry(dynid, &drv->dynids.list, node) { node 775 drivers/hv/vmbus_drv.c list_add_tail(&dynid->node, &drv->dynids.list); node 786 drivers/hv/vmbus_drv.c list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { node 787 drivers/hv/vmbus_drv.c list_del(&dynid->node); node 838 drivers/hv/vmbus_drv.c list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { node 842 drivers/hv/vmbus_drv.c list_del(&dynid->node); node 1082 drivers/hwmon/applesmc.c struct applesmc_dev_attr *node; node 1085 drivers/hwmon/applesmc.c for (node = grp->nodes; node->sda.dev_attr.attr.name; node++) node 1087 drivers/hwmon/applesmc.c &node->sda.dev_attr.attr); node 1099 drivers/hwmon/applesmc.c struct applesmc_dev_attr *node; node 1104 drivers/hwmon/applesmc.c grp->nodes = kcalloc(num + 1, sizeof(*node), GFP_KERNEL); node 1110 drivers/hwmon/applesmc.c node = &grp->nodes[i]; node 1111 drivers/hwmon/applesmc.c scnprintf(node->name, sizeof(node->name), grp->format, node 1113 drivers/hwmon/applesmc.c node->sda.index = (grp->option << 16) | (i & 0xffff); node 1114 drivers/hwmon/applesmc.c node->sda.dev_attr.show = grp->show; node 1115 drivers/hwmon/applesmc.c node->sda.dev_attr.store = grp->store; node 1116 drivers/hwmon/applesmc.c attr = &node->sda.dev_attr.attr; node 1118 drivers/hwmon/applesmc.c attr->name = node->name; node 388 drivers/hwmon/ibmpowernv.c struct device_node *node, node 412 drivers/hwmon/ibmpowernv.c if (it.phandle == node->phandle) { node 413 drivers/hwmon/ibmpowernv.c of_node_put(it.node); node 468 drivers/hwmon/max6697.c static void max6697_get_config_of(struct device_node *node, node 475 drivers/hwmon/max6697.c of_property_read_bool(node, "smbus-timeout-disable"); node 477 drivers/hwmon/max6697.c of_property_read_bool(node, "extended-range-enable"); node 479 drivers/hwmon/max6697.c of_property_read_bool(node, "beta-compensation-enable"); node 481 drivers/hwmon/max6697.c prop = of_get_property(node, "alert-mask", &len); node 484 drivers/hwmon/max6697.c prop = of_get_property(node, "over-temperature-mask", &len); node 487 drivers/hwmon/max6697.c prop = of_get_property(node, "resistance-cancellation", &len); node 494 drivers/hwmon/max6697.c prop = of_get_property(node, "transistor-ideality", &len); node 76 drivers/hwspinlock/omap_hwspinlock.c struct device_node *node = pdev->dev.of_node; node 85 drivers/hwspinlock/omap_hwspinlock.c if (!node) node 260 drivers/hwtracing/coresight/coresight-catu.c catu_init_sg_table(struct device *catu_dev, int node, node 271 drivers/hwtracing/coresight/coresight-catu.c catu_table = tmc_alloc_sg_table(catu_dev, node, nr_tpages, node 330 drivers/hwtracing/coresight/coresight-catu.c struct etr_buf *etr_buf, int node, void **pages) node 343 drivers/hwtracing/coresight/coresight-catu.c catu_table = catu_init_sg_table(&csdev->dev, node, node 377 drivers/hwtracing/coresight/coresight-etb10.c int node; node 380 drivers/hwtracing/coresight/coresight-etb10.c node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu); node 382 drivers/hwtracing/coresight/coresight-etb10.c buf = kzalloc_node(sizeof(struct cs_buffers), GFP_KERNEL, node); node 84 drivers/hwtracing/coresight/coresight-etm-perf.c int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu); node 86 drivers/hwtracing/coresight/coresight-etm-perf.c filters = kzalloc_node(sizeof(struct etm_filters), GFP_KERNEL, node); node 66 drivers/hwtracing/coresight/coresight-platform.c static void of_coresight_get_ports_legacy(const struct device_node *node, node 73 drivers/hwtracing/coresight/coresight-platform.c ep = of_graph_get_next_endpoint(node, ep); node 104 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_input_ports_node(const struct device_node *node) node 106 drivers/hwtracing/coresight/coresight-platform.c return of_get_child_by_name(node, "in-ports"); node 110 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_output_ports_node(const struct device_node *node) node 112 drivers/hwtracing/coresight/coresight-platform.c return of_get_child_by_name(node, "out-ports"); node 126 drivers/hwtracing/coresight/coresight-platform.c static void of_coresight_get_ports(const struct device_node *node, node 131 drivers/hwtracing/coresight/coresight-platform.c input_ports = of_coresight_get_input_ports_node(node); node 132 drivers/hwtracing/coresight/coresight-platform.c output_ports = of_coresight_get_output_ports_node(node); node 145 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_ports_legacy(node, nr_inport, nr_outport); node 245 drivers/hwtracing/coresight/coresight-platform.c struct device_node *node = dev->of_node; node 248 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_ports(node, &pdata->nr_inport, &pdata->nr_outport); node 258 drivers/hwtracing/coresight/coresight-platform.c parent = of_coresight_get_output_ports_node(node); node 266 drivers/hwtracing/coresight/coresight-platform.c parent = node; node 743 drivers/hwtracing/coresight/coresight-stm.c list_for_each_entry(rent, &res_list, node) { node 392 drivers/hwtracing/coresight/coresight-tmc-etf.c int node; node 395 drivers/hwtracing/coresight/coresight-tmc-etf.c node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu); node 398 drivers/hwtracing/coresight/coresight-tmc-etf.c buf = kzalloc_node(sizeof(struct cs_buffers), GFP_KERNEL, node); node 191 drivers/hwtracing/coresight/coresight-tmc-etr.c struct device *dev, int node, node 218 drivers/hwtracing/coresight/coresight-tmc-etr.c page = alloc_pages_node(node, node 291 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table->dev, sg_table->node, node 316 drivers/hwtracing/coresight/coresight-tmc-etr.c int node, node 329 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table->node = node; node 556 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_init_etr_sg_table(struct device *dev, int node, node 570 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table = tmc_alloc_sg_table(dev, node, nr_tpages, nr_dpages, pages); node 591 drivers/hwtracing/coresight/coresight-tmc-etr.c struct etr_buf *etr_buf, int node, node 671 drivers/hwtracing/coresight/coresight-tmc-etr.c struct etr_buf *etr_buf, int node, node 677 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_table = tmc_init_etr_sg_table(dev, node, node 797 drivers/hwtracing/coresight/coresight-tmc-etr.c struct etr_buf *etr_buf, int node, node 808 drivers/hwtracing/coresight/coresight-tmc-etr.c node, pages); node 827 drivers/hwtracing/coresight/coresight-tmc-etr.c int node, void **pages) node 862 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf, node, pages); node 865 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf, node, pages); node 868 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf, node, pages); node 1206 drivers/hwtracing/coresight/coresight-tmc-etr.c int node; node 1210 drivers/hwtracing/coresight/coresight-tmc-etr.c node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu); node 1217 drivers/hwtracing/coresight/coresight-tmc-etr.c 0, node, NULL); node 1228 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf = tmc_alloc_etr_buf(drvdata, size, 0, node, NULL); node 1332 drivers/hwtracing/coresight/coresight-tmc-etr.c int node; node 1336 drivers/hwtracing/coresight/coresight-tmc-etr.c node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu); node 1338 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_perf = kzalloc_node(sizeof(*etr_perf), GFP_KERNEL, node); node 214 drivers/hwtracing/coresight/coresight-tmc.h int node, void **pages); node 248 drivers/hwtracing/coresight/coresight-tmc.h int node; node 310 drivers/hwtracing/coresight/coresight-tmc.h int node, node 626 drivers/hwtracing/coresight/coresight.c struct coresight_node *node; node 654 drivers/hwtracing/coresight/coresight.c node = kzalloc(sizeof(struct coresight_node), GFP_KERNEL); node 655 drivers/hwtracing/coresight/coresight.c if (!node) node 659 drivers/hwtracing/coresight/coresight.c node->csdev = csdev; node 660 drivers/hwtracing/coresight/coresight.c list_add(&node->link, path); node 193 drivers/hwtracing/intel_th/core.c char *node; node 196 drivers/hwtracing/intel_th/core.c node = kasprintf(GFP_KERNEL, "intel_th%d/%s%d", th->id, node 199 drivers/hwtracing/intel_th/core.c node = kasprintf(GFP_KERNEL, "intel_th%d/%s", th->id, node 202 drivers/hwtracing/intel_th/core.c return node; node 86 drivers/hwtracing/stm/p_sys-t.c struct sys_t_policy_node node; node 107 drivers/hwtracing/stm/p_sys-t.c memcpy(&opriv->node, pn, sizeof(opriv->node)); node 240 drivers/hwtracing/stm/p_sys-t.c if (op->node.ts_interval && node 241 drivers/hwtracing/stm/p_sys-t.c time_after(jiffies, op->ts_jiffies + op->node.ts_interval)) { node 252 drivers/hwtracing/stm/p_sys-t.c if (op->node.clocksync_interval && node 254 drivers/hwtracing/stm/p_sys-t.c op->clocksync_jiffies + op->node.clocksync_interval)) { node 307 drivers/hwtracing/stm/p_sys-t.c if (op->node.do_len) node 325 drivers/hwtracing/stm/p_sys-t.c sz = stm_data_write(data, m, c, false, op->node.uuid.b, UUID_SIZE); node 330 drivers/hwtracing/stm/p_sys-t.c if (op->node.do_len) { node 83 drivers/hwtracing/stm/policy.c struct stp_policy_node *node = to_stp_policy_node(item); node 85 drivers/hwtracing/stm/policy.c return stp_policy_node_priv(node); node 184 drivers/hwtracing/stm/policy.c struct stp_policy_node *node = to_stp_policy_node(item); node 186 drivers/hwtracing/stm/policy.c kfree(node); node 1331 drivers/i2c/busses/i2c-img-scb.c struct device_node *node = pdev->dev.of_node; node 1375 drivers/i2c/busses/i2c-img-scb.c if (!of_property_read_u32(node, "clock-frequency", &val)) node 1380 drivers/i2c/busses/i2c-img-scb.c i2c->adap.dev.of_node = node; node 81 drivers/i2c/busses/i2c-mpc.c void (*setup)(struct device_node *node, struct mpc_i2c *i2c, u32 clock); node 201 drivers/i2c/busses/i2c-mpc.c static int mpc_i2c_get_fdr_52xx(struct device_node *node, u32 clock, node 211 drivers/i2c/busses/i2c-mpc.c *real_clk = mpc5xxx_get_bus_frequency(node) / 2048; node 216 drivers/i2c/busses/i2c-mpc.c divider = mpc5xxx_get_bus_frequency(node) / clock; node 231 drivers/i2c/busses/i2c-mpc.c *real_clk = mpc5xxx_get_bus_frequency(node) / div->divider; node 235 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_52xx(struct device_node *node, node 247 drivers/i2c/busses/i2c-mpc.c ret = mpc_i2c_get_fdr_52xx(node, clock, &i2c->real_clk); node 257 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_52xx(struct device_node *node, node 265 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_512x(struct device_node *node, node 281 drivers/i2c/busses/i2c-mpc.c pval = of_get_property(node, "reg", NULL); node 290 drivers/i2c/busses/i2c-mpc.c mpc_i2c_setup_52xx(node, i2c, clock); node 293 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_512x(struct device_node *node, node 323 drivers/i2c/busses/i2c-mpc.c struct device_node *node; node 327 drivers/i2c/busses/i2c-mpc.c node = of_find_node_by_name(NULL, "global-utilities"); node 328 drivers/i2c/busses/i2c-mpc.c if (node) { node 329 drivers/i2c/busses/i2c-mpc.c const u32 *prop = of_get_property(node, "reg", NULL); node 348 drivers/i2c/busses/i2c-mpc.c of_node_put(node); node 388 drivers/i2c/busses/i2c-mpc.c static int mpc_i2c_get_fdr_8xxx(struct device_node *node, u32 clock, node 421 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_8xxx(struct device_node *node, node 434 drivers/i2c/busses/i2c-mpc.c ret = mpc_i2c_get_fdr_8xxx(node, clock, &i2c->real_clk); node 446 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_8xxx(struct device_node *node, node 766 drivers/i2c/busses/i2c-mxs.c struct device_node *node = dev->of_node; node 769 drivers/i2c/busses/i2c-mxs.c ret = of_property_read_u32(node, "clock-frequency", &speed); node 137 drivers/i2c/busses/i2c-octeon-platdrv.c struct device_node *node = pdev->dev.of_node; node 143 drivers/i2c/busses/i2c-octeon-platdrv.c cn78xx_style = of_device_is_compatible(node, "cavium,octeon-7890-twsi"); node 182 drivers/i2c/busses/i2c-octeon-platdrv.c if (of_property_read_u32(node, "clock-frequency", &i2c->twsi_freq) && node 183 drivers/i2c/busses/i2c-octeon-platdrv.c of_property_read_u32(node, "clock-rate", &i2c->twsi_freq)) { node 244 drivers/i2c/busses/i2c-octeon-platdrv.c i2c->adap.dev.of_node = node; node 1361 drivers/i2c/busses/i2c-omap.c struct device_node *node = pdev->dev.of_node; node 1390 drivers/i2c/busses/i2c-omap.c of_property_read_u32(node, "clock-frequency", &freq); node 38 drivers/i2c/busses/i2c-parport.c struct list_head node; node 243 drivers/i2c/busses/i2c-parport.c list_add_tail(&adapter->node, &adapter_list); node 260 drivers/i2c/busses/i2c-parport.c list_for_each_entry_safe(adapter, _n, &adapter_list, node) { node 274 drivers/i2c/busses/i2c-parport.c list_del(&adapter->node); node 208 drivers/i2c/busses/i2c-powermac.c struct device_node *node) node 214 drivers/i2c/busses/i2c-powermac.c prop = of_get_property(node, "reg", &len); node 219 drivers/i2c/busses/i2c-powermac.c prop = of_get_property(node, "i2c-address", &len); node 224 drivers/i2c/busses/i2c-powermac.c if (of_node_name_eq(node, "cereal")) node 226 drivers/i2c/busses/i2c-powermac.c else if (of_node_name_eq(node, "deq")) node 229 drivers/i2c/busses/i2c-powermac.c dev_warn(&adap->dev, "No i2c address for %pOF\n", node); node 277 drivers/i2c/busses/i2c-powermac.c struct device_node *node, node 293 drivers/i2c/busses/i2c-powermac.c if (of_modalias_node(node, tmp, sizeof(tmp)) >= 0) { node 299 drivers/i2c/busses/i2c-powermac.c if (of_node_name_eq(node, "deq")) { node 310 drivers/i2c/busses/i2c-powermac.c dev_err(&adap->dev, "i2c-powermac: modalias failure on %pOF\n", node); node 318 drivers/i2c/busses/i2c-powermac.c struct device_node *node; node 329 drivers/i2c/busses/i2c-powermac.c for_each_child_of_node(adap->dev.of_node, node) { node 334 drivers/i2c/busses/i2c-powermac.c addr = i2c_powermac_get_addr(adap, bus, node); node 339 drivers/i2c/busses/i2c-powermac.c if (!pmac_i2c_match_adapter(node, adap)) node 342 drivers/i2c/busses/i2c-powermac.c dev_dbg(&adap->dev, "i2c-powermac: register %pOF\n", node); node 348 drivers/i2c/busses/i2c-powermac.c if (of_device_is_compatible(node, "pcm3052")) node 352 drivers/i2c/busses/i2c-powermac.c if (!i2c_powermac_get_type(adap, node, addr, node 359 drivers/i2c/busses/i2c-powermac.c info.irq = irq_of_parse_and_map(node, 0); node 360 drivers/i2c/busses/i2c-powermac.c info.of_node = of_node_get(node); node 365 drivers/i2c/busses/i2c-powermac.c " %pOF\n", node); node 366 drivers/i2c/busses/i2c-powermac.c of_node_put(node); node 227 drivers/i2c/busses/i2c-stm32f7.c struct list_head node; node 476 drivers/i2c/busses/i2c-stm32f7.c list_add_tail(&v->node, node 508 drivers/i2c/busses/i2c-stm32f7.c list_for_each_entry(v, &solutions, node) { node 564 drivers/i2c/busses/i2c-stm32f7.c list_for_each_entry_safe(v, _v, &solutions, node) { node 565 drivers/i2c/busses/i2c-stm32f7.c list_del(&v->node); node 119 drivers/i2c/busses/i2c-thunderx-pcidrv.c struct device_node *node) node 121 drivers/i2c/busses/i2c-thunderx-pcidrv.c if (!node) node 124 drivers/i2c/busses/i2c-thunderx-pcidrv.c i2c->alert_data.irq = irq_of_parse_and_map(node, 0); node 135 drivers/i2c/busses/i2c-thunderx-pcidrv.c struct device_node *node) node 141 drivers/i2c/busses/i2c-thunderx-pcidrv.c return thunder_i2c_smbus_setup_of(i2c, node); node 22 drivers/i2c/i2c-core-of.c int of_i2c_get_board_info(struct device *dev, struct device_node *node, node 30 drivers/i2c/i2c-core-of.c if (of_modalias_node(node, info->type, sizeof(info->type)) < 0) { node 31 drivers/i2c/i2c-core-of.c dev_err(dev, "of_i2c: modalias failure on %pOF\n", node); node 35 drivers/i2c/i2c-core-of.c ret = of_property_read_u32(node, "reg", &addr); node 37 drivers/i2c/i2c-core-of.c dev_err(dev, "of_i2c: invalid reg on %pOF\n", node); node 52 drivers/i2c/i2c-core-of.c info->of_node = node; node 54 drivers/i2c/i2c-core-of.c if (of_property_read_bool(node, "host-notify")) node 57 drivers/i2c/i2c-core-of.c if (of_get_property(node, "wakeup-source", NULL)) node 65 drivers/i2c/i2c-core-of.c struct device_node *node) node 71 drivers/i2c/i2c-core-of.c dev_dbg(&adap->dev, "of_i2c: register %pOF\n", node); node 73 drivers/i2c/i2c-core-of.c ret = of_i2c_get_board_info(&adap->dev, node, &info); node 79 drivers/i2c/i2c-core-of.c dev_err(&adap->dev, "of_i2c: Failure registering %pOF\n", node); node 87 drivers/i2c/i2c-core-of.c struct device_node *bus, *node; node 100 drivers/i2c/i2c-core-of.c for_each_available_child_of_node(bus, node) { node 101 drivers/i2c/i2c-core-of.c if (of_node_test_and_set_flag(node, OF_POPULATED)) node 104 drivers/i2c/i2c-core-of.c client = of_i2c_register_device(adap, node); node 108 drivers/i2c/i2c-core-of.c node); node 109 drivers/i2c/i2c-core-of.c of_node_clear_flag(node, OF_POPULATED); node 128 drivers/i2c/i2c-core-of.c struct i2c_client *of_find_i2c_device_by_node(struct device_node *node) node 133 drivers/i2c/i2c-core-of.c dev = bus_find_device_by_of_node(&i2c_bus_type, node); node 146 drivers/i2c/i2c-core-of.c struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node) node 151 drivers/i2c/i2c-core-of.c dev = bus_find_device(&i2c_bus_type, NULL, node, node 165 drivers/i2c/i2c-core-of.c struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node) node 169 drivers/i2c/i2c-core-of.c adapter = of_find_i2c_adapter_by_node(node); node 64 drivers/i2c/i2c-stub.c struct list_head node; node 96 drivers/i2c/i2c-stub.c list_for_each_entry(b, &chip->smbus_blocks, node) { node 107 drivers/i2c/i2c-stub.c list_add(&rb->node, &chip->smbus_blocks); node 41 drivers/i2c/muxes/i2c-mux-pinctrl.c list_for_each_entry(setting, &state->settings, node) { node 1342 drivers/i3c/master.c list_add_tail(&dev->common.node, &master->bus.devs.i3c); node 1384 drivers/i3c/master.c list_del(&dev->common.node); node 1398 drivers/i3c/master.c list_add_tail(&dev->common.node, &master->bus.devs.i2c); node 1407 drivers/i3c/master.c list_del(&dev->common.node); node 1578 drivers/i3c/master.c common.node) { node 1590 drivers/i3c/master.c common.node) { node 1642 drivers/i3c/master.c list_for_each_entry(i2cboardinfo, &master->boardinfo.i2c, node) { node 1666 drivers/i3c/master.c list_for_each_entry(i3cboardinfo, &master->boardinfo.i3c, node) { node 1936 drivers/i3c/master.c struct device_node *node, u32 *reg) node 1946 drivers/i3c/master.c ret = of_i2c_get_board_info(dev, node, &boardinfo->base); node 1963 drivers/i3c/master.c list_add_tail(&boardinfo->node, &master->boardinfo.i2c); node 1964 drivers/i3c/master.c of_node_get(node); node 1971 drivers/i3c/master.c struct device_node *node, u32 *reg) node 1994 drivers/i3c/master.c if (!of_property_read_u32(node, "assigned-address", &init_dyn_addr)) { node 2011 drivers/i3c/master.c boardinfo->of_node = of_node_get(node); node 2012 drivers/i3c/master.c list_add_tail(&boardinfo->node, &master->boardinfo.i3c); node 2018 drivers/i3c/master.c struct device_node *node) node 2023 drivers/i3c/master.c if (!master || !node) node 2026 drivers/i3c/master.c ret = of_property_read_u32_array(node, "reg", reg, ARRAY_SIZE(reg)); node 2035 drivers/i3c/master.c ret = of_i3c_master_add_i2c_boardinfo(master, node, reg); node 2037 drivers/i3c/master.c ret = of_i3c_master_add_i3c_boardinfo(master, node, reg); node 2046 drivers/i3c/master.c struct device_node *node; node 2053 drivers/i3c/master.c for_each_available_child_of_node(i3cbus_np, node) { node 2054 drivers/i3c/master.c ret = of_i3c_master_add_dev(master, node); node 2056 drivers/i3c/master.c of_node_put(node); node 2215 drivers/i3c/master.c struct list_head node; node 2241 drivers/i3c/master.c struct i3c_generic_ibi_slot, node); node 2242 drivers/i3c/master.c list_del(&slot->node); node 2307 drivers/i3c/master.c list_add_tail(&slot->node, &pool->free_slots); node 2337 drivers/i3c/master.c struct i3c_generic_ibi_slot, node); node 2339 drivers/i3c/master.c list_del(&slot->node); node 2365 drivers/i3c/master.c list_add_tail(&slot->node, &pool->free_slots); node 2445 drivers/i3c/master.c list_for_each_entry(i2cbi, &master->boardinfo.i2c, node) { node 220 drivers/i3c/master/dw-i3c-master.c struct list_head node; node 366 drivers/i3c/master/dw-i3c-master.c INIT_LIST_HEAD(&xfer->node); node 414 drivers/i3c/master/dw-i3c-master.c list_add_tail(&xfer->node, &master->xferqueue.list); node 437 drivers/i3c/master/dw-i3c-master.c list_del_init(&xfer->node); node 510 drivers/i3c/master/dw-i3c-master.c node); node 512 drivers/i3c/master/dw-i3c-master.c list_del_init(&xfer->node); node 384 drivers/i3c/master/i3c-master-cdns.c struct list_head node; node 508 drivers/i3c/master/i3c-master-cdns.c INIT_LIST_HEAD(&xfer->node); node 615 drivers/i3c/master/i3c-master-cdns.c struct cdns_i3c_xfer, node); node 617 drivers/i3c/master/i3c-master-cdns.c list_del_init(&xfer->node); node 631 drivers/i3c/master/i3c-master-cdns.c list_add_tail(&xfer->node, &master->xferqueue.list); node 661 drivers/i3c/master/i3c-master-cdns.c list_del_init(&xfer->node); node 1314 drivers/ide/ide-probe.c static int ide_port_alloc_devices(ide_hwif_t *hwif, int node) node 1320 drivers/ide/ide-probe.c drive = kzalloc_node(sizeof(*drive), GFP_KERNEL, node); node 1332 drivers/ide/ide-probe.c drive->id = kzalloc_node(SECTOR_SIZE, GFP_KERNEL, node); node 1352 drivers/ide/ide-probe.c int node = dev ? dev_to_node(dev) : -1; node 1355 drivers/ide/ide-probe.c host = kzalloc_node(sizeof(*host), GFP_KERNEL, node); node 1366 drivers/ide/ide-probe.c hwif = kzalloc_node(sizeof(*hwif), GFP_KERNEL, node); node 1370 drivers/ide/ide-probe.c if (ide_port_alloc_devices(hwif, node) < 0) { node 41 drivers/ide/ide-scan-pci.c list_add_tail(&driver->node, &ide_pci_drivers); node 62 drivers/ide/ide-scan-pci.c d = list_entry(l, struct pci_driver, node); node 106 drivers/ide/ide-scan-pci.c d = list_entry(l, struct pci_driver, node); node 57 drivers/ide/pmac.c struct device_node* node; node 874 drivers/ide/pmac.c ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, pmif->node, pmif->aapl_bus_id, node 887 drivers/ide/pmac.c ppc_md.feature_call(PMAC_FTR_IDE_RESET, pmif->node, pmif->aapl_bus_id, 1); node 888 drivers/ide/pmac.c ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, pmif->node, pmif->aapl_bus_id, 1); node 890 drivers/ide/pmac.c ppc_md.feature_call(PMAC_FTR_IDE_RESET, pmif->node, pmif->aapl_bus_id, 0); node 913 drivers/ide/pmac.c struct device_node *np = pmif->node; node 1019 drivers/ide/pmac.c struct device_node *np = pmif->node; node 1174 drivers/ide/pmac.c pmif->node = mdev->ofdev.dev.of_node; node 1287 drivers/ide/pmac.c pmif->node = np; node 834 drivers/iio/adc/at91_adc.c static int at91_adc_probe_dt_ts(struct device_node *node, node 840 drivers/iio/adc/at91_adc.c ret = of_property_read_u32(node, "atmel,adc-ts-wires", &prop); node 859 drivers/iio/adc/at91_adc.c of_property_read_u32(node, "atmel,adc-ts-pressure-threshold", &prop); node 873 drivers/iio/adc/at91_adc.c struct device_node *node = pdev->dev.of_node; node 878 drivers/iio/adc/at91_adc.c if (!node) node 884 drivers/iio/adc/at91_adc.c st->use_external = of_property_read_bool(node, "atmel,adc-use-external-triggers"); node 886 drivers/iio/adc/at91_adc.c if (of_property_read_u32(node, "atmel,adc-channels-used", &prop)) { node 893 drivers/iio/adc/at91_adc.c st->sleep_mode = of_property_read_bool(node, "atmel,adc-sleep-mode"); node 895 drivers/iio/adc/at91_adc.c if (of_property_read_u32(node, "atmel,adc-startup-time", &prop)) { node 903 drivers/iio/adc/at91_adc.c of_property_read_u32(node, "atmel,adc-sample-hold-time", &prop); node 906 drivers/iio/adc/at91_adc.c if (of_property_read_u32(node, "atmel,adc-vref", &prop)) { node 919 drivers/iio/adc/at91_adc.c st->trigger_number = of_get_child_count(node); node 930 drivers/iio/adc/at91_adc.c for_each_child_of_node(node, trig_node) { node 953 drivers/iio/adc/at91_adc.c return at91_adc_probe_dt_ts(node, st, &idev->dev); node 310 drivers/iio/adc/cc10001_adc.c struct device_node *node = pdev->dev.of_node; node 325 drivers/iio/adc/cc10001_adc.c if (!of_property_read_u32(node, "adc-reserved-channels", &ret)) { node 509 drivers/iio/adc/qcom-spmi-adc5.c struct device_node *node, node 512 drivers/iio/adc/qcom-spmi-adc5.c const char *name = node->name, *channel_name; node 517 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32(node, "reg", &chan); node 532 drivers/iio/adc/qcom-spmi-adc5.c channel_name = of_get_property(node, node 533 drivers/iio/adc/qcom-spmi-adc5.c "label", NULL) ? : node->name; node 540 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32(node, "qcom,decimation", &value); node 553 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32_array(node, "qcom,pre-scaling", varr, 2); node 567 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32(node, "qcom,hw-settle-time", &value); node 599 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32(node, "qcom,avg-samples", &value); node 612 drivers/iio/adc/qcom-spmi-adc5.c if (of_property_read_bool(node, "qcom,ratiometric")) node 669 drivers/iio/adc/qcom-spmi-adc5.c static int adc5_get_dt_data(struct adc5_chip *adc, struct device_node *node) node 680 drivers/iio/adc/qcom-spmi-adc5.c adc->nchannels = of_get_available_child_count(node); node 696 drivers/iio/adc/qcom-spmi-adc5.c id = of_match_node(adc5_match_table, node); node 703 drivers/iio/adc/qcom-spmi-adc5.c for_each_available_child_of_node(node, child) { node 731 drivers/iio/adc/qcom-spmi-adc5.c struct device_node *node = pdev->dev.of_node; node 743 drivers/iio/adc/qcom-spmi-adc5.c ret = of_property_read_u32(node, "reg", ®); node 758 drivers/iio/adc/qcom-spmi-adc5.c ret = adc5_get_dt_data(adc, node); node 777 drivers/iio/adc/qcom-spmi-adc5.c indio_dev->dev.of_node = node; node 430 drivers/iio/adc/qcom-spmi-iadc.c static int iadc_rsense_read(struct iadc_chip *iadc, struct device_node *node) node 435 drivers/iio/adc/qcom-spmi-iadc.c ret = of_property_read_u32(node, "qcom,external-resistor-micro-ohms", node 487 drivers/iio/adc/qcom-spmi-iadc.c struct device_node *node = pdev->dev.of_node; node 508 drivers/iio/adc/qcom-spmi-iadc.c ret = of_property_read_u32(node, "reg", &res); node 518 drivers/iio/adc/qcom-spmi-iadc.c ret = iadc_rsense_read(iadc, node); node 557 drivers/iio/adc/qcom-spmi-iadc.c indio_dev->dev.of_node = node; node 653 drivers/iio/adc/qcom-spmi-vadc.c struct device_node *node) node 655 drivers/iio/adc/qcom-spmi-vadc.c const char *name = node->name; node 659 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32(node, "reg", &chan); node 673 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32(node, "qcom,decimation", &value); node 686 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32_array(node, "qcom,pre-scaling", varr, 2); node 699 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32(node, "qcom,hw-settle-time", &value); node 712 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32(node, "qcom,avg-samples", &value); node 725 drivers/iio/adc/qcom-spmi-vadc.c if (of_property_read_bool(node, "qcom,ratiometric")) node 735 drivers/iio/adc/qcom-spmi-vadc.c static int vadc_get_dt_data(struct vadc_priv *vadc, struct device_node *node) node 744 drivers/iio/adc/qcom-spmi-vadc.c vadc->nchannels = of_get_available_child_count(node); node 760 drivers/iio/adc/qcom-spmi-vadc.c for_each_available_child_of_node(node, child) { node 852 drivers/iio/adc/qcom-spmi-vadc.c struct device_node *node = pdev->dev.of_node; node 864 drivers/iio/adc/qcom-spmi-vadc.c ret = of_property_read_u32(node, "reg", ®); node 884 drivers/iio/adc/qcom-spmi-vadc.c ret = vadc_get_dt_data(vadc, node); node 911 drivers/iio/adc/qcom-spmi-vadc.c indio_dev->dev.of_node = node; node 1579 drivers/iio/adc/stm32-adc.c struct device_node *node = indio_dev->dev.of_node; node 1584 drivers/iio/adc/stm32-adc.c if (of_property_read_u32(node, "assigned-resolution-bits", &res)) node 1658 drivers/iio/adc/stm32-adc.c struct device_node *node = indio_dev->dev.of_node; node 1668 drivers/iio/adc/stm32-adc.c ret = of_property_count_u32_elems(node, "st,adc-channels"); node 1676 drivers/iio/adc/stm32-adc.c ret = of_property_count_elems_of_size(node, "st,adc-diff-channels", node 1686 drivers/iio/adc/stm32-adc.c ret = of_property_read_u32_array(node, "st,adc-diff-channels", node 1698 drivers/iio/adc/stm32-adc.c ret = of_property_count_u32_elems(node, "st,min-sample-time-nsecs"); node 1709 drivers/iio/adc/stm32-adc.c of_property_for_each_u32(node, "st,adc-channels", prop, cur, val) { node 1748 drivers/iio/adc/stm32-adc.c of_property_read_u32_index(node, "st,min-sample-time-nsecs", node 220 drivers/iio/adc/stm32-dfsdm-core.c struct device_node *node = pdev->dev.of_node; node 226 drivers/iio/adc/stm32-dfsdm-core.c if (!node) node 845 drivers/iio/adc/ti-ads1015.c struct device_node *node; node 851 drivers/iio/adc/ti-ads1015.c for_each_child_of_node(client->dev.of_node, node) { node 857 drivers/iio/adc/ti-ads1015.c if (of_property_read_u32(node, "reg", &pval)) { node 859 drivers/iio/adc/ti-ads1015.c node); node 867 drivers/iio/adc/ti-ads1015.c channel, node); node 871 drivers/iio/adc/ti-ads1015.c if (!of_property_read_u32(node, "ti,gain", &pval)) { node 875 drivers/iio/adc/ti-ads1015.c node); node 876 drivers/iio/adc/ti-ads1015.c of_node_put(node); node 881 drivers/iio/adc/ti-ads1015.c if (!of_property_read_u32(node, "ti,datarate", &pval)) { node 886 drivers/iio/adc/ti-ads1015.c node); node 887 drivers/iio/adc/ti-ads1015.c of_node_put(node); node 579 drivers/iio/adc/ti_am335x_adc.c struct device_node *node = pdev->dev.of_node; node 585 drivers/iio/adc/ti_am335x_adc.c of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) { node 596 drivers/iio/adc/ti_am335x_adc.c of_property_read_u32_array(node, "ti,chan-step-avg", node 598 drivers/iio/adc/ti_am335x_adc.c of_property_read_u32_array(node, "ti,chan-step-opendelay", node 600 drivers/iio/adc/ti_am335x_adc.c of_property_read_u32_array(node, "ti,chan-step-sampledelay", node 611 drivers/iio/adc/ti_am335x_adc.c struct device_node *node = pdev->dev.of_node; node 614 drivers/iio/adc/ti_am335x_adc.c if (!node) { node 446 drivers/iio/common/ssp_sensors/ssp_dev.c struct device_node *node = dev->of_node; node 453 drivers/iio/common/ssp_sensors/ssp_dev.c data->mcu_ap_gpio = of_get_named_gpio(node, "mcu-ap-gpios", 0); node 457 drivers/iio/common/ssp_sensors/ssp_dev.c data->ap_mcu_gpio = of_get_named_gpio(node, "ap-mcu-gpios", 0); node 461 drivers/iio/common/ssp_sensors/ssp_dev.c data->mcu_reset_gpio = of_get_named_gpio(node, "mcu-reset-gpios", 0); node 475 drivers/iio/common/ssp_sensors/ssp_dev.c match = of_match_node(ssp_of_match, node); node 697 drivers/infiniband/core/cm.c struct rb_node *node = cm.listen_service_table.rb_node; node 700 drivers/infiniband/core/cm.c while (node) { node 701 drivers/infiniband/core/cm.c cm_id_priv = rb_entry(node, struct cm_id_private, service_node); node 708 drivers/infiniband/core/cm.c node = node->rb_left; node 710 drivers/infiniband/core/cm.c node = node->rb_right; node 712 drivers/infiniband/core/cm.c node = node->rb_left; node 714 drivers/infiniband/core/cm.c node = node->rb_right; node 716 drivers/infiniband/core/cm.c node = node->rb_right; node 754 drivers/infiniband/core/cm.c struct rb_node *node = cm.remote_id_table.rb_node; node 757 drivers/infiniband/core/cm.c while (node) { node 758 drivers/infiniband/core/cm.c timewait_info = rb_entry(node, struct cm_timewait_info, node 761 drivers/infiniband/core/cm.c node = node->rb_left; node 763 drivers/infiniband/core/cm.c node = node->rb_right; node 765 drivers/infiniband/core/cm.c node = node->rb_left; node 767 drivers/infiniband/core/cm.c node = node->rb_right; node 1637 drivers/infiniband/core/cma.c hlist_for_each_entry(id_priv, &bind_list->owners, node) { node 1794 drivers/infiniband/core/cma.c hlist_del(&id_priv->node); node 3267 drivers/infiniband/core/cma.c hlist_add_head(&id_priv->node, &bind_list->owners); node 3302 drivers/infiniband/core/cma.c hlist_for_each_entry(cur_id, &bind_list->owners, node) { node 3389 drivers/infiniband/core/cma.c hlist_for_each_entry(cur_id, &bind_list->owners, node) { node 57 drivers/infiniband/core/cma_priv.h struct hlist_node node; node 98 drivers/infiniband/core/multicast.c struct rb_node node; node 132 drivers/infiniband/core/multicast.c struct rb_node *node = port->table.rb_node; node 136 drivers/infiniband/core/multicast.c while (node) { node 137 drivers/infiniband/core/multicast.c group = rb_entry(node, struct mcast_group, node); node 143 drivers/infiniband/core/multicast.c node = node->rb_left; node 145 drivers/infiniband/core/multicast.c node = node->rb_right; node 161 drivers/infiniband/core/multicast.c cur_group = rb_entry(parent, struct mcast_group, node); node 174 drivers/infiniband/core/multicast.c rb_link_node(&group->node, parent, link); node 175 drivers/infiniband/core/multicast.c rb_insert_color(&group->node, &port->table); node 192 drivers/infiniband/core/multicast.c rb_erase(&group->node, &port->table); node 535 drivers/infiniband/core/multicast.c rb_erase(&group->node, &group->port->table); node 774 drivers/infiniband/core/multicast.c struct rb_node *node; node 778 drivers/infiniband/core/multicast.c for (node = rb_first(&port->table); node; node = rb_next(node)) { node 779 drivers/infiniband/core/multicast.c group = rb_entry(node, struct mcast_group, node); node 82 drivers/infiniband/core/umem_odp.c struct rb_node *node; node 88 drivers/infiniband/core/umem_odp.c for (node = rb_first_cached(&per_mm->umem_tree); node; node 89 drivers/infiniband/core/umem_odp.c node = rb_next(node)) { node 91 drivers/infiniband/core/umem_odp.c rb_entry(node, struct ib_umem_odp, interval_tree.rb); node 776 drivers/infiniband/core/umem_odp.c struct interval_tree_node *node, *next; node 782 drivers/infiniband/core/umem_odp.c for (node = interval_tree_iter_first(root, start, last - 1); node 783 drivers/infiniband/core/umem_odp.c node; node = next) { node 787 drivers/infiniband/core/umem_odp.c next = interval_tree_iter_next(node, start, last - 1); node 788 drivers/infiniband/core/umem_odp.c umem = container_of(node, struct ib_umem_odp, interval_tree); node 472 drivers/infiniband/core/uverbs_cmd.c struct rb_node node; node 494 drivers/infiniband/core/uverbs_cmd.c scan = rb_entry(parent, struct xrcd_table_entry, node); node 506 drivers/infiniband/core/uverbs_cmd.c rb_link_node(&entry->node, parent, p); node 507 drivers/infiniband/core/uverbs_cmd.c rb_insert_color(&entry->node, &dev->xrcd_tree); node 519 drivers/infiniband/core/uverbs_cmd.c entry = rb_entry(p, struct xrcd_table_entry, node); node 551 drivers/infiniband/core/uverbs_cmd.c rb_erase(&entry->node, &dev->xrcd_tree); node 169 drivers/infiniband/hw/hfi1/affinity.c int node; node 200 drivers/infiniband/hw/hfi1/affinity.c node = pcibus_to_node(dev->bus); node 201 drivers/infiniband/hw/hfi1/affinity.c if (node < 0) node 204 drivers/infiniband/hw/hfi1/affinity.c hfi1_per_node_cntr[node]++; node 218 drivers/infiniband/hw/hfi1/affinity.c for (node = 0; node < node_affinity.num_possible_nodes; node++) node 219 drivers/infiniband/hw/hfi1/affinity.c hfi1_per_node_cntr[node] = 1; node 246 drivers/infiniband/hw/hfi1/affinity.c static struct hfi1_affinity_node *node_affinity_allocate(int node) node 253 drivers/infiniband/hw/hfi1/affinity.c entry->node = node; node 270 drivers/infiniband/hw/hfi1/affinity.c static struct hfi1_affinity_node *node_affinity_lookup(int node) node 277 drivers/infiniband/hw/hfi1/affinity.c if (entry->node == node) node 500 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 557 drivers/infiniband/hw/hfi1/affinity.c hfi1_per_node_cntr[dd->node]; node 566 drivers/infiniband/hw/hfi1/affinity.c hfi1_per_node_cntr[dd->node] != 0) node 634 drivers/infiniband/hw/hfi1/affinity.c int node = pcibus_to_node(dd->pcidev->bus); node 644 drivers/infiniband/hw/hfi1/affinity.c if (node < 0) { node 646 drivers/infiniband/hw/hfi1/affinity.c node = 0; node 648 drivers/infiniband/hw/hfi1/affinity.c dd->node = node; node 650 drivers/infiniband/hw/hfi1/affinity.c local_mask = cpumask_of_node(dd->node); node 655 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 662 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_allocate(node); node 704 drivers/infiniband/hw/hfi1/affinity.c hfi1_per_node_cntr[dd->node]; node 768 drivers/infiniband/hw/hfi1/affinity.c if (dd->node < 0) node 772 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 783 drivers/infiniband/hw/hfi1/affinity.c dd->node = NUMA_NO_NODE; node 803 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 899 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 974 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); node 1037 drivers/infiniband/hw/hfi1/affinity.c int hfi1_get_proc_affinity(int node) node 1114 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(node); node 1153 drivers/infiniband/hw/hfi1/affinity.c node_mask = cpumask_of_node(node); node 1154 drivers/infiniband/hw/hfi1/affinity.c hfi1_cdbg(PROC, "Device on NUMA %u, CPUs %*pbl", node, node 1160 drivers/infiniband/hw/hfi1/affinity.c hfi1_cdbg(PROC, "Available CPUs on NUMA %u: %*pbl", node, node 95 drivers/infiniband/hw/hfi1/affinity.h int hfi1_get_proc_affinity(int node); node 100 drivers/infiniband/hw/hfi1/affinity.h int node; node 968 drivers/infiniband/hw/hfi1/file_ops.c fd->rec_cpu_num = hfi1_get_proc_affinity(dd->node); node 985 drivers/infiniband/hw/hfi1/file_ops.c uctxt->sc = sc_alloc(dd, SC_USER, uctxt->rcvhdrqentsize, dd->node); node 1213 drivers/infiniband/hw/hfi1/hfi.h int node; /* home node of this chip */ node 138 drivers/infiniband/hw/hfi1/init.c ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); node 159 drivers/infiniband/hw/hfi1/init.c rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); node 178 drivers/infiniband/hw/hfi1/init.c GFP_KERNEL, dd->node); node 1293 drivers/infiniband/hw/hfi1/init.c dd->node = NUMA_NO_NODE; node 127 drivers/infiniband/hw/hfi1/mad.c struct trap_node *node, *q; node 142 drivers/infiniband/hw/hfi1/mad.c list_for_each_entry_safe(node, q, &trap_list, list) { node 143 drivers/infiniband/hw/hfi1/mad.c list_del(&node->list); node 144 drivers/infiniband/hw/hfi1/mad.c if (node != trap) node 145 drivers/infiniband/hw/hfi1/mad.c kfree(node); node 159 drivers/infiniband/hw/hfi1/mad.c struct trap_node *node; node 183 drivers/infiniband/hw/hfi1/mad.c list_for_each_entry(node, &trap_list->list, list) { node 184 drivers/infiniband/hw/hfi1/mad.c if (node == trap) { node 185 drivers/infiniband/hw/hfi1/mad.c node->retry++; node 207 drivers/infiniband/hw/hfi1/mad.c node = NULL; node 219 drivers/infiniband/hw/hfi1/mad.c node = list_first_entry(&trap_list->list, struct trap_node, node 221 drivers/infiniband/hw/hfi1/mad.c node->in_use = 1; node 225 drivers/infiniband/hw/hfi1/mad.c return node; node 82 drivers/infiniband/hw/hfi1/mmu_rb.c INTERVAL_TREE_DEFINE(struct mmu_rb_node, node, unsigned long, __last, node 85 drivers/infiniband/hw/hfi1/mmu_rb.c static unsigned long mmu_node_start(struct mmu_rb_node *node) node 87 drivers/infiniband/hw/hfi1/mmu_rb.c return node->addr & PAGE_MASK; node 90 drivers/infiniband/hw/hfi1/mmu_rb.c static unsigned long mmu_node_last(struct mmu_rb_node *node) node 92 drivers/infiniband/hw/hfi1/mmu_rb.c return PAGE_ALIGN(node->addr + node->len) - 1; node 132 drivers/infiniband/hw/hfi1/mmu_rb.c struct rb_node *node; node 148 drivers/infiniband/hw/hfi1/mmu_rb.c while ((node = rb_first_cached(&handler->root))) { node 149 drivers/infiniband/hw/hfi1/mmu_rb.c rbnode = rb_entry(node, struct mmu_rb_node, node); node 150 drivers/infiniband/hw/hfi1/mmu_rb.c rb_erase_cached(node, &handler->root); node 164 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node; node 170 drivers/infiniband/hw/hfi1/mmu_rb.c node = __mmu_rb_search(handler, mnode->addr, mnode->len); node 171 drivers/infiniband/hw/hfi1/mmu_rb.c if (node) { node 193 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node = NULL; node 197 drivers/infiniband/hw/hfi1/mmu_rb.c node = __mmu_int_rb_iter_first(&handler->root, addr, node 200 drivers/infiniband/hw/hfi1/mmu_rb.c for (node = __mmu_int_rb_iter_first(&handler->root, addr, node 202 drivers/infiniband/hw/hfi1/mmu_rb.c node; node 203 drivers/infiniband/hw/hfi1/mmu_rb.c node = __mmu_int_rb_iter_next(node, addr, node 205 drivers/infiniband/hw/hfi1/mmu_rb.c if (handler->ops->filter(node, addr, len)) node 206 drivers/infiniband/hw/hfi1/mmu_rb.c return node; node 209 drivers/infiniband/hw/hfi1/mmu_rb.c return node; node 216 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node; node 221 drivers/infiniband/hw/hfi1/mmu_rb.c node = __mmu_rb_search(handler, addr, len); node 222 drivers/infiniband/hw/hfi1/mmu_rb.c if (node) { node 223 drivers/infiniband/hw/hfi1/mmu_rb.c if (node->addr == addr && node->len == len) node 225 drivers/infiniband/hw/hfi1/mmu_rb.c __mmu_int_rb_remove(node, &handler->root); node 226 drivers/infiniband/hw/hfi1/mmu_rb.c list_del(&node->list); /* remove from LRU list */ node 231 drivers/infiniband/hw/hfi1/mmu_rb.c *rb_node = node; node 271 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node) node 276 drivers/infiniband/hw/hfi1/mmu_rb.c trace_hfi1_mmu_rb_remove(node->addr, node->len); node 278 drivers/infiniband/hw/hfi1/mmu_rb.c __mmu_int_rb_remove(node, &handler->root); node 279 drivers/infiniband/hw/hfi1/mmu_rb.c list_del(&node->list); /* remove from LRU list */ node 282 drivers/infiniband/hw/hfi1/mmu_rb.c handler->ops->remove(handler->ops_arg, node); node 291 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node, *ptr = NULL; node 296 drivers/infiniband/hw/hfi1/mmu_rb.c for (node = __mmu_int_rb_iter_first(root, range->start, range->end-1); node 297 drivers/infiniband/hw/hfi1/mmu_rb.c node; node = ptr) { node 299 drivers/infiniband/hw/hfi1/mmu_rb.c ptr = __mmu_int_rb_iter_next(node, range->start, node 301 drivers/infiniband/hw/hfi1/mmu_rb.c trace_hfi1_mmu_mem_invalidate(node->addr, node->len); node 302 drivers/infiniband/hw/hfi1/mmu_rb.c if (handler->ops->invalidate(handler->ops_arg, node)) { node 303 drivers/infiniband/hw/hfi1/mmu_rb.c __mmu_int_rb_remove(node, root); node 305 drivers/infiniband/hw/hfi1/mmu_rb.c list_move(&node->list, &handler->del_list); node 325 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *node; node 328 drivers/infiniband/hw/hfi1/mmu_rb.c node = list_first_entry(del_list, struct mmu_rb_node, list); node 329 drivers/infiniband/hw/hfi1/mmu_rb.c list_del(&node->list); node 330 drivers/infiniband/hw/hfi1/mmu_rb.c handler->ops->remove(handler->ops_arg, node); node 56 drivers/infiniband/hw/hfi1/mmu_rb.h struct rb_node node; node 65 drivers/infiniband/hw/hfi1/mmu_rb.h bool (*filter)(struct mmu_rb_node *node, unsigned long addr, node 69 drivers/infiniband/hw/hfi1/mmu_rb.h int (*invalidate)(void *ops_arg, struct mmu_rb_node *node); node 566 drivers/infiniband/hw/hfi1/pio.c sc->hw_free = &sc->dd->cr_base[sc->node].va[gc].cr[index]; node 568 drivers/infiniband/hw/hfi1/pio.c &((struct credit_return *)sc->dd->cr_base[sc->node].dma)[gc]; node 740 drivers/infiniband/hw/hfi1/pio.c sc->node = numa; node 2023 drivers/infiniband/hw/hfi1/pio.c dd->rcd[0]->rcvhdrqentsize, dd->node); node 2032 drivers/infiniband/hw/hfi1/pio.c GFP_KERNEL, dd->node); node 2047 drivers/infiniband/hw/hfi1/pio.c dd->rcd[0]->rcvhdrqentsize, dd->node); node 2057 drivers/infiniband/hw/hfi1/pio.c sc_alloc(dd, SC_KERNEL, dd->rcd[0]->rcvhdrqentsize, dd->node); node 2127 drivers/infiniband/hw/hfi1/pio.c set_dev_node(&dd->pcidev->dev, dd->node); node 2135 drivers/infiniband/hw/hfi1/pio.c set_dev_node(&dd->pcidev->dev, dd->node); node 106 drivers/infiniband/hw/hfi1/pio.h int node; /* context home node */ node 389 drivers/infiniband/hw/hfi1/qp.c cpumask_first(cpumask_of_node(dd->node))); node 724 drivers/infiniband/hw/hfi1/qp.c priv = kzalloc_node(sizeof(*priv), GFP_KERNEL, rdi->dparms.node); node 731 drivers/infiniband/hw/hfi1/qp.c rdi->dparms.node); node 589 drivers/infiniband/hw/hfi1/ruc.c cpumask_first(cpumask_of_node(ps.ppd->dd->node)); node 842 drivers/infiniband/hw/hfi1/sdma.c struct rhash_head node; node 849 drivers/infiniband/hw/hfi1/sdma.c .head_offset = offsetof(struct sdma_rht_node, node), node 998 drivers/infiniband/hw/hfi1/sdma.c &rht_node->node, node 1065 drivers/infiniband/hw/hfi1/sdma.c &rht_node->node, node 1400 drivers/infiniband/hw/hfi1/sdma.c GFP_KERNEL, dd->node); node 1478 drivers/infiniband/hw/hfi1/sdma.c GFP_KERNEL, dd->node); node 372 drivers/infiniband/hw/hfi1/tid_rdma.c GFP_KERNEL, dd->node); node 380 drivers/infiniband/hw/hfi1/tid_rdma.c dd->node); node 392 drivers/infiniband/hw/hfi1/tid_rdma.c dd->node); node 624 drivers/infiniband/hw/hfi1/tid_rdma.c cpumask_first(cpumask_of_node(dd->node)), node 1201 drivers/infiniband/hw/hfi1/tid_rdma.c struct kern_tid_node *node = &flow->tnode[flow->tnode_cnt++]; node 1210 drivers/infiniband/hw/hfi1/tid_rdma.c node->grp = grp; node 1211 drivers/infiniband/hw/hfi1/tid_rdma.c node->map = grp->map; node 1212 drivers/infiniband/hw/hfi1/tid_rdma.c node->cnt = cnt; node 1298 drivers/infiniband/hw/hfi1/tid_rdma.c struct kern_tid_node *node = &flow->tnode[grp_num]; node 1299 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_group *grp = node->grp; node 1308 drivers/infiniband/hw/hfi1/tid_rdma.c if (node->map & BIT(i) || cnt >= node->cnt) { node 1330 drivers/infiniband/hw/hfi1/tid_rdma.c pair = !(i & 0x1) && !((node->map >> i) & 0x3) && node 1331 drivers/infiniband/hw/hfi1/tid_rdma.c node->cnt >= cnt + 2; node 1365 drivers/infiniband/hw/hfi1/tid_rdma.c struct kern_tid_node *node = &flow->tnode[grp_num]; node 1366 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_group *grp = node->grp; node 1373 drivers/infiniband/hw/hfi1/tid_rdma.c if (node->map & BIT(i) || cnt >= node->cnt) { node 5363 drivers/infiniband/hw/hfi1/tid_rdma.c cpumask_first(cpumask_of_node(ps.ppd->dd->node)); node 5414 drivers/infiniband/hw/hfi1/tid_rdma.c cpumask_first(cpumask_of_node(dd->node))); node 62 drivers/infiniband/hw/hfi1/user_exp_rcv.c static int tid_rb_insert(void *arg, struct mmu_rb_node *node); node 65 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void tid_rb_remove(void *arg, struct mmu_rb_node *node); node 74 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void clear_tid_node(struct hfi1_filedata *fd, struct tid_rb_node *node); node 193 drivers/infiniband/hw/hfi1/user_exp_rcv.c struct tid_rb_node *node, node 202 drivers/infiniband/hw/hfi1/user_exp_rcv.c pci_unmap_single(dd->pcidev, node->dma_addr, node 203 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->mmu.len, PCI_DMA_FROMDEVICE); node 204 drivers/infiniband/hw/hfi1/user_exp_rcv.c pages = &node->pages[idx]; node 755 drivers/infiniband/hw/hfi1/user_exp_rcv.c struct tid_rb_node *node; node 764 drivers/infiniband/hw/hfi1/user_exp_rcv.c node = kzalloc(sizeof(*node) + (sizeof(struct page *) * npages), node 766 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (!node) node 775 drivers/infiniband/hw/hfi1/user_exp_rcv.c kfree(node); node 779 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->mmu.addr = tbuf->vaddr + (pageidx * PAGE_SIZE); node 780 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->mmu.len = npages * PAGE_SIZE; node 781 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->phys = page_to_phys(pages[0]); node 782 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->npages = npages; node 783 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->rcventry = rcventry; node 784 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->dma_addr = phys; node 785 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->grp = grp; node 786 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->freed = false; node 787 drivers/infiniband/hw/hfi1/user_exp_rcv.c memcpy(node->pages, pages, sizeof(struct page *) * npages); node 790 drivers/infiniband/hw/hfi1/user_exp_rcv.c ret = tid_rb_insert(fd, &node->mmu); node 792 drivers/infiniband/hw/hfi1/user_exp_rcv.c ret = hfi1_mmu_rb_insert(fd->handler, &node->mmu); node 796 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->rcventry, node->mmu.addr, node->phys, ret); node 799 drivers/infiniband/hw/hfi1/user_exp_rcv.c kfree(node); node 804 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->mmu.addr, node->phys, phys); node 813 drivers/infiniband/hw/hfi1/user_exp_rcv.c struct tid_rb_node *node; node 828 drivers/infiniband/hw/hfi1/user_exp_rcv.c node = fd->entry_to_rb[rcventry]; node 829 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (!node || node->rcventry != (uctxt->expected_base + rcventry)) node 833 drivers/infiniband/hw/hfi1/user_exp_rcv.c *grp = node->grp; node 836 drivers/infiniband/hw/hfi1/user_exp_rcv.c cacheless_tid_rb_remove(fd, node); node 838 drivers/infiniband/hw/hfi1/user_exp_rcv.c hfi1_mmu_rb_remove(fd->handler, &node->mmu); node 843 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void clear_tid_node(struct hfi1_filedata *fd, struct tid_rb_node *node) node 848 drivers/infiniband/hw/hfi1/user_exp_rcv.c trace_hfi1_exp_tid_unreg(uctxt->ctxt, fd->subctxt, node->rcventry, node 849 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->npages, node->mmu.addr, node->phys, node 850 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->dma_addr); node 856 drivers/infiniband/hw/hfi1/user_exp_rcv.c hfi1_put_tid(dd, node->rcventry, PT_INVALID_FLUSH, 0, 0); node 858 drivers/infiniband/hw/hfi1/user_exp_rcv.c unpin_rcv_pages(fd, NULL, node, 0, node->npages, true); node 860 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->grp->used--; node 861 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->grp->map &= ~(1 << (node->rcventry - node->grp->base)); node 863 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (node->grp->used == node->grp->size - 1) node 864 drivers/infiniband/hw/hfi1/user_exp_rcv.c tid_group_move(node->grp, &uctxt->tid_full_list, node 866 drivers/infiniband/hw/hfi1/user_exp_rcv.c else if (!node->grp->used) node 867 drivers/infiniband/hw/hfi1/user_exp_rcv.c tid_group_move(node->grp, &uctxt->tid_used_list, node 869 drivers/infiniband/hw/hfi1/user_exp_rcv.c kfree(node); node 889 drivers/infiniband/hw/hfi1/user_exp_rcv.c struct tid_rb_node *node; node 891 drivers/infiniband/hw/hfi1/user_exp_rcv.c node = fd->entry_to_rb[rcventry - node 893 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (!node || node->rcventry != rcventry) node 896 drivers/infiniband/hw/hfi1/user_exp_rcv.c cacheless_tid_rb_remove(fd, node); node 913 drivers/infiniband/hw/hfi1/user_exp_rcv.c struct tid_rb_node *node = node 916 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (node->freed) node 919 drivers/infiniband/hw/hfi1/user_exp_rcv.c trace_hfi1_exp_tid_inval(uctxt->ctxt, fdata->subctxt, node->mmu.addr, node 920 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->rcventry, node->npages, node->dma_addr); node 921 drivers/infiniband/hw/hfi1/user_exp_rcv.c node->freed = true; node 926 drivers/infiniband/hw/hfi1/user_exp_rcv.c rcventry2tidinfo(node->rcventry - uctxt->expected_base); node 928 drivers/infiniband/hw/hfi1/user_exp_rcv.c EXP_TID_SET(LEN, node->npages); node 951 drivers/infiniband/hw/hfi1/user_exp_rcv.c static int tid_rb_insert(void *arg, struct mmu_rb_node *node) node 955 drivers/infiniband/hw/hfi1/user_exp_rcv.c container_of(node, struct tid_rb_node, mmu); node 971 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void tid_rb_remove(void *arg, struct mmu_rb_node *node) node 975 drivers/infiniband/hw/hfi1/user_exp_rcv.c container_of(node, struct tid_rb_node, mmu); node 108 drivers/infiniband/hw/hfi1/user_sdma.c static bool sdma_rb_filter(struct mmu_rb_node *node, unsigned long addr, node 969 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node, node 979 drivers/infiniband/hw/hfi1/user_sdma.c memcpy(pages, node->pages, node->npages * sizeof(*pages)); node 981 drivers/infiniband/hw/hfi1/user_sdma.c npages -= node->npages; node 991 drivers/infiniband/hw/hfi1/user_sdma.c (node->npages * PAGE_SIZE)), npages, 0, node 992 drivers/infiniband/hw/hfi1/user_sdma.c pages + node->npages); node 998 drivers/infiniband/hw/hfi1/user_sdma.c unpin_vector_pages(pq->mm, pages, node->npages, pinned); node 1001 drivers/infiniband/hw/hfi1/user_sdma.c kfree(node->pages); node 1002 drivers/infiniband/hw/hfi1/user_sdma.c node->rb.len = iovec->iov.iov_len; node 1003 drivers/infiniband/hw/hfi1/user_sdma.c node->pages = pages; node 1008 drivers/infiniband/hw/hfi1/user_sdma.c static void unpin_sdma_pages(struct sdma_mmu_node *node) node 1010 drivers/infiniband/hw/hfi1/user_sdma.c if (node->npages) { node 1011 drivers/infiniband/hw/hfi1/user_sdma.c unpin_vector_pages(node->pq->mm, node->pages, 0, node->npages); node 1012 drivers/infiniband/hw/hfi1/user_sdma.c atomic_sub(node->npages, &node->pq->n_locked); node 1021 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = NULL; node 1032 drivers/infiniband/hw/hfi1/user_sdma.c node = container_of(rb_node, struct sdma_mmu_node, rb); node 1034 drivers/infiniband/hw/hfi1/user_sdma.c atomic_inc(&node->refcount); node 1035 drivers/infiniband/hw/hfi1/user_sdma.c iovec->pages = node->pages; node 1036 drivers/infiniband/hw/hfi1/user_sdma.c iovec->npages = node->npages; node 1037 drivers/infiniband/hw/hfi1/user_sdma.c iovec->node = node; node 1042 drivers/infiniband/hw/hfi1/user_sdma.c if (!node) { node 1043 drivers/infiniband/hw/hfi1/user_sdma.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 1044 drivers/infiniband/hw/hfi1/user_sdma.c if (!node) node 1047 drivers/infiniband/hw/hfi1/user_sdma.c node->rb.addr = (unsigned long)iovec->iov.iov_base; node 1048 drivers/infiniband/hw/hfi1/user_sdma.c node->pq = pq; node 1049 drivers/infiniband/hw/hfi1/user_sdma.c atomic_set(&node->refcount, 0); node 1054 drivers/infiniband/hw/hfi1/user_sdma.c if (node->npages < npages) { node 1055 drivers/infiniband/hw/hfi1/user_sdma.c pinned = pin_sdma_pages(req, iovec, node, npages); node 1060 drivers/infiniband/hw/hfi1/user_sdma.c node->npages += pinned; node 1061 drivers/infiniband/hw/hfi1/user_sdma.c npages = node->npages; node 1063 drivers/infiniband/hw/hfi1/user_sdma.c iovec->pages = node->pages; node 1065 drivers/infiniband/hw/hfi1/user_sdma.c iovec->node = node; node 1067 drivers/infiniband/hw/hfi1/user_sdma.c ret = hfi1_mmu_rb_insert(req->pq->handler, &node->rb); node 1069 drivers/infiniband/hw/hfi1/user_sdma.c iovec->node = NULL; node 1074 drivers/infiniband/hw/hfi1/user_sdma.c unpin_sdma_pages(node); node 1075 drivers/infiniband/hw/hfi1/user_sdma.c kfree(node); node 1455 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = req->iovs[i].node; node 1457 drivers/infiniband/hw/hfi1/user_sdma.c if (!node) node 1460 drivers/infiniband/hw/hfi1/user_sdma.c req->iovs[i].node = NULL; node 1464 drivers/infiniband/hw/hfi1/user_sdma.c &node->rb); node 1466 drivers/infiniband/hw/hfi1/user_sdma.c atomic_dec(&node->refcount); node 1486 drivers/infiniband/hw/hfi1/user_sdma.c static bool sdma_rb_filter(struct mmu_rb_node *node, unsigned long addr, node 1489 drivers/infiniband/hw/hfi1/user_sdma.c return (bool)(node->addr == addr); node 1494 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = node 1497 drivers/infiniband/hw/hfi1/user_sdma.c atomic_inc(&node->refcount); node 1509 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = node 1514 drivers/infiniband/hw/hfi1/user_sdma.c if (atomic_read(&node->refcount)) node 1518 drivers/infiniband/hw/hfi1/user_sdma.c evict_data->cleared += node->npages; node 1529 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = node 1532 drivers/infiniband/hw/hfi1/user_sdma.c unpin_sdma_pages(node); node 1533 drivers/infiniband/hw/hfi1/user_sdma.c kfree(node); node 1538 drivers/infiniband/hw/hfi1/user_sdma.c struct sdma_mmu_node *node = node 1541 drivers/infiniband/hw/hfi1/user_sdma.c if (!atomic_read(&node->refcount)) node 164 drivers/infiniband/hw/hfi1/user_sdma.h struct sdma_mmu_node *node; node 1902 drivers/infiniband/hw/hfi1/verbs.c dd->verbs_dev.rdi.dparms.node = dd->node; node 110 drivers/infiniband/hw/hfi1/vnic_main.c ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt); node 45 drivers/infiniband/hw/mlx4/cm.c struct rb_node node; node 147 drivers/infiniband/hw/mlx4/cm.c struct rb_node *node = sl_id_map->rb_node; node 149 drivers/infiniband/hw/mlx4/cm.c while (node) { node 151 drivers/infiniband/hw/mlx4/cm.c rb_entry(node, struct id_map_entry, node); node 154 drivers/infiniband/hw/mlx4/cm.c node = node->rb_left; node 156 drivers/infiniband/hw/mlx4/cm.c node = node->rb_right; node 158 drivers/infiniband/hw/mlx4/cm.c node = node->rb_left; node 160 drivers/infiniband/hw/mlx4/cm.c node = node->rb_right; node 181 drivers/infiniband/hw/mlx4/cm.c rb_erase(&found_ent->node, sl_id_map); node 202 drivers/infiniband/hw/mlx4/cm.c rb_replace_node(&ent->node, &new->node, sl_id_map); node 209 drivers/infiniband/hw/mlx4/cm.c ent = rb_entry(parent, struct id_map_entry, node); node 217 drivers/infiniband/hw/mlx4/cm.c rb_link_node(&new->node, parent, link); node 218 drivers/infiniband/hw/mlx4/cm.c rb_insert_color(&new->node, sl_id_map); node 411 drivers/infiniband/hw/mlx4/cm.c struct id_map_entry, node); node 413 drivers/infiniband/hw/mlx4/cm.c rb_erase(&ent->node, sl_id_map); node 422 drivers/infiniband/hw/mlx4/cm.c rb_entry(nd, struct id_map_entry, node); node 429 drivers/infiniband/hw/mlx4/cm.c rb_erase(&map->node, sl_id_map); node 103 drivers/infiniband/hw/mlx4/mcg.c struct rb_node node; node 165 drivers/infiniband/hw/mlx4/mcg.c struct rb_node *node = ctx->mcg_table.rb_node; node 169 drivers/infiniband/hw/mlx4/mcg.c while (node) { node 170 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(node, struct mcast_group, node); node 176 drivers/infiniband/hw/mlx4/mcg.c node = node->rb_left; node 178 drivers/infiniband/hw/mlx4/mcg.c node = node->rb_right; node 193 drivers/infiniband/hw/mlx4/mcg.c cur_group = rb_entry(parent, struct mcast_group, node); node 204 drivers/infiniband/hw/mlx4/mcg.c rb_link_node(&group->node, parent, link); node 205 drivers/infiniband/hw/mlx4/mcg.c rb_insert_color(&group->node, &ctx->mcg_table); node 461 drivers/infiniband/hw/mlx4/mcg.c rb_erase(&group->node, &ctx->mcg_table); node 1069 drivers/infiniband/hw/mlx4/mcg.c rb_erase(&group->node, &group->demux->mcg_table); node 1103 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(p, struct mcast_group, node); node 1232 drivers/infiniband/hw/mlx4/mcg.c group = rb_entry(p, struct mcast_group, node); node 257 drivers/infiniband/hw/mlx5/cong.c enum mlx5_ib_cong_node_type node; node 271 drivers/infiniband/hw/mlx5/cong.c node = mlx5_ib_param_to_node(offset); node 273 drivers/infiniband/hw/mlx5/cong.c err = mlx5_cmd_query_cong_params(mdev, node, out, outlen); node 293 drivers/infiniband/hw/mlx5/cong.c enum mlx5_ib_cong_node_type node; node 312 drivers/infiniband/hw/mlx5/cong.c node = mlx5_ib_param_to_node(offset); node 313 drivers/infiniband/hw/mlx5/cong.c MLX5_SET(modify_cong_params_in, in, cong_protocol, node); node 578 drivers/infiniband/hw/mlx5/odp.c struct rb_node *node; node 581 drivers/infiniband/hw/mlx5/odp.c for (node = rb_first_cached(&per_mm->umem_tree); node; node 582 drivers/infiniband/hw/mlx5/odp.c node = rb_next(node)) { node 584 drivers/infiniband/hw/mlx5/odp.c rb_entry(node, struct ib_umem_odp, interval_tree.rb); node 61 drivers/infiniband/hw/qib/qib_user_sdma.c struct rb_node node; node 144 drivers/infiniband/hw/qib/qib_user_sdma.c struct rb_node *node = root->rb_node; node 146 drivers/infiniband/hw/qib/qib_user_sdma.c while (node) { node 147 drivers/infiniband/hw/qib/qib_user_sdma.c sdma_rb_node = rb_entry(node, struct qib_user_sdma_rb_node, node 148 drivers/infiniband/hw/qib/qib_user_sdma.c node); node 150 drivers/infiniband/hw/qib/qib_user_sdma.c node = node->rb_left; node 152 drivers/infiniband/hw/qib/qib_user_sdma.c node = node->rb_right; node 162 drivers/infiniband/hw/qib/qib_user_sdma.c struct rb_node **node = &(root->rb_node); node 166 drivers/infiniband/hw/qib/qib_user_sdma.c while (*node) { node 167 drivers/infiniband/hw/qib/qib_user_sdma.c got = rb_entry(*node, struct qib_user_sdma_rb_node, node); node 168 drivers/infiniband/hw/qib/qib_user_sdma.c parent = *node; node 170 drivers/infiniband/hw/qib/qib_user_sdma.c node = &((*node)->rb_left); node 172 drivers/infiniband/hw/qib/qib_user_sdma.c node = &((*node)->rb_right); node 177 drivers/infiniband/hw/qib/qib_user_sdma.c rb_link_node(&new->node, parent, node); node 178 drivers/infiniband/hw/qib/qib_user_sdma.c rb_insert_color(&new->node, root); node 1103 drivers/infiniband/hw/qib/qib_user_sdma.c rb_erase(&pq->sdma_rb_node->node, &qib_user_sdma_rb_root); node 1603 drivers/infiniband/hw/qib/qib_verbs.c dd->verbs_dev.rdi.dparms.node = dd->assigned_node_id; node 42 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define START(node) ((node)->start) node 43 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define LAST(node) ((node)->last) node 45 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MAKE_NODE(node, start, end, ref_cnt, flags, err, err_out) \ node 47 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c node = usnic_uiom_interval_node_alloc(start, \ node 49 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c if (!node) { \ node 55 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MARK_FOR_ADD(node, list) (list_add_tail(&node->link, list)) node 57 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MAKE_NODE_AND_APPEND(node, start, end, ref_cnt, flags, err, \ node 60 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c MAKE_NODE(node, start, end, \ node 63 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c MARK_FOR_ADD(node, list); \ node 107 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c struct usnic_uiom_interval_node *node; node 111 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c for (node = usnic_uiom_interval_tree_iter_first(root, start, last); node 112 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c node; node 113 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c node = usnic_uiom_interval_tree_iter_next(node, start, last)) node 114 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c list_add_tail(&node->link, list); node 50 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_insert(struct usnic_uiom_interval_node *node, node 53 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_remove(struct usnic_uiom_interval_node *node, node 60 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_iter_next(struct usnic_uiom_interval_node *node, node 240 drivers/infiniband/sw/rdmavt/cq.c k_wc = vzalloc_node(sz, rdi->dparms.node); node 289 drivers/infiniband/sw/rdmavt/cq.c cpumask_first(cpumask_of_node(rdi->dparms.node)); node 406 drivers/infiniband/sw/rdmavt/cq.c k_wc = vzalloc_node(sz, rdi->dparms.node); node 167 drivers/infiniband/sw/rdmavt/mmap.c ip = kmalloc_node(sizeof(*ip), GFP_KERNEL, rdi->dparms.node); node 91 drivers/infiniband/sw/rdmavt/mr.c vmalloc_node(lk_tab_size, rdi->dparms.node); node 146 drivers/infiniband/sw/rdmavt/mr.c dev->dparms.node); node 174 drivers/infiniband/sw/rdmavt/qp.c int node = rdi->dparms.node; node 181 drivers/infiniband/sw/rdmavt/qp.c rdi->wss = kzalloc_node(sizeof(*rdi->wss), GFP_KERNEL, node); node 219 drivers/infiniband/sw/rdmavt/qp.c GFP_KERNEL, node); node 419 drivers/infiniband/sw/rdmavt/qp.c rdi->dparms.node); node 429 drivers/infiniband/sw/rdmavt/qp.c GFP_KERNEL, rdi->dparms.node); node 822 drivers/infiniband/sw/rdmavt/qp.c int rvt_alloc_rq(struct rvt_rq *rq, u32 size, int node, node 830 drivers/infiniband/sw/rdmavt/qp.c rq->kwq = kzalloc_node(sizeof(*rq->kwq), GFP_KERNEL, node); node 837 drivers/infiniband/sw/rdmavt/qp.c vzalloc_node(sizeof(struct rvt_krwq) + size, node); node 1032 drivers/infiniband/sw/rdmavt/qp.c static int alloc_ud_wq_attr(struct rvt_qp *qp, int node) node 1040 drivers/infiniband/sw/rdmavt/qp.c GFP_KERNEL, node); node 1113 drivers/infiniband/sw/rdmavt/qp.c swq = vzalloc_node(array_size(sz, sqsize), rdi->dparms.node); node 1129 drivers/infiniband/sw/rdmavt/qp.c rdi->dparms.node); node 1140 drivers/infiniband/sw/rdmavt/qp.c rdi->dparms.node); node 1171 drivers/infiniband/sw/rdmavt/qp.c rdi->dparms.node, udata); node 1196 drivers/infiniband/sw/rdmavt/qp.c err = alloc_ud_wq_attr(qp, rdi->dparms.node); node 71 drivers/infiniband/sw/rdmavt/qp.h int rvt_alloc_rq(struct rvt_rq *rq, u32 size, int node, node 101 drivers/infiniband/sw/rdmavt/srq.c dev->dparms.node, udata)) { node 189 drivers/infiniband/sw/rdmavt/srq.c if (rvt_alloc_rq(&tmp_rq, size * sz, dev->dparms.node, node 291 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(parent, struct rxe_pool_entry, node); node 304 drivers/infiniband/sw/rxe/rxe_pool.c rb_link_node(&new->node, parent, link); node 305 drivers/infiniband/sw/rxe/rxe_pool.c rb_insert_color(&new->node, &pool->tree); node 319 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(parent, struct rxe_pool_entry, node); node 335 drivers/infiniband/sw/rxe/rxe_pool.c rb_link_node(&new->node, parent, link); node 336 drivers/infiniband/sw/rxe/rxe_pool.c rb_insert_color(&new->node, &pool->tree); node 360 drivers/infiniband/sw/rxe/rxe_pool.c rb_erase(&elem->node, &pool->tree); node 384 drivers/infiniband/sw/rxe/rxe_pool.c rb_erase(&elem->node, &pool->tree); node 479 drivers/infiniband/sw/rxe/rxe_pool.c struct rb_node *node = NULL; node 488 drivers/infiniband/sw/rxe/rxe_pool.c node = pool->tree.rb_node; node 490 drivers/infiniband/sw/rxe/rxe_pool.c while (node) { node 491 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(node, struct rxe_pool_entry, node); node 494 drivers/infiniband/sw/rxe/rxe_pool.c node = node->rb_left; node 496 drivers/infiniband/sw/rxe/rxe_pool.c node = node->rb_right; node 505 drivers/infiniband/sw/rxe/rxe_pool.c return node ? elem : NULL; node 510 drivers/infiniband/sw/rxe/rxe_pool.c struct rb_node *node = NULL; node 520 drivers/infiniband/sw/rxe/rxe_pool.c node = pool->tree.rb_node; node 522 drivers/infiniband/sw/rxe/rxe_pool.c while (node) { node 523 drivers/infiniband/sw/rxe/rxe_pool.c elem = rb_entry(node, struct rxe_pool_entry, node); node 529 drivers/infiniband/sw/rxe/rxe_pool.c node = node->rb_left; node 531 drivers/infiniband/sw/rxe/rxe_pool.c node = node->rb_right; node 536 drivers/infiniband/sw/rxe/rxe_pool.c if (node) node 541 drivers/infiniband/sw/rxe/rxe_pool.c return node ? elem : NULL; node 88 drivers/infiniband/sw/rxe/rxe_pool.h struct rb_node node; node 188 drivers/infiniband/sw/siw/siw_main.c int i, num_cpus, cpu, min_use, node = sdev->numa_node, tx_cpu = -1; node 190 drivers/infiniband/sw/siw/siw_main.c if (node < 0) node 193 drivers/infiniband/sw/siw/siw_main.c tx_cpumask = siw_cpu_info.tx_valid_cpus[node]; node 221 drivers/infiniband/sw/siw/siw_main.c "tx cpu %d, node %d, %d qp's\n", tx_cpu, node, min_use); node 393 drivers/infiniband/ulp/isert/ib_isert.c INIT_LIST_HEAD(&isert_conn->node); node 545 drivers/infiniband/ulp/isert/ib_isert.c list_add_tail(&isert_conn->node, &isert_np->accepted); node 606 drivers/infiniband/ulp/isert/ib_isert.c list_move_tail(&isert_conn->node, &isert_np->pending); node 637 drivers/infiniband/ulp/isert/ib_isert.c if (!list_empty(&isert_conn->node)) { node 642 drivers/infiniband/ulp/isert/ib_isert.c list_del_init(&isert_conn->node); node 739 drivers/infiniband/ulp/isert/ib_isert.c list_del_init(&isert_conn->node); node 2489 drivers/infiniband/ulp/isert/ib_isert.c struct isert_conn, node); node 2490 drivers/infiniband/ulp/isert/ib_isert.c list_del_init(&isert_conn->node); node 2525 drivers/infiniband/ulp/isert/ib_isert.c node) { node 2536 drivers/infiniband/ulp/isert/ib_isert.c node) { node 152 drivers/infiniband/ulp/isert/ib_isert.h struct list_head node; node 105 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_vnic_mac_tbl_node *node; node 112 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) { node 113 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c hash_del(&node->hlist); node 114 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c kfree(node); node 155 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_vnic_mac_tbl_node *node; node 168 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each(mactbl, bkt, node, hlist) { node 169 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct __opa_vnic_mactable_entry *nentry = &node->entry; node 172 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if ((node->index < loffset) || node 173 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c (node->index >= (loffset + lnum_entries))) node 177 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c entry = &tbl->tbl_entries[node->index - loffset]; node 207 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_vnic_mac_tbl_node *node, *new_node; node 241 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 242 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if (!node) { node 247 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c node->index = loffset + i; node 248 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c nentry = &node->entry; node 254 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c key = node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX]; node 255 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_add(new_mactbl, &node->hlist, key); node 263 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each(old_mactbl, bkt, node, hlist) { node 264 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if ((node->index >= loffset) && node 265 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c (node->index < (loffset + lnum_entries))) node 274 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c new_node->index = node->index; node 275 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(&new_node->entry, &node->entry, sizeof(node->entry)); node 301 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_vnic_mac_tbl_node *node; node 312 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each_possible(mactbl, node, hlist, key) { node 313 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct __opa_vnic_mactable_entry *entry = &node->entry; node 319 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if (!memcmp(node->entry.mac_addr, mac_hdr->h_dest, node 320 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c ARRAY_SIZE(node->entry.mac_addr))) { node 322 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c dlid = OPA_VNIC_DLID_SD_GET_DLID(node->entry.dlid_sd); node 276 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_add(hashtable, node, key) \ node 277 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h hlist_add_head(node, \ node 3778 drivers/infiniband/ulp/srp/ib_srp.c int ret, node_idx, node, cpu, i; node 3912 drivers/infiniband/ulp/srp/ib_srp.c for_each_online_node(node) { node 3924 drivers/infiniband/ulp/srp/ib_srp.c if (cpu_to_node(cpu) != node) node 48 drivers/input/evdev.c struct list_head node; node 305 drivers/input/evdev.c list_for_each_entry_rcu(client, &evdev->client_list, node) node 376 drivers/input/evdev.c list_add_tail_rcu(&client->node, &evdev->client_list); node 384 drivers/input/evdev.c list_del_rcu(&client->node); node 428 drivers/input/evdev.c list_for_each_entry(client, &evdev->client_list, node) node 263 drivers/input/gameport/gameport.c struct list_head node; node 278 drivers/input/gameport/gameport.c struct gameport_event, node); node 279 drivers/input/gameport/gameport.c list_del_init(&event->node); node 299 drivers/input/gameport/gameport.c list_for_each_entry_safe(e, next, &gameport_event_list, node) { node 309 drivers/input/gameport/gameport.c list_del_init(&e->node); node 368 drivers/input/gameport/gameport.c list_for_each_entry_reverse(event, &gameport_event_list, node) { node 395 drivers/input/gameport/gameport.c list_add_tail(&event->node, &gameport_event_list); node 414 drivers/input/gameport/gameport.c list_for_each_entry_safe(event, next, &gameport_event_list, node) { node 416 drivers/input/gameport/gameport.c list_del_init(&event->node); node 440 drivers/input/gameport/gameport.c list_for_each_entry(event, &gameport_event_list, node) { node 539 drivers/input/gameport/gameport.c INIT_LIST_HEAD(&gameport->node); node 559 drivers/input/gameport/gameport.c list_add_tail(&gameport->node, &gameport_list); node 597 drivers/input/gameport/gameport.c list_del_init(&gameport->node); node 765 drivers/input/gameport/gameport.c list_for_each_entry(gameport, &gameport_list, node) { node 37 drivers/input/gameport/ns558.c struct list_head node; node 146 drivers/input/gameport/ns558.c list_add(&ns558->node, &ns558_list); node 219 drivers/input/gameport/ns558.c list_add_tail(&ns558->node, &ns558_list); node 260 drivers/input/gameport/ns558.c list_for_each_entry_safe(ns558, safe, &ns558_list, node) { node 1165 drivers/input/input.c struct input_dev *dev = container_of(v, struct input_dev, node); node 1259 drivers/input/input.c struct input_handler *handler = container_of(v, struct input_handler, node); node 1826 drivers/input/input.c INIT_LIST_HEAD(&dev->node); node 2097 drivers/input/input.c list_del_init(&dev->node); node 2228 drivers/input/input.c list_add_tail(&dev->node, &input_dev_list); node 2230 drivers/input/input.c list_for_each_entry(handler, &input_handler_list, node) node 2300 drivers/input/input.c list_add_tail(&handler->node, &input_handler_list); node 2302 drivers/input/input.c list_for_each_entry(dev, &input_dev_list, node) node 2329 drivers/input/input.c list_del_init(&handler->node); node 66 drivers/input/joydev.c struct list_head node; node 147 drivers/input/joydev.c list_for_each_entry_rcu(client, &joydev->client_list, node) node 173 drivers/input/joydev.c list_add_tail_rcu(&client->node, &joydev->client_list); node 181 drivers/input/joydev.c list_del_rcu(&client->node); node 238 drivers/input/joydev.c list_for_each_entry(client, &joydev->client_list, node) node 251 drivers/input/keyboard/cap11xx.c struct device_node *node = dev->of_node, *child; node 253 drivers/input/keyboard/cap11xx.c int cnt = of_get_child_count(node); node 280 drivers/input/keyboard/cap11xx.c for_each_child_of_node(node, child) { node 326 drivers/input/keyboard/cap11xx.c struct device_node *node; node 378 drivers/input/keyboard/cap11xx.c node = dev->of_node; node 380 drivers/input/keyboard/cap11xx.c if (!of_property_read_u32(node, "microchip,sensor-gain", &gain32)) { node 387 drivers/input/keyboard/cap11xx.c if (of_property_read_bool(node, "microchip,irq-active-high")) { node 398 drivers/input/keyboard/cap11xx.c of_property_read_u32_array(node, "linux,keycodes", node 420 drivers/input/keyboard/cap11xx.c if (of_property_read_bool(node, "autorepeat")) node 455 drivers/input/keyboard/cap11xx.c irq = irq_of_parse_and_map(node, 0); node 243 drivers/input/keyboard/mtk-pmic-keys.c struct device_node *node = pdev->dev.of_node, *child; node 270 drivers/input/keyboard/mtk-pmic-keys.c keycount = of_get_available_child_count(node); node 276 drivers/input/keyboard/mtk-pmic-keys.c for_each_child_of_node(node, child) { node 169 drivers/input/misc/atmel_captouch.c struct device_node *node; node 206 drivers/input/misc/atmel_captouch.c node = dev->of_node; node 207 drivers/input/misc/atmel_captouch.c if (!node) { node 212 drivers/input/misc/atmel_captouch.c if (of_property_read_bool(node, "autorepeat")) node 215 drivers/input/misc/atmel_captouch.c capdev->num_btn = of_property_count_u32_elems(node, "linux,keymap"); node 219 drivers/input/misc/atmel_captouch.c err = of_property_read_u32_array(node, "linux,keycodes", node 119 drivers/input/misc/regulator-haptic.c struct device_node *node; node 122 drivers/input/misc/regulator-haptic.c node = dev->of_node; node 123 drivers/input/misc/regulator-haptic.c if(!node) { node 128 drivers/input/misc/regulator-haptic.c error = of_property_read_u32(node, "max-microvolt", &haptic->max_volt); node 134 drivers/input/misc/regulator-haptic.c error = of_property_read_u32(node, "min-microvolt", &haptic->min_volt); node 169 drivers/input/misc/twl4030-vibra.c struct device_node *node; node 174 drivers/input/misc/twl4030-vibra.c node = of_get_child_by_name(parent, "codec"); node 175 drivers/input/misc/twl4030-vibra.c if (node) { node 176 drivers/input/misc/twl4030-vibra.c of_node_put(node); node 21 drivers/input/mouse/psmouse-smbus.c struct list_head node; node 38 drivers/input/mouse/psmouse-smbus.c list_for_each_entry(smbdev, &psmouse_smbus_list, node) { node 66 drivers/input/mouse/psmouse-smbus.c list_for_each_entry_safe(smbdev, tmp, &psmouse_smbus_list, node) { node 80 drivers/input/mouse/psmouse-smbus.c list_del(&smbdev->node); node 173 drivers/input/mouse/psmouse-smbus.c list_del(&smbdev->node); node 216 drivers/input/mouse/psmouse-smbus.c list_for_each_entry_safe(smbdev, tmp, &psmouse_smbus_list, node) { node 218 drivers/input/mouse/psmouse-smbus.c list_del(&smbdev->node); node 263 drivers/input/mouse/psmouse-smbus.c list_add_tail(&smbdev->node, &psmouse_smbus_list); node 283 drivers/input/mouse/psmouse-smbus.c list_del(&smbdev->node); node 97 drivers/input/mousedev.c struct list_head node; node 270 drivers/input/mousedev.c list_for_each_entry_rcu(client, &mousedev->client_list, node) { node 507 drivers/input/mousedev.c list_add_tail_rcu(&client->node, &mousedev->client_list); node 515 drivers/input/mousedev.c list_del_rcu(&client->node); node 805 drivers/input/mousedev.c list_for_each_entry(client, &mousedev->client_list, node) node 158 drivers/input/rmi4/rmi_bus.c struct device_node *node = fn->rmi_dev->xport->dev->of_node; node 162 drivers/input/rmi4/rmi_bus.c fn->dev.of_node = of_get_child_by_name(node, of_name); node 40 drivers/input/rmi4/rmi_bus.h struct list_head node; node 44 drivers/input/rmi4/rmi_driver.c &data->function_list, node) { node 45 drivers/input/rmi4/rmi_driver.c list_del(&fn->node); node 104 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { node 119 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { node 248 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { node 281 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { node 315 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { node 850 drivers/input/rmi4/rmi_driver.c INIT_LIST_HEAD(&fn->node); node 871 drivers/input/rmi4/rmi_driver.c list_add_tail(&fn->node, &data->function_list); node 90 drivers/input/serio/gscps2.c struct list_head node; node 231 drivers/input/serio/gscps2.c list_for_each_entry(ps2port, &ps2port_list, node) { node 249 drivers/input/serio/gscps2.c list_for_each_entry(ps2port, &ps2port_list, node) { node 392 drivers/input/serio/gscps2.c list_add_tail(&ps2port->node, &ps2port_list); node 421 drivers/input/serio/gscps2.c list_del(&ps2port->node); node 583 drivers/input/serio/hil_mlc.c static inline void hilse_setup_input(hil_mlc *mlc, const struct hilse_node *node) node 586 drivers/input/serio/hil_mlc.c switch (node->act) { node 588 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; node 592 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; node 596 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; node 605 drivers/input/serio/hil_mlc.c mlc->intimeout = usecs_to_jiffies(node->arg); node 619 drivers/input/serio/hil_mlc.c const struct hilse_node *node; node 633 drivers/input/serio/hil_mlc.c node = hil_mlc_se + mlc->seidx; node 635 drivers/input/serio/hil_mlc.c switch (node->act) { node 640 drivers/input/serio/hil_mlc.c BUG_ON(node->object.func == NULL); node 641 drivers/input/serio/hil_mlc.c rc = node->object.func(mlc, node->arg); node 642 drivers/input/serio/hil_mlc.c nextidx = (rc > 0) ? node->ugly : node 643 drivers/input/serio/hil_mlc.c ((rc < 0) ? node->bad : node->good); node 654 drivers/input/serio/hil_mlc.c rc = mlc->in(mlc, node->arg); node 662 drivers/input/serio/hil_mlc.c nextidx = node->ugly; node 664 drivers/input/serio/hil_mlc.c nextidx = node->good; node 666 drivers/input/serio/hil_mlc.c nextidx = node->bad; node 673 drivers/input/serio/hil_mlc.c pack = node->object.packet; node 679 drivers/input/serio/hil_mlc.c pack = node->object.packet; node 685 drivers/input/serio/hil_mlc.c pack = node->object.packet; node 689 drivers/input/serio/hil_mlc.c if ((node + 1)->act & HILSE_IN) node 690 drivers/input/serio/hil_mlc.c hilse_setup_input(mlc, node + 1); node 718 drivers/input/serio/hil_mlc.c nextidx = mlc->cts(mlc) ? node->bad : node->good; node 144 drivers/input/serio/serio.c struct list_head node; node 159 drivers/input/serio/serio.c struct serio_event, node); node 160 drivers/input/serio/serio.c list_del_init(&event->node); node 181 drivers/input/serio/serio.c list_for_each_entry_safe(e, next, &serio_event_list, node) { node 191 drivers/input/serio/serio.c list_del_init(&e->node); node 256 drivers/input/serio/serio.c list_for_each_entry_reverse(event, &serio_event_list, node) { node 283 drivers/input/serio/serio.c list_add_tail(&event->node, &serio_event_list); node 302 drivers/input/serio/serio.c list_for_each_entry_safe(event, next, &serio_event_list, node) { node 304 drivers/input/serio/serio.c list_del_init(&event->node); node 326 drivers/input/serio/serio.c list_for_each_entry(event, &serio_event_list, node) { node 503 drivers/input/serio/serio.c INIT_LIST_HEAD(&serio->node); node 537 drivers/input/serio/serio.c list_add_tail(&serio->node, &serio_list); node 575 drivers/input/serio/serio.c list_del_init(&serio->node); node 863 drivers/input/serio/serio.c list_for_each_entry(serio, &serio_list, node) { node 38 drivers/input/serio/serio_raw.c struct list_head node; node 45 drivers/input/serio/serio_raw.c struct list_head node; node 66 drivers/input/serio/serio_raw.c list_for_each_entry(serio_raw, &serio_raw_list, node) { node 107 drivers/input/serio/serio_raw.c list_add_tail(&client->node, &serio_raw->client_list); node 130 drivers/input/serio/serio_raw.c list_del(&client->node); node 282 drivers/input/serio/serio_raw.c list_for_each_entry(client, &serio_raw->client_list, node) node 321 drivers/input/serio/serio_raw.c list_add_tail(&serio_raw->node, &serio_raw_list); node 347 drivers/input/serio/serio_raw.c list_del_init(&serio_raw->node); node 383 drivers/input/serio/serio_raw.c list_for_each_entry(client, &serio_raw->client_list, node) node 399 drivers/input/serio/serio_raw.c list_del_init(&serio_raw->node); node 1185 drivers/input/touchscreen/ads7846.c struct device_node *node = dev->of_node; node 1189 drivers/input/touchscreen/ads7846.c if (!node) { node 1206 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,vref-delay-usecs", node 1208 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,vref-mv", &pdata->vref_mv); node 1209 drivers/input/touchscreen/ads7846.c pdata->keep_vref_on = of_property_read_bool(node, "ti,keep-vref-on"); node 1211 drivers/input/touchscreen/ads7846.c pdata->swap_xy = of_property_read_bool(node, "ti,swap-xy"); node 1213 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,settle-delay-usec", node 1215 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,penirq-recheck-delay-usecs", node 1218 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,x-plate-ohms", &pdata->x_plate_ohms); node 1219 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,y-plate-ohms", &pdata->y_plate_ohms); node 1221 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,x-min", &pdata->x_min); node 1222 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,y-min", &pdata->y_min); node 1223 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,x-max", &pdata->x_max); node 1224 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,y-max", &pdata->y_max); node 1230 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,pressure-min", &pdata->pressure_min); node 1231 drivers/input/touchscreen/ads7846.c if (!of_property_read_u32(node, "touchscreen-min-pressure", &value)) node 1233 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,pressure-max", &pdata->pressure_max); node 1235 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,debounce-max", &pdata->debounce_max); node 1236 drivers/input/touchscreen/ads7846.c if (!of_property_read_u32(node, "touchscreen-average-samples", &value)) node 1238 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,debounce-tol", &pdata->debounce_tol); node 1239 drivers/input/touchscreen/ads7846.c of_property_read_u16(node, "ti,debounce-rep", &pdata->debounce_rep); node 1241 drivers/input/touchscreen/ads7846.c of_property_read_u32(node, "ti,pendown-gpio-debounce", node 1244 drivers/input/touchscreen/ads7846.c pdata->wakeup = of_property_read_bool(node, "wakeup-source") || node 1245 drivers/input/touchscreen/ads7846.c of_property_read_bool(node, "linux,wakeup"); node 606 drivers/input/touchscreen/mxs-lradc-ts.c struct device_node *node = dev->parent->of_node; node 626 drivers/input/touchscreen/mxs-lradc-ts.c ret = of_property_read_u32(node, "fsl,lradc-touchscreen-wires", node 631 drivers/input/touchscreen/mxs-lradc-ts.c if (of_property_read_u32(node, "fsl,ave-ctrl", &adapt)) { node 643 drivers/input/touchscreen/mxs-lradc-ts.c if (of_property_read_u32(node, "fsl,ave-delay", &adapt)) { node 655 drivers/input/touchscreen/mxs-lradc-ts.c if (of_property_read_u32(node, "fsl,settling", &adapt)) { node 678 drivers/input/touchscreen/mxs-lradc-ts.c virq = irq_of_parse_and_map(node, irq); node 885 drivers/input/touchscreen/sur40.c struct sur40_buffer *buf, *node; node 888 drivers/input/touchscreen/sur40.c list_for_each_entry_safe(buf, node, &sur40->buf_list, list) { node 344 drivers/input/touchscreen/ti_am335x_tsc.c struct device_node *node = pdev->dev.of_node; node 347 drivers/input/touchscreen/ti_am335x_tsc.c if (!node) node 350 drivers/input/touchscreen/ti_am335x_tsc.c err = of_property_read_u32(node, "ti,wires", &ts_dev->wires); node 362 drivers/input/touchscreen/ti_am335x_tsc.c err = of_property_read_u32(node, "ti,x-plate-resistance", node 371 drivers/input/touchscreen/ti_am335x_tsc.c err = of_property_read_u32(node, "ti,coordinate-readouts", node 375 drivers/input/touchscreen/ti_am335x_tsc.c err = of_property_read_u32(node, "ti,coordiante-readouts", node 388 drivers/input/touchscreen/ti_am335x_tsc.c err = of_property_read_u32(node, "ti,charge-delay", node 399 drivers/input/touchscreen/ti_am335x_tsc.c return of_property_read_u32_array(node, "ti,wire-config", node 38 drivers/interconnect/core.c struct icc_node *node; node 105 drivers/interconnect/core.c struct icc_node *node = dst; node 116 drivers/interconnect/core.c node->provider->users++; node 117 drivers/interconnect/core.c hlist_add_head(&path->reqs[i].req_node, &node->req_list); node 118 drivers/interconnect/core.c path->reqs[i].node = node; node 121 drivers/interconnect/core.c node = node->reverse; node 131 drivers/interconnect/core.c struct icc_node *n, *node = NULL; node 146 drivers/interconnect/core.c list_for_each_entry_safe(node, n, &traverse_list, search_list) { node 147 drivers/interconnect/core.c if (node == dst) { node 153 drivers/interconnect/core.c for (i = 0; i < node->num_links; i++) { node 154 drivers/interconnect/core.c struct icc_node *tmp = node->links[i]; node 165 drivers/interconnect/core.c tmp->reverse = node; node 200 drivers/interconnect/core.c static int aggregate_requests(struct icc_node *node) node 202 drivers/interconnect/core.c struct icc_provider *p = node->provider; node 205 drivers/interconnect/core.c node->avg_bw = 0; node 206 drivers/interconnect/core.c node->peak_bw = 0; node 209 drivers/interconnect/core.c p->pre_aggregate(node); node 211 drivers/interconnect/core.c hlist_for_each_entry(r, &node->req_list, req_node) node 212 drivers/interconnect/core.c p->aggregate(node, r->tag, r->avg_bw, r->peak_bw, node 213 drivers/interconnect/core.c &node->avg_bw, &node->peak_bw); node 225 drivers/interconnect/core.c next = path->reqs[i].node; node 284 drivers/interconnect/core.c struct icc_node *node = ERR_PTR(-EPROBE_DEFER); node 293 drivers/interconnect/core.c node = provider->xlate(spec, provider->data); node 294 drivers/interconnect/core.c if (!IS_ERR(node)) node 299 drivers/interconnect/core.c return node; node 434 drivers/interconnect/core.c struct icc_node *node; node 448 drivers/interconnect/core.c node = path->reqs[i].node; node 455 drivers/interconnect/core.c aggregate_requests(node); node 464 drivers/interconnect/core.c node = path->reqs[i].node; node 467 drivers/interconnect/core.c aggregate_requests(node); node 528 drivers/interconnect/core.c struct icc_node *node; node 541 drivers/interconnect/core.c node = path->reqs[i].node; node 543 drivers/interconnect/core.c if (!WARN_ON(!node->provider->users)) node 544 drivers/interconnect/core.c node->provider->users--; node 554 drivers/interconnect/core.c struct icc_node *node; node 557 drivers/interconnect/core.c node = node_find(id); node 558 drivers/interconnect/core.c if (node) node 559 drivers/interconnect/core.c return node; node 561 drivers/interconnect/core.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 562 drivers/interconnect/core.c if (!node) node 565 drivers/interconnect/core.c id = idr_alloc(&icc_idr, node, id, id + 1, GFP_KERNEL); node 568 drivers/interconnect/core.c kfree(node); node 572 drivers/interconnect/core.c node->id = id; node 574 drivers/interconnect/core.c return node; node 585 drivers/interconnect/core.c struct icc_node *node; node 589 drivers/interconnect/core.c node = icc_node_create_nolock(id); node 593 drivers/interconnect/core.c return node; node 603 drivers/interconnect/core.c struct icc_node *node; node 607 drivers/interconnect/core.c node = node_find(id); node 608 drivers/interconnect/core.c if (node) { node 609 drivers/interconnect/core.c idr_remove(&icc_idr, node->id); node 610 drivers/interconnect/core.c WARN_ON(!hlist_empty(&node->req_list)); node 615 drivers/interconnect/core.c kfree(node); node 632 drivers/interconnect/core.c int icc_link_create(struct icc_node *node, const int dst_id) node 638 drivers/interconnect/core.c if (!node->provider) node 653 drivers/interconnect/core.c new = krealloc(node->links, node 654 drivers/interconnect/core.c (node->num_links + 1) * sizeof(*node->links), node 661 drivers/interconnect/core.c node->links = new; node 662 drivers/interconnect/core.c node->links[node->num_links++] = dst; node 720 drivers/interconnect/core.c void icc_node_add(struct icc_node *node, struct icc_provider *provider) node 724 drivers/interconnect/core.c node->provider = provider; node 725 drivers/interconnect/core.c list_add_tail(&node->node_list, &provider->nodes); node 735 drivers/interconnect/core.c void icc_node_del(struct icc_node *node) node 739 drivers/interconnect/core.c list_del(&node->node_list); node 330 drivers/interconnect/qcom/qcs404.c static int qcom_icc_aggregate(struct icc_node *node, u32 tag, u32 avg_bw, node 417 drivers/interconnect/qcom/qcs404.c struct icc_node *node, *tmp; node 473 drivers/interconnect/qcom/qcs404.c node = icc_node_create(qnodes[i]->id); node 474 drivers/interconnect/qcom/qcs404.c if (IS_ERR(node)) { node 475 drivers/interconnect/qcom/qcs404.c ret = PTR_ERR(node); node 479 drivers/interconnect/qcom/qcs404.c node->name = qnodes[i]->name; node 480 drivers/interconnect/qcom/qcs404.c node->data = qnodes[i]; node 481 drivers/interconnect/qcom/qcs404.c icc_node_add(node, provider); node 483 drivers/interconnect/qcom/qcs404.c dev_dbg(dev, "registered node %s\n", node->name); node 487 drivers/interconnect/qcom/qcs404.c icc_link_create(node, qnodes[i]->links[j]); node 489 drivers/interconnect/qcom/qcs404.c data->nodes[i] = node; node 497 drivers/interconnect/qcom/qcs404.c list_for_each_entry_safe(node, tmp, &provider->nodes, node_list) { node 498 drivers/interconnect/qcom/qcs404.c icc_node_del(node); node 499 drivers/interconnect/qcom/qcs404.c icc_node_destroy(node->id); node 633 drivers/interconnect/qcom/sdm845.c static void qcom_icc_pre_aggregate(struct icc_node *node) node 638 drivers/interconnect/qcom/sdm845.c qn = node->data; node 646 drivers/interconnect/qcom/sdm845.c static int qcom_icc_aggregate(struct icc_node *node, u32 tag, u32 avg_bw, node 652 drivers/interconnect/qcom/sdm845.c qn = node->data; node 676 drivers/interconnect/qcom/sdm845.c struct icc_node *node; node 683 drivers/interconnect/qcom/sdm845.c node = dst; node 685 drivers/interconnect/qcom/sdm845.c node = src; node 687 drivers/interconnect/qcom/sdm845.c qp = to_qcom_provider(node->provider); node 778 drivers/interconnect/qcom/sdm845.c struct icc_node *node; node 820 drivers/interconnect/qcom/sdm845.c node = icc_node_create(qnodes[i]->id); node 821 drivers/interconnect/qcom/sdm845.c if (IS_ERR(node)) { node 822 drivers/interconnect/qcom/sdm845.c ret = PTR_ERR(node); node 826 drivers/interconnect/qcom/sdm845.c node->name = qnodes[i]->name; node 827 drivers/interconnect/qcom/sdm845.c node->data = qnodes[i]; node 828 drivers/interconnect/qcom/sdm845.c icc_node_add(node, provider); node 830 drivers/interconnect/qcom/sdm845.c dev_dbg(&pdev->dev, "registered node %p %s %d\n", node, node 831 drivers/interconnect/qcom/sdm845.c qnodes[i]->name, node->id); node 835 drivers/interconnect/qcom/sdm845.c icc_link_create(node, qnodes[i]->links[j]); node 837 drivers/interconnect/qcom/sdm845.c data->nodes[i] = node; node 858 drivers/interconnect/qcom/sdm845.c list_for_each_entry(node, &provider->nodes, node_list) { node 859 drivers/interconnect/qcom/sdm845.c icc_node_del(node); node 860 drivers/interconnect/qcom/sdm845.c icc_node_destroy(node->id); node 215 drivers/iommu/amd_iommu.c struct llist_node *node; node 220 drivers/iommu/amd_iommu.c node = dev_data_list.first; node 221 drivers/iommu/amd_iommu.c llist_for_each_entry(dev_data, node, dev_data_list) { node 3523 drivers/iommu/arm-smmu-v3.c struct acpi_iort_node *node; node 3525 drivers/iommu/arm-smmu-v3.c node = *(struct acpi_iort_node **)dev_get_platdata(dev); node 3528 drivers/iommu/arm-smmu-v3.c iort_smmu = (struct acpi_iort_smmu_v3 *)node->node_data; node 155 drivers/iommu/arm-smmu.c struct device_node *np = it->node; node 160 drivers/iommu/arm-smmu.c if (it->node == np) { node 164 drivers/iommu/arm-smmu.c it->node = np; node 188 drivers/iommu/arm-smmu.c it.node = np; node 1920 drivers/iommu/arm-smmu.c struct acpi_iort_node *node = node 1926 drivers/iommu/arm-smmu.c iort_smmu = (struct acpi_iort_smmu *)node->node_data; node 227 drivers/iommu/dma-iommu.c if (window->node.next == &bridge->dma_ranges && node 965 drivers/iommu/dma-iommu.c int node = dev_to_node(dev); node 971 drivers/iommu/dma-iommu.c page = alloc_pages_node(node, gfp, get_order(alloc_size)); node 476 drivers/iommu/dmar.c int node = acpi_map_pxm_to_node(rhsa->proximity_domain); node 478 drivers/iommu/dmar.c if (!node_online(node)) node 479 drivers/iommu/dmar.c node = NUMA_NO_NODE; node 480 drivers/iommu/dmar.c drhd->iommu->node = node; node 1064 drivers/iommu/dmar.c iommu->node = NUMA_NO_NODE; node 1480 drivers/iommu/dmar.c desc_page = alloc_pages_node(iommu->node, GFP_ATOMIC | __GFP_ZERO, node 1796 drivers/iommu/dmar.c irq = dmar_alloc_hwirq(iommu->seq_id, iommu->node, iommu); node 512 drivers/iommu/fsl_pamu.c struct device_node *node; node 519 drivers/iommu/fsl_pamu.c node = of_find_matching_node(NULL, l3_device_ids); node 520 drivers/iommu/fsl_pamu.c if (node) { node 521 drivers/iommu/fsl_pamu.c prop = of_get_property(node, "cache-stash-id", NULL); node 524 drivers/iommu/fsl_pamu.c node); node 525 drivers/iommu/fsl_pamu.c of_node_put(node); node 528 drivers/iommu/fsl_pamu.c of_node_put(node); node 534 drivers/iommu/fsl_pamu.c for_each_of_cpu_node(node) { node 535 drivers/iommu/fsl_pamu.c prop = of_get_property(node, "reg", &len); node 548 drivers/iommu/fsl_pamu.c prop = of_get_property(node, "cache-stash-id", NULL); node 551 drivers/iommu/fsl_pamu.c node); node 552 drivers/iommu/fsl_pamu.c of_node_put(node); node 555 drivers/iommu/fsl_pamu.c of_node_put(node); node 559 drivers/iommu/fsl_pamu.c prop = of_get_property(node, "next-level-cache", NULL); node 561 drivers/iommu/fsl_pamu.c pr_debug("can't find next-level-cache at %pOF\n", node); node 562 drivers/iommu/fsl_pamu.c of_node_put(node); node 565 drivers/iommu/fsl_pamu.c of_node_put(node); node 568 drivers/iommu/fsl_pamu.c node = of_find_node_by_phandle(*prop); node 569 drivers/iommu/fsl_pamu.c if (!node) { node 715 drivers/iommu/fsl_pamu.c struct device_node *node = NULL; node 718 drivers/iommu/fsl_pamu.c for_each_node_with_property(node, "fsl,liodn") { node 719 drivers/iommu/fsl_pamu.c prop = of_get_property(node, "fsl,liodn", &len); node 738 drivers/iommu/fsl_pamu.c if (of_device_is_compatible(node, "fsl,qman-portal")) node 740 drivers/iommu/fsl_pamu.c if (of_device_is_compatible(node, "fsl,qman")) node 742 drivers/iommu/fsl_pamu.c if (of_device_is_compatible(node, "fsl,bman")) node 514 drivers/iommu/intel-iommu.c void *alloc_pgtable_page(int node) node 519 drivers/iommu/intel-iommu.c page = alloc_pages_node(node, GFP_ATOMIC | __GFP_ZERO, 0); node 718 drivers/iommu/intel-iommu.c context = alloc_pgtable_page(iommu->node); node 1195 drivers/iommu/intel-iommu.c root = (struct root_entry *)alloc_pgtable_page(iommu->node); node 1756 drivers/iommu/intel-iommu.c domain->nid = iommu->node; node 1894 drivers/iommu/intel-iommu.c domain->nid = iommu->node; node 3042 drivers/iommu/intel-iommu.c new_ce = alloc_pgtable_page(iommu->node); node 4891 drivers/iommu/intel-iommu.c &adev->physical_node_list, node) { node 155 drivers/iommu/intel-pasid.c pages = alloc_pages_node(info->iommu->node, node 245 drivers/iommu/intel-pasid.c entries = alloc_pgtable_page(info->iommu->node); node 54 drivers/iommu/intel-svm.c irq = dmar_alloc_hwirq(DMAR_UNITS_SUPPORTED + iommu->seq_id, iommu->node, iommu); node 543 drivers/iommu/intel_irq_remapping.c pages = alloc_pages_node(iommu->node, GFP_KERNEL | __GFP_ZERO, node 42 drivers/iommu/iova.c iovad->cached_node = &iovad->anchor.node; node 43 drivers/iommu/iova.c iovad->cached32_node = &iovad->anchor.node; node 51 drivers/iommu/iova.c rb_link_node(&iovad->anchor.node, NULL, &iovad->rbroot.rb_node); node 52 drivers/iommu/iova.c rb_insert_color(&iovad->anchor.node, &iovad->rbroot); node 129 drivers/iommu/iova.c iovad->cached32_node = &new->node; node 131 drivers/iommu/iova.c iovad->cached_node = &new->node; node 139 drivers/iommu/iova.c cached_iova = rb_entry(iovad->cached32_node, struct iova, node); node 143 drivers/iommu/iova.c iovad->cached32_node = rb_next(&free->node); node 147 drivers/iommu/iova.c cached_iova = rb_entry(iovad->cached_node, struct iova, node); node 149 drivers/iommu/iova.c iovad->cached_node = rb_next(&free->node); node 162 drivers/iommu/iova.c struct iova *this = rb_entry(*new, struct iova, node); node 176 drivers/iommu/iova.c rb_link_node(&iova->node, parent, new); node 177 drivers/iommu/iova.c rb_insert_color(&iova->node, root); node 200 drivers/iommu/iova.c curr_iova = rb_entry(curr, struct iova, node); node 206 drivers/iommu/iova.c curr_iova = rb_entry(curr, struct iova, node); node 320 drivers/iommu/iova.c struct rb_node *node = iovad->rbroot.rb_node; node 324 drivers/iommu/iova.c while (node) { node 325 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); node 328 drivers/iommu/iova.c node = node->rb_left; node 330 drivers/iommu/iova.c node = node->rb_right; node 342 drivers/iommu/iova.c rb_erase(&iova->node, &iovad->rbroot); node 599 drivers/iommu/iova.c rbtree_postorder_for_each_entry_safe(iova, tmp, &iovad->rbroot, node) node 605 drivers/iommu/iova.c __is_range_overlap(struct rb_node *node, node 608 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); node 664 drivers/iommu/iova.c struct rb_node *node; node 674 drivers/iommu/iova.c for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) { node 675 drivers/iommu/iova.c if (__is_range_overlap(node, pfn_lo, pfn_hi)) { node 676 drivers/iommu/iova.c iova = rb_entry(node, struct iova, node); node 709 drivers/iommu/iova.c struct rb_node *node; node 712 drivers/iommu/iova.c for (node = rb_first(&from->rbroot); node; node = rb_next(node)) { node 713 drivers/iommu/iova.c struct iova *iova = rb_entry(node, struct iova, node); node 748 drivers/iommu/iova.c rb_erase(&iova->node, &iovad->rbroot); node 433 drivers/iommu/mtk_iommu_v1.c iommu_spec.np = of_node_get(it.node); node 599 drivers/iommu/mtk_iommu_v1.c larb_spec.np = of_node_get(it.node); node 40 drivers/iommu/omap-iommu.c struct list_head node; node 1263 drivers/iommu/omap-iommu.c list_for_each_entry_safe(orphan_dev, tmp, &orphan_dev_list, node) { node 1266 drivers/iommu/omap-iommu.c list_del(&orphan_dev->node); node 1707 drivers/iommu/omap-iommu.c node) { node 1719 drivers/iommu/omap-iommu.c list_add(&orphan_dev->node, &orphan_dev_list); node 108 drivers/iommu/rockchip-iommu.c struct list_head node; /* entry in rk_iommu_domain.iommus */ node 629 drivers/iommu/rockchip-iommu.c iommu = list_entry(pos, struct rk_iommu, node); node 918 drivers/iommu/rockchip-iommu.c list_del_init(&iommu->node); node 957 drivers/iommu/rockchip-iommu.c list_add_tail(&iommu->node, &rk_domain->iommus); node 353 drivers/iommu/virtio-iommu.c struct interval_tree_node *node, *next; node 358 drivers/iommu/virtio-iommu.c node = next; node 359 drivers/iommu/virtio-iommu.c mapping = container_of(node, struct viommu_mapping, iova); node 360 drivers/iommu/virtio-iommu.c next = interval_tree_iter_next(node, iova, last); node 372 drivers/iommu/virtio-iommu.c interval_tree_remove(node, &vdomain->mappings); node 392 drivers/iommu/virtio-iommu.c struct interval_tree_node *node; node 396 drivers/iommu/virtio-iommu.c node = interval_tree_iter_first(&vdomain->mappings, 0, -1UL); node 397 drivers/iommu/virtio-iommu.c while (node) { node 398 drivers/iommu/virtio-iommu.c mapping = container_of(node, struct viommu_mapping, iova); node 412 drivers/iommu/virtio-iommu.c node = interval_tree_iter_next(node, 0, -1UL); node 788 drivers/iommu/virtio-iommu.c struct interval_tree_node *node; node 792 drivers/iommu/virtio-iommu.c node = interval_tree_iter_first(&vdomain->mappings, iova, iova); node 793 drivers/iommu/virtio-iommu.c if (node) { node 794 drivers/iommu/virtio-iommu.c mapping = container_of(node, struct viommu_mapping, iova); node 140 drivers/irqchip/irq-al-fic.c static int al_fic_register(struct device_node *node, node 146 drivers/irqchip/irq-al-fic.c fic->domain = irq_domain_add_linear(node, node 200 drivers/irqchip/irq-al-fic.c static struct al_fic *al_fic_wire_init(struct device_node *node, node 225 drivers/irqchip/irq-al-fic.c ret = al_fic_register(node, fic); node 241 drivers/irqchip/irq-al-fic.c static int __init al_fic_init_dt(struct device_node *node, node 251 drivers/irqchip/irq-al-fic.c node->name); node 255 drivers/irqchip/irq-al-fic.c base = of_iomap(node, 0); node 257 drivers/irqchip/irq-al-fic.c pr_err("%s: fail to map memory\n", node->name); node 261 drivers/irqchip/irq-al-fic.c parent_irq = irq_of_parse_and_map(node, 0); node 263 drivers/irqchip/irq-al-fic.c pr_err("%s: fail to map irq\n", node->name); node 268 drivers/irqchip/irq-al-fic.c fic = al_fic_wire_init(node, node 270 drivers/irqchip/irq-al-fic.c node->name, node 274 drivers/irqchip/irq-al-fic.c node->name, node 191 drivers/irqchip/irq-alpine-msi.c struct device_node *node) node 196 drivers/irqchip/irq-alpine-msi.c gic_node = of_irq_find_parent(node); node 218 drivers/irqchip/irq-alpine-msi.c msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(node), node 230 drivers/irqchip/irq-alpine-msi.c static int alpine_msix_init(struct device_node *node, node 243 drivers/irqchip/irq-alpine-msi.c ret = of_address_to_resource(node, 0, &res); node 259 drivers/irqchip/irq-alpine-msi.c if (of_property_read_u32(node, "al,msi-base-spi", &priv->spi_first)) { node 265 drivers/irqchip/irq-alpine-msi.c if (of_property_read_u32(node, "al,msi-num-spis", &priv->num_spis)) { node 282 drivers/irqchip/irq-alpine-msi.c ret = alpine_msix_init_domains(priv, node); node 271 drivers/irqchip/irq-armada-370-xp.c static int armada_370_xp_msi_init(struct device_node *node, node 286 drivers/irqchip/irq-armada-370-xp.c pci_msi_create_irq_domain(of_node_to_fwnode(node), node 306 drivers/irqchip/irq-armada-370-xp.c static inline int armada_370_xp_msi_init(struct device_node *node, node 646 drivers/irqchip/irq-armada-370-xp.c static int __init armada_370_xp_mpic_of_init(struct device_node *node, node 653 drivers/irqchip/irq-armada-370-xp.c BUG_ON(of_address_to_resource(node, 0, &main_int_res)); node 654 drivers/irqchip/irq-armada-370-xp.c BUG_ON(of_address_to_resource(node, 1, &per_cpu_int_res)); node 658 drivers/irqchip/irq-armada-370-xp.c node->full_name)); node 661 drivers/irqchip/irq-armada-370-xp.c node->full_name)); node 678 drivers/irqchip/irq-armada-370-xp.c irq_domain_add_linear(node, nr_irqs, node 687 drivers/irqchip/irq-armada-370-xp.c armada_370_xp_msi_init(node, main_int_res.start); node 689 drivers/irqchip/irq-armada-370-xp.c parent_irq = irq_of_parse_and_map(node, 0); node 65 drivers/irqchip/irq-aspeed-i2c-ic.c static int __init aspeed_i2c_ic_of_init(struct device_node *node, node 75 drivers/irqchip/irq-aspeed-i2c-ic.c i2c_ic->base = of_iomap(node, 0); node 81 drivers/irqchip/irq-aspeed-i2c-ic.c i2c_ic->parent_irq = irq_of_parse_and_map(node, 0); node 87 drivers/irqchip/irq-aspeed-i2c-ic.c i2c_ic->irq_domain = irq_domain_add_linear(node, ASPEED_I2C_IC_NUM_BUS, node 184 drivers/irqchip/irq-aspeed-vic.c static int __init avic_of_init(struct device_node *node, node 195 drivers/irqchip/irq-aspeed-vic.c regs = of_iomap(node, 0); node 214 drivers/irqchip/irq-aspeed-vic.c vic->dom = irq_domain_add_simple(node, NUM_IRQS, 0, node 57 drivers/irqchip/irq-ath79-cpu.c struct device_node *node, struct device_node *parent) node 63 drivers/irqchip/irq-ath79-cpu.c node, "qca,ddr-wb-channels", "#qca,ddr-wb-channel-cells"); node 70 drivers/irqchip/irq-ath79-cpu.c node, "qca,ddr-wb-channel-interrupts", i, &irq); node 75 drivers/irqchip/irq-ath79-cpu.c node, "qca,ddr-wb-channels", node 84 drivers/irqchip/irq-ath79-cpu.c return mips_cpu_irq_of_init(node, parent); node 132 drivers/irqchip/irq-ath79-misc.c struct device_node *node, struct device_node *parent) node 138 drivers/irqchip/irq-ath79-misc.c irq = irq_of_parse_and_map(node, 0); node 144 drivers/irqchip/irq-ath79-misc.c base = of_iomap(node, 0); node 150 drivers/irqchip/irq-ath79-misc.c domain = irq_domain_add_linear(node, ATH79_MISC_IRQ_COUNT, node 162 drivers/irqchip/irq-ath79-misc.c struct device_node *node, struct device_node *parent) node 165 drivers/irqchip/irq-ath79-misc.c return ath79_misc_intc_of_init(node, parent); node 172 drivers/irqchip/irq-ath79-misc.c struct device_node *node, struct device_node *parent) node 175 drivers/irqchip/irq-ath79-misc.c return ath79_misc_intc_of_init(node, parent); node 112 drivers/irqchip/irq-ativic32.c int __init ativic32_init_irq(struct device_node *node, struct device_node *parent) node 130 drivers/irqchip/irq-ativic32.c root_domain = irq_domain_add_linear(node, nr_ints, node 134 drivers/irqchip/irq-ativic32.c panic("%s: unable to create IRQ domain\n", node->full_name); node 111 drivers/irqchip/irq-atmel-aic-common.c struct device_node *node = irq_domain_get_of_node(domain); node 123 drivers/irqchip/irq-atmel-aic-common.c of_property_for_each_u32(node, "atmel,external-irqs", prop, p, hwirq) { node 208 drivers/irqchip/irq-atmel-aic-common.c struct irq_domain *__init aic_common_of_init(struct device_node *node, node 223 drivers/irqchip/irq-atmel-aic-common.c reg_base = of_iomap(node, 0); node 233 drivers/irqchip/irq-atmel-aic-common.c domain = irq_domain_add_linear(node, nchips * 32, ops, aic); node 31 drivers/irqchip/irq-atmel-aic-common.h struct irq_domain *__init aic_common_of_init(struct device_node *node, node 241 drivers/irqchip/irq-atmel-aic.c static int __init aic_of_init(struct device_node *node, node 250 drivers/irqchip/irq-atmel-aic.c domain = aic_common_of_init(node, &aic_irq_ops, "atmel-aic", node 320 drivers/irqchip/irq-atmel-aic5.c static int __init aic5_of_init(struct device_node *node, node 335 drivers/irqchip/irq-atmel-aic5.c domain = aic_common_of_init(node, &aic5_irq_ops, "atmel-aic5", node 363 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d2_aic5_of_init(struct device_node *node, node 373 drivers/irqchip/irq-atmel-aic5.c return aic5_of_init(node, parent, NR_SAMA5D2_IRQS); node 379 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d3_aic5_of_init(struct device_node *node, node 382 drivers/irqchip/irq-atmel-aic5.c return aic5_of_init(node, parent, NR_SAMA5D3_IRQS); node 388 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d4_aic5_of_init(struct device_node *node, node 391 drivers/irqchip/irq-atmel-aic5.c return aic5_of_init(node, parent, NR_SAMA5D4_IRQS); node 397 drivers/irqchip/irq-atmel-aic5.c static int __init sam9x60_aic5_of_init(struct device_node *node, node 400 drivers/irqchip/irq-atmel-aic5.c return aic5_of_init(node, parent, NR_SAM9X60_IRQS); node 132 drivers/irqchip/irq-bcm2835.c static int __init armctrl_of_init(struct device_node *node, node 139 drivers/irqchip/irq-bcm2835.c base = of_iomap(node, 0); node 141 drivers/irqchip/irq-bcm2835.c panic("%pOF: unable to map IC registers\n", node); node 143 drivers/irqchip/irq-bcm2835.c intc.domain = irq_domain_add_linear(node, MAKE_HWIRQ(NR_BANKS, 0), node 146 drivers/irqchip/irq-bcm2835.c panic("%pOF: unable to create IRQ domain\n", node); node 163 drivers/irqchip/irq-bcm2835.c int parent_irq = irq_of_parse_and_map(node, 0); node 167 drivers/irqchip/irq-bcm2835.c node); node 177 drivers/irqchip/irq-bcm2835.c static int __init bcm2835_armctrl_of_init(struct device_node *node, node 180 drivers/irqchip/irq-bcm2835.c return armctrl_of_init(node, parent, false); node 183 drivers/irqchip/irq-bcm2835.c static int __init bcm2836_armctrl_of_init(struct device_node *node, node 186 drivers/irqchip/irq-bcm2835.c return armctrl_of_init(node, parent, true); node 219 drivers/irqchip/irq-bcm2836.c static int __init bcm2836_arm_irqchip_l1_intc_of_init(struct device_node *node, node 222 drivers/irqchip/irq-bcm2836.c intc.base = of_iomap(node, 0); node 224 drivers/irqchip/irq-bcm2836.c panic("%pOF: unable to map local interrupt registers\n", node); node 229 drivers/irqchip/irq-bcm2836.c intc.domain = irq_domain_add_linear(node, LAST_IRQ + 1, node 233 drivers/irqchip/irq-bcm2836.c panic("%pOF: unable to create IRQ domain\n", node); node 195 drivers/irqchip/irq-crossbar.c static int __init crossbar_of_init(struct device_node *node) node 207 drivers/irqchip/irq-crossbar.c cb->crossbar_base = of_iomap(node, 0); node 211 drivers/irqchip/irq-crossbar.c of_property_read_u32(node, "ti,max-crossbar-sources", node 219 drivers/irqchip/irq-crossbar.c of_property_read_u32(node, "ti,max-irqs", &max); node 235 drivers/irqchip/irq-crossbar.c irqsr = of_get_property(node, "ti,irqs-reserved", &size); node 240 drivers/irqchip/irq-crossbar.c of_property_read_u32_index(node, node 253 drivers/irqchip/irq-crossbar.c irqsr = of_get_property(node, "ti,irqs-skip", &size); node 258 drivers/irqchip/irq-crossbar.c of_property_read_u32_index(node, node 275 drivers/irqchip/irq-crossbar.c of_property_read_u32(node, "ti,reg-size", ®_size); node 306 drivers/irqchip/irq-crossbar.c of_property_read_u32(node, "ti,irqs-safe-map", &cb->safe_map); node 333 drivers/irqchip/irq-crossbar.c static int __init irqcrossbar_init(struct device_node *node, node 340 drivers/irqchip/irq-crossbar.c pr_err("%pOF: no parent, giving up\n", node); node 346 drivers/irqchip/irq-crossbar.c pr_err("%pOF: unable to obtain parent domain\n", node); node 350 drivers/irqchip/irq-crossbar.c err = crossbar_of_init(node); node 356 drivers/irqchip/irq-crossbar.c node, &crossbar_domain_ops, node 359 drivers/irqchip/irq-crossbar.c pr_err("%pOF: failed to allocated domain\n", node); node 60 drivers/irqchip/irq-csky-apb-intc.c static void __init ck_set_gc(struct device_node *node, void __iomem *reg_base, node 71 drivers/irqchip/irq-csky-apb-intc.c if (of_find_property(node, "csky,support-pulse-signal", NULL)) node 102 drivers/irqchip/irq-csky-apb-intc.c ck_intc_init_comm(struct device_node *node, struct device_node *parent) node 111 drivers/irqchip/irq-csky-apb-intc.c reg_base = of_iomap(node, 0); node 113 drivers/irqchip/irq-csky-apb-intc.c pr_err("C-SKY Intc unable to map: %p.\n", node); node 117 drivers/irqchip/irq-csky-apb-intc.c root_domain = irq_domain_add_linear(node, nr_irq, node 164 drivers/irqchip/irq-csky-apb-intc.c gx_intc_init(struct device_node *node, struct device_node *parent) node 168 drivers/irqchip/irq-csky-apb-intc.c ret = ck_intc_init_comm(node, parent); node 186 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base, GX_INTC_NEN31_00, 0); node 187 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base, GX_INTC_NEN63_32, 32); node 231 drivers/irqchip/irq-csky-apb-intc.c ck_intc_init(struct device_node *node, struct device_node *parent) node 235 drivers/irqchip/irq-csky-apb-intc.c ret = ck_intc_init_comm(node, parent); node 246 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base, CK_INTC_NEN31_00, 0); node 247 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base, CK_INTC_NEN63_32, 32); node 258 drivers/irqchip/irq-csky-apb-intc.c ck_dual_intc_init(struct device_node *node, struct device_node *parent) node 265 drivers/irqchip/irq-csky-apb-intc.c ret = ck_intc_init(node, parent); node 273 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base + CK_INTC_DUAL_BASE, CK_INTC_NEN31_00, 64); node 274 drivers/irqchip/irq-csky-apb-intc.c ck_set_gc(node, reg_base + CK_INTC_DUAL_BASE, CK_INTC_NEN63_32, 96); node 228 drivers/irqchip/irq-csky-mpintc.c csky_mpintc_init(struct device_node *node, struct device_node *parent) node 239 drivers/irqchip/irq-csky-mpintc.c ret = of_property_read_u32(node, "csky,num-irqs", &nr_irq); node 258 drivers/irqchip/irq-csky-mpintc.c root_domain = irq_domain_add_linear(node, nr_irq, &csky_irqdomain_ops, node 160 drivers/irqchip/irq-davinci-cp-intc.c struct device_node *node) node 217 drivers/irqchip/irq-davinci-cp-intc.c node, config->num_irqs, irq_base, 0, node 238 drivers/irqchip/irq-davinci-cp-intc.c static int __init davinci_cp_intc_of_init(struct device_node *node, node 244 drivers/irqchip/irq-davinci-cp-intc.c ret = of_address_to_resource(node, 0, &config.reg); node 251 drivers/irqchip/irq-davinci-cp-intc.c ret = of_property_read_u32(node, "ti,intc-size", &config.num_irqs); node 258 drivers/irqchip/irq-davinci-cp-intc.c return davinci_cp_intc_do_init(&config, node); node 71 drivers/irqchip/irq-digicolor.c static int __init digicolor_of_init(struct device_node *node, node 79 drivers/irqchip/irq-digicolor.c reg_base = of_iomap(node, 0); node 81 drivers/irqchip/irq-digicolor.c pr_err("%pOF: unable to map IC registers\n", node); node 89 drivers/irqchip/irq-digicolor.c ucregs = syscon_regmap_lookup_by_phandle(node, "syscon"); node 91 drivers/irqchip/irq-digicolor.c pr_err("%pOF: unable to map UC registers\n", node); node 98 drivers/irqchip/irq-digicolor.c irq_domain_add_linear(node, 64, &irq_generic_chip_ops, NULL); node 100 drivers/irqchip/irq-digicolor.c pr_err("%pOF: unable to create IRQ domain\n", node); node 108 drivers/irqchip/irq-digicolor.c pr_err("%pOF: unable to allocate IRQ gc\n", node); node 135 drivers/irqchip/irq-eznps.c static int __init nps400_of_init(struct device_node *node, node 145 drivers/irqchip/irq-eznps.c nps400_root_domain = irq_domain_add_linear(node, NPS_NR_CPU_IRQS, node 166 drivers/irqchip/irq-ftintc010.c int __init ft010_of_init_irq(struct device_node *node, node 177 drivers/irqchip/irq-ftintc010.c f->base = of_iomap(node, 0); node 184 drivers/irqchip/irq-ftintc010.c f->domain = irq_domain_add_simple(node, FT010_NUM_IRQS, 0, node 45 drivers/irqchip/irq-gic-realview.c realview_gic_of_init(struct device_node *node, struct device_node *parent) node 73 drivers/irqchip/irq-gic-realview.c return gic_of_init(node, parent); node 419 drivers/irqchip/irq-gic-v2m.c struct device_node *node = to_of_node(parent_handle); node 422 drivers/irqchip/irq-gic-v2m.c for (child = of_find_matching_node(node, gicv2m_device_id); child; node 3806 drivers/irqchip/irq-gic-v3-its.c static int __init its_of_probe(struct device_node *node) node 3811 drivers/irqchip/irq-gic-v3-its.c for (np = of_find_matching_node(node, its_device_id); np; node 3866 drivers/irqchip/irq-gic-v3-its.c int node; node 3879 drivers/irqchip/irq-gic-v3-its.c node = acpi_map_pxm_to_node(its_affinity->proximity_domain); node 3881 drivers/irqchip/irq-gic-v3-its.c if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) { node 3882 drivers/irqchip/irq-gic-v3-its.c pr_err("SRAT: Invalid NUMA node %d in ITS affinity\n", node); node 3886 drivers/irqchip/irq-gic-v3-its.c its_srat_maps[its_in_srat].numa_node = node; node 3890 drivers/irqchip/irq-gic-v3-its.c its_affinity->proximity_domain, its_affinity->its_id, node); node 1700 drivers/irqchip/irq-gic-v3.c static void __init gic_of_setup_kvm_info(struct device_node *node) node 1708 drivers/irqchip/irq-gic-v3.c gic_v3_kvm_info.maint_irq = irq_of_parse_and_map(node, 0); node 1712 drivers/irqchip/irq-gic-v3.c if (of_property_read_u32(node, "#redistributor-regions", node 1717 drivers/irqchip/irq-gic-v3.c ret = of_address_to_resource(node, gicv_idx, &r); node 1725 drivers/irqchip/irq-gic-v3.c static int __init gic_of_init(struct device_node *node, struct device_node *parent) node 1733 drivers/irqchip/irq-gic-v3.c dist_base = of_iomap(node, 0); node 1735 drivers/irqchip/irq-gic-v3.c pr_err("%pOF: unable to map gic dist registers\n", node); node 1741 drivers/irqchip/irq-gic-v3.c pr_err("%pOF: no distributor detected, giving up\n", node); node 1745 drivers/irqchip/irq-gic-v3.c if (of_property_read_u32(node, "#redistributor-regions", &nr_redist_regions)) node 1759 drivers/irqchip/irq-gic-v3.c ret = of_address_to_resource(node, 1 + i, &res); node 1760 drivers/irqchip/irq-gic-v3.c rdist_regs[i].redist_base = of_iomap(node, 1 + i); node 1762 drivers/irqchip/irq-gic-v3.c pr_err("%pOF: couldn't map region %d\n", node, i); node 1769 drivers/irqchip/irq-gic-v3.c if (of_property_read_u64(node, "redistributor-stride", &redist_stride)) node 1772 drivers/irqchip/irq-gic-v3.c gic_enable_of_quirks(node, gic_quirks, &gic_data); node 1775 drivers/irqchip/irq-gic-v3.c redist_stride, &node->fwnode); node 1779 drivers/irqchip/irq-gic-v3.c gic_populate_ppi_partitions(node); node 1782 drivers/irqchip/irq-gic-v3.c gic_of_setup_kvm_info(node); node 963 drivers/irqchip/irq-gic.c static void __init gic_init_physaddr(struct device_node *node) node 966 drivers/irqchip/irq-gic.c if (of_address_to_resource(node, 0, &res) == 0) { node 973 drivers/irqchip/irq-gic.c #define gic_init_physaddr(node) do { } while (0) node 1275 drivers/irqchip/irq-gic.c static bool gic_check_eoimode(struct device_node *node, void __iomem **base) node 1279 drivers/irqchip/irq-gic.c of_address_to_resource(node, 1, &cpuif_res); node 1353 drivers/irqchip/irq-gic.c static int gic_of_setup(struct gic_chip_data *gic, struct device_node *node) node 1355 drivers/irqchip/irq-gic.c if (!gic || !node) node 1358 drivers/irqchip/irq-gic.c gic->raw_dist_base = of_iomap(node, 0); node 1362 drivers/irqchip/irq-gic.c gic->raw_cpu_base = of_iomap(node, 1); node 1366 drivers/irqchip/irq-gic.c if (of_property_read_u32(node, "cpu-offset", &gic->percpu_offset)) node 1405 drivers/irqchip/irq-gic.c static void __init gic_of_setup_kvm_info(struct device_node *node) node 1413 drivers/irqchip/irq-gic.c gic_v2_kvm_info.maint_irq = irq_of_parse_and_map(node, 0); node 1417 drivers/irqchip/irq-gic.c ret = of_address_to_resource(node, 2, vctrl_res); node 1421 drivers/irqchip/irq-gic.c ret = of_address_to_resource(node, 3, vcpu_res); node 1430 drivers/irqchip/irq-gic.c gic_of_init(struct device_node *node, struct device_node *parent) node 1435 drivers/irqchip/irq-gic.c if (WARN_ON(!node)) node 1443 drivers/irqchip/irq-gic.c ret = gic_of_setup(gic, node); node 1451 drivers/irqchip/irq-gic.c if (gic_cnt == 0 && !gic_check_eoimode(node, &gic->raw_cpu_base)) node 1454 drivers/irqchip/irq-gic.c ret = __gic_init_bases(gic, &node->fwnode); node 1461 drivers/irqchip/irq-gic.c gic_init_physaddr(node); node 1462 drivers/irqchip/irq-gic.c gic_of_setup_kvm_info(node); node 1466 drivers/irqchip/irq-gic.c irq = irq_of_parse_and_map(node, 0); node 1471 drivers/irqchip/irq-gic.c gicv2m_init(&node->fwnode, gic_data[gic_cnt].domain); node 362 drivers/irqchip/irq-hip04.c hip04_of_init(struct device_node *node, struct device_node *parent) node 367 drivers/irqchip/irq-hip04.c if (WARN_ON(!node)) node 370 drivers/irqchip/irq-hip04.c hip04_data.dist_base = of_iomap(node, 0); node 373 drivers/irqchip/irq-hip04.c hip04_data.cpu_base = of_iomap(node, 1); node 401 drivers/irqchip/irq-hip04.c hip04_data.domain = irq_domain_add_legacy(node, nr_irqs, irq_base, node 312 drivers/irqchip/irq-i8259.c struct irq_domain * __init __init_i8259_irqs(struct device_node *node) node 321 drivers/irqchip/irq-i8259.c domain = irq_domain_add_legacy(node, 16, I8259A_IRQ_BASE, 0, node 349 drivers/irqchip/irq-i8259.c int __init i8259_of_init(struct device_node *node, struct device_node *parent) node 354 drivers/irqchip/irq-i8259.c domain = __init_i8259_irqs(node); node 356 drivers/irqchip/irq-i8259.c parent_irq = irq_of_parse_and_map(node, 0); node 300 drivers/irqchip/irq-imgpdc.c struct device_node *node = pdev->dev.of_node; node 307 drivers/irqchip/irq-imgpdc.c if (!node) node 333 drivers/irqchip/irq-imgpdc.c ret = of_property_read_u32(node, "num-perips", &val); node 345 drivers/irqchip/irq-imgpdc.c ret = of_property_read_u32(node, "num-syswakes", &val); node 382 drivers/irqchip/irq-imgpdc.c priv->domain = irq_domain_add_linear(node, 16, &irq_generic_chip_ops, node 202 drivers/irqchip/irq-imx-gpcv2.c static int __init imx_gpcv2_irqchip_init(struct device_node *node, node 212 drivers/irqchip/irq-imx-gpcv2.c pr_err("%pOF: no parent, giving up\n", node); node 216 drivers/irqchip/irq-imx-gpcv2.c id = of_match_node(gpcv2_of_match, node); node 218 drivers/irqchip/irq-imx-gpcv2.c pr_err("%pOF: unknown compatibility string\n", node); node 226 drivers/irqchip/irq-imx-gpcv2.c pr_err("%pOF: unable to get parent domain\n", node); node 232 drivers/irqchip/irq-imx-gpcv2.c pr_err("%pOF: kzalloc failed!\n", node); node 238 drivers/irqchip/irq-imx-gpcv2.c cd->gpc_base = of_iomap(node, 0); node 240 drivers/irqchip/irq-imx-gpcv2.c pr_err("%pOF: unable to map gpc registers\n", node); node 246 drivers/irqchip/irq-imx-gpcv2.c node, &gpcv2_irqchip_data_domain_ops, cd); node 287 drivers/irqchip/irq-imx-gpcv2.c of_node_clear_flag(node, OF_POPULATED); node 78 drivers/irqchip/irq-ingenic.c static int __init ingenic_intc_of_init(struct device_node *node, node 94 drivers/irqchip/irq-ingenic.c parent_irq = irq_of_parse_and_map(node, 0); node 105 drivers/irqchip/irq-ingenic.c intc->base = of_iomap(node, 0); node 111 drivers/irqchip/irq-ingenic.c domain = irq_domain_add_legacy(node, num_chips * 32, node 158 drivers/irqchip/irq-ingenic.c static int __init intc_1chip_of_init(struct device_node *node, node 161 drivers/irqchip/irq-ingenic.c return ingenic_intc_of_init(node, 1); node 166 drivers/irqchip/irq-ingenic.c static int __init intc_2chip_of_init(struct device_node *node, node 169 drivers/irqchip/irq-ingenic.c return ingenic_intc_of_init(node, 2); node 65 drivers/irqchip/irq-jcore-aic.c static int __init aic_irq_of_init(struct device_node *node, node 75 drivers/irqchip/irq-jcore-aic.c if (of_device_is_compatible(node, "jcore,aic1")) { node 79 drivers/irqchip/irq-jcore-aic.c void __iomem *base = of_iomap(node, cpu); node 103 drivers/irqchip/irq-jcore-aic.c domain = irq_domain_add_linear(node, dom_sz, &jcore_aic_irqdomain_ops, node 174 drivers/irqchip/irq-lpc32xx.c static int __init lpc32xx_of_ic_init(struct device_node *node, node 178 drivers/irqchip/irq-lpc32xx.c bool is_mic = of_device_is_compatible(node, "nxp,lpc3220-mic"); node 179 drivers/irqchip/irq-lpc32xx.c const __be32 *reg = of_get_property(node, "reg", NULL); node 186 drivers/irqchip/irq-lpc32xx.c irqc->base = of_iomap(node, 0); node 188 drivers/irqchip/irq-lpc32xx.c pr_err("%pOF: unable to map registers\n", node); node 202 drivers/irqchip/irq-lpc32xx.c irqc->domain = irq_domain_add_linear(node, NR_LPC32XX_IC_IRQS, node 216 drivers/irqchip/irq-lpc32xx.c for (i = 0; i < of_irq_count(node); i++) { node 217 drivers/irqchip/irq-lpc32xx.c parent_irq = irq_of_parse_and_map(node, i); node 103 drivers/irqchip/irq-ls1x.c static int __init ls1x_intc_of_init(struct device_node *node, node 115 drivers/irqchip/irq-ls1x.c priv->intc_base = of_iomap(node, 0); node 121 drivers/irqchip/irq-ls1x.c parent_irq = irq_of_parse_and_map(node, 0); node 129 drivers/irqchip/irq-ls1x.c priv->domain = irq_domain_add_linear(node, 32, &irq_generic_chip_ops, node 138 drivers/irqchip/irq-ls1x.c node->full_name, handle_level_irq, node 352 drivers/irqchip/irq-meson-gpio.c static int __init meson_gpio_irq_parse_dt(struct device_node *node, node 358 drivers/irqchip/irq-meson-gpio.c match = of_match_node(meson_irq_gpio_matches, node); node 364 drivers/irqchip/irq-meson-gpio.c ret = of_property_read_variable_u32_array(node, node 377 drivers/irqchip/irq-meson-gpio.c static int __init meson_gpio_irq_of_init(struct device_node *node, node 401 drivers/irqchip/irq-meson-gpio.c ctl->base = of_iomap(node, 0); node 407 drivers/irqchip/irq-meson-gpio.c ret = meson_gpio_irq_parse_dt(node, ctl); node 413 drivers/irqchip/irq-meson-gpio.c of_node_to_fwnode(node), node 208 drivers/irqchip/irq-mips-cpu.c static int mips_cpu_ipi_match(struct irq_domain *d, struct device_node *node, node 216 drivers/irqchip/irq-mips-cpu.c return (!node || (to_of_node(d->fwnode) == node)) && is_ipi; node 634 drivers/irqchip/irq-mips-gic.c int gic_ipi_domain_match(struct irq_domain *d, struct device_node *node, node 642 drivers/irqchip/irq-mips-gic.c return (!node || to_of_node(d->fwnode) == node) && is_ipi; node 671 drivers/irqchip/irq-mips-gic.c static int __init gic_of_init(struct device_node *node, node 683 drivers/irqchip/irq-mips-gic.c while (!of_property_read_u32_index(node, "mti,reserved-cpu-vectors", node 693 drivers/irqchip/irq-mips-gic.c if (of_address_to_resource(node, 0, &res)) { node 759 drivers/irqchip/irq-mips-gic.c gic_irq_domain = irq_domain_add_simple(node, GIC_NUM_LOCAL_INTRS + node 770 drivers/irqchip/irq-mips-gic.c node, &gic_ipi_domain_ops, NULL); node 778 drivers/irqchip/irq-mips-gic.c if (node && node 779 drivers/irqchip/irq-mips-gic.c !of_property_read_u32_array(node, "mti,reserved-ipi-vectors", v, 2)) { node 190 drivers/irqchip/irq-mmp.c static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node, node 373 drivers/irqchip/irq-mmp.c static int __init mmp_init_bases(struct device_node *node) node 377 drivers/irqchip/irq-mmp.c ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs); node 383 drivers/irqchip/irq-mmp.c mmp_icu_base = of_iomap(node, 0); node 390 drivers/irqchip/irq-mmp.c icu_data[0].domain = irq_domain_add_linear(node, nr_irqs, node 414 drivers/irqchip/irq-mmp.c static int __init mmp_of_init(struct device_node *node, node 419 drivers/irqchip/irq-mmp.c ret = mmp_init_bases(node); node 432 drivers/irqchip/irq-mmp.c static int __init mmp2_of_init(struct device_node *node, node 437 drivers/irqchip/irq-mmp.c ret = mmp_init_bases(node); node 450 drivers/irqchip/irq-mmp.c static int __init mmp3_of_init(struct device_node *node, node 455 drivers/irqchip/irq-mmp.c mmp_icu2_base = of_iomap(node, 1); node 461 drivers/irqchip/irq-mmp.c ret = mmp_init_bases(node); node 482 drivers/irqchip/irq-mmp.c static int __init mmp2_mux_of_init(struct device_node *node, node 493 drivers/irqchip/irq-mmp.c ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", node 506 drivers/irqchip/irq-mmp.c ret = of_property_read_variable_u32_array(node, "reg", reg, node 515 drivers/irqchip/irq-mmp.c icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0); node 520 drivers/irqchip/irq-mmp.c icu_data[i].domain = irq_domain_add_linear(node, nr_irqs, node 533 drivers/irqchip/irq-mmp.c if (!of_property_read_u32(node, "mrvl,clr-mfp-irq", node 61 drivers/irqchip/irq-mscc-ocelot.c static int __init ocelot_irq_init(struct device_node *node, node 68 drivers/irqchip/irq-mscc-ocelot.c parent_irq = irq_of_parse_and_map(node, 0); node 72 drivers/irqchip/irq-mscc-ocelot.c domain = irq_domain_add_linear(node, OCELOT_NR_IRQ, node 75 drivers/irqchip/irq-mscc-ocelot.c pr_err("%pOFn: unable to add irq domain\n", node); node 83 drivers/irqchip/irq-mscc-ocelot.c pr_err("%pOFn: unable to alloc irq domain gc\n", node); node 88 drivers/irqchip/irq-mscc-ocelot.c gc->reg_base = of_iomap(node, 0); node 90 drivers/irqchip/irq-mscc-ocelot.c pr_err("%pOFn: unable to map resource\n", node); node 243 drivers/irqchip/irq-mtk-cirq.c static int __init mtk_cirq_of_init(struct device_node *node, node 260 drivers/irqchip/irq-mtk-cirq.c cirq_data->base = of_iomap(node, 0); node 267 drivers/irqchip/irq-mtk-cirq.c ret = of_property_read_u32_index(node, "mediatek,ext-irq-range", 0, node 272 drivers/irqchip/irq-mtk-cirq.c ret = of_property_read_u32_index(node, "mediatek,ext-irq-range", 1, node 279 drivers/irqchip/irq-mtk-cirq.c irq_num, node, node 122 drivers/irqchip/irq-mtk-sysirq.c static int __init mtk_sysirq_of_init(struct device_node *node, node 139 drivers/irqchip/irq-mtk-sysirq.c while (of_get_address(node, i++, NULL, NULL)) node 167 drivers/irqchip/irq-mtk-sysirq.c ret = of_address_to_resource(node, i, &res); node 171 drivers/irqchip/irq-mtk-sysirq.c chip_data->intpol_bases[i] = of_iomap(node, i); node 173 drivers/irqchip/irq-mtk-sysirq.c pr_err("%pOF: couldn't map region %d\n", node, i); node 209 drivers/irqchip/irq-mtk-sysirq.c domain = irq_domain_add_hierarchy(domain_parent, 0, intpol_num, node, node 172 drivers/irqchip/irq-mvebu-gicp.c struct device_node *node = pdev->dev.of_node; node 187 drivers/irqchip/irq-mvebu-gicp.c ret = of_property_count_u32_elems(node, "marvell,spi-ranges"); node 202 drivers/irqchip/irq-mvebu-gicp.c of_property_read_u32_index(node, "marvell,spi-ranges", node 206 drivers/irqchip/irq-mvebu-gicp.c of_property_read_u32_index(node, "marvell,spi-ranges", node 219 drivers/irqchip/irq-mvebu-gicp.c irq_parent_dn = of_irq_find_parent(node); node 233 drivers/irqchip/irq-mvebu-gicp.c of_node_to_fwnode(node), node 239 drivers/irqchip/irq-mvebu-gicp.c plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node), node 161 drivers/irqchip/irq-mvebu-odmi.c static int __init mvebu_odmi_init(struct device_node *node, node 167 drivers/irqchip/irq-mvebu-odmi.c if (of_property_read_u32(node, "marvell,odmi-frames", &odmis_count)) node 184 drivers/irqchip/irq-mvebu-odmi.c ret = of_address_to_resource(node, i, &odmi->res); node 188 drivers/irqchip/irq-mvebu-odmi.c odmi->base = of_io_request_and_map(node, i, "odmi"); node 194 drivers/irqchip/irq-mvebu-odmi.c if (of_property_read_u32_index(node, "marvell,spi-base", node 201 drivers/irqchip/irq-mvebu-odmi.c inner_domain = irq_domain_create_linear(of_node_to_fwnode(node), node 211 drivers/irqchip/irq-mvebu-odmi.c plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node), node 124 drivers/irqchip/irq-mvebu-pic.c struct device_node *node = pdev->dev.of_node; node 144 drivers/irqchip/irq-mvebu-pic.c pic->parent_irq = irq_of_parse_and_map(node, 0); node 150 drivers/irqchip/irq-mvebu-pic.c pic->domain = irq_domain_add_linear(node, PIC_MAX_IRQS, node 370 drivers/irqchip/irq-mvebu-sei.c struct device_node *node = pdev->dev.of_node; node 404 drivers/irqchip/irq-mvebu-sei.c parent_irq = irq_of_parse_and_map(node, 0); node 411 drivers/irqchip/irq-mvebu-sei.c sei->sei_domain = irq_domain_create_linear(of_node_to_fwnode(node), node 427 drivers/irqchip/irq-mvebu-sei.c of_node_to_fwnode(node), node 441 drivers/irqchip/irq-mvebu-sei.c of_node_to_fwnode(node), node 452 drivers/irqchip/irq-mvebu-sei.c plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node), node 83 drivers/irqchip/irq-nvic.c static int __init nvic_of_init(struct device_node *node, node 93 drivers/irqchip/irq-nvic.c nvic_base = of_iomap(node, 0); node 104 drivers/irqchip/irq-nvic.c irq_domain_add_linear(node, irqs, &nvic_irq_domain_ops, NULL); node 243 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq_of(struct device_node *node) node 247 drivers/irqchip/irq-omap-intc.c omap_irq_base = of_iomap(node, 0); node 251 drivers/irqchip/irq-omap-intc.c domain = irq_domain_add_linear(node, omap_nr_irqs, node 263 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq_legacy(u32 base, struct device_node *node) node 277 drivers/irqchip/irq-omap-intc.c domain = irq_domain_add_legacy(node, omap_nr_irqs, irq_base, 0, node 297 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq(u32 base, struct device_node *node) node 307 drivers/irqchip/irq-omap-intc.c if (of_device_is_compatible(node, "ti,omap2-intc") || node 308 drivers/irqchip/irq-omap-intc.c of_device_is_compatible(node, "ti,omap3-intc")) { node 311 drivers/irqchip/irq-omap-intc.c if (of_address_to_resource(node, 0, &res)) node 315 drivers/irqchip/irq-omap-intc.c ret = omap_init_irq_legacy(base, node); node 316 drivers/irqchip/irq-omap-intc.c } else if (node) { node 317 drivers/irqchip/irq-omap-intc.c ret = omap_init_irq_of(node); node 363 drivers/irqchip/irq-omap-intc.c static int __init intc_of_init(struct device_node *node, node 371 drivers/irqchip/irq-omap-intc.c if (WARN_ON(!node)) node 374 drivers/irqchip/irq-omap-intc.c if (of_device_is_compatible(node, "ti,dm814-intc") || node 375 drivers/irqchip/irq-omap-intc.c of_device_is_compatible(node, "ti,dm816-intc") || node 376 drivers/irqchip/irq-omap-intc.c of_device_is_compatible(node, "ti,am33xx-intc")) { node 381 drivers/irqchip/irq-omap-intc.c ret = omap_init_irq(-1, of_node_get(node)); node 147 drivers/irqchip/irq-ompic.c static int __init ompic_of_init(struct device_node *node, node 160 drivers/irqchip/irq-ompic.c if (of_address_to_resource(node, 0, &res)) { node 179 drivers/irqchip/irq-ompic.c irq = irq_of_parse_and_map(node, 0); node 142 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_init(struct device_node *node, node 148 drivers/irqchip/irq-or1k-pic.c root_domain = irq_domain_add_linear(node, 32, &or1k_irq_domain_ops, node 156 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_or1200_init(struct device_node *node, node 159 drivers/irqchip/irq-or1k-pic.c return or1k_pic_init(node, &or1k_pic_or1200); node 164 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_level_init(struct device_node *node, node 167 drivers/irqchip/irq-or1k-pic.c return or1k_pic_init(node, &or1k_pic_level); node 172 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_edge_init(struct device_node *node, node 175 drivers/irqchip/irq-or1k-pic.c return or1k_pic_init(node, &or1k_pic_edge); node 192 drivers/irqchip/irq-pic32-evic.c struct device_node *node = irq_domain_get_of_node(domain); node 200 drivers/irqchip/irq-pic32-evic.c of_property_for_each_u32(node, pname, prop, p, hwirq) { node 212 drivers/irqchip/irq-pic32-evic.c static int __init pic32_of_init(struct device_node *node, node 223 drivers/irqchip/irq-pic32-evic.c evic_base = of_iomap(node, 0); node 233 drivers/irqchip/irq-pic32-evic.c evic_irq_domain = irq_domain_add_linear(node, nchips * 32, node 84 drivers/irqchip/irq-rda-intc.c static int __init rda8810_intc_init(struct device_node *node, node 87 drivers/irqchip/irq-rda-intc.c rda_intc_base = of_io_request_and_map(node, 0, "rda-intc"); node 94 drivers/irqchip/irq-rda-intc.c rda_irq_domain = irq_domain_create_linear(&node->fwnode, RDA_NR_IRQS, node 201 drivers/irqchip/irq-sifive-plic.c static int plic_find_hart_id(struct device_node *node) node 203 drivers/irqchip/irq-sifive-plic.c for (; node; node = node->parent) { node 204 drivers/irqchip/irq-sifive-plic.c if (of_device_is_compatible(node, "riscv")) node 205 drivers/irqchip/irq-sifive-plic.c return riscv_of_processor_hartid(node); node 211 drivers/irqchip/irq-sifive-plic.c static int __init plic_init(struct device_node *node, node 222 drivers/irqchip/irq-sifive-plic.c plic_regs = of_iomap(node, 0); node 227 drivers/irqchip/irq-sifive-plic.c of_property_read_u32(node, "riscv,ndev", &nr_irqs); node 231 drivers/irqchip/irq-sifive-plic.c nr_contexts = of_irq_count(node); node 238 drivers/irqchip/irq-sifive-plic.c plic_irqdomain = irq_domain_add_linear(node, nr_irqs + 1, node 250 drivers/irqchip/irq-sifive-plic.c if (of_irq_parse_one(node, i, &parent)) { node 210 drivers/irqchip/irq-sni-exiu.c static int __init exiu_dt_init(struct device_node *node, node 218 drivers/irqchip/irq-sni-exiu.c pr_err("%pOF: no parent, giving up\n", node); node 224 drivers/irqchip/irq-sni-exiu.c pr_err("%pOF: unable to obtain parent domain\n", node); node 228 drivers/irqchip/irq-sni-exiu.c if (of_address_to_resource(node, 0, &res)) { node 229 drivers/irqchip/irq-sni-exiu.c pr_err("%pOF: failed to parse memory resource\n", node); node 233 drivers/irqchip/irq-sni-exiu.c data = exiu_init(of_node_to_fwnode(node), &res); node 237 drivers/irqchip/irq-sni-exiu.c domain = irq_domain_add_hierarchy(parent_domain, 0, NUM_IRQS, node, node 240 drivers/irqchip/irq-sni-exiu.c pr_err("%pOF: failed to allocate domain\n", node); node 244 drivers/irqchip/irq-sni-exiu.c pr_info("%pOF: %d interrupts forwarded to %pOF\n", node, NUM_IRQS, node 653 drivers/irqchip/irq-stm32-exti.c struct device_node *node) node 668 drivers/irqchip/irq-stm32-exti.c host_data->base = of_iomap(node, 0); node 670 drivers/irqchip/irq-stm32-exti.c pr_err("%pOF: Unable to map registers\n", node); node 689 drivers/irqchip/irq-stm32-exti.c struct device_node *node) node 709 drivers/irqchip/irq-stm32-exti.c pr_info("%pOF: bank%d\n", node, bank_idx); node 715 drivers/irqchip/irq-stm32-exti.c struct device_node *node) node 723 drivers/irqchip/irq-stm32-exti.c host_data = stm32_exti_host_init(drv_data, node); node 727 drivers/irqchip/irq-stm32-exti.c domain = irq_domain_add_linear(node, drv_data->bank_nr * IRQS_PER_BANK, node 731 drivers/irqchip/irq-stm32-exti.c node); node 740 drivers/irqchip/irq-stm32-exti.c node); node 749 drivers/irqchip/irq-stm32-exti.c chip_data = stm32_exti_chip_init(host_data, i, node); node 768 drivers/irqchip/irq-stm32-exti.c nr_irqs = of_irq_count(node); node 770 drivers/irqchip/irq-stm32-exti.c unsigned int irq = irq_of_parse_and_map(node, i); node 107 drivers/irqchip/irq-sun4i.c static int __init sun4i_of_init(struct device_node *node, node 110 drivers/irqchip/irq-sun4i.c irq_ic_data->irq_base = of_iomap(node, 0); node 113 drivers/irqchip/irq-sun4i.c node); node 136 drivers/irqchip/irq-sun4i.c irq_ic_data->irq_domain = irq_domain_add_linear(node, 3 * 32, node 139 drivers/irqchip/irq-sun4i.c panic("%pOF: unable to create IRQ domain\n", node); node 146 drivers/irqchip/irq-sun4i.c static int __init sun4i_ic_of_init(struct device_node *node, node 158 drivers/irqchip/irq-sun4i.c return sun4i_of_init(node, parent); node 163 drivers/irqchip/irq-sun4i.c static int __init suniv_ic_of_init(struct device_node *node, node 175 drivers/irqchip/irq-sun4i.c return sun4i_of_init(node, parent); node 159 drivers/irqchip/irq-sunxi-nmi.c static int __init sunxi_sc_nmi_irq_init(struct device_node *node, node 169 drivers/irqchip/irq-sunxi-nmi.c domain = irq_domain_add_linear(node, 1, &irq_generic_chip_ops, NULL); node 183 drivers/irqchip/irq-sunxi-nmi.c irq = irq_of_parse_and_map(node, 0); node 191 drivers/irqchip/irq-sunxi-nmi.c gc->reg_base = of_io_request_and_map(node, 0, of_node_full_name(node)); node 235 drivers/irqchip/irq-sunxi-nmi.c static int __init sun6i_r_intc_irq_init(struct device_node *node, node 238 drivers/irqchip/irq-sunxi-nmi.c return sunxi_sc_nmi_irq_init(node, &sun6i_r_intc_reg_offs); node 243 drivers/irqchip/irq-sunxi-nmi.c static int __init sun6i_sc_nmi_irq_init(struct device_node *node, node 246 drivers/irqchip/irq-sunxi-nmi.c return sunxi_sc_nmi_irq_init(node, &sun6i_reg_offs); node 250 drivers/irqchip/irq-sunxi-nmi.c static int __init sun7i_sc_nmi_irq_init(struct device_node *node, node 253 drivers/irqchip/irq-sunxi-nmi.c return sunxi_sc_nmi_irq_init(node, &sun7i_reg_offs); node 257 drivers/irqchip/irq-sunxi-nmi.c static int __init sun9i_nmi_irq_init(struct device_node *node, node 260 drivers/irqchip/irq-sunxi-nmi.c return sunxi_sc_nmi_irq_init(node, &sun9i_reg_offs); node 173 drivers/irqchip/irq-tango.c struct device_node *node) node 181 drivers/irqchip/irq-tango.c irq = irq_of_parse_and_map(node, 0); node 183 drivers/irqchip/irq-tango.c panic("%pOFn: failed to get IRQ", node); node 185 drivers/irqchip/irq-tango.c err = of_address_to_resource(node, 0, &res); node 187 drivers/irqchip/irq-tango.c panic("%pOFn: failed to get address", node); node 193 drivers/irqchip/irq-tango.c dom = irq_domain_add_linear(node, 64, &irq_generic_chip_ops, chip); node 195 drivers/irqchip/irq-tango.c panic("%pOFn: failed to create irqdomain", node); node 197 drivers/irqchip/irq-tango.c err = irq_alloc_domain_generic_chips(dom, 32, 2, node->name, node 200 drivers/irqchip/irq-tango.c panic("%pOFn: failed to allocate irqchip", node); node 209 drivers/irqchip/irq-tango.c static int __init tangox_of_irq_init(struct device_node *node, node 216 drivers/irqchip/irq-tango.c base = of_iomap(node, 0); node 218 drivers/irqchip/irq-tango.c panic("%pOFn: of_iomap failed", node); node 220 drivers/irqchip/irq-tango.c of_address_to_resource(node, 0, &res); node 222 drivers/irqchip/irq-tango.c for_each_child_of_node(node, c) node 275 drivers/irqchip/irq-tegra.c static int __init tegra_ictlr_init(struct device_node *node, node 285 drivers/irqchip/irq-tegra.c pr_err("%pOF: no parent, giving up\n", node); node 291 drivers/irqchip/irq-tegra.c pr_err("%pOF: unable to obtain parent domain\n", node); node 295 drivers/irqchip/irq-tegra.c match = of_match_node(ictlr_matches, node); node 308 drivers/irqchip/irq-tegra.c base = of_iomap(node, i); node 323 drivers/irqchip/irq-tegra.c pr_err("%pOF: no valid regions, giving up\n", node); node 330 drivers/irqchip/irq-tegra.c node, num_ictlrs, soc->num_ictlrs); node 334 drivers/irqchip/irq-tegra.c node, &tegra_ictlr_domain_ops, node 337 drivers/irqchip/irq-tegra.c pr_err("%pOF: failed to allocated domain\n", node); node 345 drivers/irqchip/irq-tegra.c node, num_ictlrs * 32, parent); node 520 drivers/irqchip/irq-ti-sci-inta.c struct device_node *parent_node, *node; node 526 drivers/irqchip/irq-ti-sci-inta.c node = dev_of_node(dev); node 527 drivers/irqchip/irq-ti-sci-inta.c parent_node = of_irq_find_parent(node); node 584 drivers/irqchip/irq-ti-sci-inta.c msi_domain = ti_sci_inta_msi_create_irq_domain(of_node_to_fwnode(node), node 94 drivers/irqchip/irq-ts4800.c struct device_node *node = pdev->dev.of_node; node 111 drivers/irqchip/irq-ts4800.c parent_irq = irq_of_parse_and_map(node, 0); node 122 drivers/irqchip/irq-ts4800.c data->domain = irq_domain_add_linear(node, 8, &ts4800_ic_ops, data); node 150 drivers/irqchip/irq-versatile-fpga.c int parent_irq, u32 valid, struct device_node *node) node 173 drivers/irqchip/irq-versatile-fpga.c f->domain = irq_domain_add_simple(node, fls(valid), irq_start, node 195 drivers/irqchip/irq-versatile-fpga.c int __init fpga_irq_of_init(struct device_node *node, node 203 drivers/irqchip/irq-versatile-fpga.c if (WARN_ON(!node)) node 206 drivers/irqchip/irq-versatile-fpga.c base = of_iomap(node, 0); node 209 drivers/irqchip/irq-versatile-fpga.c if (of_property_read_u32(node, "clear-mask", &clear_mask)) node 212 drivers/irqchip/irq-versatile-fpga.c if (of_property_read_u32(node, "valid-mask", &valid_mask)) node 219 drivers/irqchip/irq-versatile-fpga.c parent_irq = irq_of_parse_and_map(node, 0); node 225 drivers/irqchip/irq-versatile-fpga.c fpga_irq_init(base, node->name, 0, parent_irq, valid_mask, node); node 232 drivers/irqchip/irq-versatile-fpga.c if (of_device_is_compatible(node, "arm,versatile-sic")) node 178 drivers/irqchip/irq-vf610-mscm-ir.c static int __init vf610_mscm_ir_of_init(struct device_node *node, node 195 drivers/irqchip/irq-vf610-mscm-ir.c mscm_ir_data->mscm_ir_base = of_io_request_and_map(node, 0, "mscm-ir"); node 202 drivers/irqchip/irq-vf610-mscm-ir.c mscm_cp_regmap = syscon_regmap_lookup_by_phandle(node, "fsl,cpucfg"); node 213 drivers/irqchip/irq-vf610-mscm-ir.c MSCM_IRSPRC_NUM, node, node 268 drivers/irqchip/irq-vic.c struct device_node *node) node 290 drivers/irqchip/irq-vic.c v->domain = irq_domain_add_simple(node, fls(valid_sources), irq, node 402 drivers/irqchip/irq-vic.c u32 vic_sources, struct device_node *node) node 428 drivers/irqchip/irq-vic.c vic_register(base, 0, irq_start, vic_sources, 0, node); node 433 drivers/irqchip/irq-vic.c struct device_node *node) node 451 drivers/irqchip/irq-vic.c vic_init_st(base, irq_start, vic_sources, node); node 468 drivers/irqchip/irq-vic.c vic_register(base, parent_irq, irq_start, vic_sources, resume_sources, node); node 506 drivers/irqchip/irq-vic.c static int __init vic_of_init(struct device_node *node, node 516 drivers/irqchip/irq-vic.c regs = of_iomap(node, 0); node 520 drivers/irqchip/irq-vic.c of_property_read_u32(node, "valid-mask", &interrupt_mask); node 521 drivers/irqchip/irq-vic.c of_property_read_u32(node, "valid-wakeup-mask", &wakeup_mask); node 526 drivers/irqchip/irq-vic.c __vic_init(regs, 0, 0, interrupt_mask, wakeup_mask, node); node 190 drivers/irqchip/irq-vt8500.c static int __init vt8500_irq_init(struct device_node *node, node 194 drivers/irqchip/irq-vt8500.c struct device_node *np = node; node 203 drivers/irqchip/irq-vt8500.c intc[active_cnt].domain = irq_domain_add_linear(node, 64, node 69 drivers/irqchip/irq-zevio.c static int __init zevio_of_init(struct device_node *node, node 79 drivers/irqchip/irq-zevio.c zevio_irq_io = of_iomap(node, 0); node 95 drivers/irqchip/irq-zevio.c zevio_irq_domain = irq_domain_add_linear(node, MAX_INTRS, node 271 drivers/irqchip/qcom-pdc.c static int qcom_pdc_init(struct device_node *node, struct device_node *parent) node 276 drivers/irqchip/qcom-pdc.c pdc_base = of_iomap(node, 0); node 278 drivers/irqchip/qcom-pdc.c pr_err("%pOF: unable to map PDC registers\n", node); node 284 drivers/irqchip/qcom-pdc.c pr_err("%pOF: unable to find PDC's parent domain\n", node); node 289 drivers/irqchip/qcom-pdc.c ret = pdc_setup_pin_mapping(node); node 291 drivers/irqchip/qcom-pdc.c pr_err("%pOF: failed to init PDC pin-hwirq mapping\n", node); node 296 drivers/irqchip/qcom-pdc.c of_fwnode_handle(node), node 299 drivers/irqchip/qcom-pdc.c pr_err("%pOF: GIC domain add failed\n", node); node 305 drivers/leds/led-class.c list_add_tail(&led_cdev->node, &leds_list); node 361 drivers/leds/led-class.c list_del(&led_cdev->node); node 248 drivers/leds/led-triggers.c list_for_each_entry(led_cdev, &leds_list, node) { node 277 drivers/leds/led-triggers.c list_for_each_entry(led_cdev, &leds_list, node) { node 602 drivers/leds/leds-max77693.c struct device_node *node = dev->of_node, *child_node; node 607 drivers/leds/leds-max77693.c of_property_read_u32(node, "maxim,boost-mode", &cfg->boost_mode); node 608 drivers/leds/leds-max77693.c of_property_read_u32(node, "maxim,boost-mvout", &cfg->boost_vout); node 609 drivers/leds/leds-max77693.c of_property_read_u32(node, "maxim,mvsys-min", &cfg->low_vsys); node 611 drivers/leds/leds-max77693.c for_each_available_child_of_node(node, child_node) { node 48 drivers/leds/trigger/ledtrig-panic.c list_for_each_entry(led_cdev, &leds_list, node) node 383 drivers/lightnvm/core.c tqueue = blk_alloc_queue_node(GFP_KERNEL, dev->q->node); node 1167 drivers/lightnvm/core.c struct nvm_dev *nvm_alloc_dev(int node) node 1171 drivers/lightnvm/core.c dev = kzalloc_node(sizeof(struct nvm_dev), GFP_KERNEL, node); node 160 drivers/macintosh/ans-lcd.c struct device_node* node; node 162 drivers/macintosh/ans-lcd.c node = of_find_node_by_name(NULL, "lcd"); node 163 drivers/macintosh/ans-lcd.c if (!node || !of_node_name_eq(node->parent, "gc")) { node 164 drivers/macintosh/ans-lcd.c of_node_put(node); node 167 drivers/macintosh/ans-lcd.c of_node_put(node); node 156 drivers/macintosh/windfarm_smu_controls.c static struct smu_fan_control *smu_fan_create(struct device_node *node, node 168 drivers/macintosh/windfarm_smu_controls.c l = of_get_property(node, "location", NULL); node 231 drivers/macintosh/windfarm_smu_controls.c v = of_get_property(node, "min-value", NULL); node 235 drivers/macintosh/windfarm_smu_controls.c v = of_get_property(node, "max-value", NULL); node 241 drivers/macintosh/windfarm_smu_controls.c reg = of_get_property(node, "reg", NULL); node 35 drivers/macintosh/windfarm_smu_sat.c struct device_node *node; node 210 drivers/macintosh/windfarm_smu_sat.c sat->node = of_node_get(dev); node 196 drivers/macintosh/windfarm_smu_sensors.c static struct smu_ad_sensor *smu_ads_create(struct device_node *node) node 205 drivers/macintosh/windfarm_smu_sensors.c l = of_get_property(node, "location", NULL); node 216 drivers/macintosh/windfarm_smu_sensors.c if (of_node_is_type(node, "temp-sensor") && node 225 drivers/macintosh/windfarm_smu_sensors.c } else if (of_node_is_type(node, "current-sensor") && node 234 drivers/macintosh/windfarm_smu_sensors.c } else if (of_node_is_type(node, "voltage-sensor") && node 243 drivers/macintosh/windfarm_smu_sensors.c } else if (of_node_is_type(node, "power-sensor") && node 255 drivers/macintosh/windfarm_smu_sensors.c v = of_get_property(node, "reg", NULL); node 264 drivers/mailbox/hi6220-mailbox.c struct device_node *node = pdev->dev.of_node; node 331 drivers/mailbox/hi6220-mailbox.c if (of_find_property(node, "hi6220,mbox-tx-noirq", NULL)) node 353 drivers/mailbox/mailbox.c list_for_each_entry(mbox, &mbox_cons, node) node 516 drivers/mailbox/mailbox.c list_add_tail(&mbox->node, &mbox_cons); node 536 drivers/mailbox/mailbox.c list_del(&mbox->node); node 676 drivers/mailbox/omap-mailbox.c struct device_node *node; node 684 drivers/mailbox/omap-mailbox.c node = of_find_node_by_phandle(phandle); node 685 drivers/mailbox/omap-mailbox.c if (!node) { node 691 drivers/mailbox/omap-mailbox.c mbox = omap_mbox_device_find(mdev, node->name); node 692 drivers/mailbox/omap-mailbox.c of_node_put(node); node 705 drivers/mailbox/omap-mailbox.c struct device_node *node = pdev->dev.of_node; node 714 drivers/mailbox/omap-mailbox.c if (!node) { node 724 drivers/mailbox/omap-mailbox.c if (of_property_read_u32(node, "ti,mbox-num-users", &num_users)) node 727 drivers/mailbox/omap-mailbox.c if (of_property_read_u32(node, "ti,mbox-num-fifos", &num_fifos)) node 730 drivers/mailbox/omap-mailbox.c info_count = of_get_available_child_count(node); node 744 drivers/mailbox/omap-mailbox.c child = of_get_next_available_child(node, child); node 436 drivers/mailbox/zynqmp-ipi-mailbox.c static int zynqmp_ipi_mbox_get_buf_res(struct device_node *node, node 442 drivers/mailbox/zynqmp-ipi-mailbox.c index = of_property_match_string(node, "reg-names", name); node 444 drivers/mailbox/zynqmp-ipi-mailbox.c ret = of_address_to_resource(node, index, res); node 474 drivers/mailbox/zynqmp-ipi-mailbox.c struct device_node *node) node 488 drivers/mailbox/zynqmp-ipi-mailbox.c ipi_mbox->dev.of_node = node; node 489 drivers/mailbox/zynqmp-ipi-mailbox.c dev_set_name(&ipi_mbox->dev, "%s", of_node_full_name(node)); node 502 drivers/mailbox/zynqmp-ipi-mailbox.c ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res); node 518 drivers/mailbox/zynqmp-ipi-mailbox.c ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res); node 541 drivers/mailbox/zynqmp-ipi-mailbox.c ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res); node 557 drivers/mailbox/zynqmp-ipi-mailbox.c ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res); node 579 drivers/mailbox/zynqmp-ipi-mailbox.c ret = of_property_read_u32(node, "xlnx,ipi-id", &ipi_mbox->remote_id); node 226 drivers/md/bcache/bcache.h struct rb_node node; node 1791 drivers/md/bcache/btree.c &dc->writeback_keys.keys, node) node 2508 drivers/md/bcache/btree.c if (RB_INSERT(&buf->keys, w, node, keybuf_cmp)) node 2550 drivers/md/bcache/btree.c w = RB_FIRST(&buf->keys, struct keybuf_key, node); node 2553 drivers/md/bcache/btree.c w = RB_LAST(&buf->keys, struct keybuf_key, node); node 2565 drivers/md/bcache/btree.c rb_erase(&w->node, &buf->keys); node 2589 drivers/md/bcache/btree.c w = RB_GREATER(&buf->keys, s, node, keybuf_nonoverlapping_cmp); node 2593 drivers/md/bcache/btree.c w = RB_NEXT(w, node); node 2611 drivers/md/bcache/btree.c w = RB_FIRST(&buf->keys, struct keybuf_key, node); node 2614 drivers/md/bcache/btree.c w = RB_NEXT(w, node); node 120 drivers/md/dm-bio-prison-v1.c rb_entry(*new, struct dm_bio_prison_cell, node); node 140 drivers/md/dm-bio-prison-v1.c rb_link_node(&cell_prealloc->node, parent, new); node 141 drivers/md/dm-bio-prison-v1.c rb_insert_color(&cell_prealloc->node, &prison->cells); node 188 drivers/md/dm-bio-prison-v1.c rb_erase(&cell->node, &prison->cells); node 216 drivers/md/dm-bio-prison-v1.c rb_erase(&cell->node, &prison->cells); node 257 drivers/md/dm-bio-prison-v1.c rb_erase(&cell->node, &prison->cells); node 266 drivers/md/dm-bio-prison-v1.c rb_erase(&cell->node, &prison->cells); node 42 drivers/md/dm-bio-prison-v1.h struct rb_node node; node 124 drivers/md/dm-bio-prison-v2.c rb_entry(*new, struct dm_bio_prison_cell_v2, node); node 143 drivers/md/dm-bio-prison-v2.c rb_link_node(&cell_prealloc->node, parent, new); node 144 drivers/md/dm-bio-prison-v2.c rb_insert_color(&cell_prealloc->node, &prison->cells); node 204 drivers/md/dm-bio-prison-v2.c rb_erase(&cell->node, &prison->cells); node 336 drivers/md/dm-bio-prison-v2.c rb_erase(&cell->node, &prison->cells); node 51 drivers/md/dm-bio-prison-v2.h struct rb_node node; node 134 drivers/md/dm-bufio.c struct rb_node node; node 254 drivers/md/dm-bufio.c b = container_of(n, struct dm_buffer, node); node 271 drivers/md/dm-bufio.c found = container_of(*new, struct dm_buffer, node); node 283 drivers/md/dm-bufio.c rb_link_node(&b->node, parent, new); node 284 drivers/md/dm-bufio.c rb_insert_color(&b->node, &c->buffer_tree); node 289 drivers/md/dm-bufio.c rb_erase(&b->node, &c->buffer_tree); node 15 drivers/md/dm-cache-background-tracker.c struct rb_node node; node 87 drivers/md/dm-cache-background-tracker.c w = container_of(*new, struct bt_work, node); node 102 drivers/md/dm-cache-background-tracker.c rb_link_node(&nw->node, parent, new); node 103 drivers/md/dm-cache-background-tracker.c rb_insert_color(&nw->node, &b->pending); node 116 drivers/md/dm-cache-background-tracker.c w = container_of(*new, struct bt_work, node); node 236 drivers/md/dm-cache-background-tracker.c rb_erase(&w->node, &b->pending); node 1509 drivers/md/dm-crypt.c #define crypt_io_from_node(node) rb_entry((node), struct dm_crypt_io, rb_node) node 18 drivers/md/dm-dust.c struct rb_node node; node 37 drivers/md/dm-dust.c struct rb_node *node = root->rb_node; node 39 drivers/md/dm-dust.c while (node) { node 40 drivers/md/dm-dust.c struct badblock *bblk = rb_entry(node, struct badblock, node); node 43 drivers/md/dm-dust.c node = node->rb_left; node 45 drivers/md/dm-dust.c node = node->rb_right; node 61 drivers/md/dm-dust.c bblk = rb_entry(parent, struct badblock, node); node 71 drivers/md/dm-dust.c rb_link_node(&new->node, parent, link); node 72 drivers/md/dm-dust.c rb_insert_color(&new->node, root); node 94 drivers/md/dm-dust.c rb_erase(&bblock->node, &dd->badblocklist); node 183 drivers/md/dm-dust.c rb_erase(&bblk->node, &dd->badblocklist); node 228 drivers/md/dm-dust.c struct rb_node *node = NULL, *nnode = NULL; node 237 drivers/md/dm-dust.c node = nnode; node 238 drivers/md/dm-dust.c nnode = rb_next(node); node 239 drivers/md/dm-dust.c rb_erase(node, tree); node 241 drivers/md/dm-dust.c kfree(node); node 140 drivers/md/dm-integrity.c struct rb_node node; node 272 drivers/md/dm-integrity.c struct rb_node node; node 1116 drivers/md/dm-integrity.c struct dm_integrity_range *range = container_of(*n, struct dm_integrity_range, node); node 1120 drivers/md/dm-integrity.c n = &range->node.rb_left; node 1122 drivers/md/dm-integrity.c n = &range->node.rb_right; node 1128 drivers/md/dm-integrity.c rb_link_node(&new_range->node, parent, n); node 1129 drivers/md/dm-integrity.c rb_insert_color(&new_range->node, &ic->in_progress); node 1136 drivers/md/dm-integrity.c rb_erase(&range->node, &ic->in_progress); node 1181 drivers/md/dm-integrity.c static void init_journal_node(struct journal_node *node) node 1183 drivers/md/dm-integrity.c RB_CLEAR_NODE(&node->node); node 1184 drivers/md/dm-integrity.c node->sector = (sector_t)-1; node 1187 drivers/md/dm-integrity.c static void add_journal_node(struct dm_integrity_c *ic, struct journal_node *node, sector_t sector) node 1192 drivers/md/dm-integrity.c node->sector = sector; node 1193 drivers/md/dm-integrity.c BUG_ON(!RB_EMPTY_NODE(&node->node)); node 1201 drivers/md/dm-integrity.c j = container_of(parent, struct journal_node, node); node 1203 drivers/md/dm-integrity.c link = &j->node.rb_left; node 1205 drivers/md/dm-integrity.c link = &j->node.rb_right; node 1208 drivers/md/dm-integrity.c rb_link_node(&node->node, parent, link); node 1209 drivers/md/dm-integrity.c rb_insert_color(&node->node, &ic->journal_tree_root); node 1212 drivers/md/dm-integrity.c static void remove_journal_node(struct dm_integrity_c *ic, struct journal_node *node) node 1214 drivers/md/dm-integrity.c BUG_ON(RB_EMPTY_NODE(&node->node)); node 1215 drivers/md/dm-integrity.c rb_erase(&node->node, &ic->journal_tree_root); node 1216 drivers/md/dm-integrity.c init_journal_node(node); node 1227 drivers/md/dm-integrity.c struct journal_node *j = container_of(n, struct journal_node, node); node 1233 drivers/md/dm-integrity.c n = j->node.rb_left; node 1235 drivers/md/dm-integrity.c n = j->node.rb_right; node 1244 drivers/md/dm-integrity.c struct journal_node *node, *next_node; node 1249 drivers/md/dm-integrity.c node = &ic->journal_tree[pos]; node 1250 drivers/md/dm-integrity.c if (unlikely(RB_EMPTY_NODE(&node->node))) node 1252 drivers/md/dm-integrity.c if (unlikely(node->sector != sector)) node 1255 drivers/md/dm-integrity.c next = rb_next(&node->node); node 1259 drivers/md/dm-integrity.c next_node = container_of(next, struct journal_node, node); node 1263 drivers/md/dm-integrity.c static bool find_newer_committed_node(struct dm_integrity_c *ic, struct journal_node *node) node 1269 drivers/md/dm-integrity.c BUG_ON(RB_EMPTY_NODE(&node->node)); node 1271 drivers/md/dm-integrity.c next = rb_next(&node->node); node 1275 drivers/md/dm-integrity.c next_node = container_of(next, struct journal_node, node); node 1277 drivers/md/dm-integrity.c if (next_node->sector != node->sector) node 240 drivers/md/dm-snap.c struct hlist_node node; node 247 drivers/md/dm-snap.c INIT_HLIST_NODE(&c->node); node 253 drivers/md/dm-snap.c return !hlist_unhashed(&c->node); node 263 drivers/md/dm-snap.c hlist_add_head(&c->node, node 274 drivers/md/dm-snap.c hlist_del(&c->node); node 286 drivers/md/dm-snap.c &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) { node 143 drivers/md/dm-stats.c static void *dm_kvzalloc(size_t alloc_size, int node) node 150 drivers/md/dm-stats.c p = kvzalloc_node(alloc_size, GFP_KERNEL | __GFP_NOMEMALLOC, node); node 125 drivers/md/dm-table.c sector_t *node; node 128 drivers/md/dm-table.c node = get_node(t, l, n); node 131 drivers/md/dm-table.c node[k] = high(t, l + 1, get_child(n, k)); node 1366 drivers/md/dm-table.c sector_t *node; node 1373 drivers/md/dm-table.c node = get_node(t, l, n); node 1376 drivers/md/dm-table.c if (node[k] >= sector) node 2108 drivers/md/dm-thin.c #define thin_pbd(node) rb_entry((node), struct dm_thin_endio_hook, rb_node) node 2136 drivers/md/dm-thin.c struct rb_node *node; node 2140 drivers/md/dm-thin.c for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) { node 2141 drivers/md/dm-thin.c pbd = thin_pbd(node); node 541 drivers/md/dm-writecache.c struct rb_node *node = wc->tree.rb_node; node 543 drivers/md/dm-writecache.c if (unlikely(!node)) node 547 drivers/md/dm-writecache.c e = container_of(node, struct wc_entry, rb_node); node 551 drivers/md/dm-writecache.c node = (read_original_sector(wc, e) >= block ? node 553 drivers/md/dm-writecache.c if (unlikely(!node)) { node 559 drivers/md/dm-writecache.c node = rb_next(&e->rb_node); node 560 drivers/md/dm-writecache.c if (unlikely(!node)) node 562 drivers/md/dm-writecache.c e = container_of(node, struct wc_entry, rb_node); node 571 drivers/md/dm-writecache.c node = rb_prev(&e->rb_node); node 573 drivers/md/dm-writecache.c node = rb_next(&e->rb_node); node 574 drivers/md/dm-writecache.c if (unlikely(!node)) node 576 drivers/md/dm-writecache.c e2 = container_of(node, struct wc_entry, rb_node); node 586 drivers/md/dm-writecache.c struct rb_node **node = &wc->tree.rb_node, *parent = NULL; node 588 drivers/md/dm-writecache.c while (*node) { node 589 drivers/md/dm-writecache.c e = container_of(*node, struct wc_entry, rb_node); node 592 drivers/md/dm-writecache.c node = &parent->rb_left; node 594 drivers/md/dm-writecache.c node = &parent->rb_right; node 596 drivers/md/dm-writecache.c rb_link_node(&ins->rb_node, parent, node); node 610 drivers/md/dm-writecache.c struct rb_node **node = &wc->freetree.rb_node, *parent = NULL; node 611 drivers/md/dm-writecache.c if (unlikely(!*node)) node 613 drivers/md/dm-writecache.c while (*node) { node 614 drivers/md/dm-writecache.c parent = *node; node 615 drivers/md/dm-writecache.c if (&e->rb_node < *node) node 616 drivers/md/dm-writecache.c node = &parent->rb_left; node 618 drivers/md/dm-writecache.c node = &parent->rb_right; node 620 drivers/md/dm-writecache.c rb_link_node(&e->rb_node, parent, node); node 800 drivers/md/dm-writecache.c struct rb_node *node = rb_next(&e->rb_node); node 811 drivers/md/dm-writecache.c if (unlikely(!node)) node 814 drivers/md/dm-writecache.c e = container_of(node, struct wc_entry, rb_node); node 1604 drivers/md/dm-writecache.c struct rb_node *node, *next_node; node 1652 drivers/md/dm-writecache.c node = rb_prev(&e->rb_node); node 1653 drivers/md/dm-writecache.c if (node) { node 1654 drivers/md/dm-writecache.c f = container_of(node, struct wc_entry, rb_node); node 101 drivers/md/dm-zoned-metadata.c struct rb_node node; node 281 drivers/md/dm-zoned-metadata.c rb_erase(&mblk->node, &zmd->mblk_rbtree); node 300 drivers/md/dm-zoned-metadata.c RB_CLEAR_NODE(&mblk->node); node 334 drivers/md/dm-zoned-metadata.c b = container_of(*new, struct dmz_mblock, node); node 340 drivers/md/dm-zoned-metadata.c rb_link_node(&mblk->node, parent, new); node 341 drivers/md/dm-zoned-metadata.c rb_insert_color(&mblk->node, root); node 352 drivers/md/dm-zoned-metadata.c struct rb_node *node = root->rb_node; node 355 drivers/md/dm-zoned-metadata.c while (node) { node 356 drivers/md/dm-zoned-metadata.c mblk = container_of(node, struct dmz_mblock, node); node 368 drivers/md/dm-zoned-metadata.c node = (mblk->no < mblk_no) ? node->rb_left : node->rb_right; node 471 drivers/md/dm-zoned-metadata.c rb_erase(&mblk->node, &zmd->mblk_rbtree); node 521 drivers/md/dm-zoned-metadata.c rb_erase(&mblk->node, &zmd->mblk_rbtree); node 2367 drivers/md/dm-zoned-metadata.c rb_erase(&mblk->node, &zmd->mblk_rbtree); node 2375 drivers/md/dm-zoned-metadata.c rb_erase(&mblk->node, &zmd->mblk_rbtree); node 2381 drivers/md/dm-zoned-metadata.c rbtree_postorder_for_each_entry_safe(mblk, next, root, node) { node 85 drivers/md/persistent-data/dm-btree.c static int insert_at(size_t value_size, struct btree_node *node, unsigned index, node 89 drivers/md/persistent-data/dm-btree.c uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); node 93 drivers/md/persistent-data/dm-btree.c index >= le32_to_cpu(node->header.max_entries)) { node 101 drivers/md/persistent-data/dm-btree.c array_insert(node->keys, sizeof(*node->keys), nr_entries, index, &key_le); node 102 drivers/md/persistent-data/dm-btree.c array_insert(value_base(node), value_size, nr_entries, index, value); node 103 drivers/md/persistent-data/dm-btree.c node->header.nr_entries = cpu_to_le32(nr_entries + 1); node 424 drivers/md/persistent-data/dm-btree.c struct dm_block *node; node 427 drivers/md/persistent-data/dm-btree.c r = bn_read_lock(info, root, &node); node 431 drivers/md/persistent-data/dm-btree.c n = dm_block_data(node); node 466 drivers/md/persistent-data/dm-btree.c dm_tm_unlock(info->tm, node); node 697 drivers/md/persistent-data/dm-btree.c struct btree_node *node; node 704 drivers/md/persistent-data/dm-btree.c node = dm_block_data(shadow_current(s)); node 719 drivers/md/persistent-data/dm-btree.c node = dm_block_data(shadow_current(s)); node 721 drivers/md/persistent-data/dm-btree.c if (node->header.nr_entries == node->header.max_entries) { node 731 drivers/md/persistent-data/dm-btree.c node = dm_block_data(shadow_current(s)); node 733 drivers/md/persistent-data/dm-btree.c i = lower_bound(node, key); node 735 drivers/md/persistent-data/dm-btree.c if (le32_to_cpu(node->header.flags) & LEAF_NODE) node 740 drivers/md/persistent-data/dm-btree.c node->keys[0] = cpu_to_le64(key); node 744 drivers/md/persistent-data/dm-btree.c root = value64(node, i); node 748 drivers/md/persistent-data/dm-btree.c if (i < 0 || le64_to_cpu(node->keys[i]) != key) node 755 drivers/md/persistent-data/dm-btree.c static bool need_insert(struct btree_node *node, uint64_t *keys, node 758 drivers/md/persistent-data/dm-btree.c return ((index >= le32_to_cpu(node->header.nr_entries)) || node 759 drivers/md/persistent-data/dm-btree.c (le64_to_cpu(node->keys[index]) != keys[level])); node 955 drivers/md/persistent-data/dm-btree.c struct dm_block *node; node 959 drivers/md/persistent-data/dm-btree.c r = bn_read_lock(info, block, &node); node 963 drivers/md/persistent-data/dm-btree.c n = dm_block_data(node); node 980 drivers/md/persistent-data/dm-btree.c dm_tm_unlock(info->tm, node); node 6777 drivers/md/raid5.c static int raid456_cpu_dead(unsigned int cpu, struct hlist_node *node) node 6779 drivers/md/raid5.c struct r5conf *conf = hlist_entry_safe(node, struct r5conf, node); node 6790 drivers/md/raid5.c cpuhp_state_remove_instance(CPUHP_MD_RAID5_PREPARE, &conf->node); node 6814 drivers/md/raid5.c static int raid456_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) node 6816 drivers/md/raid5.c struct r5conf *conf = hlist_entry_safe(node, struct r5conf, node); node 6835 drivers/md/raid5.c err = cpuhp_state_add_instance(CPUHP_MD_RAID5_PREPARE, &conf->node); node 646 drivers/md/raid5.h struct hlist_node node; node 21 drivers/media/cec/cec-priv.h #define to_cec_adapter(node) container_of(node, struct cec_adapter, devnode) node 413 drivers/media/i2c/adp1653.c struct device_node *node) node 423 drivers/media/i2c/adp1653.c child = of_get_child_by_name(node, "flash"); node 444 drivers/media/i2c/adp1653.c child = of_get_child_by_name(node, "indicator"); node 1573 drivers/media/i2c/s5c73m3/s5c73m3-core.c struct device_node *node = dev->of_node; node 1579 drivers/media/i2c/s5c73m3/s5c73m3-core.c ret = of_get_named_gpio_flags(node, prop_names[i], node 1596 drivers/media/i2c/s5c73m3/s5c73m3-core.c struct device_node *node = dev->of_node; node 1601 drivers/media/i2c/s5c73m3/s5c73m3-core.c if (!node) { node 1617 drivers/media/i2c/s5c73m3/s5c73m3-core.c if (of_property_read_u32(node, "clock-frequency", node 1628 drivers/media/i2c/s5c73m3/s5c73m3-core.c node_ep = of_graph_get_next_endpoint(node, NULL); node 1630 drivers/media/i2c/s5c73m3/s5c73m3-core.c dev_warn(dev, "no endpoint defined for node: %pOF\n", node); node 1820 drivers/media/i2c/s5k5baf.c struct device_node *node = dev->of_node; node 1825 drivers/media/i2c/s5k5baf.c ret = of_get_named_gpio_flags(node, names[i], 0, &flags); node 1839 drivers/media/i2c/s5k5baf.c struct device_node *node = dev->of_node; node 1844 drivers/media/i2c/s5k5baf.c if (!node) { node 1849 drivers/media/i2c/s5k5baf.c ret = of_property_read_u32(node, "clock-frequency", node 1861 drivers/media/i2c/s5k5baf.c node_ep = of_graph_get_next_endpoint(node, NULL); node 1863 drivers/media/i2c/s5k5baf.c dev_err(dev, "no endpoint defined at node %pOF\n", node); node 1882 drivers/media/i2c/s5k5baf.c node); node 1199 drivers/media/pci/cobalt/cobalt-v4l2.c static int cobalt_node_register(struct cobalt *cobalt, int node) node 1203 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_stream *s = cobalt->streams + node; node 1212 drivers/media/pci/cobalt/cobalt-v4l2.c "%s-%d", cobalt->v4l2_dev.name, node); node 1232 drivers/media/pci/cobalt/cobalt-v4l2.c cobalt_warn("Setting up dummy video node %d\n", node); node 1282 drivers/media/pci/cobalt/cobalt-v4l2.c node); node 1285 drivers/media/pci/cobalt/cobalt-v4l2.c cobalt_info("registered node %d\n", node); node 1292 drivers/media/pci/cobalt/cobalt-v4l2.c int node, ret; node 1295 drivers/media/pci/cobalt/cobalt-v4l2.c for (node = 0; node < COBALT_NUM_STREAMS; node++) { node 1296 drivers/media/pci/cobalt/cobalt-v4l2.c ret = cobalt_node_register(cobalt, node); node 1306 drivers/media/pci/cobalt/cobalt-v4l2.c int node; node 1309 drivers/media/pci/cobalt/cobalt-v4l2.c for (node = 0; node < COBALT_NUM_STREAMS; node++) { node 1310 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_stream *s = cobalt->streams + node; node 1050 drivers/media/pci/ivtv/ivtvfb.c oi->ivtvfb_info.node = -1; node 350 drivers/media/pci/sta2x11/sta2x11_vip.c struct vip_buffer *vip_buf, *node; node 359 drivers/media/pci/sta2x11/sta2x11_vip.c list_for_each_entry_safe(vip_buf, node, &vip->buffer_list, list) { node 71 drivers/media/platform/atmel/atmel-isi.c struct device_node *node; node 422 drivers/media/platform/atmel/atmel-isi.c struct frame_buffer *buf, *node; node 464 drivers/media/platform/atmel/atmel-isi.c list_for_each_entry_safe(buf, node, &isi->video_buffer_list, list) { node 477 drivers/media/platform/atmel/atmel-isi.c struct frame_buffer *buf, *node; node 489 drivers/media/platform/atmel/atmel-isi.c list_for_each_entry_safe(buf, node, &isi->video_buffer_list, list) { node 1101 drivers/media/platform/atmel/atmel-isi.c static int isi_graph_parse(struct atmel_isi *isi, struct device_node *node) node 1106 drivers/media/platform/atmel/atmel-isi.c ep = of_graph_get_next_endpoint(node, ep); node 1116 drivers/media/platform/atmel/atmel-isi.c isi->entity.node = remote; node 1137 drivers/media/platform/atmel/atmel-isi.c of_node_put(isi->entity.node); node 867 drivers/media/platform/exynos4-is/fimc-core.c struct device_node *node = dev->of_node; node 874 drivers/media/platform/exynos4-is/fimc-core.c if (of_property_read_bool(node, "samsung,lcd-wb")) node 881 drivers/media/platform/exynos4-is/fimc-core.c of_id = of_match_node(fimc_of_match, node); node 885 drivers/media/platform/exynos4-is/fimc-core.c ret = of_property_read_u32_array(node, "samsung,pix-limits", node 898 drivers/media/platform/exynos4-is/fimc-core.c ret = of_property_read_u32_array(node, "samsung,min-pix-sizes", node 902 drivers/media/platform/exynos4-is/fimc-core.c ret = of_property_read_u32_array(node, "samsung,min-pix-alignment", node 907 drivers/media/platform/exynos4-is/fimc-core.c ret = of_property_read_u32(node, "samsung,rotators", &args[1]); node 910 drivers/media/platform/exynos4-is/fimc-core.c v->has_mainscaler_ext = of_property_read_bool(node, node 913 drivers/media/platform/exynos4-is/fimc-core.c v->has_isp_wb = of_property_read_bool(node, "samsung,isp-wb"); node 914 drivers/media/platform/exynos4-is/fimc-core.c v->has_cam_if = of_property_read_bool(node, "samsung,cam-if"); node 915 drivers/media/platform/exynos4-is/fimc-core.c of_property_read_u32(node, "clock-frequency", clk_freq); node 916 drivers/media/platform/exynos4-is/fimc-core.c fimc->id = of_alias_get_id(node, "fimc"); node 646 drivers/media/platform/exynos4-is/fimc-core.h static inline struct regmap * fimc_get_sysreg_regmap(struct device_node *node) node 648 drivers/media/platform/exynos4-is/fimc-core.h return syscon_regmap_lookup_by_phandle(node, "samsung,sysreg"); node 651 drivers/media/platform/exynos4-is/fimc-core.h #define fimc_get_sysreg_regmap(node) (NULL) node 39 drivers/media/platform/exynos4-is/fimc-is-i2c.c struct device_node *node = pdev->dev.of_node; node 55 drivers/media/platform/exynos4-is/fimc-is-i2c.c i2c_adap->dev.of_node = node; node 25 drivers/media/platform/exynos4-is/fimc-is-sensor.c struct device_node *node) node 29 drivers/media/platform/exynos4-is/fimc-is-sensor.c of_id = of_match_node(fimc_is_sensor_of_ids, node); node 51 drivers/media/platform/exynos4-is/fimc-is-sensor.h struct device_node *node); node 165 drivers/media/platform/exynos4-is/fimc-is.c struct device_node *node) node 172 drivers/media/platform/exynos4-is/fimc-is.c sensor->drvdata = fimc_is_sensor_get_drvdata(node); node 175 drivers/media/platform/exynos4-is/fimc-is.c node); node 179 drivers/media/platform/exynos4-is/fimc-is.c ep = of_graph_get_next_endpoint(node, NULL); node 781 drivers/media/platform/exynos4-is/fimc-is.c struct device_node *node; node 803 drivers/media/platform/exynos4-is/fimc-is.c node = of_get_child_by_name(dev->of_node, "pmu"); node 804 drivers/media/platform/exynos4-is/fimc-is.c if (!node) node 807 drivers/media/platform/exynos4-is/fimc-is.c is->pmu_regs = of_iomap(node, 0); node 808 drivers/media/platform/exynos4-is/fimc-is.c of_node_put(node); node 475 drivers/media/platform/exynos4-is/media-dev.c struct device_node *node; node 493 drivers/media/platform/exynos4-is/media-dev.c for_each_available_child_of_node(parent, node) { node 496 drivers/media/platform/exynos4-is/media-dev.c if (!of_node_name_eq(node, "csis")) node 499 drivers/media/platform/exynos4-is/media-dev.c port = of_get_next_child(node, NULL); node 506 drivers/media/platform/exynos4-is/media-dev.c of_node_put(node); node 517 drivers/media/platform/exynos4-is/media-dev.c for_each_child_of_node(ports, node) { node 518 drivers/media/platform/exynos4-is/media-dev.c ret = fimc_md_parse_port_node(fmd, node, index); node 520 drivers/media/platform/exynos4-is/media-dev.c of_node_put(node); node 617 drivers/media/platform/exynos4-is/media-dev.c struct device_node *node = pdev->dev.of_node; node 620 drivers/media/platform/exynos4-is/media-dev.c id = node ? __of_get_csis_id(node) : max(0, pdev->id); node 713 drivers/media/platform/exynos4-is/media-dev.c struct device_node *node; node 716 drivers/media/platform/exynos4-is/media-dev.c for_each_available_child_of_node(parent, node) { node 720 drivers/media/platform/exynos4-is/media-dev.c pdev = of_find_device_by_node(node); node 725 drivers/media/platform/exynos4-is/media-dev.c if (of_node_name_eq(node, CSIS_OF_NODE_NAME)) node 727 drivers/media/platform/exynos4-is/media-dev.c else if (of_node_name_eq(node, FIMC_IS_OF_NODE_NAME)) node 729 drivers/media/platform/exynos4-is/media-dev.c else if (of_node_name_eq(node, FIMC_LITE_OF_NODE_NAME)) node 731 drivers/media/platform/exynos4-is/media-dev.c else if (of_node_name_eq(node, FIMC_OF_NODE_NAME) && node 732 drivers/media/platform/exynos4-is/media-dev.c !of_property_read_bool(node, "samsung,lcd-wb")) node 740 drivers/media/platform/exynos4-is/media-dev.c of_node_put(node); node 186 drivers/media/platform/exynos4-is/media-dev.h static inline bool fimc_md_is_isp_available(struct device_node *node) node 188 drivers/media/platform/exynos4-is/media-dev.h node = of_get_child_by_name(node, FIMC_IS_OF_NODE_NAME); node 189 drivers/media/platform/exynos4-is/media-dev.h return node ? of_device_is_available(node) : false; node 192 drivers/media/platform/exynos4-is/media-dev.h #define fimc_md_is_isp_available(node) (false) node 717 drivers/media/platform/exynos4-is/mipi-csis.c struct device_node *node = pdev->dev.of_node; node 721 drivers/media/platform/exynos4-is/mipi-csis.c if (of_property_read_u32(node, "clock-frequency", node 724 drivers/media/platform/exynos4-is/mipi-csis.c if (of_property_read_u32(node, "bus-width", node 728 drivers/media/platform/exynos4-is/mipi-csis.c node = of_graph_get_next_endpoint(node, NULL); node 729 drivers/media/platform/exynos4-is/mipi-csis.c if (!node) { node 735 drivers/media/platform/exynos4-is/mipi-csis.c ret = v4l2_fwnode_endpoint_parse(of_fwnode_handle(node), &endpoint); node 746 drivers/media/platform/exynos4-is/mipi-csis.c of_property_read_u32(node, "samsung,csis-hs-settle", node 748 drivers/media/platform/exynos4-is/mipi-csis.c state->wclk_ext = of_property_read_bool(node, node 754 drivers/media/platform/exynos4-is/mipi-csis.c of_node_put(node); node 1134 drivers/media/platform/marvell-ccic/mcam-core.c struct mcam_vb_buffer *buf, *node; node 1139 drivers/media/platform/marvell-ccic/mcam-core.c list_for_each_entry_safe(buf, node, &cam->buffers, queue) { node 1054 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c struct device_node *node; node 1057 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c node = of_parse_phandle(jpeg->dev->of_node, "mediatek,larb", 0); node 1058 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c if (!node) node 1060 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c pdev = of_find_device_by_node(node); node 1062 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c of_node_put(node); node 1065 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c of_node_put(node); node 40 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c int mtk_mdp_comp_get_id(struct device *dev, struct device_node *node, node 43 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c int id = of_alias_get_id(node, mtk_mdp_comp_stem[comp_type]); node 94 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c int mtk_mdp_comp_init(struct device *dev, struct device_node *node, node 106 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c comp->dev_node = of_node_get(node); node 109 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c comp->regs = of_iomap(node, 0); node 112 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c comp->clk[i] = of_clk_get(node, i); node 126 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c larb_node = of_parse_phandle(node, "mediatek,larb", 0); node 129 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c "Missing mediadek,larb phandle in %pOF node\n", node); node 55 drivers/media/platform/mtk-mdp/mtk_mdp_comp.h int mtk_mdp_comp_init(struct device *dev, struct device_node *node, node 58 drivers/media/platform/mtk-mdp/mtk_mdp_comp.h int mtk_mdp_comp_get_id(struct device *dev, struct device_node *node, node 98 drivers/media/platform/mtk-mdp/mtk_mdp_core.c struct device_node *node, *parent; node 113 drivers/media/platform/mtk-mdp/mtk_mdp_core.c node = of_get_next_child(dev->of_node, NULL); node 114 drivers/media/platform/mtk-mdp/mtk_mdp_core.c if (node) { node 115 drivers/media/platform/mtk-mdp/mtk_mdp_core.c of_node_put(node); node 123 drivers/media/platform/mtk-mdp/mtk_mdp_core.c for_each_child_of_node(parent, node) { node 129 drivers/media/platform/mtk-mdp/mtk_mdp_core.c of_id = of_match_node(mtk_mdp_comp_dt_ids, node); node 133 drivers/media/platform/mtk-mdp/mtk_mdp_core.c if (!of_device_is_available(node)) { node 135 drivers/media/platform/mtk-mdp/mtk_mdp_core.c node); node 140 drivers/media/platform/mtk-mdp/mtk_mdp_core.c comp_id = mtk_mdp_comp_get_id(dev, node, comp_type); node 143 drivers/media/platform/mtk-mdp/mtk_mdp_core.c node); node 150 drivers/media/platform/mtk-mdp/mtk_mdp_core.c of_node_put(node); node 155 drivers/media/platform/mtk-mdp/mtk_mdp_core.c ret = mtk_mdp_comp_init(dev, node, comp, comp_id); node 157 drivers/media/platform/mtk-mdp/mtk_mdp_core.c of_node_put(node); node 19 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c struct device_node *node; node 30 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c node = of_parse_phandle(pdev->dev.of_node, "mediatek,larb", 0); node 31 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c if (!node) { node 36 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c pdev = of_find_device_by_node(node); node 37 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c of_node_put(node); node 20 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c struct device_node *node; node 36 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c node = of_parse_phandle(dev->of_node, "mediatek,larb", 0); node 37 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c if (!node) { node 41 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c pdev = of_find_device_by_node(node); node 42 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c of_node_put(node); node 49 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c node = of_parse_phandle(dev->of_node, "mediatek,larb", 1); node 50 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c if (!node) { node 55 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c pdev = of_find_device_by_node(node); node 56 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c of_node_put(node); node 295 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb_node *node; node 302 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_for_each_entry(node, &inst->fb_use_list, list) { node 303 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb *fb = (struct vdec_fb *)node->fb; node 306 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, node 314 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node = list_first_entry(&inst->available_fb_node_list, node 316 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node->fb = inst->cur_fb; node 317 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->fb_use_list); node 321 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node = list_first_entry(&inst->available_fb_node_list, node 323 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node->fb = inst->cur_fb; node 324 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->fb_disp_list); node 330 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb_node *node, *tmp; node 332 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_for_each_entry_safe(node, tmp, &inst->fb_use_list, list) node 333 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->fb_free_list); node 355 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb_node *node; node 358 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node = list_first_entry(&inst->available_fb_node_list, node 360 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node->fb = fb; node 361 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->fb_free_list); node 518 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb_node *node; node 521 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node = list_first_entry_or_null(&inst->fb_disp_list, node 523 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c if (node) { node 524 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->available_fb_node_list); node 525 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c fb = (struct vdec_fb *)node->fb; node 528 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node->fb, fb->status); node 539 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c struct vdec_fb_node *node; node 542 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node = list_first_entry_or_null(&inst->fb_free_list, node 544 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c if (node) { node 545 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c list_move_tail(&node->list, &inst->available_fb_node_list); node 546 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c fb = (struct vdec_fb *)node->fb; node 549 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c node->fb, fb->status); node 218 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 220 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_for_each_entry(node, &inst->fb_use_list, list) { node 221 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c fb = (struct vdec_fb *)node->fb; node 223 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, node 234 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 237 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node = list_first_entry_or_null(&inst->available_fb_node_list, node 240 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (node) { node 241 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node->fb = fb; node 242 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->fb_free_list); node 432 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 439 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node = list_first_entry_or_null(&inst->available_fb_node_list, node 441 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (node) { node 442 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node->fb = fb; node 443 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->fb_disp_list); node 614 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 617 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node = list_first_entry_or_null(&inst->fb_disp_list, node 619 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (node) { node 620 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c fb = (struct vdec_fb *)node->fb; node 622 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->available_fb_node_list); node 624 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node->fb, fb->status); node 634 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 641 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node = list_first_entry_or_null(&inst->available_fb_node_list, node 643 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (node) { node 644 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node->fb = fb; node 645 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->fb_use_list); node 655 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node, *tmp; node 657 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_for_each_entry_safe(node, tmp, &inst->fb_use_list, list) node 658 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->fb_free_list); node 722 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c struct vdec_fb_node *node; node 725 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node = list_first_entry_or_null(&inst->fb_free_list, node 727 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (node) { node 728 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c list_move_tail(&node->list, &inst->available_fb_node_list); node 729 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c fb = (struct vdec_fb *)node->fb; node 732 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c node->fb, fb->status); node 427 drivers/media/platform/qcom/camss/camss.c struct device_node *node, node 435 drivers/media/platform/qcom/camss/camss.c v4l2_fwnode_endpoint_parse(of_fwnode_handle(node), &vep); node 468 drivers/media/platform/qcom/camss/camss.c struct device_node *node = NULL; node 472 drivers/media/platform/qcom/camss/camss.c for_each_endpoint_of_node(dev->of_node, node) { node 476 drivers/media/platform/qcom/camss/camss.c if (!of_device_is_available(node)) node 479 drivers/media/platform/qcom/camss/camss.c remote = of_graph_get_remote_port_parent(node); node 497 drivers/media/platform/qcom/camss/camss.c ret = camss_of_parse_endpoint_node(dev, node, csd); node 508 drivers/media/platform/qcom/camss/camss.c of_node_put(node); node 68 drivers/media/platform/qcom/venus/firmware.c struct device_node *node; node 79 drivers/media/platform/qcom/venus/firmware.c node = of_parse_phandle(dev->of_node, "memory-region", 0); node 80 drivers/media/platform/qcom/venus/firmware.c if (!node) { node 85 drivers/media/platform/qcom/venus/firmware.c ret = of_address_to_resource(node, 0, &r); node 126 drivers/media/platform/qcom/venus/firmware.c of_node_put(node); node 976 drivers/media/platform/rcar-vin/rcar-dma.c struct rvin_buffer *buf, *node; node 987 drivers/media/platform/rcar-vin/rcar-dma.c list_for_each_entry_safe(buf, node, &vin->buf_list, list) { node 290 drivers/media/platform/sh_vou.c struct sh_vou_buffer *buf, *node; node 297 drivers/media/platform/sh_vou.c list_for_each_entry_safe(buf, node, &vou_dev->buf_list, list) { node 338 drivers/media/platform/sh_vou.c struct sh_vou_buffer *buf, *node; node 349 drivers/media/platform/sh_vou.c list_for_each_entry_safe(buf, node, &vou_dev->buf_list, list) { node 322 drivers/media/platform/sti/bdisp/bdisp-debug.c struct bdisp_node *node; node 331 drivers/media/platform/sti/bdisp/bdisp-debug.c node = bdisp->dbg.copy_node[i]; node 332 drivers/media/platform/sti/bdisp/bdisp-debug.c if (!node) node 336 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "NIP\t0x%08X\n", node->nip); node 337 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "CIC\t0x%08X\n", node->cic); node 338 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_ins(s, node->ins); node 339 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "ACK\t0x%08X\n", node->ack); node 341 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "TBA\t0x%08X\n", node->tba); node 342 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_tty(s, node->tty); node 343 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_xy(s, node->txy, "TXY"); node 344 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sz(s, node->tsz, "TSZ"); node 347 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "S1BA\t0x%08X\n", node->s1ba); node 348 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sty(s, node->s1ty, node->s1ba, "S1TY"); node 349 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_xy(s, node->s1xy, "S1XY"); node 351 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "S2BA\t0x%08X\n", node->s2ba); node 352 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sty(s, node->s2ty, node->s2ba, "S2TY"); node 353 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_xy(s, node->s2xy, "S2XY"); node 354 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sz(s, node->s2sz, "S2SZ"); node 356 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "S3BA\t0x%08X\n", node->s3ba); node 357 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sty(s, node->s3ty, node->s3ba, "S3TY"); node 358 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_xy(s, node->s3xy, "S3XY"); node 359 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_sz(s, node->s3sz, "S3SZ"); node 363 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_fctl(s, node->fctl); node 366 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_rsf(s, node->rsf, "RSF"); node 367 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_rzi(s, node->rzi, "RZI"); node 368 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "HFP\t0x%08X\n", node->hfp); node 369 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "VFP\t0x%08X\n", node->vfp); node 371 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_rsf(s, node->y_rsf, "Y_RSF"); node 372 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_rzi(s, node->y_rzi, "Y_RZI"); node 373 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "Y_HFP\t0x%08X\n", node->y_hfp); node 374 drivers/media/platform/sti/bdisp/bdisp-debug.c seq_printf(s, "Y_VFP\t0x%08X\n", node->y_vfp); node 380 drivers/media/platform/sti/bdisp/bdisp-debug.c bdisp_dbg_dump_ivmx(s, node->ivmx0, node->ivmx1, node 381 drivers/media/platform/sti/bdisp/bdisp-debug.c node->ivmx2, node->ivmx3); node 386 drivers/media/platform/sti/bdisp/bdisp-debug.c } while ((++i < MAX_NB_NODE) && node->nip); node 394 drivers/media/platform/sti/bdisp/bdisp-debug.c struct bdisp_node *node; node 404 drivers/media/platform/sti/bdisp/bdisp-debug.c node = bdisp->dbg.copy_node[i]; node 405 drivers/media/platform/sti/bdisp/bdisp-debug.c if (!node) node 409 drivers/media/platform/sti/bdisp/bdisp-debug.c val = (u32 *)node; node 412 drivers/media/platform/sti/bdisp/bdisp-debug.c } while ((++i < MAX_NB_NODE) && node->nip); node 433 drivers/media/platform/sti/bdisp/bdisp-hw.c if (ctx && ctx->node[0]) node 436 drivers/media/platform/sti/bdisp/bdisp-hw.c ctx->node[0], ctx->node_paddr[0], node 467 drivers/media/platform/sti/bdisp/bdisp-hw.c ctx->node[i] = base; node 469 drivers/media/platform/sti/bdisp/bdisp-hw.c dev_dbg(dev, "node[%d]=0x%p (paddr=%pad)\n", i, ctx->node[i], node 759 drivers/media/platform/sti/bdisp/bdisp-hw.c struct bdisp_node *node, node 774 drivers/media/platform/sti/bdisp/bdisp-hw.c memset(node, 0, sizeof(*node)); node 789 drivers/media/platform/sti/bdisp/bdisp-hw.c node->nip = 0; node 790 drivers/media/platform/sti/bdisp/bdisp-hw.c node->cic = BLT_CIC_ALL_GRP; node 791 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ack = BLT_ACK_BYPASS_S2S3; node 796 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins = BLT_INS_S1_OFF | BLT_INS_S2_MEM | BLT_INS_S3_OFF; node 802 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins = BLT_INS_S1_OFF | BLT_INS_S3_MEM; node 804 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= BLT_INS_S2_CF; node 806 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= BLT_INS_S2_MEM; node 813 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins = BLT_INS_S3_MEM; node 815 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= BLT_INS_S2_CF | BLT_INS_S1_CF; node 817 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= BLT_INS_S2_MEM | BLT_INS_S1_MEM; node 822 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= cfg->cconv ? BLT_INS_IVMX : 0; node 824 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ins |= (cfg->scale || cfg->src_420 || cfg->dst_420) ? node 828 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tba = (t_plan == BDISP_CBCR) ? dst->paddr[1] : dst->paddr[0]; node 830 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty = dst->bytesperline; node 831 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty |= bdisp_hw_color_format(dst_fmt); node 832 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty |= BLT_TTY_DITHER; node 833 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty |= (t_plan == BDISP_CBCR) ? BLT_TTY_CHROMA : 0; node 834 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty |= cfg->hflip ? BLT_TTY_HSO : 0; node 835 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tty |= cfg->vflip ? BLT_TTY_VSO : 0; node 847 drivers/media/platform/sti/bdisp/bdisp-hw.c node->txy = cfg->vflip ? (dst_rect.height - 1) : dst_rect.top; node 848 drivers/media/platform/sti/bdisp/bdisp-hw.c node->txy <<= 16; node 849 drivers/media/platform/sti/bdisp/bdisp-hw.c node->txy |= cfg->hflip ? (dst_width - dst_x_offset - 1) : node 852 drivers/media/platform/sti/bdisp/bdisp-hw.c node->tsz = dst_rect.height << 16 | dst_rect.width; node 862 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ba = src->paddr[0]; node 864 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty = src->bytesperline; node 866 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty *= 2; node 868 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty |= bdisp_hw_color_format(src_fmt); node 870 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2xy = src_rect.top << 16 | src_rect.left; node 871 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2sz = src_rect.height << 16 | src_rect.width; node 882 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ba = src->paddr[1]; node 884 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty = src->bytesperline; node 886 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty /= 2; node 888 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty *= 2; node 890 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2ty |= bdisp_hw_color_format(src_fmt); node 892 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2xy = src_rect.top << 16 | src_rect.left; node 893 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s2sz = src_rect.height << 16 | src_rect.width; node 897 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s1ba = src->paddr[2]; node 899 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s1ty = node->s2ty; node 900 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s1xy = node->s2xy; node 904 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3ba = src->paddr[0]; node 906 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3ty = src->bytesperline; node 908 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3ty *= 2; node 909 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3ty |= bdisp_hw_color_format(src_fmt); node 913 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3xy = node->s2xy * 2; node 914 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3sz = node->s2sz * 2; node 917 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3ty |= BLT_S3TY_BLANK_ACC; node 918 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3xy = node->s2xy; node 919 drivers/media/platform/sti/bdisp/bdisp-hw.c node->s3sz = node->s2sz; node 924 drivers/media/platform/sti/bdisp/bdisp-hw.c if (node->ins & BLT_INS_SCALE) { node 930 drivers/media/platform/sti/bdisp/bdisp-hw.c node->fctl = BLT_FCTL_HV_SCALE; node 932 drivers/media/platform/sti/bdisp/bdisp-hw.c node->fctl |= BLT_FCTL_Y_HV_SCALE; node 934 drivers/media/platform/sti/bdisp/bdisp-hw.c node->fctl = BLT_FCTL_HV_SAMPLE; node 936 drivers/media/platform/sti/bdisp/bdisp-hw.c node->fctl |= BLT_FCTL_Y_HV_SAMPLE; node 951 drivers/media/platform/sti/bdisp/bdisp-hw.c node->rsf = v_inc << 16 | h_inc; node 954 drivers/media/platform/sti/bdisp/bdisp-hw.c node->rzi = BLT_RZI_DEFAULT; node 957 drivers/media/platform/sti/bdisp/bdisp-hw.c node->hfp = bdisp_hw_get_hf_addr(h_inc); node 958 drivers/media/platform/sti/bdisp/bdisp-hw.c node->vfp = bdisp_hw_get_vf_addr(v_inc); node 965 drivers/media/platform/sti/bdisp/bdisp-hw.c node->y_rsf = yv_inc << 16 | yh_inc; node 966 drivers/media/platform/sti/bdisp/bdisp-hw.c node->y_rzi = BLT_RZI_DEFAULT; node 967 drivers/media/platform/sti/bdisp/bdisp-hw.c node->y_hfp = bdisp_hw_get_hf_addr(yh_inc); node 968 drivers/media/platform/sti/bdisp/bdisp-hw.c node->y_vfp = bdisp_hw_get_vf_addr(yv_inc); node 976 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ivmx0 = ivmx[0]; node 977 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ivmx1 = ivmx[1]; node 978 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ivmx2 = ivmx[2]; node 979 drivers/media/platform/sti/bdisp/bdisp-hw.c node->ivmx3 = ivmx[3]; node 999 drivers/media/platform/sti/bdisp/bdisp-hw.c if (!ctx->node[i]) { node 1011 drivers/media/platform/sti/bdisp/bdisp-hw.c bdisp_hw_build_node(ctx, &cfg, ctx->node[nid], node 1015 drivers/media/platform/sti/bdisp/bdisp-hw.c ctx->node[nid - 1]->nip = ctx->node_paddr[nid]; node 1020 drivers/media/platform/sti/bdisp/bdisp-hw.c bdisp_hw_build_node(ctx, &cfg, ctx->node[nid], node 1022 drivers/media/platform/sti/bdisp/bdisp-hw.c ctx->node[nid - 1]->nip = ctx->node_paddr[nid]; node 1033 drivers/media/platform/sti/bdisp/bdisp-hw.c ctx->node[nid - 1]->nip = 0; node 1051 drivers/media/platform/sti/bdisp/bdisp-hw.c struct bdisp_node **node = ctx->node; node 1071 drivers/media/platform/sti/bdisp/bdisp-hw.c *copy_node[i] = *node[i]; node 1112 drivers/media/platform/sti/bdisp/bdisp-hw.c if (!ctx->node[node_id]->nip) node 120 drivers/media/platform/sti/bdisp/bdisp.h struct bdisp_node *node[MAX_NB_NODE]; node 728 drivers/media/platform/stm32/stm32-dcmi.c struct dcmi_buf *buf, *node; node 847 drivers/media/platform/stm32/stm32-dcmi.c list_for_each_entry_safe(buf, node, &dcmi->buffers, list) { node 860 drivers/media/platform/stm32/stm32-dcmi.c struct dcmi_buf *buf, *node; node 875 drivers/media/platform/stm32/stm32-dcmi.c list_for_each_entry_safe(buf, node, &dcmi->buffers, list) { node 1778 drivers/media/platform/stm32/stm32-dcmi.c static int dcmi_graph_parse(struct stm32_dcmi *dcmi, struct device_node *node) node 1783 drivers/media/platform/stm32/stm32-dcmi.c ep = of_graph_get_next_endpoint(node, ep); node 207 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c struct sun4i_csi_buffer *buf, *node; node 210 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c list_for_each_entry_safe(buf, node, &csi->buf_list, list) { node 222 drivers/media/platform/vimc/vimc-capture.c struct vimc_cap_buffer *vbuf, *node; node 226 drivers/media/platform/vimc/vimc-capture.c list_for_each_entry_safe(vbuf, node, &vcap->buf_list, list) { node 1228 drivers/media/platform/vivid/vivid-core.c dev->fb_info.node); node 1625 drivers/media/platform/vivid/vivid-core.c dev->fb_info.node); node 312 drivers/media/platform/vivid/vivid-osd.c dev->fb_info.node = -1; node 711 drivers/media/platform/xilinx/xilinx-tpg.c struct device_node *node = xtpg->xvip.dev->of_node; node 717 drivers/media/platform/xilinx/xilinx-tpg.c ports = of_get_child_by_name(node, "ports"); node 719 drivers/media/platform/xilinx/xilinx-tpg.c ports = node; node 102 drivers/media/platform/xilinx/xilinx-vip.c const struct xvip_video_format *xvip_of_get_format(struct device_node *node) node 110 drivers/media/platform/xilinx/xilinx-vip.c ret = of_property_read_u32(node, "xlnx,video-format", &vf_code); node 114 drivers/media/platform/xilinx/xilinx-vip.c ret = of_property_read_u32(node, "xlnx,video-width", &width); node 119 drivers/media/platform/xilinx/xilinx-vip.c of_property_read_string(node, "xlnx,cfa-pattern", &pattern); node 124 drivers/media/platform/xilinx/xilinx-vip.h const struct xvip_video_format *xvip_of_get_format(struct device_node *node); node 188 drivers/media/platform/xilinx/xilinx-vipp.c struct device_node *node = xdev->dev->of_node; node 203 drivers/media/platform/xilinx/xilinx-vipp.c ep = of_graph_get_next_endpoint(node, ep); node 432 drivers/media/platform/xilinx/xilinx-vipp.c struct device_node *node) node 440 drivers/media/platform/xilinx/xilinx-vipp.c ret = of_property_read_string(node, "direction", &direction); node 451 drivers/media/platform/xilinx/xilinx-vipp.c of_property_read_u32(node, "reg", &index); node 459 drivers/media/platform/xilinx/xilinx-vipp.c dev_err(xdev->dev, "%pOF initialization failed\n", node); node 311 drivers/media/platform/xilinx/xilinx-vtc.c struct device_node *node = xvtc->xvip.dev->of_node; node 313 drivers/media/platform/xilinx/xilinx-vtc.c xvtc->has_detector = of_property_read_bool(node, "xlnx,detector"); node 314 drivers/media/platform/xilinx/xilinx-vtc.c xvtc->has_generator = of_property_read_bool(node, "xlnx,generator"); node 214 drivers/media/rc/ir-hix5hd2.c struct device_node *node = pdev->dev.of_node; node 222 drivers/media/rc/ir-hix5hd2.c priv->regmap = syscon_regmap_lookup_by_phandle(node, node 258 drivers/media/rc/ir-hix5hd2.c map_name = of_get_property(node, "linux,rc-map-name", NULL); node 104 drivers/media/rc/meson-ir.c struct device_node *node = dev->of_node; node 133 drivers/media/rc/meson-ir.c map_name = of_get_property(node, "linux,rc-map-name", NULL); node 162 drivers/media/rc/meson-ir.c if (of_device_is_compatible(node, "amlogic,meson6-ir")) node 198 drivers/media/rc/meson-ir.c struct device_node *node = dev->of_node; node 208 drivers/media/rc/meson-ir.c if (of_device_is_compatible(node, "amlogic,meson6-ir")) node 1340 drivers/media/usb/gspca/gspca.c struct gspca_buffer *buf, *node; node 1344 drivers/media/usb/gspca/gspca.c list_for_each_entry_safe(buf, node, &gspca_dev->buf_list, list) { node 731 drivers/media/usb/hackrf/hackrf.c struct hackrf_buffer *buffer, *node; node 743 drivers/media/usb/hackrf/hackrf.c list_for_each_entry_safe(buffer, node, buffer_list, list) { node 1083 drivers/media/usb/s2255/s2255drv.c struct s2255_buffer *buf, *node; node 1087 drivers/media/usb/s2255/s2255drv.c list_for_each_entry_safe(buf, node, &vc->buf_list, list) { node 1243 drivers/media/usb/uvc/uvc_ctrl.c list_for_each_entry(sev, &mapping->ev_subs, node) { node 1419 drivers/media/usb/uvc/uvc_ctrl.c list_add_tail(&sev->node, &mapping->ev_subs); node 1446 drivers/media/usb/uvc/uvc_ctrl.c list_del(&sev->node); node 1494 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(sev, &ctrl->ev_subs, node) node 2111 drivers/media/v4l2-core/v4l2-ctrls.c static inline u32 node2id(struct list_head *node) node 2113 drivers/media/v4l2-core/v4l2-ctrls.c return list_entry(node, struct v4l2_ctrl_ref, node)->ctrl->id; node 2167 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry_safe(ref, next_ref, &hdl->ctrl_refs, node) { node 2168 drivers/media/v4l2-core/v4l2-ctrls.c list_del(&ref->node); node 2172 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry_safe(ctrl, next_ctrl, &hdl->ctrls, node) { node 2173 drivers/media/v4l2-core/v4l2-ctrls.c list_del(&ctrl->node); node 2174 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry_safe(sev, next_sev, &ctrl->ev_subs, node) node 2175 drivers/media/v4l2-core/v4l2-ctrls.c list_del(&sev->node); node 2199 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) { node 2302 drivers/media/v4l2-core/v4l2-ctrls.c INIT_LIST_HEAD(&new_ref->node); node 2311 drivers/media/v4l2-core/v4l2-ctrls.c list_add_tail(&new_ref->node, &hdl->ctrl_refs); node 2316 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) { node 2324 drivers/media/v4l2-core/v4l2-ctrls.c list_add(&new_ref->node, ref->node.prev); node 2469 drivers/media/v4l2-core/v4l2-ctrls.c INIT_LIST_HEAD(&ctrl->node); node 2516 drivers/media/v4l2-core/v4l2-ctrls.c list_add_tail(&ctrl->node, &hdl->ctrls); node 2691 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &add->ctrl_refs, node) { node 2864 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ctrl, &hdl->ctrls, node) node 2889 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ctrl, &hdl->ctrls, node) node 2892 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ctrl, &hdl->ctrls, node) { node 2971 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry_continue(ref, &hdl->ctrl_refs, node) { node 2978 drivers/media/v4l2-core/v4l2-ctrls.c if (&ref->node == &hdl->ctrl_refs) node 2985 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) { node 2992 drivers/media/v4l2-core/v4l2-ctrls.c if (&ref->node == &hdl->ctrl_refs) node 3121 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &from->ctrl_refs, node) { node 3160 drivers/media/v4l2-core/v4l2-ctrls.c struct v4l2_ctrl_ref, node); node 3161 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref_ctrl, &hdl->ctrl_refs, node) { node 3166 drivers/media/v4l2-core/v4l2-ctrls.c if (list_is_last(&ref_ctrl_prev->node, node 3169 drivers/media/v4l2-core/v4l2-ctrls.c ref_ctrl_prev = list_next_entry(ref_ctrl_prev, node); node 4108 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) { node 4174 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) node 4177 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry(ref, &hdl->ctrl_refs, node) { node 4330 drivers/media/v4l2-core/v4l2-ctrls.c list_add_tail(&sev->node, &ctrl->ev_subs); node 4356 drivers/media/v4l2-core/v4l2-ctrls.c list_del(&sev->node); node 30 drivers/memory/atmel-ebi.c struct list_head node; node 369 drivers/memory/atmel-ebi.c list_add_tail(&ebid->node, &ebi->devs); node 617 drivers/memory/atmel-ebi.c list_for_each_entry(ebid, &ebi->devs, node) { node 105 drivers/memory/da8xx-ddrctl.c struct device_node *node; node 112 drivers/memory/da8xx-ddrctl.c node = dev->of_node; node 60 drivers/memory/emif.c struct list_head node; node 326 drivers/memory/emif.c list_for_each_entry(emif, &device_list, node) { node 337 drivers/memory/emif.c list_for_each_entry(emif, &device_list, node) { node 1553 drivers/memory/emif.c list_add(&emif->node, &device_list); node 1813 drivers/memory/emif.c list_for_each_entry(emif, &device_list, node) node 1881 drivers/memory/emif.c list_for_each_entry(emif, &device_list, node) node 1905 drivers/memory/emif.c list_for_each_entry(emif, &device_list, node) node 158 drivers/memory/jz4780-nemc.c struct device_node *node) node 188 drivers/memory/jz4780-nemc.c if (!of_property_read_u32(node, "ingenic,nemc-bus-width", &val)) { node 204 drivers/memory/jz4780-nemc.c if (of_property_read_u32(node, "ingenic,nemc-tAS", &val) == 0) { node 216 drivers/memory/jz4780-nemc.c if (of_property_read_u32(node, "ingenic,nemc-tAH", &val) == 0) { node 228 drivers/memory/jz4780-nemc.c if (of_property_read_u32(node, "ingenic,nemc-tBP", &val) == 0) { node 240 drivers/memory/jz4780-nemc.c if (of_property_read_u32(node, "ingenic,nemc-tAW", &val) == 0) { node 252 drivers/memory/jz4780-nemc.c if (of_property_read_u32(node, "ingenic,nemc-tSTRV", &val) == 0) { node 87 drivers/memory/mvebu-devbus.c struct device_node *node, node 94 drivers/memory/mvebu-devbus.c err = of_property_read_u32(node, name, &time_ps); node 97 drivers/memory/mvebu-devbus.c node, name); node 109 drivers/memory/mvebu-devbus.c struct device_node *node, node 115 drivers/memory/mvebu-devbus.c err = of_property_read_u32(node, "devbus,bus-width", &r->bus_width); node 119 drivers/memory/mvebu-devbus.c node); node 136 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,badr-skew-ps", node 141 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,turn-off-ps", node 146 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,acc-first-ps", node 151 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,acc-next-ps", node 157 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,rd-setup-ps", node 162 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,rd-hold-ps", node 167 drivers/memory/mvebu-devbus.c err = of_property_read_u32(node, "devbus,sync-enable", node 172 drivers/memory/mvebu-devbus.c node); node 177 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,ale-wr-ps", node 182 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,wr-low-ps", node 187 drivers/memory/mvebu-devbus.c err = get_timing_param_ps(devbus, node, "devbus,wr-high-ps", node 196 drivers/memory/mvebu-devbus.c struct device_node *node, node 229 drivers/memory/mvebu-devbus.c struct device_node *node, node 266 drivers/memory/mvebu-devbus.c struct device_node *node = pdev->dev.of_node; node 301 drivers/memory/mvebu-devbus.c if (!of_property_read_bool(node, "devbus,keep-config")) { node 303 drivers/memory/mvebu-devbus.c err = devbus_get_timing_params(devbus, node, &r, &w); node 308 drivers/memory/mvebu-devbus.c if (of_device_is_compatible(node, "marvell,orion-devbus")) node 309 drivers/memory/mvebu-devbus.c devbus_orion_set_timing_params(devbus, node, &r, &w); node 311 drivers/memory/mvebu-devbus.c devbus_armada_set_timing_params(devbus, node, &r, &w); node 319 drivers/memory/mvebu-devbus.c err = of_platform_populate(node, NULL, NULL, dev); node 345 drivers/memory/tegra/mc.c struct device_node *node) node 350 drivers/memory/tegra/mc.c err = of_property_read_u32(node, "clock-frequency", &tmp); node 353 drivers/memory/tegra/mc.c "timing %pOFn: failed to read rate\n", node); node 363 drivers/memory/tegra/mc.c err = of_property_read_u32_array(node, "nvidia,emem-configuration", node 369 drivers/memory/tegra/mc.c node); node 376 drivers/memory/tegra/mc.c static int load_timings(struct tegra_mc *mc, struct device_node *node) node 380 drivers/memory/tegra/mc.c int child_count = of_get_child_count(node); node 390 drivers/memory/tegra/mc.c for_each_child_of_node(node, child) { node 405 drivers/memory/tegra/mc.c struct device_node *node; node 413 drivers/memory/tegra/mc.c for_each_child_of_node(mc->dev->of_node, node) { node 414 drivers/memory/tegra/mc.c err = of_property_read_u32(node, "nvidia,ram-code", node 419 drivers/memory/tegra/mc.c err = load_timings(mc, node); node 420 drivers/memory/tegra/mc.c of_node_put(node); node 876 drivers/memory/tegra/tegra124-emc.c struct device_node *node) node 881 drivers/memory/tegra/tegra124-emc.c err = of_property_read_u32(node, "clock-frequency", &value); node 884 drivers/memory/tegra/tegra124-emc.c node, err); node 890 drivers/memory/tegra/tegra124-emc.c err = of_property_read_u32_array(node, "nvidia,emc-configuration", node 896 drivers/memory/tegra/tegra124-emc.c node, err); node 901 drivers/memory/tegra/tegra124-emc.c err = of_property_read_u32(node, dtprop, &timing->prop); \ node 904 drivers/memory/tegra/tegra124-emc.c node, err); \ node 946 drivers/memory/tegra/tegra124-emc.c struct device_node *node) node 948 drivers/memory/tegra/tegra124-emc.c int child_count = of_get_child_count(node); node 961 drivers/memory/tegra/tegra124-emc.c for_each_child_of_node(node, child) { node 983 drivers/memory/tegra/tegra124-emc.c tegra_emc_find_node_by_ram_code(struct device_node *node, u32 ram_code) node 988 drivers/memory/tegra/tegra124-emc.c for_each_child_of_node(node, np) { node 280 drivers/memory/tegra/tegra20-emc.c struct device_node *node) node 285 drivers/memory/tegra/tegra20-emc.c if (!of_device_is_compatible(node, "nvidia,tegra20-emc-table")) { node 286 drivers/memory/tegra/tegra20-emc.c dev_err(emc->dev, "incompatible DT node: %pOF\n", node); node 290 drivers/memory/tegra/tegra20-emc.c err = of_property_read_u32(node, "clock-frequency", &rate); node 293 drivers/memory/tegra/tegra20-emc.c node, err); node 297 drivers/memory/tegra/tegra20-emc.c err = of_property_read_u32_array(node, "nvidia,emc-registers", node 303 drivers/memory/tegra/tegra20-emc.c node, err); node 314 drivers/memory/tegra/tegra20-emc.c __func__, node, timing->rate); node 334 drivers/memory/tegra/tegra20-emc.c struct device_node *node) node 341 drivers/memory/tegra/tegra20-emc.c child_count = of_get_child_count(node); node 343 drivers/memory/tegra/tegra20-emc.c dev_err(emc->dev, "no memory timings in DT node: %pOF\n", node); node 355 drivers/memory/tegra/tegra20-emc.c for_each_child_of_node(node, child) { node 576 drivers/mfd/88pm860x-core.c struct device_node *node = i2c->dev.of_node; node 627 drivers/mfd/88pm860x-core.c irq_domain_add_legacy(node, nr_irqs, chip->irq_base, 0, node 1135 drivers/mfd/88pm860x-core.c struct device_node *node = client->dev.of_node; node 1139 drivers/mfd/88pm860x-core.c if (node && !pdata) { node 1146 drivers/mfd/88pm860x-core.c ret = pm860x_dt_init(node, &client->dev, pdata); node 160 drivers/mfd/ab8500-gpadc.c struct list_head node; node 180 drivers/mfd/ab8500-gpadc.c list_for_each_entry(gpadc, &ab8500_gpadc_list, node) { node 998 drivers/mfd/ab8500-gpadc.c list_add_tail(&gpadc->node, &ab8500_gpadc_list); node 1016 drivers/mfd/ab8500-gpadc.c list_del(&gpadc->node); node 341 drivers/mfd/cros_ec_dev.c struct device_node *node; node 428 drivers/mfd/cros_ec_dev.c node = ec->ec_dev->dev->of_node; node 429 drivers/mfd/cros_ec_dev.c if (of_property_read_bool(node, "google,has-vbc-nvram")) { node 68 drivers/mfd/ipaq-micro.c list_add_tail(&msg->node, µ->queue); node 104 drivers/mfd/ipaq-micro.c node); node 105 drivers/mfd/ipaq-micro.c list_del_init(µ->msg->node); node 666 drivers/mfd/max8925-core.c struct device_node *node = chip->dev->of_node; node 692 drivers/mfd/max8925-core.c irq_domain_add_legacy(node, MAX8925_NR_IRQS, chip->irq_base, 0, node 152 drivers/mfd/max8925-i2c.c struct device_node *node = client->dev.of_node; node 154 drivers/mfd/max8925-i2c.c if (node && !pdata) { node 162 drivers/mfd/max8925-i2c.c if (max8925_dt_init(node, &client->dev, pdata)) node 105 drivers/mfd/mfd-core.c list_for_each_entry(child, &parent->children, node) { node 115 drivers/mfd/mfd-core.c list_for_each_entry(child, &parent->children, node) { node 130 drivers/mfd/mxs-lradc.c struct device_node *node = dev->of_node; node 159 drivers/mfd/mxs-lradc.c ret = of_property_read_u32(node, "fsl,lradc-touchscreen-wires", node 476 drivers/mfd/omap-usb-host.c struct device_node *node = dev->of_node; node 478 drivers/mfd/omap-usb-host.c ret = of_property_read_u32(node, "num-ports", &pdata->nports); node 496 drivers/mfd/omap-usb-host.c ret = of_property_read_string(node, prop, &mode); node 513 drivers/mfd/omap-usb-host.c pdata->single_ulpi_bypass = of_property_read_bool(node, node 392 drivers/mfd/palmas.c struct device_node *node = i2c->dev.of_node; node 396 drivers/mfd/palmas.c ret = of_property_read_u32(node, "ti,mux-pad1", &prop); node 402 drivers/mfd/palmas.c ret = of_property_read_u32(node, "ti,mux-pad2", &prop); node 409 drivers/mfd/palmas.c ret = of_property_read_u32(node, "ti,power-ctrl", &prop); node 419 drivers/mfd/palmas.c pdata->pm_off = of_property_read_bool(node, node 511 drivers/mfd/palmas.c struct device_node *node = i2c->dev.of_node; node 519 drivers/mfd/palmas.c if (node && !pdata) { node 560 drivers/mfd/palmas.c palmas->i2c_clients[i]->dev.of_node = of_node_get(node); node 681 drivers/mfd/palmas.c if (node) { node 127 drivers/mfd/ti_am335x_tscadc.c struct device_node *node; node 142 drivers/mfd/ti_am335x_tscadc.c node = of_get_child_by_name(pdev->dev.of_node, "tsc"); node 143 drivers/mfd/ti_am335x_tscadc.c of_property_read_u32(node, "ti,wires", &tsc_wires); node 144 drivers/mfd/ti_am335x_tscadc.c of_property_read_u32(node, "ti,coordiante-readouts", &readouts); node 146 drivers/mfd/ti_am335x_tscadc.c node = of_get_child_by_name(pdev->dev.of_node, "adc"); node 147 drivers/mfd/ti_am335x_tscadc.c of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) { node 1071 drivers/mfd/twl-core.c struct device_node *node = client->dev.of_node; node 1078 drivers/mfd/twl-core.c if (!node && !pdata) { node 1216 drivers/mfd/twl-core.c if (node) { node 1219 drivers/mfd/twl-core.c status = of_platform_populate(node, NULL, twl_auxdata_lookup, node 150 drivers/mfd/twl4030-audio.c struct device_node *node; node 155 drivers/mfd/twl4030-audio.c node = of_get_child_by_name(parent, "codec"); node 156 drivers/mfd/twl4030-audio.c if (node) { node 157 drivers/mfd/twl4030-audio.c of_node_put(node); node 165 drivers/mfd/twl4030-audio.c struct device_node *node) node 172 drivers/mfd/twl4030-audio.c if (!of_property_read_u32(node, "ti,enable-vibra", &vibra) && vibra) node 182 drivers/mfd/twl4030-audio.c struct device_node *node = pdev->dev.of_node; node 187 drivers/mfd/twl4030-audio.c if (!pdata && !node) { node 225 drivers/mfd/twl4030-audio.c if (twl4030_audio_has_codec(pdata, node)) { node 234 drivers/mfd/twl4030-audio.c if (twl4030_audio_has_vibra(pdata, node)) { node 678 drivers/mfd/twl4030-irq.c struct device_node *node = dev->of_node; node 693 drivers/mfd/twl4030-irq.c irq_domain_add_legacy(node, nr_irqs, irq_base, 0, node 678 drivers/mfd/twl4030-power.c struct device_node *node) node 683 drivers/mfd/twl4030-power.c if (of_property_read_bool(node, "ti,system-power-controller")) node 686 drivers/mfd/twl4030-power.c if (of_property_read_bool(node, "ti,use_poweroff")) node 885 drivers/mfd/twl4030-power.c struct device_node *node = pdev->dev.of_node; node 891 drivers/mfd/twl4030-power.c if (!pdata && !node) { node 926 drivers/mfd/twl4030-power.c if (twl4030_power_use_poweroff(pdata, node) && !pm_power_off) { node 367 drivers/mfd/twl6030-irq.c struct device_node *node = dev->of_node; node 415 drivers/mfd/twl6030-irq.c irq_domain_add_linear(node, nr_irqs, node 88 drivers/mfd/twl6040.c struct device_node *node; node 90 drivers/mfd/twl6040.c node = of_get_child_by_name(parent, "vibra"); node 91 drivers/mfd/twl6040.c if (node) { node 92 drivers/mfd/twl6040.c of_node_put(node); node 639 drivers/mfd/twl6040.c struct device_node *node = client->dev.of_node; node 644 drivers/mfd/twl6040.c if (!node) { node 716 drivers/mfd/twl6040.c twl6040->audpwron = of_get_named_gpio(node, node 771 drivers/mfd/twl6040.c if (twl6040_has_vibra(node)) { node 588 drivers/mfd/ucb1x00-core.c list_add_tail(&ucb->node, &ucb1x00_devices); node 589 drivers/mfd/ucb1x00-core.c list_for_each_entry(drv, &ucb1x00_drivers, node) { node 618 drivers/mfd/ucb1x00-core.c list_del(&ucb->node); node 642 drivers/mfd/ucb1x00-core.c list_add_tail(&drv->node, &ucb1x00_drivers); node 643 drivers/mfd/ucb1x00-core.c list_for_each_entry(ucb, &ucb1x00_devices, node) { node 655 drivers/mfd/ucb1x00-core.c list_del(&drv->node); node 50 drivers/mfd/vexpress-sysreg.c struct device_node *node = of_find_compatible_node(NULL, NULL, node 53 drivers/mfd/vexpress-sysreg.c base = of_iomap(node, 0); node 220 drivers/mfd/vexpress-sysreg.c struct device_node *node; node 223 drivers/mfd/vexpress-sysreg.c for_each_matching_node(node, vexpress_sysreg_match) node 224 drivers/mfd/vexpress-sysreg.c of_platform_device_create(node, NULL, NULL); node 50 drivers/misc/atmel_tclib.c list_for_each_entry(tc, &tc_list, node) { node 173 drivers/misc/atmel_tclib.c list_add_tail(&tc->node, &tc_list); node 49 drivers/misc/enclosure.c edev = list_prepare_entry(start, &container_list, node); node 53 drivers/misc/enclosure.c list_for_each_entry_continue(edev, &container_list, node) { node 91 drivers/misc/enclosure.c list_for_each_entry(edev, &container_list, node) { node 142 drivers/misc/enclosure.c list_add_tail(&edev->node, &container_list); node 166 drivers/misc/enclosure.c list_del(&edev->node); node 131 drivers/misc/fastrpc.c struct list_head node; node 135 drivers/misc/fastrpc.c struct list_head node; node 159 drivers/misc/fastrpc.c struct list_head node; /* list of ctxs */ node 242 drivers/misc/fastrpc.c list_for_each_entry(map, &fl->maps, node) { node 422 drivers/misc/fastrpc.c INIT_LIST_HEAD(&ctx->node); node 458 drivers/misc/fastrpc.c list_add_tail(&ctx->node, &user->pending); node 476 drivers/misc/fastrpc.c list_del(&ctx->node); node 535 drivers/misc/fastrpc.c INIT_LIST_HEAD(&a->node); node 539 drivers/misc/fastrpc.c list_add(&a->node, &buffer->attachments); node 552 drivers/misc/fastrpc.c list_del(&a->node); node 607 drivers/misc/fastrpc.c INIT_LIST_HEAD(&map->node); node 637 drivers/misc/fastrpc.c list_add_tail(&map->node, &fl->maps); node 959 drivers/misc/fastrpc.c list_del(&ctx->node); node 1071 drivers/misc/fastrpc.c list_del(&map->node); node 1145 drivers/misc/fastrpc.c list_for_each_entry_safe(ctx, n, &fl->pending, node) { node 1146 drivers/misc/fastrpc.c list_del(&ctx->node); node 1150 drivers/misc/fastrpc.c list_for_each_entry_safe(map, m, &fl->maps, node) { node 1151 drivers/misc/fastrpc.c list_del(&map->node); node 1458 drivers/misc/fastrpc.c list_for_each_entry(ctx, &user->pending, node) node 191 drivers/misc/genwqe/card_dev.c struct list_head *node, *next; node 196 drivers/misc/genwqe/card_dev.c list_for_each_safe(node, next, &cfile->map_list) { node 197 drivers/misc/genwqe/card_dev.c dma_map = list_entry(node, struct dma_mapping, card_list); node 229 drivers/misc/genwqe/card_dev.c struct list_head *node, *next; node 233 drivers/misc/genwqe/card_dev.c list_for_each_safe(node, next, &cfile->pin_list) { node 234 drivers/misc/genwqe/card_dev.c dma_map = list_entry(node, struct dma_mapping, pin_list); node 253 drivers/misc/habanalabs/debugfs.c hash_for_each(ctx->mem_hash, i, hnode, node) { node 912 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *node = inode->i_private; node 914 drivers/misc/habanalabs/debugfs.c return single_open(file, node->info_ent->show, node); node 920 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *node = file->f_inode->i_private; node 922 drivers/misc/habanalabs/debugfs.c if (node->info_ent->write) node 923 drivers/misc/habanalabs/debugfs.c return node->info_ent->write(file, buf, count, f_pos); node 70 drivers/misc/habanalabs/habanalabs.h struct hlist_node node; node 818 drivers/misc/habanalabs/habanalabs.h struct hlist_node node; node 861 drivers/misc/habanalabs/habanalabs.h struct list_head node; node 359 drivers/misc/habanalabs/memory.c list_for_each_entry_safe(va_block, tmp, va_list, node) { node 360 drivers/misc/habanalabs/memory.c list_del(&va_block->node); node 384 drivers/misc/habanalabs/memory.c list_for_each_entry(va_block, va_list, node) node 409 drivers/misc/habanalabs/memory.c prev = list_prev_entry(va_block, node); node 410 drivers/misc/habanalabs/memory.c if (&prev->node != va_list && prev->end + 1 == va_block->start) { node 413 drivers/misc/habanalabs/memory.c list_del(&va_block->node); node 418 drivers/misc/habanalabs/memory.c next = list_next_entry(va_block, node); node 419 drivers/misc/habanalabs/memory.c if (&next->node != va_list && va_block->end + 1 == next->start) { node 422 drivers/misc/habanalabs/memory.c list_del(&va_block->node); node 449 drivers/misc/habanalabs/memory.c list_for_each_entry(va_block, va_list, node) { node 472 drivers/misc/habanalabs/memory.c list_add(&va_block->node, va_list); node 474 drivers/misc/habanalabs/memory.c list_add(&va_block->node, &res->node); node 548 drivers/misc/habanalabs/memory.c list_for_each_entry(va_block, &va_range->list, node) { node 599 drivers/misc/habanalabs/memory.c list_del(&new_va_block->node); node 930 drivers/misc/habanalabs/memory.c hash_add(ctx->mem_hash, &hnode->node, ret_vaddr); node 989 drivers/misc/habanalabs/memory.c hash_for_each_possible(ctx->mem_hash, hnode, node, (unsigned long)vaddr) node 1001 drivers/misc/habanalabs/memory.c hash_del(&hnode->node); node 1087 drivers/misc/habanalabs/memory.c hash_add(ctx->mem_hash, &hnode->node, vaddr); node 1576 drivers/misc/habanalabs/memory.c va_block = list_first_entry(&va_range->list, typeof(*va_block), node); node 1629 drivers/misc/habanalabs/memory.c hash_for_each_safe(ctx->mem_hash, i, tmp_node, hnode, node) { node 20 drivers/misc/habanalabs/mmu.c hash_for_each_possible(ctx->mmu_shadow_hash, pgt_info, node, node 35 drivers/misc/habanalabs/mmu.c hash_del(&pgt_info->node); node 67 drivers/misc/habanalabs/mmu.c hash_add(ctx->mmu_shadow_hash, &pgt_info->node, shadow_addr); node 506 drivers/misc/habanalabs/mmu.c hash_for_each_safe(ctx->mmu_shadow_hash, i, tmp, pgt_info, node) { node 30 drivers/misc/ibmasm/event.c list_for_each_entry(reader, &sp->event_buffer->readers, node) node 123 drivers/misc/ibmasm/event.c list_add(&reader->node, &sp->event_buffer->readers); node 132 drivers/misc/ibmasm/event.c list_del(&reader->node); node 124 drivers/misc/ibmasm/ibmasm.h struct list_head node; node 140 drivers/misc/ibmasm/ibmasm.h struct list_head node; node 208 drivers/misc/ibmasm/ibmasmfs.c list_add(&sp->node, &service_processors); node 586 drivers/misc/ibmasm/ibmasmfs.c sp = list_entry(entry, struct service_processor, node); node 16 drivers/misc/lkdtm/bugs.c struct list_head node; node 195 drivers/misc/lkdtm/bugs.c list_add(&good.node, &test_head); node 204 drivers/misc/lkdtm/bugs.c list_add(&bad.node, &test_head); node 219 drivers/misc/lkdtm/bugs.c list_add(&item.node, &test_head); node 222 drivers/misc/lkdtm/bugs.c list_del(&item.node); node 225 drivers/misc/lkdtm/bugs.c list_add(&item.node, &test_head); node 228 drivers/misc/lkdtm/bugs.c item.node.next = redirection; node 229 drivers/misc/lkdtm/bugs.c list_del(&item.node); node 306 drivers/misc/mic/cosm/cosm_scif_server.c cdev = cosm_find_cdev_by_id(port_id.node - 1); node 112 drivers/misc/mic/cosm_client/cosm_scif_client.c port_id.node = 0; node 275 drivers/misc/mic/scif/scif_api.c msg.dst.node = conreq->msg.src.node; node 284 drivers/misc/mic/scif/scif_api.c scif_nodeqp_send(&scif_dev[conreq->msg.src.node], node 372 drivers/misc/mic/scif/scif_api.c ep->port.node = scif_info.nodeid; node 600 drivers/misc/mic/scif/scif_api.c if (!scif_dev || dst->node > scif_info.maxid) node 605 drivers/misc/mic/scif/scif_api.c remote_dev = &scif_dev[dst->node]; node 646 drivers/misc/mic/scif/scif_api.c ep->port.node = scif_info.nodeid; node 683 drivers/misc/mic/scif/scif_api.c ep->remote_dev = &scif_dev[dst->node]; node 794 drivers/misc/mic/scif/scif_api.c peer->node = conreq->msg.src.node; node 806 drivers/misc/mic/scif/scif_api.c cep->remote_dev = &scif_dev[peer->node]; node 845 drivers/misc/mic/scif/scif_api.c cep->port.node = lep->port.node; node 847 drivers/misc/mic/scif/scif_api.c cep->peer.node = peer->node; node 905 drivers/misc/mic/scif/scif_api.c msg.dst.node = conreq->msg.src.node; node 909 drivers/misc/mic/scif/scif_api.c scif_nodeqp_send(&scif_dev[conreq->msg.src.node], &msg); node 1422 drivers/misc/mic/scif/scif_api.c int node; node 1430 drivers/misc/mic/scif/scif_api.c for (node = 0; node <= scif_info.maxid; node++) { node 1431 drivers/misc/mic/scif/scif_api.c if (_scifdev_alive(&scif_dev[node])) { node 1434 drivers/misc/mic/scif/scif_api.c nodes[offset++] = node; node 20 drivers/misc/mic/scif/scif_debugfs.c int node; node 31 drivers/misc/mic/scif/scif_debugfs.c for (node = 0; node <= scif_info.maxid; node++) node 32 drivers/misc/mic/scif/scif_debugfs.c seq_printf(s, "%-16d\t%-16s\n", scif_dev[node].node, node 33 drivers/misc/mic/scif/scif_debugfs.c _scifdev_alive(&scif_dev[node]) ? node 153 drivers/misc/mic/scif/scif_epd.c scif_nodeqp_send(&scif_dev[msg->src.node], msg); node 171 drivers/misc/mic/scif/scif_epd.c ep->peer.node = msg->src.node; node 299 drivers/misc/mic/scif/scif_epd.c scif_nodeqp_send(&scif_dev[msg->src.node], msg); node 113 drivers/misc/mic/scif/scif_fd.c req.self.node = ep->port.node; node 47 drivers/misc/mic/scif/scif_main.c "SCIF INTR %d", scifdev->node); node 118 drivers/misc/mic/scif/scif_main.c scifdev->node = i; node 224 drivers/misc/mic/scif/scif_main.c scif_disconnect_node(scifdev->node, true); node 169 drivers/misc/mic/scif/scif_main.h u8 node; node 221 drivers/misc/mic/scif/scif_main.h return dev->node == scif_info.nodeid; node 268 drivers/misc/mic/scif/scif_main.h void scif_handle_remove_node(int node); node 90 drivers/misc/mic/scif/scif_mmap.c static void _scif_zap_mmaps(int node, struct list_head *head) node 98 drivers/misc/mic/scif/scif_mmap.c if (ep->remote_dev->node == node) node 108 drivers/misc/mic/scif/scif_mmap.c void scif_zap_mmaps(int node) node 110 drivers/misc/mic/scif/scif_mmap.c _scif_zap_mmaps(node, &scif_info.connected); node 111 drivers/misc/mic/scif/scif_mmap.c _scif_zap_mmaps(node, &scif_info.disconnected); node 144 drivers/misc/mic/scif/scif_mmap.c void scif_cleanup_rma_for_zombies(int node) node 152 drivers/misc/mic/scif/scif_mmap.c if (ep->remote_dev && ep->remote_dev->node == node) node 18 drivers/misc/mic/scif/scif_nm.c static void scif_invalidate_ep(int node) node 27 drivers/misc/mic/scif/scif_nm.c if (ep->remote_dev->node == node) { node 36 drivers/misc/mic/scif/scif_nm.c if (ep->remote_dev->node == node) { node 87 drivers/misc/mic/scif/scif_nm.c msg.src.node = scif_info.nodeid; node 88 drivers/misc/mic/scif/scif_nm.c msg.dst.node = SCIF_MGMT_NODE; node 89 drivers/misc/mic/scif/scif_nm.c msg.payload[0] = dev->node; node 95 drivers/misc/mic/scif/scif_nm.c msg.src.node = scif_info.nodeid; node 96 drivers/misc/mic/scif/scif_nm.c msg.dst.node = dev->node; node 123 drivers/misc/mic/scif/scif_nm.c scif_invalidate_ep(dev->node); node 124 drivers/misc/mic/scif/scif_nm.c scif_zap_mmaps(dev->node); node 125 drivers/misc/mic/scif/scif_nm.c scif_cleanup_rma_for_zombies(dev->node); node 128 drivers/misc/mic/scif/scif_nm.c if (!dev->node && scif_info.card_initiated_exit) { node 144 drivers/misc/mic/scif/scif_nm.c void scif_handle_remove_node(int node) node 146 drivers/misc/mic/scif/scif_nm.c struct scif_dev *scifdev = &scif_dev[node]; node 152 drivers/misc/mic/scif/scif_nm.c static int scif_send_rmnode_msg(int node, int remove_node) node 155 drivers/misc/mic/scif/scif_nm.c struct scif_dev *dev = &scif_dev[node]; node 158 drivers/misc/mic/scif/scif_nm.c notif_msg.src.node = scif_info.nodeid; node 159 drivers/misc/mic/scif/scif_nm.c notif_msg.dst.node = node; node 219 drivers/misc/mic/scif/scif_nm.c msg.src.node = scif_info.nodeid; node 220 drivers/misc/mic/scif/scif_nm.c msg.dst.node = SCIF_MGMT_NODE; node 239 drivers/misc/mic/scif/scif_nodeqp.c scif_dev[scif_info.nodeid].node, scifdev->node); node 298 drivers/misc/mic/scif/scif_nodeqp.c msg.dst.node = scifdev->node; node 311 drivers/misc/mic/scif/scif_nodeqp.c msg.src.node = scif_info.nodeid; node 312 drivers/misc/mic/scif/scif_nodeqp.c msg.dst.node = scifdev->node; node 416 drivers/misc/mic/scif/scif_nodeqp.c p2p->ppi_peer_id = peerdev->node; node 478 drivers/misc/mic/scif/scif_nodeqp.c if (p2p->ppi_peer_id == dev_j->node) node 498 drivers/misc/mic/scif/scif_nodeqp.c msg.src.node = dev_j->node; node 499 drivers/misc/mic/scif/scif_nodeqp.c msg.dst.node = dev_i->node; node 515 drivers/misc/mic/scif/scif_nodeqp.c msg.src.node = dev_i->node; node 516 drivers/misc/mic/scif/scif_nodeqp.c msg.dst.node = dev_j->node; node 604 drivers/misc/mic/scif/scif_nodeqp.c label, message_types[msg->uop], msg->src.node, msg->src.port, node 605 drivers/misc/mic/scif/scif_nodeqp.c msg->dst.node, msg->dst.port, msg->payload[0], msg->payload[1], node 733 drivers/misc/mic/scif/scif_nodeqp.c scif_disconnect_node(scifdev->node, false); node 777 drivers/misc/mic/scif/scif_nodeqp.c scifdev->node, msg->dst.node, msg->src.node); node 781 drivers/misc/mic/scif/scif_nodeqp.c newdev = &scif_dev[msg->src.node]; node 782 drivers/misc/mic/scif/scif_nodeqp.c newdev->node = msg->src.node; node 788 drivers/misc/mic/scif/scif_nodeqp.c "failed to setup interrupts for %d\n", msg->src.node); node 794 drivers/misc/mic/scif/scif_nodeqp.c "failed to map mmio for %d\n", msg->src.node); node 824 drivers/misc/mic/scif/scif_nodeqp.c msg->dst.node = msg->src.node; node 825 drivers/misc/mic/scif/scif_nodeqp.c msg->src.node = scif_info.nodeid; node 839 drivers/misc/mic/scif/scif_nodeqp.c "node add failed for node %d\n", msg->src.node); node 841 drivers/misc/mic/scif/scif_nodeqp.c msg->dst.node = msg->src.node; node 842 drivers/misc/mic/scif/scif_nodeqp.c msg->src.node = scif_info.nodeid; node 870 drivers/misc/mic/scif/scif_nodeqp.c __func__, __LINE__, peerdev->node, qp->qp_state); node 890 drivers/misc/mic/scif/scif_nodeqp.c struct scif_dev *dst_dev = &scif_dev[msg->dst.node]; node 894 drivers/misc/mic/scif/scif_nodeqp.c scifdev->node, msg->src.node, msg->dst.node); node 911 drivers/misc/mic/scif/scif_nodeqp.c peerdev = &scif_dev[msg->src.node]; node 913 drivers/misc/mic/scif/scif_nodeqp.c peerdev->node = msg->src.node; node 941 drivers/misc/mic/scif/scif_nodeqp.c struct scif_dev *dst_dev = &scif_dev[msg->dst.node]; node 944 drivers/misc/mic/scif/scif_nodeqp.c "SCIF_NODE_ADD_NACK received from %d\n", scifdev->node); node 958 drivers/misc/mic/scif/scif_nodeqp.c int node = msg->payload[0]; node 959 drivers/misc/mic/scif/scif_nodeqp.c struct scif_dev *scdev = &scif_dev[node]; node 962 drivers/misc/mic/scif/scif_nodeqp.c scif_handle_remove_node(node); node 990 drivers/misc/mic/scif/scif_nodeqp.c swap(msg->dst.node, msg->src.node); node 1014 drivers/misc/mic/scif/scif_nodeqp.c msg->uop, scifdev->node); node 1082 drivers/misc/mic/scif/scif_nodeqp.c msg->uop, scifdev->node); node 1227 drivers/misc/mic/scif/scif_nodeqp.c "SCIF LOOPB %d", scifdev->node); node 1266 drivers/misc/mic/scif/scif_nodeqp.c scif_info.nodeid = scifdev->node; node 1328 drivers/misc/mic/scif/scif_nodeqp.c if (p2p->ppi_peer_id == scifdev->node) { node 45 drivers/misc/mic/scif/scif_peer_bus.c spdev->dnode = scifdev->node; node 60 drivers/misc/mic/scif/scif_peer_bus.c "dnode %d: initialize_device rc %d\n", scifdev->node, ret); node 74 drivers/misc/mic/scif/scif_peer_bus.c "dnode %d: peer device_add failed\n", scifdev->node); node 84 drivers/misc/mic/scif/scif_peer_bus.c "dnode %d: dmam_pool_create failed\n", scifdev->node); node 342 drivers/misc/mic/scif/scif_rma.h void scif_zap_mmaps(int node); node 344 drivers/misc/mic/scif/scif_rma.h bool scif_rma_do_apps_have_mmaps(int node); node 346 drivers/misc/mic/scif/scif_rma.h void scif_cleanup_rma_for_zombies(int node); node 87 drivers/misc/qcom-coincell.c struct device_node *node = pdev->dev.of_node; node 102 drivers/misc/qcom-coincell.c rc = of_property_read_u32(node, "reg", &chgr.base_addr); node 106 drivers/misc/qcom-coincell.c enable = !of_property_read_bool(node, "qcom,charger-disable"); node 109 drivers/misc/qcom-coincell.c rc = of_property_read_u32(node, "qcom,rset-ohms", &rset); node 116 drivers/misc/qcom-coincell.c rc = of_property_read_u32(node, "qcom,vset-millivolts", &vset); node 45 drivers/misc/sgi-gru/gru.h int node; node 223 drivers/misc/vmw_vmci/vmci_context.c struct vmci_handle_list *node; node 233 drivers/misc/vmw_vmci/vmci_context.c list_for_each_entry_rcu(node, &sub_ctx->notifier_list, node) { node 234 drivers/misc/vmw_vmci/vmci_context.c if (!vmci_handle_is_equal(node->handle, context_handle)) node 472 drivers/misc/vmw_vmci/vmci_context.c &context->notifier_list, node) { node 473 drivers/misc/vmw_vmci/vmci_context.c list_del(¬ifier->node); node 627 drivers/misc/vmw_vmci/vmci_context.c INIT_LIST_HEAD(¬ifier->node); node 633 drivers/misc/vmw_vmci/vmci_context.c list_for_each_entry(n, &context->notifier_list, node) { node 644 drivers/misc/vmw_vmci/vmci_context.c list_add_tail_rcu(¬ifier->node, node 680 drivers/misc/vmw_vmci/vmci_context.c &context->notifier_list, node) { node 682 drivers/misc/vmw_vmci/vmci_context.c list_del_rcu(¬ifier->node); node 724 drivers/misc/vmw_vmci/vmci_context.c list_for_each_entry(entry, &context->notifier_list, node) node 36 drivers/misc/vmw_vmci/vmci_context.h struct list_head node; node 35 drivers/misc/vmw_vmci/vmci_doorbell.c struct hlist_node node; node 124 drivers/misc/vmw_vmci/vmci_doorbell.c node) { node 188 drivers/misc/vmw_vmci/vmci_doorbell.c hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]); node 201 drivers/misc/vmw_vmci/vmci_doorbell.c hlist_del_init(&entry->node); node 360 drivers/misc/vmw_vmci/vmci_doorbell.c hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], node) { node 466 drivers/misc/vmw_vmci/vmci_doorbell.c INIT_HLIST_NODE(&entry->node); node 531 drivers/misc/vmw_vmci/vmci_doorbell.c if (!hlist_unhashed(&entry->node)) { node 27 drivers/misc/vmw_vmci/vmci_event.c struct list_head node; /* on one of subscriber lists */ node 50 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) { node 58 drivers/misc/vmw_vmci/vmci_event.c list_del(&cur->node); node 73 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry(cur, &subscriber_array[e], node) { node 92 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry_rcu(cur, subscriber_list, node) { node 158 drivers/misc/vmw_vmci/vmci_event.c INIT_LIST_HEAD(&sub->node); node 179 drivers/misc/vmw_vmci/vmci_event.c list_add_rcu(&sub->node, &subscriber_array[event]); node 206 drivers/misc/vmw_vmci/vmci_event.c list_del_rcu(&s->node); node 46 drivers/misc/vmw_vmci/vmci_resource.c &vmci_resource_table.entries[idx], node) { node 123 drivers/misc/vmw_vmci/vmci_resource.c INIT_HLIST_NODE(&resource->node); node 128 drivers/misc/vmw_vmci/vmci_resource.c hlist_add_head_rcu(&resource->node, &vmci_resource_table.entries[idx]); node 146 drivers/misc/vmw_vmci/vmci_resource.c hlist_for_each_entry(r, &vmci_resource_table.entries[idx], node) { node 148 drivers/misc/vmw_vmci/vmci_resource.c hlist_del_init_rcu(&r->node); node 196 drivers/misc/vmw_vmci/vmci_resource.c WARN_ON(!hlist_unhashed(&resource->node)); node 30 drivers/misc/vmw_vmci/vmci_resource.h struct hlist_node node; node 1066 drivers/misc/xilinx_sdfec.c struct device_node *node = dev->of_node; node 1074 drivers/misc/xilinx_sdfec.c rval = of_property_read_string(node, "xlnx,sdfec-code", &fec_code); node 1085 drivers/misc/xilinx_sdfec.c rval = of_property_read_u32(node, "xlnx,sdfec-din-words", node 1095 drivers/misc/xilinx_sdfec.c rval = of_property_read_u32(node, "xlnx,sdfec-din-width", &din_width); node 1110 drivers/misc/xilinx_sdfec.c rval = of_property_read_u32(node, "xlnx,sdfec-dout-words", node 1120 drivers/misc/xilinx_sdfec.c rval = of_property_read_u32(node, "xlnx,sdfec-dout-width", &dout_width); node 161 drivers/mmc/core/block.c struct list_head node; node 2547 drivers/mmc/core/block.c list_add(&rpmb->node, &md->rpmbs); node 2647 drivers/mmc/core/block.c rpmb = list_entry(pos, struct mmc_rpmb_data, node); node 1087 drivers/mmc/core/core.c static int mmc_of_get_func_num(struct device_node *node) node 1092 drivers/mmc/core/core.c ret = of_property_read_u32(node, "reg", ®); node 1102 drivers/mmc/core/core.c struct device_node *node; node 1107 drivers/mmc/core/core.c for_each_child_of_node(host->parent->of_node, node) { node 1108 drivers/mmc/core/core.c if (mmc_of_get_func_num(node) == func_num) node 1109 drivers/mmc/core/core.c return node; node 149 drivers/mmc/host/cavium-octeon.c struct device_node *cn, *node = pdev->dev.of_node; node 177 drivers/mmc/host/cavium-octeon.c if (of_device_is_compatible(node, "cavium,octeon-7890-mmc")) { node 280 drivers/mmc/host/cavium-octeon.c for_each_child_of_node(node, cn) { node 59 drivers/mmc/host/cavium-thunderx.c struct device_node *node = pdev->dev.of_node; node 127 drivers/mmc/host/cavium-thunderx.c for_each_child_of_node(node, child_node) { node 954 drivers/mmc/host/cavium.c struct device_node *node = dev->of_node; node 959 drivers/mmc/host/cavium.c ret = of_property_read_u32(node, "reg", &id); node 961 drivers/mmc/host/cavium.c dev_err(dev, "Missing or invalid reg property on %pOF\n", node); node 966 drivers/mmc/host/cavium.c dev_err(dev, "Invalid reg property on %pOF\n", node); node 987 drivers/mmc/host/cavium.c of_property_read_u32(node, "cavium,bus-max-width", &bus_width); node 996 drivers/mmc/host/cavium.c of_property_read_u32(node, "spi-max-frequency", &mmc->f_max); node 1003 drivers/mmc/host/cavium.c of_property_read_u32(node, "cavium,cmd-clk-skew", &cmd_skew); node 1004 drivers/mmc/host/cavium.c of_property_read_u32(node, "cavium,dat-clk-skew", &dat_skew); node 155 drivers/mmc/host/dw_mmc-zx.c struct device_node *node; node 161 drivers/mmc/host/dw_mmc-zx.c node = of_parse_phandle(np, "zte,aon-syscon", 0); node 162 drivers/mmc/host/dw_mmc-zx.c if (node) { node 163 drivers/mmc/host/dw_mmc-zx.c sysc_base = syscon_node_to_regmap(node); node 164 drivers/mmc/host/dw_mmc-zx.c of_node_put(node); node 558 drivers/mmc/host/moxart-mmc.c struct device_node *node = dev->of_node; node 575 drivers/mmc/host/moxart-mmc.c ret = of_address_to_resource(node, 0, &res_mmc); node 581 drivers/mmc/host/moxart-mmc.c irq = irq_of_parse_and_map(node, 0); node 701 drivers/mmc/host/sdhci-acpi.c list_for_each_entry(child, &device->children, node) node 736 drivers/mmc/host/sdhci-of-arasan.c struct device_node *node; node 757 drivers/mmc/host/sdhci-of-arasan.c node = of_parse_phandle(pdev->dev.of_node, "arasan,soc-ctl-syscon", 0); node 758 drivers/mmc/host/sdhci-of-arasan.c if (node) { node 759 drivers/mmc/host/sdhci-of-arasan.c sdhci_arasan->soc_ctl_base = syscon_node_to_regmap(node); node 760 drivers/mmc/host/sdhci-of-arasan.c of_node_put(node); node 1129 drivers/mmc/host/sdhci-pci-core.c list_for_each_entry(child, &device->children, node) node 427 drivers/mmc/host/sdhci-s3c.c struct device_node *node = dev->of_node; node 431 drivers/mmc/host/sdhci-s3c.c if (of_property_read_u32(node, "bus-width", &max_width)) node 436 drivers/mmc/host/sdhci-s3c.c if (of_get_property(node, "broken-cd", NULL)) { node 441 drivers/mmc/host/sdhci-s3c.c if (of_get_property(node, "non-removable", NULL)) { node 446 drivers/mmc/host/sdhci-s3c.c if (of_get_named_gpio(node, "cd-gpios", 0)) node 161 drivers/mtd/nand/raw/atmel/nand-controller.c struct list_head node; node 1553 drivers/mtd/nand/raw/atmel/nand-controller.c list_del(&nand->node); node 1692 drivers/mtd/nand/raw/atmel/nand-controller.c list_add_tail(&nand->node, &nc->chips); node 1703 drivers/mtd/nand/raw/atmel/nand-controller.c list_for_each_entry_safe(nand, tmp, &nc->chips, node) { node 2569 drivers/mtd/nand/raw/atmel/nand-controller.c list_for_each_entry(nand, &nc->chips, node) { node 230 drivers/mtd/nand/raw/brcmnand/brcmnand.c struct list_head node; node 2470 drivers/mtd/nand/raw/brcmnand/brcmnand.c list_for_each_entry(host, &ctrl->host_list, node) node 2504 drivers/mtd/nand/raw/brcmnand/brcmnand.c list_for_each_entry(host, &ctrl->host_list, node) { node 2718 drivers/mtd/nand/raw/brcmnand/brcmnand.c list_add_tail(&host->node, &ctrl->host_list); node 2742 drivers/mtd/nand/raw/brcmnand/brcmnand.c list_for_each_entry(host, &ctrl->host_list, node) node 1185 drivers/mtd/nand/raw/denali.c list_for_each_entry(dchip2, &denali->chips, node) { node 1238 drivers/mtd/nand/raw/denali.c list_add_tail(&dchip->node, &denali->chips); node 1340 drivers/mtd/nand/raw/denali.c list_for_each_entry(dchip, &denali->chips, node) node 329 drivers/mtd/nand/raw/denali.h struct list_head node; node 858 drivers/mtd/nand/raw/fsl_elbc_nand.c struct device_node *node = pdev->dev.of_node; node 867 drivers/mtd/nand/raw/fsl_elbc_nand.c ret = of_address_to_resource(node, 0, &res); node 984 drivers/mtd/nand/raw/fsl_ifc_nand.c struct device_node *node = dev->dev.of_node; node 992 drivers/mtd/nand/raw/fsl_ifc_nand.c ret = of_address_to_resource(node, 0, &res); node 329 drivers/mtd/nand/raw/marvell_nand.c struct list_head node; node 2671 drivers/mtd/nand/raw/marvell_nand.c list_add_tail(&marvell_nand->node, &nfc->chips); node 2721 drivers/mtd/nand/raw/marvell_nand.c list_for_each_entry_safe(entry, temp, &nfc->chips, node) { node 2723 drivers/mtd/nand/raw/marvell_nand.c list_del(&entry->node); node 2959 drivers/mtd/nand/raw/marvell_nand.c list_for_each_entry(chip, &nfc->chips, node) node 108 drivers/mtd/nand/raw/meson_nand.c struct list_head node; node 1285 drivers/mtd/nand/raw/meson_nand.c list_add_tail(&meson_chip->node, &nfc->chips); node 1298 drivers/mtd/nand/raw/meson_nand.c struct meson_nfc_nand_chip, node); node 1306 drivers/mtd/nand/raw/meson_nand.c list_del(&meson_chip->node); node 126 drivers/mtd/nand/raw/mtk_nand.c struct list_head node; node 1426 drivers/mtd/nand/raw/mtk_nand.c list_add_tail(&chip->node, &nfc->chips); node 1586 drivers/mtd/nand/raw/mtk_nand.c node); node 1588 drivers/mtd/nand/raw/mtk_nand.c list_del(&chip->node); node 1622 drivers/mtd/nand/raw/mtk_nand.c list_for_each_entry(chip, &nfc->chips, node) { node 133 drivers/mtd/nand/raw/ndfc.c struct device_node *node) node 160 drivers/mtd/nand/raw/ndfc.c flash_np = of_get_next_child(node, NULL); node 276 drivers/mtd/nand/raw/qcom_nandc.c struct list_head node; node 436 drivers/mtd/nand/raw/qcom_nandc.c struct list_head node; node 792 drivers/mtd/nand/raw/qcom_nandc.c list_add_tail(&desc->node, &nandc->desc_list); node 953 drivers/mtd/nand/raw/qcom_nandc.c list_add_tail(&desc->node, &nandc->desc_list); node 1312 drivers/mtd/nand/raw/qcom_nandc.c list_for_each_entry(desc, &nandc->desc_list, node) node 1343 drivers/mtd/nand/raw/qcom_nandc.c list_for_each_entry_safe(desc, n, &nandc->desc_list, node) { node 1344 drivers/mtd/nand/raw/qcom_nandc.c list_del(&desc->node); node 2863 drivers/mtd/nand/raw/qcom_nandc.c list_add_tail(&host->node, &nandc->host_list); node 2986 drivers/mtd/nand/raw/qcom_nandc.c list_for_each_entry(host, &nandc->host_list, node) node 192 drivers/mtd/nand/raw/sunxi_nand.c struct list_head node; node 2010 drivers/mtd/nand/raw/sunxi_nand.c list_add_tail(&sunxi_nand->node, &nfc->chips); node 2045 drivers/mtd/nand/raw/sunxi_nand.c node); node 2048 drivers/mtd/nand/raw/sunxi_nand.c list_del(&sunxi_nand->node); node 373 drivers/mtd/ubi/fastmap.c struct rb_node *node, *node2; node 376 drivers/mtd/ubi/fastmap.c ubi_rb_for_each_entry(node, av, &ai->volumes, rb) { node 305 drivers/net/arcnet/arc-rimi.c static int node; node 310 drivers/net/arcnet/arc-rimi.c module_param(node, int, 0); node 326 drivers/net/arcnet/arc-rimi.c if (node && node != 0xff) node 327 drivers/net/arcnet/arc-rimi.c dev->dev_addr[0] = node; node 368 drivers/net/arcnet/arc-rimi.c node = ints[3]; node 122 drivers/net/arcnet/com20020-isa.c static int node = 0; node 131 drivers/net/arcnet/com20020-isa.c module_param(node, int, 0); node 153 drivers/net/arcnet/com20020-isa.c if (node && node != 0xff) node 154 drivers/net/arcnet/com20020-isa.c dev->dev_addr[0] = node; node 211 drivers/net/arcnet/com20020-isa.c node = ints[3]; node 51 drivers/net/arcnet/com20020-pci.c static int node; node 58 drivers/net/arcnet/com20020-pci.c module_param(node, int, 0); node 193 drivers/net/arcnet/com20020-pci.c dev->dev_addr[0] = node; node 88 drivers/net/arcnet/com20020_cs.c static int node; node 94 drivers/net/arcnet/com20020_cs.c module_param(node, int, 0); node 136 drivers/net/arcnet/com20020_cs.c dev->dev_addr[0] = node; node 67 drivers/net/caif/caif_serial.c struct list_head node; node 310 drivers/net/caif/caif_serial.c list_for_each_entry_safe(ser, tmp, &list, node) { node 361 drivers/net/caif/caif_serial.c list_add(&ser->node, &ser_list); node 376 drivers/net/caif/caif_serial.c list_move(&ser->node, &ser_release_list); node 212 drivers/net/ethernet/amazon/ena/ena_netdev.c int size, i, node; node 221 drivers/net/ethernet/amazon/ena/ena_netdev.c node = cpu_to_node(ena_irq->cpu); node 223 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->tx_buffer_info = vzalloc_node(size, node); node 231 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->free_ids = vzalloc_node(size, node); node 239 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->push_buf_intermediate_buf = vzalloc_node(size, node); node 358 drivers/net/ethernet/amazon/ena/ena_netdev.c int size, node, i; node 370 drivers/net/ethernet/amazon/ena/ena_netdev.c node = cpu_to_node(ena_irq->cpu); node 372 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->rx_buffer_info = vzalloc_node(size, node); node 380 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->free_ids = vzalloc_node(size, node); node 188 drivers/net/ethernet/amd/xgbe/xgbe-desc.c static void *xgbe_alloc_node(size_t size, int node) node 192 drivers/net/ethernet/amd/xgbe/xgbe-desc.c mem = kzalloc_node(size, GFP_KERNEL, node); node 200 drivers/net/ethernet/amd/xgbe/xgbe-desc.c dma_addr_t *dma, int node) node 205 drivers/net/ethernet/amd/xgbe/xgbe-desc.c set_dev_node(dev, node); node 228 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ring->node); node 235 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ring->rdata = xgbe_alloc_node(size, ring->node); node 241 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ring->rdesc, &ring->rdesc_dma, ring->rdata, ring->node); node 287 drivers/net/ethernet/amd/xgbe/xgbe-desc.c int node) node 300 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pages = alloc_pages_node(node, gfp, order); node 308 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if (!pages && (node != NUMA_NO_NODE)) { node 309 drivers/net/ethernet/amd/xgbe/xgbe-desc.c node = NUMA_NO_NODE; node 363 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ret = xgbe_alloc_pages(pdata, &ring->rx_hdr_pa, 0, ring->node); node 370 drivers/net/ethernet/amd/xgbe/xgbe-desc.c PAGE_ALLOC_COSTLY_ORDER, ring->node); node 161 drivers/net/ethernet/amd/xgbe/xgbe-drv.c static void *xgbe_alloc_node(size_t size, int node) node 165 drivers/net/ethernet/amd/xgbe/xgbe-drv.c mem = kzalloc_node(size, GFP_KERNEL, node); node 196 drivers/net/ethernet/amd/xgbe/xgbe-drv.c int node; node 204 drivers/net/ethernet/amd/xgbe/xgbe-drv.c node = cpu_to_node(cpu); node 206 drivers/net/ethernet/amd/xgbe/xgbe-drv.c channel = xgbe_alloc_node(sizeof(*channel), node); node 216 drivers/net/ethernet/amd/xgbe/xgbe-drv.c channel->node = node; node 223 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring = xgbe_alloc_node(sizeof(*ring), node); node 228 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring->node = node; node 234 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring = xgbe_alloc_node(sizeof(*ring), node); node 239 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring->node = node; node 245 drivers/net/ethernet/amd/xgbe/xgbe-drv.c "%s: cpu=%u, node=%d\n", channel->name, cpu, node); node 468 drivers/net/ethernet/amd/xgbe/xgbe.h int node; node 522 drivers/net/ethernet/amd/xgbe/xgbe.h int node; node 681 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_remove_fast(&tc_info->l2_table, &l2_node->node, node 709 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_insert_fast(l2_table, &l2_node->node, node 827 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_remove_fast(tunnel_table, &tunnel_node->node, node 861 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_insert_fast(tunnel_table, &tunnel_node->node, node 926 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c &decap_l2_node->node, node 1211 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_remove_fast(&tc_info->flow_table, &flow_node->node, node 1309 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c rc = rhashtable_insert_fast(&tc_info->flow_table, &new_node->node, node 1584 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c .head_offset = offsetof(struct bnxt_tc_flow_node, node), node 1591 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c .head_offset = offsetof(struct bnxt_tc_l2_node, node), node 1598 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c .head_offset = offsetof(struct bnxt_tc_l2_node, node), node 1605 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c .head_offset = offsetof(struct bnxt_tc_tunnel_node, node), node 134 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h struct rhash_head node; node 158 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h struct rhash_head node; node 172 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h struct rhash_head node; node 270 drivers/net/ethernet/cavium/liquidio/octeon_iq.h struct list_head node; node 613 drivers/net/ethernet/cavium/liquidio/octeon_network.h struct list_head *node; node 616 drivers/net/ethernet/cavium/liquidio/octeon_network.h node = NULL; node 618 drivers/net/ethernet/cavium/liquidio/octeon_network.h node = root->next; node 620 drivers/net/ethernet/cavium/liquidio/octeon_network.h if (node) node 621 drivers/net/ethernet/cavium/liquidio/octeon_network.h list_del(node); node 623 drivers/net/ethernet/cavium/liquidio/octeon_network.h return node; node 413 drivers/net/ethernet/cavium/liquidio/request_manager.c list_add_tail(&sc->node, &oct->response_list node 766 drivers/net/ethernet/cavium/liquidio/request_manager.c list_add_tail(&sc->node, &oct->sc_buf_pool.head); node 790 drivers/net/ethernet/cavium/liquidio/request_manager.c sc = list_entry(tmp, struct octeon_soft_command, node); node 793 drivers/net/ethernet/cavium/liquidio/request_manager.c list_del(&sc->node); node 798 drivers/net/ethernet/cavium/liquidio/request_manager.c list_add_tail(&sc->node, &zombie_sc_list->head); node 826 drivers/net/ethernet/cavium/liquidio/request_manager.c sc = list_entry(tmp, struct octeon_soft_command, node); node 935 drivers/net/ethernet/cavium/liquidio/request_manager.c list_add_tail(&sc->node, &oct->sc_buf_pool.head); node 85 drivers/net/ethernet/cavium/liquidio/response_manager.c struct octeon_soft_command, node); node 137 drivers/net/ethernet/cavium/liquidio/response_manager.c list_del(&sc->node); node 146 drivers/net/ethernet/cavium/liquidio/response_manager.c list_add_tail(&sc->node, node 186 drivers/net/ethernet/cavium/liquidio/response_manager.c list_add_tail(&sc->node, node 312 drivers/net/ethernet/cavium/thunder/nic.h u8 node; node 42 drivers/net/ethernet/cavium/thunder/nic_main.c u8 node; node 173 drivers/net/ethernet/cavium/thunder/nic_main.c mac = bgx_get_lmac_mac(nic->node, bgx_idx, lmac); node 178 drivers/net/ethernet/cavium/thunder/nic_main.c mbx.nic_cfg.node_id = nic->node; node 244 drivers/net/ethernet/cavium/thunder/nic_main.c mbx.bgx_stats.stats = bgx_get_rx_stats(nic->node, bgx_idx, node 247 drivers/net/ethernet/cavium/thunder/nic_main.c mbx.bgx_stats.stats = bgx_get_tx_stats(nic->node, bgx_idx, node 268 drivers/net/ethernet/cavium/thunder/nic_main.c lmac_cnt = bgx_get_lmac_count(nic->node, bgx); node 319 drivers/net/ethernet/cavium/thunder/nic_main.c unsigned bgx_map = bgx_get_map(nic->node); node 329 drivers/net/ethernet/cavium/thunder/nic_main.c lmac_cnt = bgx_get_lmac_count(nic->node, bgx); node 770 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_lmac_internal_loopback(nic->node, bgx_idx, lmac_idx, lbk->enable); node 858 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_lmac_rx_tx_enable(nic->node, bgx, lmac, enable); node 873 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_lmac_get_pfc(nic->node, bgx, lmac, &pfc); node 880 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_lmac_set_pfc(nic->node, bgx, lmac, cfg); node 909 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_config_timestamping(nic->node, bgx_idx, lmac, true); node 915 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_config_timestamping(nic->node, bgx_idx, lmac, false); node 940 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_get_lmac_link_state(nic->node, bgx, lmac, &link); node 1036 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_set_lmac_mac(nic->node, bgx, lmac, mbx.mac.mac_addr); node 1097 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_reset_xcast_mode(nic->node, bgx, lmac, node 1109 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_set_dmac_cam_filter(nic->node, bgx, lmac, node 1122 drivers/net/ethernet/cavium/thunder/nic_main.c bgx_set_xcast_mode(nic->node, bgx, lmac, mbx.xcast.mode); node 1345 drivers/net/ethernet/cavium/thunder/nic_main.c nic->node = nic_get_node_id(pdev); node 223 drivers/net/ethernet/cavium/thunder/nicvf_main.c nic->node = mbx.nic_cfg.node_id; node 1114 drivers/net/ethernet/cavium/thunder/nicvf_main.c cpumask_set_cpu(cpumask_local_spread(cpu, nic->node), node 169 drivers/net/ethernet/cavium/thunder/thunder_bgx.c static struct bgx *get_bgx(int node, int bgx_idx) node 171 drivers/net/ethernet/cavium/thunder/thunder_bgx.c int idx = (node * max_bgx_per_node) + bgx_idx; node 177 drivers/net/ethernet/cavium/thunder/thunder_bgx.c unsigned bgx_get_map(int node) node 183 drivers/net/ethernet/cavium/thunder/thunder_bgx.c if (bgx_vnic[(node * max_bgx_per_node) + i]) node 192 drivers/net/ethernet/cavium/thunder/thunder_bgx.c int bgx_get_lmac_count(int node, int bgx_idx) node 196 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx = get_bgx(node, bgx_idx); node 205 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_get_lmac_link_state(int node, int bgx_idx, int lmacid, void *status) node 211 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx = get_bgx(node, bgx_idx); node 223 drivers/net/ethernet/cavium/thunder/thunder_bgx.c const u8 *bgx_get_lmac_mac(int node, int bgx_idx, int lmacid) node 225 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 234 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_lmac_mac(int node, int bgx_idx, int lmacid, const u8 *mac) node 236 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 331 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_dmac_cam_filter(int node, int bgx_idx, int lmacid, node 334 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 354 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_xcast_mode(int node, int bgx_idx, int lmacid, u8 mode) node 356 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 388 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_reset_xcast_mode(int node, int bgx_idx, int lmacid, u8 vf_id) node 390 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 397 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx_set_xcast_mode(node, bgx_idx, lmacid, node 402 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_rx_tx_enable(int node, int bgx_idx, int lmacid, bool enable) node 404 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 434 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_config_timestamping(int node, int bgx_idx, int lmacid, bool enable) node 436 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 462 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_get_pfc(int node, int bgx_idx, int lmacid, void *pause) node 465 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 482 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_set_pfc(int node, int bgx_idx, int lmacid, void *pause) node 485 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct bgx *bgx = get_bgx(node, bgx_idx); node 630 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx) node 634 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx = get_bgx(node, bgx_idx); node 644 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx) node 648 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx = get_bgx(node, bgx_idx); node 657 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_internal_loopback(int node, int bgx_idx, node 664 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx = get_bgx(node, bgx_idx); node 1471 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct device_node *node = NULL; node 1482 drivers/net/ethernet/cavium/thunder/thunder_bgx.c node = to_of_node(fwn); node 1483 drivers/net/ethernet/cavium/thunder/thunder_bgx.c if (!node) node 1486 drivers/net/ethernet/cavium/thunder/thunder_bgx.c mac = of_get_mac_address(node); node 1493 drivers/net/ethernet/cavium/thunder/thunder_bgx.c phy_np = of_parse_phandle(node, "phy-handle", 0); node 1509 drivers/net/ethernet/cavium/thunder/thunder_bgx.c of_node_put(node); node 1526 drivers/net/ethernet/cavium/thunder/thunder_bgx.c of_node_put(node); node 219 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_dmac_cam_filter(int node, int bgx_idx, int lmacid, u64 mac, u8 vf); node 220 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_reset_xcast_mode(int node, int bgx_idx, int lmacid, u8 vf); node 221 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_xcast_mode(int node, int bgx_idx, int lmacid, u8 mode); node 223 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_rx_tx_enable(int node, int bgx_idx, int lmacid, bool enable); node 224 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_add_dmac_addr(u64 dmac, int node, int bgx_idx, int lmac); node 225 drivers/net/ethernet/cavium/thunder/thunder_bgx.h unsigned bgx_get_map(int node); node 226 drivers/net/ethernet/cavium/thunder/thunder_bgx.h int bgx_get_lmac_count(int node, int bgx); node 227 drivers/net/ethernet/cavium/thunder/thunder_bgx.h const u8 *bgx_get_lmac_mac(int node, int bgx_idx, int lmacid); node 228 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_lmac_mac(int node, int bgx_idx, int lmacid, const u8 *mac); node 229 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_get_lmac_link_state(int node, int bgx_idx, int lmacid, void *status); node 230 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_internal_loopback(int node, int bgx_idx, node 232 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_config_timestamping(int node, int bgx_idx, int lmacid, bool enable); node 233 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_get_pfc(int node, int bgx_idx, int lmacid, void *pause); node 234 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_set_pfc(int node, int bgx_idx, int lmacid, void *pause); node 239 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx); node 240 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx); node 3564 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int node, ret; node 3611 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c node = dev_to_node(adapter->pdev_dev); node 3613 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c newpage = alloc_pages_node(node, __GFP_NOWARN | GFP_KERNEL | node 697 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c ret = rhashtable_insert_fast(&adap->flower_tbl, &ch_flower->node, node 727 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c ret = rhashtable_remove_fast(&adap->flower_tbl, &ch_flower->node, node 830 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c .head_offset = offsetof(struct ch_tc_flower_entry, node), node 51 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.h struct rhash_head node; node 600 drivers/net/ethernet/chelsio/cxgb4/sge.c int node; node 608 drivers/net/ethernet/chelsio/cxgb4/sge.c node = dev_to_node(adap->pdev_dev); node 617 drivers/net/ethernet/chelsio/cxgb4/sge.c pg = alloc_pages_node(node, gfp | __GFP_COMP, s->fl_pg_order); node 648 drivers/net/ethernet/chelsio/cxgb4/sge.c pg = alloc_pages_node(node, gfp, 0); node 715 drivers/net/ethernet/chelsio/cxgb4/sge.c size_t stat_size, int node) node 724 drivers/net/ethernet/chelsio/cxgb4/sge.c s = kcalloc_node(sw_size, nelem, GFP_KERNEL, node); node 117 drivers/net/ethernet/cisco/enic/enic.h struct hlist_node node; node 97 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_for_each_entry_safe(n, tmp, hhead, node) { node 99 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_del(&n->node); node 117 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_for_each_entry_safe(n, tmp, hhead, node) node 139 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_for_each_entry_safe(n, tmp, hhead, node) { node 146 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_del(&n->node); node 161 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_for_each_entry(tpos, h, node) node 216 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_del(&n->node); node 241 drivers/net/ethernet/cisco/enic/enic_clsf.c INIT_HLIST_NODE(&d->node); node 242 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_add_head(&d->node, head); node 277 drivers/net/ethernet/cisco/enic/enic_clsf.c INIT_HLIST_NODE(&n->node); node 278 drivers/net/ethernet/cisco/enic/enic_clsf.c hlist_add_head(&n->node, &enic->rfs_h.ht_head[tbl_idx]); node 427 drivers/net/ethernet/cisco/enic/enic_ethtool.c hlist_for_each_entry_safe(n, tmp, hhead, node) { node 114 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c struct device_node *node; node 140 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c node = of_find_compatible_node(NULL, NULL, "fsl,dpaa2-ptp"); node 141 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c if (!node) { node 146 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c dev->of_node = node; node 148 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c base = of_iomap(node, 0); node 2068 drivers/net/ethernet/freescale/fec_main.c struct device_node *node; node 2149 drivers/net/ethernet/freescale/fec_main.c node = of_get_child_by_name(pdev->dev.of_node, "mdio"); node 2150 drivers/net/ethernet/freescale/fec_main.c err = of_mdiobus_register(fep->mii_bus, node); node 2151 drivers/net/ethernet/freescale/fec_main.c of_node_put(node); node 1109 drivers/net/ethernet/freescale/fman/fman_dtsec.c INIT_LIST_HEAD(&hash_entry->node); node 1113 drivers/net/ethernet/freescale/fman/fman_dtsec.c list_add_tail(&hash_entry->node, node 1116 drivers/net/ethernet/freescale/fman/fman_dtsec.c list_add_tail(&hash_entry->node, node 1209 drivers/net/ethernet/freescale/fman/fman_dtsec.c list_del_init(&hash_entry->node); node 1222 drivers/net/ethernet/freescale/fman/fman_dtsec.c list_del_init(&hash_entry->node); node 76 drivers/net/ethernet/freescale/fman/fman_mac.h hlist_entry_safe(ptr, struct eth_hash_entry, node) node 168 drivers/net/ethernet/freescale/fman/fman_mac.h struct list_head node; node 219 drivers/net/ethernet/freescale/fman/fman_mac.h list_del_init(&hash_entry->node); node 935 drivers/net/ethernet/freescale/fman/fman_memac.c INIT_LIST_HEAD(&hash_entry->node); node 937 drivers/net/ethernet/freescale/fman/fman_memac.c list_add_tail(&hash_entry->node, node 990 drivers/net/ethernet/freescale/fman/fman_memac.c list_del_init(&hash_entry->node); node 560 drivers/net/ethernet/freescale/fman/fman_tgec.c INIT_LIST_HEAD(&hash_entry->node); node 562 drivers/net/ethernet/freescale/fman/fman_tgec.c list_add_tail(&hash_entry->node, node 634 drivers/net/ethernet/freescale/fman/fman_tgec.c list_del_init(&hash_entry->node); node 196 drivers/net/ethernet/freescale/ucc_geth.c struct list_head *node = lh->next; node 197 drivers/net/ethernet/freescale/ucc_geth.c list_del(node); node 199 drivers/net/ethernet/freescale/ucc_geth.c return node; node 3916 drivers/net/ethernet/freescale/ucc_geth.c ugeth->node = np; node 1055 drivers/net/ethernet/freescale/ucc_geth.h struct list_head node; node 1058 drivers/net/ethernet/freescale/ucc_geth.h #define ENET_ADDR_CONT_ENTRY(ptr) list_entry(ptr, struct enet_addr_container, node) node 1223 drivers/net/ethernet/freescale/ucc_geth.h struct device_node *node; node 904 drivers/net/ethernet/hisilicon/hip04_eth.c struct device_node *node = d->of_node; node 935 drivers/net/ethernet/hisilicon/hip04_eth.c ret = of_parse_phandle_with_fixed_args(node, "port-handle", 3, 0, &arg); node 964 drivers/net/ethernet/hisilicon/hip04_eth.c priv->phy_mode = of_get_phy_mode(node); node 984 drivers/net/ethernet/hisilicon/hip04_eth.c priv->phy_node = of_parse_phandle(node, "phy-handle", 0); node 783 drivers/net/ethernet/hisilicon/hisi_femac.c struct device_node *node = dev->of_node; node 837 drivers/net/ethernet/hisilicon/hisi_femac.c ret = of_property_read_u32_array(node, node 846 drivers/net/ethernet/hisilicon/hisi_femac.c phy = of_phy_get_and_connect(ndev, node, hisi_femac_adjust_link); node 857 drivers/net/ethernet/hisilicon/hisi_femac.c mac_addr = of_get_mac_address(node); node 1096 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c struct device_node *node = dev->of_node; node 1169 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c ret = of_property_read_u32_array(node, node 1192 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c ret = of_mdiobus_register(bus, node); node 1196 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c priv->phy_mode = of_get_phy_mode(node); node 1203 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c priv->phy_node = of_parse_phandle(node, "phy-handle", 0); node 1224 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c mac_addr = of_get_mac_address(node); node 18 drivers/net/ethernet/hisilicon/hns/hnae.c hnae_list_add(spinlock_t *lock, struct list_head *node, struct list_head *head) node 23 drivers/net/ethernet/hisilicon/hns/hnae.c list_add_tail_rcu(node, head); node 27 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_list_del(spinlock_t *lock, struct list_head *node) node 32 drivers/net/ethernet/hisilicon/hns/hnae.c list_del_rcu(node); node 356 drivers/net/ethernet/hisilicon/hns/hnae.c hnae_list_add(&dev->lock, &handle->node, &dev->handle_list); node 381 drivers/net/ethernet/hisilicon/hns/hnae.c hnae_list_del(&dev->lock, &h->node); node 540 drivers/net/ethernet/hisilicon/hns/hnae.h struct list_head node; node 565 drivers/net/ethernet/hisilicon/hns/hnae.h struct list_head node; /* list to hnae_ae_dev->handle_list */ node 2587 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c static char *hns_dsaf_get_node_stats_strings(char *data, int node, node 2594 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_pad_drop_pkts", node); node 2596 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_manage_pkts", node); node 2598 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pkts", node); node 2600 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pkt_id", node); node 2602 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pause_frame", node); node 2604 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_release_buf_num", node); node 2606 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_sbm_drop_pkts", node); node 2608 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_crc_false_pkts", node); node 2610 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_bp_drop_pkts", node); node 2612 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_lookup_rslt_drop_pkts", node); node 2614 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_local_rslt_fail_pkts", node); node 2616 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_vlan_drop_pkts", node); node 2618 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "innod%d_stp_drop_pkts", node); node 2620 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c if (node < DSAF_SERVICE_NW_NUM && !is_ver1) { node 2624 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c node, i); node 2627 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c node, i); node 2632 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c snprintf(buff, ETH_GSTRING_LEN, "onnod%d_tx_pkts", node); node 2719 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c int node = port; node 2725 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c buff = hns_dsaf_get_node_stats_strings(buff, node, dsaf_dev); node 2728 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c node = port + DSAF_PPE_INODE_BASE; node 2729 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c (void)hns_dsaf_get_node_stats_strings(buff, node, dsaf_dev); node 113 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client_tmp, &hnae3_client_list, node) { node 118 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_add_tail(&client->node, &hnae3_client_list); node 121 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) { node 150 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client_tmp, &hnae3_client_list, node) { node 164 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) { node 168 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_del(&client->node); node 189 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_add_tail(&ae_algo->node, &hnae3_ae_algo_list); node 192 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) { node 216 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client, &hnae3_client_list, node) { node 243 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) { node 254 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client, &hnae3_client_list, node) node 262 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_del(&ae_algo->node); node 283 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_add_tail(&ae_dev->node, &hnae3_ae_dev_list); node 286 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_algo, &hnae3_ae_algo_list, node) { node 313 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client, &hnae3_client_list, node) { node 326 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_del(&ae_dev->node); node 347 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(ae_algo, &hnae3_ae_algo_list, node) { node 355 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_for_each_entry(client, &hnae3_client_list, node) node 363 drivers/net/ethernet/hisilicon/hns3/hnae3.c list_del(&ae_dev->node); node 211 drivers/net/ethernet/hisilicon/hns3/hnae3.h #define HNAE3_IS_TX_RING(node) \ node 212 drivers/net/ethernet/hisilicon/hns3/hnae3.h (((node)->flag & (1 << HNAE3_RING_TYPE_B)) == HNAE3_RING_TYPE_TX) node 231 drivers/net/ethernet/hisilicon/hns3/hnae3.h struct list_head node; node 237 drivers/net/ethernet/hisilicon/hns3/hnae3.h struct list_head node; node 552 drivers/net/ethernet/hisilicon/hns3/hnae3.h struct list_head node; node 4555 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c INIT_LIST_HEAD(&client.node); node 4494 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c struct hnae3_ring_chain_node *node; node 4508 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c for (node = ring_chain; node; node = node->next) { node 4512 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hnae3_get_bit(node->flag, HNAE3_RING_TYPE_B)); node 4514 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c HCLGE_TQP_ID_S, node->tqp_index); node 4517 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hnae3_get_field(node->int_gl_idx, node 5681 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c struct hlist_node *node; node 5694 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, node 5713 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c struct hlist_node *node; node 5728 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) { node 6034 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c struct hlist_node *node; node 6036 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) { node 6153 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c struct hlist_node *node; node 6161 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) { node 6172 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c hlist_for_each_entry_safe(rule, node, &del_list, rule_node) { node 7326 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_add_tail(&mac_cfg->node, list); node 7343 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(mac_cfg, tmp, list, node) { node 7351 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&mac_cfg->node); node 7367 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(mac_cfg, tmp, list, node) { node 7376 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&mac_cfg->node); node 7391 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(mac, tmp, &vport->uc_mac_list, node) { node 7392 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&mac->node); node 7396 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(mac, tmp, &vport->mc_mac_list, node) { node 7397 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&mac->node); node 8016 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_add_tail(&vlan->node, &vport->vlan_list); node 8025 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) { node 8049 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) { node 8058 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&vlan->node); node 8070 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) { node 8080 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&vlan->node); node 8095 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) { node 8096 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_del(&vlan->node); node 8126 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) { node 637 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h struct list_head node; node 648 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h struct list_head node; node 997 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c struct hnae3_ring_chain_node *node; node 1008 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c for (node = ring_chain; node; node = node->next) { node 1021 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c hnae3_get_bit(node->flag, HNAE3_RING_TYPE_B); node 1022 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c req->msg[idx_offset + 1] = node->tqp_index; node 1023 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c req->msg[idx_offset + 2] = hnae3_get_field(node->int_gl_idx, node 1031 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c !node->next) { node 653 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c struct hinic_api_cmd_cell *node; node 657 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c node = dma_alloc_coherent(&pdev->dev, chain->cell_size, &node_paddr, node 659 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c if (!node) { node 664 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c node->read.hw_wb_resp_paddr = 0; node 667 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c cell_ctxt->cell_vaddr = node; node 672 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c chain->head_node = node; node 680 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c err = alloc_cmd_buf(chain, node, cell_idx); node 693 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c *node_vaddr = node; node 697 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c dma_free_coherent(&pdev->dev, chain->cell_size, node, node_paddr); node 712 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c struct hinic_api_cmd_cell *node; node 718 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c node = cell_ctxt->cell_vaddr; node 732 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c dma_free_coherent(&pdev->dev, node_size, node, node 759 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c struct hinic_api_cmd_cell *node = NULL, *pre_node = NULL; node 765 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c err = api_cmd_create_cell(chain, cell_idx, pre_node, &node); node 771 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c pre_node = node; node 775 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c node->next_cell_paddr = cpu_to_be64(chain->head_cell_paddr); node 322 drivers/net/ethernet/huawei/hinic/hinic_main.c int i, node, err = 0; node 351 drivers/net/ethernet/huawei/hinic/hinic_main.c node = cpu_to_node(i); node 352 drivers/net/ethernet/huawei/hinic/hinic_main.c if (node == dev_to_node(&pdev->dev)) node 2340 drivers/net/ethernet/ibm/emac/core.c struct device_node *node; node 2373 drivers/net/ethernet/ibm/emac/core.c if (deps[i].node == NULL) node 2374 drivers/net/ethernet/ibm/emac/core.c deps[i].node = of_node_get(np); node 2376 drivers/net/ethernet/ibm/emac/core.c if (deps[i].node == NULL) node 2377 drivers/net/ethernet/ibm/emac/core.c deps[i].node = of_find_node_by_phandle(deps[i].phandle); node 2378 drivers/net/ethernet/ibm/emac/core.c if (deps[i].node == NULL) node 2381 drivers/net/ethernet/ibm/emac/core.c deps[i].ofdev = of_find_device_by_node(deps[i].node); node 2437 drivers/net/ethernet/ibm/emac/core.c of_node_put(deps[i].node); node 3476 drivers/net/ethernet/intel/i40e/i40e_main.c struct hlist_node *node; node 3487 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(filter, node, node 5560 drivers/net/ethernet/intel/i40e/i40e_main.c struct hlist_node *node; node 5602 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(cfilter, node, node 8369 drivers/net/ethernet/intel/i40e/i40e_main.c struct hlist_node *node; node 8371 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(cfilter, node, node 8846 drivers/net/ethernet/intel/i40e/i40e_main.c struct hlist_node *node; node 8869 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(filter, node, node 9699 drivers/net/ethernet/intel/i40e/i40e_main.c struct hlist_node *node; node 9703 drivers/net/ethernet/intel/i40e/i40e_main.c hlist_for_each_entry_safe(cfilter, node, &pf->cloud_filter_list, node 3281 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c struct hlist_node *node; node 3284 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hlist_for_each_entry_safe(cfilter, node, node 3327 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c struct hlist_node *node; node 3410 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hlist_for_each_entry_safe(cf, node, node 3193 drivers/net/ethernet/intel/ice/ice_common.c struct ice_aqc_txsched_elem_data node = { 0 }; node 3229 drivers/net/ethernet/intel/ice/ice_common.c node.parent_teid = parent->info.node_teid; node 3252 drivers/net/ethernet/intel/ice/ice_common.c node.node_teid = buf->txqs[0].q_teid; node 3253 drivers/net/ethernet/intel/ice/ice_common.c node.data.elem_type = ICE_AQC_ELEM_TYPE_LEAF; node 3257 drivers/net/ethernet/intel/ice/ice_common.c status = ice_sched_add_node(pi, hw->num_tx_sched_layers - 1, &node); node 3307 drivers/net/ethernet/intel/ice/ice_common.c struct ice_sched_node *node; node 3309 drivers/net/ethernet/intel/ice/ice_common.c node = ice_sched_find_node_by_teid(pi->root, q_teids[i]); node 3310 drivers/net/ethernet/intel/ice/ice_common.c if (!node) node 3323 drivers/net/ethernet/intel/ice/ice_common.c qg_list.parent_teid = node->info.parent_teid; node 3332 drivers/net/ethernet/intel/ice/ice_common.c ice_free_sched_node(pi, node); node 1364 drivers/net/ethernet/intel/ice/ice_dcb.c struct ice_sched_node *node, *tc_node; node 1411 drivers/net/ethernet/intel/ice/ice_dcb.c node = ice_sched_find_node_by_teid(pi->root, teid2); node 1412 drivers/net/ethernet/intel/ice/ice_dcb.c if (node) node 1413 drivers/net/ethernet/intel/ice/ice_dcb.c node->tc_num = j; node 154 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node; node 180 drivers/net/ethernet/intel/ice/ice_sched.c node = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*node), GFP_KERNEL); node 181 drivers/net/ethernet/intel/ice/ice_sched.c if (!node) node 185 drivers/net/ethernet/intel/ice/ice_sched.c node->children = devm_kcalloc(ice_hw_to_dev(hw), node 187 drivers/net/ethernet/intel/ice/ice_sched.c sizeof(*node), GFP_KERNEL); node 188 drivers/net/ethernet/intel/ice/ice_sched.c if (!node->children) { node 189 drivers/net/ethernet/intel/ice/ice_sched.c devm_kfree(ice_hw_to_dev(hw), node); node 194 drivers/net/ethernet/intel/ice/ice_sched.c node->in_use = true; node 195 drivers/net/ethernet/intel/ice/ice_sched.c node->parent = parent; node 196 drivers/net/ethernet/intel/ice/ice_sched.c node->tx_sched_layer = layer; node 197 drivers/net/ethernet/intel/ice/ice_sched.c parent->children[parent->num_children++] = node; node 198 drivers/net/ethernet/intel/ice/ice_sched.c memcpy(&node->info, &elem.generic[0], sizeof(node->info)); node 304 drivers/net/ethernet/intel/ice/ice_sched.c void ice_free_sched_node(struct ice_port_info *pi, struct ice_sched_node *node) node 314 drivers/net/ethernet/intel/ice/ice_sched.c while (node->num_children) node 315 drivers/net/ethernet/intel/ice/ice_sched.c ice_free_sched_node(pi, node->children[0]); node 318 drivers/net/ethernet/intel/ice/ice_sched.c if (node->tx_sched_layer >= hw->sw_entry_point_layer && node 319 drivers/net/ethernet/intel/ice/ice_sched.c node->info.data.elem_type != ICE_AQC_ELEM_TYPE_TC && node 320 drivers/net/ethernet/intel/ice/ice_sched.c node->info.data.elem_type != ICE_AQC_ELEM_TYPE_ROOT_PORT && node 321 drivers/net/ethernet/intel/ice/ice_sched.c node->info.data.elem_type != ICE_AQC_ELEM_TYPE_LEAF) { node 322 drivers/net/ethernet/intel/ice/ice_sched.c u32 teid = le32_to_cpu(node->info.node_teid); node 324 drivers/net/ethernet/intel/ice/ice_sched.c ice_sched_remove_elems(hw, node->parent, 1, &teid); node 326 drivers/net/ethernet/intel/ice/ice_sched.c parent = node->parent; node 333 drivers/net/ethernet/intel/ice/ice_sched.c if (parent->children[i] == node) { node 341 drivers/net/ethernet/intel/ice/ice_sched.c p = ice_sched_get_first_node(pi, node, node->tx_sched_layer); node 343 drivers/net/ethernet/intel/ice/ice_sched.c if (p->sibling == node) { node 344 drivers/net/ethernet/intel/ice/ice_sched.c p->sibling = node->sibling; node 351 drivers/net/ethernet/intel/ice/ice_sched.c if (pi->sib_head[node->tc_num][node->tx_sched_layer] == node) node 352 drivers/net/ethernet/intel/ice/ice_sched.c pi->sib_head[node->tc_num][node->tx_sched_layer] = node 353 drivers/net/ethernet/intel/ice/ice_sched.c node->sibling; node 357 drivers/net/ethernet/intel/ice/ice_sched.c if (node->children) node 358 drivers/net/ethernet/intel/ice/ice_sched.c devm_kfree(ice_hw_to_dev(hw), node->children); node 359 drivers/net/ethernet/intel/ice/ice_sched.c devm_kfree(ice_hw_to_dev(hw), node); node 878 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node; node 880 drivers/net/ethernet/intel/ice/ice_sched.c node = pi->root; node 881 drivers/net/ethernet/intel/ice/ice_sched.c while (node) { node 882 drivers/net/ethernet/intel/ice/ice_sched.c if (!node->num_children) node 884 drivers/net/ethernet/intel/ice/ice_sched.c node = node->children[0]; node 886 drivers/net/ethernet/intel/ice/ice_sched.c if (node && node->info.data.elem_type == ICE_AQC_ELEM_TYPE_LEAF) { node 887 drivers/net/ethernet/intel/ice/ice_sched.c u32 teid = le32_to_cpu(node->info.node_teid); node 891 drivers/net/ethernet/intel/ice/ice_sched.c status = ice_sched_remove_elems(pi->hw, node->parent, 1, &teid); node 893 drivers/net/ethernet/intel/ice/ice_sched.c ice_free_sched_node(pi, node); node 906 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node; node 911 drivers/net/ethernet/intel/ice/ice_sched.c node = pi->root; node 912 drivers/net/ethernet/intel/ice/ice_sched.c while (node) { node 913 drivers/net/ethernet/intel/ice/ice_sched.c if (node->tx_sched_layer >= pi->hw->sw_entry_point_layer && node 914 drivers/net/ethernet/intel/ice/ice_sched.c node->info.data.elem_type != ICE_AQC_ELEM_TYPE_TC && node 915 drivers/net/ethernet/intel/ice/ice_sched.c node->info.data.elem_type != ICE_AQC_ELEM_TYPE_ROOT_PORT) { node 916 drivers/net/ethernet/intel/ice/ice_sched.c ice_free_sched_node(pi, node); node 920 drivers/net/ethernet/intel/ice/ice_sched.c if (!node->num_children) node 922 drivers/net/ethernet/intel/ice/ice_sched.c node = node->children[0]; node 1095 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node) node 1102 drivers/net/ethernet/intel/ice/ice_sched.c if (node == child) node 1105 drivers/net/ethernet/intel/ice/ice_sched.c if (child->tx_sched_layer > node->tx_sched_layer) node 1111 drivers/net/ethernet/intel/ice/ice_sched.c if (ice_sched_find_node_in_subtree(hw, child, node)) node 1174 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node; node 1178 drivers/net/ethernet/intel/ice/ice_sched.c node = ice_sched_get_first_node(hw->port_info, tc_node, vsi_layer); node 1181 drivers/net/ethernet/intel/ice/ice_sched.c while (node) { node 1182 drivers/net/ethernet/intel/ice/ice_sched.c if (node->vsi_handle == vsi_handle) node 1183 drivers/net/ethernet/intel/ice/ice_sched.c return node; node 1184 drivers/net/ethernet/intel/ice/ice_sched.c node = node->sibling; node 1187 drivers/net/ethernet/intel/ice/ice_sched.c return node; node 1234 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *parent, *node; node 1261 drivers/net/ethernet/intel/ice/ice_sched.c node = parent; node 1262 drivers/net/ethernet/intel/ice/ice_sched.c while (node) { node 1263 drivers/net/ethernet/intel/ice/ice_sched.c node->owner = owner; node 1264 drivers/net/ethernet/intel/ice/ice_sched.c node = node->sibling; node 1288 drivers/net/ethernet/intel/ice/ice_sched.c struct ice_sched_node *node; node 1303 drivers/net/ethernet/intel/ice/ice_sched.c node = ice_sched_get_first_node(hw->port_info, tc_node, node 1306 drivers/net/ethernet/intel/ice/ice_sched.c while (node) { node 1307 drivers/net/ethernet/intel/ice/ice_sched.c if (node->num_children < hw->max_children[i]) node 1309 drivers/net/ethernet/intel/ice/ice_sched.c node = node->sibling; node 1316 drivers/net/ethernet/intel/ice/ice_sched.c if (node) node 1578 drivers/net/ethernet/intel/ice/ice_sched.c static bool ice_sched_is_leaf_node_present(struct ice_sched_node *node) node 1582 drivers/net/ethernet/intel/ice/ice_sched.c for (i = 0; i < node->num_children; i++) node 1583 drivers/net/ethernet/intel/ice/ice_sched.c if (ice_sched_is_leaf_node_present(node->children[i])) node 1586 drivers/net/ethernet/intel/ice/ice_sched.c return (node->info.data.elem_type == ICE_AQC_ELEM_TYPE_LEAF); node 42 drivers/net/ethernet/intel/ice/ice_sched.h void ice_free_sched_node(struct ice_port_info *pi, struct ice_sched_node *node); node 837 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c int node = NUMA_NO_NODE; node 850 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c node = cpu_to_node(cpu); node 857 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c GFP_KERNEL, node); node 867 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->numa_node = node; node 3728 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_online(unsigned int cpu, struct hlist_node *node) node 3731 drivers/net/ethernet/marvell/mvneta.c struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port, node 3786 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_down_prepare(unsigned int cpu, struct hlist_node *node) node 3788 drivers/net/ethernet/marvell/mvneta.c struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port, node 3808 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_dead(unsigned int cpu, struct hlist_node *node) node 3810 drivers/net/ethernet/marvell/mvneta.c struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port, node 396 drivers/net/ethernet/marvell/mvneta_bm.c struct mvneta_bm *mvneta_bm_get(struct device_node *node) node 398 drivers/net/ethernet/marvell/mvneta_bm.c struct platform_device *pdev = of_find_device_by_node(node); node 134 drivers/net/ethernet/marvell/mvneta_bm.h struct mvneta_bm *mvneta_bm_get(struct device_node *node); node 188 drivers/net/ethernet/marvell/mvneta_bm.h static inline struct mvneta_bm *mvneta_bm_get(struct device_node *node) node 352 drivers/net/ethernet/marvell/octeontx2/af/mbox.h int node; node 435 drivers/net/ethernet/marvell/octeontx2/af/mbox.h int node; node 66 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c struct hlist_node node; node 1659 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c hlist_for_each_entry(mce, &mce_list->head, node) { node 1669 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c hlist_del(&mce->node); node 1685 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c hlist_add_head(&mce->node, &mce_list->head); node 1687 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c hlist_add_behind(&mce->node, &tail->node); node 1742 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c hlist_for_each_entry(mce, &mce_list->head, node) { node 1746 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c next_mce = hlist_entry(mce->node.next, node 1747 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c struct mce, node); node 49 drivers/net/ethernet/mellanox/mlx4/en_cq.c int node) node 55 drivers/net/ethernet/mellanox/mlx4/en_cq.c cq = kzalloc_node(sizeof(*cq), GFP_KERNEL, node); node 71 drivers/net/ethernet/mellanox/mlx4/en_cq.c set_dev_node(&mdev->dev->persist->pdev->dev, node); node 2143 drivers/net/ethernet/mellanox/mlx4/en_netdev.c int node; node 2148 drivers/net/ethernet/mellanox/mlx4/en_netdev.c node = cpu_to_node(i % num_online_cpus()); node 2150 drivers/net/ethernet/mellanox/mlx4/en_netdev.c prof->tx_ring_size, i, t, node)) node 2155 drivers/net/ethernet/mellanox/mlx4/en_netdev.c TXBB_SIZE, node, i)) node 2162 drivers/net/ethernet/mellanox/mlx4/en_netdev.c node = cpu_to_node(i % num_online_cpus()); node 2164 drivers/net/ethernet/mellanox/mlx4/en_netdev.c prof->rx_ring_size, i, RX, node)) node 2169 drivers/net/ethernet/mellanox/mlx4/en_netdev.c node, i)) node 265 drivers/net/ethernet/mellanox/mlx4/en_rx.c u32 size, u16 stride, int node, int queue_index) node 272 drivers/net/ethernet/mellanox/mlx4/en_rx.c ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node); node 291 drivers/net/ethernet/mellanox/mlx4/en_rx.c ring->rx_info = kvzalloc_node(tmp, GFP_KERNEL, node); node 301 drivers/net/ethernet/mellanox/mlx4/en_rx.c set_dev_node(&mdev->dev->persist->pdev->dev, node); node 52 drivers/net/ethernet/mellanox/mlx4/en_tx.c u16 stride, int node, int queue_index) node 59 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node); node 71 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->tx_info = kvmalloc_node(tmp, GFP_KERNEL, node); node 80 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->bounce_buf = kmalloc_node(MAX_DESC_SIZE, GFP_KERNEL, node); node 91 drivers/net/ethernet/mellanox/mlx4/en_tx.c set_dev_node(&mdev->dev->persist->pdev->dev, node); node 120 drivers/net/ethernet/mellanox/mlx4/en_tx.c err = mlx4_bf_alloc(mdev->dev, &ring->bf, node); node 99 drivers/net/ethernet/mellanox/mlx4/icm.c gfp_t gfp_mask, int node) node 103 drivers/net/ethernet/mellanox/mlx4/icm.c page = alloc_pages_node(node, gfp_mask, order); node 691 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h int entries, int ring, enum cq_type mode, int node); node 714 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h int node, int queue_index); node 728 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h u32 size, u16 stride, int node, int queue_index); node 172 drivers/net/ethernet/mellanox/mlx4/pd.c int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf, int node) node 190 drivers/net/ethernet/mellanox/mlx4/pd.c uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node); node 73 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct rb_node node; node 237 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct rb_node *node = root->rb_node; node 239 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c while (node) { node 240 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_common *res = rb_entry(node, struct res_common, node 241 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c node); node 244 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c node = node->rb_left; node 246 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c node = node->rb_right; node 260 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c node); node 272 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_link_node(&res->node, parent, new); node 273 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_insert_color(&res->node, root); node 1322 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&res_arr[i]->node, root); node 1492 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&r->node, &tracker->res_tree[type]); node 4688 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&qp->com.node, node 4761 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&srq->com.node, node 4826 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&cq->com.node, node 4888 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&mpt->com.node, node 4956 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&mtt->com.node, node 5036 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c fs_rule = rb_entry(p, struct res_fs_rule, com.node); node 5095 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&fs_rule->com.node, node 5141 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&eq->com.node, node 5200 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&counter->com.node, node 5237 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&xrcd->com.node, &tracker->res_tree[RES_XRCD]); node 57 drivers/net/ethernet/mellanox/mlx5/core/alloc.c int node) node 66 drivers/net/ethernet/mellanox/mlx5/core/alloc.c set_dev_node(device, node); node 75 drivers/net/ethernet/mellanox/mlx5/core/alloc.c struct mlx5_frag_buf *buf, int node) node 88 drivers/net/ethernet/mellanox/mlx5/core/alloc.c &t, node); node 122 drivers/net/ethernet/mellanox/mlx5/core/alloc.c struct mlx5_frag_buf *buf, int node) node 139 drivers/net/ethernet/mellanox/mlx5/core/alloc.c &frag->map, node); node 181 drivers/net/ethernet/mellanox/mlx5/core/alloc.c int node) node 199 drivers/net/ethernet/mellanox/mlx5/core/alloc.c &pgdir->db_dma, node); node 234 drivers/net/ethernet/mellanox/mlx5/core/alloc.c int mlx5_db_alloc_node(struct mlx5_core_dev *dev, struct mlx5_db *db, int node) node 245 drivers/net/ethernet/mellanox/mlx5/core/alloc.c pgdir = mlx5_alloc_db_pgdir(dev, node); node 115 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h fs_get_obj(__entry->ft, fg->node.parent); node 203 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h fs_get_obj(__entry->fg, fte->node.parent); node 286 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h fs_get_obj(__entry->fte, rule->node.parent); node 312 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h fs_get_obj(__entry->fte, rule->node.parent); node 113 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct rhash_head node; node 3484 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c .head_offset = offsetof(struct mlx5e_tc_flow, node), node 3825 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c err = rhashtable_lookup_insert_fast(tc_ht, &flow->node, tc_ht_params); node 3867 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c rhashtable_remove_fast(tc_ht, &flow->node, tc_ht_params); node 52 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c struct l2addr_node node; node 497 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = vaddr->node.addr; node 529 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = vaddr->node.addr; node 558 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = vaddr->node.addr; node 606 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = vaddr->node.addr; node 646 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = vaddr->node.addr; node 692 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c struct l2addr_node *node; node 703 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c for_each_l2hash_node(node, tmp, hash, hi) { node 704 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c addr = container_of(node, struct vport_addr, node); node 724 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c struct l2addr_node *node; node 742 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c for_each_l2hash_node(node, tmp, hash, hi) { node 743 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c addr = container_of(node, struct vport_addr, node); node 809 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c struct l2addr_node *node; node 817 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c for_each_l2hash_node(node, tmp, esw->mc_table, hi) { node 818 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u8 *mac = node->addr; node 198 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h struct l2addr_node node; node 80 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c struct rb_node node; node 764 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c fs_get_obj(fg, fte->node.parent); node 862 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c struct rb_node *node = root->rb_node; node 864 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c while (node) { node 866 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c container_of(node, struct mlx5_fpga_ipsec_rule, node 867 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c node); node 870 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c node = node->rb_left; node 872 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c node = node->rb_right; node 900 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c node); node 912 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rb_link_node(&rule->node, parent, new); node 913 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rb_insert_color(&rule->node, root); node 936 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rb_erase(&rule->node, root); node 1320 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rbtree_postorder_for_each_entry_safe(r, tmp, root, node) { node 1321 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c rb_erase(&r->node, root); node 376 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c list_for_each_entry(dst, &fte->node.children, node.list) { node 487 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c list_for_each_entry(dst, &fte->node.children, node.list) { node 542 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c list_for_each_entry(dst, &fte->node.children, node.list) { node 228 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_table(struct fs_node *node); node 229 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_group(struct fs_node *node); node 230 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_fte(struct fs_node *node); node 231 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_table(struct fs_node *node); node 232 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_group(struct fs_node *node); node 233 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_fte(struct fs_node *node); node 234 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_prio(struct fs_node *node); node 235 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_ns(struct fs_node *node); node 239 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_hw_rule(struct fs_node *node); node 247 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_init_node(struct fs_node *node, node 251 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_set(&node->refcount, 1); node 252 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c INIT_LIST_HEAD(&node->list); node 253 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c INIT_LIST_HEAD(&node->children); node 254 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c init_rwsem(&node->lock); node 255 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->del_hw_func = del_hw_func; node 256 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->del_sw_func = del_sw_func; node 257 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->active = false; node 260 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_add_node(struct fs_node *node, struct fs_node *parent) node 264 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->parent = parent; node 268 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->root = node; node 270 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->root = parent->root; node 273 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static int tree_get_node(struct fs_node *node) node 275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return refcount_inc_not_zero(&node->refcount); node 278 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void nested_down_read_ref_node(struct fs_node *node, node 281 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node) { node 282 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_read_nested(&node->lock, class); node 283 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_inc(&node->refcount); node 287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void nested_down_write_ref_node(struct fs_node *node, node 290 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node) { node 291 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write_nested(&node->lock, class); node 292 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_inc(&node->refcount); node 296 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void down_write_ref_node(struct fs_node *node, bool locked) node 298 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node) { node 300 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write(&node->lock); node 301 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_inc(&node->refcount); node 305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void up_read_ref_node(struct fs_node *node) node 307 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_dec(&node->refcount); node 308 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_read(&node->lock); node 311 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void up_write_ref_node(struct fs_node *node, bool locked) node 313 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_dec(&node->refcount); node 315 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write(&node->lock); node 318 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_put_node(struct fs_node *node, bool locked) node 320 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct fs_node *parent_node = node->parent; node 322 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (refcount_dec_and_test(&node->refcount)) { node 323 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node->del_hw_func) node 324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->del_hw_func(node); node 330 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_del_init(&node->list); node 331 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node->del_sw_func) node 332 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->del_sw_func(node); node 335 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c kfree(node); node 337 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node = NULL; node 339 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!node && parent_node) node 343 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static int tree_remove_node(struct fs_node *node, bool locked) node 345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (refcount_read(&node->refcount) > 1) { node 346 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_dec(&node->refcount); node 349 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(node, locked); node 379 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_root_namespace *find_root(struct fs_node *node) node 384 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = node->root; node 391 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ns = container_of(root, struct mlx5_flow_namespace, node); node 395 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static inline struct mlx5_flow_steering *get_steering(struct fs_node *node) node 397 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(node); node 404 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static inline struct mlx5_core_dev *get_dev(struct fs_node *node) node 406 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(node); node 413 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_ns(struct fs_node *node) node 415 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c kfree(node); node 418 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_prio(struct fs_node *node) node 420 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c kfree(node); node 423 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_table(struct fs_node *node) node 430 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, node); node 431 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c dev = get_dev(&ft->node); node 432 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ft->node); node 435 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node->active) { node 442 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_table(struct fs_node *node) node 447 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, node); node 450 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); node 463 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, fte->node.parent); node 464 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 465 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c dev = get_dev(&fte->node); node 467 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ft->node); node 476 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_hw_rule(struct fs_node *node) node 481 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(rule, node); node 482 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fte, rule->node.parent); node 508 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_fte(struct fs_node *node) node 517 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fte, node); node 518 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, fte->node.parent); node 519 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 522 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c dev = get_dev(&ft->node); node 523 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ft->node); node 524 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node->active) { node 530 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node->active = 0; node 534 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_fte(struct fs_node *node) node 536 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(node); node 541 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fte, node); node 542 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, fte->node.parent); node 552 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_group(struct fs_node *node) node 559 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, node); node 560 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 561 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c dev = get_dev(&ft->node); node 564 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ft->node); node 565 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (fg->node.active && root->cmds->destroy_flow_group(root, ft, fg)) node 570 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_group(struct fs_node *node) node 572 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(node); node 577 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, node); node 578 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 607 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&fte->node, &fg->node); node 608 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add_tail(&fte->node.list, &fg->node.children); node 620 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(&ft->node); node 628 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fte->node.type = FS_TYPE_FLOW_ENTRY; node 632 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&fte->node, NULL, del_sw_fte); node 667 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fg->node.type = FS_TYPE_FLOW_GROUP; node 681 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(&ft->node); node 699 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&fg->node, del_hw_flow_group, del_sw_flow_group); node 700 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&fg->node, &ft->node); node 702 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add(&fg->node.list, prev); node 703 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c atomic_inc(&ft->node.version); node 727 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft->node.type = FS_TYPE_FLOW_TABLE; node 784 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c parent = prio->node.parent; node 785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c curr_node = &prio->node; node 810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&prio->node); node 841 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prev_prio, prev_ft->node.parent); node 850 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&prio->node); node 896 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fte, rule->node.parent); node 899 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write_ref_node(&fte->node, false); node 900 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fg, fte->node.parent); node 901 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 904 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ft->node); node 907 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fte->node, false); node 972 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (list_empty(&prio->node.children)) { node 992 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev = &prio->node.children; node 998 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev = &iter->node.list; node 1000 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add(&ft->node.list, prev); node 1008 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ns->node); node 1044 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&ft->node, del_hw_flow_table, del_sw_flow_table); node 1055 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft->node.active = true; node 1056 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write_ref_node(&fs_prio->node, false); node 1057 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&ft->node, &fs_prio->node); node 1060 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fs_prio->node, false); node 1139 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ft->node); node 1155 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write_ref_node(&ft->node, false); node 1158 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ft->node.children.prev); node 1159 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&ft->node, false); node 1165 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fg->node, false); node 1169 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fg->node.active = true; node 1183 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c rule->node.type = FS_TYPE_FLOW_DEST; node 1209 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (refcount_dec_and_test(&handle->rule[i]->node.refcount)) { node 1211 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_del(&handle->rule[i]->node.list); node 1240 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c refcount_inc(&rule->node.refcount); node 1253 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&rule->node, NULL, del_sw_hw_rule); node 1256 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add(&rule->node.list, &fte->node.children); node 1258 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add_tail(&rule->node.list, &fte->node.children); node 1300 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(ft, fg->node.parent); node 1301 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&fg->node); node 1309 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fte->node.active = true; node 1311 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c atomic_inc(&fte->node.version); node 1324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct list_head *prev = &ft->node.children; node 1345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c prev = &fg->node.list; node 1370 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ft->node); node 1402 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fg->node.active = true; node 1439 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_for_each_entry(rule, &fte->node.children, node.list) { node 1473 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c mlx5_core_warn(get_dev(&fte->node), node 1480 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c mlx5_core_warn(get_dev(&fte->node), node 1517 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (refcount_read(&handle->rule[i]->node.refcount) == 1) { node 1518 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&handle->rule[i]->node, &fte->node); node 1564 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&head->first.g->node, ft_locked); node 1567 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&iter->g->node, ft_locked); node 1592 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!tree_get_node(&g->node)) node 1606 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!tree_get_node(&g->node)) { node 1624 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version += (u64)atomic_read(&iter->g->node.version); node 1636 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&g->node, FS_LOCK_PARENT); node 1638 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_read_ref_node(&g->node, FS_LOCK_PARENT); node 1641 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!fte_tmp || !tree_get_node(&fte_tmp->node)) { node 1645 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!fte_tmp->node.active) { node 1646 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fte_tmp->node, false); node 1651 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&fte_tmp->node, FS_LOCK_CHILD); node 1654 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&g->node, false); node 1656 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_read_ref_node(&g->node); node 1669 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(&ft->node); node 1695 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fte_tmp->node, false); node 1696 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fte_tmp->node, false); node 1709 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (atomic_read(&ft->node.version) != ft_version) { node 1725 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!g->node.active) node 1728 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&g->node, FS_LOCK_PARENT); node 1732 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&g->node, false); node 1739 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&fte->node, FS_LOCK_CHILD); node 1740 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&g->node, false); node 1742 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fte->node, false); node 1743 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fte->node, false); node 1760 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_steering *steering = get_steering(&ft->node); node 1777 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_read_ref_node(&ft->node, FS_LOCK_GRANDPARENT); node 1779 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version = atomic_read(&ft->node.version); node 1785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&ft->node, false); node 1787 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_read_ref_node(&ft->node); node 1792 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_read_ref_node(&ft->node); node 1800 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&ft->node, false); node 1805 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&ft->node, FS_LOCK_GRANDPARENT); node 1810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version != atomic_read(&ft->node.version)) node 1816 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&ft->node, false); node 1820 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&g->node, FS_LOCK_PARENT); node 1821 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&ft->node, false); node 1839 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c nested_down_write_ref_node(&fte->node, FS_LOCK_CHILD); node 1840 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&g->node, false); node 1842 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fte->node, false); node 1843 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fte->node, false); node 1844 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&g->node, false); node 1848 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&g->node, false); node 1849 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&g->node, false); node 1856 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c (MLX5_CAP_FLOWTABLE(get_dev(&ft->node), nic_rx_multi_path_tirs))); node 1866 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ft->node); node 1873 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); node 1927 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(fte, handle->rule[0]->node.parent); node 1928 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c down_write_ref_node(&fte->node, false); node 1930 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_remove_node(&handle->rule[i]->node, true); node 1933 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write_ref_node(&fte->node, false); node 1935 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c del_hw_fte(&fte->node); node 1936 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c up_write(&fte->node.lock); node 1937 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(&fte->node, false); node 1948 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); node 1950 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!list_is_last(&ft->node.list, &prio->node.children)) node 1951 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return list_next_entry(ft, node.list); node 1957 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ft->node); node 2003 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_core_dev *dev = get_dev(&ft->node); node 2012 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_get_obj(prio, ft->node.parent); node 2013 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (!(list_first_entry(&prio->node.children, node 2015 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node.list) == ft)) node 2032 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c struct mlx5_flow_root_namespace *root = find_root(&ft->node); node 2041 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (tree_remove_node(&ft->node, false)) node 2042 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c mlx5_core_warn(get_dev(&ft->node), "Flow table %d wasn't destroyed, refcount > 1\n", node 2052 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (tree_remove_node(&fg->node, false)) node 2053 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c mlx5_core_warn(get_dev(&fg->node), "Flow group %d wasn't destroyed, refcount > 1\n", node 2118 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ns = list_first_entry(&fs_prio->node.children, node 2120 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c node.list); node 2164 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c fs_prio->node.type = type; node 2165 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&fs_prio->node, NULL, del_sw_prio); node 2166 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&fs_prio->node, &ns->node); node 2169 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add_tail(&fs_prio->node.list, &ns->node.children); node 2190 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c ns->node.type = FS_TYPE_NAMESPACE; node 2206 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&ns->node, NULL, del_sw_ns); node 2207 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&ns->node, &prio->node); node 2208 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_add_tail(&ns->node.list, &prio->node.children); node 2269 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c base = &fs_prio->node; node 2275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c base = &fs_ns->node; node 2305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c &fs_ns->node, node 2339 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_init_node(&ns->node, NULL, NULL); node 2340 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_add_node(&ns->node, NULL); node 2418 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c err = init_root_tree(steering, &root_fs, &steering->root_ns->ns.node); node 2435 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void clean_tree(struct fs_node *node) node 2437 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (node) { node 2441 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_get_node(node); node 2442 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c list_for_each_entry_safe(iter, temp, &node->children, list) node 2444 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_put_node(node, false); node 2445 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c tree_remove_node(node, false); node 2454 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c clean_tree(&root_ns->ns.node); node 2543 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c &steering->rdma_rx_root_ns->ns.node); node 2722 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c &steering->egress_root_ns->ns.node); node 2909 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c return find_root(&ns->node); node 3021 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c root = find_root(&ns->node); node 138 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 154 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 196 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 211 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 221 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 232 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct fs_node node; node 275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_get_obj(v, _node) {v = container_of((_node), typeof(*v), node); } node 278 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry(pos, root, node.list) node 281 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry_safe(pos, tmp, root, node.list) node 284 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry_reverse(pos, &(prio)->node.children, list) node 287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h list_for_each_entry(pos, (&(prio)->node.children), list) node 290 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(ns)->node.children) node 293 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(prio)->node.children) node 296 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(prio)->node.children) node 299 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry_safe(pos, tmp, &(prio)->node.children) node 302 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(ft)->node.children) node 305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(fg)->node.children) node 308 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h fs_list_for_each_entry(pos, &(fte)->node.children) node 75 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c struct qpn_to_netdev *node; node 77 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c hlist_for_each_entry(node, h, hlist) { node 78 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c if (node->underlay_qpn == qpn) node 79 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c return node; node 110 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c struct qpn_to_netdev *node; node 112 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); node 113 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c if (!node) { node 119 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c hlist_del_init(&node->hlist); node 121 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c kfree(node); node 129 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c struct qpn_to_netdev *node; node 131 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); node 132 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c if (!node) node 135 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c return node->netdev; node 68 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c struct l2addr_node node; node 57 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h hlist_for_each_entry(ptr, &(hash)[ix], node.hlist) \ node 58 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h if (ether_addr_equal(ptr->node.addr, mac)) {\ node 73 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h ether_addr_copy(ptr->node.addr, mac); \ node 74 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h hlist_add_head(&ptr->node.hlist, &(hash)[ix]);\ node 80 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h hlist_del(&(ptr)->node.hlist); \ node 353 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c list_for_each_entry(dst, &fte->node.children, node.list) { node 391 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c list_for_each_entry(dst, &fte->node.children, node.list) { node 63 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c struct list_head node; node 76 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c struct list_head node; node 377 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_del(&mr_route->node); node 419 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_add_tail(&mr_route->node, &mr_table->route_list); node 436 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_del(&mr_orig_route->node); node 447 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_del(&mr_route->node); node 930 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_add_tail(&mr_table->node, &mr->table_list); node 946 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_del(&mr_table->node); node 958 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_for_each_entry_safe(mr_route, tmp, &mr_table->route_list, node) node 1004 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_for_each_entry(mr_table, &mr->table_list, node) node 1005 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c list_for_each_entry(mr_route, &mr_table->route_list, node) node 461 drivers/net/ethernet/moxa/moxart_ether.c struct device_node *node = p_dev->of_node; node 472 drivers/net/ethernet/moxa/moxart_ether.c irq = irq_of_parse_and_map(node, 0); node 741 drivers/net/ethernet/qlogic/qed/qed_fcoe.c struct hlist_node node; node 782 drivers/net/ethernet/qlogic/qed/qed_fcoe.c hash_for_each_possible(cdev->connections, hash_con, node, handle) { node 898 drivers/net/ethernet/qlogic/qed/qed_fcoe.c hash_add(cdev->connections, &hash_con->node, *handle); node 918 drivers/net/ethernet/qlogic/qed/qed_fcoe.c hlist_del(&hash_con->node); node 1078 drivers/net/ethernet/qlogic/qed/qed_iscsi.c struct hlist_node node; node 1117 drivers/net/ethernet/qlogic/qed/qed_iscsi.c hash_for_each_possible(cdev->connections, hash_con, node, handle) { node 1227 drivers/net/ethernet/qlogic/qed/qed_iscsi.c hash_add(cdev->connections, &hash_con->node, *handle); node 1247 drivers/net/ethernet/qlogic/qed/qed_iscsi.c hlist_del(&hash_con->node); node 94 drivers/net/ethernet/qlogic/qede/qede_filter.c struct hlist_node node; node 175 drivers/net/ethernet/qlogic/qede/qede_filter.c INIT_HLIST_NODE(&fltr->node); node 176 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_add_head(&fltr->node, node 194 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_del(&fltr->node); node 267 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_for_each_entry_safe(fltr, temp, head, node) { node 415 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_for_each_entry(tpos, h, node) node 1358 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_for_each_entry(fltr, head, node) node 1383 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_for_each_entry(fltr, head, node) { node 1678 drivers/net/ethernet/qlogic/qede/qede_filter.c hlist_for_each_entry_safe(fltr, temp, head, node) { node 1149 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct list_head *node; node 1155 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_for_each(node, &vf->rcv_pend.wait_list) { node 1156 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c trans = list_entry(node, struct qlcnic_bc_trans, list); node 678 drivers/net/ethernet/qualcomm/emac/emac-mac.c int node = dev_to_node(adpt->netdev->dev.parent); node 682 drivers/net/ethernet/qualcomm/emac/emac-mac.c tx_q->tpd.tpbuff = kzalloc_node(size, GFP_KERNEL, node); node 719 drivers/net/ethernet/qualcomm/emac/emac-mac.c int node = dev_to_node(adpt->netdev->dev.parent); node 724 drivers/net/ethernet/qualcomm/emac/emac-mac.c rx_q->rfd.rfbuff = kzalloc_node(size, GFP_KERNEL, node); node 5540 drivers/net/ethernet/realtek/r8169_main.c int node = dev_to_node(d); node 5544 drivers/net/ethernet/realtek/r8169_main.c data = alloc_pages_node(node, GFP_KERNEL, get_order(R8169_RX_BUF_SIZE)); node 87 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c struct device_node *node = dev->of_node; node 115 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c priv->irq = irq_of_parse_and_map(node, 0); node 127 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c priv->txq[i]->irq_no = irq_of_parse_and_map(node, chan++); node 135 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c priv->rxq[i]->irq_no = irq_of_parse_and_map(node, chan++); node 142 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c priv->lpi_irq = irq_of_parse_and_map(node, chan); node 1228 drivers/net/ethernet/sfc/efx.c list_add_tail(&efx->node, &efx_primary_list); node 1231 drivers/net/ethernet/sfc/efx.c node) { node 1233 drivers/net/ethernet/sfc/efx.c list_del(&other->node); node 1238 drivers/net/ethernet/sfc/efx.c list_add_tail(&other->node, node 1246 drivers/net/ethernet/sfc/efx.c list_for_each_entry(other, &efx_primary_list, node) { node 1252 drivers/net/ethernet/sfc/efx.c list_add_tail(&efx->node, node 1261 drivers/net/ethernet/sfc/efx.c list_add_tail(&efx->node, &efx_unassociated_list); node 1269 drivers/net/ethernet/sfc/efx.c list_del(&efx->node); node 1272 drivers/net/ethernet/sfc/efx.c list_for_each_entry_safe(other, next, &efx->secondary_list, node) { node 1273 drivers/net/ethernet/sfc/efx.c list_del(&other->node); node 1276 drivers/net/ethernet/sfc/efx.c list_add_tail(&other->node, &efx_unassociated_list); node 3009 drivers/net/ethernet/sfc/efx.c INIT_LIST_HEAD(&efx->node); node 3179 drivers/net/ethernet/sfc/efx.c struct hlist_node *node; node 3184 drivers/net/ethernet/sfc/efx.c hlist_for_each(node, head) { node 3185 drivers/net/ethernet/sfc/efx.c rule = container_of(node, struct efx_arfs_rule, node); node 3198 drivers/net/ethernet/sfc/efx.c struct hlist_node *node; node 3203 drivers/net/ethernet/sfc/efx.c hlist_for_each(node, head) { node 3204 drivers/net/ethernet/sfc/efx.c rule = container_of(node, struct efx_arfs_rule, node); node 3214 drivers/net/ethernet/sfc/efx.c hlist_add_head(&rule->node, head); node 3223 drivers/net/ethernet/sfc/efx.c struct hlist_node *node; node 3228 drivers/net/ethernet/sfc/efx.c hlist_for_each(node, head) { node 3229 drivers/net/ethernet/sfc/efx.c rule = container_of(node, struct efx_arfs_rule, node); node 3239 drivers/net/ethernet/sfc/efx.c hlist_del(node); node 1168 drivers/net/ethernet/sfc/falcon/efx.c list_add_tail(&efx->node, &ef4_primary_list); node 1171 drivers/net/ethernet/sfc/falcon/efx.c node) { node 1173 drivers/net/ethernet/sfc/falcon/efx.c list_del(&other->node); node 1178 drivers/net/ethernet/sfc/falcon/efx.c list_add_tail(&other->node, node 1186 drivers/net/ethernet/sfc/falcon/efx.c list_for_each_entry(other, &ef4_primary_list, node) { node 1192 drivers/net/ethernet/sfc/falcon/efx.c list_add_tail(&efx->node, node 1201 drivers/net/ethernet/sfc/falcon/efx.c list_add_tail(&efx->node, &ef4_unassociated_list); node 1209 drivers/net/ethernet/sfc/falcon/efx.c list_del(&efx->node); node 1212 drivers/net/ethernet/sfc/falcon/efx.c list_for_each_entry_safe(other, next, &efx->secondary_list, node) { node 1213 drivers/net/ethernet/sfc/falcon/efx.c list_del(&other->node); node 1216 drivers/net/ethernet/sfc/falcon/efx.c list_add_tail(&other->node, &ef4_unassociated_list); node 2641 drivers/net/ethernet/sfc/falcon/efx.c INIT_LIST_HEAD(&efx->node); node 51 drivers/net/ethernet/sfc/falcon/mtd.c list_del(&part->node); node 80 drivers/net/ethernet/sfc/falcon/mtd.c list_add_tail(&part->node, &efx->mtd_list); node 105 drivers/net/ethernet/sfc/falcon/mtd.c node); node 107 drivers/net/ethernet/sfc/falcon/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) node 119 drivers/net/ethernet/sfc/falcon/mtd.c list_for_each_entry(part, &efx->mtd_list, node) node 767 drivers/net/ethernet/sfc/falcon/net_driver.h struct list_head node; node 915 drivers/net/ethernet/sfc/falcon/net_driver.h struct list_head node; node 51 drivers/net/ethernet/sfc/mtd.c list_del(&part->node); node 83 drivers/net/ethernet/sfc/mtd.c list_add_tail(&part->node, &efx->mtd_list); node 108 drivers/net/ethernet/sfc/mtd.c node); node 110 drivers/net/ethernet/sfc/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) node 122 drivers/net/ethernet/sfc/mtd.c list_for_each_entry(part, &efx->mtd_list, node) node 756 drivers/net/ethernet/sfc/net_driver.h struct hlist_node node; node 937 drivers/net/ethernet/sfc/net_driver.h struct list_head node; node 1106 drivers/net/ethernet/sfc/net_driver.h struct list_head node; node 873 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c struct device_node *node = priv->device->of_node; node 890 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c if (of_property_read_bool(node, "allwinner,leds-active-low")) node 914 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c if (!of_property_read_u32(node, "allwinner,tx-delay-ps", &val)) { node 932 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c if (!of_property_read_u32(node, "allwinner,rx-delay-ps", &val)) { node 1077 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c static struct regmap *sun8i_dwmac_get_syscon_from_dev(struct device_node *node) node 1083 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c syscon_node = of_parse_phandle(node, "syscon", 0); node 1012 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct device_node *node; node 1015 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c node = priv->plat->phylink_node; node 1017 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (node) node 1018 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c ret = phylink_of_phy_connect(priv->phylink, node, 0); node 1023 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (!node || ret) { node 156 drivers/net/ethernet/ti/cpsw-phy-sel.c const struct device_node *node = (const struct device_node *)data; node 157 drivers/net/ethernet/ti/cpsw-phy-sel.c return dev->of_node == node && node 163 drivers/net/ethernet/ti/cpsw-phy-sel.c struct device_node *node; node 166 drivers/net/ethernet/ti/cpsw-phy-sel.c node = of_parse_phandle(dev->of_node, "cpsw-phy-sel", 0); node 167 drivers/net/ethernet/ti/cpsw-phy-sel.c if (!node) { node 168 drivers/net/ethernet/ti/cpsw-phy-sel.c node = of_get_child_by_name(dev->of_node, "cpsw-phy-sel"); node 169 drivers/net/ethernet/ti/cpsw-phy-sel.c if (!node) { node 175 drivers/net/ethernet/ti/cpsw-phy-sel.c dev = bus_find_device(&platform_bus_type, NULL, node, match); node 177 drivers/net/ethernet/ti/cpsw-phy-sel.c dev_err(dev, "unable to find platform device for %pOF\n", node); node 187 drivers/net/ethernet/ti/cpsw-phy-sel.c of_node_put(node); node 2494 drivers/net/ethernet/ti/cpsw.c struct device_node *node = pdev->dev.of_node; node 2499 drivers/net/ethernet/ti/cpsw.c if (!node) node 2502 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "slaves", &prop)) { node 2508 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "active_slave", &prop)) { node 2521 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "cpdma_channels", &prop)) { node 2527 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "ale_entries", &prop)) { node 2533 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "bd_ram_size", &prop)) { node 2539 drivers/net/ethernet/ti/cpsw.c if (of_property_read_u32(node, "mac_control", &prop)) { node 2545 drivers/net/ethernet/ti/cpsw.c if (of_property_read_bool(node, "dual_emac")) node 2551 drivers/net/ethernet/ti/cpsw.c ret = of_platform_populate(node, NULL, NULL, &pdev->dev); node 2556 drivers/net/ethernet/ti/cpsw.c for_each_available_child_of_node(node, slave_node) { node 2670 drivers/net/ethernet/ti/cpsw.c struct device_node *node = pdev->dev.of_node; node 2674 drivers/net/ethernet/ti/cpsw.c for_each_available_child_of_node(node, slave_node) { node 536 drivers/net/ethernet/ti/cpts.c static int cpts_of_mux_clk_setup(struct cpts *cpts, struct device_node *node) node 545 drivers/net/ethernet/ti/cpts.c refclk_np = of_get_child_by_name(node, "cpts-refclk-mux"); node 612 drivers/net/ethernet/ti/cpts.c static int cpts_of_parse(struct cpts *cpts, struct device_node *node) node 617 drivers/net/ethernet/ti/cpts.c if (!of_property_read_u32(node, "cpts_clock_mult", &prop)) node 620 drivers/net/ethernet/ti/cpts.c if (!of_property_read_u32(node, "cpts_clock_shift", &prop)) node 627 drivers/net/ethernet/ti/cpts.c return cpts_of_mux_clk_setup(cpts, node); node 635 drivers/net/ethernet/ti/cpts.c struct device_node *node) node 648 drivers/net/ethernet/ti/cpts.c ret = cpts_of_parse(cpts, node); node 652 drivers/net/ethernet/ti/cpts.c cpts->refclk = devm_get_clk_from_child(dev, node, "cpts"); node 124 drivers/net/ethernet/ti/cpts.h struct device_node *node); node 149 drivers/net/ethernet/ti/cpts.h struct device_node *node) node 312 drivers/net/ethernet/ti/davinci_mdio.c struct device_node *node = pdev->dev.of_node; node 315 drivers/net/ethernet/ti/davinci_mdio.c if (!node) node 318 drivers/net/ethernet/ti/davinci_mdio.c if (of_property_read_u32(node, "bus_freq", &prop)) { node 63 drivers/net/ethernet/ti/netcp.h struct list_head node; node 194 drivers/net/ethernet/ti/netcp.h struct device *device, struct device_node *node, node 200 drivers/net/ethernet/ti/netcp.h struct device_node *node, void **intf_priv); node 235 drivers/net/ethernet/ti/netcp_core.c struct device_node *devices, *interface, *node = dev->of_node; node 244 drivers/net/ethernet/ti/netcp_core.c devices = of_get_child_by_name(node, "netcp-devices"); node 1388 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry(naddr, &netcp->addr_list, node) { node 1416 drivers/net/ethernet/ti/netcp_core.c list_add_tail(&naddr->node, &netcp->addr_list); node 1423 drivers/net/ethernet/ti/netcp_core.c list_del(&naddr->node); node 1431 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry(naddr, &netcp->addr_list, node) node 1458 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry_safe(naddr, tmp, &netcp->addr_list, node) { node 1482 drivers/net/ethernet/ti/netcp_core.c list_for_each_entry_safe(naddr, tmp, &netcp->addr_list, node) { node 1962 drivers/net/ethernet/ti/netcp_core.c struct device_node *node = dev->of_node; node 2008 drivers/net/ethernet/ti/netcp_core.c if (of_address_to_resource(node, NETCP_EFUSE_REG_INDEX, &res)) { node 2151 drivers/net/ethernet/ti/netcp_core.c struct device_node *node = pdev->dev.of_node; node 2163 drivers/net/ethernet/ti/netcp_core.c if (!node) { node 2188 drivers/net/ethernet/ti/netcp_core.c interfaces = of_get_child_by_name(node, "netcp-interfaces"); node 703 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node; node 2323 drivers/net/ethernet/ti/netcp_ethss.c phy_mode = of_get_phy_mode(slave->node); node 2998 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3004 drivers/net/ethernet/ti/netcp_ethss.c if (of_property_read_u32(node, "slave-port", &slave->slave_num)) { node 3009 drivers/net/ethernet/ti/netcp_ethss.c if (of_property_read_u32(node, "link-interface", node 3016 drivers/net/ethernet/ti/netcp_ethss.c slave->node = node; node 3021 drivers/net/ethernet/ti/netcp_ethss.c slave->phy_node = of_parse_phandle(node, "phy-handle", 0); node 3122 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3131 drivers/net/ethernet/ti/netcp_ethss.c for_each_child_of_node(node, port) { node 3227 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3233 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, XGBE_SS_REG_INDEX, &res); node 3237 drivers/net/ethernet/ti/netcp_ethss.c node, XGBE_SS_REG_INDEX); node 3248 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, XGBE_SM_REG_INDEX, &res); node 3252 drivers/net/ethernet/ti/netcp_ethss.c node, XGBE_SM_REG_INDEX); node 3263 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, XGBE_SERDES_REG_INDEX, &res); node 3267 drivers/net/ethernet/ti/netcp_ethss.c node, XGBE_SERDES_REG_INDEX); node 3335 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3341 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, GBE_SS_REG_INDEX, &res); node 3345 drivers/net/ethernet/ti/netcp_ethss.c node, GBE_SS_REG_INDEX); node 3360 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3366 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, GBE_SGMII34_REG_INDEX, &res); node 3370 drivers/net/ethernet/ti/netcp_ethss.c node, GBE_SGMII34_REG_INDEX); node 3382 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, GBE_SM_REG_INDEX, &res); node 3386 drivers/net/ethernet/ti/netcp_ethss.c node, GBE_SM_REG_INDEX); node 3458 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node) node 3492 drivers/net/ethernet/ti/netcp_ethss.c ret = of_address_to_resource(node, GBENU_SM_REG_INDEX, &res); node 3496 drivers/net/ethernet/ti/netcp_ethss.c node, GBENU_SM_REG_INDEX); node 3555 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node, void **inst_priv) node 3564 drivers/net/ethernet/ti/netcp_ethss.c if (!node) { node 3573 drivers/net/ethernet/ti/netcp_ethss.c if (of_device_is_compatible(node, "ti,netcp-gbe-5") || node 3574 drivers/net/ethernet/ti/netcp_ethss.c of_device_is_compatible(node, "ti,netcp-gbe")) { node 3576 drivers/net/ethernet/ti/netcp_ethss.c } else if (of_device_is_compatible(node, "ti,netcp-gbe-9")) { node 3578 drivers/net/ethernet/ti/netcp_ethss.c } else if (of_device_is_compatible(node, "ti,netcp-gbe-2")) { node 3581 drivers/net/ethernet/ti/netcp_ethss.c } else if (of_device_is_compatible(node, "ti,netcp-xgbe")) { node 3596 drivers/net/ethernet/ti/netcp_ethss.c if (of_find_property(node, "enable-ale", NULL)) { node 3604 drivers/net/ethernet/ti/netcp_ethss.c ret = of_property_read_u32(node, "tx-queue", node 3611 drivers/net/ethernet/ti/netcp_ethss.c ret = of_property_read_string(node, "tx-channel", node 3618 drivers/net/ethernet/ti/netcp_ethss.c if (of_node_name_eq(node, "gbe")) { node 3619 drivers/net/ethernet/ti/netcp_ethss.c ret = get_gbe_resource_version(gbe_dev, node); node 3626 drivers/net/ethernet/ti/netcp_ethss.c ret = set_gbe_ethss14_priv(gbe_dev, node); node 3628 drivers/net/ethernet/ti/netcp_ethss.c ret = set_gbenu_ethss_priv(gbe_dev, node); node 3632 drivers/net/ethernet/ti/netcp_ethss.c } else if (of_node_name_eq(node, "xgbe")) { node 3633 drivers/net/ethernet/ti/netcp_ethss.c ret = set_xgbe_ethss10_priv(gbe_dev, node); node 3639 drivers/net/ethernet/ti/netcp_ethss.c dev_err(dev, "unknown GBE node(%pOFn)\n", node); node 3646 drivers/net/ethernet/ti/netcp_ethss.c interfaces = of_get_child_by_name(node, "interfaces"); node 3684 drivers/net/ethernet/ti/netcp_ethss.c secondary_ports = of_get_child_by_name(node, "secondary-slave-ports"); node 3716 drivers/net/ethernet/ti/netcp_ethss.c cpts_node = of_get_child_by_name(node, "cpts"); node 3718 drivers/net/ethernet/ti/netcp_ethss.c cpts_node = of_node_get(node); node 3751 drivers/net/ethernet/ti/netcp_ethss.c struct device_node *node, void **intf_priv) node 3757 drivers/net/ethernet/ti/netcp_ethss.c if (!node) { node 3778 drivers/net/ethernet/ti/netcp_ethss.c if (init_slave(gbe_dev, gbe_intf->slave, node)) { node 72 drivers/net/fddi/skfp/h/sba.h struct s_sba_node_vars node[MAX_NODES] ; node 128 drivers/net/geneve.c struct geneve_dev_node *node; node 134 drivers/net/geneve.c hlist_for_each_entry_rcu(node, vni_list_head, hlist) { node 135 drivers/net/geneve.c if (eq_tun_id_and_vni((u8 *)&node->geneve->info.key.tun_id, vni) && node 136 drivers/net/geneve.c addr == node->geneve->info.key.u.ipv4.dst) node 137 drivers/net/geneve.c return node->geneve; node 147 drivers/net/geneve.c struct geneve_dev_node *node; node 153 drivers/net/geneve.c hlist_for_each_entry_rcu(node, vni_list_head, hlist) { node 154 drivers/net/geneve.c if (eq_tun_id_and_vni((u8 *)&node->geneve->info.key.tun_id, vni) && node 155 drivers/net/geneve.c ipv6_addr_equal(&addr6, &node->geneve->info.key.u.ipv6.dst)) node 156 drivers/net/geneve.c return node->geneve; node 652 drivers/net/geneve.c struct geneve_dev_node *node; node 673 drivers/net/geneve.c node = &geneve->hlist6; node 678 drivers/net/geneve.c node = &geneve->hlist4; node 680 drivers/net/geneve.c node->geneve = geneve; node 684 drivers/net/geneve.c hlist_add_head_rcu(&node->hlist, &gs->vni_list[hash]); node 284 drivers/net/hyperv/netvsc.c int node = cpu_to_node(nvchan->channel->target_cpu); node 288 drivers/net/hyperv/netvsc.c nvchan->mrc.slots = vzalloc_node(size, node); node 41 drivers/net/phy/fixed_phy.c struct list_head node; node 59 drivers/net/phy/fixed_phy.c list_for_each_entry(fp, &fmb->phys, node) { node 80 drivers/net/phy/fixed_phy.c list_for_each_entry(fp, &fmb->phys, node) { node 125 drivers/net/phy/fixed_phy.c list_for_each_entry(fp, &fmb->phys, node) { node 164 drivers/net/phy/fixed_phy.c list_add_tail(&fp->node, &fmb->phys); node 183 drivers/net/phy/fixed_phy.c list_for_each_entry_safe(fp, tmp, &fmb->phys, node) { node 185 drivers/net/phy/fixed_phy.c list_del(&fp->node); node 386 drivers/net/phy/fixed_phy.c list_for_each_entry_safe(fp, tmp, &fmb->phys, node) { node 387 drivers/net/phy/fixed_phy.c list_del(&fp->node); node 25 drivers/net/phy/mdio-thunder.c struct device_node *node; node 66 drivers/net/phy/mdio-thunder.c node = to_of_node(fwn); node 67 drivers/net/phy/mdio-thunder.c if (!node) node 70 drivers/net/phy/mdio-thunder.c err = of_address_to_resource(node, 0, &r); node 74 drivers/net/phy/mdio-thunder.c node); node 99 drivers/net/phy/mdio-thunder.c err = of_mdiobus_register(bus->mii_bus, node); node 246 drivers/net/phy/phy-core.c struct device_node *node = phydev->mdio.dev.of_node; node 252 drivers/net/phy/phy-core.c if (!node) node 255 drivers/net/phy/phy-core.c if (!of_property_read_u32(node, "max-speed", &max_speed)) node 261 drivers/net/phy/phy-core.c struct device_node *node = phydev->mdio.dev.of_node; node 267 drivers/net/phy/phy-core.c if (!node) node 270 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-100tx")) node 272 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-1000t")) node 274 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-10gt")) node 276 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-1000kx")) node 278 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-10gkx4")) node 280 drivers/net/phy/phy-core.c if (of_property_read_bool(node, "eee-broken-10gkr")) node 18 drivers/net/phy/sfp-bus.c struct list_head node; node 299 drivers/net/phy/sfp-bus.c list_for_each_entry(sfp, &sfp_buses, node) { node 310 drivers/net/phy/sfp-bus.c list_add(&new->node, &sfp_buses); node 326 drivers/net/phy/sfp-bus.c list_del(&bus->node); node 1888 drivers/net/phy/sfp.c struct device_node *node = pdev->dev.of_node; node 1892 drivers/net/phy/sfp.c id = of_match_node(sfp_of_match, node); node 1898 drivers/net/phy/sfp.c np = of_parse_phandle(node, "i2c-bus", 0); node 66 drivers/net/rionet.c struct list_head node; node 240 drivers/net/rionet.c list_for_each_entry(peer, &nets[netid].peers, node) { node 354 drivers/net/rionet.c list_for_each_entry(peer, &nets[netid].peers, node) { node 384 drivers/net/rionet.c list_for_each_entry(peer, &nets[netid].peers, node) { node 414 drivers/net/rionet.c list_for_each_entry(peer, &nets[netid].peers, node) { node 416 drivers/net/rionet.c list_del(&peer->node); node 620 drivers/net/rionet.c list_add_tail(&peer->node, &nets[netid].peers); node 649 drivers/net/rionet.c list_for_each_entry(peer, &nets[i].peers, node) { node 838 drivers/net/team/team.c struct list_head *node; node 843 drivers/net/team/team.c node = qom_list; node 847 drivers/net/team/team.c node = &cur->qom_list; node 849 drivers/net/team/team.c list_add_tail_rcu(&port->qom_list, node); node 1799 drivers/net/usb/lan78xx.c struct device_node *node; node 1829 drivers/net/usb/lan78xx.c node = of_get_child_by_name(dev->udev->dev.of_node, "mdio"); node 1830 drivers/net/usb/lan78xx.c ret = of_mdiobus_register(dev->mdiobus, node); node 1831 drivers/net/usb/lan78xx.c of_node_put(node); node 1539 drivers/net/usb/r8152.c int node = netdev->dev.parent ? dev_to_node(netdev->dev.parent) : -1; node 1544 drivers/net/usb/r8152.c rx_agg = kmalloc_node(sizeof(*rx_agg), mflags, node); node 1614 drivers/net/usb/r8152.c int node, i; node 1616 drivers/net/usb/r8152.c node = netdev->dev.parent ? dev_to_node(netdev->dev.parent) : -1; node 1636 drivers/net/usb/r8152.c buf = kmalloc_node(agg_buf_sz, GFP_KERNEL, node); node 1643 drivers/net/usb/r8152.c node); node 218 drivers/net/virtio_net.c struct hlist_node node; node 1971 drivers/net/virtio_net.c static int virtnet_cpu_online(unsigned int cpu, struct hlist_node *node) node 1973 drivers/net/virtio_net.c struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info, node 1974 drivers/net/virtio_net.c node); node 1979 drivers/net/virtio_net.c static int virtnet_cpu_dead(unsigned int cpu, struct hlist_node *node) node 1981 drivers/net/virtio_net.c struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info, node 1987 drivers/net/virtio_net.c static int virtnet_cpu_down_prep(unsigned int cpu, struct hlist_node *node) node 1989 drivers/net/virtio_net.c struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info, node 1990 drivers/net/virtio_net.c node); node 2002 drivers/net/virtio_net.c ret = cpuhp_state_add_instance_nocalls(virtionet_online, &vi->node); node 2009 drivers/net/virtio_net.c cpuhp_state_remove_instance_nocalls(virtionet_online, &vi->node); node 2015 drivers/net/virtio_net.c cpuhp_state_remove_instance_nocalls(virtionet_online, &vi->node); node 208 drivers/net/vxlan.c struct vxlan_dev_node *node; node 214 drivers/net/vxlan.c hlist_for_each_entry_rcu(node, vni_head(vs, vni), hlist) { node 215 drivers/net/vxlan.c if (node->vxlan->default_dst.remote_vni != vni) node 219 drivers/net/vxlan.c const struct vxlan_config *cfg = &node->vxlan->cfg; node 226 drivers/net/vxlan.c return node->vxlan; node 2772 drivers/net/vxlan.c struct vxlan_dev_node *node) node 2777 drivers/net/vxlan.c node->vxlan = vxlan; node 2779 drivers/net/vxlan.c hlist_add_head_rcu(&node->hlist, vni_head(vs, vni)); node 3277 drivers/net/vxlan.c struct vxlan_dev_node *node; node 3304 drivers/net/vxlan.c node = &vxlan->hlist6; node 3309 drivers/net/vxlan.c node = &vxlan->hlist4; node 3311 drivers/net/vxlan.c vxlan_vs_add_dev(vs, vxlan, node); node 46 drivers/net/wan/hd64570.c #define SCA_INTR_MSCI(node) (node ? 0x10 : 0x01) node 47 drivers/net/wan/hd64570.c #define SCA_INTR_DMAC_RX(node) (node ? 0x20 : 0x02) node 48 drivers/net/wan/hd64570.c #define SCA_INTR_DMAC_TX(node) (node ? 0x40 : 0x04) node 131 drivers/net/wan/hd64570.h #define DSR_RX(node) (DSR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET)) node 132 drivers/net/wan/hd64570.h #define DSR_TX(node) (DSR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET)) node 134 drivers/net/wan/hd64570.h #define DMR_RX(node) (DMR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET)) node 135 drivers/net/wan/hd64570.h #define DMR_TX(node) (DMR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET)) node 137 drivers/net/wan/hd64570.h #define FCT_RX(node) (FCT + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET)) node 138 drivers/net/wan/hd64570.h #define FCT_TX(node) (FCT + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET)) node 140 drivers/net/wan/hd64570.h #define DIR_RX(node) (DIR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET)) node 141 drivers/net/wan/hd64570.h #define DIR_TX(node) (DIR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET)) node 143 drivers/net/wan/hd64570.h #define DCR_RX(node) (DCR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET)) node 144 drivers/net/wan/hd64570.h #define DCR_TX(node) (DCR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET)) node 51 drivers/net/wan/lapbether.c struct list_head node; node 67 drivers/net/wan/lapbether.c list_for_each_entry_rcu(lapbeth, &lapbeth_devices, node) { node 337 drivers/net/wan/lapbether.c list_add_rcu(&lapbeth->node, &lapbeth_devices); node 353 drivers/net/wan/lapbether.c list_del_rcu(&lapbeth->node); node 433 drivers/net/wan/lapbether.c lapbeth = list_entry(entry, struct lapbethdev, node); node 58 drivers/net/wan/wanxl.c int node; /* physical port #0 - 3 */ node 94 drivers/net/wan/wanxl.c return &port->card->status->port_status[port->node]; node 298 drivers/net/wan/wanxl.c writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node), node 409 drivers/net/wan/wanxl.c writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr); node 421 drivers/net/wan/wanxl.c writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr); node 435 drivers/net/wan/wanxl.c writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), node 782 drivers/net/wan/wanxl.c port->node = i; node 987 drivers/net/wireless/ath/ath10k/core.c struct device_node *node; node 990 drivers/net/wireless/ath/ath10k/core.c node = ar->dev->of_node; node 991 drivers/net/wireless/ath/ath10k/core.c if (!node) node 994 drivers/net/wireless/ath/ath10k/core.c of_property_read_string(node, "qcom,ath10k-calibration-variant", node 1681 drivers/net/wireless/ath/ath10k/core.c struct device_node *node; node 1686 drivers/net/wireless/ath/ath10k/core.c node = ar->dev->of_node; node 1687 drivers/net/wireless/ath/ath10k/core.c if (!node) node 1693 drivers/net/wireless/ath/ath10k/core.c if (!of_get_property(node, dt_name, &data_len)) { node 1711 drivers/net/wireless/ath/ath10k/core.c ret = of_property_read_u8_array(node, dt_name, data, data_len); node 4730 drivers/net/wireless/ath/ath10k/mac.c struct device_node *node; node 4734 drivers/net/wireless/ath/ath10k/mac.c node = ar->dev->of_node; node 4735 drivers/net/wireless/ath/ath10k/mac.c if (!node) node 4738 drivers/net/wireless/ath/ath10k/mac.c ret = of_property_read_string_index(node, "ext-fem-name", 0, &fem_name); node 885 drivers/net/wireless/ath/ath10k/qmi.c sq->sq_node = service->node; node 959 drivers/net/wireless/ath/ath10k/qmi.c struct device_node *node; node 963 drivers/net/wireless/ath/ath10k/qmi.c node = of_parse_phandle(dev->of_node, "memory-region", 0); node 964 drivers/net/wireless/ath/ath10k/qmi.c if (node) { node 965 drivers/net/wireless/ath/ath10k/qmi.c ret = of_address_to_resource(node, 0, &r); node 970 drivers/net/wireless/ath/ath10k/qmi.c of_node_put(node); node 705 drivers/net/wireless/ath/ath6kl/init.c struct device_node *node; node 710 drivers/net/wireless/ath/ath6kl/init.c for_each_compatible_node(node, NULL, "atheros,ath6kl") { node 711 drivers/net/wireless/ath/ath6kl/init.c board_id = of_get_property(node, board_id_prop, NULL); node 714 drivers/net/wireless/ath/ath6kl/init.c board_id_prop, node); node 727 drivers/net/wireless/ath/ath6kl/init.c of_node_put(node); node 642 drivers/net/wireless/ath/ath6kl/sdio.c struct hif_scatter_req *node = NULL; node 647 drivers/net/wireless/ath/ath6kl/sdio.c node = list_first_entry(&ar_sdio->scat_req, node 649 drivers/net/wireless/ath/ath6kl/sdio.c list_del(&node->list); node 651 drivers/net/wireless/ath/ath6kl/sdio.c node->scat_q_depth = get_queue_depth(&ar_sdio->scat_req); node 656 drivers/net/wireless/ath/ath6kl/sdio.c return node; node 1051 drivers/net/wireless/ath/ath6kl/txrx.c struct skb_hold_q *node; node 1078 drivers/net/wireless/ath/ath6kl/txrx.c node = &rxtid->hold_q[idx]; node 1079 drivers/net/wireless/ath/ath6kl/txrx.c if ((order == 1) && (!node->skb)) node 1082 drivers/net/wireless/ath/ath6kl/txrx.c if (node->skb) { node 1083 drivers/net/wireless/ath/ath6kl/txrx.c if (node->is_amsdu) node 1085 drivers/net/wireless/ath/ath6kl/txrx.c node->skb); node 1087 drivers/net/wireless/ath/ath6kl/txrx.c skb_queue_tail(&rxtid->q, node->skb); node 1088 drivers/net/wireless/ath/ath6kl/txrx.c node->skb = NULL; node 1112 drivers/net/wireless/ath/ath6kl/txrx.c struct skb_hold_q *node; node 1175 drivers/net/wireless/ath/ath6kl/txrx.c node = &rxtid->hold_q[idx]; node 1191 drivers/net/wireless/ath/ath6kl/txrx.c dev_kfree_skb(node->skb); node 1194 drivers/net/wireless/ath/ath6kl/txrx.c node->skb = frame; node 1196 drivers/net/wireless/ath/ath6kl/txrx.c node->is_amsdu = is_amsdu; node 1197 drivers/net/wireless/ath/ath6kl/txrx.c node->seq_no = seq_no; node 1199 drivers/net/wireless/ath/ath6kl/txrx.c if (node->is_amsdu) node 2126 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c .node = {}, node 72 drivers/net/wireless/marvell/mwifiex/11n.h struct mwifiex_sta_node *node = mwifiex_get_sta_entry(priv, ptr->ra); node 74 drivers/net/wireless/marvell/mwifiex/11n.h if (unlikely(!node)) node 77 drivers/net/wireless/marvell/mwifiex/11n.h return (node->ampdu_sta[tid] != BA_STREAM_NOT_ALLOWED) ? true : false; node 171 drivers/net/wireless/marvell/mwifiex/11n.h struct mwifiex_sta_node *node) node 173 drivers/net/wireless/marvell/mwifiex/11n.h if (!node || ((priv->bss_role == MWIFIEX_BSS_ROLE_UAP) && node 179 drivers/net/wireless/marvell/mwifiex/11n.h return node->is_11n_enabled; node 185 drivers/net/wireless/marvell/mwifiex/11n.h struct mwifiex_sta_node *node = mwifiex_get_sta_entry(priv, ra); node 186 drivers/net/wireless/marvell/mwifiex/11n.h if (node) node 187 drivers/net/wireless/marvell/mwifiex/11n.h return node->is_11n_enabled; node 347 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c struct mwifiex_sta_node *node; node 373 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c node = mwifiex_get_sta_entry(priv, ta); node 374 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c if (node) node 375 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c last_seq = node->rx_seq[tid]; node 378 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c node = mwifiex_get_sta_entry(priv, ta); node 379 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c if (node) node 380 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c last_seq = node->rx_seq[tid]; node 1382 drivers/net/wireless/marvell/mwifiex/cfg80211.c struct mwifiex_sta_node *node, node 1393 drivers/net/wireless/marvell/mwifiex/cfg80211.c if (!node) node 1399 drivers/net/wireless/marvell/mwifiex/cfg80211.c jiffies_to_msecs(jiffies - node->stats.last_rx); node 1401 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->signal = node->stats.rssi; node 1402 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->signal_avg = node->stats.rssi; node 1403 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->rx_bytes = node->stats.rx_bytes; node 1404 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->tx_bytes = node->stats.tx_bytes; node 1405 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->rx_packets = node->stats.rx_packets; node 1406 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->tx_packets = node->stats.tx_packets; node 1407 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->tx_failed = node->stats.tx_failed; node 1410 drivers/net/wireless/marvell/mwifiex/cfg80211.c node->stats.last_tx_htinfo, node 1412 drivers/net/wireless/marvell/mwifiex/cfg80211.c sinfo->txrate.legacy = node->stats.last_tx_rate * 5; node 1499 drivers/net/wireless/marvell/mwifiex/cfg80211.c static struct mwifiex_sta_node *node; node 1509 drivers/net/wireless/marvell/mwifiex/cfg80211.c if (node && (&node->list == &priv->sta_list)) { node 1510 drivers/net/wireless/marvell/mwifiex/cfg80211.c node = NULL; node 1514 drivers/net/wireless/marvell/mwifiex/cfg80211.c node = list_prepare_entry(node, &priv->sta_list, list); node 1515 drivers/net/wireless/marvell/mwifiex/cfg80211.c list_for_each_entry_continue(node, &priv->sta_list, list) { node 1516 drivers/net/wireless/marvell/mwifiex/cfg80211.c ether_addr_copy(mac, node->mac_addr); node 1517 drivers/net/wireless/marvell/mwifiex/cfg80211.c return mwifiex_dump_station_info(priv, node, sinfo); node 1585 drivers/net/wireless/marvell/mwifiex/main.h struct mwifiex_sta_node *node); node 1598 drivers/net/wireless/marvell/mwifiex/main.h struct device_node *node, const char *prefix); node 1607 drivers/net/wireless/marvell/mwifiex/main.h int ies_len, struct mwifiex_sta_node *node); node 1488 drivers/net/wireless/marvell/mwifiex/sta_cmd.c struct device_node *node, const char *prefix) node 1496 drivers/net/wireless/marvell/mwifiex/sta_cmd.c for_each_property_of_node(node, prop) { node 878 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c struct mwifiex_sta_node *node = node 884 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c if (!node || reason == TDLS_ERR_LINK_NONEXISTENT) node 903 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c if (node && reason != TDLS_ERR_LINK_EXISTS) node 904 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c node->tdls_status = TDLS_SETUP_FAILURE; node 916 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c if (node) node 917 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c node->tdls_status = TDLS_SETUP_FAILURE; node 113 drivers/net/wireless/marvell/mwifiex/uap_event.c struct mwifiex_sta_node *node; node 148 drivers/net/wireless/marvell/mwifiex/uap_event.c node = mwifiex_add_sta_entry(priv, event->sta_addr); node 149 drivers/net/wireless/marvell/mwifiex/uap_event.c if (!node) { node 162 drivers/net/wireless/marvell/mwifiex/uap_event.c sinfo->assoc_req_ies_len, node); node 165 drivers/net/wireless/marvell/mwifiex/uap_event.c if (node->is_11n_enabled) node 166 drivers/net/wireless/marvell/mwifiex/uap_event.c node->ampdu_sta[i] = node 169 drivers/net/wireless/marvell/mwifiex/uap_event.c node->ampdu_sta[i] = BA_STREAM_NOT_ALLOWED; node 171 drivers/net/wireless/marvell/mwifiex/uap_event.c memset(node->rx_seq, 0xff, sizeof(node->rx_seq)); node 347 drivers/net/wireless/marvell/mwifiex/uap_event.c struct mwifiex_sta_node *node) node 349 drivers/net/wireless/marvell/mwifiex/uap_event.c if (priv->ap_11n_enabled && node->is_11n_enabled) { node 350 drivers/net/wireless/marvell/mwifiex/uap_event.c mwifiex_11n_del_rx_reorder_tbl_by_ta(priv, node->mac_addr); node 351 drivers/net/wireless/marvell/mwifiex/uap_event.c mwifiex_del_tx_ba_stream_tbl_by_ra(priv, node->mac_addr); node 353 drivers/net/wireless/marvell/mwifiex/uap_event.c mwifiex_del_sta_entry(priv, node->mac_addr); node 380 drivers/net/wireless/marvell/mwifiex/uap_txrx.c struct mwifiex_sta_node *node; node 396 drivers/net/wireless/marvell/mwifiex/uap_txrx.c node = mwifiex_get_sta_entry(priv, ta); node 397 drivers/net/wireless/marvell/mwifiex/uap_txrx.c if (node) node 398 drivers/net/wireless/marvell/mwifiex/uap_txrx.c node->stats.tx_failed++; node 415 drivers/net/wireless/marvell/mwifiex/uap_txrx.c node = mwifiex_get_sta_entry(priv, ta); node 416 drivers/net/wireless/marvell/mwifiex/uap_txrx.c if (node) node 417 drivers/net/wireless/marvell/mwifiex/uap_txrx.c node->rx_seq[uap_rx_pd->priority] = node 529 drivers/net/wireless/marvell/mwifiex/util.c struct mwifiex_sta_node *node; node 534 drivers/net/wireless/marvell/mwifiex/util.c list_for_each_entry(node, &priv->sta_list, list) { node 535 drivers/net/wireless/marvell/mwifiex/util.c if (!memcmp(node->mac_addr, mac, ETH_ALEN)) node 536 drivers/net/wireless/marvell/mwifiex/util.c return node; node 545 drivers/net/wireless/marvell/mwifiex/util.c struct mwifiex_sta_node *node; node 547 drivers/net/wireless/marvell/mwifiex/util.c list_for_each_entry(node, &priv->sta_list, list) { node 548 drivers/net/wireless/marvell/mwifiex/util.c if (node->tdls_status == status) node 549 drivers/net/wireless/marvell/mwifiex/util.c return node; node 609 drivers/net/wireless/marvell/mwifiex/util.c struct mwifiex_sta_node *node; node 615 drivers/net/wireless/marvell/mwifiex/util.c node = mwifiex_get_sta_entry(priv, mac); node 616 drivers/net/wireless/marvell/mwifiex/util.c if (node) node 619 drivers/net/wireless/marvell/mwifiex/util.c node = kzalloc(sizeof(*node), GFP_ATOMIC); node 620 drivers/net/wireless/marvell/mwifiex/util.c if (!node) node 623 drivers/net/wireless/marvell/mwifiex/util.c memcpy(node->mac_addr, mac, ETH_ALEN); node 624 drivers/net/wireless/marvell/mwifiex/util.c list_add_tail(&node->list, &priv->sta_list); node 628 drivers/net/wireless/marvell/mwifiex/util.c return node; node 636 drivers/net/wireless/marvell/mwifiex/util.c int ies_len, struct mwifiex_sta_node *node) node 648 drivers/net/wireless/marvell/mwifiex/util.c node->is_11n_enabled = 1; node 649 drivers/net/wireless/marvell/mwifiex/util.c node->max_amsdu = le16_to_cpu(ht_cap->cap_info) & node 654 drivers/net/wireless/marvell/mwifiex/util.c node->is_11n_enabled = 0; node 663 drivers/net/wireless/marvell/mwifiex/util.c struct mwifiex_sta_node *node; node 667 drivers/net/wireless/marvell/mwifiex/util.c node = mwifiex_get_sta_entry(priv, mac); node 668 drivers/net/wireless/marvell/mwifiex/util.c if (node) { node 669 drivers/net/wireless/marvell/mwifiex/util.c list_del(&node->list); node 670 drivers/net/wireless/marvell/mwifiex/util.c kfree(node); node 680 drivers/net/wireless/marvell/mwifiex/util.c struct mwifiex_sta_node *node, *tmp; node 684 drivers/net/wireless/marvell/mwifiex/util.c list_for_each_entry_safe(node, tmp, &priv->sta_list, list) { node 685 drivers/net/wireless/marvell/mwifiex/util.c list_del(&node->list); node 686 drivers/net/wireless/marvell/mwifiex/util.c kfree(node); node 140 drivers/net/wireless/marvell/mwifiex/wmm.c struct mwifiex_sta_node *node; node 166 drivers/net/wireless/marvell/mwifiex/wmm.c node = mwifiex_get_sta_entry(priv, ra); node 167 drivers/net/wireless/marvell/mwifiex/wmm.c if (node) node 168 drivers/net/wireless/marvell/mwifiex/wmm.c ra_list->tx_paused = node->tx_pause; node 170 drivers/net/wireless/marvell/mwifiex/wmm.c mwifiex_is_sta_11n_enabled(priv, node); node 172 drivers/net/wireless/marvell/mwifiex/wmm.c ra_list->max_amsdu = node->max_amsdu; node 19 drivers/net/wireless/quantenna/qtnfmac/util.c struct qtnf_sta_node *node; node 24 drivers/net/wireless/quantenna/qtnfmac/util.c list_for_each_entry(node, &list->head, list) { node 25 drivers/net/wireless/quantenna/qtnfmac/util.c if (ether_addr_equal(node->mac_addr, mac)) node 26 drivers/net/wireless/quantenna/qtnfmac/util.c return node; node 35 drivers/net/wireless/quantenna/qtnfmac/util.c struct qtnf_sta_node *node; node 40 drivers/net/wireless/quantenna/qtnfmac/util.c list_for_each_entry(node, &list->head, list) { node 42 drivers/net/wireless/quantenna/qtnfmac/util.c return node; node 52 drivers/net/wireless/quantenna/qtnfmac/util.c struct qtnf_sta_node *node; node 57 drivers/net/wireless/quantenna/qtnfmac/util.c node = qtnf_sta_list_lookup(list, mac); node 59 drivers/net/wireless/quantenna/qtnfmac/util.c if (node) node 62 drivers/net/wireless/quantenna/qtnfmac/util.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 63 drivers/net/wireless/quantenna/qtnfmac/util.c if (unlikely(!node)) node 66 drivers/net/wireless/quantenna/qtnfmac/util.c ether_addr_copy(node->mac_addr, mac); node 67 drivers/net/wireless/quantenna/qtnfmac/util.c list_add_tail(&node->list, &list->head); node 72 drivers/net/wireless/quantenna/qtnfmac/util.c return node; node 78 drivers/net/wireless/quantenna/qtnfmac/util.c struct qtnf_sta_node *node; node 81 drivers/net/wireless/quantenna/qtnfmac/util.c node = qtnf_sta_list_lookup(list, mac); node 83 drivers/net/wireless/quantenna/qtnfmac/util.c if (node) { node 84 drivers/net/wireless/quantenna/qtnfmac/util.c list_del(&node->list); node 86 drivers/net/wireless/quantenna/qtnfmac/util.c kfree(node); node 96 drivers/net/wireless/quantenna/qtnfmac/util.c struct qtnf_sta_node *node, *tmp; node 100 drivers/net/wireless/quantenna/qtnfmac/util.c list_for_each_entry_safe(node, tmp, &list->head, list) { node 101 drivers/net/wireless/quantenna/qtnfmac/util.c list_del(&node->list); node 102 drivers/net/wireless/quantenna/qtnfmac/util.c kfree(node); node 705 drivers/net/xen-netback/xenbus.c char *node; node 708 drivers/net/xen-netback/xenbus.c if (vif->credit_watch.node) node 711 drivers/net/xen-netback/xenbus.c node = kmalloc(maxlen, GFP_KERNEL); node 712 drivers/net/xen-netback/xenbus.c if (!node) node 714 drivers/net/xen-netback/xenbus.c snprintf(node, maxlen, "%s/rate", dev->nodename); node 715 drivers/net/xen-netback/xenbus.c vif->credit_watch.node = node; node 719 drivers/net/xen-netback/xenbus.c pr_err("Failed to set watcher %s\n", vif->credit_watch.node); node 720 drivers/net/xen-netback/xenbus.c kfree(node); node 721 drivers/net/xen-netback/xenbus.c vif->credit_watch.node = NULL; node 729 drivers/net/xen-netback/xenbus.c if (vif->credit_watch.node) { node 731 drivers/net/xen-netback/xenbus.c kfree(vif->credit_watch.node); node 732 drivers/net/xen-netback/xenbus.c vif->credit_watch.node = NULL; node 751 drivers/net/xen-netback/xenbus.c char *node; node 755 drivers/net/xen-netback/xenbus.c if (vif->mcast_ctrl_watch.node) { node 760 drivers/net/xen-netback/xenbus.c node = kmalloc(maxlen, GFP_KERNEL); node 761 drivers/net/xen-netback/xenbus.c if (!node) { node 765 drivers/net/xen-netback/xenbus.c snprintf(node, maxlen, "%s/request-multicast-control", node 767 drivers/net/xen-netback/xenbus.c vif->mcast_ctrl_watch.node = node; node 772 drivers/net/xen-netback/xenbus.c vif->mcast_ctrl_watch.node); node 773 drivers/net/xen-netback/xenbus.c kfree(node); node 774 drivers/net/xen-netback/xenbus.c vif->mcast_ctrl_watch.node = NULL; node 782 drivers/net/xen-netback/xenbus.c if (vif->mcast_ctrl_watch.node) { node 784 drivers/net/xen-netback/xenbus.c kfree(vif->mcast_ctrl_watch.node); node 785 drivers/net/xen-netback/xenbus.c vif->mcast_ctrl_watch.node = NULL; node 806 drivers/net/xen-netback/xenbus.c kfree(be->hotplug_status_watch.node); node 172 drivers/nfc/nfcmrvl/i2c.c static int nfcmrvl_i2c_parse_dt(struct device_node *node, node 177 drivers/nfc/nfcmrvl/i2c.c ret = nfcmrvl_parse_dt(node, pdata); node 183 drivers/nfc/nfcmrvl/i2c.c if (of_find_property(node, "i2c-int-falling", NULL)) node 188 drivers/nfc/nfcmrvl/i2c.c ret = irq_of_parse_and_map(node, 0); node 262 drivers/nfc/nfcmrvl/main.c int nfcmrvl_parse_dt(struct device_node *node, node 267 drivers/nfc/nfcmrvl/main.c reset_n_io = of_get_named_gpio(node, "reset-n-io", 0); node 276 drivers/nfc/nfcmrvl/main.c if (of_find_property(node, "hci-muxed", NULL)) node 114 drivers/nfc/nfcmrvl/nfcmrvl.h int nfcmrvl_parse_dt(struct device_node *node, node 120 drivers/nfc/nfcmrvl/spi.c static int nfcmrvl_spi_parse_dt(struct device_node *node, node 125 drivers/nfc/nfcmrvl/spi.c ret = nfcmrvl_parse_dt(node, pdata); node 131 drivers/nfc/nfcmrvl/spi.c ret = irq_of_parse_and_map(node, 0); node 70 drivers/nfc/nfcmrvl/uart.c static int nfcmrvl_uart_parse_dt(struct device_node *node, node 76 drivers/nfc/nfcmrvl/uart.c matched_node = of_get_compatible_child(node, "marvell,nfc-uart"); node 78 drivers/nfc/nfcmrvl/uart.c matched_node = of_get_compatible_child(node, "mrvl,nfc-uart"); node 586 drivers/ntb/hw/amd/ntb_hw_amd.c int rc, i, msix_count, node; node 590 drivers/ntb/hw/amd/ntb_hw_amd.c node = dev_to_node(&pdev->dev); node 596 drivers/ntb/hw/amd/ntb_hw_amd.c GFP_KERNEL, node); node 601 drivers/ntb/hw/amd/ntb_hw_amd.c GFP_KERNEL, node); node 1062 drivers/ntb/hw/amd/ntb_hw_amd.c int rc, node; node 1064 drivers/ntb/hw/amd/ntb_hw_amd.c node = dev_to_node(&pdev->dev); node 1066 drivers/ntb/hw/amd/ntb_hw_amd.c ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node); node 367 drivers/ntb/hw/intel/ntb_hw_gen1.c int rc, i, msix_count, node; node 371 drivers/ntb/hw/intel/ntb_hw_gen1.c node = dev_to_node(&pdev->dev); node 382 drivers/ntb/hw/intel/ntb_hw_gen1.c GFP_KERNEL, node); node 387 drivers/ntb/hw/intel/ntb_hw_gen1.c GFP_KERNEL, node); node 1858 drivers/ntb/hw/intel/ntb_hw_gen1.c int rc, node; node 1860 drivers/ntb/hw/intel/ntb_hw_gen1.c node = dev_to_node(&pdev->dev); node 1863 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node); node 1880 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node); node 387 drivers/ntb/ntb_transport.c int node; node 396 drivers/ntb/ntb_transport.c node = dev_to_node(&nt->ndev->dev); node 399 drivers/ntb/ntb_transport.c GFP_KERNEL, node); node 625 drivers/ntb/ntb_transport.c int node; node 657 drivers/ntb/ntb_transport.c node = dev_to_node(&ndev->dev); node 659 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); node 1249 drivers/ntb/ntb_transport.c int node; node 1269 drivers/ntb/ntb_transport.c node = dev_to_node(&ndev->dev); node 1271 drivers/ntb/ntb_transport.c nt = kzalloc_node(sizeof(*nt), GFP_KERNEL, node); node 1305 drivers/ntb/ntb_transport.c GFP_KERNEL, node); node 1352 drivers/ntb/ntb_transport.c GFP_KERNEL, node); node 1957 drivers/ntb/ntb_transport.c static bool ntb_dma_filter_fn(struct dma_chan *chan, void *node) node 1959 drivers/ntb/ntb_transport.c return dev_to_node(&chan->dev->device) == (int)(unsigned long)node; node 1988 drivers/ntb/ntb_transport.c int node; node 1995 drivers/ntb/ntb_transport.c node = dev_to_node(&ndev->dev); node 2020 drivers/ntb/ntb_transport.c (void *)(unsigned long)node); node 2026 drivers/ntb/ntb_transport.c (void *)(unsigned long)node); node 2054 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); node 2065 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); node 857 drivers/ntb/test/ntb_perf.c int node; node 859 drivers/ntb/test/ntb_perf.c node = dev_to_node(&perf->ntb->dev); node 861 drivers/ntb/test/ntb_perf.c return node == NUMA_NO_NODE || node == dev_to_node(chan->device->dev); node 3486 drivers/nvme/host/core.c int node = ctrl->numa_node, flags = GENHD_FL_EXT_DEVT, ret; node 3488 drivers/nvme/host/core.c ns = kzalloc_node(sizeof(*ns), GFP_KERNEL, node); node 3530 drivers/nvme/host/core.c disk = alloc_disk_node(0, node); node 3546 drivers/nvme/host/core.c ret = nvme_nvm_register(ns, disk_name, node); node 948 drivers/nvme/host/lightnvm.c int nvme_nvm_register(struct nvme_ns *ns, char *disk_name, int node) node 956 drivers/nvme/host/lightnvm.c dev = nvm_alloc_dev(node); node 142 drivers/nvme/host/multipath.c int node; node 147 drivers/nvme/host/multipath.c for_each_node(node) { node 148 drivers/nvme/host/multipath.c if (ns == rcu_access_pointer(head->current_path[node])) { node 149 drivers/nvme/host/multipath.c rcu_assign_pointer(head->current_path[node], NULL); node 177 drivers/nvme/host/multipath.c static struct nvme_ns *__nvme_find_path(struct nvme_ns_head *head, int node) node 187 drivers/nvme/host/multipath.c distance = node_distance(node, ns->ctrl->numa_node); node 212 drivers/nvme/host/multipath.c rcu_assign_pointer(head->current_path[node], found); node 227 drivers/nvme/host/multipath.c int node, struct nvme_ns *old) node 255 drivers/nvme/host/multipath.c rcu_assign_pointer(head->current_path[node], found); node 267 drivers/nvme/host/multipath.c int node = numa_node_id(); node 270 drivers/nvme/host/multipath.c ns = srcu_dereference(head->current_path[node], &head->srcu); node 272 drivers/nvme/host/multipath.c ns = nvme_round_robin_path(head, node, ns); node 274 drivers/nvme/host/multipath.c ns = __nvme_find_path(head, node); node 426 drivers/nvme/host/multipath.c int node, srcu_idx; node 429 drivers/nvme/host/multipath.c for_each_node(node) node 430 drivers/nvme/host/multipath.c __nvme_find_path(head, node); node 631 drivers/nvme/host/nvme.h int nvme_nvm_register(struct nvme_ns *ns, char *disk_name, int node); node 637 drivers/nvme/host/nvme.h int node) node 2784 drivers/nvme/host/pci.c int node, result = -ENOMEM; node 2789 drivers/nvme/host/pci.c node = dev_to_node(&pdev->dev); node 2790 drivers/nvme/host/pci.c if (node == NUMA_NO_NODE) node 2793 drivers/nvme/host/pci.c dev = kzalloc_node(sizeof(*dev), GFP_KERNEL, node); node 2798 drivers/nvme/host/pci.c GFP_KERNEL, node); node 2830 drivers/nvme/host/pci.c GFP_KERNEL, node); node 443 drivers/nvme/target/tcp.c struct llist_node *node; node 445 drivers/nvme/target/tcp.c node = llist_del_all(&queue->resp_list); node 446 drivers/nvme/target/tcp.c if (!node) node 449 drivers/nvme/target/tcp.c while (node) { node 450 drivers/nvme/target/tcp.c struct nvmet_tcp_cmd *cmd = llist_entry(node, node 454 drivers/nvme/target/tcp.c node = node->next; node 30 drivers/nvmem/core.c struct list_head node; node 110 drivers/nvmem/core.c list_del(&cell->node); node 121 drivers/nvmem/core.c list_for_each_entry_safe(cell, p, &nvmem->cells, node) node 128 drivers/nvmem/core.c list_add_tail(&cell->node, &cell->nvmem->cells); node 244 drivers/nvmem/core.c list_for_each_entry(table, &nvmem_cell_tables, node) { node 279 drivers/nvmem/core.c list_for_each_entry(iter, &nvmem->cells, node) { node 711 drivers/nvmem/core.c list_for_each_entry(lookup, &nvmem_lookup_list, node) { node 743 drivers/nvmem/core.c list_for_each_entry(iter, &nvmem->cells, node) { node 1281 drivers/nvmem/core.c list_add_tail(&table->node, &nvmem_cell_tables); node 1294 drivers/nvmem/core.c list_del(&table->node); node 1311 drivers/nvmem/core.c list_add_tail(&entries[i].node, &nvmem_lookup_list); node 1329 drivers/nvmem/core.c list_del(&entries[i].node); node 91 drivers/nvmem/snvs_lpgpr.c struct device_node *node = dev->of_node; node 98 drivers/nvmem/snvs_lpgpr.c if (!node) node 109 drivers/nvmem/snvs_lpgpr.c syscon_node = of_get_parent(node); node 236 drivers/of/address.c struct device_node *node, const char *name) node 241 drivers/of/address.c parser->node = node; node 242 drivers/of/address.c parser->pna = of_n_addr_cells(node); node 245 drivers/of/address.c parser->range = of_get_property(node, name, &rlen); node 255 drivers/of/address.c struct device_node *node) node 257 drivers/of/address.c return parser_init(parser, node, "ranges"); node 262 drivers/of/address.c struct device_node *node) node 264 drivers/of/address.c return parser_init(parser, node, "dma-ranges"); node 282 drivers/of/address.c range->cpu_addr = of_translate_address(parser->node, node 295 drivers/of/address.c cpu_addr = of_translate_address(parser->node, node 923 drivers/of/address.c struct device_node *node = of_node_get(np); node 929 drivers/of/address.c if (!node) node 935 drivers/of/address.c naddr = of_n_addr_cells(node); node 936 drivers/of/address.c nsize = of_n_size_cells(node); node 938 drivers/of/address.c parent = __of_get_dma_parent(node); node 939 drivers/of/address.c of_node_put(node); node 941 drivers/of/address.c node = parent; node 942 drivers/of/address.c if (!node) node 945 drivers/of/address.c ranges = of_get_property(node, "dma-ranges", &len); node 967 drivers/of/address.c pna = of_n_addr_cells(node); node 990 drivers/of/address.c of_node_put(node); node 1006 drivers/of/address.c struct device_node *node = of_node_get(np); node 1011 drivers/of/address.c while (node) { node 1012 drivers/of/address.c if (of_property_read_bool(node, "dma-coherent")) { node 1013 drivers/of/address.c of_node_put(node); node 1016 drivers/of/address.c node = of_get_next_parent(node); node 1018 drivers/of/address.c of_node_put(node); node 686 drivers/of/base.c struct device_node *of_get_parent(const struct device_node *node) node 691 drivers/of/base.c if (!node) node 695 drivers/of/base.c np = of_node_get(node->parent); node 712 drivers/of/base.c struct device_node *of_get_next_parent(struct device_node *node) node 717 drivers/of/base.c if (!node) node 721 drivers/of/base.c parent = of_node_get(node->parent); node 722 drivers/of/base.c of_node_put(node); node 728 drivers/of/base.c static struct device_node *__of_get_next_child(const struct device_node *node, node 733 drivers/of/base.c if (!node) node 736 drivers/of/base.c next = prev ? prev->sibling : node->child; node 756 drivers/of/base.c struct device_node *of_get_next_child(const struct device_node *node, node 763 drivers/of/base.c next = __of_get_next_child(node, prev); node 777 drivers/of/base.c struct device_node *of_get_next_available_child(const struct device_node *node, node 783 drivers/of/base.c if (!node) node 787 drivers/of/base.c next = prev ? prev->sibling : node->child; node 812 drivers/of/base.c struct device_node *node; node 815 drivers/of/base.c node = of_find_node_by_path("/cpus"); node 820 drivers/of/base.c else if (node) { node 821 drivers/of/base.c next = node->child; node 822 drivers/of/base.c of_node_put(node); node 873 drivers/of/base.c struct device_node *of_get_child_by_name(const struct device_node *node, node 878 drivers/of/base.c for_each_child_of_node(node, child) node 903 drivers/of/base.c struct device_node *__of_find_node_by_full_path(struct device_node *node, node 908 drivers/of/base.c while (node && *path == '/') { node 909 drivers/of/base.c struct device_node *tmp = node; node 912 drivers/of/base.c node = __of_find_node_by_path(node, path); node 918 drivers/of/base.c return node; node 1109 drivers/of/base.c const struct device_node *node) node 1118 drivers/of/base.c score = __of_device_is_compatible(node, matches->compatible, node 1137 drivers/of/base.c const struct device_node *node) node 1143 drivers/of/base.c match = __of_match_node(matches, node); node 1201 drivers/of/base.c int of_modalias_node(struct device_node *node, char *modalias, int len) node 1206 drivers/of/base.c compatible = of_get_property(node, "compatible", &cplen); node 1315 drivers/of/base.c if (it->node) { node 1316 drivers/of/base.c of_node_put(it->node); node 1317 drivers/of/base.c it->node = NULL; node 1334 drivers/of/base.c it->node = of_find_node_by_phandle(it->phandle); node 1337 drivers/of/base.c if (!it->node) { node 1343 drivers/of/base.c if (of_property_read_u32(it->node, it->cells_name, node 1356 drivers/of/base.c it->node); node 1382 drivers/of/base.c if (it->node) { node 1383 drivers/of/base.c of_node_put(it->node); node 1384 drivers/of/base.c it->node = NULL; node 1436 drivers/of/base.c out_args->np = it.node; node 1439 drivers/of/base.c of_node_put(it.node); node 1456 drivers/of/base.c of_node_put(it.node); node 32 drivers/of/dynamic.c struct device_node *of_node_get(struct device_node *node) node 34 drivers/of/dynamic.c if (node) node 35 drivers/of/dynamic.c kobject_get(&node->kobj); node 36 drivers/of/dynamic.c return node; node 45 drivers/of/dynamic.c void of_node_put(struct device_node *node) node 47 drivers/of/dynamic.c if (node) node 48 drivers/of/dynamic.c kobject_put(&node->kobj); node 328 drivers/of/dynamic.c struct device_node *node = kobj_to_device_node(kobj); node 331 drivers/of/dynamic.c if (!of_node_check_flag(node, OF_DETACHED)) { node 332 drivers/of/dynamic.c pr_err("ERROR: Bad of_node_put() on %pOF\n", node); node 336 drivers/of/dynamic.c if (!of_node_check_flag(node, OF_DYNAMIC)) node 339 drivers/of/dynamic.c if (of_node_check_flag(node, OF_OVERLAY)) { node 341 drivers/of/dynamic.c if (!of_node_check_flag(node, OF_OVERLAY_FREE_CSET)) { node 344 drivers/of/dynamic.c node); node 353 drivers/of/dynamic.c if (node->properties) node 355 drivers/of/dynamic.c __func__, node); node 358 drivers/of/dynamic.c property_list_free(node->properties); node 359 drivers/of/dynamic.c property_list_free(node->deadprops); node 361 drivers/of/dynamic.c kfree(node->full_name); node 362 drivers/of/dynamic.c kfree(node->data); node 363 drivers/of/dynamic.c kfree(node); node 423 drivers/of/dynamic.c struct device_node *node; node 425 drivers/of/dynamic.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 426 drivers/of/dynamic.c if (!node) node 428 drivers/of/dynamic.c node->full_name = kstrdup(full_name, GFP_KERNEL); node 429 drivers/of/dynamic.c if (!node->full_name) { node 430 drivers/of/dynamic.c kfree(node); node 434 drivers/of/dynamic.c of_node_set_flag(node, OF_DYNAMIC); node 435 drivers/of/dynamic.c of_node_set_flag(node, OF_DETACHED); node 436 drivers/of/dynamic.c of_node_init(node); node 445 drivers/of/dynamic.c if (__of_add_property(node, new_pp)) { node 453 drivers/of/dynamic.c return node; node 456 drivers/of/dynamic.c of_node_put(node); /* Frees the node and properties */ node 473 drivers/of/dynamic.c list_del(&ce->node); node 692 drivers/of/dynamic.c list_for_each_entry_safe_reverse(ce, cen, &ocs->entries, node) node 713 drivers/of/dynamic.c list_for_each_entry(ce, &ocs->entries, node) { node 718 drivers/of/dynamic.c node) { node 745 drivers/of/dynamic.c list_for_each_entry(ce, &ocs->entries, node) { node 816 drivers/of/dynamic.c list_for_each_entry_reverse(ce, &ocs->entries, node) { node 820 drivers/of/dynamic.c list_for_each_entry_continue(ce, &ocs->entries, node) { node 845 drivers/of/dynamic.c list_for_each_entry_reverse(ce, &ocs->entries, node) { node 926 drivers/of/dynamic.c list_add_tail(&ce->node, &ocs->entries); node 82 drivers/of/fdt.c static bool of_fdt_device_is_available(const void *blob, unsigned long node) node 84 drivers/of/fdt.c const char *status = fdt_getprop(blob, node, "status", NULL); node 476 drivers/of/fdt.c static int __init __reserved_mem_reserve_reg(unsigned long node, node 486 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "reg", &len); node 496 drivers/of/fdt.c nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL; node 512 drivers/of/fdt.c fdt_reserved_mem_save_node(node, uname, base, size); node 524 drivers/of/fdt.c static int __init __reserved_mem_check_root(unsigned long node) node 528 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "#size-cells", NULL); node 532 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "#address-cells", NULL); node 536 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "ranges", NULL); node 545 drivers/of/fdt.c static int __init __fdt_scan_reserved_mem(unsigned long node, const char *uname, node 552 drivers/of/fdt.c if (__reserved_mem_check_root(node) != 0) { node 568 drivers/of/fdt.c if (!of_fdt_device_is_available(initial_boot_params, node)) node 571 drivers/of/fdt.c err = __reserved_mem_reserve_reg(node, uname); node 572 drivers/of/fdt.c if (err == -ENOENT && of_get_flat_dt_prop(node, "size", NULL)) node 573 drivers/of/fdt.c fdt_reserved_mem_save_node(node, uname, 0, 0); node 629 drivers/of/fdt.c int __init of_scan_flat_dt(int (*it)(unsigned long node, node 661 drivers/of/fdt.c int (*it)(unsigned long node, node 667 drivers/of/fdt.c int node; node 669 drivers/of/fdt.c fdt_for_each_subnode(node, blob, parent) { node 673 drivers/of/fdt.c pathp = fdt_get_name(blob, node, NULL); node 676 drivers/of/fdt.c rc = it(node, pathp, data); node 691 drivers/of/fdt.c int __init of_get_flat_dt_subnode_by_name(unsigned long node, const char *uname) node 693 drivers/of/fdt.c return fdt_subnode_offset(initial_boot_params, node, uname); node 710 drivers/of/fdt.c const void *__init of_get_flat_dt_prop(unsigned long node, const char *name, node 713 drivers/of/fdt.c return fdt_getprop(initial_boot_params, node, name, size); node 727 drivers/of/fdt.c unsigned long node, const char *compat) node 733 drivers/of/fdt.c cp = fdt_getprop(blob, node, "compatible", &cplen); node 753 drivers/of/fdt.c int __init of_flat_dt_is_compatible(unsigned long node, const char *compat) node 755 drivers/of/fdt.c return of_fdt_is_compatible(initial_boot_params, node, compat); node 761 drivers/of/fdt.c static int __init of_flat_dt_match(unsigned long node, const char *const *compat) node 769 drivers/of/fdt.c tmp = of_fdt_is_compatible(initial_boot_params, node, *compat); node 781 drivers/of/fdt.c uint32_t __init of_get_flat_dt_phandle(unsigned long node) node 783 drivers/of/fdt.c return fdt_get_phandle(initial_boot_params, node); node 791 drivers/of/fdt.c int (*iterator)(unsigned long node, const char *uname, int depth, void *data); node 875 drivers/of/fdt.c static void __init early_init_dt_check_for_initrd(unsigned long node) node 883 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "linux,initrd-start", &len); node 888 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "linux,initrd-end", &len); node 901 drivers/of/fdt.c static inline void early_init_dt_check_for_initrd(unsigned long node) node 960 drivers/of/fdt.c int __init early_init_dt_scan_root(unsigned long node, const char *uname, node 971 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "#size-cells", NULL); node 976 drivers/of/fdt.c prop = of_get_flat_dt_prop(node, "#address-cells", NULL); node 996 drivers/of/fdt.c int __init early_init_dt_scan_memory(unsigned long node, const char *uname, node 999 drivers/of/fdt.c const char *type = of_get_flat_dt_prop(node, "device_type", NULL); node 1008 drivers/of/fdt.c reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l); node 1010 drivers/of/fdt.c reg = of_get_flat_dt_prop(node, "reg", &l); node 1015 drivers/of/fdt.c hotpluggable = of_get_flat_dt_prop(node, "hotpluggable", NULL); node 1043 drivers/of/fdt.c int __init early_init_dt_scan_chosen(unsigned long node, const char *uname, node 1056 drivers/of/fdt.c early_init_dt_check_for_initrd(node); node 1059 drivers/of/fdt.c p = of_get_flat_dt_prop(node, "bootargs", &l); node 1083 drivers/of/fdt.c rng_seed = of_get_flat_dt_prop(node, "rng-seed", &l); node 1088 drivers/of/fdt.c fdt_nop_property(initial_boot_params, node, "rng-seed"); node 246 drivers/of/fdt_address.c u64 __init of_flat_dt_translate_address(unsigned long node) node 248 drivers/of/fdt_address.c return fdt_translate_address(initial_boot_params, node); node 8 drivers/of/kobj.c static int of_node_is_initialized(struct device_node *node) node 10 drivers/of/kobj.c return node && node->kobj.state_initialized; node 14 drivers/of/kobj.c int of_node_is_attached(struct device_node *node) node 16 drivers/of/kobj.c return node && node->kobj.state_in_sysfs; node 61 drivers/of/of_private.h int of_node_is_attached(struct device_node *node); node 126 drivers/of/of_private.h struct device_node *__of_find_node_by_full_path(struct device_node *node, node 155 drivers/of/of_private.h list_for_each_entry(_te, &(_oft)->te_list, node) node 159 drivers/of/of_private.h list_for_each_entry_reverse(_te, &(_oft)->te_list, node) node 51 drivers/of/of_reserved_mem.c void __init fdt_reserved_mem_save_node(unsigned long node, const char *uname, node 61 drivers/of/of_reserved_mem.c rmem->fdt_node = node; node 74 drivers/of/of_reserved_mem.c static int __init __reserved_mem_alloc_size(unsigned long node, node 85 drivers/of/of_reserved_mem.c prop = of_get_flat_dt_prop(node, "size", &len); node 95 drivers/of/of_reserved_mem.c nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL; node 97 drivers/of/of_reserved_mem.c prop = of_get_flat_dt_prop(node, "alignment", &len); node 109 drivers/of/of_reserved_mem.c && of_flat_dt_is_compatible(node, "shared-dma-pool") node 110 drivers/of/of_reserved_mem.c && of_get_flat_dt_prop(node, "reusable", NULL) node 111 drivers/of/of_reserved_mem.c && !of_get_flat_dt_prop(node, "no-map", NULL)) { node 118 drivers/of/of_reserved_mem.c prop = of_get_flat_dt_prop(node, "alloc-ranges", &len); node 246 drivers/of/of_reserved_mem.c unsigned long node = rmem->fdt_node; node 252 drivers/of/of_reserved_mem.c nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL; node 253 drivers/of/of_reserved_mem.c prop = of_get_flat_dt_prop(node, "phandle", &len); node 255 drivers/of/of_reserved_mem.c prop = of_get_flat_dt_prop(node, "linux,phandle", &len); node 260 drivers/of/of_reserved_mem.c err = __reserved_mem_alloc_size(node, rmem->name, node 275 drivers/of/of_reserved_mem.c static inline struct reserved_mem *__find_rmem(struct device_node *node) node 279 drivers/of/of_reserved_mem.c if (!node->phandle) node 283 drivers/of/of_reserved_mem.c if (reserved_mem[i].phandle == node->phandle) node 410 drivers/of/overlay.c struct target *target, struct device_node *node) node 418 drivers/of/overlay.c node_kbasename = kbasename(node->full_name); node 430 drivers/of/overlay.c tchild->name = __of_get_property(node, "name", NULL); node 436 drivers/of/overlay.c phandle = __of_get_property(node, "phandle", &size); node 449 drivers/of/overlay.c ret = build_changeset_next_level(ovcs, &target_child, node); node 454 drivers/of/overlay.c if (node->phandle && tchild->phandle) { node 459 drivers/of/overlay.c ret = build_changeset_next_level(ovcs, &target_child, node); node 544 drivers/of/overlay.c list_for_each_entry_continue(ce_2, &ovcs->cset.entries, node) { node 578 drivers/of/overlay.c list_for_each_entry_continue(ce_2, &ovcs->cset.entries, node) { node 616 drivers/of/overlay.c list_for_each_entry(ce_1, &ovcs->cset.entries, node) { node 691 drivers/of/overlay.c struct device_node *node; node 698 drivers/of/overlay.c node = of_find_node_by_phandle(val); node 699 drivers/of/overlay.c if (!node) node 702 drivers/of/overlay.c return node; node 707 drivers/of/overlay.c node = of_find_node_by_path(path); node 708 drivers/of/overlay.c if (!node) node 711 drivers/of/overlay.c return node; node 735 drivers/of/overlay.c struct device_node *node, *overlay_node; node 767 drivers/of/overlay.c for_each_child_of_node(tree, node) { node 768 drivers/of/overlay.c overlay_node = of_get_child_by_name(node, "__overlay__"); node 775 drivers/of/overlay.c node = of_get_child_by_name(tree, "__symbols__"); node 776 drivers/of/overlay.c if (node) { node 778 drivers/of/overlay.c of_node_put(node); node 788 drivers/of/overlay.c for_each_child_of_node(tree, node) { node 789 drivers/of/overlay.c overlay_node = of_get_child_by_name(node, "__overlay__"); node 795 drivers/of/overlay.c fragment->target = find_target(node); node 809 drivers/of/overlay.c node = of_get_child_by_name(tree, "__symbols__"); node 810 drivers/of/overlay.c if (node) { node 813 drivers/of/overlay.c fragment->overlay = node; node 1111 drivers/of/overlay.c list_for_each_entry(ce, &ovcs->cset.entries, node) { node 1144 drivers/of/overlay.c list_for_each_entry(remove_ce, &remove_ovcs->cset.entries, node) { node 65 drivers/of/pdt.c static struct property * __init of_pdt_build_one_prop(phandle node, char *prev, node 90 drivers/of/pdt.c err = of_pdt_prom_ops->nextprop(node, prev, p->name); node 95 drivers/of/pdt.c p->length = of_pdt_prom_ops->getproplen(node, p->name); node 102 drivers/of/pdt.c len = of_pdt_prom_ops->getproperty(node, p->name, node 112 drivers/of/pdt.c static struct property * __init of_pdt_build_prop_list(phandle node) node 116 drivers/of/pdt.c head = tail = of_pdt_build_one_prop(node, NULL, node 117 drivers/of/pdt.c ".node", &node, sizeof(node)); node 119 drivers/of/pdt.c tail->next = of_pdt_build_one_prop(node, NULL, NULL, NULL, 0); node 122 drivers/of/pdt.c tail->next = of_pdt_build_one_prop(node, tail->name, node 130 drivers/of/pdt.c static char * __init of_pdt_get_one_property(phandle node, const char *name) node 135 drivers/of/pdt.c len = of_pdt_prom_ops->getproplen(node, name); node 138 drivers/of/pdt.c len = of_pdt_prom_ops->getproperty(node, name, buf, len); node 144 drivers/of/pdt.c static struct device_node * __init of_pdt_create_node(phandle node, node 149 drivers/of/pdt.c if (!node) node 157 drivers/of/pdt.c dp->name = of_pdt_get_one_property(node, "name"); node 158 drivers/of/pdt.c dp->phandle = node; node 160 drivers/of/pdt.c dp->properties = of_pdt_build_prop_list(node); node 170 drivers/of/pdt.c phandle node) node 176 drivers/of/pdt.c dp = of_pdt_create_node(node, parent); node 187 drivers/of/pdt.c dp->child = of_pdt_build_tree(dp, of_pdt_prom_ops->getchild(node)); node 189 drivers/of/pdt.c node = of_pdt_prom_ops->getsibling(node); node 77 drivers/of/platform.c struct device_node *node = dev->of_node; node 82 drivers/of/platform.c while (node->parent) { node 87 drivers/of/platform.c reg = of_get_property(node, "reg", NULL); node 88 drivers/of/platform.c if (reg && (addr = of_translate_address(node, reg)) != OF_BAD_ADDR) { node 90 drivers/of/platform.c addr, node, dev_name(dev)); node 96 drivers/of/platform.c kbasename(node->full_name), dev_name(dev)); node 97 drivers/of/platform.c node = node->parent; node 219 drivers/of/platform.c static struct amba_device *of_amba_device_create(struct device_node *node, node 228 drivers/of/platform.c pr_debug("Creating amba device %pOF\n", node); node 230 drivers/of/platform.c if (!of_device_is_available(node) || node 231 drivers/of/platform.c of_node_test_and_set_flag(node, OF_POPULATED)) node 243 drivers/of/platform.c dev->dev.of_node = of_node_get(node); node 244 drivers/of/platform.c dev->dev.fwnode = &node->fwnode; node 253 drivers/of/platform.c prop = of_get_property(node, "arm,primecell-periphid", NULL); node 259 drivers/of/platform.c dev->irq[i] = irq_of_parse_and_map(node, i); node 261 drivers/of/platform.c ret = of_address_to_resource(node, 0, &dev->res); node 264 drivers/of/platform.c ret, node); node 271 drivers/of/platform.c ret, node); node 280 drivers/of/platform.c of_node_clear_flag(node, OF_POPULATED); node 284 drivers/of/platform.c static struct amba_device *of_amba_device_create(struct device_node *node, node 516 drivers/of/platform.c struct device_node *node; node 526 drivers/of/platform.c for_each_matching_node(node, reserved_mem_matches) node 527 drivers/of/platform.c of_platform_device_create(node, NULL, NULL); node 529 drivers/of/platform.c node = of_find_node_by_path("/firmware"); node 530 drivers/of/platform.c if (node) { node 531 drivers/of/platform.c of_platform_populate(node, NULL, NULL, NULL); node 532 drivers/of/platform.c of_node_put(node); node 531 drivers/of/property.c int of_graph_parse_endpoint(const struct device_node *node, node 534 drivers/of/property.c struct device_node *port_node = of_get_parent(node); node 537 drivers/of/property.c __func__, node); node 541 drivers/of/property.c endpoint->local_node = node; node 547 drivers/of/property.c of_property_read_u32(node, "reg", &endpoint->id); node 565 drivers/of/property.c struct device_node *node, *port; node 567 drivers/of/property.c node = of_get_child_by_name(parent, "ports"); node 568 drivers/of/property.c if (node) node 569 drivers/of/property.c parent = node; node 581 drivers/of/property.c of_node_put(node); node 610 drivers/of/property.c struct device_node *node; node 612 drivers/of/property.c node = of_get_child_by_name(parent, "ports"); node 613 drivers/of/property.c if (node) node 614 drivers/of/property.c parent = node; node 617 drivers/of/property.c of_node_put(node); node 668 drivers/of/property.c struct device_node *node = NULL; node 670 drivers/of/property.c for_each_endpoint_of_node(parent, node) { node 671 drivers/of/property.c of_graph_parse_endpoint(node, &endpoint); node 674 drivers/of/property.c return node; node 688 drivers/of/property.c struct device_node *of_graph_get_remote_endpoint(const struct device_node *node) node 691 drivers/of/property.c return of_parse_phandle(node, "remote-endpoint", 0); node 702 drivers/of/property.c struct device_node *of_graph_get_port_parent(struct device_node *node) node 706 drivers/of/property.c if (!node) node 713 drivers/of/property.c of_node_get(node); node 716 drivers/of/property.c for (depth = 3; depth && node; depth--) { node 717 drivers/of/property.c node = of_get_next_parent(node); node 718 drivers/of/property.c if (depth == 2 && !of_node_name_eq(node, "ports")) node 721 drivers/of/property.c return node; node 733 drivers/of/property.c const struct device_node *node) node 738 drivers/of/property.c np = of_graph_get_remote_endpoint(node); node 755 drivers/of/property.c struct device_node *of_graph_get_remote_port(const struct device_node *node) node 760 drivers/of/property.c np = of_graph_get_remote_endpoint(node); node 788 drivers/of/property.c struct device_node *of_graph_get_remote_node(const struct device_node *node, node 793 drivers/of/property.c endpoint_node = of_graph_get_endpoint_by_regs(node, port, endpoint); node 796 drivers/of/property.c port, endpoint, node); node 843 drivers/of/property.c const struct device_node *node = to_of_node(fwnode); node 846 drivers/of/property.c return of_property_count_elems_of_size(node, propname, node 851 drivers/of/property.c return of_property_read_u8_array(node, propname, val, nval); node 853 drivers/of/property.c return of_property_read_u16_array(node, propname, val, nval); node 855 drivers/of/property.c return of_property_read_u32_array(node, propname, val, nval); node 857 drivers/of/property.c return of_property_read_u64_array(node, propname, val, nval); node 868 drivers/of/property.c const struct device_node *node = to_of_node(fwnode); node 871 drivers/of/property.c of_property_read_string_array(node, propname, val, nval) : node 872 drivers/of/property.c of_property_count_strings(node, propname); node 893 drivers/of/property.c const struct device_node *node = to_of_node(fwnode); node 896 drivers/of/property.c for_each_available_child_of_node(node, child) node 968 drivers/of/property.c const struct device_node *node = to_of_node(fwnode); node 969 drivers/of/property.c struct device_node *port_node = of_get_parent(node); node 974 drivers/of/property.c of_property_read_u32(node, "reg", &endpoint->id); node 24 drivers/of/resolver.c struct device_node *node; node 30 drivers/of/resolver.c for_each_of_allnodes(node) { node 31 drivers/of/resolver.c if (node->phandle != OF_PHANDLE_ILLEGAL && node 32 drivers/of/resolver.c node->phandle > phandle) node 33 drivers/of/resolver.c phandle = node->phandle; node 321 drivers/of/unittest.c struct hlist_node node; node 337 drivers/of/unittest.c hash_for_each_possible(phandle_ht, nh, node, np->phandle) { node 351 drivers/of/unittest.c hash_add(phandle_ht, &nh->node, np->phandle); node 358 drivers/of/unittest.c hash_for_each_safe(phandle_ht, i, tmp, nh, node) { node 359 drivers/of/unittest.c hash_del(&nh->node); node 38 drivers/opp/core.c list_for_each_entry(opp_dev, &opp_table->dev_list, node) node 50 drivers/opp/core.c list_for_each_entry(opp_table, &opp_tables, node) { node 230 drivers/opp/core.c list_for_each_entry(opp, &opp_table->opp_list, node) { node 309 drivers/opp/core.c list_for_each_entry(opp, &opp_table->opp_list, node) { node 386 drivers/opp/core.c list_for_each_entry(temp_opp, &opp_table->opp_list, node) { node 435 drivers/opp/core.c list_for_each_entry(temp_opp, &opp_table->opp_list, node) { node 459 drivers/opp/core.c list_for_each_entry(temp_opp, &opp_table->opp_list, node) { node 551 drivers/opp/core.c list_for_each_entry(temp_opp, &opp_table->opp_list, node) { node 609 drivers/opp/core.c list_for_each_entry(temp_opp, &opp_table->opp_list, node) { node 914 drivers/opp/core.c list_del(&opp_dev->node); node 930 drivers/opp/core.c list_add(&opp_dev->node, &opp_table->dev_list); node 994 drivers/opp/core.c list_add(&opp_table->node, &opp_tables); node 1056 drivers/opp/core.c list_for_each_entry_safe(opp_dev, temp, &opp_table->dev_list, node) { node 1069 drivers/opp/core.c list_del(&opp_table->node); node 1079 drivers/opp/core.c list_for_each_entry_safe(opp, tmp, &opp_table->opp_list, node) { node 1124 drivers/opp/core.c list_del(&opp->node); node 1181 drivers/opp/core.c list_for_each_entry(opp, &opp_table->opp_list, node) { node 1222 drivers/opp/core.c list_for_each_entry_safe(opp, temp, &opp_table->opp_list, node) { node 1255 drivers/opp/core.c INIT_LIST_HEAD(&opp->node); node 1299 drivers/opp/core.c list_for_each_entry(opp, &opp_table->opp_list, node) { node 1301 drivers/opp/core.c *head = &opp->node; node 1349 drivers/opp/core.c list_add(&new_opp->node, head); node 1982 drivers/opp/core.c list_for_each_entry(opp, &src_table->opp_list, node) { node 2066 drivers/opp/core.c list_for_each_entry(tmp_opp, &opp_table->opp_list, node) { node 224 drivers/opp/cpu.c list_for_each_entry(opp_dev, &opp_table->dev_list, node) node 153 drivers/opp/debugfs.c list_for_each_entry(new_dev, &opp_table->dev_list, node) node 51 drivers/opp/of.c list_for_each_entry(opp_table, &opp_tables, node) { node 82 drivers/opp/of.c list_for_each_entry(opp, &opp_table->opp_list, node) { node 124 drivers/opp/of.c list_for_each_entry(opp_table, &opp_tables, node) { node 693 drivers/opp/of.c list_for_each_entry(opp, &opp_table->opp_list, node) node 72 drivers/opp/opp.h struct list_head node; node 107 drivers/opp/opp.h struct list_head node; node 164 drivers/opp/opp.h struct list_head node; node 698 drivers/parisc/lba_pci.c list_for_each_entry(child, &bus->children, node) node 241 drivers/pci/controller/dwc/pci-dra7xx.c struct device_node *node = dev->of_node; node 242 drivers/pci/controller/dwc/pci-dra7xx.c struct device_node *pcie_intc_node = of_get_next_child(node, NULL); node 1018 drivers/pci/controller/dwc/pci-imx6.c struct device_node *node = dev->of_node; node 1037 drivers/pci/controller/dwc/pci-imx6.c np = of_parse_phandle(node, "fsl,imx7d-pcie-phy", 0); node 1059 drivers/pci/controller/dwc/pci-imx6.c imx6_pcie->reset_gpio = of_get_named_gpio(node, "reset-gpio", 0); node 1060 drivers/pci/controller/dwc/pci-imx6.c imx6_pcie->gpio_active_high = of_property_read_bool(node, node 1149 drivers/pci/controller/dwc/pci-imx6.c if (of_property_read_u32(node, "fsl,tx-deemph-gen1", node 1153 drivers/pci/controller/dwc/pci-imx6.c if (of_property_read_u32(node, "fsl,tx-deemph-gen2-3p5db", node 1157 drivers/pci/controller/dwc/pci-imx6.c if (of_property_read_u32(node, "fsl,tx-deemph-gen2-6db", node 1161 drivers/pci/controller/dwc/pci-imx6.c if (of_property_read_u32(node, "fsl,tx-swing-full", node 1165 drivers/pci/controller/dwc/pci-imx6.c if (of_property_read_u32(node, "fsl,tx-swing-low", node 1170 drivers/pci/controller/dwc/pci-imx6.c ret = of_property_read_u32(node, "fsl,max-link-speed", node 115 drivers/pci/controller/dwc/pcie-armada8k.c struct device_node *node = dev->of_node; node 120 drivers/pci/controller/dwc/pcie-armada8k.c pcie->phy[i] = devm_of_phy_get_by_index(dev, node, i); node 505 drivers/pci/controller/dwc/pcie-designware-host.c list_for_each_entry(child, &pp->root_bus->children, node) node 1037 drivers/pci/controller/dwc/pcie-tegra194.c list_for_each_entry(child, &pp->root_bus->children, node) { node 772 drivers/pci/controller/pci-aardvark.c struct device_node *node = dev->of_node; node 807 drivers/pci/controller/pci-aardvark.c pci_msi_create_irq_domain(of_node_to_fwnode(node), node 826 drivers/pci/controller/pci-aardvark.c struct device_node *node = dev->of_node; node 831 drivers/pci/controller/pci-aardvark.c pcie_intc_node = of_get_next_child(node, NULL); node 1043 drivers/pci/controller/pci-mvebu.c list_for_each_entry(child, &bus->children, node) node 2825 drivers/pci/controller/pci-tegra.c list_for_each_entry(child, &host->bus->children, node) node 357 drivers/pci/controller/pci-thunder-pem.c int node = acpi_get_node(root->device->handle); node 360 drivers/pci/controller/pci-thunder-pem.c if (node == NUMA_NO_NODE) node 361 drivers/pci/controller/pci-thunder-pem.c node = 0; node 364 drivers/pci/controller/pci-thunder-pem.c index -= node * PEM_MAX_DOM_IN_NODE; node 365 drivers/pci/controller/pci-thunder-pem.c res_pem->start = PEM_RES_BASE | FIELD_PREP(PEM_NODE_MASK, node) | node 215 drivers/pci/controller/pci-versatile.c list_for_each_entry(child, &bus->children, node) node 33 drivers/pci/controller/pci-xgene-msi.c struct device_node *node; node 250 drivers/pci/controller/pci-xgene-msi.c msi->msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(msi->node), node 463 drivers/pci/controller/pci-xgene-msi.c xgene_msi->node = pdev->dev.of_node; node 65 drivers/pci/controller/pci-xgene.c struct device_node *node; node 220 drivers/pci/controller/pci-xgene.c entry = list_first_entry(&list, struct resource_entry, node); node 539 drivers/pci/controller/pci-xgene.c struct device_node *np = port->node; node 622 drivers/pci/controller/pci-xgene.c port->node = of_node_get(dn); node 626 drivers/pci/controller/pci-xgene.c if (of_device_is_compatible(port->node, "apm,xgene-pcie")) node 665 drivers/pci/controller/pci-xgene.c list_for_each_entry(child, &bus->children, node) node 709 drivers/pci/controller/pcie-altera.c struct device_node *node = dev->of_node; node 712 drivers/pci/controller/pcie-altera.c pcie->irq_domain = irq_domain_add_linear(node, PCI_NUM_INTX, node 873 drivers/pci/controller/pcie-altera.c list_for_each_entry(child, &bus->children, node) node 438 drivers/pci/controller/pcie-iproc-msi.c static int iproc_msi_alloc_domains(struct device_node *node, node 446 drivers/pci/controller/pcie-iproc-msi.c msi->msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(node), node 511 drivers/pci/controller/pcie-iproc-msi.c int iproc_msi_init(struct iproc_pcie *pcie, struct device_node *node) node 517 drivers/pci/controller/pcie-iproc-msi.c if (!of_device_is_compatible(node, "brcm,iproc-msi")) node 520 drivers/pci/controller/pcie-iproc-msi.c if (!of_find_property(node, "msi-controller", NULL)) node 536 drivers/pci/controller/pcie-iproc-msi.c msi->nr_irqs = of_irq_count(node); node 577 drivers/pci/controller/pcie-iproc-msi.c if (of_find_property(node, "brcm,pcie-msi-inten", NULL)) node 592 drivers/pci/controller/pcie-iproc-msi.c unsigned int irq = irq_of_parse_and_map(node, i); node 613 drivers/pci/controller/pcie-iproc-msi.c ret = iproc_msi_alloc_domains(node, msi); node 1199 drivers/pci/controller/pcie-iproc.c head = &tmp->node; node 1564 drivers/pci/controller/pcie-iproc.c list_for_each_entry(child, &host->bus->children, node) node 114 drivers/pci/controller/pcie-iproc.h int iproc_msi_init(struct iproc_pcie *pcie, struct device_node *node); node 118 drivers/pci/controller/pcie-iproc.h struct device_node *node) node 157 drivers/pci/controller/pcie-mediatek.c int (*setup_irq)(struct mtk_pcie_port *port, struct device_node *node); node 571 drivers/pci/controller/pcie-mediatek.c struct device_node *node) node 578 drivers/pci/controller/pcie-mediatek.c pcie_intc_node = of_get_next_child(node, NULL); node 641 drivers/pci/controller/pcie-mediatek.c struct device_node *node) node 648 drivers/pci/controller/pcie-mediatek.c err = mtk_pcie_init_irq_domain(port, node); node 894 drivers/pci/controller/pcie-mediatek.c struct device_node *node, node 964 drivers/pci/controller/pcie-mediatek.c err = pcie->soc->setup_irq(port, node); node 1022 drivers/pci/controller/pcie-mediatek.c struct device_node *node = dev->of_node, *child; node 1060 drivers/pci/controller/pcie-mediatek.c for_each_available_child_of_node(node, child) { node 427 drivers/pci/controller/pcie-mobiveil.c struct device_node *node = dev->of_node; node 453 drivers/pci/controller/pcie-mobiveil.c if (of_property_read_u32(node, "apio-wins", &pcie->apio_wins)) node 456 drivers/pci/controller/pcie-mobiveil.c if (of_property_read_u32(node, "ppio-wins", &pcie->ppio_wins)) node 836 drivers/pci/controller/pcie-mobiveil.c struct device_node *node = dev->of_node; node 840 drivers/pci/controller/pcie-mobiveil.c pcie->intx_domain = irq_domain_add_linear(node, PCI_NUM_INTX, node 939 drivers/pci/controller/pcie-mobiveil.c list_for_each_entry(child, &bus->children, node) node 489 drivers/pci/controller/pcie-rcar.c list_for_each_entry(child, &bus->children, node) node 1065 drivers/pci/controller/pcie-rockchip-host.c list_for_each_entry(child, &bus->children, node) node 29 drivers/pci/controller/pcie-rockchip.c struct device_node *node = dev->of_node; node 59 drivers/pci/controller/pcie-rockchip.c err = of_property_read_u32(node, "num-lanes", &rockchip->lanes); node 67 drivers/pci/controller/pcie-rockchip.c rockchip->link_gen = of_pci_get_max_link_speed(node); node 547 drivers/pci/controller/pcie-xilinx-nwl.c struct device_node *node = dev->of_node; node 550 drivers/pci/controller/pcie-xilinx-nwl.c legacy_intc_node = of_get_next_child(node, NULL); node 888 drivers/pci/controller/pcie-xilinx-nwl.c list_for_each_entry(child, &bus->children, node) node 504 drivers/pci/controller/pcie-xilinx.c struct device_node *node = dev->of_node; node 509 drivers/pci/controller/pcie-xilinx.c pcie_intc_node = of_get_next_child(node, NULL); node 526 drivers/pci/controller/pcie-xilinx.c port->msi_domain = irq_domain_add_linear(node, node 583 drivers/pci/controller/pcie-xilinx.c struct device_node *node = dev->of_node; node 587 drivers/pci/controller/pcie-xilinx.c err = of_address_to_resource(node, 0, ®s); node 597 drivers/pci/controller/pcie-xilinx.c port->irq = irq_of_parse_and_map(node, 0); node 681 drivers/pci/controller/pcie-xilinx.c list_for_each_entry(child, &bus->children, node) node 68 drivers/pci/controller/vmd.c struct list_head node; node 147 drivers/pci/controller/vmd.c list_add_tail_rcu(&vmdirq->node, &vmdirq->irq->irq_list); node 163 drivers/pci/controller/vmd.c list_del_rcu(&vmdirq->node); node 238 drivers/pci/controller/vmd.c INIT_LIST_HEAD(&vmdirq->node); node 673 drivers/pci/controller/vmd.c sd->node = pcibus_to_node(vmd->dev->bus); node 709 drivers/pci/controller/vmd.c list_for_each_entry(child, &vmd->bus->children, node) node 726 drivers/pci/controller/vmd.c list_for_each_entry_rcu(vmdirq, &irqs->irq_list, node) node 81 drivers/pci/hotplug/acpiphp.h struct list_head node; node 152 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry_safe(slot, next, &bridge->slots, node) { node 269 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry(slot, &bridge->slots, node) node 285 drivers/pci/hotplug/acpiphp_glue.c list_add_tail(&slot->node, &bridge->slots); node 335 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry(slot, &bridge->slots, node) { node 377 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry(tmp, &bus->children, node) { node 697 drivers/pci/hotplug/acpiphp_glue.c list_for_each_entry(slot, &bridge->slots, node) { node 462 drivers/pci/hotplug/cpqphp.h struct pci_resource *node) node 464 drivers/pci/hotplug/cpqphp.h if (!node || !head) node 466 drivers/pci/hotplug/cpqphp.h node->next = *head; node 467 drivers/pci/hotplug/cpqphp.h *head = node; node 400 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_resource *node; node 425 drivers/pci/hotplug/cpqphp_ctrl.c node = *head; node 427 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length & (alignment - 1)) { node 436 drivers/pci/hotplug/cpqphp_ctrl.c temp_dword = (node->length | (alignment-1)) + 1 - alignment; node 438 drivers/pci/hotplug/cpqphp_ctrl.c split_node->base = node->base; node 441 drivers/pci/hotplug/cpqphp_ctrl.c node->length -= temp_dword; node 442 drivers/pci/hotplug/cpqphp_ctrl.c node->base += split_node->length; node 446 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node; node 449 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length < alignment) node 453 drivers/pci/hotplug/cpqphp_ctrl.c if (*head == node) { node 454 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; node 457 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) node 460 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; node 462 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; node 464 drivers/pci/hotplug/cpqphp_ctrl.c return node; node 476 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_resource *node; node 485 drivers/pci/hotplug/cpqphp_ctrl.c node = *head; node 487 drivers/pci/hotplug/cpqphp_ctrl.c while (node->next) { node 488 drivers/pci/hotplug/cpqphp_ctrl.c prevnode = node; node 489 drivers/pci/hotplug/cpqphp_ctrl.c node = node->next; node 493 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length < alignment) node 496 drivers/pci/hotplug/cpqphp_ctrl.c if (node->base & (alignment - 1)) { node 498 drivers/pci/hotplug/cpqphp_ctrl.c temp_dword = (node->base | (alignment-1)) + 1; node 499 drivers/pci/hotplug/cpqphp_ctrl.c if ((node->length - (temp_dword - node->base)) < alignment) node 502 drivers/pci/hotplug/cpqphp_ctrl.c node->length -= (temp_dword - node->base); node 503 drivers/pci/hotplug/cpqphp_ctrl.c node->base = temp_dword; node 506 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length & (alignment - 1)) node 510 drivers/pci/hotplug/cpqphp_ctrl.c return node; node 512 drivers/pci/hotplug/cpqphp_ctrl.c kfree(node); node 529 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_resource *node; node 542 drivers/pci/hotplug/cpqphp_ctrl.c for (node = *head; node; node = node->next) { node 543 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length < size) node 546 drivers/pci/hotplug/cpqphp_ctrl.c if (node->base & (size - 1)) { node 550 drivers/pci/hotplug/cpqphp_ctrl.c temp_dword = (node->base | (size-1)) + 1; node 553 drivers/pci/hotplug/cpqphp_ctrl.c if ((node->length - (temp_dword - node->base)) < size) node 561 drivers/pci/hotplug/cpqphp_ctrl.c split_node->base = node->base; node 562 drivers/pci/hotplug/cpqphp_ctrl.c split_node->length = temp_dword - node->base; node 563 drivers/pci/hotplug/cpqphp_ctrl.c node->base = temp_dword; node 564 drivers/pci/hotplug/cpqphp_ctrl.c node->length -= split_node->length; node 567 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; node 568 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; node 572 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length > size) { node 581 drivers/pci/hotplug/cpqphp_ctrl.c split_node->base = node->base + size; node 582 drivers/pci/hotplug/cpqphp_ctrl.c split_node->length = node->length - size; node 583 drivers/pci/hotplug/cpqphp_ctrl.c node->length = size; node 586 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; node 587 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; node 591 drivers/pci/hotplug/cpqphp_ctrl.c if (node->base & 0x300L) node 597 drivers/pci/hotplug/cpqphp_ctrl.c if (*head == node) { node 598 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; node 601 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) node 604 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; node 606 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; node 610 drivers/pci/hotplug/cpqphp_ctrl.c return node; node 723 drivers/pci/hotplug/cpqphp_ctrl.c struct pci_resource *node; node 733 drivers/pci/hotplug/cpqphp_ctrl.c for (node = *head; node; node = node->next) { node 735 drivers/pci/hotplug/cpqphp_ctrl.c __func__, size, node, node->base, node->length); node 736 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length < size) node 739 drivers/pci/hotplug/cpqphp_ctrl.c if (node->base & (size - 1)) { node 744 drivers/pci/hotplug/cpqphp_ctrl.c temp_dword = (node->base | (size-1)) + 1; node 747 drivers/pci/hotplug/cpqphp_ctrl.c if ((node->length - (temp_dword - node->base)) < size) node 755 drivers/pci/hotplug/cpqphp_ctrl.c split_node->base = node->base; node 756 drivers/pci/hotplug/cpqphp_ctrl.c split_node->length = temp_dword - node->base; node 757 drivers/pci/hotplug/cpqphp_ctrl.c node->base = temp_dword; node 758 drivers/pci/hotplug/cpqphp_ctrl.c node->length -= split_node->length; node 760 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; node 761 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; node 765 drivers/pci/hotplug/cpqphp_ctrl.c if (node->length > size) { node 775 drivers/pci/hotplug/cpqphp_ctrl.c split_node->base = node->base + size; node 776 drivers/pci/hotplug/cpqphp_ctrl.c split_node->length = node->length - size; node 777 drivers/pci/hotplug/cpqphp_ctrl.c node->length = size; node 780 drivers/pci/hotplug/cpqphp_ctrl.c split_node->next = node->next; node 781 drivers/pci/hotplug/cpqphp_ctrl.c node->next = split_node; node 787 drivers/pci/hotplug/cpqphp_ctrl.c if (*head == node) { node 788 drivers/pci/hotplug/cpqphp_ctrl.c *head = node->next; node 791 drivers/pci/hotplug/cpqphp_ctrl.c while (prevnode->next != node) node 794 drivers/pci/hotplug/cpqphp_ctrl.c prevnode->next = node->next; node 796 drivers/pci/hotplug/cpqphp_ctrl.c node->next = NULL; node 799 drivers/pci/hotplug/cpqphp_ctrl.c return node; node 1421 drivers/pci/hotplug/cpqphp_pci.c struct pci_resource *node; node 1428 drivers/pci/hotplug/cpqphp_pci.c node = func->io_head; node 1430 drivers/pci/hotplug/cpqphp_pci.c while (node) { node 1431 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; node 1432 drivers/pci/hotplug/cpqphp_pci.c return_resource(&(resources->io_head), node); node 1433 drivers/pci/hotplug/cpqphp_pci.c node = t_node; node 1436 drivers/pci/hotplug/cpqphp_pci.c node = func->mem_head; node 1438 drivers/pci/hotplug/cpqphp_pci.c while (node) { node 1439 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; node 1440 drivers/pci/hotplug/cpqphp_pci.c return_resource(&(resources->mem_head), node); node 1441 drivers/pci/hotplug/cpqphp_pci.c node = t_node; node 1444 drivers/pci/hotplug/cpqphp_pci.c node = func->p_mem_head; node 1446 drivers/pci/hotplug/cpqphp_pci.c while (node) { node 1447 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; node 1448 drivers/pci/hotplug/cpqphp_pci.c return_resource(&(resources->p_mem_head), node); node 1449 drivers/pci/hotplug/cpqphp_pci.c node = t_node; node 1452 drivers/pci/hotplug/cpqphp_pci.c node = func->bus_head; node 1454 drivers/pci/hotplug/cpqphp_pci.c while (node) { node 1455 drivers/pci/hotplug/cpqphp_pci.c t_node = node->next; node 1456 drivers/pci/hotplug/cpqphp_pci.c return_resource(&(resources->bus_head), node); node 1457 drivers/pci/hotplug/cpqphp_pci.c node = t_node; node 38 drivers/pci/of.c struct device_node *node; node 41 drivers/pci/of.c node = pcibios_get_phb_of_node(bus); node 43 drivers/pci/of.c node = of_node_get(bus->self->dev.of_node); node 44 drivers/pci/of.c if (node && of_property_read_bool(node, "external-facing")) node 48 drivers/pci/of.c bus->dev.of_node = node; node 106 drivers/pci/of.c static inline int __of_pci_pci_compare(struct device_node *node, node 111 drivers/pci/of.c devfn = of_pci_get_devfn(node); node 121 drivers/pci/of.c struct device_node *node, *node2; node 123 drivers/pci/of.c for_each_child_of_node(parent, node) { node 124 drivers/pci/of.c if (__of_pci_pci_compare(node, devfn)) node 125 drivers/pci/of.c return node; node 131 drivers/pci/of.c if (of_node_name_eq(node, "multifunc-device")) { node 132 drivers/pci/of.c for_each_child_of_node(node, node2) { node 134 drivers/pci/of.c of_node_put(node); node 173 drivers/pci/of.c int of_pci_parse_bus_range(struct device_node *node, struct resource *res) node 178 drivers/pci/of.c error = of_property_read_u32_array(node, "bus-range", bus_range, node 183 drivers/pci/of.c res->name = node->name; node 201 drivers/pci/of.c int of_get_pci_domain_nr(struct device_node *node) node 206 drivers/pci/of.c error = of_property_read_u32(node, "linux,pci-domain", &domain); node 545 drivers/pci/of.c int of_pci_get_max_link_speed(struct device_node *node) node 549 drivers/pci/of.c if (of_property_read_u32(node, "max-link-speed", &max_link_speed) || node 55 drivers/pci/pci-acpi.c entry = list_first_entry(&list, struct resource_entry, node); node 25 drivers/pci/pci-driver.c struct list_head node; node 71 drivers/pci/pci-driver.c list_add_tail(&dynid->node, &drv->dynids.list); node 83 drivers/pci/pci-driver.c list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { node 84 drivers/pci/pci-driver.c list_del(&dynid->node); node 183 drivers/pci/pci-driver.c list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { node 190 drivers/pci/pci-driver.c list_del(&dynid->node); node 263 drivers/pci/pci-driver.c list_for_each_entry(dynid, &drv->dynids.list, node) { node 334 drivers/pci/pci-driver.c int error, node, cpu; node 342 drivers/pci/pci-driver.c node = dev_to_node(&dev->dev); node 351 drivers/pci/pci-driver.c if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || node 355 drivers/pci/pci-driver.c cpu = cpumask_any_and(cpumask_of_node(node), cpu_online_mask); node 323 drivers/pci/pci-sysfs.c int node, ret; node 328 drivers/pci/pci-sysfs.c ret = kstrtoint(buf, 0, &node); node 332 drivers/pci/pci-sysfs.c if ((node < 0 && node != NUMA_NO_NODE) || node >= MAX_NUMNODES) node 335 drivers/pci/pci-sysfs.c if (node != NUMA_NO_NODE && !node_online(node)) node 340 drivers/pci/pci-sysfs.c node); node 342 drivers/pci/pci-sysfs.c dev->numa_node = node; node 158 drivers/pci/pci.c list_for_each_entry(tmp, &bus->children, node) { node 604 drivers/pci/pci.h int of_pci_parse_bus_range(struct device_node *node, struct resource *res); node 605 drivers/pci/pci.h int of_get_pci_domain_nr(struct device_node *node); node 606 drivers/pci/pci.h int of_pci_get_max_link_speed(struct device_node *node); node 614 drivers/pci/pci.h of_pci_parse_bus_range(struct device_node *node, struct resource *res) node 620 drivers/pci/pci.h of_get_pci_domain_nr(struct device_node *node) node 626 drivers/pci/pci.h of_pci_get_max_link_speed(struct device_node *node) node 553 drivers/pci/probe.c INIT_LIST_HEAD(&b->node); node 903 drivers/pci/probe.c list_move_tail(&window->node, &bridge->windows); node 928 drivers/pci/probe.c list_add_tail(&bus->node, &pci_root_buses); node 1066 drivers/pci/probe.c list_add_tail(&child->node, &parent->children); node 2955 drivers/pci/probe.c list_for_each_entry(child, &bus->children, node) node 52 drivers/pci/remove.c list_del(&bus->node); node 118 drivers/pci/search.c list_for_each_entry(tmp, &bus->children, node) { node 167 drivers/pci/search.c n = from ? from->node.next : pci_root_buses.next; node 169 drivers/pci/search.c b = list_entry(n, struct pci_bus, node); node 1439 drivers/pci/setup-bus.c list_for_each_entry(child, &b->children, node) node 1620 drivers/pci/setup-bus.c list_for_each_entry(child_bus, &bus->children, node) { node 1814 drivers/pci/setup-bus.c list_for_each_entry(root_bus, &pci_root_buses, node) { node 72 drivers/pci/xen-pcifront.c sd->sd.node = first_online_node; node 307 drivers/pcmcia/cistpl.c list_for_each_entry(cis, &s->cis_cache, node) { node 325 drivers/pcmcia/cistpl.c list_add(&cis->node, &s->cis_cache); node 339 drivers/pcmcia/cistpl.c list_for_each_entry(cis, &s->cis_cache, node) node 341 drivers/pcmcia/cistpl.c list_del(&cis->node); node 361 drivers/pcmcia/cistpl.c cis = list_entry(l, struct cis_cache_entry, node); node 362 drivers/pcmcia/cistpl.c list_del(&cis->node); node 385 drivers/pcmcia/cistpl.c list_for_each_entry(cis, &s->cis_cache, node) { node 39 drivers/pcmcia/cs_internal.h struct list_head node; node 81 drivers/pcmcia/ds.c struct list_head node; node 124 drivers/pcmcia/ds.c list_add_tail(&dynid->node, &pdrv->dynids.list); node 141 drivers/pcmcia/ds.c list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { node 142 drivers/pcmcia/ds.c list_del(&dynid->node); node 913 drivers/pcmcia/ds.c list_for_each_entry(dynid, &p_drv->dynids.list, node) { node 85 drivers/pcmcia/soc_common.h struct list_head node; node 1100 drivers/pcmcia/yenta_socket.c node) { node 164 drivers/perf/arm-ccn.c struct hlist_node node; node 178 drivers/perf/arm-ccn.c struct arm_ccn_component *node; node 187 drivers/perf/arm-ccn.c static int arm_ccn_node_to_xp(int node) node 189 drivers/perf/arm-ccn.c return node / CCN_NUM_XP_PORTS; node 192 drivers/perf/arm-ccn.c static int arm_ccn_node_to_xp_port(int node) node 194 drivers/perf/arm-ccn.c return node % CCN_NUM_XP_PORTS; node 232 drivers/perf/arm-ccn.c static CCN_FORMAT_ATTR(node, "config:0-7"); node 671 drivers/perf/arm-ccn.c source = &ccn->node[node_xp]; node 782 drivers/perf/arm-ccn.c if (!arm_ccn_pmu_type_eq(type, ccn->node[node_xp].type)) { node 1205 drivers/perf/arm-ccn.c static int arm_ccn_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 1207 drivers/perf/arm-ccn.c struct arm_ccn_dt *dt = hlist_entry_safe(node, struct arm_ccn_dt, node); node 1307 drivers/perf/arm-ccn.c &ccn->dt.node); node 1317 drivers/perf/arm-ccn.c &ccn->dt.node); node 1332 drivers/perf/arm-ccn.c &ccn->dt.node); node 1403 drivers/perf/arm-ccn.c component = &ccn->node[id]; node 1409 drivers/perf/arm-ccn.c component = &ccn->node[id]; node 1514 drivers/perf/arm-ccn.c ccn->node = devm_kcalloc(ccn->dev, ccn->num_nodes, sizeof(*ccn->node), node 1516 drivers/perf/arm-ccn.c ccn->xp = devm_kcalloc(ccn->dev, ccn->num_xps, sizeof(*ccn->node), node 1518 drivers/perf/arm-ccn.c if (!ccn->node || !ccn->xp) node 767 drivers/perf/arm_dsu_pmu.c static int dsu_pmu_cpu_online(unsigned int cpu, struct hlist_node *node) node 769 drivers/perf/arm_dsu_pmu.c struct dsu_pmu *dsu_pmu = hlist_entry_safe(node, struct dsu_pmu, node 785 drivers/perf/arm_dsu_pmu.c static int dsu_pmu_cpu_teardown(unsigned int cpu, struct hlist_node *node) node 788 drivers/perf/arm_dsu_pmu.c struct dsu_pmu *dsu_pmu = hlist_entry_safe(node, struct dsu_pmu, node 615 drivers/perf/arm_pmu.c static int arm_perf_starting_cpu(unsigned int cpu, struct hlist_node *node) node 617 drivers/perf/arm_pmu.c struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node); node 638 drivers/perf/arm_pmu.c static int arm_perf_teardown_cpu(unsigned int cpu, struct hlist_node *node) node 640 drivers/perf/arm_pmu.c struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node); node 757 drivers/perf/arm_pmu.c &cpu_pmu->node); node 769 drivers/perf/arm_pmu.c &cpu_pmu->node); node 778 drivers/perf/arm_pmu.c &cpu_pmu->node); node 60 drivers/perf/arm_pmu_platform.c static bool pmu_has_irq_affinity(struct device_node *node) node 62 drivers/perf/arm_pmu_platform.c return !!of_find_property(node, "interrupt-affinity", NULL); node 65 drivers/perf/arm_pmu_platform.c static int pmu_parse_irq_affinity(struct device_node *node, int i) node 75 drivers/perf/arm_pmu_platform.c if (!pmu_has_irq_affinity(node)) node 78 drivers/perf/arm_pmu_platform.c dn = of_parse_phandle(node, "interrupt-affinity", i); node 81 drivers/perf/arm_pmu_platform.c i, node); node 196 drivers/perf/arm_pmu_platform.c struct device_node *node = pdev->dev.of_node; node 210 drivers/perf/arm_pmu_platform.c if (node && (of_id = of_match_node(of_table, pdev->dev.of_node))) { node 229 drivers/perf/arm_pmu_platform.c pr_info("%pOF: failed to probe PMU!\n", node); node 246 drivers/perf/arm_pmu_platform.c pr_info("%pOF: failed to register PMU devices!\n", node); node 102 drivers/perf/arm_smmuv3_pmu.c struct hlist_node node; node 585 drivers/perf/arm_smmuv3_pmu.c static int smmu_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 590 drivers/perf/arm_smmuv3_pmu.c smmu_pmu = hlist_entry_safe(node, struct smmu_pmu, node); node 814 drivers/perf/arm_smmuv3_pmu.c &smmu_pmu->node); node 836 drivers/perf/arm_smmuv3_pmu.c cpuhp_state_remove_instance_nocalls(cpuhp_state_num, &smmu_pmu->node); node 845 drivers/perf/arm_smmuv3_pmu.c cpuhp_state_remove_instance_nocalls(cpuhp_state_num, &smmu_pmu->node); node 1071 drivers/perf/arm_spe_pmu.c static int arm_spe_pmu_cpu_startup(unsigned int cpu, struct hlist_node *node) node 1075 drivers/perf/arm_spe_pmu.c spe_pmu = hlist_entry_safe(node, struct arm_spe_pmu, hotplug_node); node 1083 drivers/perf/arm_spe_pmu.c static int arm_spe_pmu_cpu_teardown(unsigned int cpu, struct hlist_node *node) node 1087 drivers/perf/arm_spe_pmu.c spe_pmu = hlist_entry_safe(node, struct arm_spe_pmu, hotplug_node); node 72 drivers/perf/fsl_imx8_ddr_perf.c struct hlist_node node; node 518 drivers/perf/fsl_imx8_ddr_perf.c static int ddr_perf_offline_cpu(unsigned int cpu, struct hlist_node *node) node 520 drivers/perf/fsl_imx8_ddr_perf.c struct ddr_pmu *pmu = hlist_entry_safe(node, struct ddr_pmu, node); node 583 drivers/perf/fsl_imx8_ddr_perf.c ret = cpuhp_state_add_instance_nocalls(pmu->cpuhp_state, &pmu->node); node 621 drivers/perf/fsl_imx8_ddr_perf.c cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node); node 634 drivers/perf/fsl_imx8_ddr_perf.c cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node); node 374 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c &ddrc_pmu->node); node 401 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c &ddrc_pmu->node); node 413 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c &ddrc_pmu->node); node 385 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c &hha_pmu->node); node 412 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c &hha_pmu->node); node 424 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c &hha_pmu->node); node 375 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c &l3c_pmu->node); node 402 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c &l3c_pmu->node); node 414 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c &l3c_pmu->node); node 383 drivers/perf/hisilicon/hisi_uncore_pmu.c int hisi_uncore_pmu_online_cpu(unsigned int cpu, struct hlist_node *node) node 385 drivers/perf/hisilicon/hisi_uncore_pmu.c struct hisi_pmu *hisi_pmu = hlist_entry_safe(node, struct hisi_pmu, node 386 drivers/perf/hisilicon/hisi_uncore_pmu.c node); node 406 drivers/perf/hisilicon/hisi_uncore_pmu.c int hisi_uncore_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 408 drivers/perf/hisilicon/hisi_uncore_pmu.c struct hisi_pmu *hisi_pmu = hlist_entry_safe(node, struct hisi_pmu, node 409 drivers/perf/hisilicon/hisi_uncore_pmu.c node); node 67 drivers/perf/hisilicon/hisi_uncore_pmu.h struct hlist_node node; node 97 drivers/perf/hisilicon/hisi_uncore_pmu.h int hisi_uncore_pmu_online_cpu(unsigned int cpu, struct hlist_node *node); node 98 drivers/perf/hisilicon/hisi_uncore_pmu.h int hisi_uncore_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node); node 152 drivers/perf/qcom_l2_pmu.c struct hlist_node node; node 812 drivers/perf/qcom_l2_pmu.c static int l2cache_pmu_online_cpu(unsigned int cpu, struct hlist_node *node) node 817 drivers/perf/qcom_l2_pmu.c l2cache_pmu = hlist_entry_safe(node, struct l2cache_pmu, node); node 847 drivers/perf/qcom_l2_pmu.c static int l2cache_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 854 drivers/perf/qcom_l2_pmu.c l2cache_pmu = hlist_entry_safe(node, struct l2cache_pmu, node); node 993 drivers/perf/qcom_l2_pmu.c &l2cache_pmu->node); node 1012 drivers/perf/qcom_l2_pmu.c &l2cache_pmu->node); node 1023 drivers/perf/qcom_l2_pmu.c &l2cache_pmu->node); node 156 drivers/perf/qcom_l3_pmu.c struct hlist_node node; node 706 drivers/perf/qcom_l3_pmu.c static int qcom_l3_cache_pmu_online_cpu(unsigned int cpu, struct hlist_node *node) node 708 drivers/perf/qcom_l3_pmu.c struct l3cache_pmu *l3pmu = hlist_entry_safe(node, struct l3cache_pmu, node); node 717 drivers/perf/qcom_l3_pmu.c static int qcom_l3_cache_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 719 drivers/perf/qcom_l3_pmu.c struct l3cache_pmu *l3pmu = hlist_entry_safe(node, struct l3cache_pmu, node); node 790 drivers/perf/qcom_l3_pmu.c ret = cpuhp_state_add_instance(CPUHP_AP_PERF_ARM_QCOM_L3_ONLINE, &l3pmu->node); node 69 drivers/perf/thunderx2_pmu.c int node; node 579 drivers/perf/thunderx2_pmu.c cpu = cpumask_any_and(cpumask_of_node(tx2_pmu->node), node 627 drivers/perf/thunderx2_pmu.c list_for_each_entry(rentry, &list, node) { node 651 drivers/perf/thunderx2_pmu.c tx2_pmu->node = dev_to_node(dev); node 662 drivers/perf/thunderx2_pmu.c "uncore_l3c_%d", tx2_pmu->node); node 674 drivers/perf/thunderx2_pmu.c "uncore_dmc_%d", tx2_pmu->node); node 728 drivers/perf/thunderx2_pmu.c (tx2_pmu->node == cpu_to_node(cpu))) node 751 drivers/perf/thunderx2_pmu.c cpumask_of_node(tx2_pmu->node), node 803 drivers/perf/thunderx2_pmu.c if (tx2_pmu->node == dev_to_node(dev)) { node 122 drivers/perf/xgene_pmu.c struct hlist_node node; node 1794 drivers/perf/xgene_pmu.c static int xgene_pmu_online_cpu(unsigned int cpu, struct hlist_node *node) node 1796 drivers/perf/xgene_pmu.c struct xgene_pmu *xgene_pmu = hlist_entry_safe(node, struct xgene_pmu, node 1797 drivers/perf/xgene_pmu.c node); node 1808 drivers/perf/xgene_pmu.c static int xgene_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) node 1810 drivers/perf/xgene_pmu.c struct xgene_pmu *xgene_pmu = hlist_entry_safe(node, struct xgene_pmu, node 1811 drivers/perf/xgene_pmu.c node); node 1929 drivers/perf/xgene_pmu.c &xgene_pmu->node); node 1949 drivers/perf/xgene_pmu.c &xgene_pmu->node); node 1972 drivers/perf/xgene_pmu.c &xgene_pmu->node); node 126 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c struct device_node *node = dev->of_node, *child; node 133 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c if (of_get_child_count(node) == 0) { node 151 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c for_each_available_child_of_node(node, child) { node 218 drivers/phy/broadcom/phy-bcm-sr-pcie.c struct device_node *node = dev->of_node; node 235 drivers/phy/broadcom/phy-bcm-sr-pcie.c core->cdru = syscon_regmap_lookup_by_phandle(node, "brcm,sr-cdru"); node 241 drivers/phy/broadcom/phy-bcm-sr-pcie.c core->mhb = syscon_regmap_lookup_by_phandle(node, "brcm,sr-mhb"); node 286 drivers/phy/broadcom/phy-bcm-sr-usb.c static int bcm_usb_phy_create(struct device *dev, struct device_node *node, node 309 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].phy = devm_phy_create(dev, node, node 326 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg->phy = devm_phy_create(dev, node, &sr_phy_ops); node 169 drivers/phy/cadence/phy-cadence-sierra.c int i, ret, node = 0; node 232 drivers/phy/cadence/phy-cadence-sierra.c sp->phys[node].lnk_rst = node 235 drivers/phy/cadence/phy-cadence-sierra.c if (IS_ERR(sp->phys[node].lnk_rst)) { node 238 drivers/phy/cadence/phy-cadence-sierra.c ret = PTR_ERR(sp->phys[node].lnk_rst); node 243 drivers/phy/cadence/phy-cadence-sierra.c ret = cdns_sierra_get_optional(&sp->phys[node], child); node 257 drivers/phy/cadence/phy-cadence-sierra.c sp->phys[node].phy = gphy; node 258 drivers/phy/cadence/phy-cadence-sierra.c phy_set_drvdata(gphy, &sp->phys[node]); node 264 drivers/phy/cadence/phy-cadence-sierra.c node++; node 266 drivers/phy/cadence/phy-cadence-sierra.c sp->nsubnodes = node; node 278 drivers/phy/cadence/phy-cadence-sierra.c node++; node 280 drivers/phy/cadence/phy-cadence-sierra.c for (i = 0; i < node; i++) node 80 drivers/phy/phy-core.c list_add_tail(&pl->node, &phys); node 104 drivers/phy/phy-core.c list_for_each_entry(pl, &phys, node) node 107 drivers/phy/phy-core.c list_del(&pl->node); node 121 drivers/phy/phy-core.c list_for_each_entry(p, &phys, node) node 131 drivers/phy/phy-core.c static struct phy_provider *of_phy_provider_lookup(struct device_node *node) node 137 drivers/phy/phy-core.c if (phy_provider->dev->of_node == node) node 141 drivers/phy/phy-core.c if (child == node) node 834 drivers/phy/phy-core.c struct phy *phy_create(struct device *dev, struct device_node *node, node 860 drivers/phy/phy-core.c phy->dev.of_node = node ?: dev->of_node; node 910 drivers/phy/phy-core.c struct phy *devm_phy_create(struct device *dev, struct device_node *node, node 919 drivers/phy/phy-core.c phy = phy_create(dev, node, ops); node 321 drivers/phy/rockchip/phy-rockchip-inno-usb2.c struct device_node *node = rphy->dev->of_node; node 331 drivers/phy/rockchip/phy-rockchip-inno-usb2.c of_property_read_string(node, "clock-output-names", &init.name); node 351 drivers/phy/rockchip/phy-rockchip-inno-usb2.c ret = of_clk_add_provider(node, of_clk_src_simple_get, rphy->clk480m); node 363 drivers/phy/rockchip/phy-rockchip-inno-usb2.c of_clk_del_provider(node); node 373 drivers/phy/rockchip/phy-rockchip-inno-usb2.c struct device_node *node = rphy->dev->of_node; node 376 drivers/phy/rockchip/phy-rockchip-inno-usb2.c if (of_property_read_bool(node, "extcon")) { node 842 drivers/phy/samsung/phy-exynos5-usbdrd.c struct device_node *node = dev->of_node; node 888 drivers/phy/samsung/phy-exynos5-usbdrd.c channel = of_alias_get_id(node, "usbdrdphy"); node 167 drivers/phy/samsung/phy-exynos5250-sata.c struct device_node *node; node 187 drivers/phy/samsung/phy-exynos5250-sata.c node = of_parse_phandle(dev->of_node, node 189 drivers/phy/samsung/phy-exynos5250-sata.c if (!node) node 192 drivers/phy/samsung/phy-exynos5250-sata.c sata_phy->client = of_find_i2c_device_by_node(node); node 1136 drivers/phy/st/phy-miphy28lp.c static int miphy28lp_probe_resets(struct device_node *node, node 1143 drivers/phy/st/phy-miphy28lp.c of_reset_control_get_shared(node, "miphy-sw-rst"); node 397 drivers/phy/ti/phy-am654-serdes.c struct device_node *node = am654_phy->of_node; node 416 drivers/phy/ti/phy-am654-serdes.c regmap_node = of_parse_phandle(node, "ti,serdes-clk", 0); node 430 drivers/phy/ti/phy-am654-serdes.c num_parents = of_clk_get_parent_count(node); node 444 drivers/phy/ti/phy-am654-serdes.c of_clk_parent_fill(node, parent_names, num_parents); node 559 drivers/phy/ti/phy-am654-serdes.c struct device_node *node = dev->of_node; node 589 drivers/phy/ti/phy-am654-serdes.c am654_phy->of_node = node; node 603 drivers/phy/ti/phy-am654-serdes.c ret = of_property_read_string_index(node, "clock-output-names", node 621 drivers/phy/ti/phy-am654-serdes.c ret = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data); node 641 drivers/phy/ti/phy-am654-serdes.c of_clk_del_provider(node); node 649 drivers/phy/ti/phy-am654-serdes.c struct device_node *node = am654_phy->of_node; node 652 drivers/phy/ti/phy-am654-serdes.c of_clk_del_provider(node); node 143 drivers/phy/ti/phy-da8xx-usb.c struct device_node *node = dev->of_node; node 172 drivers/phy/ti/phy-da8xx-usb.c d_phy->usb11_phy = devm_phy_create(dev, node, &da8xx_usb11_phy_ops); node 178 drivers/phy/ti/phy-da8xx-usb.c d_phy->usb20_phy = devm_phy_create(dev, node, &da8xx_usb20_phy_ops); node 188 drivers/phy/ti/phy-da8xx-usb.c if (node) { node 297 drivers/phy/ti/phy-gmii-sel.c struct device_node *node = dev->of_node; node 313 drivers/phy/ti/phy-gmii-sel.c priv->regmap = syscon_node_to_regmap(node->parent); node 287 drivers/phy/ti/phy-omap-usb2.c struct device_node *node = pdev->dev.of_node; node 326 drivers/phy/ti/phy-omap-usb2.c phy->syscon_phy_power = syscon_regmap_lookup_by_phandle(node, node 333 drivers/phy/ti/phy-omap-usb2.c control_node = of_parse_phandle(node, "ctrl-module", 0); node 347 drivers/phy/ti/phy-omap-usb2.c if (of_property_read_u32_index(node, node 673 drivers/phy/ti/phy-ti-pipe3.c struct device_node *node = dev->of_node; node 677 drivers/phy/ti/phy-ti-pipe3.c phy->phy_power_syscon = syscon_regmap_lookup_by_phandle(node, node 684 drivers/phy/ti/phy-ti-pipe3.c if (of_property_read_u32_index(node, node 693 drivers/phy/ti/phy-ti-pipe3.c control_node = of_parse_phandle(node, "ctrl-module", 0); node 709 drivers/phy/ti/phy-ti-pipe3.c phy->pcs_syscon = syscon_regmap_lookup_by_phandle(node, node 716 drivers/phy/ti/phy-ti-pipe3.c if (of_property_read_u32_index(node, node 727 drivers/phy/ti/phy-ti-pipe3.c phy->dpll_reset_syscon = syscon_regmap_lookup_by_phandle(node, node 734 drivers/phy/ti/phy-ti-pipe3.c if (of_property_read_u32_index(node, node 2632 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c struct device_node *node; node 2635 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c node = of_parse_phandle(ctx->dev->of_node, node 2637 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c if (node) { node 2638 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c map = syscon_node_to_regmap(node); node 2639 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c of_node_put(node); node 2651 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c struct device_node *node; node 2654 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c node = of_parse_phandle(ctx->dev->of_node, node 2656 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c if (node) { node 2657 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c map = syscon_node_to_regmap(node->parent); node 2658 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c of_node_put(node); node 51 drivers/pinctrl/berlin/berlin.c struct device_node *node, node 64 drivers/pinctrl/berlin/berlin.c ret = of_property_read_string(node, "function", &function_name); node 67 drivers/pinctrl/berlin/berlin.c "missing function property in node %pOFn\n", node); node 71 drivers/pinctrl/berlin/berlin.c ngroups = of_property_count_strings(node, "groups"); node 74 drivers/pinctrl/berlin/berlin.c "missing groups property in node %pOFn\n", node); node 85 drivers/pinctrl/berlin/berlin.c of_property_for_each_string(node, "groups", prop, group_name) { node 108 drivers/pinctrl/core.c list_for_each_entry(pctldev, &pinctrldev_list, node) { node 127 drivers/pinctrl/core.c list_for_each_entry(pctldev, &pinctrldev_list, node) node 295 drivers/pinctrl/core.c list_for_each_entry(range, &pctldev->gpio_ranges, node) { node 333 drivers/pinctrl/core.c list_for_each_entry(pctldev, &pinctrldev_list, node) { node 336 drivers/pinctrl/core.c list_for_each_entry(range, &pctldev->gpio_ranges, node) { node 376 drivers/pinctrl/core.c list_for_each_entry(pctldev, &pinctrldev_list, node) { node 405 drivers/pinctrl/core.c list_add_tail(&range->node, &pctldev->gpio_ranges); node 466 drivers/pinctrl/core.c list_for_each_entry(range, &pctldev->gpio_ranges, node) { node 511 drivers/pinctrl/core.c list_del(&range->node); node 917 drivers/pinctrl/core.c list_for_each_entry(state, &p->states, node) node 936 drivers/pinctrl/core.c list_add_tail(&state->node, &p->states); node 1001 drivers/pinctrl/core.c list_add_tail(&setting->node, &state->settings); node 1011 drivers/pinctrl/core.c list_for_each_entry(p, &pinctrl_list, node) node 1102 drivers/pinctrl/core.c list_add_tail(&p->node, &pinctrl_list); node 1159 drivers/pinctrl/core.c list_for_each_entry_safe(state, n1, &p->states, node) { node 1160 drivers/pinctrl/core.c list_for_each_entry_safe(setting, n2, &state->settings, node) { node 1162 drivers/pinctrl/core.c list_del(&setting->node); node 1165 drivers/pinctrl/core.c list_del(&state->node); node 1172 drivers/pinctrl/core.c list_del(&p->node); node 1250 drivers/pinctrl/core.c list_for_each_entry(setting, &p->state->settings, node) { node 1260 drivers/pinctrl/core.c list_for_each_entry(setting, &state->settings, node) { node 1290 drivers/pinctrl/core.c list_for_each_entry(setting2, &state->settings, node) { node 1291 drivers/pinctrl/core.c if (&setting2->node == &setting->node) node 1446 drivers/pinctrl/core.c list_add_tail(&maps_node->node, &pinctrl_maps); node 1471 drivers/pinctrl/core.c list_for_each_entry(maps_node, &pinctrl_maps, node) { node 1473 drivers/pinctrl/core.c list_del(&maps_node->node); node 1693 drivers/pinctrl/core.c list_for_each_entry(range, &pctldev->gpio_ranges, node) { node 1725 drivers/pinctrl/core.c list_for_each_entry(pctldev, &pinctrldev_list, node) { node 1808 drivers/pinctrl/core.c list_for_each_entry(p, &pinctrl_list, node) { node 1813 drivers/pinctrl/core.c list_for_each_entry(state, &p->states, node) { node 1816 drivers/pinctrl/core.c list_for_each_entry(setting, &state->settings, node) { node 1969 drivers/pinctrl/core.c INIT_LIST_HEAD(&pctldev->node); node 2065 drivers/pinctrl/core.c list_add_tail(&pctldev->node, &pinctrldev_list); node 2159 drivers/pinctrl/core.c list_del(&pctldev->node); node 2166 drivers/pinctrl/core.c list_for_each_entry_safe(range, n, &pctldev->gpio_ranges, node) node 2167 drivers/pinctrl/core.c list_del(&range->node); node 43 drivers/pinctrl/core.h struct list_head node; node 78 drivers/pinctrl/core.h struct list_head node; node 93 drivers/pinctrl/core.h struct list_head node; node 132 drivers/pinctrl/core.h struct list_head node; node 180 drivers/pinctrl/core.h struct list_head node; node 250 drivers/pinctrl/core.h list_for_each_entry(_maps_node_, &pinctrl_maps, node) \ node 23 drivers/pinctrl/devicetree.c struct list_head node; node 53 drivers/pinctrl/devicetree.c list_for_each_entry_safe(dt_map, n1, &p->dt_maps, node) { node 55 drivers/pinctrl/devicetree.c list_del(&dt_map->node); node 93 drivers/pinctrl/devicetree.c list_add_tail(&dt_map->node, &p->dt_maps); node 491 drivers/pinctrl/mediatek/pinctrl-mtk-common.c struct device_node *node, node 507 drivers/pinctrl/mediatek/pinctrl-mtk-common.c pins = of_find_property(node, "pinmux", NULL); node 510 drivers/pinctrl/mediatek/pinctrl-mtk-common.c node); node 514 drivers/pinctrl/mediatek/pinctrl-mtk-common.c err = pinconf_generic_parse_dt_config(node, pctldev, &configs, node 543 drivers/pinctrl/mediatek/pinctrl-mtk-common.c err = of_property_read_u32_index(node, "pinmux", node 1022 drivers/pinctrl/mediatek/pinctrl-mtk-common.c struct device_node *np = pdev->dev.of_node, *node; node 1038 drivers/pinctrl/mediatek/pinctrl-mtk-common.c node = of_parse_phandle(np, "mediatek,pctl-regmap", 0); node 1039 drivers/pinctrl/mediatek/pinctrl-mtk-common.c if (node) { node 1040 drivers/pinctrl/mediatek/pinctrl-mtk-common.c pctl->regmap1 = syscon_node_to_regmap(node); node 1051 drivers/pinctrl/mediatek/pinctrl-mtk-common.c node = of_parse_phandle(np, "mediatek,pctl-regmap", 1); node 1052 drivers/pinctrl/mediatek/pinctrl-mtk-common.c if (node) { node 1053 drivers/pinctrl/mediatek/pinctrl-mtk-common.c pctl->regmap2 = syscon_node_to_regmap(node); node 426 drivers/pinctrl/mediatek/pinctrl-paris.c struct device_node *node, node 441 drivers/pinctrl/mediatek/pinctrl-paris.c pins = of_find_property(node, "pinmux", NULL); node 444 drivers/pinctrl/mediatek/pinctrl-paris.c node); node 448 drivers/pinctrl/mediatek/pinctrl-paris.c err = pinconf_generic_parse_dt_config(node, pctldev, &configs, node 477 drivers/pinctrl/mediatek/pinctrl-paris.c err = of_property_read_u32_index(node, "pinmux", i, &pinfunc); node 621 drivers/pinctrl/meson/pinctrl-meson.c struct device_node *node, char *name) node 627 drivers/pinctrl/meson/pinctrl-meson.c i = of_property_match_string(node, "reg-names", name); node 628 drivers/pinctrl/meson/pinctrl-meson.c if (of_address_to_resource(node, i, &res)) node 637 drivers/pinctrl/meson/pinctrl-meson.c "%pOFn-%s", node, node 646 drivers/pinctrl/meson/pinctrl-meson.c struct device_node *node) node 650 drivers/pinctrl/meson/pinctrl-meson.c for_each_child_of_node(node, np) { node 131 drivers/pinctrl/pinctrl-at91-pio4.c struct device_node *node; node 470 drivers/pinctrl/pinctrl-at91-pio4.c if (np->parent == atmel_pioctrl->node) node 946 drivers/pinctrl/pinctrl-at91-pio4.c atmel_pioctrl->node = dev->of_node; node 1939 drivers/pinctrl/pinctrl-ingenic.c struct device_node *node) node 1946 drivers/pinctrl/pinctrl-ingenic.c err = of_property_read_u32(node, "reg", &bank); node 1971 drivers/pinctrl/pinctrl-ingenic.c jzgc->gc.of_node = node; node 1980 drivers/pinctrl/pinctrl-ingenic.c if (of_property_read_bool(node, "gpio-ranges")) { node 1989 drivers/pinctrl/pinctrl-ingenic.c jzgc->irq = irq_of_parse_and_map(node, 0); node 2023 drivers/pinctrl/pinctrl-ingenic.c struct device_node *node; node 2126 drivers/pinctrl/pinctrl-ingenic.c for_each_child_of_node(dev->of_node, node) { node 2127 drivers/pinctrl/pinctrl-ingenic.c if (of_match_node(ingenic_gpio_of_match, node)) { node 2128 drivers/pinctrl/pinctrl-ingenic.c err = ingenic_gpio_probe(jzpc, node); node 1347 drivers/pinctrl/pinctrl-pistachio.c struct device_node *node = pctl->dev->of_node; node 1357 drivers/pinctrl/pinctrl-pistachio.c child = of_get_child_by_name(node, child_name); node 3187 drivers/pinctrl/pinctrl-rockchip.c struct device_node *node; node 3189 drivers/pinctrl/pinctrl-rockchip.c node = of_parse_phandle(bank->of_node->parent, node 3191 drivers/pinctrl/pinctrl-rockchip.c if (!node) { node 3208 drivers/pinctrl/pinctrl-rockchip.c of_node_put(node); node 3228 drivers/pinctrl/pinctrl-rockchip.c struct device_node *node = pdev->dev.of_node; node 3234 drivers/pinctrl/pinctrl-rockchip.c match = of_match_node(rockchip_pinctrl_dt_match, node); node 3237 drivers/pinctrl/pinctrl-rockchip.c for_each_child_of_node(node, np) { node 3404 drivers/pinctrl/pinctrl-rockchip.c struct device_node *np = pdev->dev.of_node, *node; node 3427 drivers/pinctrl/pinctrl-rockchip.c node = of_parse_phandle(np, "rockchip,grf", 0); node 3428 drivers/pinctrl/pinctrl-rockchip.c if (node) { node 3429 drivers/pinctrl/pinctrl-rockchip.c info->regmap_base = syscon_node_to_regmap(node); node 3463 drivers/pinctrl/pinctrl-rockchip.c node = of_parse_phandle(np, "rockchip,pmu", 0); node 3464 drivers/pinctrl/pinctrl-rockchip.c if (node) { node 3465 drivers/pinctrl/pinctrl-rockchip.c info->regmap_pmu = syscon_node_to_regmap(node); node 96 drivers/pinctrl/pinctrl-single.c struct list_head node; node 110 drivers/pinctrl/pinctrl-single.c struct list_head node; node 392 drivers/pinctrl/pinctrl-single.c frange = list_entry(pos, struct pcs_gpiofunc_range, node); node 1316 drivers/pinctrl/pinctrl-single.c static int pcs_add_gpio_func(struct device_node *node, struct pcs_device *pcs) node 1325 drivers/pinctrl/pinctrl-single.c ret = of_parse_phandle_with_args(node, propname, cellname, node 1341 drivers/pinctrl/pinctrl-single.c list_add_tail(&range->node, &pcs->gpiofuncs); node 1356 drivers/pinctrl/pinctrl-single.c struct list_head node; node 1377 drivers/pinctrl/pinctrl-single.c pcswi = list_entry(pos, struct pcs_interrupt, node); node 1458 drivers/pinctrl/pinctrl-single.c pcswi = list_entry(pos, struct pcs_interrupt, node); node 1524 drivers/pinctrl/pinctrl-single.c list_add_tail(&pcswi->node, &pcs->irqs); node 986 drivers/pinctrl/samsung/pinctrl-samsung.c struct device_node *node = pdev->dev.of_node; node 990 drivers/pinctrl/samsung/pinctrl-samsung.c id = of_alias_get_id(node, "pinctrl"); node 1010 drivers/pinctrl/samsung/pinctrl-samsung.c struct device_node *node = pdev->dev.of_node; node 1075 drivers/pinctrl/samsung/pinctrl-samsung.c for_each_child_of_node(node, np) { node 271 drivers/pinctrl/samsung/pinctrl-samsung.h struct list_head node; node 506 drivers/pinctrl/stm32/pinctrl-stm32.c struct device_node *node, node 523 drivers/pinctrl/stm32/pinctrl-stm32.c pins = of_find_property(node, "pinmux", NULL); node 526 drivers/pinctrl/stm32/pinctrl-stm32.c node); node 530 drivers/pinctrl/stm32/pinctrl-stm32.c err = pinconf_generic_parse_dt_config(node, pctldev, &configs, node 559 drivers/pinctrl/stm32/pinctrl-stm32.c err = of_property_read_u32_index(node, "pinmux", node 154 drivers/pinctrl/sunxi/pinctrl-sunxi.c static bool sunxi_pctrl_has_bias_prop(struct device_node *node) node 156 drivers/pinctrl/sunxi/pinctrl-sunxi.c return of_find_property(node, "bias-pull-up", NULL) || node 157 drivers/pinctrl/sunxi/pinctrl-sunxi.c of_find_property(node, "bias-pull-down", NULL) || node 158 drivers/pinctrl/sunxi/pinctrl-sunxi.c of_find_property(node, "bias-disable", NULL) || node 159 drivers/pinctrl/sunxi/pinctrl-sunxi.c of_find_property(node, "allwinner,pull", NULL); node 162 drivers/pinctrl/sunxi/pinctrl-sunxi.c static bool sunxi_pctrl_has_drive_prop(struct device_node *node) node 164 drivers/pinctrl/sunxi/pinctrl-sunxi.c return of_find_property(node, "drive-strength", NULL) || node 165 drivers/pinctrl/sunxi/pinctrl-sunxi.c of_find_property(node, "allwinner,drive", NULL); node 168 drivers/pinctrl/sunxi/pinctrl-sunxi.c static int sunxi_pctrl_parse_bias_prop(struct device_node *node) node 173 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_find_property(node, "bias-pull-up", NULL)) node 176 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_find_property(node, "bias-pull-down", NULL)) node 179 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_find_property(node, "bias-disable", NULL)) node 183 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_property_read_u32(node, "allwinner,pull", &val)) node 198 drivers/pinctrl/sunxi/pinctrl-sunxi.c static int sunxi_pctrl_parse_drive_prop(struct device_node *node) node 203 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (!of_property_read_u32(node, "drive-strength", &val)) { node 217 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_property_read_u32(node, "allwinner,drive", &val)) node 223 drivers/pinctrl/sunxi/pinctrl-sunxi.c static const char *sunxi_pctrl_parse_function_prop(struct device_node *node) node 229 drivers/pinctrl/sunxi/pinctrl-sunxi.c ret = of_property_read_string(node, "function", &function); node 234 drivers/pinctrl/sunxi/pinctrl-sunxi.c ret = of_property_read_string(node, "allwinner,function", &function); node 241 drivers/pinctrl/sunxi/pinctrl-sunxi.c static const char *sunxi_pctrl_find_pins_prop(struct device_node *node, node 247 drivers/pinctrl/sunxi/pinctrl-sunxi.c count = of_property_count_strings(node, "pins"); node 254 drivers/pinctrl/sunxi/pinctrl-sunxi.c count = of_property_count_strings(node, "allwinner,pins"); node 263 drivers/pinctrl/sunxi/pinctrl-sunxi.c static unsigned long *sunxi_pctrl_build_pin_config(struct device_node *node, node 270 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (sunxi_pctrl_has_drive_prop(node)) node 272 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (sunxi_pctrl_has_bias_prop(node)) node 285 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (sunxi_pctrl_has_drive_prop(node)) { node 286 drivers/pinctrl/sunxi/pinctrl-sunxi.c int drive = sunxi_pctrl_parse_drive_prop(node); node 296 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (sunxi_pctrl_has_bias_prop(node)) { node 297 drivers/pinctrl/sunxi/pinctrl-sunxi.c int pull = sunxi_pctrl_parse_bias_prop(node); node 320 drivers/pinctrl/sunxi/pinctrl-sunxi.c struct device_node *node, node 334 drivers/pinctrl/sunxi/pinctrl-sunxi.c function = sunxi_pctrl_parse_function_prop(node); node 337 drivers/pinctrl/sunxi/pinctrl-sunxi.c node); node 341 drivers/pinctrl/sunxi/pinctrl-sunxi.c pin_prop = sunxi_pctrl_find_pins_prop(node, &npins); node 344 drivers/pinctrl/sunxi/pinctrl-sunxi.c node); node 360 drivers/pinctrl/sunxi/pinctrl-sunxi.c pinconfig = sunxi_pctrl_build_pin_config(node, &configlen); node 366 drivers/pinctrl/sunxi/pinctrl-sunxi.c of_property_for_each_string(node, pin_prop, prop, group) { node 1089 drivers/pinctrl/sunxi/pinctrl-sunxi.c struct device_node *node, node 1318 drivers/pinctrl/sunxi/pinctrl-sunxi.c struct device_node *node) node 1327 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (of_clk_get_parent_count(node) != 3) node 1331 drivers/pinctrl/sunxi/pinctrl-sunxi.c if (!of_find_property(node, "input-debounce", NULL)) node 1346 drivers/pinctrl/sunxi/pinctrl-sunxi.c ret = of_property_read_u32_index(node, "input-debounce", node 1383 drivers/pinctrl/sunxi/pinctrl-sunxi.c struct device_node *node = pdev->dev.of_node; node 1502 drivers/pinctrl/sunxi/pinctrl-sunxi.c ret = of_clk_get_parent_count(node); node 1530 drivers/pinctrl/sunxi/pinctrl-sunxi.c pctl->domain = irq_domain_add_linear(node, node 1561 drivers/pinctrl/sunxi/pinctrl-sunxi.c sunxi_pinctrl_setup_debounce(pctl, node); node 29 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c struct list_head node; node 705 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c list_for_each_entry(r, &priv->reg_regions, node) { node 721 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c list_for_each_entry(r, &priv->reg_regions, node) { node 760 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c list_add_tail(®ion->node, &priv->reg_regions); node 49 drivers/platform/chrome/cros_ec_chardev.c struct list_head node; node 117 drivers/platform/chrome/cros_ec_chardev.c list_add_tail(&event->node, &priv->events); node 149 drivers/platform/chrome/cros_ec_chardev.c event = list_first_entry(&priv->events, struct ec_event, node); node 150 drivers/platform/chrome/cros_ec_chardev.c list_del(&event->node); node 263 drivers/platform/chrome/cros_ec_chardev.c list_for_each_entry_safe(event, e, &priv->events, node) { node 264 drivers/platform/chrome/cros_ec_chardev.c list_del(&event->node); node 28 drivers/platform/olpc/olpc-ec.c struct list_head node; node 87 drivers/platform/olpc/olpc-ec.c desc = list_first_entry(&ec->cmd_q, struct ec_cmd_desc, node); node 88 drivers/platform/olpc/olpc-ec.c list_del(&desc->node); node 118 drivers/platform/olpc/olpc-ec.c INIT_LIST_HEAD(&desc->node); node 121 drivers/platform/olpc/olpc-ec.c list_add_tail(&desc->node, &ec->cmd_q); node 853 drivers/platform/x86/asus-wmi.c static int asus_register_rfkill_notifier(struct asus_wmi *asus, char *node) node 858 drivers/platform/x86/asus-wmi.c status = acpi_get_handle(NULL, node, &handle); node 865 drivers/platform/x86/asus-wmi.c pr_warn("Failed to register notify on %s\n", node); node 870 drivers/platform/x86/asus-wmi.c static void asus_unregister_rfkill_notifier(struct asus_wmi *asus, char *node) node 875 drivers/platform/x86/asus-wmi.c status = acpi_get_handle(NULL, node, &handle); node 882 drivers/platform/x86/asus-wmi.c pr_err("Error removing rfkill notify handler %s\n", node); node 2362 drivers/platform/x86/asus-wmi.c struct asus_wmi_debugfs_node *node = inode->i_private; node 2364 drivers/platform/x86/asus-wmi.c return single_open(file, node->show, node->asus); node 2396 drivers/platform/x86/asus-wmi.c struct asus_wmi_debugfs_node *node = &asus_wmi_debug_files[i]; node 2398 drivers/platform/x86/asus-wmi.c node->asus = asus; node 2399 drivers/platform/x86/asus-wmi.c debugfs_create_file(node->name, S_IFREG | S_IRUGO, node 2400 drivers/platform/x86/asus-wmi.c asus->debug.root, node, node 634 drivers/platform/x86/eeepc-laptop.c static void eeepc_rfkill_hotplug_update(struct eeepc_laptop *eeepc, char *node) node 639 drivers/platform/x86/eeepc-laptop.c status = acpi_get_handle(NULL, node, &handle); node 656 drivers/platform/x86/eeepc-laptop.c char *node) node 661 drivers/platform/x86/eeepc-laptop.c status = acpi_get_handle(NULL, node, &handle); node 671 drivers/platform/x86/eeepc-laptop.c pr_warn("Failed to register notify on %s\n", node); node 682 drivers/platform/x86/eeepc-laptop.c char *node) node 687 drivers/platform/x86/eeepc-laptop.c status = acpi_get_handle(NULL, node, &handle); node 697 drivers/platform/x86/eeepc-laptop.c node); node 176 drivers/platform/x86/intel_cht_int33fe.c if (mux_ref.node) { node 177 drivers/platform/x86/intel_cht_int33fe.c fwnode_handle_put(software_node_fwnode(mux_ref.node)); node 178 drivers/platform/x86/intel_cht_int33fe.c mux_ref.node = NULL; node 204 drivers/platform/x86/intel_cht_int33fe.c mux_ref.node = software_node_find_by_name(NULL, "intel-xhci-usb-sw"); node 205 drivers/platform/x86/intel_cht_int33fe.c if (!mux_ref.node) { node 228 drivers/platform/x86/intel_menlow.c struct list_head node; node 395 drivers/platform/x86/intel_menlow.c list_add_tail(&attr->node, &intel_menlow_attr_list); node 472 drivers/platform/x86/intel_menlow.c list_for_each_entry_safe(pos, next, &intel_menlow_attr_list, node) { node 473 drivers/platform/x86/intel_menlow.c list_del(&pos->node); node 6863 drivers/platform/x86/thinkpad_acpi.c list_for_each_entry(child, &device->children, node) { node 192 drivers/pnp/pnpbios/core.c struct pnp_bios_node *node; node 198 drivers/pnp/pnpbios/core.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 199 drivers/pnp/pnpbios/core.c if (!node) node 201 drivers/pnp/pnpbios/core.c if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) { node 202 drivers/pnp/pnpbios/core.c kfree(node); node 205 drivers/pnp/pnpbios/core.c pnpbios_read_resources_from_node(dev, node); node 207 drivers/pnp/pnpbios/core.c kfree(node); node 214 drivers/pnp/pnpbios/core.c struct pnp_bios_node *node; node 221 drivers/pnp/pnpbios/core.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 222 drivers/pnp/pnpbios/core.c if (!node) node 224 drivers/pnp/pnpbios/core.c if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) { node 225 drivers/pnp/pnpbios/core.c kfree(node); node 228 drivers/pnp/pnpbios/core.c if (pnpbios_write_resources_to_node(dev, node) < 0) { node 229 drivers/pnp/pnpbios/core.c kfree(node); node 232 drivers/pnp/pnpbios/core.c ret = pnp_bios_set_dev_node(node->handle, (char)PNPMODE_DYNAMIC, node); node 233 drivers/pnp/pnpbios/core.c kfree(node); node 239 drivers/pnp/pnpbios/core.c static void pnpbios_zero_data_stream(struct pnp_bios_node *node) node 241 drivers/pnp/pnpbios/core.c unsigned char *p = (char *)node->data; node 242 drivers/pnp/pnpbios/core.c unsigned char *end = (char *)(node->data + node->size); node 266 drivers/pnp/pnpbios/core.c struct pnp_bios_node *node; node 273 drivers/pnp/pnpbios/core.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 274 drivers/pnp/pnpbios/core.c if (!node) node 277 drivers/pnp/pnpbios/core.c if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) { node 278 drivers/pnp/pnpbios/core.c kfree(node); node 281 drivers/pnp/pnpbios/core.c pnpbios_zero_data_stream(node); node 283 drivers/pnp/pnpbios/core.c ret = pnp_bios_set_dev_node(dev->number, (char)PNPMODE_DYNAMIC, node); node 284 drivers/pnp/pnpbios/core.c kfree(node); node 299 drivers/pnp/pnpbios/core.c static int __init insert_device(struct pnp_bios_node *node) node 309 drivers/pnp/pnpbios/core.c if (dev->number == node->handle) node 313 drivers/pnp/pnpbios/core.c pnp_eisa_id_to_string(node->eisa_id & PNP_EISA_ID_MASK, id); node 314 drivers/pnp/pnpbios/core.c dev = pnp_alloc_dev(&pnpbios_protocol, node->handle, id); node 318 drivers/pnp/pnpbios/core.c pnpbios_parse_data_stream(dev, node); node 320 drivers/pnp/pnpbios/core.c dev->flags = node->flags; node 341 drivers/pnp/pnpbios/core.c pnpbios_interface_attach_device(node); node 351 drivers/pnp/pnpbios/core.c struct pnp_bios_node *node; node 353 drivers/pnp/pnpbios/core.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 354 drivers/pnp/pnpbios/core.c if (!node) node 364 drivers/pnp/pnpbios/core.c (&nodenum, (char)PNPMODE_DYNAMIC, node)) node 368 drivers/pnp/pnpbios/core.c (&nodenum, (char)PNPMODE_STATIC, node)) node 372 drivers/pnp/pnpbios/core.c if (insert_device(node) == 0) node 382 drivers/pnp/pnpbios/core.c kfree(node); node 153 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_parse_data_stream(struct pnp_dev *dev, struct pnp_bios_node * node); node 154 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_read_resources_from_node(struct pnp_dev *dev, struct pnp_bios_node *node); node 155 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_write_resources_to_node(struct pnp_dev *dev, struct pnp_bios_node *node); node 162 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_interface_attach_device(struct pnp_bios_node * node); node 166 drivers/pnp/pnpbios/pnpbios.h static inline int pnpbios_interface_attach_device(struct pnp_bios_node * node) { return 0; } node 125 drivers/pnp/pnpbios/proc.c struct pnp_bios_node *node; node 128 drivers/pnp/pnpbios/proc.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 129 drivers/pnp/pnpbios/proc.c if (!node) node 135 drivers/pnp/pnpbios/proc.c if (pnp_bios_get_dev_node(&nodenum, PNPMODE_DYNAMIC, node)) node 138 drivers/pnp/pnpbios/proc.c node->handle, node->eisa_id, node 139 drivers/pnp/pnpbios/proc.c node->type_code, node->flags); node 149 drivers/pnp/pnpbios/proc.c kfree(node); node 156 drivers/pnp/pnpbios/proc.c struct pnp_bios_node *node; node 161 drivers/pnp/pnpbios/proc.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 162 drivers/pnp/pnpbios/proc.c if (!node) node 164 drivers/pnp/pnpbios/proc.c if (pnp_bios_get_dev_node(&nodenum, boot, node)) { node 165 drivers/pnp/pnpbios/proc.c kfree(node); node 168 drivers/pnp/pnpbios/proc.c len = node->size - sizeof(struct pnp_bios_node); node 169 drivers/pnp/pnpbios/proc.c seq_write(m, node->data, len); node 170 drivers/pnp/pnpbios/proc.c kfree(node); node 183 drivers/pnp/pnpbios/proc.c struct pnp_bios_node *node; node 188 drivers/pnp/pnpbios/proc.c node = kzalloc(node_info.max_node_size, GFP_KERNEL); node 189 drivers/pnp/pnpbios/proc.c if (!node) node 191 drivers/pnp/pnpbios/proc.c if (pnp_bios_get_dev_node(&nodenum, boot, node)) { node 195 drivers/pnp/pnpbios/proc.c if (count != node->size - sizeof(struct pnp_bios_node)) { node 199 drivers/pnp/pnpbios/proc.c if (copy_from_user(node->data, buf, count)) { node 203 drivers/pnp/pnpbios/proc.c if (pnp_bios_set_dev_node(node->handle, boot, node) != 0) { node 209 drivers/pnp/pnpbios/proc.c kfree(node); node 222 drivers/pnp/pnpbios/proc.c int pnpbios_interface_attach_device(struct pnp_bios_node *node) node 226 drivers/pnp/pnpbios/proc.c sprintf(name, "%02x", node->handle); node 232 drivers/pnp/pnpbios/proc.c (void *)(long)(node->handle)); node 238 drivers/pnp/pnpbios/proc.c (void *)(long)(node->handle + 0x100))) node 771 drivers/pnp/pnpbios/rsparser.c struct pnp_bios_node *node) node 773 drivers/pnp/pnpbios/rsparser.c unsigned char *p = (char *)node->data; node 774 drivers/pnp/pnpbios/rsparser.c unsigned char *end = (char *)(node->data + node->size); node 789 drivers/pnp/pnpbios/rsparser.c struct pnp_bios_node *node) node 791 drivers/pnp/pnpbios/rsparser.c unsigned char *p = (char *)node->data; node 792 drivers/pnp/pnpbios/rsparser.c unsigned char *end = (char *)(node->data + node->size); node 801 drivers/pnp/pnpbios/rsparser.c struct pnp_bios_node *node) node 803 drivers/pnp/pnpbios/rsparser.c unsigned char *p = (char *)node->data; node 804 drivers/pnp/pnpbios/rsparser.c unsigned char *end = (char *)(node->data + node->size); node 87 drivers/power/avs/smartreflex.c list_for_each_entry(sr_info, &sr_list, node) { node 228 drivers/power/avs/smartreflex.c list_del(&sr_info->node); node 678 drivers/power/avs/smartreflex.c list_for_each_entry(sr_info, &sr_list, node) node 881 drivers/power/avs/smartreflex.c list_add(&sr_info->node, &sr_list); node 945 drivers/power/avs/smartreflex.c list_del(&sr_info->node); node 974 drivers/power/avs/smartreflex.c list_del(&sr_info->node); node 94 drivers/power/supply/ab8500_btemp.c struct list_head node; node 126 drivers/power/supply/ab8500_btemp.c return list_first_entry(&ab8500_btemp_list, struct ab8500_btemp, node); node 1102 drivers/power/supply/ab8500_btemp.c list_add_tail(&di->node, &ab8500_btemp_list); node 200 drivers/power/supply/ab8500_fg.c struct list_head node; node 249 drivers/power/supply/ab8500_fg.c node); node 2993 drivers/power/supply/ab8500_fg.c list_del(&di->node); node 3210 drivers/power/supply/ab8500_fg.c list_add_tail(&di->node, &ab8500_fg_list); node 73 drivers/power/supply/cpcap-battery.c struct list_head node; node 552 drivers/power/supply/cpcap-battery.c list_for_each_entry(d, &ddata->irq_list, node) { node 619 drivers/power/supply/cpcap-battery.c list_add(&d->node, &ddata->irq_list); node 145 drivers/power/supply/cpcap-charger.c struct list_head node; node 517 drivers/power/supply/cpcap-charger.c list_add(&d->node, &ddata->irq_list); node 336 drivers/powercap/powercap_sys.c list_for_each_entry(pos, &powercap_cntrl_list, node) { node 621 drivers/powercap/powercap_sys.c INIT_LIST_HEAD(&control_type->node); node 633 drivers/powercap/powercap_sys.c list_add_tail(&control_type->node, &powercap_cntrl_list); node 649 drivers/powercap/powercap_sys.c list_for_each_entry(pos, &powercap_cntrl_list, node) { node 651 drivers/powercap/powercap_sys.c list_del(&control_type->node); node 396 drivers/ptp/ptp_qoriq.c struct device_node *node) node 407 drivers/ptp/ptp_qoriq.c clk = of_clk_get(node, 0); node 451 drivers/ptp/ptp_qoriq.c struct device_node *node = ptp_qoriq->dev->of_node; node 457 drivers/ptp/ptp_qoriq.c if (!node) node 463 drivers/ptp/ptp_qoriq.c if (of_property_read_u32(node, "fsl,cksel", &ptp_qoriq->cksel)) node 466 drivers/ptp/ptp_qoriq.c if (of_property_read_bool(node, "fsl,extts-fifo")) node 471 drivers/ptp/ptp_qoriq.c if (of_property_read_u32(node, node 473 drivers/ptp/ptp_qoriq.c of_property_read_u32(node, node 475 drivers/ptp/ptp_qoriq.c of_property_read_u32(node, node 477 drivers/ptp/ptp_qoriq.c of_property_read_u32(node, node 479 drivers/ptp/ptp_qoriq.c of_property_read_u32(node, node 481 drivers/ptp/ptp_qoriq.c of_property_read_u32(node, node 485 drivers/ptp/ptp_qoriq.c if (ptp_qoriq_auto_config(ptp_qoriq, node)) node 489 drivers/ptp/ptp_qoriq.c if (of_property_read_bool(node, "little-endian")) { node 498 drivers/ptp/ptp_qoriq.c if (of_device_is_compatible(node, "fsl,etsec-ptp")) { node 198 drivers/pwm/pwm-lp3943.c struct device_node *node = dev->of_node; node 205 drivers/pwm/pwm-lp3943.c if (!node) node 218 drivers/pwm/pwm-lp3943.c if (!of_get_property(node, name[i], &proplen)) node 230 drivers/pwm/pwm-lp3943.c err = of_property_read_u32_array(node, name[i], output, node 24 drivers/pwm/pwm-tipwmss.c struct device_node *node = pdev->dev.of_node; node 29 drivers/pwm/pwm-tipwmss.c ret = of_platform_populate(node, NULL, NULL, &pdev->dev); node 122 drivers/rapidio/devices/rio_mport_cdev.c struct list_head node; node 164 drivers/rapidio/devices/rio_mport_cdev.c struct list_head node; node 385 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&map->node, &md->mappings); node 401 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(map, &md->mappings, node) { node 475 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(map, _map, &md->mappings, node) { node 536 drivers/rapidio/devices/rio_mport_cdev.c struct list_head node; node 760 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&req->node, &priv->async_list); node 904 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(map, &md->mappings, node) { node 1032 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(req, &priv->async_list, node) { node 1034 drivers/rapidio/devices/rio_mport_cdev.c list_del(&req->node); node 1082 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&req->node, &priv->async_list); node 1109 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&map->node, &md->mappings); node 1156 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(map, _map, &md->mappings, node) { node 1236 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&map->node, &md->mappings); node 1261 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(map, &md->mappings, node) { node 1342 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(map, _map, &md->mappings, node) { node 1971 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(req, req_next, &list, node) { node 1975 drivers/rapidio/devices/rio_mport_cdev.c list_del(&req->node); node 2040 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(map, _map, &chdev->mappings, node) { node 2152 drivers/rapidio/devices/rio_mport_cdev.c list_del(&map->node); node 2209 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(map, &md->mappings, node) { node 2429 drivers/rapidio/devices/rio_mport_cdev.c list_add_tail(&md->node, &mport_devs); node 2514 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry_safe(map, _map, &md->mappings, node) { node 2569 drivers/rapidio/devices/rio_mport_cdev.c list_for_each_entry(chdev, &mport_devs, node) { node 2572 drivers/rapidio/devices/rio_mport_cdev.c list_del(&chdev->node); node 413 drivers/rapidio/devices/tsi721.c list_for_each_entry(dbell, &mport->dbells, node) { node 1173 drivers/rapidio/devices/tsi721.c list_add_tail(&map->node, &ib_win->mappings); node 1208 drivers/rapidio/devices/tsi721.c list_add_tail(&map->node, &ib_win->mappings); node 1266 drivers/rapidio/devices/tsi721.c &ib_win->mappings, node) { node 1268 drivers/rapidio/devices/tsi721.c list_del(&map->node); node 839 drivers/rapidio/devices/tsi721.h struct list_head node; node 897 drivers/rapidio/rio-scan.c list_for_each_entry(rswitch, &net->switches, node) { node 1029 drivers/rapidio/rio-scan.c list_for_each_entry(rswitch, &net->switches, node) { node 36 drivers/rapidio/rio.c struct list_head node; node 113 drivers/rapidio/rio.c INIT_LIST_HEAD(&net->node); node 131 drivers/rapidio/rio.c list_add_tail(&net->node, &rio_nets); node 141 drivers/rapidio/rio.c if (!list_empty(&net->node)) node 142 drivers/rapidio/rio.c list_del(&net->node); node 186 drivers/rapidio/rio.c list_add_tail(&rdev->rswitch->node, node 212 drivers/rapidio/rio.c list_del(&rdev->rswitch->node); node 416 drivers/rapidio/rio.c list_add_tail(&dbell->node, &mport->dbells); node 480 drivers/rapidio/rio.c list_for_each_entry(dbell, &mport->dbells, node) { node 482 drivers/rapidio/rio.c list_del(&dbell->node); node 574 drivers/rapidio/rio.c list_add_tail(&pwrite->node, &mport->pwrites); node 597 drivers/rapidio/rio.c list_for_each_entry(pwrite, &mport->pwrites, node) { node 599 drivers/rapidio/rio.c list_del(&pwrite->node); node 1211 drivers/rapidio/rio.c list_for_each_entry(pwrite, &mport->pwrites, node) node 1941 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) { node 1984 drivers/rapidio/rio.c list_for_each_entry(scan, &rio_scans, node) { node 2012 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) { node 2020 drivers/rapidio/rio.c list_add_tail(&scan->node, &rio_scans); node 2052 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) node 2057 drivers/rapidio/rio.c list_for_each_entry(scan, &rio_scans, node) { node 2059 drivers/rapidio/rio.c list_del(&scan->node); node 2081 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) { node 2157 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) { node 2193 drivers/rapidio/rio.c list_for_each_entry(port, &rio_mports, node) { node 2253 drivers/rapidio/rio.c list_for_each_entry(scan, &rio_scans, node) { node 2262 drivers/rapidio/rio.c list_add_tail(&port->node, &rio_mports); node 2321 drivers/rapidio/rio.c list_del(&port->node); node 139 drivers/rapidio/rio_cm.c struct list_head node; node 197 drivers/rapidio/rio_cm.c struct list_head node; node 208 drivers/rapidio/rio_cm.c struct list_head node; node 407 drivers/rapidio/rio_cm.c list_add_tail(&req->node, &ch->accept_queue); node 675 drivers/rapidio/rio_cm.c list_for_each_entry_safe(req, _req, &cm->tx_reqs, node) { node 676 drivers/rapidio/rio_cm.c list_del(&req->node); node 718 drivers/rapidio/rio_cm.c list_add_tail(&treq->node, &cm->tx_reqs); node 1140 drivers/rapidio/rio_cm.c req = list_first_entry(&ch->accept_queue, struct conn_req, node); node 1141 drivers/rapidio/rio_cm.c list_del(&req->node); node 1154 drivers/rapidio/rio_cm.c list_for_each_entry(peer, &new_ch->cmdev->peers, node) { node 1596 drivers/rapidio/rio_cm.c list_for_each_entry(peer, &cm->peers, node) { node 1811 drivers/rapidio/rio_cm.c list_for_each_entry(peer, &cm->peers, node) { node 1973 drivers/rapidio/rio_cm.c list_add_tail(&peer->node, &cm->peers); node 2020 drivers/rapidio/rio_cm.c list_for_each_entry(peer, &cm->peers, node) { node 2024 drivers/rapidio/rio_cm.c list_del(&peer->node); node 2230 drivers/rapidio/rio_cm.c list_for_each_entry_safe(peer, temp, &cm->peers, node) { node 2232 drivers/rapidio/rio_cm.c list_del(&peer->node); node 1459 drivers/regulator/core.c struct regulator_map *node; node 1470 drivers/regulator/core.c list_for_each_entry(node, ®ulator_map_list, list) { node 1471 drivers/regulator/core.c if (node->dev_name && consumer_dev_name) { node 1472 drivers/regulator/core.c if (strcmp(node->dev_name, consumer_dev_name) != 0) node 1474 drivers/regulator/core.c } else if (node->dev_name || consumer_dev_name) { node 1478 drivers/regulator/core.c if (strcmp(node->supply, supply) != 0) node 1483 drivers/regulator/core.c dev_name(&node->regulator->dev), node 1484 drivers/regulator/core.c node->regulator->desc->name, node 1490 drivers/regulator/core.c node = kzalloc(sizeof(struct regulator_map), GFP_KERNEL); node 1491 drivers/regulator/core.c if (node == NULL) node 1494 drivers/regulator/core.c node->regulator = rdev; node 1495 drivers/regulator/core.c node->supply = supply; node 1498 drivers/regulator/core.c node->dev_name = kstrdup(consumer_dev_name, GFP_KERNEL); node 1499 drivers/regulator/core.c if (node->dev_name == NULL) { node 1500 drivers/regulator/core.c kfree(node); node 1505 drivers/regulator/core.c list_add(&node->list, ®ulator_map_list); node 1511 drivers/regulator/core.c struct regulator_map *node, *n; node 1513 drivers/regulator/core.c list_for_each_entry_safe(node, n, ®ulator_map_list, list) { node 1514 drivers/regulator/core.c if (rdev == node->regulator) { node 1515 drivers/regulator/core.c list_del(&node->list); node 1516 drivers/regulator/core.c kfree(node->dev_name); node 1517 drivers/regulator/core.c kfree(node); node 1716 drivers/regulator/core.c struct device_node *node; node 1724 drivers/regulator/core.c node = of_get_regulator(dev, supply); node 1725 drivers/regulator/core.c if (node) { node 1726 drivers/regulator/core.c r = of_find_regulator_by_node(node); node 667 drivers/regulator/da9063-regulator.c struct device_node *node; node 674 drivers/regulator/da9063-regulator.c node = of_get_child_by_name(pdev->dev.parent->of_node, "regulators"); node 675 drivers/regulator/da9063-regulator.c if (!node) { node 680 drivers/regulator/da9063-regulator.c num = of_regulator_match(&pdev->dev, node, da9063_matches, node 682 drivers/regulator/da9063-regulator.c of_node_put(node); node 256 drivers/regulator/da9211-regulator.c struct device_node *node; node 259 drivers/regulator/da9211-regulator.c node = of_get_child_by_name(dev->of_node, "regulators"); node 260 drivers/regulator/da9211-regulator.c if (!node) { node 265 drivers/regulator/da9211-regulator.c num = of_regulator_match(dev, node, da9211_matches, node 267 drivers/regulator/da9211-regulator.c of_node_put(node); node 61 drivers/regulator/internal.h struct device_node **node); node 81 drivers/regulator/internal.h struct device_node **node) node 231 drivers/regulator/mc13783-regulator.c #define MC13783_DEFINE(prefix, name, node, reg, vsel_reg, voltages) \ node 232 drivers/regulator/mc13783-regulator.c MC13xxx_DEFINE(MC13783_REG_, name, node, reg, vsel_reg, voltages, \ node 235 drivers/regulator/mc13783-regulator.c #define MC13783_FIXED_DEFINE(prefix, name, node, reg, voltages) \ node 236 drivers/regulator/mc13783-regulator.c MC13xxx_FIXED_DEFINE(MC13783_REG_, name, node, reg, voltages, \ node 239 drivers/regulator/mc13783-regulator.c #define MC13783_GPO_DEFINE(prefix, name, node, reg, voltages) \ node 240 drivers/regulator/mc13783-regulator.c MC13xxx_GPO_DEFINE(MC13783_REG_, name, node, reg, voltages, \ node 425 drivers/regulator/mc13783-regulator.c struct device_node *node = NULL; node 431 drivers/regulator/mc13783-regulator.c node = mc13xxx_data[i].node; node 441 drivers/regulator/mc13783-regulator.c config.of_node = node; node 249 drivers/regulator/mc13892-regulator.c #define MC13892_FIXED_DEFINE(name, node, reg, voltages) \ node 250 drivers/regulator/mc13892-regulator.c MC13xxx_FIXED_DEFINE(MC13892_, name, node, reg, voltages, \ node 253 drivers/regulator/mc13892-regulator.c #define MC13892_GPO_DEFINE(name, node, reg, voltages) \ node 254 drivers/regulator/mc13892-regulator.c MC13xxx_GPO_DEFINE(MC13892_, name, node, reg, voltages, \ node 257 drivers/regulator/mc13892-regulator.c #define MC13892_SW_DEFINE(name, node, reg, vsel_reg, voltages) \ node 258 drivers/regulator/mc13892-regulator.c MC13xxx_DEFINE(MC13892_, name, node, reg, vsel_reg, voltages, \ node 261 drivers/regulator/mc13892-regulator.c #define MC13892_DEFINE_REGU(name, node, reg, vsel_reg, voltages) \ node 262 drivers/regulator/mc13892-regulator.c MC13xxx_DEFINE(MC13892_, name, node, reg, vsel_reg, voltages, \ node 595 drivers/regulator/mc13892-regulator.c struct device_node *node = NULL; node 601 drivers/regulator/mc13892-regulator.c node = mc13xxx_data[i].node; node 611 drivers/regulator/mc13892-regulator.c config.of_node = node; node 195 drivers/regulator/mc13xxx-regulator-core.c p->node = child; node 274 drivers/regulator/of_regulator.c struct device_node *node, node 279 drivers/regulator/of_regulator.c if (!node) node 286 drivers/regulator/of_regulator.c if (of_get_regulation_constraints(dev, node, &init_data, desc)) node 325 drivers/regulator/of_regulator.c int of_regulator_match(struct device *dev, struct device_node *node, node 335 drivers/regulator/of_regulator.c if (!dev || !node) node 355 drivers/regulator/of_regulator.c for_each_child_of_node(node, child) { node 433 drivers/regulator/of_regulator.c struct device_node **node) node 464 drivers/regulator/of_regulator.c *node = child; node 488 drivers/regulator/of_regulator.c struct device_node *node = rdev->dev.of_node; node 491 drivers/regulator/of_regulator.c n_phandles = of_count_phandle_with_args(node, node 547 drivers/regulator/of_regulator.c struct device_node *node = rdev->dev.of_node; node 564 drivers/regulator/of_regulator.c c_node = of_parse_phandle(node, node 580 drivers/regulator/of_regulator.c if (!of_coupling_find_node(c_node, node, &index)) { node 620 drivers/regulator/of_regulator.c struct device_node *node = rdev->dev.of_node; node 624 drivers/regulator/of_regulator.c c_node = of_parse_phandle(node, "regulator-coupled-with", index); node 1467 drivers/regulator/palmas-regulator.c struct device_node *node, node 1475 drivers/regulator/palmas-regulator.c regulators = of_get_child_by_name(node, "regulators"); node 1550 drivers/regulator/palmas-regulator.c pdata->ldo6_vibrator = of_property_read_bool(node, "ti,ldo6-vibrator"); node 1599 drivers/regulator/palmas-regulator.c struct device_node *node = pdev->dev.of_node; node 1622 drivers/regulator/palmas-regulator.c if (of_device_is_compatible(node, "ti,tps659038-pmic")) { node 1634 drivers/regulator/palmas-regulator.c ret = palmas_dt_to_pdata(&pdev->dev, node, pdata, driver_data); node 414 drivers/regulator/qcom-rpmh-regulator.c struct device_node *node, const char *pmic_id, node 427 drivers/regulator/qcom-rpmh-regulator.c if (of_node_name_eq(node, rpmh_data->name)) node 431 drivers/regulator/qcom-rpmh-regulator.c dev_err(dev, "Unknown regulator %pOFn\n", node); node 441 drivers/regulator/qcom-rpmh-regulator.c node, rpmh_resource_name); node 459 drivers/regulator/qcom-rpmh-regulator.c vreg->always_wait_for_ack = of_property_read_bool(node, node 467 drivers/regulator/qcom-rpmh-regulator.c init_data = of_get_regulator_init_data(dev, node, &vreg->rdesc); node 480 drivers/regulator/qcom-rpmh-regulator.c reg_config.of_node = node; node 487 drivers/regulator/qcom-rpmh-regulator.c node, ret); node 492 drivers/regulator/qcom-rpmh-regulator.c node, rpmh_resource_name, vreg->addr); node 885 drivers/regulator/qcom-rpmh-regulator.c struct device_node *node; node 900 drivers/regulator/qcom-rpmh-regulator.c for_each_available_child_of_node(dev->of_node, node) { node 903 drivers/regulator/qcom-rpmh-regulator.c of_node_put(node); node 907 drivers/regulator/qcom-rpmh-regulator.c ret = rpmh_regulator_init_vreg(vreg, dev, node, pmic_id, node 910 drivers/regulator/qcom-rpmh-regulator.c of_node_put(node); node 654 drivers/regulator/qcom_rpm-regulator.c struct device_node *node, node 669 drivers/regulator/qcom_rpm-regulator.c ret = of_property_read_u32(node, key, &freq); node 686 drivers/regulator/qcom_rpm-regulator.c static int rpm_reg_of_parse(struct device_node *node, node 699 drivers/regulator/qcom_rpm-regulator.c if (of_property_read_bool(node, key)) { node 708 drivers/regulator/qcom_rpm-regulator.c ret = rpm_reg_of_parse_freq(dev, node, vreg); node 715 drivers/regulator/qcom_rpm-regulator.c pwm = !of_property_read_bool(node, key); node 728 drivers/regulator/qcom_rpm-regulator.c ret = of_property_read_u32(node, key, &val); node 374 drivers/regulator/qcom_spmi-regulator.c struct list_head node; node 1699 drivers/regulator/qcom_spmi-regulator.c struct device_node *node, struct spmi_regulator_init_data *data) node 1710 drivers/regulator/qcom_spmi-regulator.c of_property_read_u32(node, "qcom,ocp-max-retries", node 1712 drivers/regulator/qcom_spmi-regulator.c of_property_read_u32(node, "qcom,ocp-retry-delay", node 1714 drivers/regulator/qcom_spmi-regulator.c of_property_read_u32(node, "qcom,pin-ctrl-enable", node 1716 drivers/regulator/qcom_spmi-regulator.c of_property_read_u32(node, "qcom,pin-ctrl-hpm", &data->pin_ctrl_hpm); node 1717 drivers/regulator/qcom_spmi-regulator.c of_property_read_u32(node, "qcom,vs-soft-start-strength", node 1731 drivers/regulator/qcom_spmi-regulator.c static int spmi_regulator_of_parse(struct device_node *node, node 1740 drivers/regulator/qcom_spmi-regulator.c spmi_regulator_get_dt_config(vreg, node, &data); node 1966 drivers/regulator/qcom_spmi-regulator.c struct device_node *node = pdev->dev.of_node; node 1986 drivers/regulator/qcom_spmi-regulator.c if (of_find_property(node, "qcom,saw-reg", &lenp)) { node 1987 drivers/regulator/qcom_spmi-regulator.c syscon = of_parse_phandle(node, "qcom,saw-reg", 0); node 1997 drivers/regulator/qcom_spmi-regulator.c reg_node = of_get_child_by_name(node, reg->name); node 2036 drivers/regulator/qcom_spmi-regulator.c reg_node = of_get_child_by_name(node, reg->name); node 2064 drivers/regulator/qcom_spmi-regulator.c INIT_LIST_HEAD(&vreg->node); node 2065 drivers/regulator/qcom_spmi-regulator.c list_add(&vreg->node, vreg_list); node 2071 drivers/regulator/qcom_spmi-regulator.c list_for_each_entry(vreg, vreg_list, node) node 2082 drivers/regulator/qcom_spmi-regulator.c list_for_each_entry(vreg, vreg_list, node) node 165 drivers/regulator/tps65086-regulator.c static int tps65086_of_parse_cb(struct device_node *node, node 172 drivers/regulator/tps65086-regulator.c if (of_property_read_bool(node, "ti,regulator-step-size-25mv")) { node 196 drivers/regulator/tps65086-regulator.c if (desc->id <= BUCK6 && of_property_read_bool(node, "ti,regulator-decay")) { node 285 drivers/remoteproc/imx_rproc.c struct device_node *node; node 288 drivers/remoteproc/imx_rproc.c node = of_parse_phandle(np, "memory-region", a); node 289 drivers/remoteproc/imx_rproc.c err = of_address_to_resource(node, 0, &res); node 32 drivers/remoteproc/qcom_common.c glink->edge = qcom_glink_smem_register(glink->dev, glink->node); node 54 drivers/remoteproc/qcom_common.c glink->node = of_get_child_by_name(dev->parent->of_node, "glink-edge"); node 55 drivers/remoteproc/qcom_common.c if (!glink->node) node 73 drivers/remoteproc/qcom_common.c if (!glink->node) node 77 drivers/remoteproc/qcom_common.c of_node_put(glink->node); node 128 drivers/remoteproc/qcom_common.c smd->edge = qcom_smd_register_edge(smd->dev, smd->node); node 150 drivers/remoteproc/qcom_common.c smd->node = of_get_child_by_name(dev->parent->of_node, "smd-edge"); node 151 drivers/remoteproc/qcom_common.c if (!smd->node) node 169 drivers/remoteproc/qcom_common.c if (!smd->node) node 173 drivers/remoteproc/qcom_common.c of_node_put(smd->node); node 15 drivers/remoteproc/qcom_common.h struct device_node *node; node 23 drivers/remoteproc/qcom_common.h struct device_node *node; node 382 drivers/remoteproc/qcom_q6v5_adsp.c struct device_node *node; node 386 drivers/remoteproc/qcom_q6v5_adsp.c node = of_parse_phandle(adsp->dev->of_node, "memory-region", 0); node 387 drivers/remoteproc/qcom_q6v5_adsp.c if (!node) { node 392 drivers/remoteproc/qcom_q6v5_adsp.c ret = of_address_to_resource(node, 0, &r); node 1363 drivers/remoteproc/qcom_q6v5_mss.c struct device_node *node; node 1368 drivers/remoteproc/qcom_q6v5_mss.c node = of_parse_phandle(child, "memory-region", 0); node 1369 drivers/remoteproc/qcom_q6v5_mss.c ret = of_address_to_resource(node, 0, &r); node 1374 drivers/remoteproc/qcom_q6v5_mss.c of_node_put(node); node 1386 drivers/remoteproc/qcom_q6v5_mss.c node = of_parse_phandle(child, "memory-region", 0); node 1387 drivers/remoteproc/qcom_q6v5_mss.c ret = of_address_to_resource(node, 0, &r); node 1392 drivers/remoteproc/qcom_q6v5_mss.c of_node_put(node); node 222 drivers/remoteproc/qcom_q6v5_pas.c struct device_node *node; node 226 drivers/remoteproc/qcom_q6v5_pas.c node = of_parse_phandle(adsp->dev->of_node, "memory-region", 0); node 227 drivers/remoteproc/qcom_q6v5_pas.c if (!node) { node 232 drivers/remoteproc/qcom_q6v5_pas.c ret = of_address_to_resource(node, 0, &r); node 502 drivers/remoteproc/qcom_q6v5_wcss.c struct device_node *node; node 505 drivers/remoteproc/qcom_q6v5_wcss.c node = of_parse_phandle(dev->of_node, "memory-region", 0); node 506 drivers/remoteproc/qcom_q6v5_wcss.c if (node) node 507 drivers/remoteproc/qcom_q6v5_wcss.c rmem = of_reserved_mem_lookup(node); node 508 drivers/remoteproc/qcom_q6v5_wcss.c of_node_put(node); node 25 drivers/remoteproc/qcom_sysmon.c struct list_head node; node 402 drivers/remoteproc/qcom_sysmon.c sysmon->ssctl.sq_node = svc->node; node 555 drivers/remoteproc/qcom_sysmon.c list_add(&sysmon->node, &sysmon_list); node 572 drivers/remoteproc/qcom_sysmon.c list_del(&sysmon->node); node 606 drivers/remoteproc/qcom_sysmon.c list_for_each_entry(sysmon, &sysmon_list, node) { node 432 drivers/remoteproc/qcom_wcnss.c struct device_node *node; node 436 drivers/remoteproc/qcom_wcnss.c node = of_parse_phandle(wcnss->dev->of_node, "memory-region", 0); node 437 drivers/remoteproc/qcom_wcnss.c if (!node) { node 442 drivers/remoteproc/qcom_wcnss.c ret = of_address_to_resource(node, 0, &r); node 201 drivers/remoteproc/remoteproc_core.c list_for_each_entry(carveout, &rproc->carveouts, node) { node 257 drivers/remoteproc/remoteproc_core.c list_for_each_entry(carveout, &rproc->carveouts, node) { node 553 drivers/remoteproc/remoteproc_core.c list_add_tail(&rvdev->node, &rproc->rvdevs); node 583 drivers/remoteproc/remoteproc_core.c list_del(&rvdev->node); node 642 drivers/remoteproc/remoteproc_core.c list_add_tail(&trace->node, &rproc->traces); node 718 drivers/remoteproc/remoteproc_core.c list_add_tail(&mapping->node, &rproc->mappings); node 809 drivers/remoteproc/remoteproc_core.c list_add_tail(&mapping->node, &rproc->mappings); node 941 drivers/remoteproc/remoteproc_core.c list_add_tail(&mem->node, &rproc->carveouts); node 1098 drivers/remoteproc/remoteproc_core.c list_for_each_entry(subdev, &rproc->subdevs, node) { node 1109 drivers/remoteproc/remoteproc_core.c list_for_each_entry_continue_reverse(subdev, &rproc->subdevs, node) { node 1122 drivers/remoteproc/remoteproc_core.c list_for_each_entry(subdev, &rproc->subdevs, node) { node 1133 drivers/remoteproc/remoteproc_core.c list_for_each_entry_continue_reverse(subdev, &rproc->subdevs, node) { node 1145 drivers/remoteproc/remoteproc_core.c list_for_each_entry_reverse(subdev, &rproc->subdevs, node) { node 1155 drivers/remoteproc/remoteproc_core.c list_for_each_entry_reverse(subdev, &rproc->subdevs, node) { node 1180 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) { node 1239 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->dump_segments, node) { node 1240 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); node 1260 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(trace, ttmp, &rproc->traces, node) { node 1263 drivers/remoteproc/remoteproc_core.c list_del(&trace->node); node 1268 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->mappings, node) { node 1278 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); node 1283 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) { node 1286 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); node 1291 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(rvdev, rvtmp, &rproc->rvdevs, node) node 1514 drivers/remoteproc/remoteproc_core.c list_add_tail(&segment->node, &rproc->dump_segments); node 1552 drivers/remoteproc/remoteproc_core.c list_add_tail(&segment->node, &rproc->dump_segments); node 1580 drivers/remoteproc/remoteproc_core.c list_for_each_entry(segment, &rproc->dump_segments, node) { node 1609 drivers/remoteproc/remoteproc_core.c list_for_each_entry(segment, &rproc->dump_segments, node) { node 1859 drivers/remoteproc/remoteproc_core.c list_for_each_entry(r, &rproc_list, node) { node 1929 drivers/remoteproc/remoteproc_core.c list_add(&rproc->node, &rproc_list); node 2144 drivers/remoteproc/remoteproc_core.c list_del(&rproc->node); node 2162 drivers/remoteproc/remoteproc_core.c list_add_tail(&subdev->node, &rproc->subdevs); node 2173 drivers/remoteproc/remoteproc_core.c list_del(&subdev->node); node 290 drivers/remoteproc/remoteproc_debugfs.c list_for_each_entry(carveout, &rproc->carveouts, node) { node 23 drivers/remoteproc/remoteproc_internal.h struct list_head node; node 130 drivers/remoteproc/st_remoteproc.c rmem = of_reserved_mem_lookup(it.node); node 137 drivers/remoteproc/st_remoteproc.c if (strcmp(it.node->name, "vdev0buffer")) { node 144 drivers/remoteproc/st_remoteproc.c it.node->name); node 150 drivers/remoteproc/st_remoteproc.c it.node->name); node 212 drivers/remoteproc/stm32_rproc.c rmem = of_reserved_mem_lookup(it.node); node 225 drivers/remoteproc/stm32_rproc.c if (strcmp(it.node->name, "vdev0buffer")) { node 232 drivers/remoteproc/stm32_rproc.c it.node->name); node 242 drivers/remoteproc/stm32_rproc.c it.node->name); node 368 drivers/remoteproc/stm32_rproc.c list_for_each_entry(trace, &rproc->traces, node) { node 371 drivers/remoteproc/stm32_rproc.c list_for_each_entry(segment, &rproc->dump_segments, node) { node 600 drivers/reset/core.c struct reset_control *__of_reset_control_get(struct device_node *node, node 610 drivers/reset/core.c if (!node) node 614 drivers/reset/core.c index = of_property_match_string(node, node 622 drivers/reset/core.c ret = of_parse_phandle_with_args(node, "resets", "#reset-cells", node 840 drivers/reset/core.c static int of_reset_control_get_count(struct device_node *node) node 844 drivers/reset/core.c if (!node) node 847 drivers/reset/core.c count = of_count_phandle_with_args(node, "resets", "#reset-cells"); node 49 drivers/rpmsg/qcom_glink_native.c struct list_head node; node 75 drivers/rpmsg/qcom_glink_native.c struct list_head node; node 254 drivers/rpmsg/qcom_glink_native.c list_for_each_entry_safe(intent, tmp, &channel->done_intents, node) { node 491 drivers/rpmsg/qcom_glink_native.c list_for_each_entry_safe(intent, tmp, &channel->done_intents, node) { node 492 drivers/rpmsg/qcom_glink_native.c list_del(&intent->node); node 531 drivers/rpmsg/qcom_glink_native.c list_add_tail(&intent->node, &channel->done_intents); node 772 drivers/rpmsg/qcom_glink_native.c INIT_LIST_HEAD(&dcmd->node); node 777 drivers/rpmsg/qcom_glink_native.c list_add_tail(&dcmd->node, &glink->rx_queue); node 1338 drivers/rpmsg/qcom_glink_native.c static struct device_node *qcom_glink_match_channel(struct device_node *node, node 1346 drivers/rpmsg/qcom_glink_native.c for_each_available_child_of_node(node, child) { node 1385 drivers/rpmsg/qcom_glink_native.c struct device_node *node; node 1431 drivers/rpmsg/qcom_glink_native.c node = qcom_glink_match_channel(glink->dev->of_node, name); node 1432 drivers/rpmsg/qcom_glink_native.c rpdev->dev.of_node = node; node 1528 drivers/rpmsg/qcom_glink_native.c struct glink_defer_cmd, node); node 1529 drivers/rpmsg/qcom_glink_native.c list_del(&dcmd->node); node 1573 drivers/rpmsg/qcom_glink_native.c list_for_each_entry_safe(dcmd, tmp, &glink->rx_queue, node) node 186 drivers/rpmsg/qcom_glink_smem.c struct device_node *node) node 202 drivers/rpmsg/qcom_glink_smem.c dev->of_node = node; node 204 drivers/rpmsg/qcom_glink_smem.c dev_set_name(dev, "%s:%pOFn", dev_name(parent->parent), node); node 1321 drivers/rpmsg/qcom_smd.c struct device_node *node, node 1335 drivers/rpmsg/qcom_smd.c edge->of_node = of_node_get(node); node 1338 drivers/rpmsg/qcom_smd.c ret = of_property_read_u32(node, key, &edge->edge_id); node 1346 drivers/rpmsg/qcom_smd.c of_property_read_u32(node, key, &edge->remote_pid); node 1357 drivers/rpmsg/qcom_smd.c syscon_np = of_parse_phandle(node, "qcom,ipc", 0); node 1368 drivers/rpmsg/qcom_smd.c ret = of_property_read_u32_index(node, key, 1, &edge->ipc_offset); node 1374 drivers/rpmsg/qcom_smd.c ret = of_property_read_u32_index(node, key, 2, &edge->ipc_bit); node 1381 drivers/rpmsg/qcom_smd.c ret = of_property_read_string(node, "label", &edge->name); node 1383 drivers/rpmsg/qcom_smd.c edge->name = node->name; node 1385 drivers/rpmsg/qcom_smd.c irq = irq_of_parse_and_map(node, 0); node 1393 drivers/rpmsg/qcom_smd.c node->name, edge); node 1445 drivers/rpmsg/qcom_smd.c struct device_node *node) node 1458 drivers/rpmsg/qcom_smd.c edge->dev.of_node = node; node 1460 drivers/rpmsg/qcom_smd.c dev_set_name(&edge->dev, "%s:%pOFn", dev_name(parent), node); node 1468 drivers/rpmsg/qcom_smd.c ret = qcom_smd_parse_edge(&edge->dev, node, edge); node 1525 drivers/rpmsg/qcom_smd.c struct device_node *node; node 1533 drivers/rpmsg/qcom_smd.c for_each_available_child_of_node(pdev->dev.of_node, node) node 1534 drivers/rpmsg/qcom_smd.c qcom_smd_register_edge(&pdev->dev, node); node 400 drivers/rtc/interface.c alarm->time = rtc_ktime_to_tm(rtc->aie_timer.node.expires); node 471 drivers/rtc/interface.c rtc->aie_timer.node.expires = rtc_tm_to_ktime(alarm->time); node 500 drivers/rtc/interface.c rtc->aie_timer.node.expires = rtc_tm_to_ktime(alarm->time); node 505 drivers/rtc/interface.c rtc->aie_timer.node.expires)) { node 507 drivers/rtc/interface.c timerqueue_add(&rtc->timerqueue, &rtc->aie_timer.node); node 576 drivers/rtc/interface.c rtc->uie_rtctimer.node.expires = ktime_add(now, onesec); node 808 drivers/rtc/interface.c timerqueue_add(&rtc->timerqueue, &timer->node); node 810 drivers/rtc/interface.c if (!next || ktime_before(timer->node.expires, next->expires)) { node 814 drivers/rtc/interface.c alarm.time = rtc_ktime_to_tm(timer->node.expires); node 821 drivers/rtc/interface.c timerqueue_del(&rtc->timerqueue, &timer->node); node 855 drivers/rtc/interface.c timerqueue_del(&rtc->timerqueue, &timer->node); node 858 drivers/rtc/interface.c if (next == &timer->node) { node 906 drivers/rtc/interface.c timer = container_of(next, struct rtc_timer, node); node 907 drivers/rtc/interface.c timerqueue_del(&rtc->timerqueue, &timer->node); node 916 drivers/rtc/interface.c timer->node.expires = ktime_add(timer->node.expires, node 919 drivers/rtc/interface.c timerqueue_add(&rtc->timerqueue, &timer->node); node 940 drivers/rtc/interface.c timer = container_of(next, struct rtc_timer, node); node 941 drivers/rtc/interface.c timerqueue_del(&rtc->timerqueue, &timer->node); node 965 drivers/rtc/interface.c timerqueue_init(&timer->node); node 988 drivers/rtc/interface.c timer->node.expires = expires; node 244 drivers/rtc/rtc-88pm80x.c struct device_node *node = pdev->dev.of_node; node 247 drivers/rtc/rtc-88pm80x.c if (!pdata && !node) { node 189 drivers/rtc/rtc-ab-eoz9.c static int abeoz9_trickle_parse_dt(struct device_node *node) node 193 drivers/rtc/rtc-ab-eoz9.c if (of_property_read_u32(node, "trickle-resistor-ohms", &ohms)) node 210 drivers/rtc/rtc-ab-eoz9.c static int abeoz9_rtc_setup(struct device *dev, struct device_node *node) node 243 drivers/rtc/rtc-ab-eoz9.c ret = abeoz9_trickle_parse_dt(node); node 142 drivers/rtc/rtc-bq32k.c static int trickle_charger_of_init(struct device *dev, struct device_node *node) node 148 drivers/rtc/rtc-bq32k.c if (of_property_read_u32(node, "trickle-resistor-ohms" , &ohms)) node 158 drivers/rtc/rtc-bq32k.c if (of_property_read_bool(node, "trickle-diode-disable")) { node 168 drivers/rtc/rtc-bq32k.c if (!of_property_read_bool(node, "trickle-diode-disable")) { node 1374 drivers/rtc/rtc-cmos.c struct device_node *node = pdev->dev.of_node; node 1377 drivers/rtc/rtc-cmos.c if (!node) node 1380 drivers/rtc/rtc-cmos.c val = of_get_property(node, "ctrl-reg", NULL); node 1384 drivers/rtc/rtc-cmos.c val = of_get_property(node, "freq-reg", NULL); node 1505 drivers/rtc/rtc-ds1307.c struct device_node *node = ds1307->dev->of_node; node 1530 drivers/rtc/rtc-ds1307.c of_property_read_string_index(node, "clock-output-names", i, node 1540 drivers/rtc/rtc-ds1307.c if (!node) node 1543 drivers/rtc/rtc-ds1307.c of_clk_add_provider(node, of_clk_src_onecell_get, onecell); node 395 drivers/rtc/rtc-hym8563.c struct device_node *node = client->dev.of_node; node 413 drivers/rtc/rtc-hym8563.c of_property_read_string(node, "clock-output-names", &init.name); node 419 drivers/rtc/rtc-hym8563.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 547 drivers/rtc/rtc-m41t80.c struct device_node *node = client->dev.of_node; node 570 drivers/rtc/rtc-m41t80.c of_property_read_string(node, "clock-output-names", &init.name); node 575 drivers/rtc/rtc-m41t80.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 506 drivers/rtc/rtc-pcf8563.c struct device_node *node = client->dev.of_node; node 526 drivers/rtc/rtc-pcf8563.c of_property_read_string(node, "clock-output-names", &init.name); node 532 drivers/rtc/rtc-pcf8563.c of_clk_add_provider(node, of_clk_src_simple_get, clk); node 491 drivers/rtc/rtc-rv8803.c struct device_node *node = client->dev.of_node; node 495 drivers/rtc/rtc-rv8803.c if (!node) node 507 drivers/rtc/rtc-rv8803.c if (of_property_read_bool(node, "epson,vdet-disable")) node 510 drivers/rtc/rtc-rv8803.c if (of_property_read_bool(node, "trickle-diode-disable")) node 453 drivers/rtc/rtc-sc27xx.c rtc_ktime_to_tm(rtc->rtc->aie_timer.node.expires); node 598 drivers/rtc/rtc-sc27xx.c struct device_node *node = pdev->dev.of_node; node 610 drivers/rtc/rtc-sc27xx.c ret = of_property_read_u32(node, "reg", &rtc->base); node 214 drivers/rtc/rtc-sun6i.c static void __init sun6i_rtc_clk_init(struct device_node *node, node 241 drivers/rtc/rtc-sun6i.c rtc->base = of_io_request_and_map(node, 0, of_node_full_name(node)); node 264 drivers/rtc/rtc-sun6i.c if (!of_get_property(node, "clocks", NULL)) node 269 drivers/rtc/rtc-sun6i.c of_property_read_string_index(node, "clock-output-names", 2, node 283 drivers/rtc/rtc-sun6i.c parents[1] = of_clk_get_parent_name(node, 0); node 288 drivers/rtc/rtc-sun6i.c init.num_parents = of_clk_get_parent_count(node) + 1; node 289 drivers/rtc/rtc-sun6i.c of_property_read_string_index(node, "clock-output-names", 0, node 298 drivers/rtc/rtc-sun6i.c of_property_read_string_index(node, "clock-output-names", 1, node 316 drivers/rtc/rtc-sun6i.c of_clk_add_hw_provider(node, of_clk_hw_onecell_get, clk_data); node 328 drivers/rtc/rtc-sun6i.c static void __init sun6i_a31_rtc_clk_init(struct device_node *node) node 330 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun6i_a31_rtc_data); node 341 drivers/rtc/rtc-sun6i.c static void __init sun8i_a23_rtc_clk_init(struct device_node *node) node 343 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun8i_a23_rtc_data); node 356 drivers/rtc/rtc-sun6i.c static void __init sun8i_h3_rtc_clk_init(struct device_node *node) node 358 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun8i_h3_rtc_data); node 376 drivers/rtc/rtc-sun6i.c static void __init sun50i_h6_rtc_clk_init(struct device_node *node) node 378 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun50i_h6_rtc_data); node 392 drivers/rtc/rtc-sun6i.c static void __init sun8i_r40_rtc_clk_init(struct device_node *node) node 394 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun8i_r40_rtc_data); node 404 drivers/rtc/rtc-sun6i.c static void __init sun8i_v3_rtc_clk_init(struct device_node *node) node 406 drivers/rtc/rtc-sun6i.c sun6i_rtc_clk_init(node, &sun8i_v3_rtc_data); node 182 drivers/s390/char/tape.h struct list_head node; node 339 drivers/s390/char/tape_core.c list_for_each_entry(tmp, &tape_device_list, node) { node 349 drivers/s390/char/tape_core.c list_add_tail(&device->node, &tmp->node); node 359 drivers/s390/char/tape_core.c list_del_init(&device->node); node 543 drivers/s390/char/tape_core.c INIT_LIST_HEAD(&device->node); node 600 drivers/s390/char/tape_core.c list_for_each_entry(tmp, &tape_device_list, node) { node 440 drivers/s390/cio/cmf.c struct ccw_device_private *node; node 455 drivers/s390/cio/cmf.c list_for_each_entry(node, &cmb_area.list, cmb_list) { node 457 drivers/s390/cio/cmf.c data = node->cmb; node 468 drivers/s390/cio/cmf.c list_add_tail(&cdev->private->cmb_list, &node->cmb_list); node 343 drivers/s390/crypto/vfio_ap_ops.c list_add(&matrix_mdev->node, &matrix_dev->mdev_list); node 358 drivers/s390/crypto/vfio_ap_ops.c list_del(&matrix_mdev->node); node 542 drivers/s390/crypto/vfio_ap_ops.c list_for_each_entry(lstdev, &matrix_dev->mdev_list, node) { node 1042 drivers/s390/crypto/vfio_ap_ops.c list_for_each_entry(m, &matrix_dev->mdev_list, node) { node 82 drivers/s390/crypto/vfio_ap_private.h struct list_head node; node 130 drivers/s390/virtio/virtio_ccw.c struct list_head node; node 302 drivers/s390/virtio/virtio_ccw.c list_for_each_entry(info, &vcdev->virtqueues, node) node 456 drivers/s390/virtio/virtio_ccw.c list_del(&info->node); node 587 drivers/s390/virtio/virtio_ccw.c list_add(&info->node, &vcdev->virtqueues); node 1057 drivers/s390/virtio/virtio_ccw.c list_for_each_entry(info, &vcdev->virtqueues, node) { node 1857 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y symbol_node_t *node; node 1892 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y for(node = expression->referenced_syms.slh_first; node 1893 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y node != NULL; node 1894 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y node = node->links.sle_next) { node 1895 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y if ((node->symbol->type == MASK node 1896 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y || node->symbol->type == FIELD node 1897 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y || node->symbol->type == ENUM node 1898 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y || node->symbol->type == ENUM_ENTRY) node 1899 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y && symlist_search(&node->symbol->info.finfo->symrefs, node 1904 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y node->symbol->name, symbol->name); node 314 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c symbol_node_t *node; node 317 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c while((node = SLIST_FIRST(symlist_src2)) != NULL) { node 319 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c SLIST_INSERT_HEAD(symlist_dest, node, links); node 3948 drivers/scsi/bfa/bfa_svc.c bfa_fcport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node) node 3951 drivers/scsi/bfa/bfa_svc.c if (node) node 541 drivers/scsi/bfa/bfa_svc.h wwn_t bfa_fcport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node); node 64 drivers/scsi/device_handler/scsi_dh_alua.c struct list_head node; node 84 drivers/scsi/device_handler/scsi_dh_alua.c struct list_head node; node 116 drivers/scsi/device_handler/scsi_dh_alua.c list_del(&pg->node); node 186 drivers/scsi/device_handler/scsi_dh_alua.c list_for_each_entry(pg, &port_group_list, node) { node 241 drivers/scsi/device_handler/scsi_dh_alua.c INIT_LIST_HEAD(&pg->node); node 254 drivers/scsi/device_handler/scsi_dh_alua.c list_add(&pg->node, &port_group_list); node 363 drivers/scsi/device_handler/scsi_dh_alua.c list_del_rcu(&h->node); node 372 drivers/scsi/device_handler/scsi_dh_alua.c list_add_rcu(&h->node, &pg->dh_list); node 660 drivers/scsi/device_handler/scsi_dh_alua.c &tmp_pg->dh_list, node) { node 707 drivers/scsi/device_handler/scsi_dh_alua.c list_for_each_entry_rcu(h, &pg->dh_list, node) { node 1124 drivers/scsi/device_handler/scsi_dh_alua.c INIT_LIST_HEAD(&h->node); node 1154 drivers/scsi/device_handler/scsi_dh_alua.c list_del_rcu(&h->node); node 155 drivers/scsi/device_handler/scsi_dh_rdac.c struct list_head node; /* list of all controllers */ node 185 drivers/scsi/device_handler/scsi_dh_rdac.c struct list_head node; node 322 drivers/scsi/device_handler/scsi_dh_rdac.c list_del(&ctlr->node); node 331 drivers/scsi/device_handler/scsi_dh_rdac.c list_for_each_entry(tmp, &ctlr_list, node) { node 356 drivers/scsi/device_handler/scsi_dh_rdac.c list_add(&ctlr->node, &ctlr_list); node 425 drivers/scsi/device_handler/scsi_dh_rdac.c list_for_each_entry_rcu(tmp, &h->ctlr->dh_list, node) { node 456 drivers/scsi/device_handler/scsi_dh_rdac.c list_add_rcu(&h->node, &h->ctlr->dh_list); node 780 drivers/scsi/device_handler/scsi_dh_rdac.c list_del_rcu(&h->node); node 552 drivers/scsi/isci/init.c INIT_LIST_HEAD(&idev->node); node 260 drivers/scsi/isci/port.c node) { node 1053 drivers/scsi/isci/remote_device.c list_del_init(&idev->node); node 1527 drivers/scsi/isci/remote_device.c if (WARN_ONCE(!list_empty(&idev->node), "found non-idle remote device\n")) node 1631 drivers/scsi/isci/remote_device.c INIT_LIST_HEAD(&isci_device->node); node 1636 drivers/scsi/isci/remote_device.c list_add_tail(&isci_device->node, &isci_port->remote_dev_list); node 95 drivers/scsi/isci/remote_device.h struct list_head node; node 54 drivers/scsi/lpfc/lpfc_bsg.c struct list_head node; node 114 drivers/scsi/lpfc/lpfc_bsg.c struct list_head node; node 807 drivers/scsi/lpfc/lpfc_bsg.c list_del(&evt->node); node 810 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_get.next, typeof(*ed), node); node 811 drivers/scsi/lpfc/lpfc_bsg.c list_del(&ed->node); node 817 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_see.next, typeof(*ed), node); node 818 drivers/scsi/lpfc/lpfc_bsg.c list_del(&ed->node); node 966 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry(evt, &phba->ct_ev_waiters, node) { node 1123 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt_dat->node, &evt->events_to_see); node 1232 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry(evt, &phba->ct_ev_waiters, node) { node 1242 drivers/scsi/lpfc/lpfc_bsg.c if (&evt->node == &phba->ct_ev_waiters) { node 1265 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); node 1317 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry_safe(evt, evt_next, &phba->ct_ev_waiters, node) { node 1324 drivers/scsi/lpfc/lpfc_bsg.c struct event_data, node); node 1325 drivers/scsi/lpfc/lpfc_bsg.c list_del(&evt_dat->node); node 2670 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); node 2761 drivers/scsi/lpfc/lpfc_bsg.c node))->immed_dat; node 3220 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); node 3339 drivers/scsi/lpfc/lpfc_bsg.c typeof(*evdat), node); node 11089 drivers/scsi/lpfc/lpfc_init.c static int lpfc_cpu_offline(unsigned int cpu, struct hlist_node *node) node 11091 drivers/scsi/lpfc/lpfc_init.c struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp); node 11115 drivers/scsi/lpfc/lpfc_init.c static int lpfc_cpu_online(unsigned int cpu, struct hlist_node *node) node 11117 drivers/scsi/lpfc/lpfc_init.c struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp); node 412 drivers/scsi/mac53c94.c struct device_node *node = macio_get_of_node(mdev); node 452 drivers/scsi/mac53c94.c printk(KERN_ERR "mac53c94: ioremap failed for %pOF\n", node); node 456 drivers/scsi/mac53c94.c clkprop = of_get_property(node, "clock-frequency", &proplen); node 459 drivers/scsi/mac53c94.c "assuming 25MHz\n", node); node 473 drivers/scsi/mac53c94.c "command space for %pOF\n", node); node 486 drivers/scsi/mac53c94.c state->intr, node); node 1574 drivers/scsi/qla2xxx/qla_target.c void *node; node 1602 drivers/scsi/qla2xxx/qla_target.c btree_for_each_safe64(&tgt->lun_qpair_map, key, node) node 4167 drivers/scsi/qla2xxx/qla_target.c void *node; node 4176 drivers/scsi/qla2xxx/qla_target.c btree_for_each_safe64(&tgt->lun_qpair_map, key, node) node 6446 drivers/scsi/qla2xxx/qla_target.c struct scsi_qla_host *node; node 6449 drivers/scsi/qla2xxx/qla_target.c btree_for_each_safe32(&ha->tgt.host_map, key, node) node 768 drivers/scsi/qla2xxx/tcm_qla2xxx.c void *node; node 772 drivers/scsi/qla2xxx/tcm_qla2xxx.c node = btree_remove32(&lport->lport_fcport_map, nacl->nport_id); node 773 drivers/scsi/qla2xxx/tcm_qla2xxx.c if (WARN_ON(node && (node != se_nacl))) { node 781 drivers/scsi/qla2xxx/tcm_qla2xxx.c node, GFP_ATOMIC); node 1652 drivers/scsi/qla2xxx/tcm_qla2xxx.c struct se_node_acl *node; node 1666 drivers/scsi/qla2xxx/tcm_qla2xxx.c btree_for_each_safe32(&lport->lport_fcport_map, key, node) node 33 drivers/scsi/raid_class.c struct list_head node; node 101 drivers/scsi/raid_class.c list_for_each_entry_safe(rc, next, &rd->component_list, node) { node 102 drivers/scsi/raid_class.c list_del(&rc->node); node 235 drivers/scsi/raid_class.c INIT_LIST_HEAD(&rc->node); node 242 drivers/scsi/raid_class.c list_add_tail(&rc->node, &rd->component_list); node 251 drivers/scsi/raid_class.c list_del(&rc->node); node 30 drivers/scsi/scsi_devinfo.c struct list_head node; /* our node for being on the master list */ node 270 drivers/scsi/scsi_devinfo.c list_for_each_entry(devinfo_table, &scsi_dev_info_list, node) node 627 drivers/scsi/scsi_devinfo.c list_entry(dl->top, struct scsi_dev_info_list_table, node); node 652 drivers/scsi/scsi_devinfo.c node); node 666 drivers/scsi/scsi_devinfo.c list_entry(dl->top, struct scsi_dev_info_list_table, node); node 678 drivers/scsi/scsi_devinfo.c node); node 793 drivers/scsi/scsi_devinfo.c INIT_LIST_HEAD(&devinfo_table->node); node 797 drivers/scsi/scsi_devinfo.c list_add_tail(&devinfo_table->node, &scsi_dev_info_list); node 822 drivers/scsi/scsi_devinfo.c list_del(&devinfo_table->node); node 2436 drivers/scsi/scsi_lib.c evt = list_entry(this, struct scsi_event, node); node 2437 drivers/scsi/scsi_lib.c list_del(&evt->node); node 2466 drivers/scsi/scsi_lib.c list_add_tail(&evt->node, &sdev->event_list); node 2487 drivers/scsi/scsi_lib.c INIT_LIST_HEAD(&evt->node); node 459 drivers/scsi/scsi_sysfs.c evt = list_entry(this, struct scsi_event, node); node 460 drivers/scsi/scsi_sysfs.c list_del(&evt->node); node 7863 drivers/scsi/smartpqi/smartpqi_init.c int node, cp_node; node 7880 drivers/scsi/smartpqi/smartpqi_init.c node = dev_to_node(&pci_dev->dev); node 7881 drivers/scsi/smartpqi/smartpqi_init.c if (node == NUMA_NO_NODE) { node 7888 drivers/scsi/smartpqi/smartpqi_init.c ctrl_info = pqi_alloc_ctrl_info(node); node 79 drivers/scsi/virtio_scsi.c struct hlist_node node; node 421 drivers/sh/clk/core.c if (clk->node.next || clk->node.prev) node 438 drivers/sh/clk/core.c list_add(&clk->node, &clock_list); node 456 drivers/sh/clk/core.c list_del(&clk->node); node 466 drivers/sh/clk/core.c list_for_each_entry(clkp, &clock_list, node) node 576 drivers/sh/clk/core.c list_for_each_entry(clkp, &clock_list, node) { node 613 drivers/sh/clk/core.c list_for_each_entry(clk, &clock_list, node) node 153 drivers/siox/siox-core.c list_for_each_entry(sdevice, &smaster->devices, node) { node 194 drivers/siox/siox-core.c list_for_each_entry(sdevice, &smaster->devices, node) { node 362 drivers/siox/siox-core.c list_for_each_entry(sdevice, &smaster->devices, node) { node 750 drivers/siox/siox-core.c struct siox_device, node); node 751 drivers/siox/siox-core.c list_del(&sdevice->node); node 818 drivers/siox/siox-core.c list_add_tail(&sdevice->node, &smaster->devices); node 858 drivers/siox/siox-core.c sdevice = container_of(smaster->devices.prev, struct siox_device, node); node 859 drivers/siox/siox-core.c list_del(&sdevice->node); node 156 drivers/slimbus/core.c struct device_node *node) node 166 drivers/slimbus/core.c if (node) node 167 drivers/slimbus/core.c sbdev->dev.of_node = of_node_get(node); node 180 drivers/slimbus/core.c struct device_node *node) node 190 drivers/slimbus/core.c ret = slim_add_device(ctrl, sbdev, node); node 202 drivers/slimbus/core.c struct device_node *node; node 207 drivers/slimbus/core.c for_each_child_of_node(ctrl->dev->of_node, node) { node 214 drivers/slimbus/core.c compat = of_get_property(node, "compatible", NULL); node 225 drivers/slimbus/core.c ret = of_property_read_u32_array(node, "reg", reg, 2); node 237 drivers/slimbus/core.c sbdev = slim_alloc_device(ctrl, &e_addr, node); node 1137 drivers/slimbus/qcom-ngd-ctrl.c struct device_node *node; node 1139 drivers/slimbus/qcom-ngd-ctrl.c for_each_child_of_node(ctrl->ngd->pdev->dev.of_node, node) { node 1140 drivers/slimbus/qcom-ngd-ctrl.c sbdev = of_slim_get_device(&ctrl->ctrl, node); node 1263 drivers/slimbus/qcom-ngd-ctrl.c qmi->svc_info.sq_node = service->node; node 1331 drivers/slimbus/qcom-ngd-ctrl.c struct device_node *node; node 1336 drivers/slimbus/qcom-ngd-ctrl.c for_each_available_child_of_node(parent->of_node, node) { node 1337 drivers/slimbus/qcom-ngd-ctrl.c if (of_property_read_u32(node, "reg", &id)) node 1342 drivers/slimbus/qcom-ngd-ctrl.c of_node_put(node); node 1349 drivers/slimbus/qcom-ngd-ctrl.c of_node_put(node); node 1355 drivers/slimbus/qcom-ngd-ctrl.c ngd->pdev->dev.of_node = node; node 343 drivers/slimbus/slimbus.h struct list_head node; node 116 drivers/slimbus/stream.c list_add_tail(&rt->node, &dev->stream_list); node 469 drivers/slimbus/stream.c list_del(&stream->node); node 200 drivers/soc/aspeed/aspeed-lpc-ctrl.c struct device_node *node; node 212 drivers/soc/aspeed/aspeed-lpc-ctrl.c node = of_parse_phandle(dev->of_node, "flash", 0); node 213 drivers/soc/aspeed/aspeed-lpc-ctrl.c if (!node) { node 216 drivers/soc/aspeed/aspeed-lpc-ctrl.c rc = of_address_to_resource(node, 1, &resm); node 217 drivers/soc/aspeed/aspeed-lpc-ctrl.c of_node_put(node); node 231 drivers/soc/aspeed/aspeed-lpc-ctrl.c node = of_parse_phandle(dev->of_node, "memory-region", 0); node 232 drivers/soc/aspeed/aspeed-lpc-ctrl.c if (!node) { node 235 drivers/soc/aspeed/aspeed-lpc-ctrl.c rc = of_address_to_resource(node, 0, &resm); node 236 drivers/soc/aspeed/aspeed-lpc-ctrl.c of_node_put(node); node 337 drivers/soc/aspeed/aspeed-p2a-ctrl.c struct device_node *node; node 349 drivers/soc/aspeed/aspeed-p2a-ctrl.c node = of_parse_phandle(dev->of_node, "memory-region", 0); node 350 drivers/soc/aspeed/aspeed-p2a-ctrl.c if (node) { node 351 drivers/soc/aspeed/aspeed-p2a-ctrl.c rc = of_address_to_resource(node, 0, &resm); node 352 drivers/soc/aspeed/aspeed-p2a-ctrl.c of_node_put(node); node 103 drivers/soc/fsl/dpaa2-console.c static int dpaa2_generic_console_open(struct inode *node, struct file *fp, node 168 drivers/soc/fsl/dpaa2-console.c static int dpaa2_mc_console_open(struct inode *node, struct file *fp) node 170 drivers/soc/fsl/dpaa2-console.c return dpaa2_generic_console_open(node, fp, node 175 drivers/soc/fsl/dpaa2-console.c static int dpaa2_aiop_console_open(struct inode *node, struct file *fp) node 177 drivers/soc/fsl/dpaa2-console.c return dpaa2_generic_console_open(node, fp, node 182 drivers/soc/fsl/dpaa2-console.c static int dpaa2_console_close(struct inode *node, struct file *fp) node 24 drivers/soc/fsl/dpio/dpio-service.c struct list_head node; node 74 drivers/soc/fsl/dpio/dpio-service.c d = list_entry(dpio_list.next, struct dpaa2_io, node); node 75 drivers/soc/fsl/dpio/dpio-service.c list_del(&d->node); node 76 drivers/soc/fsl/dpio/dpio-service.c list_add_tail(&d->node, &dpio_list); node 134 drivers/soc/fsl/dpio/dpio-service.c INIT_LIST_HEAD(&obj->node); node 147 drivers/soc/fsl/dpio/dpio-service.c list_add_tail(&obj->node, &dpio_list); node 170 drivers/soc/fsl/dpio/dpio-service.c list_del(&d->node); node 273 drivers/soc/fsl/dpio/dpio-service.c list_add(&ctx->node, &d->notifications); node 305 drivers/soc/fsl/dpio/dpio-service.c list_del(&ctx->node); node 212 drivers/soc/fsl/qbman/bman_ccsr.c struct device_node *node = dev->of_node; node 222 drivers/soc/fsl/qbman/bman_ccsr.c node); node 264 drivers/soc/fsl/qbman/bman_ccsr.c dev_info(dev, "Can't get %pOF IRQ\n", node); node 271 drivers/soc/fsl/qbman/bman_ccsr.c ret, node); node 100 drivers/soc/fsl/qbman/bman_portal.c struct device_node *node = dev->of_node; node 124 drivers/soc/fsl/qbman/bman_portal.c dev_err(dev, "Can't get %pOF property 'reg::CE'\n", node); node 131 drivers/soc/fsl/qbman/bman_portal.c dev_err(dev, "Can't get %pOF property 'reg::CI'\n", node); node 1481 drivers/soc/fsl/qbman/qman.c list_for_each_entry(cgr, &p->cgr_cbs, node) node 2464 drivers/soc/fsl/qbman/qman.c list_add(&cgr->node, &p->cgr_cbs); node 2504 drivers/soc/fsl/qbman/qman.c list_del(&cgr->node); node 2509 drivers/soc/fsl/qbman/qman.c list_for_each_entry(i, &p->cgr_cbs, node) node 2515 drivers/soc/fsl/qbman/qman.c list_add(&cgr->node, &p->cgr_cbs); node 2526 drivers/soc/fsl/qbman/qman.c list_add(&cgr->node, &p->cgr_cbs); node 755 drivers/soc/fsl/qbman/qman_ccsr.c struct device_node *node = dev->of_node; node 766 drivers/soc/fsl/qbman/qman_ccsr.c node); node 844 drivers/soc/fsl/qbman/qman_ccsr.c node); node 851 drivers/soc/fsl/qbman/qman_ccsr.c ret, node); node 233 drivers/soc/fsl/qbman/qman_portal.c struct device_node *node = dev->of_node; node 258 drivers/soc/fsl/qbman/qman_portal.c dev_err(dev, "Can't get %pOF property 'reg::CE'\n", node); node 265 drivers/soc/fsl/qbman/qman_portal.c dev_err(dev, "Can't get %pOF property 'reg::CI'\n", node); node 269 drivers/soc/fsl/qbman/qman_portal.c err = of_property_read_u32(node, "cell-index", &val); node 271 drivers/soc/fsl/qbman/qman_portal.c dev_err(dev, "Can't get %pOF property 'cell-index'\n", node); node 154 drivers/soc/fsl/qbman/qman_test_stash.c struct list_head node; node 165 drivers/soc/fsl/qbman/qman_test_stash.c struct list_head node; node 323 drivers/soc/fsl/qbman/qman_test_stash.c list_add_tail(&hp_cpu->node, &hp_cpu_list); node 337 drivers/soc/fsl/qbman/qman_test_stash.c list_add_tail(&handler->node, &hp_cpu->handlers); node 348 drivers/soc/fsl/qbman/qman_test_stash.c list_del(&hp_cpu->node); node 353 drivers/soc/fsl/qbman/qman_test_stash.c node); node 368 drivers/soc/fsl/qbman/qman_test_stash.c list_del(&handler->node); node 448 drivers/soc/fsl/qbman/qman_test_stash.c list_for_each_entry(hp_cpu, &hp_cpu_list, node) { node 454 drivers/soc/fsl/qbman/qman_test_stash.c struct hp_handler, node); node 457 drivers/soc/fsl/qbman/qman_test_stash.c hp_cpu->iterator->node.next, node 458 drivers/soc/fsl/qbman/qman_test_stash.c struct hp_handler, node); node 476 drivers/soc/fsl/qbman/qman_test_stash.c hp_cpu = list_first_entry(&hp_cpu_list, struct hp_cpu, node); node 477 drivers/soc/fsl/qbman/qman_test_stash.c handler = list_first_entry(&hp_cpu->handlers, struct hp_handler, node); node 493 drivers/soc/fsl/qbman/qman_test_stash.c list_for_each_entry(hp_cpu, &hp_cpu_list, node) { node 497 drivers/soc/fsl/qbman/qman_test_stash.c struct hp_handler, node); node 500 drivers/soc/fsl/qbman/qman_test_stash.c hp_cpu->iterator->node.next, node 501 drivers/soc/fsl/qbman/qman_test_stash.c struct hp_handler, node); node 244 drivers/soc/fsl/qe/qe_ic.c static int qe_ic_host_match(struct irq_domain *h, struct device_node *node, node 249 drivers/soc/fsl/qe/qe_ic.c return of_node == NULL || of_node == node; node 316 drivers/soc/fsl/qe/qe_ic.c void __init qe_ic_init(struct device_node *node, unsigned int flags, node 324 drivers/soc/fsl/qe/qe_ic.c ret = of_address_to_resource(node, 0, &res); node 332 drivers/soc/fsl/qe/qe_ic.c qe_ic->irqhost = irq_domain_add_linear(node, NR_QE_IC_INTS, node 343 drivers/soc/fsl/qe/qe_ic.c qe_ic->virq_high = irq_of_parse_and_map(node, 0); node 344 drivers/soc/fsl/qe/qe_ic.c qe_ic->virq_low = irq_of_parse_and_map(node, 1); node 30 drivers/soc/qcom/apr.c struct list_head node; node 93 drivers/soc/qcom/apr.c list_add_tail(&abuf->node, &apr->rx_list); node 177 drivers/soc/qcom/apr.c list_for_each_entry_safe(abuf, b, &apr->rx_list, node) { node 180 drivers/soc/qcom/apr.c list_del(&abuf->node); node 306 drivers/soc/qcom/apr.c struct device_node *node; node 308 drivers/soc/qcom/apr.c for_each_child_of_node(dev->of_node, node) { node 311 drivers/soc/qcom/apr.c if (of_property_read_u32(node, "reg", &id.svc_id)) node 316 drivers/soc/qcom/apr.c if (apr_add_device(dev, node, &id)) node 453 drivers/soc/qcom/qcom_aoss.c struct device_node *node) node 455 drivers/soc/qcom/qcom_aoss.c char *cdev_name = (char *)node->name; node 461 drivers/soc/qcom/qcom_aoss.c (qmp->dev, node, node 127 drivers/soc/qcom/qcom_gsbi.c struct device_node *node = pdev->dev.of_node; node 148 drivers/soc/qcom/qcom_gsbi.c gsbi->tcsr = syscon_regmap_lookup_by_phandle(node, "syscon-tcsr"); node 151 drivers/soc/qcom/qcom_gsbi.c tcsr_node = of_parse_phandle(node, "syscon-tcsr", 0); node 163 drivers/soc/qcom/qcom_gsbi.c if (of_property_read_u32(node, "cell-index", &gsbi_num)) { node 173 drivers/soc/qcom/qcom_gsbi.c if (of_property_read_u32(node, "qcom,mode", &gsbi->mode)) { node 179 drivers/soc/qcom/qcom_gsbi.c of_property_read_u32(node, "qcom,crci", &gsbi->crci); node 216 drivers/soc/qcom/qcom_gsbi.c ret = of_platform_populate(node, NULL, NULL, &pdev->dev); node 33 drivers/soc/qcom/qmi_interface.c unsigned int node, unsigned int port) node 43 drivers/soc/qcom/qmi_interface.c if (!node && !port) node 53 drivers/soc/qcom/qmi_interface.c svc->node = node; node 73 drivers/soc/qcom/qmi_interface.c unsigned int node, unsigned int port) node 80 drivers/soc/qcom/qmi_interface.c if (node != -1 && svc->node != node) node 103 drivers/soc/qcom/qmi_interface.c unsigned int node) node 107 drivers/soc/qcom/qmi_interface.c qmi_recv_del_server(qmi, node, -1); node 110 drivers/soc/qcom/qmi_interface.c ops->bye(qmi, node); node 122 drivers/soc/qcom/qmi_interface.c unsigned int node, unsigned int port) node 127 drivers/soc/qcom/qmi_interface.c ops->del_client(qmi, node, port); node 142 drivers/soc/qcom/qmi_interface.c qmi_recv_bye(qmi, le32_to_cpu(pkt->client.node)); node 148 drivers/soc/qcom/qmi_interface.c le32_to_cpu(pkt->server.node), node 153 drivers/soc/qcom/qmi_interface.c le32_to_cpu(pkt->server.node), node 158 drivers/soc/qcom/qmi_interface.c le32_to_cpu(pkt->client.node), node 239 drivers/soc/qcom/qmi_interface.c pkt.server.node = cpu_to_le32(qmi->sq.sq_node); node 173 drivers/soc/qcom/rmtfs_mem.c struct device_node *node = pdev->dev.of_node; node 181 drivers/soc/qcom/rmtfs_mem.c rmem = of_reserved_mem_lookup(node); node 187 drivers/soc/qcom/rmtfs_mem.c ret = of_property_read_u32(node, "qcom,client-id", &client_id); node 229 drivers/soc/qcom/rmtfs_mem.c ret = of_property_read_u32(node, "qcom,vmid", &vmid); node 91 drivers/soc/qcom/smp2p.c struct list_head node; node 202 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) { node 213 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) { node 307 drivers/soc/qcom/smp2p.c struct device_node *node) node 309 drivers/soc/qcom/smp2p.c entry->domain = irq_domain_add_linear(node, 32, &smp2p_irq_ops, entry); node 343 drivers/soc/qcom/smp2p.c struct device_node *node) node 357 drivers/soc/qcom/smp2p.c entry->state = qcom_smem_state_register(node, &smp2p_state_ops, entry); node 444 drivers/soc/qcom/smp2p.c struct device_node *node; node 500 drivers/soc/qcom/smp2p.c for_each_available_child_of_node(pdev->dev.of_node, node) { node 510 drivers/soc/qcom/smp2p.c ret = of_property_read_string(node, "qcom,entry-name", &entry->name); node 514 drivers/soc/qcom/smp2p.c if (of_property_read_bool(node, "interrupt-controller")) { node 515 drivers/soc/qcom/smp2p.c ret = qcom_smp2p_inbound_entry(smp2p, entry, node); node 519 drivers/soc/qcom/smp2p.c list_add(&entry->node, &smp2p->inbound); node 521 drivers/soc/qcom/smp2p.c ret = qcom_smp2p_outbound_entry(smp2p, entry, node); node 525 drivers/soc/qcom/smp2p.c list_add(&entry->node, &smp2p->outbound); node 545 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) node 548 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->outbound, node) node 568 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) node 571 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->outbound, node) node 346 drivers/soc/qcom/smsm.c struct device_node *node = smsm->dev->of_node; node 352 drivers/soc/qcom/smsm.c syscon = of_parse_phandle(node, key, 0); node 360 drivers/soc/qcom/smsm.c ret = of_property_read_u32_index(node, key, 1, &host->ipc_offset); node 366 drivers/soc/qcom/smsm.c ret = of_property_read_u32_index(node, key, 2, &host->ipc_bit); node 383 drivers/soc/qcom/smsm.c struct device_node *node) node 388 drivers/soc/qcom/smsm.c irq = irq_of_parse_and_map(node, 0); node 403 drivers/soc/qcom/smsm.c entry->domain = irq_domain_add_linear(node, 32, &smsm_irq_ops, entry); node 458 drivers/soc/qcom/smsm.c struct device_node *node; node 551 drivers/soc/qcom/smsm.c for_each_available_child_of_node(pdev->dev.of_node, node) { node 552 drivers/soc/qcom/smsm.c if (!of_property_read_bool(node, "interrupt-controller")) node 555 drivers/soc/qcom/smsm.c ret = of_property_read_u32(node, "reg", &id); node 571 drivers/soc/qcom/smsm.c ret = smsm_inbound_entry(smsm, entry, node); node 382 drivers/soc/rockchip/pm_domains.c struct device_node *node) node 391 drivers/soc/rockchip/pm_domains.c error = of_property_read_u32(node, "reg", &id); node 395 drivers/soc/rockchip/pm_domains.c node, error); node 401 drivers/soc/rockchip/pm_domains.c node, id); node 408 drivers/soc/rockchip/pm_domains.c node, id); node 419 drivers/soc/rockchip/pm_domains.c pd->num_clks = of_clk_get_parent_count(node); node 427 drivers/soc/rockchip/pm_domains.c node, pd->num_clks); node 432 drivers/soc/rockchip/pm_domains.c pd->clks[i].clk = of_clk_get(node, i); node 437 drivers/soc/rockchip/pm_domains.c node, i, error); node 446 drivers/soc/rockchip/pm_domains.c pd->num_qos = of_count_phandle_with_args(node, "pm_qos", node 470 drivers/soc/rockchip/pm_domains.c qos_node = of_parse_phandle(node, "pm_qos", j); node 489 drivers/soc/rockchip/pm_domains.c node, error); node 493 drivers/soc/rockchip/pm_domains.c pd->genpd.name = node->name; node 623 drivers/soc/rockchip/pm_domains.c struct device_node *node; node 682 drivers/soc/rockchip/pm_domains.c for_each_available_child_of_node(np, node) { node 683 drivers/soc/rockchip/pm_domains.c error = rockchip_pm_add_one_domain(pmu, node); node 686 drivers/soc/rockchip/pm_domains.c node, error); node 687 drivers/soc/rockchip/pm_domains.c of_node_put(node); node 691 drivers/soc/rockchip/pm_domains.c error = rockchip_pm_add_subdomain(pmu, node); node 694 drivers/soc/rockchip/pm_domains.c node, error); node 695 drivers/soc/rockchip/pm_domains.c of_node_put(node); node 95 drivers/soc/samsung/pm_domains.c static __init const char *exynos_get_domain_name(struct device_node *node) node 99 drivers/soc/samsung/pm_domains.c if (of_property_read_string(node, "label", &name) < 0) node 100 drivers/soc/samsung/pm_domains.c name = kbasename(node->full_name); node 165 drivers/soc/sunxi/sunxi_sram.c static const struct sunxi_sram_data *sunxi_sram_of_parse(struct device_node *node, node 175 drivers/soc/sunxi/sunxi_sram.c ret = of_parse_phandle_with_fixed_args(node, "allwinner,sram", 1, 0, node 537 drivers/soc/ti/knav_dma.c struct device_node *node, node 545 drivers/soc/ti/knav_dma.c ret = of_address_to_resource(node, index, &res); node 548 drivers/soc/ti/knav_dma.c node, index); node 555 drivers/soc/ti/knav_dma.c index, node); node 623 drivers/soc/ti/knav_dma.c struct device_node *node = dma_node; node 658 drivers/soc/ti/knav_dma.c dma->reg_global = pktdma_get_regs(dma, node, 0, &size); node 666 drivers/soc/ti/knav_dma.c dma->reg_tx_chan = pktdma_get_regs(dma, node, 1, &size); node 671 drivers/soc/ti/knav_dma.c dma->reg_rx_chan = pktdma_get_regs(dma, node, 2, &size); node 676 drivers/soc/ti/knav_dma.c dma->reg_tx_sched = pktdma_get_regs(dma, node, 3, &size); node 681 drivers/soc/ti/knav_dma.c dma->reg_rx_flow = pktdma_get_regs(dma, node, 4, &size); node 689 drivers/soc/ti/knav_dma.c dma->enable_all = (of_get_property(node, "ti,enable-all", NULL) != NULL); node 690 drivers/soc/ti/knav_dma.c dma->loopback = (of_get_property(node, "ti,loop-back", NULL) != NULL); node 692 drivers/soc/ti/knav_dma.c ret = of_property_read_u32(node, "ti,rx-retry-timeout", &timeout); node 704 drivers/soc/ti/knav_dma.c strcpy(dma->name, node->name); node 740 drivers/soc/ti/knav_dma.c struct device_node *node = pdev->dev.of_node; node 744 drivers/soc/ti/knav_dma.c if (!node) { node 767 drivers/soc/ti/knav_dma.c for_each_child_of_node(node, child) { node 768 drivers/soc/ti/knav_dma.c ret = dma_init(node, child); node 383 drivers/soc/ti/knav_qmss.h struct device_node *node, node 472 drivers/soc/ti/knav_qmss_acc.c struct device_node *node, node 487 drivers/soc/ti/knav_qmss_acc.c ret = of_property_read_u32_array(node, "accumulator", config, 5); node 524 drivers/soc/ti/knav_qmss_acc.c if (of_get_property(node, "multi-queue", NULL)) { node 769 drivers/soc/ti/knav_qmss_queue.c struct list_head *node; node 827 drivers/soc/ti/knav_qmss_queue.c node = ®ion->pools; node 835 drivers/soc/ti/knav_qmss_queue.c node = &pi->region_inst; node 843 drivers/soc/ti/knav_qmss_queue.c list_add_tail(&pool->region_inst, node); node 1073 drivers/soc/ti/knav_qmss_queue.c static const char *knav_queue_find_name(struct device_node *node) node 1077 drivers/soc/ti/knav_qmss_queue.c if (of_property_read_string(node, "label", &name) < 0) node 1078 drivers/soc/ti/knav_qmss_queue.c name = node->name; node 1146 drivers/soc/ti/knav_qmss_queue.c struct device_node *node = pdev->dev.of_node; node 1162 drivers/soc/ti/knav_qmss_queue.c if (!of_property_read_u32_array(node, name , temp, 2)) { node 1217 drivers/soc/ti/knav_qmss_queue.c struct device_node *node) node 1232 drivers/soc/ti/knav_qmss_queue.c range->name = knav_queue_find_name(node); node 1233 drivers/soc/ti/knav_qmss_queue.c ret = of_property_read_u32_array(node, "qrange", temp, 2); node 1246 drivers/soc/ti/knav_qmss_queue.c if (of_irq_parse_one(node, i, &oirq)) node 1274 drivers/soc/ti/knav_qmss_queue.c if (of_get_property(node, "qalloc-by-id", NULL)) node 1277 drivers/soc/ti/knav_qmss_queue.c if (of_get_property(node, "accumulator", NULL)) { node 1278 drivers/soc/ti/knav_qmss_queue.c ret = knav_init_acc_range(kdev, node, range); node 1377 drivers/soc/ti/knav_qmss_queue.c struct device_node *node, int index) node 1383 drivers/soc/ti/knav_qmss_queue.c ret = of_address_to_resource(node, index, &res); node 1386 drivers/soc/ti/knav_qmss_queue.c node, index); node 1393 drivers/soc/ti/knav_qmss_queue.c index, node); node 1761 drivers/soc/ti/knav_qmss_queue.c struct device_node *node = pdev->dev.of_node; node 1768 drivers/soc/ti/knav_qmss_queue.c if (!node) { node 1798 drivers/soc/ti/knav_qmss_queue.c if (of_property_read_u32_array(node, "queue-range", temp, 2)) { node 1807 drivers/soc/ti/knav_qmss_queue.c qmgrs = of_get_child_by_name(node, "qmgrs"); node 1819 drivers/soc/ti/knav_qmss_queue.c pdsps = of_get_child_by_name(node, "pdsps"); node 1832 drivers/soc/ti/knav_qmss_queue.c queue_pools = of_get_child_by_name(node, "queue-pools"); node 1861 drivers/soc/ti/knav_qmss_queue.c regions = of_get_child_by_name(node, "descriptor-regions"); node 123 drivers/soundwire/bus.c list_del_init(&slave->node); node 415 drivers/soundwire/bus.c list_for_each_entry(slave, &bus->slaves, node) { node 559 drivers/soundwire/bus.c list_for_each_entry_safe(slave, _s, &bus->slaves, node) { node 166 drivers/soundwire/mipi_disco.c struct fwnode_handle *node; node 182 drivers/soundwire/mipi_disco.c node = device_get_named_child_node(&slave->dev, name); node 183 drivers/soundwire/mipi_disco.c if (!node) { node 188 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-port-max-wordlength", node 190 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-port-min-wordlength", node 193 drivers/soundwire/mipi_disco.c nval = fwnode_property_count_u32(node, "mipi-sdw-port-wordlength-configs"); node 203 drivers/soundwire/mipi_disco.c fwnode_property_read_u32_array(node, node 208 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-data-port-type", node 211 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, node 215 drivers/soundwire/mipi_disco.c dpn[i].simple_ch_prep_sm = fwnode_property_read_bool(node, node 218 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, node 222 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, node 226 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-min-channel-number", node 229 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-max-channel-number", node 232 drivers/soundwire/mipi_disco.c nval = fwnode_property_count_u32(node, "mipi-sdw-channel-number-list"); node 241 drivers/soundwire/mipi_disco.c fwnode_property_read_u32_array(node, node 246 drivers/soundwire/mipi_disco.c nval = fwnode_property_count_u32(node, "mipi-sdw-channel-combination-list"); node 256 drivers/soundwire/mipi_disco.c fwnode_property_read_u32_array(node, node 262 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, node 265 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-max-async-buffer", node 268 drivers/soundwire/mipi_disco.c dpn[i].block_pack_mode = fwnode_property_read_bool(node, node 271 drivers/soundwire/mipi_disco.c fwnode_property_read_u32(node, "mipi-sdw-port-encoding-type", node 45 drivers/soundwire/slave.c list_add_tail(&slave->node, &bus->slaves); node 57 drivers/soundwire/slave.c list_del(&slave->node); node 83 drivers/soundwire/slave.c list_for_each_entry(adev, &parent->children, node) { node 128 drivers/soundwire/slave.c struct device_node *node; node 130 drivers/soundwire/slave.c for_each_child_of_node(bus->dev->of_node, node) { node 137 drivers/soundwire/slave.c compat = of_get_property(node, "compatible", NULL); node 150 drivers/soundwire/slave.c addr = of_get_property(node, "reg", &len); node 164 drivers/soundwire/slave.c sdw_slave_add(bus, &id, of_fwnode_handle(node)); node 819 drivers/spi/spi-davinci.c struct device_node *node = pdev->dev.of_node; node 843 drivers/spi/spi-davinci.c of_property_read_u32(node, "num-cs", &num_cs); node 845 drivers/spi/spi-davinci.c of_property_read_u32(node, "ti,davinci-spi-intr-line", &intr_line); node 142 drivers/spi/spi-jcore.c struct device_node *node = pdev->dev.of_node; node 159 drivers/spi/spi-jcore.c master->dev.of_node = node; node 140 drivers/spi/spi-omap2-mcspi.c struct list_head node; node 1055 drivers/spi/spi-omap2-mcspi.c list_add_tail(&cs->node, &ctx->cs); node 1089 drivers/spi/spi-omap2-mcspi.c list_del(&cs->node); node 1275 drivers/spi/spi-omap2-mcspi.c list_for_each_entry(cs, &ctx->cs, node) { node 1349 drivers/spi/spi-omap2-mcspi.c list_for_each_entry(cs, &ctx->cs, node) { node 1399 drivers/spi/spi-omap2-mcspi.c struct device_node *node = pdev->dev.of_node; node 1402 drivers/spi/spi-omap2-mcspi.c if (of_property_read_bool(node, "spi-slave")) node 1420 drivers/spi/spi-omap2-mcspi.c master->dev.of_node = node; node 1434 drivers/spi/spi-omap2-mcspi.c of_property_read_u32(node, "ti,spi-num-cs", &num_cs); node 1436 drivers/spi/spi-omap2-mcspi.c if (of_get_property(node, "ti,pindir-d0-out-d1-in", NULL)) node 3659 drivers/spi/spi.c struct spi_device *of_find_spi_device_by_node(struct device_node *node) node 3661 drivers/spi/spi.c struct device *dev = bus_find_device_by_of_node(&spi_bus_type, node); node 3670 drivers/spi/spi.c static struct spi_controller *of_find_spi_controller_by_node(struct device_node *node) node 3674 drivers/spi/spi.c dev = class_find_device_by_of_node(&spi_master_class, node); node 3676 drivers/spi/spi.c dev = class_find_device_by_of_node(&spi_slave_class, node); node 451 drivers/spmi/spmi.c struct device_node *node; node 457 drivers/spmi/spmi.c for_each_available_child_of_node(ctrl->dev.of_node, node) { node 461 drivers/spmi/spmi.c dev_dbg(&ctrl->dev, "adding child %pOF\n", node); node 463 drivers/spmi/spmi.c err = of_property_read_u32_array(node, "reg", reg, 2); node 467 drivers/spmi/spmi.c node, err); node 474 drivers/spmi/spmi.c node); node 479 drivers/spmi/spmi.c dev_err(&ctrl->dev, "invalid usid on node %pOF\n", node); node 489 drivers/spmi/spmi.c sdev->dev.of_node = node; node 379 drivers/staging/android/ion/ion.c plist_for_each_entry(heap, &dev->heaps, node) { node 435 drivers/staging/android/ion/ion.c plist_for_each_entry(heap, &dev->heaps, node) { node 634 drivers/staging/android/ion/ion.c plist_node_init(&heap->node, -heap->id); node 635 drivers/staging/android/ion/ion.c plist_add(&heap->node, &dev->heaps); node 140 drivers/staging/android/ion/ion.h struct plist_node node; node 349 drivers/staging/android/vsoc.c struct fd_scoped_permission_node *node) node 351 drivers/staging/android/vsoc.c if (node) { node 353 drivers/staging/android/vsoc.c &node->permission); node 355 drivers/staging/android/vsoc.c list_del(&node->list); node 357 drivers/staging/android/vsoc.c kfree(node); node 545 drivers/staging/android/vsoc.c struct fd_scoped_permission_node *node = NULL; node 547 drivers/staging/android/vsoc.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 549 drivers/staging/android/vsoc.c if (!node) node 551 drivers/staging/android/vsoc.c INIT_LIST_HEAD(&node->list); node 554 drivers/staging/android/vsoc.c node, node 558 drivers/staging/android/vsoc.c list_add(&node->list, &vsoc_dev.permissions); node 561 drivers/staging/android/vsoc.c kfree(node); node 569 drivers/staging/android/vsoc.c struct fd_scoped_permission_node *node = node 572 drivers/staging/android/vsoc.c if (!node) node 576 drivers/staging/android/vsoc.c &node->permission, sizeof(node->permission))) node 1059 drivers/staging/android/vsoc.c struct fd_scoped_permission_node *node = NULL; node 1069 drivers/staging/android/vsoc.c node = private_data->fd_scoped_permission_node; node 1070 drivers/staging/android/vsoc.c if (node) { node 1076 drivers/staging/android/vsoc.c do_destroy_fd_scoped_permission_node(owner_region_p, node); node 909 drivers/staging/fbtft/fbtft-core.c struct device_node *node = par->info->device->of_node; node 915 drivers/staging/fbtft/fbtft-core.c if (!node) node 918 drivers/staging/fbtft/fbtft-core.c prop = of_find_property(node, "init", NULL); node 1142 drivers/staging/fbtft/fbtft-core.c static u32 fbtft_of_value(struct device_node *node, const char *propname) node 1147 drivers/staging/fbtft/fbtft-core.c ret = of_property_read_u32(node, propname, &val); node 1156 drivers/staging/fbtft/fbtft-core.c struct device_node *node = dev->of_node; node 1159 drivers/staging/fbtft/fbtft-core.c if (!node) { node 1168 drivers/staging/fbtft/fbtft-core.c pdata->display.width = fbtft_of_value(node, "width"); node 1169 drivers/staging/fbtft/fbtft-core.c pdata->display.height = fbtft_of_value(node, "height"); node 1170 drivers/staging/fbtft/fbtft-core.c pdata->display.regwidth = fbtft_of_value(node, "regwidth"); node 1171 drivers/staging/fbtft/fbtft-core.c pdata->display.buswidth = fbtft_of_value(node, "buswidth"); node 1172 drivers/staging/fbtft/fbtft-core.c pdata->display.backlight = fbtft_of_value(node, "backlight"); node 1173 drivers/staging/fbtft/fbtft-core.c pdata->display.bpp = fbtft_of_value(node, "bpp"); node 1174 drivers/staging/fbtft/fbtft-core.c pdata->display.debug = fbtft_of_value(node, "debug"); node 1175 drivers/staging/fbtft/fbtft-core.c pdata->rotate = fbtft_of_value(node, "rotate"); node 1176 drivers/staging/fbtft/fbtft-core.c pdata->bgr = of_property_read_bool(node, "bgr"); node 1177 drivers/staging/fbtft/fbtft-core.c pdata->fps = fbtft_of_value(node, "fps"); node 1178 drivers/staging/fbtft/fbtft-core.c pdata->txbuflen = fbtft_of_value(node, "txbuflen"); node 1179 drivers/staging/fbtft/fbtft-core.c pdata->startbyte = fbtft_of_value(node, "startbyte"); node 1180 drivers/staging/fbtft/fbtft-core.c of_property_read_string(node, "gamma", (const char **)&pdata->gamma); node 1182 drivers/staging/fbtft/fbtft-core.c if (of_find_property(node, "led-gpios", NULL)) node 1184 drivers/staging/fbtft/fbtft-core.c if (of_find_property(node, "init", NULL)) node 1250 drivers/staging/fieldbus/anybuss/host.c struct device_node *node; node 1254 drivers/staging/fieldbus/anybuss/host.c for_each_child_of_node(dev->of_node, node) { node 1255 drivers/staging/fieldbus/anybuss/host.c if (anybus_of_get_host_idx(node) == host_idx) node 1256 drivers/staging/fieldbus/anybuss/host.c return node; node 187 drivers/staging/fieldbus/dev_core.c static int fieldbus_release(struct inode *node, struct file *filp) node 30 drivers/staging/greybus/authentication.c struct list_head node; node 69 drivers/staging/greybus/authentication.c list_for_each_entry(cap, &cap_list, node) { node 317 drivers/staging/greybus/authentication.c list_add(&cap->node, &cap_list); node 356 drivers/staging/greybus/authentication.c list_del(&cap->node); node 390 drivers/staging/greybus/authentication.c list_del(&cap->node); node 27 drivers/staging/greybus/fw-download.c struct list_head node; node 99 drivers/staging/greybus/fw-download.c list_for_each_entry(fw_req, &fw_download->fw_requests, node) { node 122 drivers/staging/greybus/fw-download.c list_del(&fw_req->node); node 202 drivers/staging/greybus/fw-download.c list_add(&fw_req->node, &fw_download->fw_requests); node 453 drivers/staging/greybus/fw-download.c list_for_each_entry(fw_req, &fw_download->fw_requests, node) node 458 drivers/staging/greybus/fw-download.c list_for_each_entry_safe(fw_req, tmp, &fw_download->fw_requests, node) { node 27 drivers/staging/greybus/fw-management.c struct list_head node; node 89 drivers/staging/greybus/fw-management.c list_for_each_entry(fw_mgmt, &fw_mgmt_list, node) { node 610 drivers/staging/greybus/fw-management.c list_add(&fw_mgmt->node, &fw_mgmt_list); node 650 drivers/staging/greybus/fw-management.c list_del(&fw_mgmt->node); node 684 drivers/staging/greybus/fw-management.c list_del(&fw_mgmt->node); node 245 drivers/staging/greybus/tools/loopback_test.c int open_sysfs(const char *sys_pfx, const char *node, int flags) node 250 drivers/staging/greybus/tools/loopback_test.c snprintf(path, sizeof(path), "%s%s", sys_pfx, node); node 259 drivers/staging/greybus/tools/loopback_test.c int read_sysfs_int_fd(int fd, const char *sys_pfx, const char *node) node 264 drivers/staging/greybus/tools/loopback_test.c fprintf(stderr, "unable to read from %s%s %s\n", sys_pfx, node, node 272 drivers/staging/greybus/tools/loopback_test.c float read_sysfs_float_fd(int fd, const char *sys_pfx, const char *node) node 278 drivers/staging/greybus/tools/loopback_test.c fprintf(stderr, "unable to read from %s%s %s\n", sys_pfx, node, node 286 drivers/staging/greybus/tools/loopback_test.c int read_sysfs_int(const char *sys_pfx, const char *node) node 290 drivers/staging/greybus/tools/loopback_test.c fd = open_sysfs(sys_pfx, node, O_RDONLY); node 291 drivers/staging/greybus/tools/loopback_test.c val = read_sysfs_int_fd(fd, sys_pfx, node); node 296 drivers/staging/greybus/tools/loopback_test.c float read_sysfs_float(const char *sys_pfx, const char *node) node 301 drivers/staging/greybus/tools/loopback_test.c fd = open_sysfs(sys_pfx, node, O_RDONLY); node 302 drivers/staging/greybus/tools/loopback_test.c val = read_sysfs_float_fd(fd, sys_pfx, node); node 307 drivers/staging/greybus/tools/loopback_test.c void write_sysfs_val(const char *sys_pfx, const char *node, int val) node 312 drivers/staging/greybus/tools/loopback_test.c fd = open_sysfs(sys_pfx, node, O_RDWR); node 315 drivers/staging/greybus/tools/loopback_test.c fprintf(stderr, "unable to write to %s%s %s\n", sys_pfx, node, node 573 drivers/staging/greybus/tools/loopback_test.c int is_loopback_device(const char *path, const char *node) node 577 drivers/staging/greybus/tools/loopback_test.c snprintf(file, MAX_SYSFS_PATH, "%s%s/iteration_count", path, node); node 74 drivers/staging/media/imx/imx-media-dev.c struct device_node *node = dev->of_node; node 82 drivers/staging/media/imx/imx-media-dev.c ret = imx_media_add_of_subdevs(imxmd, node); node 1193 drivers/staging/media/imx/imx7-media-csi.c struct device_node *node = dev->of_node; node 1238 drivers/staging/media/imx/imx7-media-csi.c ret = imx_media_of_add_csi(imxmd, node); node 828 drivers/staging/media/imx/imx7-mipi-csis.c struct device_node *node = pdev->dev.of_node; node 830 drivers/staging/media/imx/imx7-mipi-csis.c if (of_property_read_u32(node, "clock-frequency", node 840 drivers/staging/media/imx/imx7-mipi-csis.c of_property_read_u32(node, "fsl,csis-hs-settle", &state->hs_settle); node 56 drivers/staging/media/ipu3/ipu3-v4l2.c unsigned int node; node 86 drivers/staging/media/ipu3/ipu3-v4l2.c node = imgu_map_node(imgu, i); node 88 drivers/staging/media/ipu3/ipu3-v4l2.c if (node == IMGU_NODE_STAT_3A || node == IMGU_NODE_PARAMS) node 90 drivers/staging/media/ipu3/ipu3-v4l2.c fmts[i] = imgu_pipe->queue_enabled[node] ? node 91 drivers/staging/media/ipu3/ipu3-v4l2.c &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp : NULL; node 311 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 313 drivers/staging/media/ipu3/ipu3-v4l2.c unsigned int queue = imgu_node_to_queue(node->id); node 327 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 329 drivers/staging/media/ipu3/ipu3-v4l2.c unsigned int queue = imgu_node_to_queue(node->id); node 341 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 343 drivers/staging/media/ipu3/ipu3-v4l2.c unsigned int queue = imgu_node_to_queue(node->id); node 351 drivers/staging/media/ipu3/ipu3-v4l2.c need_bytes = node->vdev_fmt.fmt.meta.buffersize; node 353 drivers/staging/media/ipu3/ipu3-v4l2.c need_bytes = node->vdev_fmt.fmt.pix_mp.plane_fmt[0].sizeimage; node 365 drivers/staging/media/ipu3/ipu3-v4l2.c list_add_tail(&buf->vid_buf.list, &node->buffers); node 371 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_queue_buffers(imgu, false, node->pipe); node 374 drivers/staging/media/ipu3/ipu3-v4l2.c node->pipe, node->id); node 384 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 386 drivers/staging/media/ipu3/ipu3-v4l2.c const struct v4l2_format *fmt = &node->vdev_fmt; node 408 drivers/staging/media/ipu3/ipu3-v4l2.c INIT_LIST_HEAD(&node->buffers); node 418 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node; node 430 drivers/staging/media/ipu3/ipu3-v4l2.c node = &imgu->imgu_pipe[p].nodes[i]; node 432 drivers/staging/media/ipu3/ipu3-v4l2.c __func__, p, i, node->name, node->enabled); node 433 drivers/staging/media/ipu3/ipu3-v4l2.c if (node == except) node 435 drivers/staging/media/ipu3/ipu3-v4l2.c if (node->enabled && !vb2_start_streaming_called(&node->vbq)) node 444 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node, node 451 drivers/staging/media/ipu3/ipu3-v4l2.c list_for_each_entry_safe(b, b0, &node->buffers, list) { node 463 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 469 drivers/staging/media/ipu3/ipu3-v4l2.c node->name, node->pipe, node->id); node 476 drivers/staging/media/ipu3/ipu3-v4l2.c if (!node->enabled) { node 482 drivers/staging/media/ipu3/ipu3-v4l2.c pipe = node->pipe; node 484 drivers/staging/media/ipu3/ipu3-v4l2.c r = media_pipeline_start(&node->vdev.entity, &imgu_pipe->pipeline); node 489 drivers/staging/media/ipu3/ipu3-v4l2.c if (!imgu_all_nodes_streaming(imgu, node)) node 508 drivers/staging/media/ipu3/ipu3-v4l2.c media_pipeline_stop(&node->vdev.entity); node 510 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_return_all_buffers(imgu, node, VB2_BUF_STATE_QUEUED); node 520 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = node 525 drivers/staging/media/ipu3/ipu3-v4l2.c WARN_ON(!node->enabled); node 527 drivers/staging/media/ipu3/ipu3-v4l2.c pipe = node->pipe; node 528 drivers/staging/media/ipu3/ipu3-v4l2.c dev_dbg(dev, "Try to stream off node [%u][%u]", pipe, node->id); node 536 drivers/staging/media/ipu3/ipu3-v4l2.c if (imgu->streaming && imgu_all_nodes_streaming(imgu, node)) { node 544 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_return_all_buffers(imgu, node, VB2_BUF_STATE_ERROR); node 545 drivers/staging/media/ipu3/ipu3-v4l2.c media_pipeline_stop(&node->vdev.entity); node 587 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 591 drivers/staging/media/ipu3/ipu3-v4l2.c snprintf(cap->bus_info, sizeof(cap->bus_info), "PCI:%s", node->name); node 638 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 640 drivers/staging/media/ipu3/ipu3-v4l2.c f->fmt = node->vdev_fmt.fmt; node 649 drivers/staging/media/ipu3/ipu3-v4l2.c static int imgu_fmt(struct imgu_device *imgu, unsigned int pipe, int node, node 662 drivers/staging/media/ipu3/ipu3-v4l2.c dev_dbg(dev, "set fmt node [%u][%u](try = %u)", pipe, node, try); node 721 drivers/staging/media/ipu3/ipu3-v4l2.c css_q = imgu_node_to_queue(node); node 740 drivers/staging/media/ipu3/ipu3-v4l2.c f->fmt = imgu_pipe->nodes[node].vdev_fmt.fmt; node 776 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 781 drivers/staging/media/ipu3/ipu3-v4l2.c pix_mp->width, pix_mp->height, node->id); node 787 drivers/staging/media/ipu3/ipu3-v4l2.c return imgu_fmt(imgu, node->pipe, node->id, f, true); node 794 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 799 drivers/staging/media/ipu3/ipu3-v4l2.c pix_mp->width, pix_mp->height, node->id); node 805 drivers/staging/media/ipu3/ipu3-v4l2.c return imgu_fmt(imgu, node->pipe, node->id, f, false); node 822 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 826 drivers/staging/media/ipu3/ipu3-v4l2.c if (fmt->index > 0 || fmt->type != node->vbq.type) node 838 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = file_to_intel_imgu_node(file); node 840 drivers/staging/media/ipu3/ipu3-v4l2.c if (f->type != node->vbq.type) node 843 drivers/staging/media/ipu3/ipu3-v4l2.c f->fmt = node->vdev_fmt.fmt; node 1055 drivers/staging/media/ipu3/ipu3-v4l2.c static void imgu_node_to_v4l2(u32 node, struct video_device *vdev, node 1061 drivers/staging/media/ipu3/ipu3-v4l2.c WARN_ON(node >= IMGU_NODE_NUM); node 1063 drivers/staging/media/ipu3/ipu3-v4l2.c switch (node) { node 1160 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_video_device *node = &imgu_pipe->nodes[node_num]; node 1161 drivers/staging/media/ipu3/ipu3-v4l2.c struct video_device *vdev = &node->vdev; node 1162 drivers/staging/media/ipu3/ipu3-v4l2.c struct vb2_queue *vbq = &node->vbq; node 1188 drivers/staging/media/ipu3/ipu3-v4l2.c mutex_init(&node->lock); node 1189 drivers/staging/media/ipu3/ipu3-v4l2.c INIT_LIST_HEAD(&node->buffers); node 1192 drivers/staging/media/ipu3/ipu3-v4l2.c node->pad_fmt = def_bus_fmt; node 1193 drivers/staging/media/ipu3/ipu3-v4l2.c node->id = node_num; node 1194 drivers/staging/media/ipu3/ipu3-v4l2.c node->pipe = pipe; node 1195 drivers/staging/media/ipu3/ipu3-v4l2.c imgu_node_to_v4l2(node_num, vdev, &node->vdev_fmt); node 1196 drivers/staging/media/ipu3/ipu3-v4l2.c if (node->vdev_fmt.type == node 1198 drivers/staging/media/ipu3/ipu3-v4l2.c node->vdev_fmt.type == node 1200 drivers/staging/media/ipu3/ipu3-v4l2.c def_pix_fmt.pixelformat = node->output ? node 1203 drivers/staging/media/ipu3/ipu3-v4l2.c node->vdev_fmt.fmt.pix_mp = def_pix_fmt; node 1207 drivers/staging/media/ipu3/ipu3-v4l2.c r = media_entity_pads_init(&vdev->entity, 1, &node->vdev_pad); node 1210 drivers/staging/media/ipu3/ipu3-v4l2.c mutex_destroy(&node->lock); node 1213 drivers/staging/media/ipu3/ipu3-v4l2.c node->vdev_pad.flags = node->output ? node 1218 drivers/staging/media/ipu3/ipu3-v4l2.c vbq->type = node->vdev_fmt.type; node 1230 drivers/staging/media/ipu3/ipu3-v4l2.c vbq->lock = &node->lock; node 1240 drivers/staging/media/ipu3/ipu3-v4l2.c IMGU_NAME, pipe, node->name); node 1243 drivers/staging/media/ipu3/ipu3-v4l2.c vdev->lock = &node->lock; node 1245 drivers/staging/media/ipu3/ipu3-v4l2.c vdev->queue = &node->vbq; node 1246 drivers/staging/media/ipu3/ipu3-v4l2.c vdev->vfl_dir = node->output ? VFL_DIR_TX : VFL_DIR_RX; node 1257 drivers/staging/media/ipu3/ipu3-v4l2.c if (node->enabled) node 1259 drivers/staging/media/ipu3/ipu3-v4l2.c if (node->output) { node 1276 drivers/staging/media/ipu3/ipu3-v4l2.c unsigned int pipe, int node) node 1281 drivers/staging/media/ipu3/ipu3-v4l2.c for (i = 0; i < node; i++) { node 51 drivers/staging/media/ipu3/ipu3.c unsigned int imgu_node_to_queue(unsigned int node) node 53 drivers/staging/media/ipu3/ipu3.c return imgu_node_map[node].css_queue; node 110 drivers/staging/media/ipu3/ipu3.c unsigned int i, k, node; node 116 drivers/staging/media/ipu3/ipu3.c node = imgu_map_node(imgu, i); node 117 drivers/staging/media/ipu3/ipu3.c if (!imgu_pipe->queue_enabled[node] || i == IMGU_QUEUE_MASTER) node 128 drivers/staging/media/ipu3/ipu3.c meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta; node 129 drivers/staging/media/ipu3/ipu3.c mpix = &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp; node 131 drivers/staging/media/ipu3/ipu3.c if (node == IMGU_NODE_STAT_3A || node == IMGU_NODE_PARAMS) node 204 drivers/staging/media/ipu3/ipu3.c unsigned int node, node 210 drivers/staging/media/ipu3/ipu3.c if (WARN_ON(node >= IMGU_NODE_NUM)) node 214 drivers/staging/media/ipu3/ipu3.c list_for_each_entry(buf, &imgu_pipe->nodes[node].buffers, vid_buf.list) { node 220 drivers/staging/media/ipu3/ipu3.c return imgu_dummybufs_get(imgu, imgu_node_map[node].css_queue, pipe); node 229 drivers/staging/media/ipu3/ipu3.c unsigned int node; node 245 drivers/staging/media/ipu3/ipu3.c for (node = IMGU_NODE_NUM - 1; node 247 drivers/staging/media/ipu3/ipu3.c node = node ? node - 1 : IMGU_NODE_NUM - 1) { node 248 drivers/staging/media/ipu3/ipu3.c if (node == IMGU_NODE_VF && node 253 drivers/staging/media/ipu3/ipu3.c } else if (node == IMGU_NODE_PARAMS && node 254 drivers/staging/media/ipu3/ipu3.c imgu_pipe->nodes[node].enabled) { node 259 drivers/staging/media/ipu3/ipu3.c if (list_empty(&imgu_pipe->nodes[node].buffers)) node 262 drivers/staging/media/ipu3/ipu3.c ivb = list_first_entry(&imgu_pipe->nodes[node].buffers, node 278 drivers/staging/media/ipu3/ipu3.c } else if (imgu_pipe->queue_enabled[node]) { node 280 drivers/staging/media/ipu3/ipu3.c imgu_queue_getbuf(imgu, node, pipe); node 297 drivers/staging/media/ipu3/ipu3.c imgu_node_map[node].name, node 316 drivers/staging/media/ipu3/ipu3.c node, r); node 322 drivers/staging/media/ipu3/ipu3.c for (node = 0; node < IMGU_NODE_NUM; node++) { node 325 drivers/staging/media/ipu3/ipu3.c if (!imgu_pipe->queue_enabled[node]) node 330 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->nodes[node].buffers, node 512 drivers/staging/media/ipu3/ipu3.c unsigned int node, pipe; node 529 drivers/staging/media/ipu3/ipu3.c node = imgu_map_node(imgu, b->queue); node 537 drivers/staging/media/ipu3/ipu3.c imgu_node_map[node].name, node 547 drivers/staging/media/ipu3/ipu3.c if (!imgu_pipe->nodes[node].output) { node 552 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->nodes[node].sequence); node 157 drivers/staging/media/ipu3/ipu3.h unsigned int imgu_node_to_queue(unsigned int node); node 412 drivers/staging/mt7621-dma/mtk-hsdma.c list_del(&desc->vdesc.node); node 284 drivers/staging/mt7621-pci/pci-mt7621.c struct device_node *node = dev->of_node; node 289 drivers/staging/mt7621-pci/pci-mt7621.c if (of_pci_range_parser_init(&parser, node)) { node 310 drivers/staging/mt7621-pci/pci-mt7621.c of_pci_range_to_resource(&range, node, res); node 313 drivers/staging/mt7621-pci/pci-mt7621.c err = of_pci_parse_bus_range(node, &pcie->busn); node 316 drivers/staging/mt7621-pci/pci-mt7621.c pcie->busn.name = node->name; node 326 drivers/staging/mt7621-pci/pci-mt7621.c struct device_node *node, node 374 drivers/staging/mt7621-pci/pci-mt7621.c struct device_node *node = dev->of_node, *child; node 384 drivers/staging/mt7621-pci/pci-mt7621.c err = of_address_to_resource(node, 0, ®s); node 400 drivers/staging/mt7621-pci/pci-mt7621.c for_each_available_child_of_node(node, child) { node 270 drivers/staging/nvec/nvec.c list_add_tail(&msg->node, &nvec->tx_data); node 382 drivers/staging/nvec/nvec.c msg = list_first_entry(&nvec->tx_data, struct nvec_msg, node); node 397 drivers/staging/nvec/nvec.c list_del_init(&msg->node); node 445 drivers/staging/nvec/nvec.c msg = list_first_entry(&nvec->rx_data, struct nvec_msg, node); node 446 drivers/staging/nvec/nvec.c list_del_init(&msg->node); node 511 drivers/staging/nvec/nvec.c list_add_tail(&nvec->rx->node, &nvec->rx_data); node 555 drivers/staging/nvec/nvec.c list_add_tail(&nvec->tx->node, &nvec->tx_data); node 558 drivers/staging/nvec/nvec.c node); node 95 drivers/staging/nvec/nvec.h struct list_head node; node 273 drivers/staging/octeon-usb/octeon-hcd.c struct list_head node; node 317 drivers/staging/octeon-usb/octeon-hcd.c struct list_head node; node 1134 drivers/staging/octeon-usb/octeon-hcd.c list_add_tail(&pipe->node, &usb->idle_pipes); node 1355 drivers/staging/octeon-usb/octeon-hcd.c node); node 1500 drivers/staging/octeon-usb/octeon-hcd.c node); node 1847 drivers/staging/octeon-usb/octeon-hcd.c list_for_each_entry(pipe, list, node) { node 1850 drivers/staging/octeon-usb/octeon-hcd.c node); node 1945 drivers/staging/octeon-usb/octeon-hcd.c list_for_each_entry(pipe, &usb->active_pipes[ttype], node) { node 2092 drivers/staging/octeon-usb/octeon-hcd.c list_del(&transaction->node); node 2094 drivers/staging/octeon-usb/octeon-hcd.c list_move_tail(&pipe->node, &usb->idle_pipes); node 2160 drivers/staging/octeon-usb/octeon-hcd.c list_add_tail(&transaction->node, &pipe->transactions); node 2162 drivers/staging/octeon-usb/octeon-hcd.c list_add_tail(&transaction->node, &pipe->transactions); node 2163 drivers/staging/octeon-usb/octeon-hcd.c list_move_tail(&pipe->node, node 2304 drivers/staging/octeon-usb/octeon-hcd.c if (list_first_entry(&pipe->transactions, typeof(*transaction), node) == node 2346 drivers/staging/octeon-usb/octeon-hcd.c list_for_each_entry_safe(transaction, next, &pipe->transactions, node) { node 2371 drivers/staging/octeon-usb/octeon-hcd.c list_del(&pipe->node); node 2678 drivers/staging/octeon-usb/octeon-hcd.c node); node 616 drivers/staging/octeon/ethernet.c struct device_node *node = NULL; node 621 drivers/staging/octeon/ethernet.c node = of_get_next_child(parent, node); node 622 drivers/staging/octeon/ethernet.c if (!node) node 624 drivers/staging/octeon/ethernet.c addr = of_get_property(node, "reg", &size); node 628 drivers/staging/octeon/ethernet.c return node; node 447 drivers/staging/ralink-gdma/ralink-gdma.c list_del(&desc->vdesc.node); node 78 drivers/staging/rtl8712/rtl871x_event.h unsigned char *node; node 519 drivers/staging/rtl8723bs/hal/rtl8723b_cmd.c memcpy(pSSIDListPkt, &pwrctl->pno_ssid_list->node[i].SSID, node 86 drivers/staging/rtl8723bs/include/rtw_event.h unsigned char *node; node 179 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h struct pno_ssid node[MAX_PNO_LIST_COUNT]; node 163 drivers/staging/speakup/spk_types.h struct list_head node; node 372 drivers/staging/speakup/synth.c list_for_each_entry(tmp, &synths, node) { node 455 drivers/staging/speakup/synth.c list_for_each_entry(tmp, &synths, node) { node 466 drivers/staging/speakup/synth.c list_add_tail(&in_synth->node, &synths); node 478 drivers/staging/speakup/synth.c list_del(&in_synth->node); node 165 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { node 180 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { node 263 drivers/staging/uwb/beacon.c list_add(&bce->node, &rc->uwb_beca.list); node 278 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { node 293 drivers/staging/uwb/beacon.c list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) { node 294 drivers/staging/uwb/beacon.c list_del(&bce->node); node 352 drivers/staging/uwb/lc-dev.c list_del(&uwb_dev->bce->node); node 20 drivers/staging/uwb/pal.c INIT_LIST_HEAD(&pal->node); node 53 drivers/staging/uwb/pal.c list_add(&pal->node, &rc->pals); node 105 drivers/staging/uwb/pal.c list_del(&pal->node); node 35 drivers/staging/uwb/radio.c list_for_each_entry(pal, &rc->pals, node) { node 172 drivers/staging/uwb/radio.c list_for_each_entry(pal, &rc->pals, node) { node 760 drivers/staging/uwb/rsv.c list_for_each_entry(pal, &rc->pals, node) { node 267 drivers/staging/uwb/uwb-internal.h struct list_head node; node 440 drivers/staging/uwb/uwb.h struct list_head node; node 206 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_debugfs.c struct vchiq_debugfs_node *node = node 209 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_debugfs.c debugfs_remove_recursive(node->dentry); node 774 drivers/target/iscsi/iscsi_target_stat.c CONFIGFS_ATTR_RO(iscsi_stat_sess_, node); node 110 drivers/target/target_core_user.c struct list_head node; node 1396 drivers/target/target_core_user.c INIT_LIST_HEAD(&udev->node); node 1960 drivers/target/target_core_user.c list_add(&udev->node, &root_udev); node 1997 drivers/target/target_core_user.c list_del(&udev->node); node 2645 drivers/target/target_core_user.c list_for_each_entry(udev, &root_udev, node) { node 141 drivers/tc/tc.c list_add_tail(&tdev->node, &tbus->devices); node 89 drivers/thermal/cpu_cooling.c struct list_head node; node 634 drivers/thermal/cpu_cooling.c list_add(&cpufreq_cdev->node, &cpufreq_cdev_list); node 735 drivers/thermal/cpu_cooling.c list_del(&cpufreq_cdev->node); node 352 drivers/thermal/qcom/qcom-spmi-temp-alarm.c struct device_node *node; node 357 drivers/thermal/qcom/qcom-spmi-temp-alarm.c node = pdev->dev.of_node; node 372 drivers/thermal/qcom/qcom-spmi-temp-alarm.c ret = of_property_read_u32(node, "reg", &res); node 431 drivers/thermal/qcom/qcom-spmi-temp-alarm.c IRQF_ONESHOT, node->name, chip); node 152 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) { node 193 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) { node 618 drivers/thermal/thermal_core.c list_for_each_entry(cdev, &thermal_cdev_list, node) { node 638 drivers/thermal/thermal_core.c list_for_each_entry(cdev, &thermal_cdev_list, node) { node 697 drivers/thermal/thermal_core.c list_for_each_entry(pos1, &thermal_tz_list, node) { node 701 drivers/thermal/thermal_core.c list_for_each_entry(pos2, &thermal_cdev_list, node) { node 902 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) { node 994 drivers/thermal/thermal_core.c list_add(&cdev->node, &thermal_cdev_list); node 1001 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) node 1132 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_cdev_list, node) node 1140 drivers/thermal/thermal_core.c list_del(&cdev->node); node 1143 drivers/thermal/thermal_core.c list_for_each_entry(tz, &thermal_tz_list, node) { node 1183 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_cdev_list, node) { node 1194 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_cdev_list, node) { node 1342 drivers/thermal/thermal_core.c list_add_tail(&tz->node, &thermal_tz_list); node 1387 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) node 1395 drivers/thermal/thermal_core.c list_del(&tz->node); node 1398 drivers/thermal/thermal_core.c list_for_each_entry(cdev, &thermal_cdev_list, node) { node 1448 drivers/thermal/thermal_core.c list_for_each_entry(pos, &thermal_tz_list, node) node 1574 drivers/thermal/thermal_core.c list_for_each_entry(tz, &thermal_tz_list, node) { node 26 drivers/thermal/thermal_hwmon.c struct list_head node; node 93 drivers/thermal/thermal_hwmon.c list_for_each_entry(hwmon, &thermal_hwmon_list, node) { node 193 drivers/thermal/thermal_hwmon.c list_add_tail(&hwmon->node, &thermal_hwmon_list); node 244 drivers/thermal/thermal_hwmon.c list_del(&hwmon->node); node 820 drivers/thermal/ti-soc-thermal/ti-bandgap.c struct device_node *node = pdev->dev.of_node; node 827 drivers/thermal/ti-soc-thermal/ti-bandgap.c if (!node) { node 863 drivers/thermal/ti-soc-thermal/ti-bandgap.c bgp->tshut_gpio = of_get_gpio(node, 0); node 915 drivers/tty/hvc/hvcs.c list_for_each_entry(pi, &head, node) node 523 drivers/tty/serdev/core.c struct device_node *node; node 528 drivers/tty/serdev/core.c for_each_available_child_of_node(ctrl->dev.of_node, node) { node 529 drivers/tty/serdev/core.c if (!of_get_property(node, "compatible", NULL)) node 532 drivers/tty/serdev/core.c dev_dbg(&ctrl->dev, "adding child %pOF\n", node); node 538 drivers/tty/serdev/core.c serdev->dev.of_node = node; node 84 drivers/tty/serial/8250/8250_core.c struct hlist_node node; node 167 drivers/tty/serial/8250/8250_core.c hlist_del(&i->node); node 184 drivers/tty/serial/8250/8250_core.c i = hlist_entry(n, struct irq_info, node); node 197 drivers/tty/serial/8250/8250_core.c hlist_add_head(&i->node, h); node 236 drivers/tty/serial/8250/8250_core.c i = hlist_entry(n, struct irq_info, node); node 232 drivers/tty/serial/earlycon.c unsigned long node, node 243 drivers/tty/serial/earlycon.c addr = of_flat_dt_translate_address(node); node 250 drivers/tty/serial/earlycon.c val = of_get_flat_dt_prop(node, "reg-offset", NULL); node 255 drivers/tty/serial/earlycon.c val = of_get_flat_dt_prop(node, "reg-shift", NULL); node 258 drivers/tty/serial/earlycon.c big_endian = of_get_flat_dt_prop(node, "big-endian", NULL) != NULL || node 260 drivers/tty/serial/earlycon.c of_get_flat_dt_prop(node, "native-endian", NULL) != NULL); node 261 drivers/tty/serial/earlycon.c val = of_get_flat_dt_prop(node, "reg-io-width", NULL); node 279 drivers/tty/serial/earlycon.c val = of_get_flat_dt_prop(node, "current-speed", NULL); node 283 drivers/tty/serial/earlycon.c val = of_get_flat_dt_prop(node, "clock-frequency", NULL); node 821 drivers/tty/serial/lantiq.c struct device_node *node = pdev->dev.of_node; node 848 drivers/tty/serial/lantiq.c line = of_alias_get_id(node, "serial"); node 737 drivers/tty/serial/pmac_zilog.c PMAC_FTR_SCC_ENABLE, uap->node, uap->port_type, 1); node 741 drivers/tty/serial/pmac_zilog.c PMAC_FTR_MODEM_ENABLE, uap->node, 0, 1); node 751 drivers/tty/serial/pmac_zilog.c PMAC_FTR_MODEM_ENABLE, uap->node, 0, 0); node 754 drivers/tty/serial/pmac_zilog.c pmac_call_feature(PMAC_FTR_SCC_ENABLE, uap->node, uap->port_type, 0); node 1399 drivers/tty/serial/pmac_zilog.c struct device_node *np = uap->node; node 1532 drivers/tty/serial/pmac_zilog.c np = uap->node; node 1536 drivers/tty/serial/pmac_zilog.c uap->node = NULL; node 1552 drivers/tty/serial/pmac_zilog.c if (pmz_ports[i].node == mdev->ofdev.dev.of_node) node 1569 drivers/tty/serial/pmac_zilog.c uap->node); node 1672 drivers/tty/serial/pmac_zilog.c pmz_ports[count].node = node_a; node 1673 drivers/tty/serial/pmac_zilog.c pmz_ports[count+1].node = node_b; node 1928 drivers/tty/serial/pmac_zilog.c if (uport->node != NULL) node 2006 drivers/tty/serial/pmac_zilog.c if (uap->node == NULL) node 27 drivers/tty/serial/pmac_zilog.h struct device_node *node; node 1662 drivers/tty/serial/serial-tegra.c struct device_node *node; node 1666 drivers/tty/serial/serial-tegra.c node = of_find_matching_node(NULL, tegra_uart_of_match); node 1667 drivers/tty/serial/serial-tegra.c if (node) node 1668 drivers/tty/serial/serial-tegra.c match = of_match_node(tegra_uart_of_match, node); node 575 drivers/tty/vcc.c u64 node; node 620 drivers/tty/vcc.c node = vio_vdev_node(hp, vdev); node 621 drivers/tty/vcc.c if (node == MDESC_NODE_NULL) { node 627 drivers/tty/vcc.c domain = mdesc_get_property(hp, node, "vcc-domain-name", NULL); node 121 drivers/tty/vt/vt.c int node; node 4017 drivers/tty/vt/vt.c con_driver->node = i; node 4032 drivers/tty/vt/vt.c MKDEV(0, con_driver->node), node 4034 drivers/tty/vt/vt.c "vtcon%i", con_driver->node); node 4112 drivers/tty/vt/vt.c device_destroy(vtconsole_class, MKDEV(0, con_driver->node)); node 4120 drivers/tty/vt/vt.c con_driver->node = 0; node 4186 drivers/tty/vt/vt.c MKDEV(0, con->node), node 4188 drivers/tty/vt/vt.c "vtcon%i", con->node); node 250 drivers/uio/uio_fsl_elbc_gpcm.c static int get_of_data(struct fsl_elbc_gpcm *priv, struct device_node *node, node 259 drivers/uio/uio_fsl_elbc_gpcm.c ret = of_address_to_resource(node, 0, res); node 266 drivers/uio/uio_fsl_elbc_gpcm.c ret = of_property_read_u32(node, "reg", &priv->bank); node 273 drivers/uio/uio_fsl_elbc_gpcm.c ret = of_property_read_u32(node, "elbc-gpcm-br", reg_br); node 280 drivers/uio/uio_fsl_elbc_gpcm.c ret = of_property_read_u32(node, "elbc-gpcm-or", reg_or); node 288 drivers/uio/uio_fsl_elbc_gpcm.c if (of_property_read_string(node, "device_type", &type) == 0) node 292 drivers/uio/uio_fsl_elbc_gpcm.c *irq = irq_of_parse_and_map(node, 0); node 300 drivers/uio/uio_fsl_elbc_gpcm.c if (of_property_read_string(node, "uio_name", &dt_name) != 0) node 311 drivers/uio/uio_fsl_elbc_gpcm.c struct device_node *node = pdev->dev.of_node; node 334 drivers/uio/uio_fsl_elbc_gpcm.c ret = get_of_data(priv, node, &res, ®_br_new, ®_or_new, node 390 drivers/uio/uio_fsl_elbc_gpcm.c info->mem[0].name = kasprintf(GFP_KERNEL, "%pOFn", node); node 105 drivers/uio/uio_pdrv_genirq.c struct device_node *node = pdev->dev.of_node; node 111 drivers/uio/uio_pdrv_genirq.c if (node) { node 122 drivers/uio/uio_pdrv_genirq.c if (!of_property_read_string(node, "linux,uio-name", &name)) node 126 drivers/uio/uio_pdrv_genirq.c "%pOFn", node); node 31 drivers/usb/c67x00/c67x00-sched.c struct list_head node; node 254 drivers/usb/c67x00/c67x00-sched.c INIT_LIST_HEAD(&ep_data->node); node 268 drivers/usb/c67x00/c67x00-sched.c if (list_empty(&ep_data->node)) { node 269 drivers/usb/c67x00/c67x00-sched.c list_add(&ep_data->node, &c67x00->list[type]); node 273 drivers/usb/c67x00/c67x00-sched.c list_for_each_entry(prev, &c67x00->list[type], node) { node 276 drivers/usb/c67x00/c67x00-sched.c list_add(&ep_data->node, prev->node.prev); node 297 drivers/usb/c67x00/c67x00-sched.c list_del(&ep_data->node); node 783 drivers/usb/c67x00/c67x00-sched.c list_for_each_entry(ep_data, &c67x00->list[type], node) { node 158 drivers/usb/chipidea/debug.c struct td_node *node, *tmpnode; node 169 drivers/usb/chipidea/debug.c list_for_each_entry_safe(node, tmpnode, &req->tds, td) { node 172 drivers/usb/chipidea/debug.c (u32)node->dma, node 178 drivers/usb/chipidea/debug.c *((u32 *)node->ptr + j)); node 345 drivers/usb/chipidea/udc.c struct td_node *lastnode, *node = kzalloc(sizeof(struct td_node), node 348 drivers/usb/chipidea/udc.c if (node == NULL) node 351 drivers/usb/chipidea/udc.c node->ptr = dma_pool_zalloc(hwep->td_pool, GFP_ATOMIC, &node->dma); node 352 drivers/usb/chipidea/udc.c if (node->ptr == NULL) { node 353 drivers/usb/chipidea/udc.c kfree(node); node 357 drivers/usb/chipidea/udc.c node->ptr->token = cpu_to_le32(length << __ffs(TD_TOTAL_BYTES)); node 358 drivers/usb/chipidea/udc.c node->ptr->token &= cpu_to_le32(TD_TOTAL_BYTES); node 359 drivers/usb/chipidea/udc.c node->ptr->token |= cpu_to_le32(TD_STATUS_ACTIVE); node 366 drivers/usb/chipidea/udc.c node->ptr->token |= cpu_to_le32(mul << __ffs(TD_MULTO)); node 371 drivers/usb/chipidea/udc.c node->ptr->page[0] = cpu_to_le32(temp); node 375 drivers/usb/chipidea/udc.c node->ptr->page[i] = cpu_to_le32(page); node 385 drivers/usb/chipidea/udc.c lastnode->ptr->next = cpu_to_le32(node->dma); node 388 drivers/usb/chipidea/udc.c INIT_LIST_HEAD(&node->td); node 389 drivers/usb/chipidea/udc.c list_add_tail(&node->td, &hwreq->tds); node 529 drivers/usb/chipidea/udc.c struct td_node *node) node 531 drivers/usb/chipidea/udc.c hwep->qh.ptr->td.next = cpu_to_le32(node->dma); node 549 drivers/usb/chipidea/udc.c struct td_node *node, *tmpnode; node 559 drivers/usb/chipidea/udc.c list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { node 560 drivers/usb/chipidea/udc.c tmptoken = le32_to_cpu(node->ptr->token); node 566 drivers/usb/chipidea/udc.c reprime_dtd(ci, hwep, node); node 601 drivers/usb/chipidea/udc.c hwep->pending_td = node; node 602 drivers/usb/chipidea/udc.c list_del_init(&node->td); node 627 drivers/usb/chipidea/udc.c struct td_node *node, *tmpnode; node 639 drivers/usb/chipidea/udc.c list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { node 640 drivers/usb/chipidea/udc.c dma_pool_free(hwep->td_pool, node->ptr, node->dma); node 641 drivers/usb/chipidea/udc.c list_del_init(&node->td); node 642 drivers/usb/chipidea/udc.c node->ptr = NULL; node 643 drivers/usb/chipidea/udc.c kfree(node); node 1359 drivers/usb/chipidea/udc.c struct td_node *node, *tmpnode; node 1371 drivers/usb/chipidea/udc.c list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { node 1372 drivers/usb/chipidea/udc.c dma_pool_free(hwep->td_pool, node->ptr, node->dma); node 1373 drivers/usb/chipidea/udc.c list_del_init(&node->td); node 1374 drivers/usb/chipidea/udc.c node->ptr = NULL; node 1375 drivers/usb/chipidea/udc.c kfree(node); node 1418 drivers/usb/chipidea/udc.c struct td_node *node, *tmpnode; node 1429 drivers/usb/chipidea/udc.c list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { node 1430 drivers/usb/chipidea/udc.c dma_pool_free(hwep->td_pool, node->ptr, node->dma); node 1431 drivers/usb/chipidea/udc.c list_del(&node->td); node 1432 drivers/usb/chipidea/udc.c kfree(node); node 282 drivers/usb/common/common.c struct device_node *node; node 285 drivers/usb/common/common.c node = of_parse_phandle(dev->of_node, "companion", 0); node 286 drivers/usb/common/common.c if (node) node 287 drivers/usb/common/common.c pdev = of_find_device_by_node(node); node 289 drivers/usb/common/common.c of_node_put(node); node 64 drivers/usb/core/driver.c INIT_LIST_HEAD(&dynid->node); node 99 drivers/usb/core/driver.c list_add_tail(&dynid->node, &dynids->list); node 119 drivers/usb/core/driver.c list_for_each_entry(dynid, &dynids->list, node) node 164 drivers/usb/core/driver.c list_for_each_entry_safe(dynid, n, &usb_driver->dynids.list, node) { node 169 drivers/usb/core/driver.c list_del(&dynid->node); node 224 drivers/usb/core/driver.c list_for_each_entry_safe(dynid, n, &usb_drv->dynids.list, node) { node 225 drivers/usb/core/driver.c list_del(&dynid->node); node 237 drivers/usb/core/driver.c list_for_each_entry(dynid, &drv->dynids.list, node) { node 2051 drivers/usb/core/message.c struct list_head node; node 2063 drivers/usb/core/message.c list_del(&req->node); node 2081 drivers/usb/core/message.c list_for_each_entry(req, &set_config_list, node) { node 2120 drivers/usb/core/message.c list_add(&req->node, &set_config_list); node 27 drivers/usb/core/of.c struct device_node *node; node 30 drivers/usb/core/of.c for_each_child_of_node(hub->dev.of_node, node) { node 31 drivers/usb/core/of.c if (of_property_read_u32(node, "reg", ®)) node 35 drivers/usb/core/of.c return node; node 94 drivers/usb/core/of.c struct device_node *node; node 97 drivers/usb/core/of.c for_each_child_of_node(udev->dev.of_node, node) { node 98 drivers/usb/core/of.c if (of_property_read_u32_array(node, "reg", reg, 2)) node 102 drivers/usb/core/of.c return node; node 134 drivers/usb/core/usb-acpi.c list_for_each_entry(adev, &parent->children, node) { node 1082 drivers/usb/dwc3/core.c struct device_node *node = dev->of_node; node 1085 drivers/usb/dwc3/core.c if (node) { node 53 drivers/usb/dwc3/dwc3-exynos.c struct device_node *node = dev->of_node; node 112 drivers/usb/dwc3/dwc3-exynos.c if (node) { node 113 drivers/usb/dwc3/dwc3-exynos.c ret = of_platform_populate(node, NULL, NULL, dev); node 82 drivers/usb/dwc3/dwc3-keystone.c struct device_node *node = pdev->dev.of_node; node 108 drivers/usb/dwc3/dwc3-keystone.c if (of_device_is_compatible(node, "ti,am654-dwc3")) node 128 drivers/usb/dwc3/dwc3-keystone.c error = of_platform_populate(node, NULL, NULL, dev); node 157 drivers/usb/dwc3/dwc3-keystone.c struct device_node *node = pdev->dev.of_node; node 159 drivers/usb/dwc3/dwc3-keystone.c if (!of_device_is_compatible(node, "ti,am654-dwc3")) node 370 drivers/usb/dwc3/dwc3-omap.c struct device_node *node = omap->dev->of_node; node 380 drivers/usb/dwc3/dwc3-omap.c if (of_device_is_compatible(node, "ti,am437x-dwc3")) { node 392 drivers/usb/dwc3/dwc3-omap.c struct device_node *node = omap->dev->of_node; node 397 drivers/usb/dwc3/dwc3-omap.c of_property_read_u32(node, "utmi-mode", &utmi_mode); node 416 drivers/usb/dwc3/dwc3-omap.c struct device_node *node = omap->dev->of_node; node 419 drivers/usb/dwc3/dwc3-omap.c if (of_property_read_bool(node, "extcon")) { node 451 drivers/usb/dwc3/dwc3-omap.c struct device_node *node = pdev->dev.of_node; node 464 drivers/usb/dwc3/dwc3-omap.c if (!node) { node 483 drivers/usb/dwc3/dwc3-omap.c if (of_property_read_bool(node, "vbus-supply")) { node 513 drivers/usb/dwc3/dwc3-omap.c ret = of_platform_populate(node, NULL, NULL, dev); node 200 drivers/usb/dwc3/dwc3-st.c struct device_node *node = dev->of_node, *child; node 214 drivers/usb/dwc3/dwc3-st.c regmap = syscon_regmap_lookup_by_phandle(node, "st,syscfg"); node 254 drivers/usb/dwc3/dwc3-st.c child = of_get_child_by_name(node, "dwc3"); node 262 drivers/usb/dwc3/dwc3-st.c ret = of_platform_populate(node, NULL, NULL, dev); node 911 drivers/usb/dwc3/gadget.c dma_addr_t dma, unsigned length, unsigned chain, unsigned node, node 928 drivers/usb/dwc3/gadget.c if (!node) { node 1024 drivers/usb/dwc3/gadget.c struct dwc3_request *req, unsigned chain, unsigned node) node 1051 drivers/usb/dwc3/gadget.c __dwc3_prepare_one_trb(dep, trb, dma, length, chain, node, node 33 drivers/usb/gadget/legacy/hid.c struct list_head node; node 100 drivers/usb/gadget/legacy/hid.c list_for_each_entry(e, &hidg_func_list, node) { node 113 drivers/usb/gadget/legacy/hid.c list_for_each_entry(n, &hidg_func_list, node) { node 145 drivers/usb/gadget/legacy/hid.c list_for_each_entry(n, &hidg_func_list, node) { node 195 drivers/usb/gadget/legacy/hid.c list_for_each_entry(m, &hidg_func_list, node) { node 207 drivers/usb/gadget/legacy/hid.c list_for_each_entry(n, &hidg_func_list, node) { node 233 drivers/usb/gadget/legacy/hid.c list_add_tail(&entry->node, &hidg_func_list); node 242 drivers/usb/gadget/legacy/hid.c list_for_each_entry_safe(e, n, &hidg_func_list, node) { node 243 drivers/usb/gadget/legacy/hid.c list_del(&e->node); node 181 drivers/usb/gadget/udc/fsl_qe_udc.h struct list_head node; node 249 drivers/usb/gadget/udc/fsl_qe_udc.h INIT_LIST_HEAD(&(frm->node)); node 174 drivers/usb/host/fhci-hcd.c list_for_each_entry_safe(ed, next_ed, &fhci->empty_eds, node) { node 175 drivers/usb/host/fhci-hcd.c list_del(&ed->node); node 179 drivers/usb/host/fhci-hcd.c list_for_each_entry_safe(td, next_td, &fhci->empty_tds, node) { node 180 drivers/usb/host/fhci-hcd.c list_del(&td->node); node 563 drivers/usb/host/fhci-hcd.c struct device_node *node = dev->of_node; node 579 drivers/usb/host/fhci-hcd.c sprop = of_get_property(node, "mode", NULL); node 593 drivers/usb/host/fhci-hcd.c iprop = of_get_property(node, "hub-power-budget", &size); node 598 drivers/usb/host/fhci-hcd.c ret = of_address_to_resource(node, 0, &usb_regs); node 613 drivers/usb/host/fhci-hcd.c iprop = of_get_property(node, "reg", &size); node 636 drivers/usb/host/fhci-hcd.c gpio = of_get_gpio_flags(node, i, &flags); node 671 drivers/usb/host/fhci-hcd.c fhci->pins[j] = qe_pin_request(node, j); node 695 drivers/usb/host/fhci-hcd.c usb_irq = irq_of_parse_and_map(node, 0); node 703 drivers/usb/host/fhci-hcd.c sprop = of_get_property(node, "fsl,fullspeed-clock", NULL); node 713 drivers/usb/host/fhci-hcd.c sprop = of_get_property(node, "fsl,lowspeed-clock", NULL); node 26 drivers/usb/host/fhci-mem.c INIT_LIST_HEAD(&td->node); node 34 drivers/usb/host/fhci-mem.c INIT_LIST_HEAD(&ed->node); node 42 drivers/usb/host/fhci-mem.c td = list_entry(fhci->empty_tds.next, struct td, node); node 58 drivers/usb/host/fhci-mem.c list_add(&td->node, &fhci->empty_tds); node 66 drivers/usb/host/fhci-mem.c ed = list_entry(fhci->empty_eds.next, struct ed, node); node 82 drivers/usb/host/fhci-mem.c list_add(&ed->node, &fhci->empty_eds); node 64 drivers/usb/host/fhci-q.c list_add_tail(&td->node, &ed->td_list); node 75 drivers/usb/host/fhci-q.c td = list_entry(ed->td_list.next, struct td, node); node 112 drivers/usb/host/fhci-q.c td = list_entry(ed->td_list.next, struct td, node); node 118 drivers/usb/host/fhci-q.c node); node 132 drivers/usb/host/fhci-q.c td = list_entry(p_list->done_list.next, struct td, node); node 145 drivers/usb/host/fhci-q.c list_del_init(&td->node); node 149 drivers/usb/host/fhci-q.c ed->td_head = list_entry(ed->td_list.next, struct td, node); node 155 drivers/usb/host/fhci-q.c list_add_tail(&td->node, &usb->hc_list->done_list); node 168 drivers/usb/host/fhci-q.c list_del_init(&urb_priv->tds[i]->node); node 174 drivers/usb/host/fhci-q.c ed->td_head = list_entry(ed->td_list.next, struct td, node); node 184 drivers/usb/host/fhci-q.c list_del_init(&ed->node); node 66 drivers/usb/host/fhci-sched.c node); node 257 drivers/usb/host/fhci-sched.c struct list_head *node = list->next; node 260 drivers/usb/host/fhci-sched.c list_move_tail(node, list); node 284 drivers/usb/host/fhci-sched.c list_for_each_entry(ed, list, node) { node 874 drivers/usb/host/fhci-sched.c list_add(&ed->node, &fhci->hc_list->ctrl_list); node 877 drivers/usb/host/fhci-sched.c list_add(&ed->node, &fhci->hc_list->bulk_list); node 880 drivers/usb/host/fhci-sched.c list_add(&ed->node, &fhci->hc_list->intr_list); node 883 drivers/usb/host/fhci-sched.c list_add(&ed->node, &fhci->hc_list->iso_list); node 331 drivers/usb/host/fhci.h struct list_head node; node 352 drivers/usb/host/fhci.h struct list_head node; node 221 drivers/usb/host/uhci-debug.c struct urb_priv, node); node 229 drivers/usb/host/uhci-debug.c list_for_each_entry(urbp, &qh->queue, node) { node 506 drivers/usb/host/uhci-debug.c head = &qh->node; node 510 drivers/usb/host/uhci-debug.c qh = list_entry(tmp, struct uhci_qh, node); node 159 drivers/usb/host/uhci-hcd.h struct list_head node; /* Node in the list of QHs */ node 484 drivers/usb/host/uhci-hcd.h struct list_head node; /* Node in the QH's urbp list */ node 55 drivers/usb/host/uhci-q.c lqh = list_entry(uhci->skel_async_qh->node.prev, node 56 drivers/usb/host/uhci-q.c struct uhci_qh, node); node 67 drivers/usb/host/uhci-q.c lqh = list_entry(uhci->skel_async_qh->node.prev, node 68 drivers/usb/host/uhci-q.c struct uhci_qh, node); node 261 drivers/usb/host/uhci-q.c INIT_LIST_HEAD(&qh->node); node 298 drivers/usb/host/uhci-q.c list_del(&qh->node); node 334 drivers/usb/host/uhci-q.c if (qh->queue.next != &urbp->node) { node 338 drivers/usb/host/uhci-q.c purbp = list_entry(urbp->node.prev, struct urb_priv, node); node 383 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); node 393 drivers/usb/host/uhci-q.c urbp = list_prepare_entry(urbp, &qh->queue, node); node 394 drivers/usb/host/uhci-q.c list_for_each_entry_continue(urbp, &qh->queue, node) { node 415 drivers/usb/host/uhci-q.c pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe; node 426 drivers/usb/host/uhci-q.c list_add_tail(&qh->node, &uhci->skel_iso_qh->node); node 439 drivers/usb/host/uhci-q.c list_add_tail(&qh->node, &uhci->skelqh[qh->skel]->node); node 441 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); node 459 drivers/usb/host/uhci-q.c list_for_each_entry_reverse(pqh, &uhci->skel_async_qh->node, node) { node 463 drivers/usb/host/uhci-q.c list_add(&qh->node, &pqh->node); node 488 drivers/usb/host/uhci-q.c struct urb_priv, node); node 506 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, node 507 drivers/usb/host/uhci-q.c node); node 508 drivers/usb/host/uhci-q.c list_del(&qh->node); node 525 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); node 538 drivers/usb/host/uhci-q.c pqh = list_entry(qh->node.prev, struct uhci_qh, node); node 570 drivers/usb/host/uhci-q.c if (list_empty(&uhci->skel_unlink_qh->node) || uhci->is_stopped) node 575 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, node 576 drivers/usb/host/uhci-q.c node); node 577 drivers/usb/host/uhci-q.c list_move_tail(&qh->node, &uhci->skel_unlink_qh->node); node 591 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, node 592 drivers/usb/host/uhci-q.c node); node 593 drivers/usb/host/uhci-q.c list_move(&qh->node, &uhci->idle_qh_list); node 734 drivers/usb/host/uhci-q.c INIT_LIST_HEAD(&urbp->node); node 745 drivers/usb/host/uhci-q.c if (!list_empty(&urbp->node)) node 1298 drivers/usb/host/uhci-q.c struct urb_priv, node)->urb; node 1456 drivers/usb/host/uhci-q.c list_add_tail(&urbp->node, &qh->queue); node 1462 drivers/usb/host/uhci-q.c if (qh->queue.next == &urbp->node && !qh->is_stopped) { node 1534 drivers/usb/host/uhci-q.c urbp->node.prev == &qh->queue && node 1535 drivers/usb/host/uhci-q.c urbp->node.next != &qh->queue) { node 1536 drivers/usb/host/uhci-q.c struct urb *nurb = list_entry(urbp->node.next, node 1537 drivers/usb/host/uhci-q.c struct urb_priv, node)->urb; node 1545 drivers/usb/host/uhci-q.c list_del_init(&urbp->node); node 1582 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); node 1615 drivers/usb/host/uhci-q.c list_for_each_entry(urbp, &qh->queue, node) { node 1641 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); node 1691 drivers/usb/host/uhci-q.c urbp = list_entry(qh->queue.next, struct urb_priv, node); node 1762 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(uhci->skelqh[i]->node.next, node 1763 drivers/usb/host/uhci-q.c struct uhci_qh, node); node 1765 drivers/usb/host/uhci-q.c uhci->next_qh = list_entry(qh->node.next, node 1766 drivers/usb/host/uhci-q.c struct uhci_qh, node); node 1772 drivers/usb/host/uhci-q.c list_entry(qh->queue.next, struct urb_priv, node)); node 1789 drivers/usb/host/uhci-q.c if (list_empty(&uhci->skel_unlink_qh->node)) node 428 drivers/usb/host/xhci-mtk.c struct device_node *node = dev->of_node; node 462 drivers/usb/host/xhci-mtk.c mtk->lpm_support = of_property_read_bool(node, "usb3-lpm-capable"); node 464 drivers/usb/host/xhci-mtk.c of_property_read_u32(node, "mediatek,u3p-dis-msk", node 467 drivers/usb/host/xhci-mtk.c ret = usb_wakeup_of_property_parse(mtk, node); node 102 drivers/usb/host/xhci-rcar.c struct device_node *node = dev->of_node; node 104 drivers/usb/host/xhci-rcar.c return of_device_is_compatible(node, "renesas,xhci-r8a7790") || node 105 drivers/usb/host/xhci-rcar.c of_device_is_compatible(node, "renesas,xhci-r8a7791") || node 106 drivers/usb/host/xhci-rcar.c of_device_is_compatible(node, "renesas,xhci-r8a7793") || node 107 drivers/usb/host/xhci-rcar.c of_device_is_compatible(node, "renesas,rcar-gen2-xhci"); node 216 drivers/usb/mtu3/mtu3_plat.c struct device_node *node = pdev->dev.of_node; node 247 drivers/usb/mtu3/mtu3_plat.c ssusb->num_phys = of_count_phandle_with_args(node, node 259 drivers/usb/mtu3/mtu3_plat.c ssusb->phys[i] = devm_of_phy_get_by_index(dev, node, i); node 279 drivers/usb/mtu3/mtu3_plat.c ret = ssusb_wakeup_of_property_parse(ssusb, node); node 286 drivers/usb/mtu3/mtu3_plat.c of_property_read_u32(node, "mediatek,u3p-dis-msk", node 299 drivers/usb/mtu3/mtu3_plat.c otg_sx->is_u3_drd = of_property_read_bool(node, "mediatek,usb3-drd"); node 301 drivers/usb/mtu3/mtu3_plat.c of_property_read_bool(node, "enable-manual-drd"); node 302 drivers/usb/mtu3/mtu3_plat.c otg_sx->role_sw_used = of_property_read_bool(node, "usb-role-switch"); node 304 drivers/usb/mtu3/mtu3_plat.c if (!otg_sx->role_sw_used && of_property_read_bool(node, "extcon")) { node 322 drivers/usb/mtu3/mtu3_plat.c struct device_node *node = pdev->dev.of_node; node 375 drivers/usb/mtu3/mtu3_plat.c ret = ssusb_host_init(ssusb, node); node 388 drivers/usb/mtu3/mtu3_plat.c ret = ssusb_host_init(ssusb, node); node 2058 drivers/usb/musb/musb_core.c struct list_head node; node 2073 drivers/usb/musb/musb_core.c list_for_each_entry_safe(w, _w, &musb->pending_list, node) { node 2082 drivers/usb/musb/musb_core.c list_del(&w->node); node 2121 drivers/usb/musb/musb_core.c list_add_tail(&w->node, &musb->pending_list); node 123 drivers/usb/phy/phy-am335x-control.c const struct device_node *node = (const struct device_node *)data; node 124 drivers/usb/phy/phy-am335x-control.c return dev->of_node == node && node 130 drivers/usb/phy/phy-am335x-control.c struct device_node *node; node 133 drivers/usb/phy/phy-am335x-control.c node = of_parse_phandle(dev->of_node, "ti,ctrl_mod", 0); node 134 drivers/usb/phy/phy-am335x-control.c if (!node) node 137 drivers/usb/phy/phy-am335x-control.c dev = bus_find_device(&platform_bus_type, NULL, node, match); node 138 drivers/usb/phy/phy-am335x-control.c of_node_put(node); node 217 drivers/usb/phy/phy-generic.c struct device_node *node = dev->of_node; node 219 drivers/usb/phy/phy-generic.c if (of_property_read_u32(node, "clock-frequency", &clk_rate)) node 222 drivers/usb/phy/phy-generic.c needs_vcc = of_property_read_bool(node, "vcc-supply"); node 223 drivers/usb/phy/phy-generic.c needs_clk = of_property_read_bool(node, "clocks"); node 147 drivers/usb/phy/phy-isp1301.c const struct device_node *node = (const struct device_node *)data; node 148 drivers/usb/phy/phy-isp1301.c return (dev->of_node == node) && node 152 drivers/usb/phy/phy-isp1301.c struct i2c_client *isp1301_get_client(struct device_node *node) node 154 drivers/usb/phy/phy-isp1301.c if (node) { /* reference of ISP1301 I2C node via DT */ node 156 drivers/usb/phy/phy-isp1301.c node, match); node 52 drivers/usb/phy/phy.c static struct usb_phy *__of_usb_find_phy(struct device_node *node) node 56 drivers/usb/phy/phy.c if (!of_device_is_available(node)) node 60 drivers/usb/phy/phy.c if (node != phy->dev->of_node) node 487 drivers/usb/phy/phy.c struct device_node *node, node 502 drivers/usb/phy/phy.c phy = __of_usb_find_phy(node); node 548 drivers/usb/phy/phy.c struct device_node *node; node 556 drivers/usb/phy/phy.c node = of_parse_phandle(dev->of_node, phandle, index); node 557 drivers/usb/phy/phy.c if (!node) { node 562 drivers/usb/phy/phy.c phy = devm_usb_get_phy_by_node(dev, node, NULL); node 563 drivers/usb/phy/phy.c of_node_put(node); node 24 drivers/usb/renesas_usbhs/fifo.c INIT_LIST_HEAD(&pkt->node); node 67 drivers/usb/renesas_usbhs/fifo.c list_move_tail(&pkt->node, &pipe->list); node 89 drivers/usb/renesas_usbhs/fifo.c list_del_init(&pkt->node); node 94 drivers/usb/renesas_usbhs/fifo.c return list_first_entry_or_null(&pipe->list, struct usbhs_pkt, node); node 46 drivers/usb/renesas_usbhs/fifo.h struct list_head node; node 678 drivers/usb/renesas_usbhs/mod_gadget.c WARN_ON(!list_empty(&ureq->pkt.node)); node 157 drivers/usb/serial/bus.c list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { node 158 drivers/usb/serial/bus.c list_del(&dynid->node); node 626 drivers/usb/serial/usb-serial.c list_for_each_entry(dynid, &drv->dynids.list, node) { node 1621 drivers/usb/typec/tcpm/fusb302.c struct device_node *node; node 1624 drivers/usb/typec/tcpm/fusb302.c node = chip->dev->of_node; node 1625 drivers/usb/typec/tcpm/fusb302.c chip->gpio_int_n = of_get_named_gpio(node, "fcs,int_n", 0); node 15 drivers/usb/usbip/usbip_event.c struct list_head node; node 48 drivers/usb/usbip/usbip_event.c ue = list_first_entry(&event_list, struct usbip_event, node); node 49 drivers/usb/usbip/usbip_event.c list_del(&ue->node); node 155 drivers/usb/usbip/usbip_event.c list_for_each_entry_reverse(ue, &event_list, node) { node 166 drivers/usb/usbip/usbip_event.c list_add_tail(&ue->node, &event_list); node 722 drivers/vfio/platform/vfio_platform_common.c void __vfio_platform_register_reset(struct vfio_platform_reset_node *node) node 725 drivers/vfio/platform/vfio_platform_common.c list_add(&node->link, &reset_list); node 84 drivers/vfio/vfio_iommu_type1.c struct rb_node node; node 111 drivers/vfio/vfio_iommu_type1.c struct rb_node node; node 137 drivers/vfio/vfio_iommu_type1.c struct rb_node *node = iommu->dma_list.rb_node; node 139 drivers/vfio/vfio_iommu_type1.c while (node) { node 140 drivers/vfio/vfio_iommu_type1.c struct vfio_dma *dma = rb_entry(node, struct vfio_dma, node); node 143 drivers/vfio/vfio_iommu_type1.c node = node->rb_left; node 145 drivers/vfio/vfio_iommu_type1.c node = node->rb_right; node 160 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(parent, struct vfio_dma, node); node 168 drivers/vfio/vfio_iommu_type1.c rb_link_node(&new->node, parent, link); node 169 drivers/vfio/vfio_iommu_type1.c rb_insert_color(&new->node, &iommu->dma_list); node 174 drivers/vfio/vfio_iommu_type1.c rb_erase(&old->node, &iommu->dma_list); node 183 drivers/vfio/vfio_iommu_type1.c struct rb_node *node = dma->pfn_list.rb_node; node 185 drivers/vfio/vfio_iommu_type1.c while (node) { node 186 drivers/vfio/vfio_iommu_type1.c vpfn = rb_entry(node, struct vfio_pfn, node); node 189 drivers/vfio/vfio_iommu_type1.c node = node->rb_left; node 191 drivers/vfio/vfio_iommu_type1.c node = node->rb_right; node 207 drivers/vfio/vfio_iommu_type1.c vpfn = rb_entry(parent, struct vfio_pfn, node); node 215 drivers/vfio/vfio_iommu_type1.c rb_link_node(&new->node, parent, link); node 216 drivers/vfio/vfio_iommu_type1.c rb_insert_color(&new->node, &dma->pfn_list); node 221 drivers/vfio/vfio_iommu_type1.c rb_erase(&old->node, &dma->pfn_list); node 1057 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *node; node 1059 drivers/vfio/vfio_iommu_type1.c list_for_each_entry(node, iova, list) { node 1060 drivers/vfio/vfio_iommu_type1.c if (start >= node->start && end <= node->end) node 1203 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); node 1497 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *node, *next; node 1503 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(node, next, iova, list) { node 1504 drivers/vfio/vfio_iommu_type1.c if (start < node->start) node 1506 drivers/vfio/vfio_iommu_type1.c if (start >= node->start && start < node->end) { node 1507 drivers/vfio/vfio_iommu_type1.c node->start = start; node 1511 drivers/vfio/vfio_iommu_type1.c list_del(&node->list); node 1512 drivers/vfio/vfio_iommu_type1.c kfree(node); node 1516 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(node, next, iova, list) { node 1517 drivers/vfio/vfio_iommu_type1.c if (end > node->end) node 1519 drivers/vfio/vfio_iommu_type1.c if (end > node->start && end <= node->end) { node 1520 drivers/vfio/vfio_iommu_type1.c node->end = end; node 1524 drivers/vfio/vfio_iommu_type1.c list_del(&node->list); node 1525 drivers/vfio/vfio_iommu_type1.c kfree(node); node 1856 drivers/vfio/vfio_iommu_type1.c struct rb_node *node; node 1858 drivers/vfio/vfio_iommu_type1.c while ((node = rb_first(&iommu->dma_list))) node 1859 drivers/vfio/vfio_iommu_type1.c vfio_remove_dma(iommu, rb_entry(node, struct vfio_dma, node)); node 1871 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); node 1876 drivers/vfio/vfio_iommu_type1.c node); node 1893 drivers/vfio/vfio_iommu_type1.c dma = rb_entry(n, struct vfio_dma, node); node 1912 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *node; node 1929 drivers/vfio/vfio_iommu_type1.c node = list_first_entry(iova_copy, struct vfio_iova, list); node 1930 drivers/vfio/vfio_iommu_type1.c node->start = start; node 1931 drivers/vfio/vfio_iommu_type1.c node = list_last_entry(iova_copy, struct vfio_iova, list); node 1932 drivers/vfio/vfio_iommu_type1.c node->end = end; node 1946 drivers/vfio/vfio_iommu_type1.c struct vfio_iova *node; node 1963 drivers/vfio/vfio_iommu_type1.c node = list_first_entry(iova_copy, struct vfio_iova, list); node 1964 drivers/vfio/vfio_iommu_type1.c start = node->start; node 1965 drivers/vfio/vfio_iommu_type1.c node = list_last_entry(iova_copy, struct vfio_iova, list); node 1966 drivers/vfio/vfio_iommu_type1.c end = node->end; node 265 drivers/vhost/vhost.c llist_add(&work->node, &dev->work_list); node 335 drivers/vhost/vhost.c struct llist_node *node; node 350 drivers/vhost/vhost.c node = llist_del_all(&dev->work_list); node 351 drivers/vhost/vhost.c if (!node) node 354 drivers/vhost/vhost.c node = llist_reverse_order(node); node 357 drivers/vhost/vhost.c llist_for_each_entry_safe(work, work_next, node, node) { node 617 drivers/vhost/vhost.c struct vhost_umem_node *node) node 619 drivers/vhost/vhost.c vhost_umem_interval_tree_remove(node, &umem->umem_tree); node 620 drivers/vhost/vhost.c list_del(&node->link); node 621 drivers/vhost/vhost.c kfree(node); node 627 drivers/vhost/vhost.c struct vhost_umem_node *node, *tmp; node 632 drivers/vhost/vhost.c list_for_each_entry_safe(node, tmp, &umem->umem_list, link) node 633 drivers/vhost/vhost.c vhost_umem_free(umem, node); node 640 drivers/vhost/vhost.c struct vhost_msg_node *node, *n; node 644 drivers/vhost/vhost.c list_for_each_entry_safe(node, n, &dev->read_list, node) { node 645 drivers/vhost/vhost.c list_del(&node->node); node 646 drivers/vhost/vhost.c kfree(node); node 649 drivers/vhost/vhost.c list_for_each_entry_safe(node, n, &dev->pending_list, node) { node 650 drivers/vhost/vhost.c list_del(&node->node); node 651 drivers/vhost/vhost.c kfree(node); node 715 drivers/vhost/vhost.c struct vhost_umem_node *node; node 720 drivers/vhost/vhost.c list_for_each_entry(node, &umem->umem_list, link) { node 721 drivers/vhost/vhost.c unsigned long a = node->userspace_addr; node 723 drivers/vhost/vhost.c if (vhost_overflow(node->userspace_addr, node->size)) node 728 drivers/vhost/vhost.c node->size)) node 731 drivers/vhost/vhost.c node->start, node 732 drivers/vhost/vhost.c node->size)) node 742 drivers/vhost/vhost.c const struct vhost_umem_node *node = vq->meta_iotlb[type]; node 744 drivers/vhost/vhost.c if (!node) node 747 drivers/vhost/vhost.c return (void *)(uintptr_t)(node->userspace_addr + addr - node->start); node 1021 drivers/vhost/vhost.c struct vhost_umem_node *tmp, *node; node 1026 drivers/vhost/vhost.c node = kmalloc(sizeof(*node), GFP_ATOMIC); node 1027 drivers/vhost/vhost.c if (!node) node 1035 drivers/vhost/vhost.c node->start = start; node 1036 drivers/vhost/vhost.c node->size = size; node 1037 drivers/vhost/vhost.c node->last = end; node 1038 drivers/vhost/vhost.c node->userspace_addr = userspace_addr; node 1039 drivers/vhost/vhost.c node->perm = perm; node 1040 drivers/vhost/vhost.c INIT_LIST_HEAD(&node->link); node 1041 drivers/vhost/vhost.c list_add_tail(&node->link, &umem->umem_list); node 1042 drivers/vhost/vhost.c vhost_umem_interval_tree_insert(node, &umem->umem_tree); node 1051 drivers/vhost/vhost.c struct vhost_umem_node *node; node 1053 drivers/vhost/vhost.c while ((node = vhost_umem_interval_tree_iter_first(&umem->umem_tree, node 1055 drivers/vhost/vhost.c vhost_umem_free(umem, node); node 1061 drivers/vhost/vhost.c struct vhost_msg_node *node, *n; node 1065 drivers/vhost/vhost.c list_for_each_entry_safe(node, n, &d->pending_list, node) { node 1066 drivers/vhost/vhost.c struct vhost_iotlb_msg *vq_msg = &node->msg.iotlb; node 1070 drivers/vhost/vhost.c vhost_poll_queue(&node->vq->poll); node 1071 drivers/vhost/vhost.c list_del(&node->node); node 1072 drivers/vhost/vhost.c kfree(node); node 1205 drivers/vhost/vhost.c struct vhost_msg_node *node; node 1217 drivers/vhost/vhost.c node = vhost_dequeue_msg(dev, &dev->read_list); node 1218 drivers/vhost/vhost.c if (node) node 1239 drivers/vhost/vhost.c if (node) { node 1241 drivers/vhost/vhost.c void *start = &node->msg; node 1243 drivers/vhost/vhost.c switch (node->msg.type) { node 1245 drivers/vhost/vhost.c size = sizeof(node->msg); node 1246 drivers/vhost/vhost.c msg = &node->msg.iotlb; node 1249 drivers/vhost/vhost.c size = sizeof(node->msg_v2); node 1250 drivers/vhost/vhost.c msg = &node->msg_v2.iotlb; node 1259 drivers/vhost/vhost.c kfree(node); node 1262 drivers/vhost/vhost.c vhost_enqueue_msg(dev, &dev->pending_list, node); node 1272 drivers/vhost/vhost.c struct vhost_msg_node *node; node 1276 drivers/vhost/vhost.c node = vhost_new_msg(vq, v2 ? VHOST_IOTLB_MSG_V2 : VHOST_IOTLB_MSG); node 1277 drivers/vhost/vhost.c if (!node) node 1281 drivers/vhost/vhost.c node->msg_v2.type = VHOST_IOTLB_MSG_V2; node 1282 drivers/vhost/vhost.c msg = &node->msg_v2.iotlb; node 1284 drivers/vhost/vhost.c msg = &node->msg.iotlb; node 1291 drivers/vhost/vhost.c vhost_enqueue_msg(dev, &dev->read_list, node); node 1308 drivers/vhost/vhost.c const struct vhost_umem_node *node, node 1314 drivers/vhost/vhost.c if (likely(node->perm & access)) node 1315 drivers/vhost/vhost.c vq->meta_iotlb[type] = node; node 1321 drivers/vhost/vhost.c const struct vhost_umem_node *node; node 1329 drivers/vhost/vhost.c node = vhost_umem_interval_tree_iter_first(&umem->umem_tree, node 1332 drivers/vhost/vhost.c if (node == NULL || node->start > addr) { node 1335 drivers/vhost/vhost.c } else if (!(node->perm & access)) { node 1342 drivers/vhost/vhost.c size = node->size - addr + node->start; node 1345 drivers/vhost/vhost.c vhost_vq_meta_update(vq, node, type); node 2043 drivers/vhost/vhost.c const struct vhost_umem_node *node; node 2057 drivers/vhost/vhost.c node = vhost_umem_interval_tree_iter_first(&umem->umem_tree, node 2059 drivers/vhost/vhost.c if (node == NULL || node->start > addr) { node 2066 drivers/vhost/vhost.c } else if (!(node->perm & access)) { node 2072 drivers/vhost/vhost.c size = node->size - addr + node->start; node 2075 drivers/vhost/vhost.c (node->userspace_addr + addr - node->start); node 2580 drivers/vhost/vhost.c struct vhost_msg_node *node = kmalloc(sizeof *node, GFP_KERNEL); node 2581 drivers/vhost/vhost.c if (!node) node 2585 drivers/vhost/vhost.c memset(&node->msg, 0, sizeof node->msg); node 2586 drivers/vhost/vhost.c node->vq = vq; node 2587 drivers/vhost/vhost.c node->msg.type = type; node 2588 drivers/vhost/vhost.c return node; node 2593 drivers/vhost/vhost.c struct vhost_msg_node *node) node 2596 drivers/vhost/vhost.c list_add_tail(&node->node, head); node 2606 drivers/vhost/vhost.c struct vhost_msg_node *node = NULL; node 2610 drivers/vhost/vhost.c node = list_first_entry(head, struct vhost_msg_node, node 2611 drivers/vhost/vhost.c node); node 2612 drivers/vhost/vhost.c list_del(&node->node); node 2616 drivers/vhost/vhost.c return node; node 21 drivers/vhost/vhost.h struct llist_node node; node 55 drivers/vhost/vhost.h #define START(node) ((node)->start) node 56 drivers/vhost/vhost.h #define LAST(node) ((node)->last) node 156 drivers/vhost/vhost.h struct list_head node; node 220 drivers/vhost/vhost.h struct vhost_msg_node *node); node 178 drivers/video/backlight/arcxcnn_bl.c struct device_node *node = dev->of_node; node 183 drivers/video/backlight/arcxcnn_bl.c if (!node) node 186 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_string(node, "label", &lp->pdata->name); node 190 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "default-brightness", &prog_val); node 194 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,led-config-0", &prog_val); node 198 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,led-config-1", &prog_val); node 202 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,dim-freq", &prog_val); node 206 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,comp-config", &prog_val); node 210 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,filter-config", &prog_val); node 214 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32(node, "arc,trim-config", &prog_val); node 218 drivers/video/backlight/arcxcnn_bl.c ret = of_property_count_u32_elems(node, "led-sources"); node 226 drivers/video/backlight/arcxcnn_bl.c ret = of_property_read_u32_array(node, "led-sources", sources, node 52 drivers/video/backlight/backlight.c int node = evdata->info->node; node 66 drivers/video/backlight/backlight.c !bd->fb_bl_on[node]) { node 67 drivers/video/backlight/backlight.c bd->fb_bl_on[node] = true; node 74 drivers/video/backlight/backlight.c bd->fb_bl_on[node]) { node 75 drivers/video/backlight/backlight.c bd->fb_bl_on[node] = false; node 592 drivers/video/backlight/backlight.c struct backlight_device *of_find_backlight_by_node(struct device_node *node) node 596 drivers/video/backlight/backlight.c dev = class_find_device(backlight_class, NULL, node, of_parent_match); node 82 drivers/video/backlight/gpio_backlight.c struct device_node *node = gbl->dev->of_node; node 85 drivers/video/backlight/gpio_backlight.c if (!node || !node->phandle) node 374 drivers/video/backlight/lm3630a_bl.c static int lm3630a_parse_led_sources(struct fwnode_handle *node, node 380 drivers/video/backlight/lm3630a_bl.c num_sources = fwnode_property_count_u32(node, "led-sources"); node 386 drivers/video/backlight/lm3630a_bl.c ret = fwnode_property_read_u32_array(node, "led-sources", sources, node 402 drivers/video/backlight/lm3630a_bl.c struct fwnode_handle *node, int *seen_led_sources) node 409 drivers/video/backlight/lm3630a_bl.c ret = fwnode_property_read_u32(node, "reg", &bank); node 416 drivers/video/backlight/lm3630a_bl.c led_sources = lm3630a_parse_led_sources(node, BIT(bank)); node 425 drivers/video/backlight/lm3630a_bl.c linear = fwnode_property_read_bool(node, node 447 drivers/video/backlight/lm3630a_bl.c ret = fwnode_property_read_string(node, "label", &label); node 455 drivers/video/backlight/lm3630a_bl.c ret = fwnode_property_read_u32(node, "default-brightness", node 464 drivers/video/backlight/lm3630a_bl.c ret = fwnode_property_read_u32(node, "max-brightness", &val); node 479 drivers/video/backlight/lm3630a_bl.c struct fwnode_handle *node; node 481 drivers/video/backlight/lm3630a_bl.c device_for_each_child_node(pchip->dev, node) { node 482 drivers/video/backlight/lm3630a_bl.c ret = lm3630a_parse_bank(pdata, node, &seen_led_sources); node 348 drivers/video/backlight/lp855x_bl.c struct device_node *node = dev->of_node; node 352 drivers/video/backlight/lp855x_bl.c if (!node) { node 361 drivers/video/backlight/lp855x_bl.c of_property_read_string(node, "bl-name", &pdata->name); node 362 drivers/video/backlight/lp855x_bl.c of_property_read_u8(node, "dev-ctrl", &pdata->device_control); node 363 drivers/video/backlight/lp855x_bl.c of_property_read_u8(node, "init-brt", &pdata->initial_brightness); node 364 drivers/video/backlight/lp855x_bl.c of_property_read_u32(node, "pwm-period", &pdata->period_ns); node 367 drivers/video/backlight/lp855x_bl.c rom_length = of_get_child_count(node); node 377 drivers/video/backlight/lp855x_bl.c for_each_child_of_node(node, child) { node 228 drivers/video/backlight/pwm_bl.c struct device_node *node = dev->of_node; node 238 drivers/video/backlight/pwm_bl.c if (!node) node 247 drivers/video/backlight/pwm_bl.c of_property_read_u32(node, "post-pwm-on-delay-ms", node 249 drivers/video/backlight/pwm_bl.c of_property_read_u32(node, "pwm-off-delay-ms", &data->pwm_off_delay); node 257 drivers/video/backlight/pwm_bl.c prop = of_find_property(node, "brightness-levels", &length); node 272 drivers/video/backlight/pwm_bl.c ret = of_property_read_u32_array(node, "brightness-levels", node 278 drivers/video/backlight/pwm_bl.c ret = of_property_read_u32(node, "default-brightness-level", node 290 drivers/video/backlight/pwm_bl.c of_property_read_u32(node, "num-interpolated-steps", node 417 drivers/video/backlight/pwm_bl.c struct device_node *node = pb->dev->of_node; node 420 drivers/video/backlight/pwm_bl.c if (!node || !node->phandle) node 451 drivers/video/backlight/pwm_bl.c struct device_node *node = pdev->dev.of_node; node 530 drivers/video/backlight/pwm_bl.c if (IS_ERR(pb->pwm) && PTR_ERR(pb->pwm) != -EPROBE_DEFER && !node) { node 187 drivers/video/backlight/tps65217_bl.c struct device_node *node; node 191 drivers/video/backlight/tps65217_bl.c node = of_get_child_by_name(tps->dev->of_node, "backlight"); node 192 drivers/video/backlight/tps65217_bl.c if (!node) node 202 drivers/video/backlight/tps65217_bl.c if (!of_property_read_u32(node, "isel", &val)) { node 215 drivers/video/backlight/tps65217_bl.c if (!of_property_read_u32(node, "fdim", &val)) { node 241 drivers/video/backlight/tps65217_bl.c if (!of_property_read_u32(node, "default-brightness", &val)) { node 252 drivers/video/backlight/tps65217_bl.c of_node_put(node); node 257 drivers/video/backlight/tps65217_bl.c of_node_put(node); node 556 drivers/video/fbdev/amba-clcd.c static int clcdfb_of_get_dpi_panel_mode(struct device_node *node, node 563 drivers/video/fbdev/amba-clcd.c err = of_get_display_timing(node, "panel-timing", &timing); node 565 drivers/video/fbdev/amba-clcd.c pr_err("%pOF: problems parsing panel-timing (%d)\n", node, err); node 849 drivers/video/fbdev/amba-clcd.c struct device_node *node = dev->dev.of_node; node 854 drivers/video/fbdev/amba-clcd.c board->name = of_node_full_name(node); node 858 drivers/video/fbdev/amba-clcd.c if (of_find_property(node, "memory-region", NULL)) { node 598 drivers/video/fbdev/arkfb.c rv = svga_check_timings (&ark_timing_regs, var, info->node); node 785 drivers/video/fbdev/arkfb.c hmul, info->node); node 1212 drivers/video/fbdev/atmel_lcdfb.c info->node, info->fix.mmio_start, sinfo->mmio, sinfo->irq_base); node 1851 drivers/video/fbdev/aty/aty128fb.c snprintf(name, sizeof(name), "aty128bl%d", info->node); node 2219 drivers/video/fbdev/aty/atyfb_base.c snprintf(name, sizeof(name), "atybl%d", info->node); node 2722 drivers/video/fbdev/aty/atyfb_base.c info->node, info->fix.id, par->bus_type == ISA ? "ISA" : "PCI"); node 149 drivers/video/fbdev/aty/radeon_backlight.c snprintf(name, sizeof(name), "radeonbl%d", rinfo->info->node); node 422 drivers/video/fbdev/chipsfb.c p->node, p->fix.smem_len / 1024); node 181 drivers/video/fbdev/core/fb_defio.c struct list_head *node, *next; node 197 drivers/video/fbdev/core/fb_defio.c list_for_each_safe(node, next, &fbdefio->pagelist) { node 198 drivers/video/fbdev/core/fb_defio.c list_del(node); node 3090 drivers/video/fbdev/core/fbcon.c int idx = info->node; node 3147 drivers/video/fbdev/core/fbcon.c idx = info->node; node 3178 drivers/video/fbdev/core/fbcon.c int i, idx = info->node; node 3209 drivers/video/fbdev/core/fbcon.c info->fix.id, info->node); node 3210 drivers/video/fbdev/core/fbcon.c primary_device = info->node; node 3217 drivers/video/fbdev/core/fbcon.c "fb%i, to tty %i-%i\n", info->node, node 3238 drivers/video/fbdev/core/fbcon.c idx = info->node; node 3321 drivers/video/fbdev/core/fbcon.c info->node == con2fb_map[i]) { node 3335 drivers/video/fbdev/core/fbcon.c info->node == con2fb_map[fg_console]) { node 727 drivers/video/fbdev/core/fbmem.c seq_printf(m, "%d %s\n", fi->node, fi->fix.id); node 1595 drivers/video/fbdev/core/fbmem.c fb_info->node = i; node 1664 drivers/video/fbdev/core/fbmem.c int i = fb_info->node; node 1680 drivers/video/fbdev/core/fbmem.c i = fb_info->node; node 1704 drivers/video/fbdev/core/fbmem.c registered_fb[fb_info->node] = NULL; node 170 drivers/video/fbdev/core/svgalib.c void svga_dump_var(struct fb_var_screeninfo *var, int node) node 172 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.vmode : 0x%X\n", node, var->vmode); node 173 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.xres : %d\n", node, var->xres); node 174 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.yres : %d\n", node, var->yres); node 175 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.bits_per_pixel: %d\n", node, var->bits_per_pixel); node 176 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.xres_virtual : %d\n", node, var->xres_virtual); node 177 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.yres_virtual : %d\n", node, var->yres_virtual); node 178 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.left_margin : %d\n", node, var->left_margin); node 179 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.right_margin : %d\n", node, var->right_margin); node 180 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.upper_margin : %d\n", node, var->upper_margin); node 181 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.lower_margin : %d\n", node, var->lower_margin); node 182 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.hsync_len : %d\n", node, var->hsync_len); node 183 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.vsync_len : %d\n", node, var->vsync_len); node 184 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.sync : 0x%X\n", node, var->sync); node 185 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: var.pixclock : %d\n\n", node, var->pixclock); node 381 drivers/video/fbdev/core/svgalib.c int svga_compute_pll(const struct svga_pll *pll, u32 f_wanted, u16 *m, u16 *n, u16 *r, int node) node 386 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: ideal frequency: %d kHz\n", node, (unsigned int) f_wanted); node 433 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: found frequency: %d kHz (VCO %d kHz)\n", node, (int) (f_current >> ar), (int) f_current); node 434 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: m = %d n = %d r = %d\n", node, (unsigned int) *m, (unsigned int) *n, (unsigned int) *r); node 443 drivers/video/fbdev/core/svgalib.c int svga_check_timings(const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, int node) node 512 drivers/video/fbdev/core/svgalib.c u32 hmul, u32 hdiv, u32 vmul, u32 vdiv, u32 hborder, int node) node 519 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal total : %d\n", node, value); node 524 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal display : %d\n", node, value); node 529 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal blank start: %d\n", node, value); node 534 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal blank end : %d\n", node, value); node 539 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal sync start : %d\n", node, value); node 544 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: horizontal sync end : %d\n", node, value); node 549 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical total : %d\n", node, value); node 554 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical display : %d\n", node, value); node 559 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical blank start : %d\n", node, value); node 564 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical blank end : %d\n", node, value); node 569 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical sync start : %d\n", node, value); node 574 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: vertical sync end : %d\n", node, value); node 581 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: positive horizontal sync\n", node); node 584 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: negative horizontal sync\n", node); node 588 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: positive vertical sync\n", node); node 591 drivers/video/fbdev/core/svgalib.c pr_debug("fb%d: negative vertical sync\n\n", node); node 520 drivers/video/fbdev/ep93xx-fb.c info->node = -1; node 297 drivers/video/fbdev/geode/gx1fb_core.c info->node = -1; node 311 drivers/video/fbdev/geode/gxfb_core.c info->node = -1; node 432 drivers/video/fbdev/geode/lxfb_core.c info->node = -1; node 476 drivers/video/fbdev/grvga.c info->node, info->var.xres, info->var.yres, info->var.bits_per_pixel, node 2084 drivers/video/fbdev/i810/i810_main.c info->node, node 113 drivers/video/fbdev/matrox/i2c-matroxfb.c minfo->fbcon.node); node 215 drivers/video/fbdev/matrox/i2c-matroxfb.c .node = LIST_HEAD_INIT(i2c_matroxfb.node), node 1955 drivers/video/fbdev/matrox/matroxfb_base.c #define matroxfb_driver_l(x) list_entry(x, struct matroxfb_driver, node) node 1959 drivers/video/fbdev/matrox/matroxfb_base.c list_add(&drv->node, &matroxfb_driver_list); node 1979 drivers/video/fbdev/matrox/matroxfb_base.c list_del(&drv->node); node 2003 drivers/video/fbdev/matrox/matroxfb_base.c drv = matroxfb_driver_l(drv->node.next)) { node 507 drivers/video/fbdev/matrox/matroxfb_base.h struct list_head node; node 660 drivers/video/fbdev/matrox/matroxfb_crtc2.c minfo->fbcon.node, m2info->fbcon.node); node 683 drivers/video/fbdev/matrox/matroxfb_crtc2.c id = m2info->fbcon.node; node 728 drivers/video/fbdev/metronomefb.c " memory\n", info->node, videomemorysize >> 10); node 79 drivers/video/fbdev/mmp/core.c list_add_tail(&panel->node, &panel_list); node 82 drivers/video/fbdev/mmp/core.c list_for_each_entry(path, &path_list, node) { node 107 drivers/video/fbdev/mmp/core.c list_del(&panel->node); node 109 drivers/video/fbdev/mmp/core.c list_for_each_entry(path, &path_list, node) { node 134 drivers/video/fbdev/mmp/core.c list_for_each_entry(path, &path_list, node) { node 176 drivers/video/fbdev/mmp/core.c list_for_each_entry(panel, &panel_list, node) { node 204 drivers/video/fbdev/mmp/core.c list_add_tail(&path->node, &path_list); node 226 drivers/video/fbdev/mmp/core.c list_del(&path->node); node 509 drivers/video/fbdev/mmp/fb/mmpfb.c info->node = -1; node 633 drivers/video/fbdev/mmp/fb/mmpfb.c info->node, info->fix.id); node 102 drivers/video/fbdev/nvidia/nv_backlight.c snprintf(name, sizeof(name), "nvidiabl%d", info->node); node 213 drivers/video/fbdev/omap2/omapfb/displays/connector-analog-tv.c struct device_node *node = pdev->dev.of_node; node 216 drivers/video/fbdev/omap2/omapfb/displays/connector-analog-tv.c in = omapdss_of_find_source_for_first_ep(node); node 238 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c struct device_node *node = pdev->dev.of_node; node 243 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c in = omapdss_of_find_source_for_first_ep(node); node 251 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c adapter_node = of_parse_phandle(node, "ddc-i2c-bus", 0); node 208 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c struct device_node *node = pdev->dev.of_node; node 213 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c gpio = of_get_named_gpio(node, "hpd-gpios", 0); node 219 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c in = omapdss_of_find_source_for_first_ep(node); node 182 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c struct device_node *node = pdev->dev.of_node; node 190 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c if (node == NULL) { node 207 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c in = omapdss_of_find_source_for_first_ep(node); node 168 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c struct device_node *node = pdev->dev.of_node; node 172 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c gpio = of_get_named_gpio(node, "powerdown-gpios", 0); node 181 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c in = omapdss_of_find_source_for_first_ep(node); node 203 drivers/video/fbdev/omap2/omapfb/displays/encoder-tpd12s015.c struct device_node *node = pdev->dev.of_node; node 206 drivers/video/fbdev/omap2/omapfb/displays/encoder-tpd12s015.c in = omapdss_of_find_source_for_first_ep(node); node 200 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c struct device_node *node = pdev->dev.of_node; node 215 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c r = of_get_display_timing(node, "panel-timing", &timing); node 224 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c in = omapdss_of_find_source_for_first_ep(node); node 1120 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c struct device_node *node = pdev->dev.of_node; node 1125 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c gpio = of_get_named_gpio(node, "reset-gpios", 0); node 1132 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c gpio = of_get_named_gpio(node, "te-gpios", 0); node 1140 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c in = omapdss_of_find_source_for_first_ep(node); node 241 drivers/video/fbdev/omap2/omapfb/displays/panel-lgphilips-lb035q02.c struct device_node *node = spi->dev.of_node; node 256 drivers/video/fbdev/omap2/omapfb/displays/panel-lgphilips-lb035q02.c in = omapdss_of_find_source_for_first_ep(node); node 233 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c struct device_node *node = spi->dev.of_node; node 238 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c gpio = of_get_named_gpio(node, "reset-gpios", 0); node 248 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c in = omapdss_of_find_source_for_first_ep(node); node 214 drivers/video/fbdev/omap2/omapfb/displays/panel-sharp-ls037v7dw01.c struct device_node *node = pdev->dev.of_node; node 249 drivers/video/fbdev/omap2/omapfb/displays/panel-sharp-ls037v7dw01.c in = omapdss_of_find_source_for_first_ep(node); node 358 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td028ttec1.c struct device_node *node = spi->dev.of_node; node 362 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td028ttec1.c in = omapdss_of_find_source_for_first_ep(node); node 465 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c struct device_node *node = spi->dev.of_node; node 470 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c gpio = of_get_named_gpio(node, "reset-gpios", 0); node 477 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c in = omapdss_of_find_source_for_first_ep(node); node 5072 drivers/video/fbdev/omap2/omapfb/dss/dsi.c struct device_node *node = pdev->dev.of_node; node 5081 drivers/video/fbdev/omap2/omapfb/dss/dsi.c ep = omapdss_of_get_first_endpoint(node); node 120 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c static struct device_node *omapdss_of_get_remote_port(const struct device_node *node) node 124 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c np = of_graph_get_remote_endpoint(node); node 152 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c omapdss_of_find_source_for_first_ep(struct device_node *node) node 158 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c ep = omapdss_of_get_first_endpoint(node); node 531 drivers/video/fbdev/omap2/omapfb/dss/hdmi4.c struct device_node *node = pdev->dev.of_node; node 535 drivers/video/fbdev/omap2/omapfb/dss/hdmi4.c ep = omapdss_of_get_first_endpoint(node); node 561 drivers/video/fbdev/omap2/omapfb/dss/hdmi5.c struct device_node *node = pdev->dev.of_node; node 565 drivers/video/fbdev/omap2/omapfb/dss/hdmi5.c ep = omapdss_of_get_first_endpoint(node); node 28 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c struct device_node *node; node 44 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_update_prop(struct device_node *node, char *compat, node 57 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c of_update_property(node, prop); node 80 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_omapify_node(struct device_node *node) node 87 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c prop = of_find_property(node, "compatible", NULL); node 108 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c omapdss_update_prop(node, new_compat, new_len); node 111 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_add_to_list(struct device_node *node, bool root) node 116 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c n->node = node; node 122 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static bool __init omapdss_list_contains(const struct device_node *node) node 127 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c if (n->node == node) node 134 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_walk_device(struct device_node *node, bool root) node 138 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c omapdss_add_to_list(node, root); node 144 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c n = of_get_child_by_name(node, "ports"); node 146 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c n = of_get_child_by_name(node, "port"); node 153 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c while ((n = of_graph_get_next_endpoint(node, n)) != NULL) { node 208 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c omapdss_omapify_node(n->node); node 211 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c of_node_put(n->node); node 804 drivers/video/fbdev/omap2/omapfb/dss/venc.c struct device_node *node = pdev->dev.of_node; node 809 drivers/video/fbdev/omap2/omapfb/dss/venc.c ep = omapdss_of_get_first_endpoint(node); node 229 drivers/video/fbdev/pmag-aa-fb.c info->node, info->fix.id, dev_name(dev)); node 652 drivers/video/fbdev/pxa168fb.c info->node = -1; node 892 drivers/video/fbdev/pxafb.c ofb->fb.node = -1; node 1830 drivers/video/fbdev/pxafb.c fbi->fb.node = -1; node 340 drivers/video/fbdev/riva/fbdev.c snprintf(name, sizeof(name), "rivabl%d", info->node); node 990 drivers/video/fbdev/s3c2410fb.c fbinfo->node, fbinfo->fix.id); node 466 drivers/video/fbdev/s3fb.c 1000000000 / pixclock, &m, &n, &r, info->node); node 584 drivers/video/fbdev/s3fb.c rv = svga_check_timings (&s3_timing_regs, var, info->node); node 591 drivers/video/fbdev/s3fb.c info->node); node 892 drivers/video/fbdev/s3fb.c hmul, info->node); node 494 drivers/video/fbdev/simplefb.c dev_info(&pdev->dev, "fb%d: simplefb registered!\n", info->node); node 1926 drivers/video/fbdev/sm501fb.c dev_info(info->dev, "fb%d: %s frame buffer\n", fbi->node, fbi->fix.id); node 1097 drivers/video/fbdev/smscufx.c info->node, user, info, dev->fb_count); node 1133 drivers/video/fbdev/smscufx.c int node = info->node; node 1150 drivers/video/fbdev/smscufx.c pr_debug("fb_info for /dev/fb%d has been freed", node); node 1177 drivers/video/fbdev/smscufx.c info->node, user, dev->fb_count); node 1719 drivers/video/fbdev/smscufx.c " Using %dK framebuffer memory\n", info->node, node 1811 drivers/video/fbdev/smscufx.c struct list_head *node; node 1828 drivers/video/fbdev/smscufx.c node = dev->urbs.list.next; /* have reserved one with sem */ node 1829 drivers/video/fbdev/smscufx.c list_del_init(node); node 1833 drivers/video/fbdev/smscufx.c unode = list_entry(node, struct urb_node, entry); node 1840 drivers/video/fbdev/smscufx.c kfree(node); node 595 drivers/video/fbdev/ssd1307fb.c struct device_node *node = client->dev.of_node; node 602 drivers/video/fbdev/ssd1307fb.c if (!node) { node 638 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,width", &par->width)) node 641 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,height", &par->height)) node 644 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,page-offset", &par->page_offset)) node 647 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,com-offset", &par->com_offset)) node 650 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,prechargep1", &par->prechargep1)) node 653 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,prechargep2", &par->prechargep2)) node 656 drivers/video/fbdev/ssd1307fb.c if (!of_property_read_u8_array(node, "solomon,lookup-table", node 661 drivers/video/fbdev/ssd1307fb.c par->seg_remap = !of_property_read_bool(node, "solomon,segment-no-remap"); node 662 drivers/video/fbdev/ssd1307fb.c par->com_seq = of_property_read_bool(node, "solomon,com-seq"); node 663 drivers/video/fbdev/ssd1307fb.c par->com_lrremap = of_property_read_bool(node, "solomon,com-lrremap"); node 664 drivers/video/fbdev/ssd1307fb.c par->com_invdir = of_property_read_bool(node, "solomon,com-invdir"); node 666 drivers/video/fbdev/ssd1307fb.c of_property_read_bool(node, "solomon,area-color-enable"); node 667 drivers/video/fbdev/ssd1307fb.c par->low_power = of_property_read_bool(node, "solomon,low-power"); node 673 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,dclk-div", &par->dclk_div)) node 675 drivers/video/fbdev/ssd1307fb.c if (of_property_read_u32(node, "solomon,dclk-frq", &par->dclk_frq)) node 745 drivers/video/fbdev/ssd1307fb.c snprintf(bl_name, sizeof(bl_name), "ssd1307fb%d", info->node); node 759 drivers/video/fbdev/ssd1307fb.c dev_info(&client->dev, "fb%d: %s framebuffer device registered, using %d bytes of video memory\n", info->node, info->fix.id, vmem_size); node 1698 drivers/video/fbdev/tridentfb.c info->node, info->fix.id, info->var.xres, node 1847 drivers/video/fbdev/udlfb.c struct list_head *node; node 1857 drivers/video/fbdev/udlfb.c node = dlfb->urbs.list.next; /* have reserved one with sem */ node 1858 drivers/video/fbdev/udlfb.c list_del_init(node); node 1862 drivers/video/fbdev/udlfb.c unode = list_entry(node, struct urb_node, entry); node 1869 drivers/video/fbdev/udlfb.c kfree(node); node 1778 drivers/video/fbdev/via/viafbdev.c viafbinfo->node = 0; node 1888 drivers/video/fbdev/via/viafbdev.c viafbinfo->node, viafbinfo->fix.id, default_var.xres, node 309 drivers/video/fbdev/vt8500lcdfb.c fbi->fb.node = -1; node 257 drivers/video/fbdev/vt8623fb.c rv = svga_compute_pll(&vt8623_pll, 1000000000 / pixclock, &m, &n, &r, info->node); node 360 drivers/video/fbdev/vt8623fb.c rv = svga_check_timings (&vt8623_timing_regs, var, info->node); node 505 drivers/video/fbdev/vt8623fb.c 1, info->node); node 713 drivers/video/fbdev/w100fb.c info->node = -1; node 296 drivers/video/fbdev/wm8505fb.c fbi->fb.node = -1; node 796 drivers/virt/fsl_hypervisor.c struct device_node *node; node 799 drivers/virt/fsl_hypervisor.c node = of_find_node_by_path("/hypervisor"); node 800 drivers/virt/fsl_hypervisor.c if (!node) node 803 drivers/virt/fsl_hypervisor.c ret = of_find_property(node, "fsl,hv-version", NULL) != NULL; node 805 drivers/virt/fsl_hypervisor.c of_node_put(node); node 100 drivers/virtio/virtio_mmio.c struct list_head node; node 302 drivers/virtio/virtio_mmio.c list_for_each_entry(info, &vm_dev->virtqueues, node) node 320 drivers/virtio/virtio_mmio.c list_del(&info->node); node 438 drivers/virtio/virtio_mmio.c list_add(&info->node, &vm_dev->virtqueues); node 67 drivers/virtio/virtio_pci_common.c list_for_each_entry(info, &vp_dev->virtqueues, node) { node 197 drivers/virtio/virtio_pci_common.c list_add(&info->node, &vp_dev->virtqueues); node 200 drivers/virtio/virtio_pci_common.c INIT_LIST_HEAD(&info->node); node 218 drivers/virtio/virtio_pci_common.c list_del(&info->node); node 36 drivers/virtio/virtio_pci_common.h struct list_head node; node 444 drivers/w1/w1_netlink.c struct w1_cb_node *node = container_of(async_cmd, struct w1_cb_node, node 446 drivers/w1/w1_netlink.c u16 mlen = node->msg->len; node 449 drivers/w1/w1_netlink.c struct w1_slave *sl = node->sl; node 450 drivers/w1/w1_netlink.c struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)node->msg->data; node 453 drivers/w1/w1_netlink.c dev->priv = node->block; node 456 drivers/w1/w1_netlink.c node->block->cur_msg = node->msg; node 468 drivers/w1/w1_netlink.c w1_netlink_check_send(node->block); node 470 drivers/w1/w1_netlink.c w1_netlink_queue_status(node->block, node->msg, cmd, err); node 479 drivers/w1/w1_netlink.c w1_netlink_queue_status(node->block, node->msg, cmd, err); node 495 drivers/w1/w1_netlink.c w1_unref_block(node->block); node 544 drivers/w1/w1_netlink.c struct w1_cb_node *node = NULL; node 615 drivers/w1/w1_netlink.c node = (struct w1_cb_node *)(block->request_cn.data + cn->len); node 627 drivers/w1/w1_netlink.c block->first_cn = (struct cn_msg *)(node + node_count); node 679 drivers/w1/w1_netlink.c node->async.cb = w1_process_cb; node 680 drivers/w1/w1_netlink.c node->block = block; node 681 drivers/w1/w1_netlink.c node->msg = (struct w1_netlink_msg *)((u8 *)&block->request_cn + node 683 drivers/w1/w1_netlink.c node->sl = sl; node 684 drivers/w1/w1_netlink.c node->dev = dev; node 687 drivers/w1/w1_netlink.c list_add_tail(&node->async.async_entry, &dev->async_list); node 690 drivers/w1/w1_netlink.c ++node; node 133 drivers/watchdog/octeon-wdt-main.c int node = cpu_to_node(cpu); node 138 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1); node 147 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1); node 279 drivers/watchdog/octeon-wdt-main.c unsigned int node = cvmx_get_node_num(); node 289 drivers/watchdog/octeon-wdt-main.c ciu_wdog.u64 = cvmx_read_csr_node(node, CVMX_CIU_WDOGX(lcore)); node 302 drivers/watchdog/octeon-wdt-main.c int node; node 306 drivers/watchdog/octeon-wdt-main.c node = cpu_to_node(cpu); node 312 drivers/watchdog/octeon-wdt-main.c domain = octeon_irq_get_block_domain(node, node 325 drivers/watchdog/octeon-wdt-main.c int node; node 330 drivers/watchdog/octeon-wdt-main.c node = cpu_to_node(cpu); node 333 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1); node 337 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64); node 348 drivers/watchdog/octeon-wdt-main.c int node; node 353 drivers/watchdog/octeon-wdt-main.c node = cpu_to_node(cpu); node 359 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64); node 365 drivers/watchdog/octeon-wdt-main.c domain = octeon_irq_get_block_domain(node, WD_BLOCK_NUMBER); node 391 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1); node 397 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64); node 406 drivers/watchdog/octeon-wdt-main.c int node; node 413 drivers/watchdog/octeon-wdt-main.c node = cpu_to_node(cpu); node 414 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1); node 458 drivers/watchdog/octeon-wdt-main.c int node; node 470 drivers/watchdog/octeon-wdt-main.c node = cpu_to_node(cpu); node 471 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1); node 475 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_WDOGX(coreid), ciu_wdog.u64); node 476 drivers/watchdog/octeon-wdt-main.c cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1); node 501 drivers/watchdog/orion_wdt.c struct device_node *node = pdev->dev.of_node; node 513 drivers/watchdog/orion_wdt.c if (of_device_is_compatible(node, "marvell,orion-wdt")) { node 520 drivers/watchdog/orion_wdt.c } else if (of_device_is_compatible(node, "marvell,armada-370-wdt") || node 521 drivers/watchdog/orion_wdt.c of_device_is_compatible(node, "marvell,armada-xp-wdt")) { node 528 drivers/watchdog/orion_wdt.c } else if (of_device_is_compatible(node, "marvell,armada-375-wdt") || node 529 drivers/watchdog/orion_wdt.c of_device_is_compatible(node, "marvell,armada-380-wdt")) { node 95 drivers/watchdog/sun4v_wdt.c u64 node; node 114 drivers/watchdog/sun4v_wdt.c node = mdesc_node_by_name(handle, MDESC_NODE_NULL, "platform"); node 116 drivers/watchdog/sun4v_wdt.c if (node == MDESC_NODE_NULL) node 127 drivers/watchdog/sun4v_wdt.c value = mdesc_get_property(handle, node, "watchdog-resolution", NULL); node 135 drivers/watchdog/sun4v_wdt.c value = mdesc_get_property(handle, node, "watchdog-max-timeout", NULL); node 27 drivers/watchdog/wdat_wdt.c struct list_head node; node 128 drivers/watchdog/wdat_wdt.c list_for_each_entry(instr, wdat->instructions[action], node) { node 392 drivers/watchdog/wdat_wdt.c INIT_LIST_HEAD(&instr->node); node 436 drivers/watchdog/wdat_wdt.c list_add_tail(&instr->node, instructions); node 90 drivers/xen/cpu_hotplug.c .node = "cpu", node 84 drivers/xen/evtchn.c struct rb_node node; node 116 drivers/xen/evtchn.c this = rb_entry(*new, struct user_evtchn, node); node 128 drivers/xen/evtchn.c rb_link_node(&evtchn->node, parent, new); node 129 drivers/xen/evtchn.c rb_insert_color(&evtchn->node, &u->evtchns); node 137 drivers/xen/evtchn.c rb_erase(&evtchn->node, &u->evtchns); node 143 drivers/xen/evtchn.c struct rb_node *node = u->evtchns.rb_node; node 145 drivers/xen/evtchn.c while (node) { node 148 drivers/xen/evtchn.c evtchn = rb_entry(node, struct user_evtchn, node); node 151 drivers/xen/evtchn.c node = node->rb_left; node 153 drivers/xen/evtchn.c node = node->rb_right; node 673 drivers/xen/evtchn.c struct rb_node *node; node 675 drivers/xen/evtchn.c while ((node = u->evtchns.rb_node)) { node 678 drivers/xen/evtchn.c evtchn = rb_entry(node, struct user_evtchn, node); node 314 drivers/xen/manage.c .node = "control/sysrq", node 320 drivers/xen/manage.c .node = "control/shutdown", node 333 drivers/xen/manage.c char node[FEATURE_PATH_SIZE]; node 353 drivers/xen/manage.c snprintf(node, FEATURE_PATH_SIZE, "feature-%s", node 355 drivers/xen/manage.c err = xenbus_printf(XBT_NIL, "control", node, "%u", 1); node 358 drivers/xen/manage.c err, node); node 106 drivers/xen/xen-balloon.c .node = "memory/target", node 117 drivers/xen/xenbus/xenbus.h void xenbus_dev_changed(const char *node, struct xen_bus_type *bus); node 122 drivers/xen/xenbus/xenbus_client.c watch->node = path; node 128 drivers/xen/xenbus/xenbus_client.c watch->node = NULL; node 555 drivers/xen/xenbus/xenbus_client.c struct xenbus_map_node *node; node 569 drivers/xen/xenbus/xenbus_client.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 570 drivers/xen/xenbus/xenbus_client.c if (!node) node 573 drivers/xen/xenbus/xenbus_client.c err = alloc_xenballooned_pages(nr_pages, node->hvm.pages); node 577 drivers/xen/xenbus/xenbus_client.c gnttab_foreach_grant(node->hvm.pages, nr_grefs, node 581 drivers/xen/xenbus/xenbus_client.c err = __xenbus_map_ring(dev, gnt_ref, nr_grefs, node->handles, node 583 drivers/xen/xenbus/xenbus_client.c node->nr_handles = nr_grefs; node 588 drivers/xen/xenbus/xenbus_client.c addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP, node 595 drivers/xen/xenbus/xenbus_client.c node->hvm.addr = addr; node 598 drivers/xen/xenbus/xenbus_client.c list_add(&node->next, &xenbus_valloc_pages); node 606 drivers/xen/xenbus/xenbus_client.c xenbus_unmap_ring(dev, node->handles, nr_grefs, info.addrs); node 612 drivers/xen/xenbus/xenbus_client.c free_xenballooned_pages(nr_pages, node->hvm.pages); node 614 drivers/xen/xenbus/xenbus_client.c kfree(node); node 681 drivers/xen/xenbus/xenbus_client.c struct xenbus_map_node *node; node 694 drivers/xen/xenbus/xenbus_client.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 695 drivers/xen/xenbus/xenbus_client.c if (!node) node 700 drivers/xen/xenbus/xenbus_client.c kfree(node); node 707 drivers/xen/xenbus/xenbus_client.c err = __xenbus_map_ring(dev, gnt_refs, nr_grefs, node->handles, node 714 drivers/xen/xenbus/xenbus_client.c node->nr_handles = nr_grefs; node 715 drivers/xen/xenbus/xenbus_client.c node->pv.area = area; node 718 drivers/xen/xenbus/xenbus_client.c list_add(&node->next, &xenbus_valloc_pages); node 730 drivers/xen/xenbus/xenbus_client.c kfree(node); node 736 drivers/xen/xenbus/xenbus_client.c struct xenbus_map_node *node; node 744 drivers/xen/xenbus/xenbus_client.c list_for_each_entry(node, &xenbus_valloc_pages, next) { node 745 drivers/xen/xenbus/xenbus_client.c if (node->pv.area->addr == vaddr) { node 746 drivers/xen/xenbus/xenbus_client.c list_del(&node->next); node 750 drivers/xen/xenbus/xenbus_client.c node = NULL; node 754 drivers/xen/xenbus/xenbus_client.c if (!node) { node 760 drivers/xen/xenbus/xenbus_client.c for (i = 0; i < node->nr_handles; i++) { node 768 drivers/xen/xenbus/xenbus_client.c unmap[i].handle = node->handles[i]; node 776 drivers/xen/xenbus/xenbus_client.c for (i = 0; i < node->nr_handles; i++) { node 781 drivers/xen/xenbus/xenbus_client.c node->handles[i], unmap[i].status); node 788 drivers/xen/xenbus/xenbus_client.c free_vm_area(node->pv.area); node 791 drivers/xen/xenbus/xenbus_client.c node->pv.area, node->nr_handles); node 793 drivers/xen/xenbus/xenbus_client.c kfree(node); node 824 drivers/xen/xenbus/xenbus_client.c struct xenbus_map_node *node; node 832 drivers/xen/xenbus/xenbus_client.c list_for_each_entry(node, &xenbus_valloc_pages, next) { node 833 drivers/xen/xenbus/xenbus_client.c addr = node->hvm.addr; node 835 drivers/xen/xenbus/xenbus_client.c list_del(&node->next); node 839 drivers/xen/xenbus/xenbus_client.c node = addr = NULL; node 843 drivers/xen/xenbus/xenbus_client.c if (!node) { node 849 drivers/xen/xenbus/xenbus_client.c nr_pages = XENBUS_PAGES(node->nr_handles); node 851 drivers/xen/xenbus/xenbus_client.c gnttab_foreach_grant(node->hvm.pages, node->nr_handles, node 855 drivers/xen/xenbus/xenbus_client.c rv = xenbus_unmap_ring(dev, node->handles, node->nr_handles, node 859 drivers/xen/xenbus/xenbus_client.c free_xenballooned_pages(nr_pages, node->hvm.pages); node 864 drivers/xen/xenbus/xenbus_client.c kfree(node); node 235 drivers/xen/xenbus/xenbus_dev_frontend.c kfree(watch->watch.node); node 249 drivers/xen/xenbus/xenbus_dev_frontend.c watch->watch.node = kstrdup(path, GFP_KERNEL); node 250 drivers/xen/xenbus/xenbus_dev_frontend.c if (watch->watch.node == NULL) node 534 drivers/xen/xenbus/xenbus_dev_frontend.c !strcmp(watch->watch.node, path)) { node 113 drivers/xen/xenbus/xenbus_probe.c if (dev->otherend_watch.node) { node 115 drivers/xen/xenbus/xenbus_probe.c kfree(dev->otherend_watch.node); node 116 drivers/xen/xenbus/xenbus_probe.c dev->otherend_watch.node = NULL; node 191 drivers/xen/xenbus/xenbus_probe.c state, xenbus_strstate(state), dev->otherend_watch.node, path); node 556 drivers/xen/xenbus/xenbus_probe.c void xenbus_dev_changed(const char *node, struct xen_bus_type *bus) node 563 drivers/xen/xenbus/xenbus_probe.c if (char_count(node, '/') < 2) node 566 drivers/xen/xenbus/xenbus_probe.c exists = xenbus_exists(XBT_NIL, node, ""); node 568 drivers/xen/xenbus/xenbus_probe.c xenbus_cleanup_devices(node, &bus->bus); node 573 drivers/xen/xenbus/xenbus_probe.c p = strchr(node, '/') + 1; node 577 drivers/xen/xenbus/xenbus_probe.c rootlen = strsep_len(node, '/', bus->levels); node 580 drivers/xen/xenbus/xenbus_probe.c root = kasprintf(GFP_KERNEL, "%.*s", rootlen, node); node 215 drivers/xen/xenbus/xenbus_probe_backend.c .node = "backend", node 167 drivers/xen/xenbus/xenbus_probe_frontend.c .node = "device", node 368 drivers/xen/xenbus/xenbus_probe_frontend.c be_watch.node = kasprintf(GFP_NOIO | __GFP_HIGH, "%s/state", be); node 369 drivers/xen/xenbus/xenbus_probe_frontend.c if (!be_watch.node) node 397 drivers/xen/xenbus/xenbus_probe_frontend.c kfree(be_watch.node); node 423 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node, unsigned int *num) node 428 drivers/xen/xenbus/xenbus_xs.c path = join(dir, node); node 443 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node) node 448 drivers/xen/xenbus/xenbus_xs.c d = xenbus_directory(t, dir, node, &dir_n); node 461 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node, unsigned int *len) node 466 drivers/xen/xenbus/xenbus_xs.c path = join(dir, node); node 480 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node, const char *string) node 486 drivers/xen/xenbus/xenbus_xs.c path = join(dir, node); node 503 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node) node 508 drivers/xen/xenbus/xenbus_xs.c path = join(dir, node); node 519 drivers/xen/xenbus/xenbus_xs.c int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node) node 524 drivers/xen/xenbus/xenbus_xs.c path = join(dir, node); node 569 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node, const char *fmt, ...) node 575 drivers/xen/xenbus/xenbus_xs.c val = xenbus_read(t, dir, node, NULL); node 591 drivers/xen/xenbus/xenbus_xs.c unsigned int xenbus_read_unsigned(const char *dir, const char *node, node 597 drivers/xen/xenbus/xenbus_xs.c ret = xenbus_scanf(XBT_NIL, dir, node, "%u", &val); node 607 drivers/xen/xenbus/xenbus_xs.c const char *dir, const char *node, const char *fmt, ...) node 620 drivers/xen/xenbus/xenbus_xs.c ret = xenbus_write(t, dir, node, buf); node 775 drivers/xen/xenbus/xenbus_xs.c err = xs_watch(watch->node, token); node 804 drivers/xen/xenbus/xenbus_xs.c err = xs_unwatch(watch->node, token); node 806 drivers/xen/xenbus/xenbus_xs.c pr_warn("Failed to release watch %s: %i\n", watch->node, err); node 852 drivers/xen/xenbus/xenbus_xs.c xs_watch(watch->node, token); node 88 fs/afs/afs.h __s8 node[6]; /* spatially unique node ID (MAC addr) */ node 203 fs/afs/afs.h __be32 node[6]; node 463 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]); node 574 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]); node 619 fs/afs/cmservice.c reply.ia.uuid[loop + 5] = htonl((s8) call->net->uuid.node[loop]); node 83 fs/afs/vlclient.c uuid->node[j] = (u8)ntohl(xdr->node[j]); node 303 fs/afs/vlclient.c r->uuid.node[i] = htonl(u->node[i]); node 100 fs/befs/btree.c struct befs_btree_node *node, node 103 fs/befs/btree.c static int befs_leafnode(struct befs_btree_node *node); node 105 fs/befs/btree.c static fs16 *befs_bt_keylen_index(struct befs_btree_node *node); node 107 fs/befs/btree.c static fs64 *befs_bt_valarray(struct befs_btree_node *node); node 109 fs/befs/btree.c static char *befs_bt_keydata(struct befs_btree_node *node); node 112 fs/befs/btree.c struct befs_btree_node *node, node 116 fs/befs/btree.c struct befs_btree_node *node, node 192 fs/befs/btree.c struct befs_btree_node *node, befs_off_t node_off) node 198 fs/befs/btree.c if (node->bh) node 199 fs/befs/btree.c brelse(node->bh); node 201 fs/befs/btree.c node->bh = befs_read_datastream(sb, ds, node_off, &off); node 202 fs/befs/btree.c if (!node->bh) { node 209 fs/befs/btree.c node->od_node = node 210 fs/befs/btree.c (befs_btree_nodehead *) ((void *) node->bh->b_data + off); node 212 fs/befs/btree.c befs_dump_index_node(sb, node->od_node); node 214 fs/befs/btree.c node->head.left = fs64_to_cpu(sb, node->od_node->left); node 215 fs/befs/btree.c node->head.right = fs64_to_cpu(sb, node->od_node->right); node 216 fs/befs/btree.c node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow); node 217 fs/befs/btree.c node->head.all_key_count = node 218 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_count); node 219 fs/befs/btree.c node->head.all_key_length = node 220 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_length); node 330 fs/befs/btree.c befs_find_key(struct super_block *sb, struct befs_btree_node *node, node 345 fs/befs/btree.c last = node->head.all_key_count - 1; node 346 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, last, &keylen); node 354 fs/befs/btree.c valarray = befs_bt_valarray(node); node 363 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, mid, &keylen); node 599 fs/befs/btree.c befs_leafnode(struct befs_btree_node *node) node 602 fs/befs/btree.c if (node->head.overflow == BEFS_BT_INVAL) node 622 fs/befs/btree.c befs_bt_keylen_index(struct befs_btree_node *node) node 626 fs/befs/btree.c (sizeof (befs_btree_nodehead) + node->head.all_key_length); node 632 fs/befs/btree.c return (fs16 *) ((void *) node->od_node + off); node 643 fs/befs/btree.c befs_bt_valarray(struct befs_btree_node *node) node 645 fs/befs/btree.c void *keylen_index_start = (void *) befs_bt_keylen_index(node); node 646 fs/befs/btree.c size_t keylen_index_size = node->head.all_key_count * sizeof (fs16); node 659 fs/befs/btree.c befs_bt_keydata(struct befs_btree_node *node) node 661 fs/befs/btree.c return (char *) ((void *) node->od_node + sizeof (befs_btree_nodehead)); node 675 fs/befs/btree.c befs_bt_get_key(struct super_block *sb, struct befs_btree_node *node, node 682 fs/befs/btree.c if (index < 0 || index > node->head.all_key_count) { node 687 fs/befs/btree.c keystart = befs_bt_keydata(node); node 688 fs/befs/btree.c keylen_index = befs_bt_keylen_index(node); node 248 fs/befs/debug.c befs_dump_index_node(const struct super_block *sb, befs_btree_nodehead *node) node 253 fs/befs/debug.c befs_debug(sb, " left %016LX", fs64_to_cpu(sb, node->left)); node 254 fs/befs/debug.c befs_debug(sb, " right %016LX", fs64_to_cpu(sb, node->right)); node 255 fs/befs/debug.c befs_debug(sb, " overflow %016LX", fs64_to_cpu(sb, node->overflow)); node 257 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_count)); node 259 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_length)); node 539 fs/btrfs/backref.c root_level = btrfs_header_level(root->node); node 586 fs/btrfs/backref.c unode_aux_to_inode_list(struct ulist_node *node) node 588 fs/btrfs/backref.c if (!node) node 590 fs/btrfs/backref.c return (struct extent_inode_elem *)(uintptr_t)node->aux; node 618 fs/btrfs/backref.c struct ulist_node *node; node 675 fs/btrfs/backref.c node = ulist_next(parents, &uiter); node 676 fs/btrfs/backref.c ref->parent = node ? node->val : 0; node 677 fs/btrfs/backref.c ref->inode_list = unode_aux_to_inode_list(node); node 680 fs/btrfs/backref.c while ((node = ulist_next(parents, &uiter))) { node 691 fs/btrfs/backref.c new_ref->parent = node->val; node 692 fs/btrfs/backref.c new_ref->inode_list = unode_aux_to_inode_list(node); node 720 fs/btrfs/backref.c struct rb_node *node; node 722 fs/btrfs/backref.c while ((node = rb_first_cached(&tree->root))) { node 723 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); node 724 fs/btrfs/backref.c rb_erase_cached(node, &tree->root); node 764 fs/btrfs/backref.c struct btrfs_delayed_ref_node *node; node 777 fs/btrfs/backref.c node = rb_entry(n, struct btrfs_delayed_ref_node, node 779 fs/btrfs/backref.c if (node->seq > seq) node 782 fs/btrfs/backref.c switch (node->action) { node 788 fs/btrfs/backref.c count = node->ref_mod; node 791 fs/btrfs/backref.c count = node->ref_mod * -1; node 797 fs/btrfs/backref.c switch (node->type) { node 802 fs/btrfs/backref.c ref = btrfs_delayed_node_to_tree_ref(node); node 805 fs/btrfs/backref.c node->bytenr, count, sc, node 813 fs/btrfs/backref.c ref = btrfs_delayed_node_to_tree_ref(node); node 816 fs/btrfs/backref.c ref->parent, node->bytenr, count, node 823 fs/btrfs/backref.c ref = btrfs_delayed_node_to_data_ref(node); node 839 fs/btrfs/backref.c &key, 0, node->bytenr, count, sc, node 847 fs/btrfs/backref.c ref = btrfs_delayed_node_to_data_ref(node); node 850 fs/btrfs/backref.c node->bytenr, count, sc, node 1126 fs/btrfs/backref.c struct rb_node *node; node 1252 fs/btrfs/backref.c node = rb_first_cached(&preftrees.direct.root); node 1253 fs/btrfs/backref.c while (node) { node 1254 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); node 1255 fs/btrfs/backref.c node = rb_next(&ref->rbnode); node 1341 fs/btrfs/backref.c struct ulist_node *node = NULL; node 1346 fs/btrfs/backref.c while ((node = ulist_next(blocks, &uiter))) { node 1347 fs/btrfs/backref.c if (!node->aux) node 1349 fs/btrfs/backref.c eie = unode_aux_to_inode_list(node); node 1351 fs/btrfs/backref.c node->aux = 0; node 1405 fs/btrfs/backref.c struct ulist_node *node = NULL; node 1427 fs/btrfs/backref.c node = ulist_next(tmp, &uiter); node 1428 fs/btrfs/backref.c if (!node) node 1430 fs/btrfs/backref.c bytenr = node->val; node 1474 fs/btrfs/backref.c struct ulist_node *node; node 1510 fs/btrfs/backref.c node = ulist_next(tmp, &uiter); node 1511 fs/btrfs/backref.c if (!node) node 1513 fs/btrfs/backref.c bytenr = node->val; node 254 fs/btrfs/block-group.c struct rb_node *node; node 266 fs/btrfs/block-group.c node = rb_next(&cache->cache_node); node 268 fs/btrfs/block-group.c if (node) { node 269 fs/btrfs/block-group.c cache = rb_entry(node, struct btrfs_block_group_cache, node 132 fs/btrfs/ctree.c eb = rcu_dereference(root->node); node 161 fs/btrfs/ctree.c if (eb == root->node) node 180 fs/btrfs/ctree.c if (eb == root->node) node 288 fs/btrfs/ctree.c struct rb_node node; node 346 fs/btrfs/ctree.c struct rb_node *node; node 379 fs/btrfs/ctree.c for (node = rb_first(tm_root); node; node = next) { node 380 fs/btrfs/ctree.c next = rb_next(node); node 381 fs/btrfs/ctree.c tm = rb_entry(node, struct tree_mod_elem, node); node 384 fs/btrfs/ctree.c rb_erase(node, tm_root); node 413 fs/btrfs/ctree.c cur = rb_entry(*new, struct tree_mod_elem, node); node 427 fs/btrfs/ctree.c rb_link_node(&tm->node, parent, new); node 428 fs/btrfs/ctree.c rb_insert_color(&tm->node, tm_root); node 486 fs/btrfs/ctree.c RB_CLEAR_NODE(&tm->node); node 578 fs/btrfs/ctree.c if (tm_list[i] && !RB_EMPTY_NODE(&tm_list[i]->node)) node 579 fs/btrfs/ctree.c rb_erase(&tm_list[i]->node, &eb->fs_info->tree_mod_log); node 602 fs/btrfs/ctree.c rb_erase(&tm_list[j]->node, node 685 fs/btrfs/ctree.c struct rb_node *node; node 691 fs/btrfs/ctree.c node = tm_root->rb_node; node 692 fs/btrfs/ctree.c while (node) { node 693 fs/btrfs/ctree.c cur = rb_entry(node, struct tree_mod_elem, node); node 695 fs/btrfs/ctree.c node = node->rb_left; node 697 fs/btrfs/ctree.c node = node->rb_right; node 699 fs/btrfs/ctree.c node = node->rb_left; node 705 fs/btrfs/ctree.c node = node->rb_left; node 711 fs/btrfs/ctree.c node = node->rb_right; node 805 fs/btrfs/ctree.c if (tm_list[i] && !RB_EMPTY_NODE(&tm_list[i]->node)) node 806 fs/btrfs/ctree.c rb_erase(&tm_list[i]->node, &fs_info->tree_mod_log); node 875 fs/btrfs/ctree.c buf != root->node && buf != root->commit_root && node 1118 fs/btrfs/ctree.c if (buf == root->node) { node 1125 fs/btrfs/ctree.c ret = tree_mod_log_insert_root(root->node, cow, 1); node 1127 fs/btrfs/ctree.c rcu_assign_pointer(root->node, cow); node 1279 fs/btrfs/ctree.c next = rb_next(&tm->node); node 1282 fs/btrfs/ctree.c tm = rb_entry(next, struct tree_mod_elem, node); node 1899 fs/btrfs/ctree.c ret = tree_mod_log_insert_root(root->node, child, 1); node 1901 fs/btrfs/ctree.c rcu_assign_pointer(root->node, child); node 2215 fs/btrfs/ctree.c struct extent_buffer *node; node 2232 fs/btrfs/ctree.c node = path->nodes[level]; node 2234 fs/btrfs/ctree.c search = btrfs_node_blockptr(node, slot); node 2244 fs/btrfs/ctree.c nritems = btrfs_header_nritems(node); node 2258 fs/btrfs/ctree.c btrfs_node_key(node, &disk_key, nr); node 2262 fs/btrfs/ctree.c search = btrfs_node_blockptr(node, nr); node 3398 fs/btrfs/ctree.c BUG_ON(path->nodes[level-1] != root->node); node 3407 fs/btrfs/ctree.c root->node->start, 0); node 3423 fs/btrfs/ctree.c old = root->node; node 3424 fs/btrfs/ctree.c ret = tree_mod_log_insert_root(root->node, c, 0); node 3426 fs/btrfs/ctree.c rcu_assign_pointer(root->node, c); node 3508 fs/btrfs/ctree.c if (c == root->node) { node 4926 fs/btrfs/ctree.c if (nritems == 0 && parent == root->node) { node 4927 fs/btrfs/ctree.c BUG_ON(btrfs_header_level(root->node) != 1); node 4929 fs/btrfs/ctree.c btrfs_set_header_level(root->node, 0); node 5023 fs/btrfs/ctree.c if (leaf == root->node) { node 475 fs/btrfs/ctree.h struct rb_node node; node 961 fs/btrfs/ctree.h struct extent_buffer *node; node 2622 fs/btrfs/ctree.h struct extent_buffer *node, node 2692 fs/btrfs/ctree.h struct extent_buffer *node); node 71 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node; node 73 fs/btrfs/delayed-inode.c node = READ_ONCE(btrfs_inode->delayed_node); node 74 fs/btrfs/delayed-inode.c if (node) { node 75 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); node 76 fs/btrfs/delayed-inode.c return node; node 80 fs/btrfs/delayed-inode.c node = radix_tree_lookup(&root->delayed_nodes_tree, ino); node 82 fs/btrfs/delayed-inode.c if (node) { node 84 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); /* can be accessed */ node 85 fs/btrfs/delayed-inode.c BUG_ON(btrfs_inode->delayed_node != node); node 87 fs/btrfs/delayed-inode.c return node; node 106 fs/btrfs/delayed-inode.c if (refcount_inc_not_zero(&node->refs)) { node 107 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); node 108 fs/btrfs/delayed-inode.c btrfs_inode->delayed_node = node; node 110 fs/btrfs/delayed-inode.c node = NULL; node 114 fs/btrfs/delayed-inode.c return node; node 125 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node; node 131 fs/btrfs/delayed-inode.c node = btrfs_get_delayed_node(btrfs_inode); node 132 fs/btrfs/delayed-inode.c if (node) node 133 fs/btrfs/delayed-inode.c return node; node 135 fs/btrfs/delayed-inode.c node = kmem_cache_zalloc(delayed_node_cache, GFP_NOFS); node 136 fs/btrfs/delayed-inode.c if (!node) node 138 fs/btrfs/delayed-inode.c btrfs_init_delayed_node(node, root, ino); node 141 fs/btrfs/delayed-inode.c refcount_set(&node->refs, 2); node 145 fs/btrfs/delayed-inode.c kmem_cache_free(delayed_node_cache, node); node 150 fs/btrfs/delayed-inode.c ret = radix_tree_insert(&root->delayed_nodes_tree, ino, node); node 153 fs/btrfs/delayed-inode.c kmem_cache_free(delayed_node_cache, node); node 157 fs/btrfs/delayed-inode.c btrfs_inode->delayed_node = node; node 161 fs/btrfs/delayed-inode.c return node; node 170 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node, node 174 fs/btrfs/delayed-inode.c if (test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) { node 175 fs/btrfs/delayed-inode.c if (!list_empty(&node->p_list)) node 176 fs/btrfs/delayed-inode.c list_move_tail(&node->p_list, &root->prepare_list); node 178 fs/btrfs/delayed-inode.c list_add_tail(&node->p_list, &root->prepare_list); node 180 fs/btrfs/delayed-inode.c list_add_tail(&node->n_list, &root->node_list); node 181 fs/btrfs/delayed-inode.c list_add_tail(&node->p_list, &root->prepare_list); node 182 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); /* inserted into list */ node 184 fs/btrfs/delayed-inode.c set_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags); node 191 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 194 fs/btrfs/delayed-inode.c if (test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) { node 196 fs/btrfs/delayed-inode.c refcount_dec(&node->refs); /* not in the list */ node 197 fs/btrfs/delayed-inode.c list_del_init(&node->n_list); node 198 fs/btrfs/delayed-inode.c if (!list_empty(&node->p_list)) node 199 fs/btrfs/delayed-inode.c list_del_init(&node->p_list); node 200 fs/btrfs/delayed-inode.c clear_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags); node 209 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node = NULL; node 216 fs/btrfs/delayed-inode.c node = list_entry(p, struct btrfs_delayed_node, n_list); node 217 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); node 221 fs/btrfs/delayed-inode.c return node; node 225 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 231 fs/btrfs/delayed-inode.c delayed_root = node->root->fs_info->delayed_root; node 233 fs/btrfs/delayed-inode.c if (!test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) { node 238 fs/btrfs/delayed-inode.c } else if (list_is_last(&node->n_list, &delayed_root->node_list)) node 241 fs/btrfs/delayed-inode.c p = node->n_list.next; node 285 fs/btrfs/delayed-inode.c static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node) node 287 fs/btrfs/delayed-inode.c __btrfs_release_delayed_node(node, 0); node 294 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node = NULL; node 302 fs/btrfs/delayed-inode.c node = list_entry(p, struct btrfs_delayed_node, p_list); node 303 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); node 307 fs/btrfs/delayed-inode.c return node; node 311 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 313 fs/btrfs/delayed-inode.c __btrfs_release_delayed_node(node, 1); node 346 fs/btrfs/delayed-inode.c struct rb_node *node, *prev_node = NULL; node 350 fs/btrfs/delayed-inode.c node = root->rb_node; node 352 fs/btrfs/delayed-inode.c while (node) { node 353 fs/btrfs/delayed-inode.c delayed_item = rb_entry(node, struct btrfs_delayed_item, node 355 fs/btrfs/delayed-inode.c prev_node = node; node 358 fs/btrfs/delayed-inode.c node = node->rb_right; node 360 fs/btrfs/delayed-inode.c node = node->rb_left; node 370 fs/btrfs/delayed-inode.c else if ((node = rb_prev(prev_node)) != NULL) { node 371 fs/btrfs/delayed-inode.c *prev = rb_entry(node, struct btrfs_delayed_item, node 382 fs/btrfs/delayed-inode.c else if ((node = rb_next(prev_node)) != NULL) { node 383 fs/btrfs/delayed-inode.c *next = rb_entry(node, struct btrfs_delayed_item, node 403 fs/btrfs/delayed-inode.c struct rb_node **p, *node; node 417 fs/btrfs/delayed-inode.c node = &ins->rb_node; node 435 fs/btrfs/delayed-inode.c rb_link_node(node, parent_node, p); node 436 fs/btrfs/delayed-inode.c rb_insert_color_cached(node, root, leftmost); node 450 fs/btrfs/delayed-inode.c static int __btrfs_add_delayed_insertion_item(struct btrfs_delayed_node *node, node 453 fs/btrfs/delayed-inode.c return __btrfs_add_delayed_item(node, item, node 457 fs/btrfs/delayed-inode.c static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node, node 460 fs/btrfs/delayed-inode.c return __btrfs_add_delayed_item(node, item, node 606 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 647 fs/btrfs/delayed-inode.c node->bytes_reserved = num_bytes; node 662 fs/btrfs/delayed-inode.c node->bytes_reserved = num_bytes; node 669 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node, node 674 fs/btrfs/delayed-inode.c if (!node->bytes_reserved) node 679 fs/btrfs/delayed-inode.c node->inode_id, node->bytes_reserved, 0); node 681 fs/btrfs/delayed-inode.c node->bytes_reserved); node 683 fs/btrfs/delayed-inode.c btrfs_qgroup_free_meta_prealloc(node->root, node 684 fs/btrfs/delayed-inode.c node->bytes_reserved); node 686 fs/btrfs/delayed-inode.c btrfs_qgroup_convert_reserved_meta(node->root, node 687 fs/btrfs/delayed-inode.c node->bytes_reserved); node 688 fs/btrfs/delayed-inode.c node->bytes_reserved = 0; node 838 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 844 fs/btrfs/delayed-inode.c mutex_lock(&node->mutex); node 845 fs/btrfs/delayed-inode.c curr = __btrfs_first_delayed_insertion_item(node); node 866 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 870 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 940 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 947 fs/btrfs/delayed-inode.c mutex_lock(&node->mutex); node 948 fs/btrfs/delayed-inode.c curr = __btrfs_first_delayed_deletion_item(node); node 968 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 976 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 981 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1015 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 1025 fs/btrfs/delayed-inode.c key.objectid = node->inode_id; node 1029 fs/btrfs/delayed-inode.c if (test_bit(BTRFS_DELAYED_NODE_DEL_IREF, &node->flags)) node 1047 fs/btrfs/delayed-inode.c write_extent_buffer(leaf, &node->inode_item, (unsigned long)inode_item, node 1051 fs/btrfs/delayed-inode.c if (!test_bit(BTRFS_DELAYED_NODE_DEL_IREF, &node->flags)) node 1059 fs/btrfs/delayed-inode.c if (key.objectid != node->inode_id) node 1073 fs/btrfs/delayed-inode.c btrfs_release_delayed_iref(node); node 1077 fs/btrfs/delayed-inode.c btrfs_delayed_inode_release_metadata(fs_info, node, (ret < 0)); node 1078 fs/btrfs/delayed-inode.c btrfs_release_delayed_inode(node); node 1104 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 1108 fs/btrfs/delayed-inode.c mutex_lock(&node->mutex); node 1109 fs/btrfs/delayed-inode.c if (!test_bit(BTRFS_DELAYED_NODE_INODE_DIRTY, &node->flags)) { node 1110 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1114 fs/btrfs/delayed-inode.c ret = __btrfs_update_delayed_inode(trans, root, path, node); node 1115 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1122 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node) node 1126 fs/btrfs/delayed-inode.c ret = btrfs_insert_delayed_items(trans, path, node->root, node); node 1130 fs/btrfs/delayed-inode.c ret = btrfs_delete_delayed_items(trans, path, node->root, node); node 1134 fs/btrfs/delayed-inode.c ret = btrfs_update_delayed_inode(trans, node->root, path, node); node 1493 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node, node 1498 fs/btrfs/delayed-inode.c mutex_lock(&node->mutex); node 1499 fs/btrfs/delayed-inode.c item = __btrfs_lookup_delayed_insertion_item(node, key); node 1501 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1505 fs/btrfs/delayed-inode.c btrfs_delayed_item_release_metadata(node->root, item); node 1507 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1514 fs/btrfs/delayed-inode.c struct btrfs_delayed_node *node; node 1519 fs/btrfs/delayed-inode.c node = btrfs_get_or_create_delayed_node(dir); node 1520 fs/btrfs/delayed-inode.c if (IS_ERR(node)) node 1521 fs/btrfs/delayed-inode.c return PTR_ERR(node); node 1527 fs/btrfs/delayed-inode.c ret = btrfs_delete_delayed_insertion_item(trans->fs_info, node, node 1552 fs/btrfs/delayed-inode.c mutex_lock(&node->mutex); node 1553 fs/btrfs/delayed-inode.c ret = __btrfs_add_delayed_deletion_item(node, item); node 1557 fs/btrfs/delayed-inode.c index, node->root->root_key.objectid, node 1558 fs/btrfs/delayed-inode.c node->inode_id, ret); node 1562 fs/btrfs/delayed-inode.c mutex_unlock(&node->mutex); node 1564 fs/btrfs/delayed-inode.c btrfs_release_delayed_node(node); node 207 fs/btrfs/delayed-ref.c if (ref1->node.type == BTRFS_TREE_BLOCK_REF_KEY) { node 227 fs/btrfs/delayed-ref.c if (ref1->node.type == BTRFS_EXTENT_DATA_REF_KEY) { node 279 fs/btrfs/delayed-ref.c struct rb_node *node) node 288 fs/btrfs/delayed-ref.c ins = rb_entry(node, struct btrfs_delayed_ref_head, href_node); node 305 fs/btrfs/delayed-ref.c rb_link_node(node, parent_node, p); node 306 fs/btrfs/delayed-ref.c rb_insert_color_cached(node, root, leftmost); node 314 fs/btrfs/delayed-ref.c struct rb_node *node = &ins->ref_node; node 336 fs/btrfs/delayed-ref.c rb_link_node(node, parent_node, p); node 337 fs/btrfs/delayed-ref.c rb_insert_color_cached(node, root, leftmost); node 437 fs/btrfs/delayed-ref.c struct rb_node *node = rb_next(&ref->ref_node); node 440 fs/btrfs/delayed-ref.c while (!done && node) { node 443 fs/btrfs/delayed-ref.c next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); node 444 fs/btrfs/delayed-ref.c node = rb_next(node); node 483 fs/btrfs/delayed-ref.c struct rb_node *node; node 506 fs/btrfs/delayed-ref.c for (node = rb_first_cached(&head->ref_tree); node; node 507 fs/btrfs/delayed-ref.c node = rb_next(node)) { node 508 fs/btrfs/delayed-ref.c ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); node 554 fs/btrfs/delayed-ref.c struct rb_node *node; node 556 fs/btrfs/delayed-ref.c node = rb_next(&head->href_node); node 557 fs/btrfs/delayed-ref.c if (!node) { node 563 fs/btrfs/delayed-ref.c head = rb_entry(node, struct btrfs_delayed_ref_head, node 962 fs/btrfs/delayed-ref.c init_delayed_ref_common(fs_info, &ref->node, bytenr, num_bytes, node 984 fs/btrfs/delayed-ref.c ret = insert_delayed_ref(trans, delayed_refs, head_ref, &ref->node); node 993 fs/btrfs/delayed-ref.c trace_add_delayed_tree_ref(fs_info, &ref->node, ref, node 1038 fs/btrfs/delayed-ref.c init_delayed_ref_common(fs_info, &ref->node, bytenr, num_bytes, node 1080 fs/btrfs/delayed-ref.c ret = insert_delayed_ref(trans, delayed_refs, head_ref, &ref->node); node 1089 fs/btrfs/delayed-ref.c trace_add_delayed_data_ref(trans->fs_info, &ref->node, ref, node 124 fs/btrfs/delayed-ref.h struct btrfs_delayed_ref_node node; node 131 fs/btrfs/delayed-ref.h struct btrfs_delayed_ref_node node; node 381 fs/btrfs/delayed-ref.h btrfs_delayed_node_to_tree_ref(struct btrfs_delayed_ref_node *node) node 383 fs/btrfs/delayed-ref.h return container_of(node, struct btrfs_delayed_tree_ref, node); node 387 fs/btrfs/delayed-ref.h btrfs_delayed_node_to_data_ref(struct btrfs_delayed_ref_node *node) node 389 fs/btrfs/delayed-ref.h return container_of(node, struct btrfs_delayed_data_ref, node); node 1125 fs/btrfs/disk-io.c root->node = NULL; node 1256 fs/btrfs/disk-io.c root->node = leaf; node 1330 fs/btrfs/disk-io.c root->node = leaf; node 1332 fs/btrfs/disk-io.c btrfs_mark_buffer_dirty(root->node); node 1333 fs/btrfs/disk-io.c btrfs_tree_unlock(root->node); node 1372 fs/btrfs/disk-io.c btrfs_set_root_node(&log_root->root_item, log_root->node); node 1414 fs/btrfs/disk-io.c root->node = read_tree_block(fs_info, node 1417 fs/btrfs/disk-io.c if (IS_ERR(root->node)) { node 1418 fs/btrfs/disk-io.c ret = PTR_ERR(root->node); node 1420 fs/btrfs/disk-io.c } else if (!btrfs_buffer_uptodate(root->node, generation, 0)) { node 1422 fs/btrfs/disk-io.c free_extent_buffer(root->node); node 1856 fs/btrfs/disk-io.c btrfs_header_generation(info->tree_root->node)) node 1869 fs/btrfs/disk-io.c btrfs_set_backup_tree_root(root_backup, info->tree_root->node->start); node 1871 fs/btrfs/disk-io.c btrfs_header_generation(info->tree_root->node)); node 1874 fs/btrfs/disk-io.c btrfs_header_level(info->tree_root->node)); node 1876 fs/btrfs/disk-io.c btrfs_set_backup_chunk_root(root_backup, info->chunk_root->node->start); node 1878 fs/btrfs/disk-io.c btrfs_header_generation(info->chunk_root->node)); node 1880 fs/btrfs/disk-io.c btrfs_header_level(info->chunk_root->node)); node 1882 fs/btrfs/disk-io.c btrfs_set_backup_extent_root(root_backup, info->extent_root->node->start); node 1884 fs/btrfs/disk-io.c btrfs_header_generation(info->extent_root->node)); node 1886 fs/btrfs/disk-io.c btrfs_header_level(info->extent_root->node)); node 1892 fs/btrfs/disk-io.c if (info->fs_root && info->fs_root->node) { node 1894 fs/btrfs/disk-io.c info->fs_root->node->start); node 1896 fs/btrfs/disk-io.c btrfs_header_generation(info->fs_root->node)); node 1898 fs/btrfs/disk-io.c btrfs_header_level(info->fs_root->node)); node 1901 fs/btrfs/disk-io.c btrfs_set_backup_dev_root(root_backup, info->dev_root->node->start); node 1903 fs/btrfs/disk-io.c btrfs_header_generation(info->dev_root->node)); node 1905 fs/btrfs/disk-io.c btrfs_header_level(info->dev_root->node)); node 1907 fs/btrfs/disk-io.c btrfs_set_backup_csum_root(root_backup, info->csum_root->node->start); node 1909 fs/btrfs/disk-io.c btrfs_header_generation(info->csum_root->node)); node 1911 fs/btrfs/disk-io.c btrfs_header_level(info->csum_root->node)); node 2011 fs/btrfs/disk-io.c free_extent_buffer(root->node); node 2013 fs/btrfs/disk-io.c root->node = NULL; node 2047 fs/btrfs/disk-io.c free_extent_buffer(gang[0]->node); node 2264 fs/btrfs/disk-io.c log_tree_root->node = read_tree_block(fs_info, bytenr, node 2267 fs/btrfs/disk-io.c if (IS_ERR(log_tree_root->node)) { node 2269 fs/btrfs/disk-io.c ret = PTR_ERR(log_tree_root->node); node 2272 fs/btrfs/disk-io.c } else if (!extent_buffer_uptodate(log_tree_root->node)) { node 2274 fs/btrfs/disk-io.c free_extent_buffer(log_tree_root->node); node 2283 fs/btrfs/disk-io.c free_extent_buffer(log_tree_root->node); node 2987 fs/btrfs/disk-io.c chunk_root->node = read_tree_block(fs_info, node 2990 fs/btrfs/disk-io.c if (IS_ERR(chunk_root->node) || node 2991 fs/btrfs/disk-io.c !extent_buffer_uptodate(chunk_root->node)) { node 2993 fs/btrfs/disk-io.c if (!IS_ERR(chunk_root->node)) node 2994 fs/btrfs/disk-io.c free_extent_buffer(chunk_root->node); node 2995 fs/btrfs/disk-io.c chunk_root->node = NULL; node 2998 fs/btrfs/disk-io.c btrfs_set_root_node(&chunk_root->root_item, chunk_root->node); node 3001 fs/btrfs/disk-io.c read_extent_buffer(chunk_root->node, fs_info->chunk_tree_uuid, node 3002 fs/btrfs/disk-io.c btrfs_header_chunk_tree_uuid(chunk_root->node), BTRFS_UUID_SIZE); node 3025 fs/btrfs/disk-io.c tree_root->node = read_tree_block(fs_info, node 3028 fs/btrfs/disk-io.c if (IS_ERR(tree_root->node) || node 3029 fs/btrfs/disk-io.c !extent_buffer_uptodate(tree_root->node)) { node 3031 fs/btrfs/disk-io.c if (!IS_ERR(tree_root->node)) node 3032 fs/btrfs/disk-io.c free_extent_buffer(tree_root->node); node 3033 fs/btrfs/disk-io.c tree_root->node = NULL; node 3037 fs/btrfs/disk-io.c btrfs_set_root_node(&tree_root->root_item, tree_root->node); node 3876 fs/btrfs/disk-io.c free_extent_buffer(root->reloc_root->node); node 3898 fs/btrfs/disk-io.c free_extent_buffer(root->node); node 4268 fs/btrfs/disk-io.c struct rb_node *node; node 4282 fs/btrfs/disk-io.c while ((node = rb_first_cached(&delayed_refs->href_root)) != NULL) { node 4287 fs/btrfs/disk-io.c head = rb_entry(node, struct btrfs_delayed_ref_head, node 40 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, u64 parent, node 52 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 1432 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 1441 fs/btrfs/extent-tree.c u64 bytenr = node->bytenr; node 1442 fs/btrfs/extent-tree.c u64 num_bytes = node->num_bytes; node 1488 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 1499 fs/btrfs/extent-tree.c ins.objectid = node->bytenr; node 1500 fs/btrfs/extent-tree.c ins.offset = node->num_bytes; node 1503 fs/btrfs/extent-tree.c ref = btrfs_delayed_node_to_data_ref(node); node 1504 fs/btrfs/extent-tree.c trace_run_delayed_data_ref(trans->fs_info, node, ref, node->action); node 1506 fs/btrfs/extent-tree.c if (node->type == BTRFS_SHARED_DATA_REF_KEY) node 1510 fs/btrfs/extent-tree.c if (node->action == BTRFS_ADD_DELAYED_REF && insert_reserved) { node 1516 fs/btrfs/extent-tree.c node->ref_mod); node 1517 fs/btrfs/extent-tree.c } else if (node->action == BTRFS_ADD_DELAYED_REF) { node 1518 fs/btrfs/extent-tree.c ret = __btrfs_inc_extent_ref(trans, node, parent, ref_root, node 1520 fs/btrfs/extent-tree.c node->ref_mod, extent_op); node 1521 fs/btrfs/extent-tree.c } else if (node->action == BTRFS_DROP_DELAYED_REF) { node 1522 fs/btrfs/extent-tree.c ret = __btrfs_free_extent(trans, node, parent, node 1524 fs/btrfs/extent-tree.c ref->offset, node->ref_mod, node 1638 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 1647 fs/btrfs/extent-tree.c ref = btrfs_delayed_node_to_tree_ref(node); node 1648 fs/btrfs/extent-tree.c trace_run_delayed_tree_ref(trans->fs_info, node, ref, node->action); node 1650 fs/btrfs/extent-tree.c if (node->type == BTRFS_SHARED_BLOCK_REF_KEY) node 1654 fs/btrfs/extent-tree.c if (node->ref_mod != 1) { node 1657 fs/btrfs/extent-tree.c node->bytenr, node->ref_mod, node->action, ref_root, node 1661 fs/btrfs/extent-tree.c if (node->action == BTRFS_ADD_DELAYED_REF && insert_reserved) { node 1663 fs/btrfs/extent-tree.c ret = alloc_reserved_tree_block(trans, node, extent_op); node 1664 fs/btrfs/extent-tree.c } else if (node->action == BTRFS_ADD_DELAYED_REF) { node 1665 fs/btrfs/extent-tree.c ret = __btrfs_inc_extent_ref(trans, node, parent, ref_root, node 1667 fs/btrfs/extent-tree.c } else if (node->action == BTRFS_DROP_DELAYED_REF) { node 1668 fs/btrfs/extent-tree.c ret = __btrfs_free_extent(trans, node, parent, ref_root, node 1678 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 1686 fs/btrfs/extent-tree.c btrfs_pin_extent(trans->fs_info, node->bytenr, node 1687 fs/btrfs/extent-tree.c node->num_bytes, 1); node 1691 fs/btrfs/extent-tree.c if (node->type == BTRFS_TREE_BLOCK_REF_KEY || node 1692 fs/btrfs/extent-tree.c node->type == BTRFS_SHARED_BLOCK_REF_KEY) node 1693 fs/btrfs/extent-tree.c ret = run_delayed_tree_ref(trans, node, extent_op, node 1695 fs/btrfs/extent-tree.c else if (node->type == BTRFS_EXTENT_DATA_REF_KEY || node 1696 fs/btrfs/extent-tree.c node->type == BTRFS_SHARED_DATA_REF_KEY) node 1697 fs/btrfs/extent-tree.c ret = run_delayed_data_ref(trans, node, extent_op, node 1702 fs/btrfs/extent-tree.c btrfs_pin_extent(trans->fs_info, node->bytenr, node 1703 fs/btrfs/extent-tree.c node->num_bytes, 1); node 2165 fs/btrfs/extent-tree.c struct rb_node *node; node 2196 fs/btrfs/extent-tree.c node = rb_first_cached(&delayed_refs->href_root); node 2197 fs/btrfs/extent-tree.c if (!node) { node 2201 fs/btrfs/extent-tree.c head = rb_entry(node, struct btrfs_delayed_ref_head, node 2250 fs/btrfs/extent-tree.c struct rb_node *node; node 2293 fs/btrfs/extent-tree.c for (node = rb_first_cached(&head->ref_tree); node; node 2294 fs/btrfs/extent-tree.c node = rb_next(node)) { node 2295 fs/btrfs/extent-tree.c ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); node 2949 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, u64 parent, node 2968 fs/btrfs/extent-tree.c u64 bytenr = node->bytenr; node 2969 fs/btrfs/extent-tree.c u64 num_bytes = node->num_bytes; node 4264 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *node, node 4281 fs/btrfs/extent-tree.c ref = btrfs_delayed_node_to_tree_ref(node); node 4283 fs/btrfs/extent-tree.c extent_key.objectid = node->bytenr; node 4289 fs/btrfs/extent-tree.c extent_key.offset = node->num_bytes; node 4292 fs/btrfs/extent-tree.c num_bytes = node->num_bytes; node 4324 fs/btrfs/extent-tree.c if (node->type == BTRFS_SHARED_BLOCK_REF_KEY) { node 5088 fs/btrfs/extent-tree.c if (eb == root->node) { node 5247 fs/btrfs/extent-tree.c level = btrfs_header_level(root->node); node 5276 fs/btrfs/extent-tree.c level = btrfs_header_level(root->node); node 5400 fs/btrfs/extent-tree.c free_extent_buffer(root->node); node 5433 fs/btrfs/extent-tree.c struct extent_buffer *node, node 5462 fs/btrfs/extent-tree.c btrfs_assert_tree_locked(node); node 5463 fs/btrfs/extent-tree.c level = btrfs_header_level(node); node 5464 fs/btrfs/extent-tree.c path->nodes[level] = node; node 273 fs/btrfs/extent_io.c struct rb_node *node; node 276 fs/btrfs/extent_io.c node = rb_first(&tree->state); node 277 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 329 fs/btrfs/extent_io.c struct rb_node *node, node 357 fs/btrfs/extent_io.c rb_link_node(node, parent, p); node 358 fs/btrfs/extent_io.c rb_insert_color(node, root); node 523 fs/btrfs/extent_io.c struct rb_node *node; node 535 fs/btrfs/extent_io.c node = tree_insert(&tree->state, NULL, end, &state->rb_node, p, parent); node 536 fs/btrfs/extent_io.c if (node) { node 538 fs/btrfs/extent_io.c found = rb_entry(node, struct extent_state, rb_node); node 565 fs/btrfs/extent_io.c struct rb_node *node; node 575 fs/btrfs/extent_io.c node = tree_insert(&tree->state, &orig->rb_node, prealloc->end, node 577 fs/btrfs/extent_io.c if (node) { node 676 fs/btrfs/extent_io.c struct rb_node *node; node 727 fs/btrfs/extent_io.c node = tree_search(tree, start); node 728 fs/btrfs/extent_io.c if (!node) node 730 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 845 fs/btrfs/extent_io.c struct rb_node *node; node 856 fs/btrfs/extent_io.c node = tree_search(tree, start); node 858 fs/btrfs/extent_io.c if (!node) node 861 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 879 fs/btrfs/extent_io.c node = rb_next(node); node 944 fs/btrfs/extent_io.c struct rb_node *node; node 971 fs/btrfs/extent_io.c node = &state->rb_node; node 979 fs/btrfs/extent_io.c node = tree_search_for_insert(tree, start, &p, &parent); node 980 fs/btrfs/extent_io.c if (!node) { node 992 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1176 fs/btrfs/extent_io.c struct rb_node *node; node 1207 fs/btrfs/extent_io.c node = &state->rb_node; node 1216 fs/btrfs/extent_io.c node = tree_search_for_insert(tree, start, &p, &parent); node 1217 fs/btrfs/extent_io.c if (!node) { node 1231 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1493 fs/btrfs/extent_io.c struct rb_node *node; node 1500 fs/btrfs/extent_io.c node = tree_search(tree, start); node 1501 fs/btrfs/extent_io.c if (!node) node 1505 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1509 fs/btrfs/extent_io.c node = rb_next(node); node 1510 fs/btrfs/extent_io.c if (!node) node 1579 fs/btrfs/extent_io.c struct rb_node *node, *prev = NULL, *next; node 1585 fs/btrfs/extent_io.c node = __etree_search(tree, start, &next, &prev, NULL, NULL); node 1586 fs/btrfs/extent_io.c if (!node && !next && !prev) { node 1594 fs/btrfs/extent_io.c } else if (!node && !next) { node 1603 fs/btrfs/extent_io.c } else if (!node) { node 1604 fs/btrfs/extent_io.c node = next; node 1610 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1661 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1669 fs/btrfs/extent_io.c node = rb_next(node); node 1670 fs/btrfs/extent_io.c if (!node) node 1687 fs/btrfs/extent_io.c struct rb_node *node; node 1699 fs/btrfs/extent_io.c node = tree_search(tree, cur_start); node 1700 fs/btrfs/extent_io.c if (!node) { node 1706 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 1724 fs/btrfs/extent_io.c node = rb_next(node); node 1728 fs/btrfs/extent_io.c if (!node) node 1966 fs/btrfs/extent_io.c struct rb_node *node; node 1985 fs/btrfs/extent_io.c node = tree_search(tree, cur_start); node 1986 fs/btrfs/extent_io.c if (!node) node 1990 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 2008 fs/btrfs/extent_io.c node = rb_next(node); node 2009 fs/btrfs/extent_io.c if (!node) node 2024 fs/btrfs/extent_io.c struct rb_node *node; node 2033 fs/btrfs/extent_io.c node = tree_search(tree, start); node 2034 fs/btrfs/extent_io.c if (!node) { node 2038 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 2052 fs/btrfs/extent_io.c struct rb_node *node; node 2061 fs/btrfs/extent_io.c node = tree_search(tree, start); node 2062 fs/btrfs/extent_io.c if (!node) { node 2066 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 2087 fs/btrfs/extent_io.c struct rb_node *node; node 2093 fs/btrfs/extent_io.c node = &cached->rb_node; node 2095 fs/btrfs/extent_io.c node = tree_search(tree, start); node 2096 fs/btrfs/extent_io.c while (node && start <= end) { node 2097 fs/btrfs/extent_io.c state = rb_entry(node, struct extent_state, rb_node); node 2122 fs/btrfs/extent_io.c node = rb_next(node); node 2123 fs/btrfs/extent_io.c if (!node) { node 254 fs/btrfs/file.c struct rb_node *node; node 257 fs/btrfs/file.c node = rb_first(&fs_info->defrag_inodes); node 258 fs/btrfs/file.c while (node) { node 259 fs/btrfs/file.c rb_erase(node, &fs_info->defrag_inodes); node 260 fs/btrfs/file.c defrag = rb_entry(node, struct inode_defrag, rb_node); node 265 fs/btrfs/file.c node = rb_first(&fs_info->defrag_inodes); node 929 fs/btrfs/free-space-cache.c struct rb_node *node = rb_first(&ctl->free_space_offset); node 939 fs/btrfs/free-space-cache.c if (!node && cluster) { node 942 fs/btrfs/free-space-cache.c node = rb_first(&cluster->root); node 947 fs/btrfs/free-space-cache.c while (node) { node 950 fs/btrfs/free-space-cache.c e = rb_entry(node, struct btrfs_free_space, offset_index); node 962 fs/btrfs/free-space-cache.c node = rb_next(node); node 963 fs/btrfs/free-space-cache.c if (!node && cluster) { node 964 fs/btrfs/free-space-cache.c node = rb_first(&cluster->root); node 1450 fs/btrfs/free-space-cache.c struct rb_node *node, int bitmap) node 1494 fs/btrfs/free-space-cache.c rb_link_node(node, parent, p); node 1495 fs/btrfs/free-space-cache.c rb_insert_color(node, root); node 1812 fs/btrfs/free-space-cache.c struct rb_node *node; node 1824 fs/btrfs/free-space-cache.c for (node = &entry->offset_index; node; node = rb_next(node)) { node 1825 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2073 fs/btrfs/free-space-cache.c struct rb_node *node; node 2080 fs/btrfs/free-space-cache.c node = rb_first(&cluster->root); node 2081 fs/btrfs/free-space-cache.c if (!node) { node 2086 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2546 fs/btrfs/free-space-cache.c struct rb_node *node; node 2556 fs/btrfs/free-space-cache.c node = rb_first(&cluster->root); node 2557 fs/btrfs/free-space-cache.c while (node) { node 2560 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2561 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 2585 fs/btrfs/free-space-cache.c struct rb_node *node; node 2587 fs/btrfs/free-space-cache.c while ((node = rb_last(&ctl->free_space_offset)) != NULL) { node 2588 fs/btrfs/free-space-cache.c info = rb_entry(node, struct btrfs_free_space, offset_index); node 2756 fs/btrfs/free-space-cache.c struct rb_node *node; node 2766 fs/btrfs/free-space-cache.c node = rb_first(&cluster->root); node 2767 fs/btrfs/free-space-cache.c if (!node) node 2770 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2778 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 2779 fs/btrfs/free-space-cache.c if (!node) node 2781 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, node 2792 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 2793 fs/btrfs/free-space-cache.c if (!node) node 2795 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, node 2927 fs/btrfs/free-space-cache.c struct rb_node *node; node 2943 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 2944 fs/btrfs/free-space-cache.c if (!node) node 2946 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2954 fs/btrfs/free-space-cache.c for (node = rb_next(&entry->offset_index); node; node 2955 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index)) { node 2956 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2978 fs/btrfs/free-space-cache.c node = &first->offset_index; node 2987 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); node 2988 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 2997 fs/btrfs/free-space-cache.c } while (node && entry != last); node 3200 fs/btrfs/free-space-cache.c struct rb_node *node; node 3227 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); node 3228 fs/btrfs/free-space-cache.c if (!node) { node 3233 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, node 1144 fs/btrfs/free-space-tree.c struct rb_node *node; node 1160 fs/btrfs/free-space-tree.c node = rb_first(&fs_info->block_group_cache_tree); node 1161 fs/btrfs/free-space-tree.c while (node) { node 1162 fs/btrfs/free-space-tree.c block_group = rb_entry(node, struct btrfs_block_group_cache, node 1167 fs/btrfs/free-space-tree.c node = rb_next(node); node 1249 fs/btrfs/free-space-tree.c btrfs_tree_lock(free_space_root->node); node 1250 fs/btrfs/free-space-tree.c btrfs_clean_tree_block(free_space_root->node); node 1251 fs/btrfs/free-space-tree.c btrfs_tree_unlock(free_space_root->node); node 1252 fs/btrfs/free-space-tree.c btrfs_free_tree_block(trans, free_space_root, free_space_root->node, node 1255 fs/btrfs/free-space-tree.c free_extent_buffer(free_space_root->node); node 2447 fs/btrfs/inode.c struct rb_node node; node 2523 fs/btrfs/inode.c entry = rb_entry(parent, struct sa_defrag_extent_backref, node); node 2532 fs/btrfs/inode.c rb_link_node(&backref->node, parent, p); node 2533 fs/btrfs/inode.c rb_insert_color(&backref->node, root); node 2932 fs/btrfs/inode.c struct rb_node *node; node 2946 fs/btrfs/inode.c node = rb_first(&new->root); node 2947 fs/btrfs/inode.c if (!node) node 2949 fs/btrfs/inode.c rb_erase(node, &new->root); node 2951 fs/btrfs/inode.c backref = rb_entry(node, struct sa_defrag_extent_backref, node); node 4445 fs/btrfs/inode.c struct rb_node *node; node 4456 fs/btrfs/inode.c node = root->inode_tree.rb_node; node 4458 fs/btrfs/inode.c while (node) { node 4459 fs/btrfs/inode.c prev = node; node 4460 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); node 4463 fs/btrfs/inode.c node = node->rb_left; node 4465 fs/btrfs/inode.c node = node->rb_right; node 4469 fs/btrfs/inode.c if (!node) { node 4473 fs/btrfs/inode.c node = prev; node 4479 fs/btrfs/inode.c while (node) { node 4480 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); node 4500 fs/btrfs/inode.c node = rb_next(node); node 5428 fs/btrfs/inode.c struct rb_node *node; node 5437 fs/btrfs/inode.c node = rb_first_cached(&map_tree->map); node 5438 fs/btrfs/inode.c em = rb_entry(node, struct extent_map, rb_node); node 5475 fs/btrfs/inode.c node = rb_first(&io_tree->state); node 5476 fs/btrfs/inode.c state = rb_entry(node, struct extent_state, rb_node); node 10753 fs/btrfs/inode.c entry = rb_entry(parent, struct btrfs_swapfile_pin, node); node 10766 fs/btrfs/inode.c rb_link_node(&sp->node, parent, p); node 10767 fs/btrfs/inode.c rb_insert_color(&sp->node, &fs_info->swapfile_pins); node 10777 fs/btrfs/inode.c struct rb_node *node, *next; node 10780 fs/btrfs/inode.c node = rb_first(&fs_info->swapfile_pins); node 10781 fs/btrfs/inode.c while (node) { node 10782 fs/btrfs/inode.c next = rb_next(node); node 10783 fs/btrfs/inode.c sp = rb_entry(node, struct btrfs_swapfile_pin, node); node 10785 fs/btrfs/inode.c rb_erase(&sp->node, &fs_info->swapfile_pins); node 10790 fs/btrfs/inode.c node = next; node 32 fs/btrfs/ordered-data.c struct rb_node *node) node 50 fs/btrfs/ordered-data.c rb_link_node(node, parent, p); node 51 fs/btrfs/ordered-data.c rb_insert_color(node, root); node 181 fs/btrfs/ordered-data.c struct rb_node *node; node 219 fs/btrfs/ordered-data.c node = tree_insert(&tree->tree, file_offset, node 221 fs/btrfs/ordered-data.c if (node) node 308 fs/btrfs/ordered-data.c struct rb_node *node; node 318 fs/btrfs/ordered-data.c node = tree_search(tree, *file_offset); node 319 fs/btrfs/ordered-data.c if (!node) { node 324 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 378 fs/btrfs/ordered-data.c struct rb_node *node; node 390 fs/btrfs/ordered-data.c node = tree_search(tree, file_offset); node 391 fs/btrfs/ordered-data.c if (!node) { node 396 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 467 fs/btrfs/ordered-data.c struct rb_node *node; node 482 fs/btrfs/ordered-data.c node = &entry->rb_node; node 483 fs/btrfs/ordered-data.c rb_erase(node, &tree->tree); node 484 fs/btrfs/ordered-data.c RB_CLEAR_NODE(node); node 485 fs/btrfs/ordered-data.c if (tree->last == node) node 716 fs/btrfs/ordered-data.c struct rb_node *node; node 721 fs/btrfs/ordered-data.c node = tree_search(tree, file_offset); node 722 fs/btrfs/ordered-data.c if (!node) node 725 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 742 fs/btrfs/ordered-data.c struct rb_node *node; node 747 fs/btrfs/ordered-data.c node = tree_search(tree, file_offset); node 748 fs/btrfs/ordered-data.c if (!node) { node 749 fs/btrfs/ordered-data.c node = tree_search(tree, file_offset + len); node 750 fs/btrfs/ordered-data.c if (!node) node 755 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 764 fs/btrfs/ordered-data.c node = rb_next(node); node 765 fs/btrfs/ordered-data.c if (!node) node 783 fs/btrfs/ordered-data.c struct rb_node *node; node 788 fs/btrfs/ordered-data.c node = tree_search(tree, file_offset); node 789 fs/btrfs/ordered-data.c if (!node) node 792 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 810 fs/btrfs/ordered-data.c struct rb_node *node; node 866 fs/btrfs/ordered-data.c node = rb_prev(&ordered->rb_node); node 878 fs/btrfs/ordered-data.c node = prev; node 880 fs/btrfs/ordered-data.c for (; node; node = rb_prev(node)) { node 881 fs/btrfs/ordered-data.c test = rb_entry(node, struct btrfs_ordered_extent, rb_node); node 18 fs/btrfs/props.c struct hlist_node node; node 51 fs/btrfs/props.c hlist_for_each_entry(h, handlers, node) node 446 fs/btrfs/props.c hash_add(prop_handlers_ht, &p->node, h); node 177 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); node 198 fs/btrfs/qgroup.c qgroup = rb_entry(parent, struct btrfs_qgroup, node); node 217 fs/btrfs/qgroup.c rb_link_node(&qgroup->node, parent, p); node 218 fs/btrfs/qgroup.c rb_insert_color(&qgroup->node, &fs_info->qgroup_tree); node 254 fs/btrfs/qgroup.c rb_erase(&qgroup->node, &fs_info->qgroup_tree); node 520 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); node 1042 fs/btrfs/qgroup.c free_extent_buffer(quota_root->node); node 1103 fs/btrfs/qgroup.c btrfs_tree_lock(quota_root->node); node 1104 fs/btrfs/qgroup.c btrfs_clean_tree_block(quota_root->node); node 1105 fs/btrfs/qgroup.c btrfs_tree_unlock(quota_root->node); node 1106 fs/btrfs/qgroup.c btrfs_free_tree_block(trans, quota_root, quota_root->node, 0, 1); node 1108 fs/btrfs/qgroup.c free_extent_buffer(quota_root->node); node 1559 fs/btrfs/qgroup.c node); node 1574 fs/btrfs/qgroup.c rb_link_node(&record->node, parent_node, p); node 1575 fs/btrfs/qgroup.c rb_insert_color(&record->node, &delayed_refs->dirty_extent_root); node 2513 fs/btrfs/qgroup.c struct rb_node *node; node 2520 fs/btrfs/qgroup.c while ((node = rb_first(&delayed_refs->dirty_extent_root))) { node 2521 fs/btrfs/qgroup.c record = rb_entry(node, struct btrfs_qgroup_extent_record, node 2522 fs/btrfs/qgroup.c node); node 2571 fs/btrfs/qgroup.c rb_erase(node, &delayed_refs->dirty_extent_root); node 3300 fs/btrfs/qgroup.c qgroup = rb_entry(n, struct btrfs_qgroup, node); node 3818 fs/btrfs/qgroup.c node) node 3903 fs/btrfs/qgroup.c node); node 3929 fs/btrfs/qgroup.c rb_link_node(&block->node, parent, cur); node 3930 fs/btrfs/qgroup.c rb_insert_color(&block->node, &blocks->blocks[level]); node 3955 fs/btrfs/qgroup.c struct rb_node *node; node 3972 fs/btrfs/qgroup.c node = blocks->blocks[level].rb_node; node 3974 fs/btrfs/qgroup.c while (node) { node 3975 fs/btrfs/qgroup.c block = rb_entry(node, struct btrfs_qgroup_swapped_block, node); node 3977 fs/btrfs/qgroup.c node = node->rb_left; node 3979 fs/btrfs/qgroup.c node = node->rb_right; node 3990 fs/btrfs/qgroup.c rb_erase(&block->node, &blocks->blocks[level]); node 4036 fs/btrfs/qgroup.c rbtree_postorder_for_each_entry_safe(entry, next, root, node) { node 107 fs/btrfs/qgroup.h struct rb_node node; node 125 fs/btrfs/qgroup.h struct rb_node node; node 218 fs/btrfs/qgroup.h struct rb_node node; /* tree of qgroups */ node 935 fs/btrfs/reada.c struct extent_buffer *node; node 954 fs/btrfs/reada.c node = btrfs_root_node(root); node 955 fs/btrfs/reada.c start = node->start; node 956 fs/btrfs/reada.c generation = btrfs_header_generation(node); node 957 fs/btrfs/reada.c free_extent_buffer(node); node 22 fs/btrfs/ref-verify.c struct rb_node node; node 36 fs/btrfs/ref-verify.c struct rb_node node; node 71 fs/btrfs/ref-verify.c struct rb_node node; node 84 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct block_entry, node); node 93 fs/btrfs/ref-verify.c rb_link_node(&be->node, parent_node, p); node 94 fs/btrfs/ref-verify.c rb_insert_color(&be->node, root); node 105 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); node 125 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct root_entry, node); node 134 fs/btrfs/ref-verify.c rb_link_node(&re->node, parent_node, p); node 135 fs/btrfs/ref-verify.c rb_insert_color(&re->node, root); node 171 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct ref_entry, node); node 181 fs/btrfs/ref-verify.c rb_link_node(&ref->node, parent_node, p); node 182 fs/btrfs/ref-verify.c rb_insert_color(&ref->node, root); node 194 fs/btrfs/ref-verify.c entry = rb_entry(n, struct root_entry, node); node 240 fs/btrfs/ref-verify.c re = rb_entry(n, struct root_entry, node); node 241 fs/btrfs/ref-verify.c rb_erase(&re->node, &be->roots); node 246 fs/btrfs/ref-verify.c ref = rb_entry(n, struct ref_entry, node); node 247 fs/btrfs/ref-verify.c rb_erase(&ref->node, &be->refs); node 644 fs/btrfs/ref-verify.c ref = rb_entry(n, struct ref_entry, node); node 652 fs/btrfs/ref-verify.c re = rb_entry(n, struct root_entry, node); node 831 fs/btrfs/ref-verify.c rb_erase(&exist->node, &be->refs); node 904 fs/btrfs/ref-verify.c be = rb_entry(n, struct block_entry, node); node 905 fs/btrfs/ref-verify.c rb_erase(&be->node, &fs_info->block_tree); node 924 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); node 949 fs/btrfs/ref-verify.c n = &be->node; node 951 fs/btrfs/ref-verify.c be = rb_entry(n, struct block_entry, node); node 970 fs/btrfs/ref-verify.c rb_erase(&be->node, &fs_info->block_tree); node 83 fs/btrfs/relocation.c struct backref_node *node[2]; node 190 fs/btrfs/relocation.c struct backref_node *node); node 192 fs/btrfs/relocation.c struct backref_node *node); node 213 fs/btrfs/relocation.c struct backref_node *node; node 217 fs/btrfs/relocation.c node = list_entry(cache->detached.next, node 219 fs/btrfs/relocation.c remove_backref_node(cache, node); node 223 fs/btrfs/relocation.c node = list_entry(cache->leaves.next, node 225 fs/btrfs/relocation.c remove_backref_node(cache, node); node 241 fs/btrfs/relocation.c struct backref_node *node; node 243 fs/btrfs/relocation.c node = kzalloc(sizeof(*node), GFP_NOFS); node 244 fs/btrfs/relocation.c if (node) { node 245 fs/btrfs/relocation.c INIT_LIST_HEAD(&node->list); node 246 fs/btrfs/relocation.c INIT_LIST_HEAD(&node->upper); node 247 fs/btrfs/relocation.c INIT_LIST_HEAD(&node->lower); node 248 fs/btrfs/relocation.c RB_CLEAR_NODE(&node->rb_node); node 251 fs/btrfs/relocation.c return node; node 255 fs/btrfs/relocation.c struct backref_node *node) node 257 fs/btrfs/relocation.c if (node) { node 259 fs/btrfs/relocation.c kfree(node); node 283 fs/btrfs/relocation.c struct rb_node *node) node 301 fs/btrfs/relocation.c rb_link_node(node, parent, p); node 302 fs/btrfs/relocation.c rb_insert_color(node, root); node 340 fs/btrfs/relocation.c static struct backref_node *walk_up_backref(struct backref_node *node, node 347 fs/btrfs/relocation.c while (!list_empty(&node->upper)) { node 348 fs/btrfs/relocation.c edge = list_entry(node->upper.next, node 351 fs/btrfs/relocation.c node = edge->node[UPPER]; node 353 fs/btrfs/relocation.c BUG_ON(node->detached); node 355 fs/btrfs/relocation.c return node; node 370 fs/btrfs/relocation.c lower = edge->node[LOWER]; node 379 fs/btrfs/relocation.c return edge->node[UPPER]; node 385 fs/btrfs/relocation.c static void unlock_node_buffer(struct backref_node *node) node 387 fs/btrfs/relocation.c if (node->locked) { node 388 fs/btrfs/relocation.c btrfs_tree_unlock(node->eb); node 389 fs/btrfs/relocation.c node->locked = 0; node 393 fs/btrfs/relocation.c static void drop_node_buffer(struct backref_node *node) node 395 fs/btrfs/relocation.c if (node->eb) { node 396 fs/btrfs/relocation.c unlock_node_buffer(node); node 397 fs/btrfs/relocation.c free_extent_buffer(node->eb); node 398 fs/btrfs/relocation.c node->eb = NULL; node 403 fs/btrfs/relocation.c struct backref_node *node) node 405 fs/btrfs/relocation.c BUG_ON(!list_empty(&node->upper)); node 407 fs/btrfs/relocation.c drop_node_buffer(node); node 408 fs/btrfs/relocation.c list_del(&node->list); node 409 fs/btrfs/relocation.c list_del(&node->lower); node 410 fs/btrfs/relocation.c if (!RB_EMPTY_NODE(&node->rb_node)) node 411 fs/btrfs/relocation.c rb_erase(&node->rb_node, &tree->rb_root); node 412 fs/btrfs/relocation.c free_backref_node(tree, node); node 419 fs/btrfs/relocation.c struct backref_node *node) node 424 fs/btrfs/relocation.c if (!node) node 427 fs/btrfs/relocation.c BUG_ON(!node->lowest && !node->detached); node 428 fs/btrfs/relocation.c while (!list_empty(&node->upper)) { node 429 fs/btrfs/relocation.c edge = list_entry(node->upper.next, struct backref_edge, node 431 fs/btrfs/relocation.c upper = edge->node[UPPER]; node 437 fs/btrfs/relocation.c BUG_ON(!list_empty(&node->upper)); node 438 fs/btrfs/relocation.c drop_backref_node(cache, node); node 439 fs/btrfs/relocation.c node = upper; node 440 fs/btrfs/relocation.c node->lowest = 1; node 453 fs/btrfs/relocation.c drop_backref_node(cache, node); node 457 fs/btrfs/relocation.c struct backref_node *node, u64 bytenr) node 460 fs/btrfs/relocation.c rb_erase(&node->rb_node, &cache->rb_root); node 461 fs/btrfs/relocation.c node->bytenr = bytenr; node 462 fs/btrfs/relocation.c rb_node = tree_insert(&cache->rb_root, node->bytenr, &node->rb_node); node 473 fs/btrfs/relocation.c struct backref_node *node; node 490 fs/btrfs/relocation.c node = list_entry(cache->detached.next, node 492 fs/btrfs/relocation.c remove_backref_node(cache, node); node 496 fs/btrfs/relocation.c node = list_entry(cache->changed.next, node 498 fs/btrfs/relocation.c list_del_init(&node->list); node 499 fs/btrfs/relocation.c BUG_ON(node->pending); node 500 fs/btrfs/relocation.c update_backref_node(cache, node, node->new_bytenr); node 508 fs/btrfs/relocation.c list_for_each_entry(node, &cache->pending[level], list) { node 509 fs/btrfs/relocation.c BUG_ON(!node->pending); node 510 fs/btrfs/relocation.c if (node->bytenr == node->new_bytenr) node 512 fs/btrfs/relocation.c update_backref_node(cache, node, node->new_bytenr); node 582 fs/btrfs/relocation.c struct mapping_node *node; node 588 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); node 589 fs/btrfs/relocation.c root = (struct btrfs_root *)node->data; node 694 fs/btrfs/relocation.c struct backref_node *node = NULL; node 717 fs/btrfs/relocation.c node = alloc_backref_node(cache); node 718 fs/btrfs/relocation.c if (!node) { node 723 fs/btrfs/relocation.c node->bytenr = bytenr; node 724 fs/btrfs/relocation.c node->level = level; node 725 fs/btrfs/relocation.c node->lowest = 1; node 726 fs/btrfs/relocation.c cur = node; node 757 fs/btrfs/relocation.c exist = edge->node[UPPER]; node 870 fs/btrfs/relocation.c edge->node[LOWER] = cur; node 871 fs/btrfs/relocation.c edge->node[UPPER] = upper; node 1006 fs/btrfs/relocation.c edge->node[LOWER] = lower; node 1007 fs/btrfs/relocation.c edge->node[UPPER] = upper; node 1036 fs/btrfs/relocation.c cur = edge->node[UPPER]; node 1044 fs/btrfs/relocation.c ASSERT(node->checked); node 1045 fs/btrfs/relocation.c cowonly = node->cowonly; node 1047 fs/btrfs/relocation.c rb_node = tree_insert(&cache->rb_root, node->bytenr, node 1048 fs/btrfs/relocation.c &node->rb_node); node 1050 fs/btrfs/relocation.c backref_tree_panic(rb_node, -EEXIST, node->bytenr); node 1051 fs/btrfs/relocation.c list_add_tail(&node->lower, &cache->leaves); node 1054 fs/btrfs/relocation.c list_for_each_entry(edge, &node->upper, list[LOWER]) node 1060 fs/btrfs/relocation.c upper = edge->node[UPPER]; node 1063 fs/btrfs/relocation.c lower = edge->node[LOWER]; node 1118 fs/btrfs/relocation.c if (upper == node) node 1119 fs/btrfs/relocation.c node = NULL; node 1129 fs/btrfs/relocation.c lower = edge->node[LOWER]; node 1158 fs/btrfs/relocation.c lower = edge->node[LOWER]; node 1159 fs/btrfs/relocation.c upper = edge->node[UPPER]; node 1184 fs/btrfs/relocation.c if (lower == node) node 1185 fs/btrfs/relocation.c node = NULL; node 1189 fs/btrfs/relocation.c remove_backref_node(cache, node); node 1192 fs/btrfs/relocation.c ASSERT(!node || !node->detached); node 1193 fs/btrfs/relocation.c return node; node 1208 fs/btrfs/relocation.c struct backref_node *node = NULL; node 1219 fs/btrfs/relocation.c node = rb_entry(rb_node, struct backref_node, rb_node); node 1220 fs/btrfs/relocation.c if (node->detached) node 1221 fs/btrfs/relocation.c node = NULL; node 1223 fs/btrfs/relocation.c BUG_ON(node->new_bytenr != reloc_root->node->start); node 1226 fs/btrfs/relocation.c if (!node) { node 1230 fs/btrfs/relocation.c node = rb_entry(rb_node, struct backref_node, node 1232 fs/btrfs/relocation.c BUG_ON(node->detached); node 1236 fs/btrfs/relocation.c if (!node) node 1243 fs/btrfs/relocation.c new_node->bytenr = dest->node->start; node 1244 fs/btrfs/relocation.c new_node->level = node->level; node 1245 fs/btrfs/relocation.c new_node->lowest = node->lowest; node 1249 fs/btrfs/relocation.c if (!node->lowest) { node 1250 fs/btrfs/relocation.c list_for_each_entry(edge, &node->lower, list[UPPER]) { node 1255 fs/btrfs/relocation.c new_edge->node[UPPER] = new_node; node 1256 fs/btrfs/relocation.c new_edge->node[LOWER] = edge->node[LOWER]; node 1272 fs/btrfs/relocation.c &new_edge->node[LOWER]->upper); node 1294 fs/btrfs/relocation.c struct mapping_node *node; node 1297 fs/btrfs/relocation.c node = kmalloc(sizeof(*node), GFP_NOFS); node 1298 fs/btrfs/relocation.c if (!node) node 1301 fs/btrfs/relocation.c node->bytenr = root->commit_root->start; node 1302 fs/btrfs/relocation.c node->data = root; node 1306 fs/btrfs/relocation.c node->bytenr, &node->rb_node); node 1311 fs/btrfs/relocation.c node->bytenr); node 1326 fs/btrfs/relocation.c struct mapping_node *node = NULL; node 1329 fs/btrfs/relocation.c if (rc && root->node) { node 1334 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); node 1335 fs/btrfs/relocation.c rb_erase(&node->rb_node, &rc->reloc_root_tree.rb_root); node 1336 fs/btrfs/relocation.c RB_CLEAR_NODE(&node->rb_node); node 1339 fs/btrfs/relocation.c if (!node) node 1341 fs/btrfs/relocation.c BUG_ON((struct btrfs_root *)node->data != root); node 1347 fs/btrfs/relocation.c kfree(node); node 1358 fs/btrfs/relocation.c struct mapping_node *node = NULL; node 1365 fs/btrfs/relocation.c node = rb_entry(rb_node, struct mapping_node, rb_node); node 1366 fs/btrfs/relocation.c rb_erase(&node->rb_node, &rc->reloc_root_tree.rb_root); node 1370 fs/btrfs/relocation.c if (!node) node 1372 fs/btrfs/relocation.c BUG_ON((struct btrfs_root *)node->data != root); node 1375 fs/btrfs/relocation.c node->bytenr = root->node->start; node 1377 fs/btrfs/relocation.c node->bytenr, &node->rb_node); node 1380 fs/btrfs/relocation.c backref_tree_panic(rb_node, -EEXIST, node->bytenr); node 1426 fs/btrfs/relocation.c ret = btrfs_copy_root(trans, root, root->node, &eb, node 1532 fs/btrfs/relocation.c if (reloc_root->commit_root != reloc_root->node) { node 1534 fs/btrfs/relocation.c btrfs_set_root_node(root_item, reloc_root->node); node 1553 fs/btrfs/relocation.c struct rb_node *node; node 1560 fs/btrfs/relocation.c node = root->inode_tree.rb_node; node 1562 fs/btrfs/relocation.c while (node) { node 1563 fs/btrfs/relocation.c prev = node; node 1564 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); node 1567 fs/btrfs/relocation.c node = node->rb_left; node 1569 fs/btrfs/relocation.c node = node->rb_right; node 1573 fs/btrfs/relocation.c if (!node) { node 1577 fs/btrfs/relocation.c node = prev; node 1583 fs/btrfs/relocation.c while (node) { node 1584 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); node 1595 fs/btrfs/relocation.c node = rb_next(node); node 2292 fs/btrfs/relocation.c extent_buffer_get(reloc_root->node); node 2293 fs/btrfs/relocation.c path->nodes[level] = reloc_root->node; node 2494 fs/btrfs/relocation.c free_extent_buffer(reloc_root->node); node 2496 fs/btrfs/relocation.c reloc_root->node = NULL; node 2614 fs/btrfs/relocation.c struct backref_node *node, node 2621 fs/btrfs/relocation.c next = node; node 2637 fs/btrfs/relocation.c if (next->new_bytenr != root->node->start) { node 2640 fs/btrfs/relocation.c next->new_bytenr = root->node->start; node 2651 fs/btrfs/relocation.c if (!next || next->level <= node->level) node 2657 fs/btrfs/relocation.c next = node; node 2663 fs/btrfs/relocation.c next = edges[index]->node[UPPER]; node 2675 fs/btrfs/relocation.c struct btrfs_root *select_one_root(struct backref_node *node) node 2683 fs/btrfs/relocation.c next = node; node 2697 fs/btrfs/relocation.c if (next != node) node 2701 fs/btrfs/relocation.c if (!next || next->level <= node->level) node 2712 fs/btrfs/relocation.c struct backref_node *node, int reserve) node 2715 fs/btrfs/relocation.c struct backref_node *next = node; node 2721 fs/btrfs/relocation.c BUG_ON(reserve && node->processed); node 2726 fs/btrfs/relocation.c if (next->processed && (reserve || next != node)) node 2737 fs/btrfs/relocation.c next = edge->node[UPPER]; node 2746 fs/btrfs/relocation.c struct backref_node *node) node 2754 fs/btrfs/relocation.c num_bytes = calcu_metadata_size(rc, node, 1) * 2; node 2794 fs/btrfs/relocation.c struct backref_node *node, node 2811 fs/btrfs/relocation.c BUG_ON(lowest && node->eb); node 2813 fs/btrfs/relocation.c path->lowest_level = node->level + 1; node 2814 fs/btrfs/relocation.c rc->backref_cache.path[node->level] = node; node 2815 fs/btrfs/relocation.c list_for_each_entry(edge, &node->upper, list[LOWER]) { node 2821 fs/btrfs/relocation.c upper = edge->node[UPPER]; node 2835 fs/btrfs/relocation.c if (node->eb->start == bytenr) node 2877 fs/btrfs/relocation.c if (bytenr != node->bytenr) { node 2880 fs/btrfs/relocation.c bytenr, node->bytenr, slot, node 2886 fs/btrfs/relocation.c if (node->eb->start == bytenr) node 2906 fs/btrfs/relocation.c if (!node->eb) { node 2915 fs/btrfs/relocation.c BUG_ON(node->eb != eb); node 2918 fs/btrfs/relocation.c node->eb->start); node 2924 fs/btrfs/relocation.c node->eb->start, blocksize, node 2927 fs/btrfs/relocation.c btrfs_init_tree_ref(&ref, node->level, node 2944 fs/btrfs/relocation.c if (!err && node->pending) { node 2945 fs/btrfs/relocation.c drop_node_buffer(node); node 2946 fs/btrfs/relocation.c list_move_tail(&node->list, &rc->backref_cache.changed); node 2947 fs/btrfs/relocation.c node->pending = 0; node 2957 fs/btrfs/relocation.c struct backref_node *node, node 2962 fs/btrfs/relocation.c btrfs_node_key_to_cpu(node->eb, &key, 0); node 2963 fs/btrfs/relocation.c return do_relocation(trans, rc, node, &key, path, 0); node 2972 fs/btrfs/relocation.c struct backref_node *node; node 2978 fs/btrfs/relocation.c node = list_entry(cache->pending[level].next, node 2980 fs/btrfs/relocation.c list_move_tail(&node->list, &list); node 2981 fs/btrfs/relocation.c BUG_ON(!node->pending); node 2984 fs/btrfs/relocation.c ret = link_to_upper(trans, rc, node, path); node 3002 fs/btrfs/relocation.c struct backref_node *node) node 3005 fs/btrfs/relocation.c if (node->level == 0 || node 3006 fs/btrfs/relocation.c in_block_group(node->bytenr, rc->block_group)) { node 3008 fs/btrfs/relocation.c mark_block_processed(rc, node->bytenr, blocksize); node 3010 fs/btrfs/relocation.c node->processed = 1; node 3018 fs/btrfs/relocation.c struct backref_node *node) node 3020 fs/btrfs/relocation.c struct backref_node *next = node; node 3039 fs/btrfs/relocation.c next = edge->node[UPPER]; node 3083 fs/btrfs/relocation.c struct backref_node *node, node 3090 fs/btrfs/relocation.c if (!node) node 3093 fs/btrfs/relocation.c BUG_ON(node->processed); node 3094 fs/btrfs/relocation.c root = select_one_root(node); node 3096 fs/btrfs/relocation.c update_processed_blocks(rc, node); node 3101 fs/btrfs/relocation.c ret = reserve_metadata_space(trans, rc, node); node 3108 fs/btrfs/relocation.c BUG_ON(node->new_bytenr); node 3109 fs/btrfs/relocation.c BUG_ON(!list_empty(&node->list)); node 3112 fs/btrfs/relocation.c node->new_bytenr = root->node->start; node 3113 fs/btrfs/relocation.c node->root = root; node 3114 fs/btrfs/relocation.c list_add_tail(&node->list, &rc->backref_cache.changed); node 3116 fs/btrfs/relocation.c path->lowest_level = node->level; node 3123 fs/btrfs/relocation.c update_processed_blocks(rc, node); node 3125 fs/btrfs/relocation.c ret = do_relocation(trans, rc, node, key, path, 1); node 3128 fs/btrfs/relocation.c if (ret || node->level == 0 || node->cowonly) node 3129 fs/btrfs/relocation.c remove_backref_node(&rc->backref_cache, node); node 3141 fs/btrfs/relocation.c struct backref_node *node; node 3171 fs/btrfs/relocation.c node = build_backref_tree(rc, &block->key, node 3173 fs/btrfs/relocation.c if (IS_ERR(node)) { node 3174 fs/btrfs/relocation.c err = PTR_ERR(node); node 3178 fs/btrfs/relocation.c ret = relocate_tree_block(trans, rc, node, &block->key, node 4712 fs/btrfs/relocation.c struct backref_node *node; node 4733 fs/btrfs/relocation.c node = rc->backref_cache.path[level]; node 4734 fs/btrfs/relocation.c BUG_ON(node->bytenr != buf->start && node 4735 fs/btrfs/relocation.c node->new_bytenr != buf->start); node 4737 fs/btrfs/relocation.c drop_node_buffer(node); node 4739 fs/btrfs/relocation.c node->eb = cow; node 4740 fs/btrfs/relocation.c node->new_bytenr = cow->start; node 4742 fs/btrfs/relocation.c if (!node->pending) { node 4743 fs/btrfs/relocation.c list_move_tail(&node->list, node 4745 fs/btrfs/relocation.c node->pending = 1; node 4749 fs/btrfs/relocation.c __mark_block_processed(rc, node); node 110 fs/btrfs/root-tree.c struct extent_buffer *node) node 112 fs/btrfs/root-tree.c btrfs_set_root_bytenr(item, node->start); node 113 fs/btrfs/root-tree.c btrfs_set_root_level(item, btrfs_header_level(node)); node 114 fs/btrfs/root-tree.c btrfs_set_root_generation(item, btrfs_header_generation(node)); node 203 fs/btrfs/scrub.c struct rb_node node; node 333 fs/btrfs/scrub.c entry = rb_entry(parent, struct full_stripe_lock, node); node 354 fs/btrfs/scrub.c rb_link_node(&ret->node, parent, p); node 355 fs/btrfs/scrub.c rb_insert_color(&ret->node, &locks_root->root); node 369 fs/btrfs/scrub.c struct rb_node *node; node 374 fs/btrfs/scrub.c node = locks_root->root.rb_node; node 375 fs/btrfs/scrub.c while (node) { node 376 fs/btrfs/scrub.c entry = rb_entry(node, struct full_stripe_lock, node); node 378 fs/btrfs/scrub.c node = node->rb_left; node 380 fs/btrfs/scrub.c node = node->rb_right; node 514 fs/btrfs/scrub.c rb_erase(&fstripe_lock->node, &locks_root->root); node 223 fs/btrfs/send.c struct rb_node node; node 232 fs/btrfs/send.c struct rb_node node; node 244 fs/btrfs/send.c struct rb_node node; node 2867 fs/btrfs/send.c entry = rb_entry(parent, struct orphan_dir_info, node); node 2884 fs/btrfs/send.c rb_link_node(&odi->node, parent, p); node 2885 fs/btrfs/send.c rb_insert_color(&odi->node, &sctx->orphan_dirs); node 2896 fs/btrfs/send.c entry = rb_entry(n, struct orphan_dir_info, node); node 2919 fs/btrfs/send.c rb_erase(&odi->node, &sctx->orphan_dirs); node 3042 fs/btrfs/send.c entry = rb_entry(parent, struct waiting_dir_move, node); node 3053 fs/btrfs/send.c rb_link_node(&dm->node, parent, p); node 3054 fs/btrfs/send.c rb_insert_color(&dm->node, &sctx->waiting_dir_moves); node 3065 fs/btrfs/send.c entry = rb_entry(n, struct waiting_dir_move, node); node 3081 fs/btrfs/send.c rb_erase(&dm->node, &sctx->waiting_dir_moves); node 3108 fs/btrfs/send.c RB_CLEAR_NODE(&pm->node); node 3112 fs/btrfs/send.c entry = rb_entry(parent, struct pending_dir_move, node); node 3141 fs/btrfs/send.c rb_link_node(&pm->node, parent, p); node 3142 fs/btrfs/send.c rb_insert_color(&pm->node, &sctx->pending_dir_moves); node 3160 fs/btrfs/send.c entry = rb_entry(n, struct pending_dir_move, node); node 3357 fs/btrfs/send.c if (!RB_EMPTY_NODE(&m->node)) node 3358 fs/btrfs/send.c rb_erase(&m->node, &sctx->pending_dir_moves); node 3375 fs/btrfs/send.c if (!RB_EMPTY_NODE(&moves->node)) { node 3376 fs/btrfs/send.c rb_erase(&moves->node, &sctx->pending_dir_moves); node 3377 fs/btrfs/send.c RB_CLEAR_NODE(&moves->node); node 6968 fs/btrfs/send.c sctx->parent_root->node != sctx->parent_root->commit_root) node 6972 fs/btrfs/send.c if (sctx->clone_roots[i].root->node != node 7313 fs/btrfs/send.c pm = rb_entry(n, struct pending_dir_move, node); node 7330 fs/btrfs/send.c dm = rb_entry(n, struct waiting_dir_move, node); node 7331 fs/btrfs/send.c rb_erase(&dm->node, &sctx->waiting_dir_moves); node 7341 fs/btrfs/send.c odi = rb_entry(n, struct orphan_dir_info, node); node 197 fs/btrfs/tests/btrfs-tests.c if (root->node) { node 199 fs/btrfs/tests/btrfs-tests.c free_extent_buffer(root->node); node 13 fs/btrfs/tests/extent-map-tests.c struct rb_node *node; node 16 fs/btrfs/tests/extent-map-tests.c node = rb_first_cached(&em_tree->map); node 17 fs/btrfs/tests/extent-map-tests.c em = rb_entry(node, struct extent_map, rb_node); node 465 fs/btrfs/tests/free-space-tree-tests.c root->node = alloc_test_extent_buffer(root->fs_info, nodesize); node 466 fs/btrfs/tests/free-space-tree-tests.c if (IS_ERR(root->node)) { node 468 fs/btrfs/tests/free-space-tree-tests.c ret = PTR_ERR(root->node); node 471 fs/btrfs/tests/free-space-tree-tests.c btrfs_set_header_level(root->node, 0); node 472 fs/btrfs/tests/free-space-tree-tests.c btrfs_set_header_nritems(root->node, 0); node 519 fs/btrfs/tests/free-space-tree-tests.c if (btrfs_header_nritems(root->node) != 0) { node 21 fs/btrfs/tests/inode-tests.c struct extent_buffer *leaf = root->node; node 54 fs/btrfs/tests/inode-tests.c struct extent_buffer *leaf = root->node; node 254 fs/btrfs/tests/inode-tests.c root->node = alloc_dummy_extent_buffer(fs_info, nodesize); node 255 fs/btrfs/tests/inode-tests.c if (!root->node) { node 260 fs/btrfs/tests/inode-tests.c btrfs_set_header_nritems(root->node, 0); node 261 fs/btrfs/tests/inode-tests.c btrfs_set_header_level(root->node, 0); node 857 fs/btrfs/tests/inode-tests.c root->node = alloc_dummy_extent_buffer(fs_info, nodesize); node 858 fs/btrfs/tests/inode-tests.c if (!root->node) { node 863 fs/btrfs/tests/inode-tests.c btrfs_set_header_nritems(root->node, 0); node 864 fs/btrfs/tests/inode-tests.c btrfs_set_header_level(root->node, 0); node 486 fs/btrfs/tests/qgroup-tests.c root->node = alloc_test_extent_buffer(root->fs_info, nodesize); node 487 fs/btrfs/tests/qgroup-tests.c if (IS_ERR(root->node)) { node 489 fs/btrfs/tests/qgroup-tests.c ret = PTR_ERR(root->node); node 492 fs/btrfs/tests/qgroup-tests.c btrfs_set_header_level(root->node, 0); node 493 fs/btrfs/tests/qgroup-tests.c btrfs_set_header_nritems(root->node, 0); node 293 fs/btrfs/transaction.c WARN_ON(!force && root->commit_root != root->node); node 1097 fs/btrfs/transaction.c if (old_root_bytenr == root->node->start && node 1101 fs/btrfs/transaction.c btrfs_set_root_node(&root->root_item, root->node); node 1252 fs/btrfs/transaction.c if (root->commit_root != root->node) { node 1256 fs/btrfs/transaction.c root->node); node 2252 fs/btrfs/transaction.c fs_info->tree_root->node); node 2257 fs/btrfs/transaction.c fs_info->chunk_root->node); node 2383 fs/btrfs/transaction.c if (btrfs_header_backref_rev(root->node) < node 1445 fs/btrfs/tree-checker.c int btrfs_check_node(struct extent_buffer *node) node 1447 fs/btrfs/tree-checker.c struct btrfs_fs_info *fs_info = node->fs_info; node 1448 fs/btrfs/tree-checker.c unsigned long nr = btrfs_header_nritems(node); node 1451 fs/btrfs/tree-checker.c int level = btrfs_header_level(node); node 1456 fs/btrfs/tree-checker.c generic_err(node, 0, node 1464 fs/btrfs/tree-checker.c btrfs_header_owner(node), node->start, node 1471 fs/btrfs/tree-checker.c bytenr = btrfs_node_blockptr(node, slot); node 1472 fs/btrfs/tree-checker.c btrfs_node_key_to_cpu(node, &key, slot); node 1473 fs/btrfs/tree-checker.c btrfs_node_key_to_cpu(node, &next_key, slot + 1); node 1476 fs/btrfs/tree-checker.c generic_err(node, slot, node 1482 fs/btrfs/tree-checker.c generic_err(node, slot, node 1490 fs/btrfs/tree-checker.c generic_err(node, slot, node 24 fs/btrfs/tree-checker.h int btrfs_check_node(struct extent_buffer *node); node 45 fs/btrfs/tree-defrag.c level = btrfs_header_level(root->node); node 2807 fs/btrfs/tree-log.c if (path->nodes[*level] == root->node) node 2870 fs/btrfs/tree-log.c level = btrfs_header_level(log->node); node 2872 fs/btrfs/tree-log.c path->nodes[level] = log->node; node 2873 fs/btrfs/tree-log.c extent_buffer_get(log->node); node 3123 fs/btrfs/tree-log.c btrfs_set_root_node(&log->root_item, log->node); node 3242 fs/btrfs/tree-log.c log_root_tree->node->start); node 3244 fs/btrfs/tree-log.c btrfs_header_level(log_root_tree->node)); node 3317 fs/btrfs/tree-log.c free_extent_buffer(log->node); node 6247 fs/btrfs/tree-log.c log->node->start, node 6248 fs/btrfs/tree-log.c log->node->len); node 6249 fs/btrfs/tree-log.c free_extent_buffer(log->node); node 6287 fs/btrfs/tree-log.c free_extent_buffer(log->node); node 6320 fs/btrfs/tree-log.c free_extent_buffer(log_root_tree->node); node 63 fs/btrfs/ulist.c struct ulist_node *node; node 66 fs/btrfs/ulist.c list_for_each_entry_safe(node, next, &ulist->nodes, list) { node 67 fs/btrfs/ulist.c kfree(node); node 135 fs/btrfs/ulist.c static void ulist_rbtree_erase(struct ulist *ulist, struct ulist_node *node) node 137 fs/btrfs/ulist.c rb_erase(&node->rb_node, &ulist->root); node 138 fs/btrfs/ulist.c list_del(&node->list); node 139 fs/btrfs/ulist.c kfree(node); node 195 fs/btrfs/ulist.c struct ulist_node *node; node 197 fs/btrfs/ulist.c node = ulist_rbtree_search(ulist, val); node 198 fs/btrfs/ulist.c if (node) { node 200 fs/btrfs/ulist.c *old_aux = node->aux; node 203 fs/btrfs/ulist.c node = kmalloc(sizeof(*node), gfp_mask); node 204 fs/btrfs/ulist.c if (!node) node 207 fs/btrfs/ulist.c node->val = val; node 208 fs/btrfs/ulist.c node->aux = aux; node 210 fs/btrfs/ulist.c ret = ulist_rbtree_insert(ulist, node); node 212 fs/btrfs/ulist.c list_add_tail(&node->list, &ulist->nodes); node 230 fs/btrfs/ulist.c struct ulist_node *node; node 232 fs/btrfs/ulist.c node = ulist_rbtree_search(ulist, val); node 234 fs/btrfs/ulist.c if (!node) node 237 fs/btrfs/ulist.c if (node->aux != aux) node 241 fs/btrfs/ulist.c ulist_rbtree_erase(ulist, node); node 263 fs/btrfs/ulist.c struct ulist_node *node; node 274 fs/btrfs/ulist.c node = list_entry(uiter->cur_list, struct ulist_node, list); node 275 fs/btrfs/ulist.c return node; node 7796 fs/btrfs/volumes.c struct rb_node *node; node 7800 fs/btrfs/volumes.c for (node = rb_first_cached(&em_tree->map); node; node = rb_next(node)) { node 7801 fs/btrfs/volumes.c em = rb_entry(node, struct extent_map, rb_node); node 7914 fs/btrfs/volumes.c struct rb_node *node; node 7917 fs/btrfs/volumes.c node = fs_info->swapfile_pins.rb_node; node 7918 fs/btrfs/volumes.c while (node) { node 7919 fs/btrfs/volumes.c sp = rb_entry(node, struct btrfs_swapfile_pin, node); node 7921 fs/btrfs/volumes.c node = node->rb_left; node 7923 fs/btrfs/volumes.c node = node->rb_right; node 7928 fs/btrfs/volumes.c return node != NULL; node 1829 fs/ceph/addr.c perm = rb_entry(*p, struct ceph_pool_perm, node); node 1863 fs/ceph/addr.c perm = rb_entry(parent, struct ceph_pool_perm, node); node 1971 fs/ceph/addr.c rb_link_node(&perm->node, parent, p); node 1972 fs/ceph/addr.c rb_insert_color(&perm->node, &mdsc->pool_perm_tree); node 2060 fs/ceph/addr.c perm = rb_entry(n, struct ceph_pool_perm, node); node 126 fs/ceph/inode.c frag = rb_entry(parent, struct ceph_inode_frag, node); node 145 fs/ceph/inode.c rb_link_node(&frag->node, parent, p); node 146 fs/ceph/inode.c rb_insert_color(&frag->node, &ci->i_fragtree); node 162 fs/ceph/inode.c rb_entry(n, struct ceph_inode_frag, node); node 266 fs/ceph/inode.c rb_erase(&frag->node, &ci->i_fragtree); node 341 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); node 371 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); node 383 fs/ceph/inode.c rb_erase(&frag->node, &ci->i_fragtree); node 402 fs/ceph/inode.c frag = rb_entry(rb_node, struct ceph_inode_frag, node); node 407 fs/ceph/inode.c rb_erase(&frag->node, &ci->i_fragtree); node 574 fs/ceph/inode.c frag = rb_entry(n, struct ceph_inode_frag, node); node 3522 fs/ceph/mds_client.c rb_entry(p, struct ceph_snap_realm, node); node 315 fs/ceph/mds_client.h struct rb_node node; node 323 fs/ceph/mds_client.h struct rb_node node; node 336 fs/ceph/mds_client.h struct rb_node node; node 85 fs/ceph/quota.c struct rb_node **node, *parent = NULL; node 88 fs/ceph/quota.c node = &(mdsc->quotarealms_inodes.rb_node); node 89 fs/ceph/quota.c while (*node) { node 90 fs/ceph/quota.c parent = *node; node 91 fs/ceph/quota.c qri = container_of(*node, struct ceph_quotarealm_inode, node); node 94 fs/ceph/quota.c node = &((*node)->rb_left); node 96 fs/ceph/quota.c node = &((*node)->rb_right); node 108 fs/ceph/quota.c rb_link_node(&qri->node, parent, node); node 109 fs/ceph/quota.c rb_insert_color(&qri->node, &mdsc->quotarealms_inodes); node 177 fs/ceph/quota.c struct rb_node *node; node 185 fs/ceph/quota.c node = rb_first(&mdsc->quotarealms_inodes); node 186 fs/ceph/quota.c qri = rb_entry(node, struct ceph_quotarealm_inode, node); node 187 fs/ceph/quota.c rb_erase(node, &mdsc->quotarealms_inodes); node 92 fs/ceph/snap.c r = rb_entry(parent, struct ceph_snap_realm, node); node 101 fs/ceph/snap.c rb_link_node(&new->node, parent, p); node 102 fs/ceph/snap.c rb_insert_color(&new->node, root); node 147 fs/ceph/snap.c r = rb_entry(n, struct ceph_snap_realm, node); node 181 fs/ceph/snap.c rb_erase(&realm->node, &mdsc->snap_realms); node 1021 fs/ceph/snap.c exist = rb_entry(*p, struct ceph_snapid_map, node); node 1059 fs/ceph/snap.c exist = rb_entry(*p, struct ceph_snapid_map, node); node 1072 fs/ceph/snap.c rb_link_node(&sm->node, parent, p); node 1073 fs/ceph/snap.c rb_insert_color(&sm->node, &mdsc->snapid_map_tree); node 1093 fs/ceph/snap.c if (!RB_EMPTY_NODE(&sm->node)) { node 1121 fs/ceph/snap.c rb_erase(&sm->node, &mdsc->snapid_map_tree); node 1143 fs/ceph/snap.c sm = rb_entry(p, struct ceph_snapid_map, node); node 238 fs/ceph/super.h struct rb_node node; node 255 fs/ceph/super.h struct rb_node node; node 801 fs/ceph/super.h struct rb_node node; node 448 fs/ceph/xattr.c xattr = rb_entry(parent, struct ceph_inode_xattr, node); node 521 fs/ceph/xattr.c rb_link_node(&xattr->node, parent, p); node 522 fs/ceph/xattr.c rb_insert_color(&xattr->node, &ci->i_xattrs.index); node 544 fs/ceph/xattr.c xattr = rb_entry(parent, struct ceph_inode_xattr, node); node 582 fs/ceph/xattr.c rb_erase(&xattr->node, &ci->i_xattrs.index); node 607 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); node 631 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); node 780 fs/ceph/xattr.c xattr = rb_entry(p, struct ceph_inode_xattr, node); node 5167 fs/cifs/connect.c struct rb_node *node; node 5173 fs/cifs/connect.c while ((node = rb_first(root))) { node 5174 fs/cifs/connect.c tlink = rb_entry(node, struct tcon_link, tl_rbnode); node 5177 fs/cifs/connect.c rb_erase(node, root); node 5337 fs/cifs/connect.c struct rb_node *node = root->rb_node; node 5340 fs/cifs/connect.c while (node) { node 5341 fs/cifs/connect.c tlink = rb_entry(node, struct tcon_link, tl_rbnode); node 5344 fs/cifs/connect.c node = node->rb_left; node 5346 fs/cifs/connect.c node = node->rb_right; node 5473 fs/cifs/connect.c struct rb_node *node; node 5485 fs/cifs/connect.c node = rb_first(root); node 5486 fs/cifs/connect.c while (node != NULL) { node 5487 fs/cifs/connect.c tmp = node; node 5488 fs/cifs/connect.c node = rb_next(tmp); node 732 fs/cifs/misc.c struct dfs_info3_param *node = (*target_nodes)+i; node 734 fs/cifs/misc.c node->flags = le32_to_cpu(rsp->DFSFlags); node 744 fs/cifs/misc.c node->path_consumed = cifs_utf16_bytes(tmp, node 749 fs/cifs/misc.c node->path_consumed = le16_to_cpu(rsp->PathConsumed); node 751 fs/cifs/misc.c node->server_type = le16_to_cpu(ref->ServerType); node 752 fs/cifs/misc.c node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags); node 757 fs/cifs/misc.c node->path_name = cifs_strndup_from_utf16(temp, max_len, node 759 fs/cifs/misc.c if (!node->path_name) { node 767 fs/cifs/misc.c node->node_name = cifs_strndup_from_utf16(temp, max_len, node 769 fs/cifs/misc.c if (!node->node_name) { node 774 fs/cifs/misc.c node->ttl = le32_to_cpu(ref->TimeToLive); node 2219 fs/dcache.c struct hlist_bl_node *node; node 2242 fs/dcache.c hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) { node 2341 fs/dcache.c struct hlist_bl_node *node; node 2367 fs/dcache.c hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) { node 2518 fs/dcache.c struct hlist_bl_node *node; node 2568 fs/dcache.c hlist_bl_for_each_entry(dentry, node, b, d_u.d_in_lookup_hash) { node 780 fs/dlm/config.c struct dlm_config_node *nodes, *node; node 802 fs/dlm/config.c node = nodes; node 804 fs/dlm/config.c node->nodeid = nd->nodeid; node 805 fs/dlm/config.c node->weight = nd->weight; node 806 fs/dlm/config.c node->new = nd->new; node 807 fs/dlm/config.c node->comm_seq = nd->comm_seq; node 808 fs/dlm/config.c node++; node 422 fs/dlm/debug_fs.c struct rb_node *node; node 454 fs/dlm/debug_fs.c for (node = rb_first(tree); node; node = rb_next(node)) { node 455 fs/dlm/debug_fs.c r = rb_entry(node, struct dlm_rsb, res_hashnode); node 486 fs/dlm/debug_fs.c node = rb_first(tree); node 487 fs/dlm/debug_fs.c r = rb_entry(node, struct dlm_rsb, res_hashnode); node 35 fs/dlm/dir.c uint32_t node; node 40 fs/dlm/dir.c node = (hash >> 16) % ls->ls_total_weight; node 41 fs/dlm/dir.c return ls->ls_node_array[node]; node 448 fs/dlm/lock.c struct rb_node *node = tree->rb_node; node 452 fs/dlm/lock.c while (node) { node 453 fs/dlm/lock.c r = rb_entry(node, struct dlm_rsb, res_hashnode); node 456 fs/dlm/lock.c node = node->rb_left; node 458 fs/dlm/lock.c node = node->rb_right; node 314 fs/dlm/member.c static int dlm_add_member(struct dlm_ls *ls, struct dlm_config_node *node) node 323 fs/dlm/member.c error = dlm_lowcomms_connect_node(node->nodeid); node 329 fs/dlm/member.c memb->nodeid = node->nodeid; node 330 fs/dlm/member.c memb->weight = node->weight; node 331 fs/dlm/member.c memb->comm_seq = node->comm_seq; node 527 fs/dlm/member.c struct dlm_config_node *node; node 541 fs/dlm/member.c node = find_config_node(rv, memb->nodeid); node 542 fs/dlm/member.c if (node && !node->new) node 545 fs/dlm/member.c if (!node) { node 550 fs/dlm/member.c memb->nodeid, memb->comm_seq, node->comm_seq); node 562 fs/dlm/member.c node = &rv->nodes[i]; node 563 fs/dlm/member.c if (dlm_is_member(ls, node->nodeid)) node 565 fs/dlm/member.c dlm_add_member(ls, node); node 566 fs/dlm/member.c log_rinfo(ls, "add member %d", node->nodeid); node 400 fs/ecryptfs/ecryptfs_kernel.h struct list_head node; node 53 fs/ecryptfs/messaging.c *msg_ctx = list_entry(p, struct ecryptfs_msg_ctx, node); node 73 fs/ecryptfs/messaging.c list_move(&msg_ctx->node, &ecryptfs_msg_ctx_alloc_list); node 86 fs/ecryptfs/messaging.c list_move(&(msg_ctx->node), &ecryptfs_msg_ctx_free_list); node 390 fs/ecryptfs/messaging.c INIT_LIST_HEAD(&ecryptfs_msg_ctx_arr[i].node); node 399 fs/ecryptfs/messaging.c list_add_tail(&ecryptfs_msg_ctx_arr[i].node, node 335 fs/ext2/balloc.c struct rb_node *node = &rsv->rsv_node; node 357 fs/ext2/balloc.c rb_link_node(node, parent, p); node 358 fs/ext2/balloc.c rb_insert_color(node, root); node 24 fs/ext4/block_validity.c struct rb_node node; node 58 fs/ext4/block_validity.c &system_blks->root, node) node 72 fs/ext4/block_validity.c struct rb_node **n = &system_blks->root.rb_node, *node; node 77 fs/ext4/block_validity.c entry = rb_entry(parent, struct ext4_system_zone, node); node 89 fs/ext4/block_validity.c node); node 101 fs/ext4/block_validity.c new_node = &new_entry->node; node 108 fs/ext4/block_validity.c node = rb_prev(new_node); node 109 fs/ext4/block_validity.c if (node) { node 110 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); node 114 fs/ext4/block_validity.c rb_erase(node, &system_blks->root); node 120 fs/ext4/block_validity.c node = rb_next(new_node); node 121 fs/ext4/block_validity.c if (node) { node 122 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); node 125 fs/ext4/block_validity.c rb_erase(node, &system_blks->root); node 134 fs/ext4/block_validity.c struct rb_node *node; node 139 fs/ext4/block_validity.c node = rb_first(&sbi->system_blks->root); node 140 fs/ext4/block_validity.c while (node) { node 141 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); node 145 fs/ext4/block_validity.c node = rb_next(node); node 175 fs/ext4/block_validity.c entry = rb_entry(n, struct ext4_system_zone, node); node 181 fs/ext4/extents_status.c struct rb_node *node; node 185 fs/ext4/extents_status.c node = rb_first(&tree->root); node 186 fs/ext4/extents_status.c while (node) { node 188 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 192 fs/ext4/extents_status.c node = rb_next(node); node 213 fs/ext4/extents_status.c struct rb_node *node = root->rb_node; node 216 fs/ext4/extents_status.c while (node) { node 217 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 219 fs/ext4/extents_status.c node = node->rb_left; node 221 fs/ext4/extents_status.c node = node->rb_right; node 230 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 231 fs/ext4/extents_status.c return node ? rb_entry(node, struct extent_status, rb_node) : node 263 fs/ext4/extents_status.c struct rb_node *node; node 286 fs/ext4/extents_status.c while ((node = rb_next(&es1->rb_node)) != NULL) { node 287 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); node 533 fs/ext4/extents_status.c struct rb_node *node; node 535 fs/ext4/extents_status.c node = rb_prev(&es->rb_node); node 536 fs/ext4/extents_status.c if (!node) node 539 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); node 557 fs/ext4/extents_status.c struct rb_node *node; node 559 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 560 fs/ext4/extents_status.c if (!node) node 563 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); node 568 fs/ext4/extents_status.c rb_erase(node, &tree->root); node 908 fs/ext4/extents_status.c struct rb_node *node; node 929 fs/ext4/extents_status.c node = tree->root.rb_node; node 930 fs/ext4/extents_status.c while (node) { node 931 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, rb_node); node 933 fs/ext4/extents_status.c node = node->rb_left; node 935 fs/ext4/extents_status.c node = node->rb_right; node 953 fs/ext4/extents_status.c node = rb_next(&es1->rb_node); node 954 fs/ext4/extents_status.c if (node) { node 955 fs/ext4/extents_status.c es1 = rb_entry(node, struct extent_status, node 996 fs/ext4/extents_status.c struct rb_node *node; node 1011 fs/ext4/extents_status.c node = rb_prev(&es->rb_node); node 1012 fs/ext4/extents_status.c rc->left_es = node ? rb_entry(node, node 1119 fs/ext4/extents_status.c struct rb_node *node = root->rb_node; node 1122 fs/ext4/extents_status.c while (node) { node 1123 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); node 1125 fs/ext4/extents_status.c node = node->rb_left; node 1127 fs/ext4/extents_status.c node = node->rb_right; node 1134 fs/ext4/extents_status.c node = rb_next(&pr->rb_node); node 1135 fs/ext4/extents_status.c return node ? rb_entry(node, struct pending_reservation, node 1164 fs/ext4/extents_status.c struct rb_node *node; node 1195 fs/ext4/extents_status.c node = rb_prev(&es->rb_node); node 1196 fs/ext4/extents_status.c if (!node) node 1198 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 1204 fs/ext4/extents_status.c node = rb_next(&right_es->rb_node); node 1205 fs/ext4/extents_status.c es = node ? rb_entry(node, struct extent_status, node 1215 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 1216 fs/ext4/extents_status.c if (!node) node 1218 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, node 1258 fs/ext4/extents_status.c node = rb_next(&pr->rb_node); node 1261 fs/ext4/extents_status.c if (!node) node 1263 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, node 1289 fs/ext4/extents_status.c struct rb_node *node; node 1364 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 1365 fs/ext4/extents_status.c if (node) node 1366 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 1374 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 1377 fs/ext4/extents_status.c if (!node) { node 1381 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 1677 fs/ext4/extents_status.c struct rb_node *node; node 1690 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 1706 fs/ext4/extents_status.c if (!node) node 1708 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 1750 fs/ext4/extents_status.c struct rb_node *node; node 1755 fs/ext4/extents_status.c node = rb_first(&tree->root); node 1756 fs/ext4/extents_status.c while (node) { node 1757 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 1758 fs/ext4/extents_status.c node = rb_next(node); node 1772 fs/ext4/extents_status.c struct rb_node *node; node 1777 fs/ext4/extents_status.c node = rb_first(&tree->root); node 1778 fs/ext4/extents_status.c while (node) { node 1779 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); node 1781 fs/ext4/extents_status.c node = rb_next(node); node 1822 fs/ext4/extents_status.c struct rb_node *node; node 1826 fs/ext4/extents_status.c node = (&tree->root)->rb_node; node 1828 fs/ext4/extents_status.c while (node) { node 1829 fs/ext4/extents_status.c pr = rb_entry(node, struct pending_reservation, rb_node); node 1831 fs/ext4/extents_status.c node = node->rb_left; node 1833 fs/ext4/extents_status.c node = node->rb_right; node 2026 fs/ext4/extents_status.c struct rb_node *node; node 2054 fs/ext4/extents_status.c node = rb_next(&es->rb_node); node 2055 fs/ext4/extents_status.c if (!node) node 2057 fs/ext4/extents_status.c es = rb_entry(node, struct extent_status, rb_node); node 4656 fs/ext4/mballoc.c struct rb_node **n = &db->bb_free_root.rb_node, *node; node 4695 fs/ext4/mballoc.c node = rb_prev(new_node); node 4696 fs/ext4/mballoc.c if (node) { node 4697 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); node 4702 fs/ext4/mballoc.c node = rb_next(new_node); node 4703 fs/ext4/mballoc.c if (node) { node 4704 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); node 33 fs/f2fs/extent_cache.c struct rb_node *node = root->rb_root.rb_node; node 36 fs/f2fs/extent_cache.c while (node) { node 37 fs/f2fs/extent_cache.c re = rb_entry(node, struct rb_entry, rb_node); node 40 fs/f2fs/extent_cache.c node = node->rb_left; node 42 fs/f2fs/extent_cache.c node = node->rb_right; node 302 fs/f2fs/extent_cache.c struct rb_node *node, *next; node 306 fs/f2fs/extent_cache.c node = rb_first_cached(&et->root); node 307 fs/f2fs/extent_cache.c while (node) { node 308 fs/f2fs/extent_cache.c next = rb_next(node); node 309 fs/f2fs/extent_cache.c en = rb_entry(node, struct extent_node, rb_node); node 311 fs/f2fs/extent_cache.c node = next; node 576 fs/f2fs/extent_cache.c struct rb_node *node = rb_next(&en->rb_node); node 578 fs/f2fs/extent_cache.c next_en = rb_entry_safe(node, struct extent_node, node 2280 fs/f2fs/f2fs.h static inline __le32 *blkaddr_in_node(struct f2fs_node *node) node 2282 fs/f2fs/f2fs.h return RAW_IS_INODE(node) ? node->i.i_addr : node->dn.addr; node 141 fs/f2fs/inode.c struct f2fs_node *node = F2FS_NODE(page); node 142 fs/f2fs/inode.c struct f2fs_inode *ri = &node->i; node 143 fs/f2fs/inode.c __le32 ino = node->footer.ino; node 1322 fs/f2fs/segment.c struct rb_node *node; node 1378 fs/f2fs/segment.c node = rb_next(&prev_dc->rb_node); node 1379 fs/f2fs/segment.c next_dc = rb_entry_safe(node, struct discard_cmd, rb_node); node 1428 fs/f2fs/segment.c struct rb_node *node; node 1445 fs/f2fs/segment.c node = rb_next(&dc->rb_node); node 1448 fs/f2fs/segment.c dc = rb_entry_safe(node, struct discard_cmd, rb_node); node 2774 fs/f2fs/segment.c struct rb_node *node; node 2800 fs/f2fs/segment.c node = rb_next(&dc->rb_node); node 2803 fs/f2fs/segment.c dc = rb_entry_safe(node, struct discard_cmd, rb_node); node 304 fs/file_table.c struct llist_node *node = llist_del_all(&delayed_fput_list); node 307 fs/file_table.c llist_for_each_entry_safe(f, t, node, f_u.fu_llist) node 1234 fs/fs-writeback.c struct list_head *pos, *node; node 1271 fs/fs-writeback.c list_for_each_prev_safe(pos, node, &tmp) { node 18 fs/hfs/bnode.c void hfs_bnode_read(struct hfs_bnode *node, void *buf, node 23 fs/hfs/bnode.c off += node->page_offset; node 24 fs/hfs/bnode.c page = node->page[0]; node 30 fs/hfs/bnode.c u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) node 34 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 2); node 38 fs/hfs/bnode.c u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) node 42 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 1); node 46 fs/hfs/bnode.c void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) node 51 fs/hfs/bnode.c tree = node->tree; node 52 fs/hfs/bnode.c if (node->type == HFS_NODE_LEAF || node 54 fs/hfs/bnode.c key_len = hfs_bnode_read_u8(node, off) + 1; node 58 fs/hfs/bnode.c hfs_bnode_read(node, key, off, key_len); node 61 fs/hfs/bnode.c void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len) node 65 fs/hfs/bnode.c off += node->page_offset; node 66 fs/hfs/bnode.c page = node->page[0]; node 73 fs/hfs/bnode.c void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data) node 77 fs/hfs/bnode.c hfs_bnode_write(node, &v, off, 2); node 80 fs/hfs/bnode.c void hfs_bnode_write_u8(struct hfs_bnode *node, int off, u8 data) node 83 fs/hfs/bnode.c hfs_bnode_write(node, &data, off, 1); node 86 fs/hfs/bnode.c void hfs_bnode_clear(struct hfs_bnode *node, int off, int len) node 90 fs/hfs/bnode.c off += node->page_offset; node 91 fs/hfs/bnode.c page = node->page[0]; node 117 fs/hfs/bnode.c void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len) node 125 fs/hfs/bnode.c src += node->page_offset; node 126 fs/hfs/bnode.c dst += node->page_offset; node 127 fs/hfs/bnode.c page = node->page[0]; node 134 fs/hfs/bnode.c void hfs_bnode_dump(struct hfs_bnode *node) node 140 fs/hfs/bnode.c hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this); node 141 fs/hfs/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc)); node 146 fs/hfs/bnode.c off = node->tree->node_size - 2; node 148 fs/hfs/bnode.c key_off = hfs_bnode_read_u16(node, off); node 150 fs/hfs/bnode.c if (i && node->type == HFS_NODE_INDEX) { node 153 fs/hfs/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS) node 154 fs/hfs/bnode.c tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1; node 156 fs/hfs/bnode.c tmp = node->tree->max_key_len + 1; node 158 fs/hfs/bnode.c tmp, hfs_bnode_read_u8(node, key_off)); node 159 fs/hfs/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4); node 161 fs/hfs/bnode.c } else if (i && node->type == HFS_NODE_LEAF) { node 164 fs/hfs/bnode.c tmp = hfs_bnode_read_u8(node, key_off); node 171 fs/hfs/bnode.c void hfs_bnode_unlink(struct hfs_bnode *node) node 177 fs/hfs/bnode.c tree = node->tree; node 178 fs/hfs/bnode.c if (node->prev) { node 179 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->prev); node 182 fs/hfs/bnode.c tmp->next = node->next; node 186 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF) node 187 fs/hfs/bnode.c tree->leaf_head = node->next; node 189 fs/hfs/bnode.c if (node->next) { node 190 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->next); node 193 fs/hfs/bnode.c tmp->prev = node->prev; node 197 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF) node 198 fs/hfs/bnode.c tree->leaf_tail = node->prev; node 201 fs/hfs/bnode.c if (!node->prev && !node->next) { node 204 fs/hfs/bnode.c if (!node->parent) { node 208 fs/hfs/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags); node 220 fs/hfs/bnode.c struct hfs_bnode *node; node 227 fs/hfs/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)]; node 228 fs/hfs/bnode.c node; node = node->next_hash) { node 229 fs/hfs/bnode.c if (node->this == cnid) { node 230 fs/hfs/bnode.c return node; node 238 fs/hfs/bnode.c struct hfs_bnode *node, *node2; node 251 fs/hfs/bnode.c node = kzalloc(size, GFP_KERNEL); node 252 fs/hfs/bnode.c if (!node) node 254 fs/hfs/bnode.c node->tree = tree; node 255 fs/hfs/bnode.c node->this = cnid; node 256 fs/hfs/bnode.c set_bit(HFS_BNODE_NEW, &node->flags); node 257 fs/hfs/bnode.c atomic_set(&node->refcnt, 1); node 259 fs/hfs/bnode.c node->tree->cnid, node->this); node 260 fs/hfs/bnode.c init_waitqueue_head(&node->lock_wq); node 265 fs/hfs/bnode.c node->next_hash = tree->node_hash[hash]; node 266 fs/hfs/bnode.c tree->node_hash[hash] = node; node 270 fs/hfs/bnode.c kfree(node); node 279 fs/hfs/bnode.c node->page_offset = off & ~PAGE_MASK; node 288 fs/hfs/bnode.c node->page[i] = page; node 291 fs/hfs/bnode.c return node; node 293 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 294 fs/hfs/bnode.c return node; node 297 fs/hfs/bnode.c void hfs_bnode_unhash(struct hfs_bnode *node) node 302 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 303 fs/hfs/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)]; node 304 fs/hfs/bnode.c *p && *p != node; p = &(*p)->next_hash) node 307 fs/hfs/bnode.c *p = node->next_hash; node 308 fs/hfs/bnode.c node->tree->node_hash_cnt--; node 314 fs/hfs/bnode.c struct hfs_bnode *node; node 320 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num); node 321 fs/hfs/bnode.c if (node) { node 322 fs/hfs/bnode.c hfs_bnode_get(node); node 324 fs/hfs/bnode.c wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags)); node 325 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 327 fs/hfs/bnode.c return node; node 330 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num); node 331 fs/hfs/bnode.c if (!node) node 333 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 335 fs/hfs/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags)) node 336 fs/hfs/bnode.c return node; node 338 fs/hfs/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); node 339 fs/hfs/bnode.c node->prev = be32_to_cpu(desc->prev); node 340 fs/hfs/bnode.c node->next = be32_to_cpu(desc->next); node 341 fs/hfs/bnode.c node->num_recs = be16_to_cpu(desc->num_recs); node 342 fs/hfs/bnode.c node->type = desc->type; node 343 fs/hfs/bnode.c node->height = desc->height; node 344 fs/hfs/bnode.c kunmap(node->page[0]); node 346 fs/hfs/bnode.c switch (node->type) { node 349 fs/hfs/bnode.c if (node->height != 0) node 353 fs/hfs/bnode.c if (node->height != 1) node 357 fs/hfs/bnode.c if (node->height <= 1 || node->height > tree->depth) node 365 fs/hfs/bnode.c off = hfs_bnode_read_u16(node, rec_off); node 368 fs/hfs/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) { node 370 fs/hfs/bnode.c next_off = hfs_bnode_read_u16(node, rec_off); node 376 fs/hfs/bnode.c if (node->type != HFS_NODE_INDEX && node 377 fs/hfs/bnode.c node->type != HFS_NODE_LEAF) node 379 fs/hfs/bnode.c key_size = hfs_bnode_read_u8(node, off) + 1; node 383 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 384 fs/hfs/bnode.c wake_up(&node->lock_wq); node 385 fs/hfs/bnode.c return node; node 388 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 389 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 390 fs/hfs/bnode.c wake_up(&node->lock_wq); node 391 fs/hfs/bnode.c hfs_bnode_put(node); node 395 fs/hfs/bnode.c void hfs_bnode_free(struct hfs_bnode *node) node 399 fs/hfs/bnode.c for (i = 0; i < node->tree->pages_per_bnode; i++) node 400 fs/hfs/bnode.c if (node->page[i]) node 401 fs/hfs/bnode.c put_page(node->page[i]); node 402 fs/hfs/bnode.c kfree(node); node 407 fs/hfs/bnode.c struct hfs_bnode *node; node 412 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num); node 414 fs/hfs/bnode.c if (node) { node 417 fs/hfs/bnode.c return node; node 419 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num); node 420 fs/hfs/bnode.c if (!node) node 422 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) { node 423 fs/hfs/bnode.c hfs_bnode_put(node); node 427 fs/hfs/bnode.c pagep = node->page; node 428 fs/hfs/bnode.c memset(kmap(*pagep) + node->page_offset, 0, node 437 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 438 fs/hfs/bnode.c wake_up(&node->lock_wq); node 440 fs/hfs/bnode.c return node; node 443 fs/hfs/bnode.c void hfs_bnode_get(struct hfs_bnode *node) node 445 fs/hfs/bnode.c if (node) { node 446 fs/hfs/bnode.c atomic_inc(&node->refcnt); node 448 fs/hfs/bnode.c node->tree->cnid, node->this, node 449 fs/hfs/bnode.c atomic_read(&node->refcnt)); node 454 fs/hfs/bnode.c void hfs_bnode_put(struct hfs_bnode *node) node 456 fs/hfs/bnode.c if (node) { node 457 fs/hfs/bnode.c struct hfs_btree *tree = node->tree; node 461 fs/hfs/bnode.c node->tree->cnid, node->this, node 462 fs/hfs/bnode.c atomic_read(&node->refcnt)); node 463 fs/hfs/bnode.c BUG_ON(!atomic_read(&node->refcnt)); node 464 fs/hfs/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) node 467 fs/hfs/bnode.c if (!node->page[i]) node 469 fs/hfs/bnode.c mark_page_accessed(node->page[i]); node 472 fs/hfs/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) { node 473 fs/hfs/bnode.c hfs_bnode_unhash(node); node 475 fs/hfs/bnode.c hfs_bmap_free(node); node 476 fs/hfs/bnode.c hfs_bnode_free(node); node 19 fs/hfs/brec.c u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) node 24 fs/hfs/brec.c dataoff = node->tree->node_size - (rec + 2) * 2; node 25 fs/hfs/brec.c hfs_bnode_read(node, retval, dataoff, 4); node 31 fs/hfs/brec.c u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) node 35 fs/hfs/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) node 38 fs/hfs/brec.c if ((node->type == HFS_NODE_INDEX) && node 39 fs/hfs/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) { node 40 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) node 41 fs/hfs/brec.c retval = node->tree->max_key_len + 2; node 43 fs/hfs/brec.c retval = node->tree->max_key_len + 1; node 45 fs/hfs/brec.c recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2); node 48 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) { node 49 fs/hfs/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2; node 50 fs/hfs/brec.c if (retval > node->tree->max_key_len + 2) { node 55 fs/hfs/brec.c retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1; node 56 fs/hfs/brec.c if (retval > node->tree->max_key_len + 1) { node 68 fs/hfs/brec.c struct hfs_bnode *node, *new_node; node 78 fs/hfs/brec.c node = hfs_bnode_find(tree, tree->leaf_head); node 79 fs/hfs/brec.c if (IS_ERR(node)) node 80 fs/hfs/brec.c return PTR_ERR(node); node 81 fs/hfs/brec.c fd->bnode = node; node 91 fs/hfs/brec.c node = fd->bnode; node 92 fs/hfs/brec.c hfs_bnode_dump(node); node 94 fs/hfs/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2; node 95 fs/hfs/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off); node 107 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) { node 111 fs/hfs/brec.c node->num_recs++; node 113 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 114 fs/hfs/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size); node 122 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2); node 123 fs/hfs/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size); node 128 fs/hfs/brec.c hfs_bnode_move(node, data_off + size, data_off, node 132 fs/hfs/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len); node 133 fs/hfs/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); node 134 fs/hfs/brec.c hfs_bnode_dump(node); node 140 fs/hfs/brec.c if (!rec && new_node != node) { node 141 fs/hfs/brec.c hfs_bnode_read_key(node, fd->search_key, data_off + size); node 180 fs/hfs/brec.c struct hfs_bnode *node, *parent; node 184 fs/hfs/brec.c node = fd->bnode; node 187 fs/hfs/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2; node 189 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) { node 193 fs/hfs/brec.c hfs_bnode_dump(node); node 196 fs/hfs/brec.c if (!--node->num_recs) { node 197 fs/hfs/brec.c hfs_bnode_unlink(node); node 198 fs/hfs/brec.c if (!node->parent) node 200 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent); node 203 fs/hfs/brec.c hfs_bnode_put(node); node 204 fs/hfs/brec.c node = fd->bnode = parent; node 206 fs/hfs/brec.c __hfs_brec_find(node, fd); node 209 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 216 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, rec_off); node 217 fs/hfs/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size); node 222 fs/hfs/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size, node 225 fs/hfs/brec.c hfs_bnode_dump(node); node 234 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *next_node; node 240 fs/hfs/brec.c node = fd->bnode; node 244 fs/hfs/brec.c hfs_bnode_get(node); node 246 fs/hfs/brec.c node->this, new_node->this, node->next); node 247 fs/hfs/brec.c new_node->next = node->next; node 248 fs/hfs/brec.c new_node->prev = node->this; node 249 fs/hfs/brec.c new_node->parent = node->parent; node 250 fs/hfs/brec.c new_node->type = node->type; node 251 fs/hfs/brec.c new_node->height = node->height; node 253 fs/hfs/brec.c if (node->next) node 254 fs/hfs/brec.c next_node = hfs_bnode_find(tree, node->next); node 259 fs/hfs/brec.c hfs_bnode_put(node); node 264 fs/hfs/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14; node 268 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 272 fs/hfs/brec.c if (++num_recs < node->num_recs) node 275 fs/hfs/brec.c hfs_bnode_put(node); node 288 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 290 fs/hfs/brec.c hfs_bnode_put(node); node 297 fs/hfs/brec.c new_node->num_recs = node->num_recs - num_recs; node 298 fs/hfs/brec.c node->num_recs = num_recs; node 309 fs/hfs/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off); node 314 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start); node 326 fs/hfs/brec.c node->next = new_node->this; node 327 fs/hfs/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc)); node 328 fs/hfs/brec.c node_desc.next = cpu_to_be32(node->next); node 329 fs/hfs/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs); node 330 fs/hfs/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc)); node 339 fs/hfs/brec.c } else if (node->this == tree->leaf_tail) { node 345 fs/hfs/brec.c hfs_bnode_dump(node); node 347 fs/hfs/brec.c hfs_bnode_put(node); node 355 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *parent; node 361 fs/hfs/brec.c node = fd->bnode; node 363 fs/hfs/brec.c if (!node->parent) node 367 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent); node 378 fs/hfs/brec.c newkeylen = (hfs_bnode_read_u8(node, 14) | 1) + 1; node 417 fs/hfs/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen); node 422 fs/hfs/brec.c hfs_bnode_put(node); node 423 fs/hfs/brec.c node = parent; node 443 fs/hfs/brec.c if (new_node == node) node 446 fs/hfs/brec.c hfs_bnode_read_key(node, fd->search_key, 14); node 451 fs/hfs/brec.c if (!rec && node->parent) node 454 fs/hfs/brec.c fd->bnode = node; node 460 fs/hfs/brec.c struct hfs_bnode *node, *new_node; node 465 fs/hfs/brec.c node = NULL; node 467 fs/hfs/brec.c node = hfs_bnode_find(tree, tree->root); node 468 fs/hfs/brec.c if (IS_ERR(node)) node 469 fs/hfs/brec.c return PTR_ERR(node); node 473 fs/hfs/brec.c hfs_bnode_put(node); node 502 fs/hfs/brec.c if (node) { node 504 fs/hfs/brec.c node->parent = tree->root; node 505 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF || node 507 fs/hfs/brec.c key_size = hfs_bnode_read_u8(node, 14) + 1; node 510 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size); node 517 fs/hfs/brec.c cnid = cpu_to_be32(node->this); node 523 fs/hfs/brec.c hfs_bnode_put(node); node 139 fs/hfs/btree.c struct hfs_bnode *node; node 146 fs/hfs/btree.c while ((node = tree->node_hash[i])) { node 147 fs/hfs/btree.c tree->node_hash[i] = node->next_hash; node 148 fs/hfs/btree.c if (atomic_read(&node->refcnt)) node 150 fs/hfs/btree.c node->tree->cnid, node->this, node 151 fs/hfs/btree.c atomic_read(&node->refcnt)); node 152 fs/hfs/btree.c hfs_bnode_free(node); node 163 fs/hfs/btree.c struct hfs_bnode *node; node 166 fs/hfs/btree.c node = hfs_bnode_find(tree, 0); node 167 fs/hfs/btree.c if (IS_ERR(node)) node 171 fs/hfs/btree.c page = node->page[0]; node 185 fs/hfs/btree.c hfs_bnode_put(node); node 191 fs/hfs/btree.c struct hfs_bnode *node; node 195 fs/hfs/btree.c node = hfs_bnode_create(tree, idx); node 196 fs/hfs/btree.c if (IS_ERR(node)) node 197 fs/hfs/btree.c return node; node 206 fs/hfs/btree.c node->type = HFS_NODE_MAP; node 207 fs/hfs/btree.c node->num_recs = 1; node 208 fs/hfs/btree.c hfs_bnode_clear(node, 0, tree->node_size); node 215 fs/hfs/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc)); node 216 fs/hfs/btree.c hfs_bnode_write_u16(node, 14, 0x8000); node 217 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14); node 218 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); node 220 fs/hfs/btree.c return node; node 249 fs/hfs/btree.c struct hfs_bnode *node, *next_node; node 263 fs/hfs/btree.c node = hfs_bnode_find(tree, nidx); node 264 fs/hfs/btree.c if (IS_ERR(node)) node 265 fs/hfs/btree.c return node; node 266 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off16); node 269 fs/hfs/btree.c off += node->page_offset; node 270 fs/hfs/btree.c pagep = node->page + (off >> PAGE_SHIFT); node 287 fs/hfs/btree.c hfs_bnode_put(node); node 301 fs/hfs/btree.c nidx = node->next; node 304 fs/hfs/btree.c next_node = hfs_bmap_new_bmap(node, idx); node 307 fs/hfs/btree.c hfs_bnode_put(node); node 310 fs/hfs/btree.c node = next_node; node 312 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off16); node 314 fs/hfs/btree.c off += node->page_offset; node 315 fs/hfs/btree.c pagep = node->page + (off >> PAGE_SHIFT); node 321 fs/hfs/btree.c void hfs_bmap_free(struct hfs_bnode *node) node 329 fs/hfs/btree.c hfs_dbg(BNODE_MOD, "btree_free_node: %u\n", node->this); node 330 fs/hfs/btree.c tree = node->tree; node 331 fs/hfs/btree.c nidx = node->this; node 332 fs/hfs/btree.c node = hfs_bnode_find(tree, 0); node 333 fs/hfs/btree.c if (IS_ERR(node)) node 335 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off); node 340 fs/hfs/btree.c i = node->next; node 344 fs/hfs/btree.c node->this); node 345 fs/hfs/btree.c hfs_bnode_put(node); node 348 fs/hfs/btree.c hfs_bnode_put(node); node 349 fs/hfs/btree.c node = hfs_bnode_find(tree, i); node 350 fs/hfs/btree.c if (IS_ERR(node)) node 352 fs/hfs/btree.c if (node->type != HFS_NODE_MAP) { node 355 fs/hfs/btree.c node->this, node->type); node 356 fs/hfs/btree.c hfs_bnode_put(node); node 359 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off); node 361 fs/hfs/btree.c off += node->page_offset + nidx / 8; node 362 fs/hfs/btree.c page = node->page[off >> PAGE_SHIFT]; node 369 fs/hfs/btree.c node->this, node->type); node 371 fs/hfs/btree.c hfs_bnode_put(node); node 377 fs/hfs/btree.c hfs_bnode_put(node); node 87 fs/hfs/btree.h extern void hfs_bmap_free(struct hfs_bnode *node); node 77 fs/hfs/inode.c struct hfs_bnode *node; node 99 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx); node 100 fs/hfs/inode.c if (!node) node 102 fs/hfs/inode.c else if (atomic_read(&node->refcnt)) node 104 fs/hfs/inode.c if (res && node) { node 105 fs/hfs/inode.c hfs_bnode_unhash(node); node 106 fs/hfs/inode.c hfs_bnode_free(node); node 114 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx++); node 115 fs/hfs/inode.c if (!node) node 117 fs/hfs/inode.c if (atomic_read(&node->refcnt)) { node 121 fs/hfs/inode.c hfs_bnode_unhash(node); node 122 fs/hfs/inode.c hfs_bnode_free(node); node 22 fs/hfsplus/bnode.c void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len) node 27 fs/hfsplus/bnode.c off += node->page_offset; node 28 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_SHIFT); node 43 fs/hfsplus/bnode.c u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) node 47 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 2); node 51 fs/hfsplus/bnode.c u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) node 55 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 1); node 59 fs/hfsplus/bnode.c void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) node 64 fs/hfsplus/bnode.c tree = node->tree; node 65 fs/hfsplus/bnode.c if (node->type == HFS_NODE_LEAF || node 67 fs/hfsplus/bnode.c node->tree->cnid == HFSPLUS_ATTR_CNID) node 68 fs/hfsplus/bnode.c key_len = hfs_bnode_read_u16(node, off) + 2; node 72 fs/hfsplus/bnode.c hfs_bnode_read(node, key, off, key_len); node 75 fs/hfsplus/bnode.c void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len) node 80 fs/hfsplus/bnode.c off += node->page_offset; node 81 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_SHIFT); node 98 fs/hfsplus/bnode.c void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data) node 102 fs/hfsplus/bnode.c hfs_bnode_write(node, &v, off, 2); node 105 fs/hfsplus/bnode.c void hfs_bnode_clear(struct hfs_bnode *node, int off, int len) node 110 fs/hfsplus/bnode.c off += node->page_offset; node 111 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_SHIFT); node 185 fs/hfsplus/bnode.c void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len) node 193 fs/hfsplus/bnode.c src += node->page_offset; node 194 fs/hfsplus/bnode.c dst += node->page_offset; node 197 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_SHIFT); node 200 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_SHIFT); node 247 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_SHIFT); node 249 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_SHIFT); node 298 fs/hfsplus/bnode.c void hfs_bnode_dump(struct hfs_bnode *node) node 304 fs/hfsplus/bnode.c hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this); node 305 fs/hfsplus/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc)); node 310 fs/hfsplus/bnode.c off = node->tree->node_size - 2; node 312 fs/hfsplus/bnode.c key_off = hfs_bnode_read_u16(node, off); node 314 fs/hfsplus/bnode.c if (i && node->type == HFS_NODE_INDEX) { node 317 fs/hfsplus/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS || node 318 fs/hfsplus/bnode.c node->tree->cnid == HFSPLUS_ATTR_CNID) node 319 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off) + 2; node 321 fs/hfsplus/bnode.c tmp = node->tree->max_key_len + 2; node 323 fs/hfsplus/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4); node 325 fs/hfsplus/bnode.c } else if (i && node->type == HFS_NODE_LEAF) { node 328 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off); node 335 fs/hfsplus/bnode.c void hfs_bnode_unlink(struct hfs_bnode *node) node 341 fs/hfsplus/bnode.c tree = node->tree; node 342 fs/hfsplus/bnode.c if (node->prev) { node 343 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->prev); node 346 fs/hfsplus/bnode.c tmp->next = node->next; node 351 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF) node 352 fs/hfsplus/bnode.c tree->leaf_head = node->next; node 354 fs/hfsplus/bnode.c if (node->next) { node 355 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->next); node 358 fs/hfsplus/bnode.c tmp->prev = node->prev; node 363 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF) node 364 fs/hfsplus/bnode.c tree->leaf_tail = node->prev; node 367 fs/hfsplus/bnode.c if (!node->prev && !node->next) node 369 fs/hfsplus/bnode.c if (!node->parent) { node 373 fs/hfsplus/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags); node 385 fs/hfsplus/bnode.c struct hfs_bnode *node; node 393 fs/hfsplus/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)]; node 394 fs/hfsplus/bnode.c node; node = node->next_hash) node 395 fs/hfsplus/bnode.c if (node->this == cnid) node 396 fs/hfsplus/bnode.c return node; node 402 fs/hfsplus/bnode.c struct hfs_bnode *node, *node2; node 416 fs/hfsplus/bnode.c node = kzalloc(size, GFP_KERNEL); node 417 fs/hfsplus/bnode.c if (!node) node 419 fs/hfsplus/bnode.c node->tree = tree; node 420 fs/hfsplus/bnode.c node->this = cnid; node 421 fs/hfsplus/bnode.c set_bit(HFS_BNODE_NEW, &node->flags); node 422 fs/hfsplus/bnode.c atomic_set(&node->refcnt, 1); node 424 fs/hfsplus/bnode.c node->tree->cnid, node->this); node 425 fs/hfsplus/bnode.c init_waitqueue_head(&node->lock_wq); node 430 fs/hfsplus/bnode.c node->next_hash = tree->node_hash[hash]; node 431 fs/hfsplus/bnode.c tree->node_hash[hash] = node; node 435 fs/hfsplus/bnode.c kfree(node); node 445 fs/hfsplus/bnode.c node->page_offset = off & ~PAGE_MASK; node 454 fs/hfsplus/bnode.c node->page[i] = page; node 457 fs/hfsplus/bnode.c return node; node 459 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 460 fs/hfsplus/bnode.c return node; node 463 fs/hfsplus/bnode.c void hfs_bnode_unhash(struct hfs_bnode *node) node 468 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 469 fs/hfsplus/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)]; node 470 fs/hfsplus/bnode.c *p && *p != node; p = &(*p)->next_hash) node 473 fs/hfsplus/bnode.c *p = node->next_hash; node 474 fs/hfsplus/bnode.c node->tree->node_hash_cnt--; node 480 fs/hfsplus/bnode.c struct hfs_bnode *node; node 486 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num); node 487 fs/hfsplus/bnode.c if (node) { node 488 fs/hfsplus/bnode.c hfs_bnode_get(node); node 490 fs/hfsplus/bnode.c wait_event(node->lock_wq, node 491 fs/hfsplus/bnode.c !test_bit(HFS_BNODE_NEW, &node->flags)); node 492 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 494 fs/hfsplus/bnode.c return node; node 497 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num); node 498 fs/hfsplus/bnode.c if (!node) node 500 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 502 fs/hfsplus/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags)) node 503 fs/hfsplus/bnode.c return node; node 505 fs/hfsplus/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node 506 fs/hfsplus/bnode.c node->page_offset); node 507 fs/hfsplus/bnode.c node->prev = be32_to_cpu(desc->prev); node 508 fs/hfsplus/bnode.c node->next = be32_to_cpu(desc->next); node 509 fs/hfsplus/bnode.c node->num_recs = be16_to_cpu(desc->num_recs); node 510 fs/hfsplus/bnode.c node->type = desc->type; node 511 fs/hfsplus/bnode.c node->height = desc->height; node 512 fs/hfsplus/bnode.c kunmap(node->page[0]); node 514 fs/hfsplus/bnode.c switch (node->type) { node 517 fs/hfsplus/bnode.c if (node->height != 0) node 521 fs/hfsplus/bnode.c if (node->height != 1) node 525 fs/hfsplus/bnode.c if (node->height <= 1 || node->height > tree->depth) node 533 fs/hfsplus/bnode.c off = hfs_bnode_read_u16(node, rec_off); node 536 fs/hfsplus/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) { node 538 fs/hfsplus/bnode.c next_off = hfs_bnode_read_u16(node, rec_off); node 544 fs/hfsplus/bnode.c if (node->type != HFS_NODE_INDEX && node 545 fs/hfsplus/bnode.c node->type != HFS_NODE_LEAF) node 547 fs/hfsplus/bnode.c key_size = hfs_bnode_read_u16(node, off) + 2; node 551 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 552 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 553 fs/hfsplus/bnode.c return node; node 556 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 557 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 558 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 559 fs/hfsplus/bnode.c hfs_bnode_put(node); node 563 fs/hfsplus/bnode.c void hfs_bnode_free(struct hfs_bnode *node) node 567 fs/hfsplus/bnode.c for (i = 0; i < node->tree->pages_per_bnode; i++) node 568 fs/hfsplus/bnode.c if (node->page[i]) node 569 fs/hfsplus/bnode.c put_page(node->page[i]); node 570 fs/hfsplus/bnode.c kfree(node); node 575 fs/hfsplus/bnode.c struct hfs_bnode *node; node 580 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num); node 582 fs/hfsplus/bnode.c if (node) { node 585 fs/hfsplus/bnode.c return node; node 587 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num); node 588 fs/hfsplus/bnode.c if (!node) node 590 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) { node 591 fs/hfsplus/bnode.c hfs_bnode_put(node); node 595 fs/hfsplus/bnode.c pagep = node->page; node 596 fs/hfsplus/bnode.c memset(kmap(*pagep) + node->page_offset, 0, node 605 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 606 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 608 fs/hfsplus/bnode.c return node; node 611 fs/hfsplus/bnode.c void hfs_bnode_get(struct hfs_bnode *node) node 613 fs/hfsplus/bnode.c if (node) { node 614 fs/hfsplus/bnode.c atomic_inc(&node->refcnt); node 616 fs/hfsplus/bnode.c node->tree->cnid, node->this, node 617 fs/hfsplus/bnode.c atomic_read(&node->refcnt)); node 622 fs/hfsplus/bnode.c void hfs_bnode_put(struct hfs_bnode *node) node 624 fs/hfsplus/bnode.c if (node) { node 625 fs/hfsplus/bnode.c struct hfs_btree *tree = node->tree; node 629 fs/hfsplus/bnode.c node->tree->cnid, node->this, node 630 fs/hfsplus/bnode.c atomic_read(&node->refcnt)); node 631 fs/hfsplus/bnode.c BUG_ON(!atomic_read(&node->refcnt)); node 632 fs/hfsplus/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) node 635 fs/hfsplus/bnode.c if (!node->page[i]) node 637 fs/hfsplus/bnode.c mark_page_accessed(node->page[i]); node 640 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) { node 641 fs/hfsplus/bnode.c hfs_bnode_unhash(node); node 644 fs/hfsplus/bnode.c hfs_bnode_clear(node, 0, tree->node_size); node 645 fs/hfsplus/bnode.c hfs_bmap_free(node); node 646 fs/hfsplus/bnode.c hfs_bnode_free(node); node 20 fs/hfsplus/brec.c u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) node 25 fs/hfsplus/brec.c dataoff = node->tree->node_size - (rec + 2) * 2; node 26 fs/hfsplus/brec.c hfs_bnode_read(node, retval, dataoff, 4); node 32 fs/hfsplus/brec.c u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) node 36 fs/hfsplus/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) node 39 fs/hfsplus/brec.c if ((node->type == HFS_NODE_INDEX) && node 40 fs/hfsplus/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS) && node 41 fs/hfsplus/brec.c (node->tree->cnid != HFSPLUS_ATTR_CNID)) { node 42 fs/hfsplus/brec.c retval = node->tree->max_key_len + 2; node 44 fs/hfsplus/brec.c recoff = hfs_bnode_read_u16(node, node 45 fs/hfsplus/brec.c node->tree->node_size - (rec + 1) * 2); node 48 fs/hfsplus/brec.c if (recoff > node->tree->node_size - 2) { node 53 fs/hfsplus/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2; node 54 fs/hfsplus/brec.c if (retval > node->tree->max_key_len + 2) { node 66 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node; node 76 fs/hfsplus/brec.c node = hfs_bnode_find(tree, tree->leaf_head); node 77 fs/hfsplus/brec.c if (IS_ERR(node)) node 78 fs/hfsplus/brec.c return PTR_ERR(node); node 79 fs/hfsplus/brec.c fd->bnode = node; node 89 fs/hfsplus/brec.c node = fd->bnode; node 90 fs/hfsplus/brec.c hfs_bnode_dump(node); node 92 fs/hfsplus/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2; node 93 fs/hfsplus/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off); node 105 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) { node 109 fs/hfsplus/brec.c node->num_recs++; node 111 fs/hfsplus/brec.c hfs_bnode_write_u16(node, node 113 fs/hfsplus/brec.c node->num_recs); node 114 fs/hfsplus/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size); node 122 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2); node 123 fs/hfsplus/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size); node 128 fs/hfsplus/brec.c hfs_bnode_move(node, data_off + size, data_off, node 132 fs/hfsplus/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len); node 133 fs/hfsplus/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); node 134 fs/hfsplus/brec.c hfs_bnode_dump(node); node 140 fs/hfsplus/brec.c if (!rec && new_node != node) { node 141 fs/hfsplus/brec.c hfs_bnode_read_key(node, fd->search_key, data_off + size); node 182 fs/hfsplus/brec.c struct hfs_bnode *node, *parent; node 186 fs/hfsplus/brec.c node = fd->bnode; node 189 fs/hfsplus/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2; node 191 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) { node 195 fs/hfsplus/brec.c hfs_bnode_dump(node); node 198 fs/hfsplus/brec.c if (!--node->num_recs) { node 199 fs/hfsplus/brec.c hfs_bnode_unlink(node); node 200 fs/hfsplus/brec.c if (!node->parent) node 202 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent); node 205 fs/hfsplus/brec.c hfs_bnode_put(node); node 206 fs/hfsplus/brec.c node = fd->bnode = parent; node 208 fs/hfsplus/brec.c __hfs_brec_find(node, fd, hfs_find_rec_by_key); node 211 fs/hfsplus/brec.c hfs_bnode_write_u16(node, node 213 fs/hfsplus/brec.c node->num_recs); node 220 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, rec_off); node 221 fs/hfsplus/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size); node 226 fs/hfsplus/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size, node 229 fs/hfsplus/brec.c hfs_bnode_dump(node); node 238 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node, *next_node; node 244 fs/hfsplus/brec.c node = fd->bnode; node 248 fs/hfsplus/brec.c hfs_bnode_get(node); node 250 fs/hfsplus/brec.c node->this, new_node->this, node->next); node 251 fs/hfsplus/brec.c new_node->next = node->next; node 252 fs/hfsplus/brec.c new_node->prev = node->this; node 253 fs/hfsplus/brec.c new_node->parent = node->parent; node 254 fs/hfsplus/brec.c new_node->type = node->type; node 255 fs/hfsplus/brec.c new_node->height = node->height; node 257 fs/hfsplus/brec.c if (node->next) node 258 fs/hfsplus/brec.c next_node = hfs_bnode_find(tree, node->next); node 263 fs/hfsplus/brec.c hfs_bnode_put(node); node 268 fs/hfsplus/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14; node 272 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 276 fs/hfsplus/brec.c if (++num_recs < node->num_recs) node 279 fs/hfsplus/brec.c hfs_bnode_put(node); node 292 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 294 fs/hfsplus/brec.c hfs_bnode_put(node); node 301 fs/hfsplus/brec.c new_node->num_recs = node->num_recs - num_recs; node 302 fs/hfsplus/brec.c node->num_recs = num_recs; node 313 fs/hfsplus/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off); node 318 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start); node 330 fs/hfsplus/brec.c node->next = new_node->this; node 331 fs/hfsplus/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc)); node 332 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(node->next); node 333 fs/hfsplus/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs); node 334 fs/hfsplus/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc)); node 343 fs/hfsplus/brec.c } else if (node->this == tree->leaf_tail) { node 349 fs/hfsplus/brec.c hfs_bnode_dump(node); node 351 fs/hfsplus/brec.c hfs_bnode_put(node); node 359 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node, *parent; node 365 fs/hfsplus/brec.c node = fd->bnode; node 367 fs/hfsplus/brec.c if (!node->parent) node 371 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent); node 383 fs/hfsplus/brec.c newkeylen = hfs_bnode_read_u16(node, 14) + 2; node 423 fs/hfsplus/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen); node 426 fs/hfsplus/brec.c hfs_bnode_put(node); node 427 fs/hfsplus/brec.c node = parent; node 447 fs/hfsplus/brec.c if (new_node == node) node 450 fs/hfsplus/brec.c hfs_bnode_read_key(node, fd->search_key, 14); node 455 fs/hfsplus/brec.c if (!rec && node->parent) node 458 fs/hfsplus/brec.c fd->bnode = node; node 464 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node; node 469 fs/hfsplus/brec.c node = NULL; node 471 fs/hfsplus/brec.c node = hfs_bnode_find(tree, tree->root); node 472 fs/hfsplus/brec.c if (IS_ERR(node)) node 473 fs/hfsplus/brec.c return PTR_ERR(node); node 477 fs/hfsplus/brec.c hfs_bnode_put(node); node 506 fs/hfsplus/brec.c if (node) { node 508 fs/hfsplus/brec.c node->parent = tree->root; node 509 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF || node 512 fs/hfsplus/brec.c key_size = hfs_bnode_read_u16(node, 14) + 2; node 515 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size); node 522 fs/hfsplus/brec.c cnid = cpu_to_be32(node->this); node 528 fs/hfsplus/brec.c hfs_bnode_put(node); node 260 fs/hfsplus/btree.c struct hfs_bnode *node; node 267 fs/hfsplus/btree.c while ((node = tree->node_hash[i])) { node 268 fs/hfsplus/btree.c tree->node_hash[i] = node->next_hash; node 269 fs/hfsplus/btree.c if (atomic_read(&node->refcnt)) node 272 fs/hfsplus/btree.c node->tree->cnid, node->this, node 273 fs/hfsplus/btree.c atomic_read(&node->refcnt)); node 274 fs/hfsplus/btree.c hfs_bnode_free(node); node 285 fs/hfsplus/btree.c struct hfs_bnode *node; node 288 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0); node 289 fs/hfsplus/btree.c if (IS_ERR(node)) node 293 fs/hfsplus/btree.c page = node->page[0]; node 308 fs/hfsplus/btree.c hfs_bnode_put(node); node 315 fs/hfsplus/btree.c struct hfs_bnode *node; node 319 fs/hfsplus/btree.c node = hfs_bnode_create(tree, idx); node 320 fs/hfsplus/btree.c if (IS_ERR(node)) node 321 fs/hfsplus/btree.c return node; node 328 fs/hfsplus/btree.c node->type = HFS_NODE_MAP; node 329 fs/hfsplus/btree.c node->num_recs = 1; node 330 fs/hfsplus/btree.c hfs_bnode_clear(node, 0, tree->node_size); node 337 fs/hfsplus/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc)); node 338 fs/hfsplus/btree.c hfs_bnode_write_u16(node, 14, 0x8000); node 339 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14); node 340 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); node 342 fs/hfsplus/btree.c return node; node 375 fs/hfsplus/btree.c struct hfs_bnode *node, *next_node; node 389 fs/hfsplus/btree.c node = hfs_bnode_find(tree, nidx); node 390 fs/hfsplus/btree.c if (IS_ERR(node)) node 391 fs/hfsplus/btree.c return node; node 392 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off16); node 395 fs/hfsplus/btree.c off += node->page_offset; node 396 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_SHIFT); node 413 fs/hfsplus/btree.c hfs_bnode_put(node); node 428 fs/hfsplus/btree.c nidx = node->next; node 431 fs/hfsplus/btree.c next_node = hfs_bmap_new_bmap(node, idx); node 434 fs/hfsplus/btree.c hfs_bnode_put(node); node 437 fs/hfsplus/btree.c node = next_node; node 439 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off16); node 441 fs/hfsplus/btree.c off += node->page_offset; node 442 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_SHIFT); node 448 fs/hfsplus/btree.c void hfs_bmap_free(struct hfs_bnode *node) node 456 fs/hfsplus/btree.c hfs_dbg(BNODE_MOD, "btree_free_node: %u\n", node->this); node 457 fs/hfsplus/btree.c BUG_ON(!node->this); node 458 fs/hfsplus/btree.c tree = node->tree; node 459 fs/hfsplus/btree.c nidx = node->this; node 460 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0); node 461 fs/hfsplus/btree.c if (IS_ERR(node)) node 463 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off); node 468 fs/hfsplus/btree.c i = node->next; node 473 fs/hfsplus/btree.c node->this); node 474 fs/hfsplus/btree.c hfs_bnode_put(node); node 477 fs/hfsplus/btree.c hfs_bnode_put(node); node 478 fs/hfsplus/btree.c node = hfs_bnode_find(tree, i); node 479 fs/hfsplus/btree.c if (IS_ERR(node)) node 481 fs/hfsplus/btree.c if (node->type != HFS_NODE_MAP) { node 485 fs/hfsplus/btree.c node->this, node->type); node 486 fs/hfsplus/btree.c hfs_bnode_put(node); node 489 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off); node 491 fs/hfsplus/btree.c off += node->page_offset + nidx / 8; node 492 fs/hfsplus/btree.c page = node->page[off >> PAGE_SHIFT]; node 500 fs/hfsplus/btree.c node->this, node->type); node 502 fs/hfsplus/btree.c hfs_bnode_put(node); node 508 fs/hfsplus/btree.c hfs_bnode_put(node); node 401 fs/hfsplus/hfsplus_fs.h void hfs_bmap_free(struct hfs_bnode *node); node 404 fs/hfsplus/hfsplus_fs.h void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len); node 405 fs/hfsplus/hfsplus_fs.h u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off); node 406 fs/hfsplus/hfsplus_fs.h u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off); node 407 fs/hfsplus/hfsplus_fs.h void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off); node 408 fs/hfsplus/hfsplus_fs.h void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len); node 409 fs/hfsplus/hfsplus_fs.h void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data); node 410 fs/hfsplus/hfsplus_fs.h void hfs_bnode_clear(struct hfs_bnode *node, int off, int len); node 413 fs/hfsplus/hfsplus_fs.h void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len); node 414 fs/hfsplus/hfsplus_fs.h void hfs_bnode_dump(struct hfs_bnode *node); node 415 fs/hfsplus/hfsplus_fs.h void hfs_bnode_unlink(struct hfs_bnode *node); node 417 fs/hfsplus/hfsplus_fs.h void hfs_bnode_unhash(struct hfs_bnode *node); node 419 fs/hfsplus/hfsplus_fs.h void hfs_bnode_free(struct hfs_bnode *node); node 421 fs/hfsplus/hfsplus_fs.h void hfs_bnode_get(struct hfs_bnode *node); node 422 fs/hfsplus/hfsplus_fs.h void hfs_bnode_put(struct hfs_bnode *node); node 426 fs/hfsplus/hfsplus_fs.h u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off); node 427 fs/hfsplus/hfsplus_fs.h u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec); node 71 fs/hfsplus/inode.c struct hfs_bnode *node; node 95 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx); node 96 fs/hfsplus/inode.c if (!node) node 98 fs/hfsplus/inode.c else if (atomic_read(&node->refcnt)) node 100 fs/hfsplus/inode.c if (res && node) { node 101 fs/hfsplus/inode.c hfs_bnode_unhash(node); node 102 fs/hfsplus/inode.c hfs_bnode_free(node); node 111 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx++); node 112 fs/hfsplus/inode.c if (!node) node 114 fs/hfsplus/inode.c if (atomic_read(&node->refcnt)) { node 118 fs/hfsplus/inode.c hfs_bnode_unhash(node); node 119 fs/hfsplus/inode.c hfs_bnode_free(node); node 61 fs/hpfs/anode.c secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsigned fsecno) node 73 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1; node 76 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return -1; node 79 fs/hpfs/anode.c a = node; node 101 fs/hpfs/anode.c fnod?'f':'a', node); node 113 fs/hpfs/anode.c hpfs_error(s, "empty file %08x, trying to add sector %08x", node, fsecno); node 117 fs/hpfs/anode.c se = !fnod ? node : (node + 16384) & ~16383; node 125 fs/hpfs/anode.c up = a != node ? le32_to_cpu(anode->up) : -1; node 131 fs/hpfs/anode.c if (a == node && fnod) { node 132 fs/hpfs/anode.c anode->up = cpu_to_le32(node); node 163 fs/hpfs/anode.c if ((a == node && fnod) || na == -1) return se; node 169 fs/hpfs/anode.c if (up != node || !fnod) { node 188 fs/hpfs/anode.c if (up == node && fnod) node 197 fs/hpfs/anode.c up = up != node ? le32_to_cpu(anode->up) : -1; node 221 fs/hpfs/anode.c anode->up = cpu_to_le32(node); node 228 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) { node 234 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) { node 240 fs/hpfs/anode.c ranode->up = cpu_to_le32(node); node 405 fs/hpfs/anode.c anode_secno node = f; node 432 fs/hpfs/anode.c hpfs_error(s, "internal btree %08x doesn't end with -1", node); node 445 fs/hpfs/anode.c node = le32_to_cpu(btree->u.internal[i].down); node 448 fs/hpfs/anode.c if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree")) node 450 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return; node 81 fs/jffs2/debug.c struct jffs2_full_dnode *fn = frag->node; node 99 fs/jffs2/debug.c && frag_prev(frag)->size < PAGE_SIZE && frag_prev(frag)->node) { node 106 fs/jffs2/debug.c && frag_next(frag)->size < PAGE_SIZE && frag_next(frag)->node) { node 711 fs/jffs2/debug.c if (this->node) node 713 fs/jffs2/debug.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw), node 714 fs/jffs2/debug.c ref_flags(this->node->raw), this, frag_left(this), frag_right(this), node 775 fs/jffs2/debug.c union jffs2_node_union node; node 783 fs/jffs2/debug.c ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node); node 790 fs/jffs2/debug.c printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic)); node 791 fs/jffs2/debug.c printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype)); node 792 fs/jffs2/debug.c printk(JFFS2_DBG "totlen:\t%#08x\n", je32_to_cpu(node.u.totlen)); node 793 fs/jffs2/debug.c printk(JFFS2_DBG "hdr_crc:\t%#08x\n", je32_to_cpu(node.u.hdr_crc)); node 795 fs/jffs2/debug.c crc = crc32(0, &node.u, sizeof(node.u) - 4); node 796 fs/jffs2/debug.c if (crc != je32_to_cpu(node.u.hdr_crc)) { node 801 fs/jffs2/debug.c if (je16_to_cpu(node.u.magic) != JFFS2_MAGIC_BITMASK && node 802 fs/jffs2/debug.c je16_to_cpu(node.u.magic) != JFFS2_OLD_MAGIC_BITMASK) node 805 fs/jffs2/debug.c je16_to_cpu(node.u.magic), JFFS2_MAGIC_BITMASK); node 809 fs/jffs2/debug.c switch(je16_to_cpu(node.u.nodetype)) { node 814 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.i.ino)); node 815 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.i.version)); node 816 fs/jffs2/debug.c printk(JFFS2_DBG "mode:\t%#08x\n", node.i.mode.m); node 817 fs/jffs2/debug.c printk(JFFS2_DBG "uid:\t%#04x\n", je16_to_cpu(node.i.uid)); node 818 fs/jffs2/debug.c printk(JFFS2_DBG "gid:\t%#04x\n", je16_to_cpu(node.i.gid)); node 819 fs/jffs2/debug.c printk(JFFS2_DBG "isize:\t%#08x\n", je32_to_cpu(node.i.isize)); node 820 fs/jffs2/debug.c printk(JFFS2_DBG "atime:\t%#08x\n", je32_to_cpu(node.i.atime)); node 821 fs/jffs2/debug.c printk(JFFS2_DBG "mtime:\t%#08x\n", je32_to_cpu(node.i.mtime)); node 822 fs/jffs2/debug.c printk(JFFS2_DBG "ctime:\t%#08x\n", je32_to_cpu(node.i.ctime)); node 823 fs/jffs2/debug.c printk(JFFS2_DBG "offset:\t%#08x\n", je32_to_cpu(node.i.offset)); node 824 fs/jffs2/debug.c printk(JFFS2_DBG "csize:\t%#08x\n", je32_to_cpu(node.i.csize)); node 825 fs/jffs2/debug.c printk(JFFS2_DBG "dsize:\t%#08x\n", je32_to_cpu(node.i.dsize)); node 826 fs/jffs2/debug.c printk(JFFS2_DBG "compr:\t%#02x\n", node.i.compr); node 827 fs/jffs2/debug.c printk(JFFS2_DBG "usercompr:\t%#02x\n", node.i.usercompr); node 828 fs/jffs2/debug.c printk(JFFS2_DBG "flags:\t%#04x\n", je16_to_cpu(node.i.flags)); node 829 fs/jffs2/debug.c printk(JFFS2_DBG "data_crc:\t%#08x\n", je32_to_cpu(node.i.data_crc)); node 830 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.i.node_crc)); node 832 fs/jffs2/debug.c crc = crc32(0, &node.i, sizeof(node.i) - 8); node 833 fs/jffs2/debug.c if (crc != je32_to_cpu(node.i.node_crc)) { node 842 fs/jffs2/debug.c printk(JFFS2_DBG "pino:\t%#08x\n", je32_to_cpu(node.d.pino)); node 843 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.d.version)); node 844 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.d.ino)); node 845 fs/jffs2/debug.c printk(JFFS2_DBG "mctime:\t%#08x\n", je32_to_cpu(node.d.mctime)); node 846 fs/jffs2/debug.c printk(JFFS2_DBG "nsize:\t%#02x\n", node.d.nsize); node 847 fs/jffs2/debug.c printk(JFFS2_DBG "type:\t%#02x\n", node.d.type); node 848 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.d.node_crc)); node 849 fs/jffs2/debug.c printk(JFFS2_DBG "name_crc:\t%#08x\n", je32_to_cpu(node.d.name_crc)); node 851 fs/jffs2/debug.c node.d.name[node.d.nsize] = '\0'; node 852 fs/jffs2/debug.c printk(JFFS2_DBG "name:\t\"%s\"\n", node.d.name); node 854 fs/jffs2/debug.c crc = crc32(0, &node.d, sizeof(node.d) - 8); node 855 fs/jffs2/debug.c if (crc != je32_to_cpu(node.d.node_crc)) { node 535 fs/jffs2/gc.c if (frag->node && frag->node->raw == raw) { node 536 fs/jffs2/gc.c fn = frag->node; node 540 fs/jffs2/gc.c if (nrfrags == frag->node->frags) node 549 fs/jffs2/gc.c frag->node->raw = f->inocache->nodes; node 595 fs/jffs2/gc.c union jffs2_node_union *node; node 624 fs/jffs2/gc.c node = kmalloc(rawlen, GFP_KERNEL); node 625 fs/jffs2/gc.c if (!node) node 628 fs/jffs2/gc.c ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node); node 634 fs/jffs2/gc.c crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4); node 635 fs/jffs2/gc.c if (je32_to_cpu(node->u.hdr_crc) != crc) { node 637 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->u.hdr_crc), crc); node 641 fs/jffs2/gc.c switch(je16_to_cpu(node->u.nodetype)) { node 643 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->i)-8); node 644 fs/jffs2/gc.c if (je32_to_cpu(node->i.node_crc) != crc) { node 646 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->i.node_crc), node 651 fs/jffs2/gc.c if (je32_to_cpu(node->i.dsize)) { node 652 fs/jffs2/gc.c crc = crc32(0, node->i.data, je32_to_cpu(node->i.csize)); node 653 fs/jffs2/gc.c if (je32_to_cpu(node->i.data_crc) != crc) { node 656 fs/jffs2/gc.c je32_to_cpu(node->i.data_crc), crc); node 663 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->d)-8); node 664 fs/jffs2/gc.c if (je32_to_cpu(node->d.node_crc) != crc) { node 667 fs/jffs2/gc.c je32_to_cpu(node->d.node_crc), crc); node 671 fs/jffs2/gc.c if (strnlen(node->d.name, node->d.nsize) != node->d.nsize) { node 677 fs/jffs2/gc.c if (node->d.nsize) { node 678 fs/jffs2/gc.c crc = crc32(0, node->d.name, node->d.nsize); node 679 fs/jffs2/gc.c if (je32_to_cpu(node->d.name_crc) != crc) { node 682 fs/jffs2/gc.c je32_to_cpu(node->d.name_crc), crc); node 691 fs/jffs2/gc.c ref_offset(raw), je16_to_cpu(node->u.nodetype)); node 700 fs/jffs2/gc.c ret = jffs2_flash_write(c, phys_ofs, rawlen, &retlen, (char *)node); node 751 fs/jffs2/gc.c kfree(node); node 1143 fs/jffs2/gc.c if (frag->node == fn) { node 1144 fs/jffs2/gc.c frag->node = new_fn; node 1217 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) { node 1227 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw; node 1273 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) { node 1283 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw; node 286 fs/jffs2/malloc.c xd->node = (void *)xd; node 306 fs/jffs2/malloc.c ref->node = (void *)ref; node 93 fs/jffs2/nodelist.c if (frag->node && (frag->ofs & (PAGE_SIZE - 1)) == 0) { node 96 fs/jffs2/nodelist.c frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE; node 104 fs/jffs2/nodelist.c if (this->node) { node 105 fs/jffs2/nodelist.c this->node->frags--; node 106 fs/jffs2/nodelist.c if (!this->node->frags) { node 109 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size); node 110 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, this->node->raw); node 111 fs/jffs2/nodelist.c jffs2_free_full_dnode(this->node); node 114 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); node 115 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 157 fs/jffs2/nodelist.c newfrag->node = fn; node 173 fs/jffs2/nodelist.c if (lastend < newfrag->node->ofs) { node 177 fs/jffs2/nodelist.c holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend); node 221 fs/jffs2/nodelist.c this = jffs2_lookup_node_frag(root, newfrag->node->ofs); node 225 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this); node 241 fs/jffs2/nodelist.c if (this->node) node 242 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 243 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 249 fs/jffs2/nodelist.c if (this->node) node 252 fs/jffs2/nodelist.c ref_offset(this->node->raw), ref_flags(this->node->raw)); node 265 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 266 fs/jffs2/nodelist.c if (this->node) node 267 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 273 fs/jffs2/nodelist.c if (this->node) node 275 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw)); node 281 fs/jffs2/nodelist.c newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size, node 285 fs/jffs2/nodelist.c if (this->node) node 286 fs/jffs2/nodelist.c this->node->frags++; node 352 fs/jffs2/nodelist.c if (this->node) node 353 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 354 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 374 fs/jffs2/nodelist.c newfrag->node->frags = 1; node 390 fs/jffs2/nodelist.c if (prev->node) node 391 fs/jffs2/nodelist.c mark_ref_normal(prev->node->raw); node 399 fs/jffs2/nodelist.c if (next->node) node 400 fs/jffs2/nodelist.c mark_ref_normal(next->node->raw); node 571 fs/jffs2/nodelist.c if (frag->node && !(--frag->node->frags)) { node 575 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, frag->node->raw); node 577 fs/jffs2/nodelist.c jffs2_free_full_dnode(frag->node); node 272 fs/jffs2/nodelist.h struct jffs2_full_dnode *node; /* NULL for holes */ node 329 fs/jffs2/nodelist.h struct rb_node *node = rb_first(root); node 331 fs/jffs2/nodelist.h if (!node) node 334 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); node 339 fs/jffs2/nodelist.h struct rb_node *node = rb_last(root); node 341 fs/jffs2/nodelist.h if (!node) node 344 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); node 191 fs/jffs2/read.c } else if (unlikely(!frag->node)) { node 210 fs/jffs2/read.c ref_offset(frag->node->raw), node 211 fs/jffs2/read.c ref_flags(frag->node->raw)); node 212 fs/jffs2/read.c ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen); node 402 fs/jffs2/readinode.c static void eat_last(struct rb_root *root, struct rb_node *node) node 404 fs/jffs2/readinode.c struct rb_node *parent = rb_parent(node); node 408 fs/jffs2/readinode.c BUG_ON(node->rb_right); node 412 fs/jffs2/readinode.c else if (node == parent->rb_left) node 417 fs/jffs2/readinode.c *link = node->rb_left; node 418 fs/jffs2/readinode.c if (node->rb_left) node 419 fs/jffs2/readinode.c node->rb_left->__rb_parent_color = node->__rb_parent_color; node 970 fs/jffs2/readinode.c union jffs2_node_union *node; node 1041 fs/jffs2/readinode.c node = (union jffs2_node_union *)buf; node 1044 fs/jffs2/readinode.c if (je32_to_cpu(node->u.hdr_crc) != crc32(0, node, sizeof(node->u)-4)) { node 1046 fs/jffs2/readinode.c ref_offset(ref), je16_to_cpu(node->u.magic), node 1047 fs/jffs2/readinode.c je16_to_cpu(node->u.nodetype), node 1048 fs/jffs2/readinode.c je32_to_cpu(node->u.totlen), node 1049 fs/jffs2/readinode.c je32_to_cpu(node->u.hdr_crc)); node 1054 fs/jffs2/readinode.c if (je16_to_cpu(node->u.magic) != JFFS2_MAGIC_BITMASK) { node 1057 fs/jffs2/readinode.c je16_to_cpu(node->u.magic), ref_offset(ref)); node 1062 fs/jffs2/readinode.c switch (je16_to_cpu(node->u.nodetype)) { node 1073 fs/jffs2/readinode.c err = read_direntry(c, ref, &node->d, retlen, rii); node 1088 fs/jffs2/readinode.c err = read_dnode(c, ref, &node->i, len, rii); node 1102 fs/jffs2/readinode.c err = read_unknown(c, ref, &node->u); node 1300 fs/jffs2/readinode.c f->metadata = frag_first(&f->fragtree)->node; node 364 fs/jffs2/scan.c raw->next_in_ino = xd->node->next_in_ino; node 365 fs/jffs2/scan.c xd->node->next_in_ino = raw; node 444 fs/jffs2/scan.c struct jffs2_unknown_node *node; node 638 fs/jffs2/scan.c if (jeb->offset + c->sector_size < ofs + sizeof(*node)) { node 642 fs/jffs2/scan.c sizeof(*node)); node 648 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + sizeof(*node)) { node 659 fs/jffs2/scan.c node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs]; node 721 fs/jffs2/scan.c if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) { node 729 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) { node 736 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) { node 744 fs/jffs2/scan.c if (je16_to_cpu(node->magic) != JFFS2_MAGIC_BITMASK) { node 749 fs/jffs2/scan.c je16_to_cpu(node->magic)); node 756 fs/jffs2/scan.c crcnode.magic = node->magic; node 757 fs/jffs2/scan.c crcnode.nodetype = cpu_to_je16( je16_to_cpu(node->nodetype) | JFFS2_NODE_ACCURATE); node 758 fs/jffs2/scan.c crcnode.totlen = node->totlen; node 761 fs/jffs2/scan.c if (hdr_crc != je32_to_cpu(node->hdr_crc)) { node 764 fs/jffs2/scan.c ofs, je16_to_cpu(node->magic), node 765 fs/jffs2/scan.c je16_to_cpu(node->nodetype), node 766 fs/jffs2/scan.c je32_to_cpu(node->totlen), node 767 fs/jffs2/scan.c je32_to_cpu(node->hdr_crc), node 775 fs/jffs2/scan.c if (ofs + je32_to_cpu(node->totlen) > jeb->offset + c->sector_size) { node 778 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen)); node 786 fs/jffs2/scan.c if (!(je16_to_cpu(node->nodetype) & JFFS2_NODE_ACCURATE)) { node 790 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 792 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 796 fs/jffs2/scan.c switch(je16_to_cpu(node->nodetype)) { node 807 fs/jffs2/scan.c node = (void *)buf; node 809 fs/jffs2/scan.c err = jffs2_scan_inode_node(c, jeb, (void *)node, ofs, s); node 811 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 815 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 818 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, node 824 fs/jffs2/scan.c node = (void *)buf; node 826 fs/jffs2/scan.c err = jffs2_scan_dirent_node(c, jeb, (void *)node, ofs, s); node 828 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 833 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 836 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, node 842 fs/jffs2/scan.c node = (void *)buf; node 844 fs/jffs2/scan.c err = jffs2_scan_xattr_node(c, jeb, (void *)node, ofs, s); node 847 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 850 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 853 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, node 859 fs/jffs2/scan.c node = (void *)buf; node 861 fs/jffs2/scan.c err = jffs2_scan_xref_node(c, jeb, (void *)node, ofs, s); node 864 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 870 fs/jffs2/scan.c if (je32_to_cpu(node->totlen) != c->cleanmarker_size) { node 872 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen), node 892 fs/jffs2/scan.c jffs2_sum_add_padding_mem(s, je32_to_cpu(node->totlen)); node 893 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 895 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 899 fs/jffs2/scan.c switch (je16_to_cpu(node->nodetype) & JFFS2_COMPAT_MASK) { node 902 fs/jffs2/scan.c je16_to_cpu(node->nodetype), ofs); node 906 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 908 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 913 fs/jffs2/scan.c je16_to_cpu(node->nodetype), ofs); node 918 fs/jffs2/scan.c je16_to_cpu(node->nodetype), ofs); node 919 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 921 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 926 fs/jffs2/scan.c je16_to_cpu(node->nodetype), ofs); node 928 fs/jffs2/scan.c jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(node->totlen)), NULL); node 932 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 251 fs/jffs2/summary.c union jffs2_node_union *node; node 259 fs/jffs2/summary.c node = invecs[0].iov_base; node 263 fs/jffs2/summary.c switch (je16_to_cpu(node->u.nodetype)) { node 271 fs/jffs2/summary.c temp->nodetype = node->i.nodetype; node 272 fs/jffs2/summary.c temp->inode = node->i.ino; node 273 fs/jffs2/summary.c temp->version = node->i.version; node 275 fs/jffs2/summary.c temp->totlen = node->i.totlen; node 283 fs/jffs2/summary.c kmalloc(sizeof(struct jffs2_sum_dirent_mem) + node->d.nsize, GFP_KERNEL); node 288 fs/jffs2/summary.c temp->nodetype = node->d.nodetype; node 289 fs/jffs2/summary.c temp->totlen = node->d.totlen; node 291 fs/jffs2/summary.c temp->pino = node->d.pino; node 292 fs/jffs2/summary.c temp->version = node->d.version; node 293 fs/jffs2/summary.c temp->ino = node->d.ino; node 294 fs/jffs2/summary.c temp->nsize = node->d.nsize; node 295 fs/jffs2/summary.c temp->type = node->d.type; node 300 fs/jffs2/summary.c memcpy(temp->name,node->d.name,node->d.nsize); node 304 fs/jffs2/summary.c memcpy(temp->name,invecs[1].iov_base,node->d.nsize); node 321 fs/jffs2/summary.c temp->nodetype = node->x.nodetype; node 322 fs/jffs2/summary.c temp->xid = node->x.xid; node 323 fs/jffs2/summary.c temp->version = node->x.version; node 324 fs/jffs2/summary.c temp->totlen = node->x.totlen; node 335 fs/jffs2/summary.c temp->nodetype = node->r.nodetype; node 344 fs/jffs2/summary.c c->summary->sum_padded += je32_to_cpu(node->u.totlen); node 510 fs/jffs2/summary.c raw->next_in_ino = xd->node->next_in_ino; node 511 fs/jffs2/summary.c xd->node->next_in_ino = raw; node 543 fs/jffs2/summary.c *pseudo_random += ref->node->flash_offset; node 185 fs/jffs2/wbuf.c union jffs2_node_union *node) node 191 fs/jffs2/wbuf.c node, je16_to_cpu(node->u.magic), je16_to_cpu(node->u.nodetype)); node 193 fs/jffs2/wbuf.c BUG_ON(je16_to_cpu(node->u.magic) != 0x1985 && node 194 fs/jffs2/wbuf.c je16_to_cpu(node->u.magic) != 0); node 196 fs/jffs2/wbuf.c switch (je16_to_cpu(node->u.nodetype)) { node 202 fs/jffs2/wbuf.c frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset)); node 205 fs/jffs2/wbuf.c while (!frag->node || frag->node->raw != raw) { node 209 fs/jffs2/wbuf.c dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node); node 210 fs/jffs2/wbuf.c return &frag->node->raw; node 223 fs/jffs2/wbuf.c je16_to_cpu(node->u.nodetype)); node 478 fs/jffs2/wbuf.c BUG_ON(xd->node != raw); node 479 fs/jffs2/wbuf.c adjust_ref = &xd->node; node 484 fs/jffs2/wbuf.c BUG_ON(xr->node != raw); node 485 fs/jffs2/wbuf.c adjust_ref = &xr->node; node 77 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 143 fs/jffs2/xattr.c offset = ref_offset(xd->node); node 144 fs/jffs2/xattr.c if (ref_flags(xd->node) == REF_PRISTINE) node 184 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 191 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((xd->node==raw) ? REF_PRISTINE : REF_NORMAL); node 212 fs/jffs2/xattr.c BUG_ON(ref_flags(xd->node) != REF_PRISTINE); node 220 fs/jffs2/xattr.c ret = jffs2_flash_read(c, ref_offset(xd->node)+sizeof(struct jffs2_raw_xattr), node 225 fs/jffs2/xattr.c ret, length, readlen, ref_offset(xd->node)); node 235 fs/jffs2/xattr.c ref_offset(xd->node), xd->data_crc, crc); node 410 fs/jffs2/xattr.c if (xd->node == (void *)xd) { node 454 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_UNCHECKED) node 456 fs/jffs2/xattr.c offset = ref_offset(ref->node); node 490 fs/jffs2/xattr.c for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) { node 497 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL); node 502 fs/jffs2/xattr.c ref->ino, ref->xid, ref_offset(ref->node)); node 794 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_PRISTINE) { node 796 fs/jffs2/xattr.c BUG_ON(ref->node->next_in_ino != (void *)ref); node 797 fs/jffs2/xattr.c ref->node->next_in_ino = NULL; node 798 fs/jffs2/xattr.c jffs2_mark_node_obsolete(c, ref->node); node 810 fs/jffs2/xattr.c raw = ref->node; node 813 fs/jffs2/xattr.c raw->next_in_ino = tmp->node; node 814 fs/jffs2/xattr.c tmp->node = raw; node 816 fs/jffs2/xattr.c raw->next_in_ino = tmp->node->next_in_ino; node 817 fs/jffs2/xattr.c tmp->node->next_in_ino = raw; node 1226 fs/jffs2/xattr.c if (xd->node != raw) node 1236 fs/jffs2/xattr.c old_ofs = ref_offset(xd->node); node 1247 fs/jffs2/xattr.c xd->xid, xd->version, old_ofs, ref_offset(xd->node)); node 1262 fs/jffs2/xattr.c BUG_ON(!ref->node); node 1264 fs/jffs2/xattr.c if (ref->node != raw) node 1269 fs/jffs2/xattr.c old_ofs = ref_offset(ref->node); node 1270 fs/jffs2/xattr.c totlen = ref_totlen(c, c->gcblock, ref->node); node 1281 fs/jffs2/xattr.c ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node)); node 1304 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 1312 fs/jffs2/xattr.c | ((xd->node == (void *)raw) ? REF_PRISTINE : REF_NORMAL); node 1325 fs/jffs2/xattr.c if (atomic_read(&xd->refcnt) || xd->node != (void *)xd) node 1337 fs/jffs2/xattr.c if (ref->node != (void *)ref) node 26 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node; node 48 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node; node 350 fs/kernfs/dir.c struct rb_node **node = &kn->parent->dir.children.rb_node; node 353 fs/kernfs/dir.c while (*node) { node 357 fs/kernfs/dir.c pos = rb_to_kn(*node); node 358 fs/kernfs/dir.c parent = *node; node 361 fs/kernfs/dir.c node = &pos->rb.rb_left; node 363 fs/kernfs/dir.c node = &pos->rb.rb_right; node 369 fs/kernfs/dir.c rb_link_node(&kn->rb, parent, node); node 835 fs/kernfs/dir.c struct rb_node *node = parent->dir.children.rb_node; node 848 fs/kernfs/dir.c while (node) { node 852 fs/kernfs/dir.c kn = rb_to_kn(node); node 855 fs/kernfs/dir.c node = node->rb_left; node 857 fs/kernfs/dir.c node = node->rb_right; node 1621 fs/kernfs/dir.c struct rb_node *node = parent->dir.children.rb_node; node 1622 fs/kernfs/dir.c while (node) { node 1623 fs/kernfs/dir.c pos = rb_to_kn(node); node 1626 fs/kernfs/dir.c node = node->rb_left; node 1628 fs/kernfs/dir.c node = node->rb_right; node 1635 fs/kernfs/dir.c struct rb_node *node = rb_next(&pos->rb); node 1636 fs/kernfs/dir.c if (!node) node 1639 fs/kernfs/dir.c pos = rb_to_kn(node); node 1650 fs/kernfs/dir.c struct rb_node *node = rb_next(&pos->rb); node 1651 fs/kernfs/dir.c if (!node) node 1654 fs/kernfs/dir.c pos = rb_to_kn(node); node 884 fs/kernfs/file.c list_for_each_entry(info, &kernfs_root(kn)->supers, node) { node 71 fs/kernfs/kernfs-internal.h struct list_head node; node 36 fs/kernfs/mount.c struct kernfs_node *node = kernfs_dentry_node(dentry); node 37 fs/kernfs/mount.c struct kernfs_root *root = kernfs_root(node); node 41 fs/kernfs/mount.c return scops->show_path(sf, node, root); node 305 fs/kernfs/mount.c INIT_LIST_HEAD(&info->node); node 325 fs/kernfs/mount.c list_add(&info->node, &info->root->supers); node 353 fs/kernfs/mount.c list_del(&info->node); node 133 fs/mbcache.c struct hlist_bl_node *node; node 139 fs/mbcache.c node = entry->e_hash_list.next; node 141 fs/mbcache.c node = hlist_bl_first(head); node 142 fs/mbcache.c while (node) { node 143 fs/mbcache.c entry = hlist_bl_entry(node, struct mb_cache_entry, node 149 fs/mbcache.c node = node->next; node 201 fs/mbcache.c struct hlist_bl_node *node; node 207 fs/mbcache.c hlist_bl_for_each_entry(entry, node, head, e_hash_list) { node 229 fs/mbcache.c struct hlist_bl_node *node; node 235 fs/mbcache.c hlist_bl_for_each_entry(entry, node, head, e_hash_list) { node 1115 fs/namespace.c struct llist_node *node = llist_del_all(&delayed_mntput_list); node 1118 fs/namespace.c llist_for_each_entry_safe(m, t, node, mnt_llist) node 152 fs/nfs/blocklayout/blocklayout.c container_of(be->be_device, struct pnfs_block_dev, node); node 590 fs/nfs/blocklayout/blocklayout.c struct nfs4_deviceid_node *node; node 594 fs/nfs/blocklayout/blocklayout.c node = nfs4_find_get_deviceid(server, id, cred, gfp_mask); node 595 fs/nfs/blocklayout/blocklayout.c if (!node) node 598 fs/nfs/blocklayout/blocklayout.c if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags) == 0) node 599 fs/nfs/blocklayout/blocklayout.c return node; node 603 fs/nfs/blocklayout/blocklayout.c if (!time_in_range(node->timestamp_unavailable, start, end)) { node 604 fs/nfs/blocklayout/blocklayout.c nfs4_delete_deviceid(node->ld, node->nfs_client, id); node 102 fs/nfs/blocklayout/blocklayout.h struct nfs4_deviceid_node node; node 46 fs/nfs/blocklayout/dev.c container_of(d, struct pnfs_block_dev, node); node 49 fs/nfs/blocklayout/dev.c kfree_rcu(dev, node.rcu); node 499 fs/nfs/blocklayout/dev.c struct nfs4_deviceid_node *node = NULL; node 537 fs/nfs/blocklayout/dev.c node = &top->node; node 538 fs/nfs/blocklayout/dev.c nfs4_init_deviceid_node(node, server, &pdev->dev_id); node 540 fs/nfs/blocklayout/dev.c nfs4_mark_deviceid_unavailable(node); node 547 fs/nfs/blocklayout/dev.c return node; node 13 fs/nfs/blocklayout/extent_tree.c ext_node(struct rb_node *node) node 15 fs/nfs/blocklayout/extent_tree.c return rb_entry(node, struct pnfs_block_extent, be_node); node 21 fs/nfs/blocklayout/extent_tree.c struct rb_node *node = rb_first(root); node 22 fs/nfs/blocklayout/extent_tree.c return node ? ext_node(node) : NULL; node 28 fs/nfs/blocklayout/extent_tree.c struct rb_node *node = rb_prev(&be->be_node); node 29 fs/nfs/blocklayout/extent_tree.c return node ? ext_node(node) : NULL; node 35 fs/nfs/blocklayout/extent_tree.c struct rb_node *node = rb_next(&be->be_node); node 36 fs/nfs/blocklayout/extent_tree.c return node ? ext_node(node) : NULL; node 48 fs/nfs/blocklayout/extent_tree.c struct rb_node *node = root->rb_node; node 51 fs/nfs/blocklayout/extent_tree.c while (node) { node 52 fs/nfs/blocklayout/extent_tree.c be = ext_node(node); node 54 fs/nfs/blocklayout/extent_tree.c node = node->rb_left; node 56 fs/nfs/blocklayout/extent_tree.c node = node->rb_right; node 325 fs/nfs/blocklayout/extent_tree.c struct rb_node *node; node 328 fs/nfs/blocklayout/extent_tree.c node = root->rb_node; node 329 fs/nfs/blocklayout/extent_tree.c while (node) { node 330 fs/nfs/blocklayout/extent_tree.c be = ext_node(node); node 332 fs/nfs/blocklayout/extent_tree.c node = node->rb_left; node 334 fs/nfs/blocklayout/extent_tree.c node = node->rb_right; node 246 fs/nfs/filelayout/filelayout.c filelayout_test_devid_unavailable(struct nfs4_deviceid_node *node) node 248 fs/nfs/filelayout/filelayout.c return filelayout_test_devid_invalid(node) || node 249 fs/nfs/filelayout/filelayout.c nfs4_test_deviceid_unavailable(node); node 255 fs/nfs/filelayout/filelayout.c struct nfs4_deviceid_node *node = FILELAYOUT_DEVID_NODE(lseg); node 257 fs/nfs/filelayout/filelayout.c return filelayout_test_devid_unavailable(node); node 96 fs/nfs/filelayout/filelayout.h filelayout_test_devid_invalid(struct nfs4_deviceid_node *node) node 98 fs/nfs/filelayout/filelayout.h return test_bit(NFS_DEVICEID_INVALID, &node->flags); node 102 fs/nfs/filelayout/filelayout.h filelayout_test_devid_unavailable(struct nfs4_deviceid_node *node); node 136 fs/nfs/flexfilelayout/flexfilelayout.h FF_LAYOUT_MIRROR_DS(struct nfs4_deviceid_node *node) node 138 fs/nfs/flexfilelayout/flexfilelayout.h return container_of(node, struct nfs4_ff_layout_ds, id_node); node 326 fs/nfs/flexfilelayout/flexfilelayoutdev.c struct nfs4_deviceid_node *node; node 329 fs/nfs/flexfilelayout/flexfilelayoutdev.c node = nfs4_find_get_deviceid(NFS_SERVER(lo->plh_inode), node 332 fs/nfs/flexfilelayout/flexfilelayoutdev.c if (node) node 333 fs/nfs/flexfilelayout/flexfilelayoutdev.c mirror_ds = FF_LAYOUT_MIRROR_DS(node); node 338 fs/nfs/flexfilelayout/flexfilelayoutdev.c nfs4_put_deviceid_node(node); node 152 fs/nfs/fscache.c xkey = rb_entry(parent, struct nfs_fscache_key, node); node 183 fs/nfs/fscache.c rb_link_node(&key->node, parent, p); node 184 fs/nfs/fscache.c rb_insert_color(&key->node, &nfs_fscache_keys); node 223 fs/nfs/fscache.c rb_erase(&nfss->fscache_key->node, &nfs_fscache_keys); node 22 fs/nfs/fscache.h struct rb_node node; node 338 fs/nfs/pnfs.h struct hlist_node node; node 357 fs/nfs/pnfs.h void nfs4_mark_deviceid_available(struct nfs4_deviceid_node *node); node 358 fs/nfs/pnfs.h void nfs4_mark_deviceid_unavailable(struct nfs4_deviceid_node *node); node 359 fs/nfs/pnfs.h bool nfs4_test_deviceid_unavailable(struct nfs4_deviceid_node *node); node 457 fs/nfs/pnfs.h pnfs_generic_mark_devid_invalid(struct nfs4_deviceid_node *node) node 459 fs/nfs/pnfs.h set_bit(NFS_DEVICEID_INVALID, &node->flags); node 83 fs/nfs/pnfs_dev.c hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node) node 208 fs/nfs/pnfs_dev.c hlist_add_head_rcu(&new->node, &nfs4_deviceid_cache[hash]); node 238 fs/nfs/pnfs_dev.c hlist_del_init_rcu(&d->node); node 251 fs/nfs/pnfs_dev.c INIT_HLIST_NODE(&d->node); node 287 fs/nfs/pnfs_dev.c nfs4_mark_deviceid_available(struct nfs4_deviceid_node *node) node 289 fs/nfs/pnfs_dev.c if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags)) { node 290 fs/nfs/pnfs_dev.c clear_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags); node 297 fs/nfs/pnfs_dev.c nfs4_mark_deviceid_unavailable(struct nfs4_deviceid_node *node) node 299 fs/nfs/pnfs_dev.c node->timestamp_unavailable = jiffies; node 301 fs/nfs/pnfs_dev.c set_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags); node 307 fs/nfs/pnfs_dev.c nfs4_test_deviceid_unavailable(struct nfs4_deviceid_node *node) node 309 fs/nfs/pnfs_dev.c if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags)) { node 314 fs/nfs/pnfs_dev.c if (time_in_range(node->timestamp_unavailable, start, end)) node 316 fs/nfs/pnfs_dev.c clear_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags); node 331 fs/nfs/pnfs_dev.c hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node) node 333 fs/nfs/pnfs_dev.c hlist_del_init_rcu(&d->node); node 372 fs/nfs/pnfs_dev.c hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[i], node) node 2657 fs/nfsd/nfs4state.c struct rb_node *node = root->rb_node; node 2660 fs/nfsd/nfs4state.c while (node) { node 2661 fs/nfsd/nfs4state.c clp = rb_entry(node, struct nfs4_client, cl_namenode); node 2664 fs/nfsd/nfs4state.c node = node->rb_left; node 2666 fs/nfsd/nfs4state.c node = node->rb_right; node 73 fs/nilfs2/btree.c static int nilfs_btree_node_get_flags(const struct nilfs_btree_node *node) node 75 fs/nilfs2/btree.c return node->bn_flags; node 79 fs/nilfs2/btree.c nilfs_btree_node_set_flags(struct nilfs_btree_node *node, int flags) node 81 fs/nilfs2/btree.c node->bn_flags = flags; node 84 fs/nilfs2/btree.c static int nilfs_btree_node_root(const struct nilfs_btree_node *node) node 86 fs/nilfs2/btree.c return nilfs_btree_node_get_flags(node) & NILFS_BTREE_NODE_ROOT; node 89 fs/nilfs2/btree.c static int nilfs_btree_node_get_level(const struct nilfs_btree_node *node) node 91 fs/nilfs2/btree.c return node->bn_level; node 95 fs/nilfs2/btree.c nilfs_btree_node_set_level(struct nilfs_btree_node *node, int level) node 97 fs/nilfs2/btree.c node->bn_level = level; node 100 fs/nilfs2/btree.c static int nilfs_btree_node_get_nchildren(const struct nilfs_btree_node *node) node 102 fs/nilfs2/btree.c return le16_to_cpu(node->bn_nchildren); node 106 fs/nilfs2/btree.c nilfs_btree_node_set_nchildren(struct nilfs_btree_node *node, int nchildren) node 108 fs/nilfs2/btree.c node->bn_nchildren = cpu_to_le16(nchildren); node 122 fs/nilfs2/btree.c nilfs_btree_node_dkeys(const struct nilfs_btree_node *node) node 124 fs/nilfs2/btree.c return (__le64 *)((char *)(node + 1) + node 125 fs/nilfs2/btree.c (nilfs_btree_node_root(node) ? node 130 fs/nilfs2/btree.c nilfs_btree_node_dptrs(const struct nilfs_btree_node *node, int ncmax) node 132 fs/nilfs2/btree.c return (__le64 *)(nilfs_btree_node_dkeys(node) + ncmax); node 136 fs/nilfs2/btree.c nilfs_btree_node_get_key(const struct nilfs_btree_node *node, int index) node 138 fs/nilfs2/btree.c return le64_to_cpu(*(nilfs_btree_node_dkeys(node) + index)); node 142 fs/nilfs2/btree.c nilfs_btree_node_set_key(struct nilfs_btree_node *node, int index, __u64 key) node 144 fs/nilfs2/btree.c *(nilfs_btree_node_dkeys(node) + index) = cpu_to_le64(key); node 148 fs/nilfs2/btree.c nilfs_btree_node_get_ptr(const struct nilfs_btree_node *node, int index, node 151 fs/nilfs2/btree.c return le64_to_cpu(*(nilfs_btree_node_dptrs(node, ncmax) + index)); node 155 fs/nilfs2/btree.c nilfs_btree_node_set_ptr(struct nilfs_btree_node *node, int index, __u64 ptr, node 158 fs/nilfs2/btree.c *(nilfs_btree_node_dptrs(node, ncmax) + index) = cpu_to_le64(ptr); node 161 fs/nilfs2/btree.c static void nilfs_btree_node_init(struct nilfs_btree_node *node, int flags, node 169 fs/nilfs2/btree.c nilfs_btree_node_set_flags(node, flags); node 170 fs/nilfs2/btree.c nilfs_btree_node_set_level(node, level); node 171 fs/nilfs2/btree.c nilfs_btree_node_set_nchildren(node, nchildren); node 173 fs/nilfs2/btree.c dkeys = nilfs_btree_node_dkeys(node); node 174 fs/nilfs2/btree.c dptrs = nilfs_btree_node_dptrs(node, ncmax); node 238 fs/nilfs2/btree.c static void nilfs_btree_node_insert(struct nilfs_btree_node *node, int index, node 245 fs/nilfs2/btree.c dkeys = nilfs_btree_node_dkeys(node); node 246 fs/nilfs2/btree.c dptrs = nilfs_btree_node_dptrs(node, ncmax); node 247 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 257 fs/nilfs2/btree.c nilfs_btree_node_set_nchildren(node, nchildren); node 261 fs/nilfs2/btree.c static void nilfs_btree_node_delete(struct nilfs_btree_node *node, int index, node 270 fs/nilfs2/btree.c dkeys = nilfs_btree_node_dkeys(node); node 271 fs/nilfs2/btree.c dptrs = nilfs_btree_node_dptrs(node, ncmax); node 274 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 287 fs/nilfs2/btree.c nilfs_btree_node_set_nchildren(node, nchildren); node 290 fs/nilfs2/btree.c static int nilfs_btree_node_lookup(const struct nilfs_btree_node *node, node 298 fs/nilfs2/btree.c high = nilfs_btree_node_get_nchildren(node) - 1; node 303 fs/nilfs2/btree.c nkey = nilfs_btree_node_get_key(node, index); node 317 fs/nilfs2/btree.c if (nilfs_btree_node_get_level(node) > NILFS_BTREE_LEVEL_NODE_MIN) { node 338 fs/nilfs2/btree.c static int nilfs_btree_node_broken(const struct nilfs_btree_node *node, node 345 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 346 fs/nilfs2/btree.c flags = nilfs_btree_node_get_flags(node); node 347 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 370 fs/nilfs2/btree.c static int nilfs_btree_root_broken(const struct nilfs_btree_node *node, node 376 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 377 fs/nilfs2/btree.c flags = nilfs_btree_node_get_flags(node); node 378 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 436 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 439 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 442 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 445 fs/nilfs2/btree.c return node; node 449 fs/nilfs2/btree.c struct nilfs_btree_node *node, int level) node 451 fs/nilfs2/btree.c if (unlikely(nilfs_btree_node_get_level(node) != level)) { node 456 fs/nilfs2/btree.c nilfs_btree_node_get_level(node), level); node 463 fs/nilfs2/btree.c struct nilfs_btree_node *node; /* parent node */ node 493 fs/nilfs2/btree.c ptr2 = nilfs_btree_node_get_ptr(ra->node, i, ra->ncmax); node 540 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 545 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 546 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 547 fs/nilfs2/btree.c if (level < minlevel || nilfs_btree_node_get_nchildren(node) <= 0) node 550 fs/nilfs2/btree.c found = nilfs_btree_node_lookup(node, key, &index); node 551 fs/nilfs2/btree.c ptr = nilfs_btree_node_get_ptr(node, index, node 561 fs/nilfs2/btree.c p.node = nilfs_btree_get_node(btree, path, level + 1, node 572 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 573 fs/nilfs2/btree.c if (nilfs_btree_bad_node(btree, node, level)) node 576 fs/nilfs2/btree.c found = nilfs_btree_node_lookup(node, key, &index); node 580 fs/nilfs2/btree.c ptr = nilfs_btree_node_get_ptr(node, index, ncmax); node 601 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 605 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 606 fs/nilfs2/btree.c index = nilfs_btree_node_get_nchildren(node) - 1; node 609 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 610 fs/nilfs2/btree.c ptr = nilfs_btree_node_get_ptr(node, index, node 620 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 621 fs/nilfs2/btree.c if (nilfs_btree_bad_node(btree, node, level)) node 623 fs/nilfs2/btree.c index = nilfs_btree_node_get_nchildren(node) - 1; node 624 fs/nilfs2/btree.c ptr = nilfs_btree_node_get_ptr(node, index, ncmax); node 629 fs/nilfs2/btree.c *keyp = nilfs_btree_node_get_key(node, index); node 650 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 658 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 660 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 663 fs/nilfs2/btree.c if (index < nilfs_btree_node_get_nchildren(node)) { node 665 fs/nilfs2/btree.c *nextkey = nilfs_btree_node_get_key(node, index); node 696 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 724 fs/nilfs2/btree.c node = nilfs_btree_get_node(btree, path, level, &ncmax); node 727 fs/nilfs2/btree.c while (index < nilfs_btree_node_get_nchildren(node)) { node 728 fs/nilfs2/btree.c if (nilfs_btree_node_get_key(node, index) != node 731 fs/nilfs2/btree.c ptr2 = nilfs_btree_node_get_ptr(node, index, ncmax); node 747 fs/nilfs2/btree.c p.node = nilfs_btree_get_node(btree, path, level + 1, &p.ncmax); node 750 fs/nilfs2/btree.c if (p.index >= nilfs_btree_node_get_nchildren(p.node) || node 751 fs/nilfs2/btree.c nilfs_btree_node_get_key(p.node, p.index) != key + cnt) node 753 fs/nilfs2/btree.c ptr2 = nilfs_btree_node_get_ptr(p.node, p.index, p.ncmax); node 763 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 802 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 806 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 808 fs/nilfs2/btree.c nilfs_btree_node_insert(node, path[level].bp_index, node 815 fs/nilfs2/btree.c nilfs_btree_node_get_key(node, node 818 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 819 fs/nilfs2/btree.c nilfs_btree_node_insert(node, path[level].bp_index, node 829 fs/nilfs2/btree.c struct nilfs_btree_node *node, *left; node 832 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 834 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 846 fs/nilfs2/btree.c nilfs_btree_node_move_left(left, node, n, ncblk, ncblk); node 854 fs/nilfs2/btree.c nilfs_btree_node_get_key(node, 0)); node 875 fs/nilfs2/btree.c struct nilfs_btree_node *node, *right; node 878 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 880 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 892 fs/nilfs2/btree.c nilfs_btree_node_move_right(node, right, n, ncblk, ncblk); node 908 fs/nilfs2/btree.c path[level].bp_index -= nilfs_btree_node_get_nchildren(node); node 922 fs/nilfs2/btree.c struct nilfs_btree_node *node, *right; node 925 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 927 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 937 fs/nilfs2/btree.c nilfs_btree_node_move_right(node, right, n, ncblk, ncblk); node 945 fs/nilfs2/btree.c path[level].bp_index -= nilfs_btree_node_get_nchildren(node); node 1000 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 1009 fs/nilfs2/btree.c node = nilfs_btree_get_node(btree, path, level, &ncmax); node 1010 fs/nilfs2/btree.c return nilfs_btree_node_get_ptr(node, node 1018 fs/nilfs2/btree.c node = nilfs_btree_get_node(btree, path, level, &ncmax); node 1019 fs/nilfs2/btree.c return nilfs_btree_node_get_ptr(node, path[level].bp_index, node 1052 fs/nilfs2/btree.c struct nilfs_btree_node *node, *parent, *sib; node 1076 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1077 fs/nilfs2/btree.c if (nilfs_btree_node_get_nchildren(node) < ncblk) { node 1144 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 1145 fs/nilfs2/btree.c if (nilfs_btree_node_get_nchildren(node) < node 1252 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 1256 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1258 fs/nilfs2/btree.c nilfs_btree_node_delete(node, path[level].bp_index, node 1264 fs/nilfs2/btree.c nilfs_btree_node_get_key(node, 0)); node 1266 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 1267 fs/nilfs2/btree.c nilfs_btree_node_delete(node, path[level].bp_index, node 1277 fs/nilfs2/btree.c struct nilfs_btree_node *node, *left; node 1282 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1284 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 1290 fs/nilfs2/btree.c nilfs_btree_node_move_right(left, node, n, ncblk, ncblk); node 1298 fs/nilfs2/btree.c nilfs_btree_node_get_key(node, 0)); node 1309 fs/nilfs2/btree.c struct nilfs_btree_node *node, *right; node 1314 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1316 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 1322 fs/nilfs2/btree.c nilfs_btree_node_move_left(node, right, n, ncblk, ncblk); node 1342 fs/nilfs2/btree.c struct nilfs_btree_node *node, *left; node 1347 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1351 fs/nilfs2/btree.c n = nilfs_btree_node_get_nchildren(node); node 1353 fs/nilfs2/btree.c nilfs_btree_node_move_left(left, node, n, ncblk, ncblk); node 1368 fs/nilfs2/btree.c struct nilfs_btree_node *node, *right; node 1373 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1379 fs/nilfs2/btree.c nilfs_btree_node_move_left(node, right, n, ncblk, ncblk); node 1426 fs/nilfs2/btree.c struct nilfs_btree_node *node, *parent, *sib; node 1438 fs/nilfs2/btree.c node = nilfs_btree_get_nonroot_node(path, level); node 1440 fs/nilfs2/btree.c nilfs_btree_node_get_ptr(node, dindex, ncblk); node 1446 fs/nilfs2/btree.c if (nilfs_btree_node_get_nchildren(node) > ncmin) { node 1507 fs/nilfs2/btree.c if (nilfs_btree_node_get_nchildren(node) - 1 <= node 1527 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 1529 fs/nilfs2/btree.c nilfs_btree_node_get_ptr(node, dindex, node 1640 fs/nilfs2/btree.c struct nilfs_btree_node *root, *node; node 1649 fs/nilfs2/btree.c node = root; node 1660 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; node 1666 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 1667 fs/nilfs2/btree.c maxkey = nilfs_btree_node_get_key(node, nchildren - 1); node 1669 fs/nilfs2/btree.c nilfs_btree_node_get_key(node, nchildren - 2) : 0; node 1680 fs/nilfs2/btree.c struct nilfs_btree_node *node, *root; node 1690 fs/nilfs2/btree.c node = root; node 1701 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; node 1705 fs/nilfs2/btree.c node = NULL; node 1709 fs/nilfs2/btree.c nchildren = nilfs_btree_node_get_nchildren(node); node 1712 fs/nilfs2/btree.c dkeys = nilfs_btree_node_dkeys(node); node 1713 fs/nilfs2/btree.c dptrs = nilfs_btree_node_dptrs(node, ncmax); node 1787 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 1807 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; node 1809 fs/nilfs2/btree.c nilfs_btree_node_init(node, 0, 1, n, ncblk, keys, ptrs); node 1810 fs/nilfs2/btree.c nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); node 1819 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 1821 fs/nilfs2/btree.c nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 2, 1, node 1828 fs/nilfs2/btree.c node = nilfs_btree_get_root(btree); node 1829 fs/nilfs2/btree.c nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 1, n, node 1832 fs/nilfs2/btree.c nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, node 2055 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 2066 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; node 2067 fs/nilfs2/btree.c key = nilfs_btree_node_get_key(node, 0); node 2068 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 2106 fs/nilfs2/btree.c struct nilfs_btree_node *node, *cnode; node 2111 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; node 2112 fs/nilfs2/btree.c key = nilfs_btree_node_get_key(node, 0); node 2113 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 2250 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 2259 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)(*bh)->b_data; node 2260 fs/nilfs2/btree.c key = nilfs_btree_node_get_key(node, 0); node 2261 fs/nilfs2/btree.c level = nilfs_btree_node_get_level(node); node 2288 fs/nilfs2/btree.c struct nilfs_btree_node *node; node 2298 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)(*bh)->b_data; node 2299 fs/nilfs2/btree.c key = nilfs_btree_node_get_key(node, 0); node 241 fs/notify/fsnotify.c struct hlist_node *node = NULL; node 245 fs/notify/fsnotify.c node = srcu_dereference(conn->list.first, &fsnotify_mark_srcu); node 247 fs/notify/fsnotify.c return hlist_entry_safe(node, struct fsnotify_mark, obj_list); node 252 fs/notify/fsnotify.c struct hlist_node *node = NULL; node 255 fs/notify/fsnotify.c node = srcu_dereference(mark->obj_list.next, node 258 fs/notify/fsnotify.c return hlist_entry_safe(node, struct fsnotify_mark, obj_list); node 578 fs/ocfs2/alloc.c struct ocfs2_path_item *node; node 584 fs/ocfs2/alloc.c node = &path->p_node[i]; node 586 fs/ocfs2/alloc.c brelse(node->bh); node 587 fs/ocfs2/alloc.c node->bh = NULL; node 588 fs/ocfs2/alloc.c node->el = NULL; node 741 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node, node 748 fs/ocfs2/cluster/heartbeat.c (f->hc_func)(node, idx, f->hc_data); node 794 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node, node 799 fs/ocfs2/cluster/heartbeat.c BUG_ON((!node) && (type != O2HB_NODE_DOWN_CB)); node 802 fs/ocfs2/cluster/heartbeat.c event->hn_node = node; node 815 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node; node 818 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num); node 819 fs/ocfs2/cluster/heartbeat.c if (!node) node 832 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node, node 842 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node); node 894 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node; node 908 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num); node 909 fs/ocfs2/cluster/heartbeat.c if (!node) { node 990 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node, node 1043 fs/ocfs2/cluster/heartbeat.c node, slot->ds_node_num); node 1065 fs/ocfs2/cluster/heartbeat.c if (node) node 1066 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node); node 34 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 40 fs/ocfs2/cluster/nodemanager.c node = o2nm_single_cluster->cl_nodes[node_num]; node 41 fs/ocfs2/cluster/nodemanager.c if (node) node 42 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 45 fs/ocfs2/cluster/nodemanager.c return node; node 73 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node, *ret = NULL; node 79 fs/ocfs2/cluster/nodemanager.c node = rb_entry(parent, struct o2nm_node, nd_ip_node); node 81 fs/ocfs2/cluster/nodemanager.c cmp = memcmp(&ip_needle, &node->nd_ipv4_address, node 88 fs/ocfs2/cluster/nodemanager.c ret = node; node 103 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 110 fs/ocfs2/cluster/nodemanager.c node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL); node 111 fs/ocfs2/cluster/nodemanager.c if (node) node 112 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 116 fs/ocfs2/cluster/nodemanager.c return node; node 120 fs/ocfs2/cluster/nodemanager.c void o2nm_node_put(struct o2nm_node *node) node 122 fs/ocfs2/cluster/nodemanager.c config_item_put(&node->nd_item); node 126 fs/ocfs2/cluster/nodemanager.c void o2nm_node_get(struct o2nm_node *node) node 128 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 160 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 161 fs/ocfs2/cluster/nodemanager.c kfree(node); node 169 fs/ocfs2/cluster/nodemanager.c static struct o2nm_cluster *to_o2nm_cluster_from_node(struct o2nm_node *node) node 173 fs/ocfs2/cluster/nodemanager.c if (node->nd_item.ci_parent) node 174 fs/ocfs2/cluster/nodemanager.c return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent); node 188 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 205 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || node 206 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) node 210 fs/ocfs2/cluster/nodemanager.c cluster = to_o2nm_cluster_from_node(node); node 220 fs/ocfs2/cluster/nodemanager.c &node->nd_set_attributes)) node 223 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[tmp] = node; node 224 fs/ocfs2/cluster/nodemanager.c node->nd_num = tmp; node 243 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 256 fs/ocfs2/cluster/nodemanager.c if (test_and_set_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) node 258 fs/ocfs2/cluster/nodemanager.c node->nd_ipv4_port = htons(tmp); node 272 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 291 fs/ocfs2/cluster/nodemanager.c cluster = to_o2nm_cluster_from_node(node); node 302 fs/ocfs2/cluster/nodemanager.c &node->nd_set_attributes)) node 305 fs/ocfs2/cluster/nodemanager.c rb_link_node(&node->nd_ip_node, parent, p); node 306 fs/ocfs2/cluster/nodemanager.c rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree); node 314 fs/ocfs2/cluster/nodemanager.c memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr)); node 327 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 341 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || node 342 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) || node 343 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) node 347 fs/ocfs2/cluster/nodemanager.c cluster = to_o2nm_cluster_from_node(node); node 356 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node != node->nd_num) { node 363 fs/ocfs2/cluster/nodemanager.c ret = o2net_start_listening(node); node 369 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node == node->nd_num) { node 370 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node); node 374 fs/ocfs2/cluster/nodemanager.c node->nd_local = tmp; node 375 fs/ocfs2/cluster/nodemanager.c if (node->nd_local) { node 377 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node = node->nd_num; node 586 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 591 fs/ocfs2/cluster/nodemanager.c node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL); node 592 fs/ocfs2/cluster/nodemanager.c if (node == NULL) node 595 fs/ocfs2/cluster/nodemanager.c strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */ node 596 fs/ocfs2/cluster/nodemanager.c config_item_init_type_name(&node->nd_item, name, &o2nm_node_type); node 597 fs/ocfs2/cluster/nodemanager.c spin_lock_init(&node->nd_lock); node 601 fs/ocfs2/cluster/nodemanager.c return &node->nd_item; node 607 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 610 fs/ocfs2/cluster/nodemanager.c if (cluster->cl_nodes[node->nd_num] == node) { node 611 fs/ocfs2/cluster/nodemanager.c o2net_disconnect_node(node); node 614 fs/ocfs2/cluster/nodemanager.c (cluster->cl_local_node == node->nd_num)) { node 617 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node); node 626 fs/ocfs2/cluster/nodemanager.c if (node->nd_ipv4_address) node 627 fs/ocfs2/cluster/nodemanager.c rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree); node 630 fs/ocfs2/cluster/nodemanager.c if (cluster->cl_nodes[node->nd_num] == node) { node 631 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[node->nd_num] = NULL; node 632 fs/ocfs2/cluster/nodemanager.c clear_bit(node->nd_num, cluster->cl_nodes_bitmap); node 637 fs/ocfs2/cluster/nodemanager.c config_item_name(&node->nd_item)); node 65 fs/ocfs2/cluster/nodemanager.h void o2nm_node_get(struct o2nm_node *node); node 66 fs/ocfs2/cluster/nodemanager.h void o2nm_node_put(struct o2nm_node *node); node 164 fs/ocfs2/cluster/quorum.c static void o2quo_set_hold(struct o2quo_state *qs, u8 node) node 168 fs/ocfs2/cluster/quorum.c if (!test_and_set_bit(node, qs->qs_hold_bm)) { node 171 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 172 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds); node 176 fs/ocfs2/cluster/quorum.c static void o2quo_clear_hold(struct o2quo_state *qs, u8 node) node 180 fs/ocfs2/cluster/quorum.c if (test_and_clear_bit(node, qs->qs_hold_bm)) { node 181 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1); node 189 fs/ocfs2/cluster/quorum.c node, qs->qs_holds); node 197 fs/ocfs2/cluster/quorum.c void o2quo_hb_up(u8 node) node 205 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 206 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node); node 207 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_hb_bm); node 209 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); node 211 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_conn_bm)) node 212 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 214 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 221 fs/ocfs2/cluster/quorum.c void o2quo_hb_down(u8 node) node 230 fs/ocfs2/cluster/quorum.c node, qs->qs_heartbeating); node 231 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node); node 232 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_hb_bm); node 234 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); node 236 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 246 fs/ocfs2/cluster/quorum.c void o2quo_hb_still_up(u8 node) node 252 fs/ocfs2/cluster/quorum.c mlog(0, "node %u\n", node); node 255 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 265 fs/ocfs2/cluster/quorum.c void o2quo_conn_up(u8 node) node 273 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 274 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node); node 275 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_conn_bm); node 277 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); node 279 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_hb_bm)) node 280 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 282 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 291 fs/ocfs2/cluster/quorum.c void o2quo_conn_err(u8 node) node 297 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_conn_bm)) { node 301 fs/ocfs2/cluster/quorum.c node, qs->qs_connected); node 303 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_conn_bm); node 305 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_hb_bm)) node 306 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 309 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); node 14 fs/ocfs2/cluster/quorum.h void o2quo_hb_up(u8 node); node 15 fs/ocfs2/cluster/quorum.h void o2quo_hb_down(u8 node); node 16 fs/ocfs2/cluster/quorum.h void o2quo_hb_still_up(u8 node); node 17 fs/ocfs2/cluster/quorum.h void o2quo_conn_up(u8 node); node 18 fs/ocfs2/cluster/quorum.h void o2quo_conn_err(u8 node); node 135 fs/ocfs2/cluster/tcp.c u32 msgkey, struct task_struct *task, u8 node) node 141 fs/ocfs2/cluster/tcp.c nst->st_node = node; node 412 fs/ocfs2/cluster/tcp.c static struct o2net_sock_container *sc_alloc(struct o2nm_node *node) node 424 fs/ocfs2/cluster/tcp.c o2nm_node_get(node); node 425 fs/ocfs2/cluster/tcp.c sc->sc_node = node; node 428 fs/ocfs2/cluster/tcp.c status = o2nm_depend_item(&node->nd_item); node 431 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 998 fs/ocfs2/cluster/tcp.c int node, ret; node 1003 fs/ocfs2/cluster/tcp.c for (node = 0; node < O2NM_MAX_NODES; ++node) { node 1004 fs/ocfs2/cluster/tcp.c if (!o2net_tx_can_proceed(o2net_nn_from_num(node), &sc, &ret)) node 1007 fs/ocfs2/cluster/tcp.c set_bit(node, map); node 1568 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL, *mynode = NULL; node 1587 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_num(o2net_num_from_nn(nn)); node 1588 fs/ocfs2/cluster/tcp.c if (node == NULL) node 1613 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node); node 1661 fs/ocfs2/cluster/tcp.c remoteaddr.sin_addr.s_addr = node->nd_ipv4_address; node 1662 fs/ocfs2/cluster/tcp.c remoteaddr.sin_port = node->nd_ipv4_port; node 1681 fs/ocfs2/cluster/tcp.c if (node) node 1682 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 1719 fs/ocfs2/cluster/tcp.c void o2net_disconnect_node(struct o2nm_node *node) node 1721 fs/ocfs2/cluster/tcp.c struct o2net_node *nn = o2net_nn_from_num(node->nd_num); node 1737 fs/ocfs2/cluster/tcp.c static void o2net_hb_node_down_cb(struct o2nm_node *node, int node_num, node 1742 fs/ocfs2/cluster/tcp.c if (!node) node 1746 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node); node 1751 fs/ocfs2/cluster/tcp.c static void o2net_hb_node_up_cb(struct o2nm_node *node, int node_num, node 1758 fs/ocfs2/cluster/tcp.c BUG_ON(!node); node 1809 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL; node 1855 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_ip(sin.sin_addr.s_addr); node 1856 fs/ocfs2/cluster/tcp.c if (node == NULL) { node 1864 fs/ocfs2/cluster/tcp.c if (o2nm_this_node() >= node->nd_num) { node 1873 fs/ocfs2/cluster/tcp.c node->nd_name, node 1874 fs/ocfs2/cluster/tcp.c node->nd_num, &sin.sin_addr.s_addr, node 1882 fs/ocfs2/cluster/tcp.c if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) { node 1885 fs/ocfs2/cluster/tcp.c node->nd_name, &sin.sin_addr.s_addr, node 1891 fs/ocfs2/cluster/tcp.c nn = o2net_nn_from_num(node->nd_num); node 1902 fs/ocfs2/cluster/tcp.c node->nd_name, &sin.sin_addr.s_addr, node 1907 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node); node 1930 fs/ocfs2/cluster/tcp.c if (node) node 1931 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 2064 fs/ocfs2/cluster/tcp.c int o2net_start_listening(struct o2nm_node *node) node 2078 fs/ocfs2/cluster/tcp.c ret = o2net_open_listening_sock(node->nd_ipv4_address, node 2079 fs/ocfs2/cluster/tcp.c node->nd_ipv4_port); node 2084 fs/ocfs2/cluster/tcp.c o2quo_conn_up(node->nd_num); node 2091 fs/ocfs2/cluster/tcp.c void o2net_stop_listening(struct o2nm_node *node) node 2106 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = o2nm_get_node_by_num(i); node 2107 fs/ocfs2/cluster/tcp.c if (node) { node 2108 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node); node 2109 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 2121 fs/ocfs2/cluster/tcp.c o2quo_conn_err(node->nd_num); node 100 fs/ocfs2/cluster/tcp.h int o2net_start_listening(struct o2nm_node *node); node 101 fs/ocfs2/cluster/tcp.h void o2net_stop_listening(struct o2nm_node *node); node 102 fs/ocfs2/cluster/tcp.h void o2net_disconnect_node(struct o2nm_node *node); node 190 fs/ocfs2/dlm/dlmast.c lock->ml.node == dlm->node_num ? "master" : node 219 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); node 239 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node == dlm->node_num); node 255 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); node 281 fs/ocfs2/dlm/dlmast.c u8 node; node 295 fs/ocfs2/dlm/dlmast.c node = past->node_idx; node 323 fs/ocfs2/dlm/dlmast.c locklen, name, node); node 334 fs/ocfs2/dlm/dlmast.c locklen, name, node); node 383 fs/ocfs2/dlm/dlmast.c locklen, name, node); node 444 fs/ocfs2/dlm/dlmast.c res->lockname.len, res->lockname.name, lock->ml.node, msg_type, node 465 fs/ocfs2/dlm/dlmast.c lock->ml.node, &status); node 469 fs/ocfs2/dlm/dlmast.c lock->ml.node); node 473 fs/ocfs2/dlm/dlmast.c "node is dead!\n", lock->ml.node); node 477 fs/ocfs2/dlm/dlmast.c "DLM_MIGRATING!\n", lock->ml.node); node 481 fs/ocfs2/dlm/dlmast.c lock->ml.node, status); node 253 fs/ocfs2/dlm/dlmcommon.h u8 node) node 257 fs/ocfs2/dlm/dlmcommon.h dlm->joining_node = node; node 337 fs/ocfs2/dlm/dlmcommon.h u8 node; node 819 fs/ocfs2/dlm/dlmcommon.h struct dlm_lock * dlm_new_lock(int type, u8 node, u64 cookie, node 852 fs/ocfs2/dlm/dlmcommon.h int dlm_is_node_dead(struct dlm_ctxt *dlm, u8 node); node 853 fs/ocfs2/dlm/dlmcommon.h void dlm_wait_for_node_death(struct dlm_ctxt *dlm, u8 node, int timeout); node 854 fs/ocfs2/dlm/dlmcommon.h void dlm_wait_for_node_recovery(struct dlm_ctxt *dlm, u8 node, int timeout); node 952 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data); node 953 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data); node 194 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.node == dlm->node_num) node 486 fs/ocfs2/dlm/dlmconvert.c tmp_lock->ml.node == cnv->node_idx) { node 66 fs/ocfs2/dlm/dlmdebug.c lock->ml.type, lock->ml.convert_type, lock->ml.node, node 474 fs/ocfs2/dlm/dlmdebug.c lock->ml.node, node 665 fs/ocfs2/dlm/dlmdebug.c struct dlm_reco_node_data *node; node 791 fs/ocfs2/dlm/dlmdebug.c list_for_each_entry(node, &dlm->reco.node_data, list) { node 792 fs/ocfs2/dlm/dlmdebug.c switch (node->state) { node 819 fs/ocfs2/dlm/dlmdebug.c node->node_num, state); node 490 fs/ocfs2/dlm/dlmdomain.c unsigned int node; node 496 fs/ocfs2/dlm/dlmdomain.c node = exit_msg->node_idx; node 497 fs/ocfs2/dlm/dlmdomain.c mlog(0, "%s: Node %u sent a begin exit domain message\n", dlm->name, node); node 500 fs/ocfs2/dlm/dlmdomain.c set_bit(node, dlm->exit_domain_map); node 533 fs/ocfs2/dlm/dlmdomain.c int node = -1, num = 0; node 538 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node 539 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 540 fs/ocfs2/dlm/dlmdomain.c printk("%d ", node); node 550 fs/ocfs2/dlm/dlmdomain.c unsigned int node; node 558 fs/ocfs2/dlm/dlmdomain.c node = exit_msg->node_idx; node 561 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map); node 562 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->exit_domain_map); node 563 fs/ocfs2/dlm/dlmdomain.c printk(KERN_NOTICE "o2dlm: Node %u leaves domain %s ", node, dlm->name); node 567 fs/ocfs2/dlm/dlmdomain.c dlm_hb_event_notify_attached(dlm, node, 0); node 577 fs/ocfs2/dlm/dlmdomain.c unsigned int node) node 583 fs/ocfs2/dlm/dlmdomain.c msg_type, node); node 589 fs/ocfs2/dlm/dlmdomain.c sizeof(leave_msg), node, NULL); node 592 fs/ocfs2/dlm/dlmdomain.c "to node %u on domain %s\n", status, msg_type, node, node 600 fs/ocfs2/dlm/dlmdomain.c int node = -1; node 614 fs/ocfs2/dlm/dlmdomain.c node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node + 1); node 615 fs/ocfs2/dlm/dlmdomain.c if (node >= O2NM_MAX_NODES) node 617 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 621 fs/ocfs2/dlm/dlmdomain.c dlm_send_one_domain_exit(dlm, DLM_BEGIN_EXIT_DOMAIN_MSG, node); node 629 fs/ocfs2/dlm/dlmdomain.c int node, clear_node, status; node 639 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node 652 fs/ocfs2/dlm/dlmdomain.c node); node 657 fs/ocfs2/dlm/dlmdomain.c "to node %d\n", status, node); node 670 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map); node 728 fs/ocfs2/dlm/dlmdomain.c static int dlm_query_join_proto_check(char *proto_type, int node, node 739 fs/ocfs2/dlm/dlmdomain.c node, proto_type, node 749 fs/ocfs2/dlm/dlmdomain.c node, proto_type, node 1219 fs/ocfs2/dlm/dlmdomain.c struct o2nm_node *node; node 1233 fs/ocfs2/dlm/dlmdomain.c node = o2nm_get_node_by_num(i); node 1234 fs/ocfs2/dlm/dlmdomain.c if (!node) node 1236 fs/ocfs2/dlm/dlmdomain.c qn->qn_nodes[count].ni_nodenum = node->nd_num; node 1237 fs/ocfs2/dlm/dlmdomain.c qn->qn_nodes[count].ni_ipv4_port = node->nd_ipv4_port; node 1238 fs/ocfs2/dlm/dlmdomain.c qn->qn_nodes[count].ni_ipv4_address = node->nd_ipv4_address; node 1239 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Node %3d, %pI4:%u\n", node->nd_num, node 1240 fs/ocfs2/dlm/dlmdomain.c &(node->nd_ipv4_address), ntohs(node->nd_ipv4_port)); node 1242 fs/ocfs2/dlm/dlmdomain.c o2nm_node_put(node); node 1353 fs/ocfs2/dlm/dlmdomain.c unsigned int node) node 1364 fs/ocfs2/dlm/dlmdomain.c &cancel_msg, sizeof(cancel_msg), node, node 1369 fs/ocfs2/dlm/dlmdomain.c node); node 1383 fs/ocfs2/dlm/dlmdomain.c int node; node 1394 fs/ocfs2/dlm/dlmdomain.c node = -1; node 1395 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES, node 1396 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 1397 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 1400 fs/ocfs2/dlm/dlmdomain.c tmpstat = dlm_send_one_join_cancel(dlm, node); node 1403 fs/ocfs2/dlm/dlmdomain.c "node %d\n", tmpstat, node); node 1415 fs/ocfs2/dlm/dlmdomain.c int node, node 1423 fs/ocfs2/dlm/dlmdomain.c mlog(0, "querying node %d\n", node); node 1436 fs/ocfs2/dlm/dlmdomain.c sizeof(join_msg), node, &join_resp); node 1440 fs/ocfs2/dlm/dlmdomain.c node); node 1468 fs/ocfs2/dlm/dlmdomain.c node); node 1478 fs/ocfs2/dlm/dlmdomain.c node, node 1487 fs/ocfs2/dlm/dlmdomain.c packet.code, node); node 1494 fs/ocfs2/dlm/dlmdomain.c mlog(0, "status %d, node %d response is %d\n", status, node, node 1502 fs/ocfs2/dlm/dlmdomain.c unsigned int node) node 1508 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending join assert to node %u\n", node); node 1516 fs/ocfs2/dlm/dlmdomain.c &assert_msg, sizeof(assert_msg), node, node 1521 fs/ocfs2/dlm/dlmdomain.c node); node 1531 fs/ocfs2/dlm/dlmdomain.c int status, node, live; node 1534 fs/ocfs2/dlm/dlmdomain.c node = -1; node 1535 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES, node 1536 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 1537 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 1544 fs/ocfs2/dlm/dlmdomain.c status = dlm_send_one_join_assert(dlm, node); node 1547 fs/ocfs2/dlm/dlmdomain.c live = test_bit(node, dlm->live_nodes_map); node 1552 fs/ocfs2/dlm/dlmdomain.c "join on node %d\n", status, node); node 1593 fs/ocfs2/dlm/dlmdomain.c int status = 0, tmpstat, node; node 1618 fs/ocfs2/dlm/dlmdomain.c node = -1; node 1619 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(ctxt->live_map, O2NM_MAX_NODES, node 1620 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 1621 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 1624 fs/ocfs2/dlm/dlmdomain.c status = dlm_request_join(dlm, node, &response); node 1633 fs/ocfs2/dlm/dlmdomain.c set_bit(node, ctxt->yes_resp_map); node 49 fs/ocfs2/dlm/dlmlock.c u8 node, u64 cookie); node 117 fs/ocfs2/dlm/dlmlock.c lock->ml.node != dlm->node_num) { node 146 fs/ocfs2/dlm/dlmlock.c lock->ml.node); node 157 fs/ocfs2/dlm/dlmlock.c lock->ml.node); node 379 fs/ocfs2/dlm/dlmlock.c u8 node, u64 cookie) node 388 fs/ocfs2/dlm/dlmlock.c newlock->ml.node = node; node 407 fs/ocfs2/dlm/dlmlock.c struct dlm_lock * dlm_new_lock(int type, u8 node, u64 cookie, node 427 fs/ocfs2/dlm/dlmlock.c dlm_init_lock(lock, type, node, cookie); node 42 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, node 46 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, node 362 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, int idx) node 376 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, int idx) node 1215 fs/ocfs2/dlm/dlmmaster.c int node; node 1224 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc); node 1225 fs/ocfs2/dlm/dlmmaster.c while (node >= 0) { node 1230 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "node %d up while restarting\n", node); node 1234 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->response_map); node 1235 fs/ocfs2/dlm/dlmmaster.c set_bit(node, mle->vote_map); node 1237 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "node down! %d\n", node); node 1243 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->maybe_map); node 1245 fs/ocfs2/dlm/dlmmaster.c if (node == lowest) { node 1248 fs/ocfs2/dlm/dlmmaster.c "waiting on it!\n", node); node 1292 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc); node 2260 fs/ocfs2/dlm/dlmmaster.c u8 node; node 2271 fs/ocfs2/dlm/dlmmaster.c node = deref->node_idx; node 2278 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node); node 2299 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) { node 2300 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(dlm, res, node); node 2312 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node); node 2328 fs/ocfs2/dlm/dlmmaster.c item->u.dl.deref_node = node; node 2355 fs/ocfs2/dlm/dlmmaster.c u8 node; node 2363 fs/ocfs2/dlm/dlmmaster.c node = deref->node_idx; node 2370 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node); node 2391 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node); node 2411 fs/ocfs2/dlm/dlmmaster.c struct dlm_lock_resource *res, u8 node) node 2428 fs/ocfs2/dlm/dlmmaster.c &deref, sizeof(deref), node, &r); node 2432 fs/ocfs2/dlm/dlmmaster.c lockname, ret, node); node 2436 fs/ocfs2/dlm/dlmmaster.c dlm->name, namelen, lockname, node, r); node 2445 fs/ocfs2/dlm/dlmmaster.c u8 node; node 2450 fs/ocfs2/dlm/dlmmaster.c node = item->u.dl.deref_node; node 2455 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) { node 2456 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(dlm, res, node); node 2461 fs/ocfs2/dlm/dlmmaster.c dlm_drop_lockres_ref_done(dlm, res, node); node 2465 fs/ocfs2/dlm/dlmmaster.c dlm->name, res->lockname.len, res->lockname.name, node); node 2470 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node); node 2510 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { node 2940 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { node 2942 fs/ocfs2/dlm/dlmmaster.c lock->ml.node); node 2949 fs/ocfs2/dlm/dlmmaster.c lock->ml.node); node 2997 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node == dlm->node_num) node 2999 fs/ocfs2/dlm/dlmmaster.c if (test_bit(lock->ml.node, dlm->exit_domain_map)) node 3001 fs/ocfs2/dlm/dlmmaster.c nodenum = lock->ml.node; node 326 fs/ocfs2/dlm/dlmrecovery.c int dlm_is_node_dead(struct dlm_ctxt *dlm, u8 node) node 330 fs/ocfs2/dlm/dlmrecovery.c dead = !test_bit(node, dlm->domain_map); node 337 fs/ocfs2/dlm/dlmrecovery.c static int dlm_is_node_recovered(struct dlm_ctxt *dlm, u8 node) node 341 fs/ocfs2/dlm/dlmrecovery.c recovered = !test_bit(node, dlm->recovery_map); node 347 fs/ocfs2/dlm/dlmrecovery.c void dlm_wait_for_node_death(struct dlm_ctxt *dlm, u8 node, int timeout) node 349 fs/ocfs2/dlm/dlmrecovery.c if (dlm_is_node_dead(dlm, node)) node 353 fs/ocfs2/dlm/dlmrecovery.c "domain %s\n", node, dlm->name); node 357 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node), node 361 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node)); node 364 fs/ocfs2/dlm/dlmrecovery.c void dlm_wait_for_node_recovery(struct dlm_ctxt *dlm, u8 node, int timeout) node 366 fs/ocfs2/dlm/dlmrecovery.c if (dlm_is_node_recovered(dlm, node)) node 370 fs/ocfs2/dlm/dlmrecovery.c "domain %s\n", node, dlm->name); node 374 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node), node 378 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node)); node 1058 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 1209 fs/ocfs2/dlm/dlmrecovery.c lock->ml.node); node 1232 fs/ocfs2/dlm/dlmrecovery.c ml->node = lock->ml.node; node 1250 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.node = dlm->node_num; node 1263 fs/ocfs2/dlm/dlmrecovery.c *nodenum = ml->node; node 1836 fs/ocfs2/dlm/dlmrecovery.c if (ml->node == dlm->node_num) { node 1865 fs/ocfs2/dlm/dlmrecovery.c ml->node, ml->list, ml->flags, ml->type, node 1871 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node != ml->node) { node 1878 fs/ocfs2/dlm/dlmrecovery.c lock->ml.node); node 1885 fs/ocfs2/dlm/dlmrecovery.c ml->node, ml->list, ml->flags, ml->type, node 1917 fs/ocfs2/dlm/dlmrecovery.c newlock = dlm_new_lock(ml->type, ml->node, node 2011 fs/ocfs2/dlm/dlmrecovery.c ml->type, ml->convert_type, ml->node, node 2034 fs/ocfs2/dlm/dlmrecovery.c res->lockname.len, res->lockname.name, ml->node); node 2035 fs/ocfs2/dlm/dlmrecovery.c dlm_lockres_set_refmap_bit(dlm, res, ml->node); node 2244 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == search_node) { node 2279 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2288 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2297 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2364 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2488 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data) node 2509 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data) node 42 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock) ((lock)->ml.node != (dlm)->node_num) node 419 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type, target->ml.node); node 480 fs/ocfs2/dlm/dlmthread.c target->ml.type, target->ml.node); node 592 fs/ocfs2/dlm/dlmthread.c lock->ml.type, lock->ml.node); node 601 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { node 651 fs/ocfs2/dlm/dlmthread.c hi, lock->ml.node); node 653 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { node 468 fs/ocfs2/dlm/dlmunlock.c lock->ml.node == unlock->node_idx) { node 530 fs/ocfs2/refcounttree.c struct rb_node *node; node 534 fs/ocfs2/refcounttree.c while ((node = rb_last(root)) != NULL) { node 535 fs/ocfs2/refcounttree.c tree = rb_entry(node, struct ocfs2_refcount_tree, rf_node); node 81 fs/ocfs2/reservations.c struct rb_node *node; node 88 fs/ocfs2/reservations.c node = rb_first(&resmap->m_reservations); node 89 fs/ocfs2/reservations.c while (node) { node 90 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); node 97 fs/ocfs2/reservations.c node = rb_next(node); node 140 fs/ocfs2/reservations.c struct rb_node *node; node 143 fs/ocfs2/reservations.c node = rb_first(&resmap->m_reservations); node 144 fs/ocfs2/reservations.c while (node) { node 145 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); node 175 fs/ocfs2/reservations.c node = rb_next(node); node 274 fs/ocfs2/reservations.c struct rb_node *node; node 279 fs/ocfs2/reservations.c while ((node = rb_last(&resmap->m_reservations)) != NULL) { node 280 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); node 363 fs/ocfs2/reservations.c struct rb_node *node = resmap->m_reservations.rb_node; node 367 fs/ocfs2/reservations.c if (!node) node 370 fs/ocfs2/reservations.c node = rb_first(&resmap->m_reservations); node 371 fs/ocfs2/reservations.c while (node) { node 372 fs/ocfs2/reservations.c resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); node 384 fs/ocfs2/reservations.c node = rb_next(node); node 394 fs/ocfs2/stack_o2cb.c unsigned int *node) node 405 fs/ocfs2/stack_o2cb.c *node = node_num; node 463 fs/ocfs2/stackglue.c unsigned int *node) node 465 fs/ocfs2/stackglue.c return active_stack->sp_ops->this_node(conn, node); node 153 fs/ocfs2/stackglue.h unsigned int *node); node 264 fs/ocfs2/stackglue.h unsigned int *node); node 134 fs/ocfs2/uptodate.c struct rb_node *node; node 137 fs/ocfs2/uptodate.c while ((node = rb_last(root)) != NULL) { node 138 fs/ocfs2/uptodate.c item = rb_entry(node, struct ocfs2_meta_cache_item, c_node); node 33 fs/openpromfs/inode.c struct device_node *node; node 195 fs/openpromfs/inode.c dp = oi->u.node; node 210 fs/openpromfs/inode.c ent_data.node = child; node 269 fs/openpromfs/inode.c struct device_node *dp = oi->u.node; node 399 fs/openpromfs/inode.c oi->u.node = of_find_node_by_path("/"); node 228 fs/orangefs/orangefs-kernel.h struct hlist_node node; node 135 fs/orangefs/super.c hash_for_each_safe(orangefs_inode->xattr_cache, i, tmp, cx, node) { node 136 fs/orangefs/super.c hlist_del(&cx->node); node 72 fs/orangefs/xattr.c hlist_for_each_entry_safe(cx, tmp, h, node) { node 178 fs/orangefs/xattr.c hash_add(orangefs_inode->xattr_cache, &cx->node, node 232 fs/orangefs/xattr.c hash_add(orangefs_inode->xattr_cache, &cx->node, node 295 fs/orangefs/xattr.c hlist_for_each_entry_safe(cx, tmp, h, node) { node 297 fs/orangefs/xattr.c hlist_del(&cx->node); node 386 fs/orangefs/xattr.c hlist_for_each_entry_safe(cx, tmp, h, node) { node 388 fs/orangefs/xattr.c hlist_del(&cx->node); node 24 fs/overlayfs/readdir.c struct rb_node node; node 63 fs/overlayfs/readdir.c return rb_entry(n, struct ovl_cache_entry, node); node 81 fs/overlayfs/readdir.c newp = &tmp->node.rb_right; node 83 fs/overlayfs/readdir.c newp = &tmp->node.rb_left; node 95 fs/overlayfs/readdir.c struct rb_node *node = root->rb_node; node 98 fs/overlayfs/readdir.c while (node) { node 99 fs/overlayfs/readdir.c struct ovl_cache_entry *p = ovl_cache_entry_from_node(node); node 103 fs/overlayfs/readdir.c node = p->node.rb_right; node 105 fs/overlayfs/readdir.c node = p->node.rb_left; node 191 fs/overlayfs/readdir.c rb_link_node(&p->node, parent, newp); node 192 fs/overlayfs/readdir.c rb_insert_color(&p->node, rdd->root); node 589 fs/overlayfs/readdir.c rb_link_node(&p->node, parent, newp); node 590 fs/overlayfs/readdir.c rb_insert_color(&p->node, root); node 72 fs/proc/generic.c struct rb_node *node = dir->subdir.rb_node; node 74 fs/proc/generic.c while (node) { node 75 fs/proc/generic.c struct proc_dir_entry *de = rb_entry(node, node 81 fs/proc/generic.c node = node->rb_left; node 83 fs/proc/generic.c node = node->rb_right; node 114 fs/proc/proc_sysctl.c struct rb_node *node = dir->root.rb_node; node 116 fs/proc/proc_sysctl.c while (node) node 122 fs/proc/proc_sysctl.c ctl_node = rb_entry(node, struct ctl_node, node); node 124 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; node 129 fs/proc/proc_sysctl.c node = node->rb_left; node 131 fs/proc/proc_sysctl.c node = node->rb_right; node 142 fs/proc/proc_sysctl.c struct rb_node *node = &head->node[entry - head->ctl_table].node; node 156 fs/proc/proc_sysctl.c parent_node = rb_entry(parent, struct ctl_node, node); node 158 fs/proc/proc_sysctl.c parent_entry = &parent_head->ctl_table[parent_node - parent_head->node]; node 174 fs/proc/proc_sysctl.c rb_link_node(node, parent, p); node 175 fs/proc/proc_sysctl.c rb_insert_color(node, &head->parent->root); node 181 fs/proc/proc_sysctl.c struct rb_node *node = &head->node[entry - head->ctl_table].node; node 183 fs/proc/proc_sysctl.c rb_erase(node, &head->parent->root); node 188 fs/proc/proc_sysctl.c struct ctl_node *node, struct ctl_table *table) node 199 fs/proc/proc_sysctl.c head->node = node; node 201 fs/proc/proc_sysctl.c if (node) { node 203 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++, node++) node 204 fs/proc/proc_sysctl.c node->header = head; node 274 fs/proc/proc_sysctl.c struct hlist_node *node; node 279 fs/proc/proc_sysctl.c node = hlist_first_rcu(&head->inodes); node 280 fs/proc/proc_sysctl.c if (!node) node 282 fs/proc/proc_sysctl.c ei = hlist_entry(node, struct proc_inode, sysctl_inodes); node 384 fs/proc/proc_sysctl.c static struct ctl_node *first_usable_entry(struct rb_node *node) node 388 fs/proc/proc_sysctl.c for (;node; node = rb_next(node)) { node 389 fs/proc/proc_sysctl.c ctl_node = rb_entry(node, struct ctl_node, node); node 408 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; node 418 fs/proc/proc_sysctl.c struct ctl_node *ctl_node = &head->node[entry - head->ctl_table]; node 423 fs/proc/proc_sysctl.c ctl_node = first_usable_entry(rb_next(&ctl_node->node)); node 428 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; node 969 fs/proc/proc_sysctl.c struct ctl_node *node; node 978 fs/proc/proc_sysctl.c node = (struct ctl_node *)(new + 1); node 979 fs/proc/proc_sysctl.c table = (struct ctl_table *)(node + 1); node 985 fs/proc/proc_sysctl.c init_header(&new->header, set->dir.header.root, set, node, table); node 1167 fs/proc/proc_sysctl.c struct ctl_node *node; node 1187 fs/proc/proc_sysctl.c node = (struct ctl_node *)(links + 1); node 1188 fs/proc/proc_sysctl.c link_table = (struct ctl_table *)(node + nr_entries); node 1199 fs/proc/proc_sysctl.c init_header(links, dir->header.root, dir->header.set, node, link_table); node 1325 fs/proc/proc_sysctl.c struct ctl_node *node; node 1336 fs/proc/proc_sysctl.c node = (struct ctl_node *)(header + 1); node 1337 fs/proc/proc_sysctl.c init_header(header, root, set, node, table); node 1659 fs/proc/task_mmu.c unsigned long node[MAX_NUMNODES]; node 1691 fs/proc/task_mmu.c md->node[page_to_nid(page)] += nr_pages; node 1884 fs/proc/task_mmu.c if (md->node[nid]) node 1885 fs/proc/task_mmu.c seq_printf(m, " N%d=%lu", nid, md->node[nid]); node 290 fs/quota/dquot.c struct hlist_node *node; node 293 fs/quota/dquot.c hlist_for_each (node, dquot_hash+hashent) { node 294 fs/quota/dquot.c dquot = hlist_entry(node, struct dquot, dq_hash); node 936 fs/seq_file.c struct hlist_node *node; node 938 fs/seq_file.c hlist_for_each(node, head) node 940 fs/seq_file.c return node; node 973 fs/seq_file.c struct hlist_node *node = v; node 979 fs/seq_file.c return node->next; node 997 fs/seq_file.c struct hlist_node *node; node 999 fs/seq_file.c __hlist_for_each_rcu(node, head) node 1001 fs/seq_file.c return node; node 1044 fs/seq_file.c struct hlist_node *node = v; node 1050 fs/seq_file.c return rcu_dereference(node->next); node 1065 fs/seq_file.c struct hlist_node *node; node 1068 fs/seq_file.c hlist_for_each(node, per_cpu_ptr(head, *cpu)) { node 1070 fs/seq_file.c return node; node 1090 fs/seq_file.c struct hlist_node *node = v; node 1094 fs/seq_file.c if (node->next) node 1095 fs/seq_file.c return node->next; node 292 fs/super.c WARN_ON(s->s_dentry_lru.node); node 293 fs/super.c WARN_ON(s->s_inode_lru.node); node 30 fs/ubifs/auth.c int __ubifs_node_calc_hash(const struct ubifs_info *c, const void *node, node 33 fs/ubifs/auth.c const struct ubifs_ch *ch = node; node 39 fs/ubifs/auth.c err = crypto_shash_digest(shash, node, le32_to_cpu(ch->len), hash); node 78 fs/ubifs/auth.c int ubifs_prepare_auth_node(struct ubifs_info *c, void *node, node 81 fs/ubifs/auth.c struct ubifs_auth_node *auth = node; node 152 fs/ubifs/auth.c void ubifs_bad_hash(const struct ubifs_info *c, const void *node, const u8 *hash, node 161 fs/ubifs/auth.c __ubifs_node_calc_hash(c, node, calc); node 178 fs/ubifs/auth.c int __ubifs_node_check_hash(const struct ubifs_info *c, const void *node, node 184 fs/ubifs/auth.c err = __ubifs_node_calc_hash(c, node, calc); node 234 fs/ubifs/auth.c signode = snod->node; node 401 fs/ubifs/auth.c static int ubifs_node_calc_hmac(const struct ubifs_info *c, const void *node, node 418 fs/ubifs/auth.c err = crypto_shash_update(shash, node + 8, ofs_hmac - 8); node 424 fs/ubifs/auth.c err = crypto_shash_update(shash, node + ofs_hmac + hmac_len, node 445 fs/ubifs/auth.c int __ubifs_node_insert_hmac(const struct ubifs_info *c, void *node, int len, node 448 fs/ubifs/auth.c return ubifs_node_calc_hmac(c, node, len, ofs_hmac, node + ofs_hmac); node 461 fs/ubifs/auth.c int __ubifs_node_verify_hmac(const struct ubifs_info *c, const void *node, node 472 fs/ubifs/auth.c err = ubifs_node_calc_hmac(c, node, len, ofs_hmac, hmac); node 478 fs/ubifs/auth.c err = crypto_memneq(hmac, node + ofs_hmac, hmac_len); node 294 fs/ubifs/debug.c void ubifs_dump_node(const struct ubifs_info *c, const void *node) node 298 fs/ubifs/debug.c const struct ubifs_ch *ch = node; node 305 fs/ubifs/debug.c (void *)node, UBIFS_CH_SZ, 1); node 310 fs/ubifs/debug.c dump_ch(node); node 315 fs/ubifs/debug.c const struct ubifs_pad_node *pad = node; node 322 fs/ubifs/debug.c const struct ubifs_sb_node *sup = node; node 359 fs/ubifs/debug.c const struct ubifs_mst_node *mst = node; node 401 fs/ubifs/debug.c const struct ubifs_ref_node *ref = node; node 410 fs/ubifs/debug.c const struct ubifs_ino_node *ino = node; node 444 fs/ubifs/debug.c const struct ubifs_dent_node *dent = node; node 469 fs/ubifs/debug.c const struct ubifs_data_node *dn = node; node 486 fs/ubifs/debug.c const struct ubifs_trun_node *trun = node; node 497 fs/ubifs/debug.c const struct ubifs_idx_node *idx = node; node 521 fs/ubifs/debug.c const struct ubifs_orph_node *orph = node; node 806 fs/ubifs/debug.c ubifs_dump_node(c, snod->node); node 837 fs/ubifs/debug.c ubifs_dump_node(c, snod->node); node 1979 fs/ubifs/debug.c void *node; node 1990 fs/ubifs/debug.c node = kmalloc(zbr->len, GFP_NOFS); node 1991 fs/ubifs/debug.c if (!node) node 1994 fs/ubifs/debug.c err = ubifs_tnc_read_node(c, zbr, node); node 2003 fs/ubifs/debug.c fscki = add_inode(c, priv, node); node 2020 fs/ubifs/debug.c ch = node; node 2030 fs/ubifs/debug.c struct ubifs_data_node *dn = node; node 2059 fs/ubifs/debug.c struct ubifs_dent_node *dent = node; node 2107 fs/ubifs/debug.c kfree(node); node 2112 fs/ubifs/debug.c ubifs_dump_node(c, node); node 2114 fs/ubifs/debug.c kfree(node); node 2316 fs/ubifs/debug.c ubifs_dump_node(c, sa->node); node 2321 fs/ubifs/debug.c ubifs_dump_node(c, sb->node); node 2352 fs/ubifs/debug.c ubifs_dump_node(c, sa->node); node 2353 fs/ubifs/debug.c ubifs_dump_node(c, sb->node); node 2384 fs/ubifs/debug.c ubifs_dump_node(c, sa->node); node 2390 fs/ubifs/debug.c ubifs_dump_node(c, sb->node); node 2440 fs/ubifs/debug.c ubifs_dump_node(c, sa->node); node 2442 fs/ubifs/debug.c ubifs_dump_node(c, sb->node); node 245 fs/ubifs/debug.h void ubifs_dump_node(const struct ubifs_info *c, const void *node); node 312 fs/ubifs/gc.c err = ubifs_wbuf_write_nolock(wbuf, snod->node, snod->len); node 371 fs/ubifs/gc.c snod->node, snod->len); node 403 fs/ubifs/gc.c snod->node, snod->len); node 554 fs/ubifs/gc.c struct ubifs_idx_node *idx = snod->node; node 356 fs/ubifs/io.c void ubifs_init_node(struct ubifs_info *c, void *node, int len, int pad) node 358 fs/ubifs/io.c struct ubifs_ch *ch = node; node 372 fs/ubifs/io.c ubifs_pad(c, node + len, pad); node 376 fs/ubifs/io.c void ubifs_crc_node(struct ubifs_info *c, void *node, int len) node 378 fs/ubifs/io.c struct ubifs_ch *ch = node; node 381 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); node 400 fs/ubifs/io.c int ubifs_prepare_node_hmac(struct ubifs_info *c, void *node, int len, node 405 fs/ubifs/io.c ubifs_init_node(c, node, len, pad); node 408 fs/ubifs/io.c err = ubifs_node_insert_hmac(c, node, len, hmac_offs); node 413 fs/ubifs/io.c ubifs_crc_node(c, node, len); node 429 fs/ubifs/io.c void ubifs_prepare_node(struct ubifs_info *c, void *node, int len, int pad) node 435 fs/ubifs/io.c ubifs_prepare_node_hmac(c, node, len, 0, pad); node 448 fs/ubifs/io.c void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last) node 451 fs/ubifs/io.c struct ubifs_ch *ch = node; node 464 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); node 225 fs/ubifs/journal.c static int ubifs_hash_nodes(struct ubifs_info *c, void *node, node 232 fs/ubifs/journal.c const struct ubifs_ch *ch = node; node 243 fs/ubifs/journal.c err = ubifs_shash_update(c, hash, (void *)node, nodelen); node 247 fs/ubifs/journal.c node += ALIGN(nodelen, 8); node 251 fs/ubifs/journal.c return ubifs_prepare_auth_node(c, node, hash); node 612 fs/ubifs/log.c void *node) node 614 fs/ubifs/log.c struct ubifs_ch *ch = node; node 627 fs/ubifs/log.c memcpy(buf + *offs, node, len); node 666 fs/ubifs/log.c struct ubifs_ref_node *ref = snod->node; node 674 fs/ubifs/log.c &offs, snod->node); node 684 fs/ubifs/log.c snod->node); node 1128 fs/ubifs/lprops.c struct ubifs_idx_node *idx = snod->node; node 69 fs/ubifs/master.c const void *node = mst; node 75 fs/ubifs/master.c crypto_shash_digest(shash, node + sizeof(struct ubifs_ch), node 110 fs/ubifs/master.c memcpy(c->mst_node, snod->node, snod->len); node 129 fs/ubifs/master.c if (ubifs_compare_master_node(c, c->mst_node, snod->node)) node 230 fs/ubifs/misc.h const union ubifs_key *key, void *node) node 232 fs/ubifs/misc.h return ubifs_tnc_locate(c, key, node, NULL, NULL); node 642 fs/ubifs/orphan.c ubifs_dump_node(c, snod->node); node 647 fs/ubifs/orphan.c orph = snod->node; node 670 fs/ubifs/orphan.c ubifs_dump_node(c, snod->node); node 841 fs/ubifs/orphan.c struct ubifs_ino_node *node; node 925 fs/ubifs/orphan.c err = ubifs_tnc_read_node(c, zbr, ci->node); node 930 fs/ubifs/orphan.c if (ci->node->nlink == 0) node 954 fs/ubifs/orphan.c orph = snod->node; node 1013 fs/ubifs/orphan.c ci.node = kmalloc(UBIFS_MAX_INO_NODE_SZ, GFP_NOFS); node 1014 fs/ubifs/orphan.c if (!ci.node) { node 1041 fs/ubifs/orphan.c kfree(ci.node); node 570 fs/ubifs/recovery.c ch = snod->node; node 615 fs/ubifs/replay.c struct ubifs_auth_node *auth = snod->node; node 632 fs/ubifs/replay.c err = crypto_shash_update(log_hash, snod->node, node 737 fs/ubifs/replay.c ubifs_node_calc_hash(c, snod->node, hash); node 745 fs/ubifs/replay.c struct ubifs_ino_node *ino = snod->node; node 757 fs/ubifs/replay.c struct ubifs_data_node *dn = snod->node; node 770 fs/ubifs/replay.c struct ubifs_dent_node *dent = snod->node; node 784 fs/ubifs/replay.c struct ubifs_trun_node *trun = snod->node; node 838 fs/ubifs/replay.c ubifs_dump_node(c, snod->node); node 994 fs/ubifs/replay.c const struct ubifs_cs_node *node; node 1016 fs/ubifs/replay.c node = sleb->buf; node 1031 fs/ubifs/replay.c if (le64_to_cpu(node->cmt_no) != c->cmt_no) { node 1034 fs/ubifs/replay.c (unsigned long long)le64_to_cpu(node->cmt_no), node 1039 fs/ubifs/replay.c c->cs_sqnum = le64_to_cpu(node->ch.sqnum); node 1046 fs/ubifs/replay.c err = ubifs_shash_update(c, c->log_hash, node, UBIFS_CS_NODE_SZ); node 1088 fs/ubifs/replay.c const struct ubifs_ref_node *ref = snod->node; node 1136 fs/ubifs/replay.c ubifs_dump_node(c, snod->node); node 196 fs/ubifs/scan.c snod->node = buf; node 356 fs/ubifs/scan.c struct ubifs_scan_node *node; node 361 fs/ubifs/scan.c node = list_entry(head->next, struct ubifs_scan_node, list); node 362 fs/ubifs/scan.c list_del(&node->list); node 363 fs/ubifs/scan.c kfree(node); node 28 fs/ubifs/tnc.c struct ubifs_zbranch *zbr, void *node); node 306 fs/ubifs/tnc.c const void *node) node 310 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node; node 323 fs/ubifs/tnc.c lnc_node = kmemdup(node, zbr->len, GFP_NOFS); node 342 fs/ubifs/tnc.c void *node) node 349 fs/ubifs/tnc.c err = ubifs_validate_entry(c, node); node 352 fs/ubifs/tnc.c ubifs_dump_node(c, node); node 356 fs/ubifs/tnc.c zbr->leaf = node; node 385 fs/ubifs/tnc.c void *node) node 394 fs/ubifs/tnc.c memcpy(node, zbr->leaf, zbr->len); node 399 fs/ubifs/tnc.c err = fallible_read_node(c, &zbr->key, zbr, node); node 409 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zbr, node); node 415 fs/ubifs/tnc.c err = lnc_add(c, zbr, node); node 498 fs/ubifs/tnc.c struct ubifs_zbranch *zbr, void *node) node 504 fs/ubifs/tnc.c ret = try_read_node(c, node, key_type(c, key), zbr); node 507 fs/ubifs/tnc.c struct ubifs_dent_node *dent = node; node 1442 fs/ubifs/tnc.c void *node, int *lnum, int *offs) node 1468 fs/ubifs/tnc.c err = tnc_read_hashed_node(c, zt, node); node 1472 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zt, node); node 1482 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, &zbr, node); node 1486 fs/ubifs/tnc.c err = fallible_read_node(c, key, &zbr, node); node 1810 fs/ubifs/tnc.c void *node, const struct fscrypt_name *nm) node 1837 fs/ubifs/tnc.c err = tnc_read_hashed_node(c, &znode->zbranch[n], node); node 1858 fs/ubifs/tnc.c void *node, const struct fscrypt_name *nm) node 1861 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node; node 1867 fs/ubifs/tnc.c err = ubifs_tnc_lookup(c, key, node); node 1880 fs/ubifs/tnc.c return do_lookup_nm(c, key, node, nm); node 1961 fs/ubifs/tnc.c void *node, uint32_t cookie) node 1964 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node; node 1973 fs/ubifs/tnc.c err = ubifs_tnc_lookup(c, key, node); node 1984 fs/ubifs/tnc.c return do_lookup_dh(c, key, node, cookie); node 256 fs/ubifs/tnc_commit.c idx = snod->node; node 462 fs/ubifs/tnc_misc.c void *node) node 474 fs/ubifs/tnc_misc.c err = ubifs_read_node_wbuf(wbuf, node, type, zbr->len, node 477 fs/ubifs/tnc_misc.c err = ubifs_read_node(c, node, type, zbr->len, zbr->lnum, node 486 fs/ubifs/tnc_misc.c key_read(c, node + UBIFS_KEY_OFFSET, &key1); node 492 fs/ubifs/tnc_misc.c ubifs_dump_node(c, node); node 496 fs/ubifs/tnc_misc.c err = ubifs_node_check_hash(c, node, zbr->hash); node 498 fs/ubifs/tnc_misc.c ubifs_bad_hash(c, node, zbr->hash, zbr->lnum, zbr->offs); node 308 fs/ubifs/ubifs.h void *node; node 1561 fs/ubifs/ubifs.h int ubifs_prepare_auth_node(struct ubifs_info *c, void *node, node 1594 fs/ubifs/ubifs.h void ubifs_bad_hash(const struct ubifs_info *c, const void *node, node 1717 fs/ubifs/ubifs.h int ubifs_write_node(struct ubifs_info *c, void *node, int len, int lnum, node 1726 fs/ubifs/ubifs.h int ubifs_prepare_node_hmac(struct ubifs_info *c, void *node, int len, node 1728 fs/ubifs/ubifs.h void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last); node 1821 fs/ubifs/ubifs.h void *node, const struct fscrypt_name *nm); node 1823 fs/ubifs/ubifs.h void *node, uint32_t secondary_hash); node 1825 fs/ubifs/ubifs.h void *node, int *lnum, int *offs); node 1873 fs/ubifs/ubifs.h void *node); node 366 fs/unicode/mkutf8data.c struct node *parent; node 382 fs/unicode/mkutf8data.c struct node *node; node 385 fs/unicode/mkutf8data.c node = tree->root; node 386 fs/unicode/mkutf8data.c while (!leaf && node) { node 387 fs/unicode/mkutf8data.c if (node->nextbyte) node 389 fs/unicode/mkutf8data.c if (*key & (1 << (node->bitnum & 7))) { node 391 fs/unicode/mkutf8data.c if (node->rightnode == NODE) { node 392 fs/unicode/mkutf8data.c node = node->right; node 393 fs/unicode/mkutf8data.c } else if (node->rightnode == LEAF) { node 394 fs/unicode/mkutf8data.c leaf = node->right; node 396 fs/unicode/mkutf8data.c node = NULL; node 400 fs/unicode/mkutf8data.c if (node->leftnode == NODE) { node 401 fs/unicode/mkutf8data.c node = node->left; node 402 fs/unicode/mkutf8data.c } else if (node->leftnode == LEAF) { node 403 fs/unicode/mkutf8data.c leaf = node->left; node 405 fs/unicode/mkutf8data.c node = NULL; node 419 fs/unicode/mkutf8data.c struct node *node; node 435 fs/unicode/mkutf8data.c node = tree->root; node 437 fs/unicode/mkutf8data.c while (node) { node 440 fs/unicode/mkutf8data.c indent, "", node, node 441 fs/unicode/mkutf8data.c node->bitnum, node->nextbyte, node 442 fs/unicode/mkutf8data.c node->left, node->right, node 443 fs/unicode/mkutf8data.c node->keymask, node->keybits); node 445 fs/unicode/mkutf8data.c if (!(node->left && node->right)) node 448 fs/unicode/mkutf8data.c while (node) { node 449 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 452 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 453 fs/unicode/mkutf8data.c assert(node->left); node 454 fs/unicode/mkutf8data.c tree->leaf_print(node->left, node 457 fs/unicode/mkutf8data.c } else if (node->left) { node 458 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 460 fs/unicode/mkutf8data.c node = node->left; node 466 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 467 fs/unicode/mkutf8data.c assert(node->right); node 468 fs/unicode/mkutf8data.c tree->leaf_print(node->right, node 471 fs/unicode/mkutf8data.c } else if (node->right) { node 472 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 474 fs/unicode/mkutf8data.c node = node->right; node 480 fs/unicode/mkutf8data.c node = node->parent; node 492 fs/unicode/mkutf8data.c static struct node *alloc_node(struct node *parent) node 494 fs/unicode/mkutf8data.c struct node *node; node 497 fs/unicode/mkutf8data.c node = malloc(sizeof(*node)); node 498 fs/unicode/mkutf8data.c node->left = node->right = NULL; node 499 fs/unicode/mkutf8data.c node->parent = parent; node 500 fs/unicode/mkutf8data.c node->leftnode = NODE; node 501 fs/unicode/mkutf8data.c node->rightnode = NODE; node 502 fs/unicode/mkutf8data.c node->keybits = 0; node 503 fs/unicode/mkutf8data.c node->keymask = 0; node 504 fs/unicode/mkutf8data.c node->mark = 0; node 505 fs/unicode/mkutf8data.c node->index = 0; node 506 fs/unicode/mkutf8data.c node->offset = -1; node 507 fs/unicode/mkutf8data.c node->size = 4; node 509 fs/unicode/mkutf8data.c if (node->parent) { node 512 fs/unicode/mkutf8data.c node->bitnum = bitnum + 7 + 8; node 513 fs/unicode/mkutf8data.c node->nextbyte = 1; node 515 fs/unicode/mkutf8data.c node->bitnum = bitnum - 1; node 516 fs/unicode/mkutf8data.c node->nextbyte = 0; node 519 fs/unicode/mkutf8data.c node->bitnum = 7; node 520 fs/unicode/mkutf8data.c node->nextbyte = 0; node 523 fs/unicode/mkutf8data.c return node; node 535 fs/unicode/mkutf8data.c struct node *node; node 536 fs/unicode/mkutf8data.c struct node *parent; node 542 fs/unicode/mkutf8data.c node = NULL; node 549 fs/unicode/mkutf8data.c *cursor = alloc_node(node); node 550 fs/unicode/mkutf8data.c node = *cursor; node 551 fs/unicode/mkutf8data.c if (node->nextbyte) node 553 fs/unicode/mkutf8data.c if (*key & (1 << (node->bitnum & 7))) node 554 fs/unicode/mkutf8data.c cursor = &node->right; node 556 fs/unicode/mkutf8data.c cursor = &node->left; node 562 fs/unicode/mkutf8data.c while (node) { node 563 fs/unicode/mkutf8data.c if (*key & (1 << (node->bitnum & 7))) node 564 fs/unicode/mkutf8data.c node->rightnode = LEAF; node 566 fs/unicode/mkutf8data.c node->leftnode = LEAF; node 567 fs/unicode/mkutf8data.c if (node->nextbyte) node 569 fs/unicode/mkutf8data.c if (node->leftnode == NODE || node->rightnode == NODE) node 571 fs/unicode/mkutf8data.c assert(node->left); node 572 fs/unicode/mkutf8data.c assert(node->right); node 574 fs/unicode/mkutf8data.c if (! tree->leaf_equal(node->left, node->right)) node 577 fs/unicode/mkutf8data.c leaf = node->left; node 579 fs/unicode/mkutf8data.c parent = node->parent; node 584 fs/unicode/mkutf8data.c } else if (parent->left == node) { node 591 fs/unicode/mkutf8data.c parent->keymask |= (1 << node->bitnum); node 593 fs/unicode/mkutf8data.c } else if (parent->right == node) { node 600 fs/unicode/mkutf8data.c parent->keymask |= (1 << node->bitnum); node 601 fs/unicode/mkutf8data.c parent->keybits |= (1 << node->bitnum); node 607 fs/unicode/mkutf8data.c free(node); node 608 fs/unicode/mkutf8data.c node = parent; node 612 fs/unicode/mkutf8data.c while (node) { node 613 fs/unicode/mkutf8data.c parent = node->parent; node 617 fs/unicode/mkutf8data.c if (node->keymask == 0) { node 624 fs/unicode/mkutf8data.c assert((parent->keymask & node->keymask) == 0); node 625 fs/unicode/mkutf8data.c parent->keymask |= node->keymask; node 627 fs/unicode/mkutf8data.c parent->keybits |= node->keybits; node 631 fs/unicode/mkutf8data.c node = parent; node 656 fs/unicode/mkutf8data.c struct node *node; node 657 fs/unicode/mkutf8data.c struct node *left; node 658 fs/unicode/mkutf8data.c struct node *right; node 659 fs/unicode/mkutf8data.c struct node *parent; node 677 fs/unicode/mkutf8data.c node = tree->root; node 678 fs/unicode/mkutf8data.c while (node) { node 679 fs/unicode/mkutf8data.c if (node->nextbyte) node 681 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) node 683 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) node 685 fs/unicode/mkutf8data.c if (!node->left) node 687 fs/unicode/mkutf8data.c if (!node->right) node 689 fs/unicode/mkutf8data.c left = node->left; node 690 fs/unicode/mkutf8data.c right = node->right; node 733 fs/unicode/mkutf8data.c parent = node->parent; node 734 fs/unicode/mkutf8data.c left = node->left; node 735 fs/unicode/mkutf8data.c right = node->right; node 736 fs/unicode/mkutf8data.c if (parent->left == node) node 738 fs/unicode/mkutf8data.c else if (parent->right == node) node 743 fs/unicode/mkutf8data.c left->keymask |= (1 << node->bitnum); node 744 fs/unicode/mkutf8data.c node->left = NULL; node 745 fs/unicode/mkutf8data.c while (node) { node 746 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 749 fs/unicode/mkutf8data.c if (node->leftnode == NODE && node->left) { node 750 fs/unicode/mkutf8data.c left = node->left; node 751 fs/unicode/mkutf8data.c free(node); node 753 fs/unicode/mkutf8data.c node = left; node 754 fs/unicode/mkutf8data.c } else if (node->rightnode == NODE && node->right) { node 755 fs/unicode/mkutf8data.c right = node->right; node 756 fs/unicode/mkutf8data.c free(node); node 758 fs/unicode/mkutf8data.c node = right; node 760 fs/unicode/mkutf8data.c node = NULL; node 764 fs/unicode/mkutf8data.c node = parent; node 766 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 770 fs/unicode/mkutf8data.c if (node->left && node->right) node 772 fs/unicode/mkutf8data.c if (node->left) { node 773 fs/unicode/mkutf8data.c left = node->left; node 774 fs/unicode/mkutf8data.c node->keymask |= left->keymask; node 775 fs/unicode/mkutf8data.c node->keybits |= left->keybits; node 777 fs/unicode/mkutf8data.c if (node->right) { node 778 fs/unicode/mkutf8data.c right = node->right; node 779 fs/unicode/mkutf8data.c node->keymask |= right->keymask; node 780 fs/unicode/mkutf8data.c node->keybits |= right->keybits; node 782 fs/unicode/mkutf8data.c node->keymask |= (1 << node->bitnum); node 783 fs/unicode/mkutf8data.c node = node->parent; node 785 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 790 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 792 fs/unicode/mkutf8data.c node->leftnode == NODE && node 793 fs/unicode/mkutf8data.c node->left) { node 795 fs/unicode/mkutf8data.c node = node->left; node 797 fs/unicode/mkutf8data.c node->rightnode == NODE && node 798 fs/unicode/mkutf8data.c node->right) { node 800 fs/unicode/mkutf8data.c node = node->right; node 804 fs/unicode/mkutf8data.c node = node->parent; node 817 fs/unicode/mkutf8data.c struct node *node; node 818 fs/unicode/mkutf8data.c struct node *n; node 831 fs/unicode/mkutf8data.c node = tree->root; node 833 fs/unicode/mkutf8data.c while (node) { node 834 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 837 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 838 fs/unicode/mkutf8data.c assert(node->left); node 839 fs/unicode/mkutf8data.c if (tree->leaf_mark(node->left)) { node 840 fs/unicode/mkutf8data.c n = node; node 847 fs/unicode/mkutf8data.c } else if (node->left) { node 848 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 849 fs/unicode/mkutf8data.c node = node->left; node 855 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 856 fs/unicode/mkutf8data.c assert(node->right); node 857 fs/unicode/mkutf8data.c if (tree->leaf_mark(node->right)) { node 858 fs/unicode/mkutf8data.c n = node; node 865 fs/unicode/mkutf8data.c } else if (node->right) { node 866 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 867 fs/unicode/mkutf8data.c node = node->right; node 873 fs/unicode/mkutf8data.c node = node->parent; node 879 fs/unicode/mkutf8data.c node = tree->root; node 881 fs/unicode/mkutf8data.c while (node) { node 882 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 885 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 886 fs/unicode/mkutf8data.c assert(node->left); node 887 fs/unicode/mkutf8data.c if (tree->leaf_mark(node->left)) { node 888 fs/unicode/mkutf8data.c n = node; node 895 fs/unicode/mkutf8data.c } else if (node->left) { node 896 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 897 fs/unicode/mkutf8data.c node = node->left; node 898 fs/unicode/mkutf8data.c if (!node->mark && node->parent->mark) { node 900 fs/unicode/mkutf8data.c node->mark = 1; node 907 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 908 fs/unicode/mkutf8data.c assert(node->right); node 909 fs/unicode/mkutf8data.c if (tree->leaf_mark(node->right)) { node 910 fs/unicode/mkutf8data.c n = node; node 917 fs/unicode/mkutf8data.c } else if (node->right) { node 918 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 919 fs/unicode/mkutf8data.c node = node->right; node 920 fs/unicode/mkutf8data.c if (!node->mark && node->parent->mark && node 921 fs/unicode/mkutf8data.c !node->parent->left) { node 923 fs/unicode/mkutf8data.c node->mark = 1; node 930 fs/unicode/mkutf8data.c node = node->parent; node 944 fs/unicode/mkutf8data.c struct node *node; node 966 fs/unicode/mkutf8data.c node = tree->root; node 968 fs/unicode/mkutf8data.c while (node) { node 969 fs/unicode/mkutf8data.c if (!node->mark) node 972 fs/unicode/mkutf8data.c if (node->index != index) node 973 fs/unicode/mkutf8data.c node->index = index; node 974 fs/unicode/mkutf8data.c index += node->size; node 976 fs/unicode/mkutf8data.c while (node) { node 977 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 978 fs/unicode/mkutf8data.c if (node->mark && (leftmask & bitmask) == 0) { node 980 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 981 fs/unicode/mkutf8data.c assert(node->left); node 982 fs/unicode/mkutf8data.c *tree->leaf_index(tree, node->left) = node 984 fs/unicode/mkutf8data.c index += tree->leaf_size(node->left); node 986 fs/unicode/mkutf8data.c } else if (node->left) { node 987 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 989 fs/unicode/mkutf8data.c node = node->left; node 993 fs/unicode/mkutf8data.c if (node->mark && (rightmask & bitmask) == 0) { node 995 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 996 fs/unicode/mkutf8data.c assert(node->right); node 997 fs/unicode/mkutf8data.c *tree->leaf_index(tree, node->right) = index; node 998 fs/unicode/mkutf8data.c index += tree->leaf_size(node->right); node 1000 fs/unicode/mkutf8data.c } else if (node->right) { node 1001 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 1003 fs/unicode/mkutf8data.c node = node->right; node 1009 fs/unicode/mkutf8data.c node = node->parent; node 1025 fs/unicode/mkutf8data.c static int mark_subtree(struct node *node) node 1029 fs/unicode/mkutf8data.c if (!node || node->mark) node 1031 fs/unicode/mkutf8data.c node->mark = 1; node 1032 fs/unicode/mkutf8data.c node->index = node->parent->index; node 1034 fs/unicode/mkutf8data.c if (node->leftnode == NODE) node 1035 fs/unicode/mkutf8data.c changed += mark_subtree(node->left); node 1036 fs/unicode/mkutf8data.c if (node->rightnode == NODE) node 1037 fs/unicode/mkutf8data.c changed += mark_subtree(node->right); node 1051 fs/unicode/mkutf8data.c struct node *node; node 1052 fs/unicode/mkutf8data.c struct node *right; node 1053 fs/unicode/mkutf8data.c struct node *n; node 1077 fs/unicode/mkutf8data.c node = tree->root; node 1079 fs/unicode/mkutf8data.c while (node) { node 1080 fs/unicode/mkutf8data.c if (!node->mark) node 1083 fs/unicode/mkutf8data.c if (!node->left || !node->right) { node 1086 fs/unicode/mkutf8data.c if (node->rightnode == NODE) { node 1093 fs/unicode/mkutf8data.c right = node->right; node 1098 fs/unicode/mkutf8data.c while (n->bitnum != node->bitnum) { node 1112 fs/unicode/mkutf8data.c if (n->bitnum != node->bitnum) node 1121 fs/unicode/mkutf8data.c offset = right->index - node->index; node 1123 fs/unicode/mkutf8data.c offset = *tree->leaf_index(tree, node->right); node 1124 fs/unicode/mkutf8data.c offset -= node->index; node 1136 fs/unicode/mkutf8data.c if (node->size != size || node->offset != offset) { node 1137 fs/unicode/mkutf8data.c node->size = size; node 1138 fs/unicode/mkutf8data.c node->offset = offset; node 1142 fs/unicode/mkutf8data.c while (node) { node 1143 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 1145 fs/unicode/mkutf8data.c if (node->mark && (leftmask & bitmask) == 0) { node 1147 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 1148 fs/unicode/mkutf8data.c assert(node->left); node 1149 fs/unicode/mkutf8data.c } else if (node->left) { node 1150 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 1152 fs/unicode/mkutf8data.c node = node->left; node 1156 fs/unicode/mkutf8data.c if (node->mark && (rightmask & bitmask) == 0) { node 1159 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 1160 fs/unicode/mkutf8data.c assert(node->right); node 1161 fs/unicode/mkutf8data.c } else if (node->right) { node 1162 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 1164 fs/unicode/mkutf8data.c node = node->right; node 1172 fs/unicode/mkutf8data.c node = node->parent; node 1187 fs/unicode/mkutf8data.c struct node *node; node 1219 fs/unicode/mkutf8data.c node = tree->root; node 1221 fs/unicode/mkutf8data.c while (node) { node 1222 fs/unicode/mkutf8data.c if (!node->mark) node 1224 fs/unicode/mkutf8data.c assert(node->offset != -1); node 1225 fs/unicode/mkutf8data.c assert(node->index == index); node 1228 fs/unicode/mkutf8data.c if (node->nextbyte) node 1230 fs/unicode/mkutf8data.c byte |= (node->bitnum & BITNUM); node 1231 fs/unicode/mkutf8data.c if (node->left && node->right) { node 1232 fs/unicode/mkutf8data.c if (node->leftnode == NODE) node 1234 fs/unicode/mkutf8data.c if (node->rightnode == NODE) node 1236 fs/unicode/mkutf8data.c if (node->offset <= 0xff) node 1238 fs/unicode/mkutf8data.c else if (node->offset <= 0xffff) node 1243 fs/unicode/mkutf8data.c offset = node->offset; node 1252 fs/unicode/mkutf8data.c } else if (node->left) { node 1253 fs/unicode/mkutf8data.c if (node->leftnode == NODE) node 1258 fs/unicode/mkutf8data.c } else if (node->right) { node 1260 fs/unicode/mkutf8data.c if (node->rightnode == NODE) node 1269 fs/unicode/mkutf8data.c while (node) { node 1270 fs/unicode/mkutf8data.c bitmask = 1 << node->bitnum; node 1271 fs/unicode/mkutf8data.c if (node->mark && (leftmask & bitmask) == 0) { node 1273 fs/unicode/mkutf8data.c if (node->leftnode == LEAF) { node 1274 fs/unicode/mkutf8data.c assert(node->left); node 1275 fs/unicode/mkutf8data.c data = tree->leaf_emit(node->left, node 1277 fs/unicode/mkutf8data.c size = tree->leaf_size(node->left); node 1281 fs/unicode/mkutf8data.c } else if (node->left) { node 1282 fs/unicode/mkutf8data.c assert(node->leftnode == NODE); node 1284 fs/unicode/mkutf8data.c node = node->left; node 1288 fs/unicode/mkutf8data.c if (node->mark && (rightmask & bitmask) == 0) { node 1290 fs/unicode/mkutf8data.c if (node->rightnode == LEAF) { node 1291 fs/unicode/mkutf8data.c assert(node->right); node 1292 fs/unicode/mkutf8data.c data = tree->leaf_emit(node->right, node 1294 fs/unicode/mkutf8data.c size = tree->leaf_size(node->right); node 1298 fs/unicode/mkutf8data.c } else if (node->right) { node 1299 fs/unicode/mkutf8data.c assert(node->rightnode == NODE); node 1301 fs/unicode/mkutf8data.c node = node->right; node 1307 fs/unicode/mkutf8data.c node = node->parent; node 2713 fs/unicode/mkutf8data.c int node; node 2719 fs/unicode/mkutf8data.c node = 1; node 2721 fs/unicode/mkutf8data.c while (node) { node 2733 fs/unicode/mkutf8data.c node = (*trie & RIGHTNODE); node 2742 fs/unicode/mkutf8data.c node = (*trie & TRIENODE); node 2752 fs/unicode/mkutf8data.c node = (*trie & LEFTNODE); node 2759 fs/unicode/mkutf8data.c node = (*trie & TRIENODE); node 326 fs/unicode/utf8-norm.c int node; node 334 fs/unicode/utf8-norm.c node = 1; node 335 fs/unicode/utf8-norm.c while (node) { node 347 fs/unicode/utf8-norm.c node = (*trie & RIGHTNODE); node 356 fs/unicode/utf8-norm.c node = (*trie & TRIENODE); node 366 fs/unicode/utf8-norm.c node = (*trie & LEFTNODE); node 373 fs/unicode/utf8-norm.c node = (*trie & TRIENODE); node 1085 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_da_intnode *node; node 1122 fs/xfs/libxfs/xfs_attr_leaf.c node = bp1->b_addr; node 1123 fs/xfs/libxfs/xfs_attr_leaf.c dp->d_ops->node_hdr_from_disk(&icnodehdr, node); node 1124 fs/xfs/libxfs/xfs_attr_leaf.c btree = dp->d_ops->node_tree_p(node); node 1134 fs/xfs/libxfs/xfs_attr_leaf.c dp->d_ops->node_hdr_to_disk(node, &icnodehdr); node 335 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 351 fs/xfs/libxfs/xfs_da_btree.c node = bp->b_addr; node 366 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_to_disk(node, &ichdr); node 368 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size)); node 385 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 504 fs/xfs/libxfs/xfs_da_btree.c node = oldblk->bp->b_addr; node 505 fs/xfs/libxfs/xfs_da_btree.c if (node->hdr.info.forw) { node 506 fs/xfs/libxfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.forw) != addblk->blkno) { node 510 fs/xfs/libxfs/xfs_da_btree.c node = addblk->bp->b_addr; node 511 fs/xfs/libxfs/xfs_da_btree.c node->hdr.info.back = cpu_to_be32(oldblk->blkno); node 513 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info, node 514 fs/xfs/libxfs/xfs_da_btree.c sizeof(node->hdr.info))); node 516 fs/xfs/libxfs/xfs_da_btree.c node = oldblk->bp->b_addr; node 517 fs/xfs/libxfs/xfs_da_btree.c if (node->hdr.info.back) { node 518 fs/xfs/libxfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.back) != addblk->blkno) { node 522 fs/xfs/libxfs/xfs_da_btree.c node = addblk->bp->b_addr; node 523 fs/xfs/libxfs/xfs_da_btree.c node->hdr.info.forw = cpu_to_be32(oldblk->blkno); node 525 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info, node 526 fs/xfs/libxfs/xfs_da_btree.c sizeof(node->hdr.info))); node 544 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 574 fs/xfs/libxfs/xfs_da_btree.c node = bp->b_addr; node 616 fs/xfs/libxfs/xfs_da_btree.c memcpy(node, oldroot, size); node 619 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da3_intnode *node3 = (struct xfs_da3_intnode *)node; node 639 fs/xfs/libxfs/xfs_da_btree.c node = bp->b_addr; node 640 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 641 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 647 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_to_disk(node, &nodehdr); node 661 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, sizeof(xfs_da_node_entry_t) * 2)); node 678 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 688 fs/xfs/libxfs/xfs_da_btree.c node = oldblk->bp->b_addr; node 689 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 735 fs/xfs/libxfs/xfs_da_btree.c node = oldblk->bp->b_addr; node 736 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 915 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 923 fs/xfs/libxfs/xfs_da_btree.c node = oldblk->bp->b_addr; node 924 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 925 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 944 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &btree[oldblk->index], node 948 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_to_disk(node, &nodehdr); node 950 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size)); node 1152 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1174 fs/xfs/libxfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info; node 1175 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1232 fs/xfs/libxfs/xfs_da_btree.c node = bp->b_addr; node 1233 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&thdr, node); node 1275 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1279 fs/xfs/libxfs/xfs_da_btree.c node = bp->b_addr; node 1280 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1285 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 1299 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1330 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; node 1331 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1332 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 1338 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &btree[blk->index], node 1353 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1362 fs/xfs/libxfs/xfs_da_btree.c node = drop_blk->bp->b_addr; node 1363 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1371 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 1377 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &btree[index], tmp)); node 1382 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &btree[index], sizeof(btree[index]))); node 1384 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_to_disk(node, &nodehdr); node 1386 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size)); node 1489 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1552 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; node 1553 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1554 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 1881 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_intnode *node; node 1904 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; node 1905 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1906 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 1965 fs/xfs/libxfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info; node 1966 fs/xfs/libxfs/xfs_da_btree.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 1967 fs/xfs/libxfs/xfs_da_btree.c btree = dp->d_ops->node_tree_p(node); node 161 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = ifp->if_u1.if_root; node 168 fs/xfs/libxfs/xfs_iext_tree.c node = node->ptrs[0]; node 169 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node); node 172 fs/xfs/libxfs/xfs_iext_tree.c return node; node 179 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = ifp->if_u1.if_root; node 187 fs/xfs/libxfs/xfs_iext_tree.c if (!node->ptrs[i]) node 189 fs/xfs/libxfs/xfs_iext_tree.c node = node->ptrs[i - 1]; node 190 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node); node 193 fs/xfs/libxfs/xfs_iext_tree.c return node; node 277 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 281 fs/xfs/libxfs/xfs_iext_tree.c if (node->keys[n] > offset) node 283 fs/xfs/libxfs/xfs_iext_tree.c if (node->keys[n] < offset) node 309 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = ifp->if_u1.if_root; node 317 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_key_cmp(node, i, offset) > 0) node 320 fs/xfs/libxfs/xfs_iext_tree.c node = node->ptrs[i - 1]; node 321 fs/xfs/libxfs/xfs_iext_tree.c if (!node) node 325 fs/xfs/libxfs/xfs_iext_tree.c return node; node 330 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 336 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_key_cmp(node, i, offset) > 0) node 345 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 351 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_key_cmp(node, i, offset) > 0) node 360 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 366 fs/xfs/libxfs/xfs_iext_tree.c if (node->keys[i] == XFS_IEXT_KEY_INVALID) node 401 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = kmem_zalloc(NODE_SIZE, KM_NOFS); node 407 fs/xfs/libxfs/xfs_iext_tree.c node->keys[0] = xfs_iext_leaf_key(prev, 0); node 408 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[0] = prev; node 414 fs/xfs/libxfs/xfs_iext_tree.c node->keys[0] = prev->keys[0]; node 415 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[0] = prev; node 419 fs/xfs/libxfs/xfs_iext_tree.c node->keys[i] = XFS_IEXT_KEY_INVALID; node 421 fs/xfs/libxfs/xfs_iext_tree.c ifp->if_u1.if_root = node; node 433 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = ifp->if_u1.if_root; node 438 fs/xfs/libxfs/xfs_iext_tree.c if (i > 0 && xfs_iext_key_cmp(node, i, old_offset) > 0) node 440 fs/xfs/libxfs/xfs_iext_tree.c if (node->keys[i] == old_offset) node 441 fs/xfs/libxfs/xfs_iext_tree.c node->keys[i] = new_offset; node 443 fs/xfs/libxfs/xfs_iext_tree.c node = node->ptrs[i - 1]; node 444 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node); node 447 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node == ptr); node 456 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node = *nodep; node 472 fs/xfs/libxfs/xfs_iext_tree.c new->keys[i] = node->keys[nr_keep + i]; node 473 fs/xfs/libxfs/xfs_iext_tree.c new->ptrs[i] = node->ptrs[nr_keep + i]; node 475 fs/xfs/libxfs/xfs_iext_tree.c node->keys[nr_keep + i] = XFS_IEXT_KEY_INVALID; node 476 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[nr_keep + i] = NULL; node 499 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, *new; node 507 fs/xfs/libxfs/xfs_iext_tree.c node = xfs_iext_find_level(ifp, offset, level); node 508 fs/xfs/libxfs/xfs_iext_tree.c pos = xfs_iext_node_insert_pos(node, offset); node 509 fs/xfs/libxfs/xfs_iext_tree.c nr_entries = xfs_iext_node_nr_entries(node, pos); node 511 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(pos >= nr_entries || xfs_iext_key_cmp(node, pos, offset) != 0); node 515 fs/xfs/libxfs/xfs_iext_tree.c new = xfs_iext_split_node(&node, &pos, &nr_entries); node 521 fs/xfs/libxfs/xfs_iext_tree.c if (node != new && pos == 0 && nr_entries > 0) node 522 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_update_node(ifp, node->keys[0], offset, level, node); node 525 fs/xfs/libxfs/xfs_iext_tree.c node->keys[i] = node->keys[i - 1]; node 526 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[i] = node->ptrs[i - 1]; node 528 fs/xfs/libxfs/xfs_iext_tree.c node->keys[pos] = offset; node 529 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[pos] = ptr; node 675 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 684 fs/xfs/libxfs/xfs_iext_tree.c return node; node 692 fs/xfs/libxfs/xfs_iext_tree.c prev->keys[nr_prev + i] = node->keys[i]; node 693 fs/xfs/libxfs/xfs_iext_tree.c prev->ptrs[nr_prev + i] = node->ptrs[i]; node 695 fs/xfs/libxfs/xfs_iext_tree.c return node; node 710 fs/xfs/libxfs/xfs_iext_tree.c node->keys[nr_entries + i] = next->keys[i]; node 711 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[nr_entries + i] = next->ptrs[i]; node 728 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, *parent; node 732 fs/xfs/libxfs/xfs_iext_tree.c node = xfs_iext_find_level(ifp, offset, level); node 733 fs/xfs/libxfs/xfs_iext_tree.c pos = xfs_iext_node_pos(node, offset); node 735 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node->ptrs[pos]); node 736 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node->ptrs[pos] == victim); node 739 fs/xfs/libxfs/xfs_iext_tree.c nr_entries = xfs_iext_node_nr_entries(node, pos) - 1; node 740 fs/xfs/libxfs/xfs_iext_tree.c offset = node->keys[0]; node 742 fs/xfs/libxfs/xfs_iext_tree.c node->keys[i] = node->keys[i + 1]; node 743 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[i] = node->ptrs[i + 1]; node 745 fs/xfs/libxfs/xfs_iext_tree.c node->keys[nr_entries] = XFS_IEXT_KEY_INVALID; node 746 fs/xfs/libxfs/xfs_iext_tree.c node->ptrs[nr_entries] = NULL; node 749 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_update_node(ifp, offset, node->keys[0], level, node); node 750 fs/xfs/libxfs/xfs_iext_tree.c offset = node->keys[0]; node 767 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(parent->ptrs[pos] == node); node 769 fs/xfs/libxfs/xfs_iext_tree.c node = xfs_iext_rebalance_node(parent, &pos, node, nr_entries); node 770 fs/xfs/libxfs/xfs_iext_tree.c if (node) { node 771 fs/xfs/libxfs/xfs_iext_tree.c victim = node; node 772 fs/xfs/libxfs/xfs_iext_tree.c node = parent; node 780 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(node == ifp->if_u1.if_root); node 781 fs/xfs/libxfs/xfs_iext_tree.c ifp->if_u1.if_root = node->ptrs[0]; node 783 fs/xfs/libxfs/xfs_iext_tree.c kmem_free(node); node 1025 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_node *node, node 1032 fs/xfs/libxfs/xfs_iext_tree.c if (node->keys[i] == XFS_IEXT_KEY_INVALID) node 1034 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_destroy_node(node->ptrs[i], level - 1); node 1038 fs/xfs/libxfs/xfs_iext_tree.c kmem_free(node); node 333 fs/xfs/scrub/dabtree.c struct xfs_da_intnode *node; node 435 fs/xfs/scrub/dabtree.c node = blk->bp->b_addr; node 436 fs/xfs/scrub/dabtree.c ip->d_ops->node_hdr_from_disk(&nodehdr, node); node 437 fs/xfs/scrub/dabtree.c btree = ip->d_ops->node_tree_p(node); node 199 fs/xfs/xfs_attr_inactive.c xfs_da_intnode_t *node; node 215 fs/xfs/xfs_attr_inactive.c node = bp->b_addr; node 216 fs/xfs/xfs_attr_inactive.c dp->d_ops->node_hdr_from_disk(&ichdr, node); node 222 fs/xfs/xfs_attr_inactive.c btree = dp->d_ops->node_tree_p(node); node 286 fs/xfs/xfs_attr_inactive.c node = bp->b_addr; node 287 fs/xfs/xfs_attr_inactive.c btree = dp->d_ops->node_tree_p(node); node 202 fs/xfs/xfs_attr_list.c struct xfs_da_intnode *node; node 220 fs/xfs/xfs_attr_list.c node = bp->b_addr; node 221 fs/xfs/xfs_attr_list.c magic = be16_to_cpu(node->hdr.info.magic); node 228 fs/xfs/xfs_attr_list.c node, sizeof(*node)); node 232 fs/xfs/xfs_attr_list.c dp->d_ops->node_hdr_from_disk(&nodehdr, node); node 246 fs/xfs/xfs_attr_list.c btree = dp->d_ops->node_tree_p(node); node 283 fs/xfs/xfs_attr_list.c struct xfs_da_intnode *node; node 308 fs/xfs/xfs_attr_list.c node = bp->b_addr; node 309 fs/xfs/xfs_attr_list.c switch (be16_to_cpu(node->hdr.info.magic)) { node 328 include/acpi/acpi_bus.h struct list_head node; node 356 include/acpi/acpi_bus.h struct list_head node; node 476 include/acpi/acpi_bus.h struct list_head node; node 1147 include/acpi/actbl2.h #define ACPI_NFIT_BUILD_DEVICE_HANDLE(dimm, channel, memory, socket, node) \ node 1152 include/acpi/actbl2.h ((node) << ACPI_NFIT_NODE_ID_OFFSET)) node 38 include/asm-generic/topology.h #define set_numa_node(node) node 41 include/asm-generic/topology.h #define set_cpu_numa_node(cpu, node) node 49 include/asm-generic/topology.h #define cpumask_of_node(node) ((node) == 0 ? cpu_online_mask : cpu_none_mask) node 51 include/asm-generic/topology.h #define cpumask_of_node(node) ((void)(node), cpu_online_mask) node 69 include/asm-generic/topology.h #define set_numa_mem(node) node 72 include/asm-generic/topology.h #define set_cpu_numa_mem(cpu, node) node 118 include/clocksource/timer-ti-dm.h struct list_head node; node 60 include/drm/drm_flip_work.h struct list_head node; node 106 include/drm/drm_mipi_dsi.h struct mipi_dsi_host *of_find_mipi_dsi_host_by_node(struct device_node *node); node 156 include/drm/drm_mipi_dsi.h struct device_node *node; node 196 include/drm/drm_mm.h void (*color_adjust)(const struct drm_mm_node *node, node 254 include/drm/drm_mm.h static inline bool drm_mm_node_allocated(const struct drm_mm_node *node) node 256 include/drm/drm_mm.h return node->allocated; node 289 include/drm/drm_mm.h static inline bool drm_mm_hole_follows(const struct drm_mm_node *node) node 291 include/drm/drm_mm.h return node->hole_size; node 402 include/drm/drm_mm.h int drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node); node 404 include/drm/drm_mm.h struct drm_mm_node *node, node 430 include/drm/drm_mm.h drm_mm_insert_node_generic(struct drm_mm *mm, struct drm_mm_node *node, node 435 include/drm/drm_mm.h return drm_mm_insert_node_in_range(mm, node, node 455 include/drm/drm_mm.h struct drm_mm_node *node, node 458 include/drm/drm_mm.h return drm_mm_insert_node_generic(mm, node, size, 0, 0, 0); node 461 include/drm/drm_mm.h void drm_mm_remove_node(struct drm_mm_node *node); node 542 include/drm/drm_mm.h struct drm_mm_node *node); node 544 include/drm/drm_mm.h struct drm_mm_node *node); node 27 include/drm/drm_of.h struct device_node *node); node 31 include/drm/drm_of.h int drm_of_encoder_active_endpoint(struct device_node *node, node 55 include/drm/drm_of.h struct device_node *node) node 67 include/drm/drm_of.h static inline int drm_of_encoder_active_endpoint(struct device_node *node, node 110 include/drm/drm_of.h static inline int drm_of_encoder_active_endpoint_id(struct device_node *node, node 114 include/drm/drm_of.h int ret = drm_of_encoder_active_endpoint(node, encoder, node 120 include/drm/drm_of.h static inline int drm_of_encoder_active_port_id(struct device_node *node, node 124 include/drm/drm_of.h int ret = drm_of_encoder_active_endpoint(node, encoder, node 72 include/drm/drm_vma_manager.h struct drm_vma_offset_node *node, unsigned long pages); node 74 include/drm/drm_vma_manager.h struct drm_vma_offset_node *node); node 76 include/drm/drm_vma_manager.h int drm_vma_node_allow(struct drm_vma_offset_node *node, struct drm_file *tag); node 77 include/drm/drm_vma_manager.h void drm_vma_node_revoke(struct drm_vma_offset_node *node, node 79 include/drm/drm_vma_manager.h bool drm_vma_node_is_allowed(struct drm_vma_offset_node *node, node 99 include/drm/drm_vma_manager.h struct drm_vma_offset_node *node; node 101 include/drm/drm_vma_manager.h node = drm_vma_offset_lookup_locked(mgr, start, pages); node 102 include/drm/drm_vma_manager.h return (node && node->vm_node.start == start) ? node : NULL; node 147 include/drm/drm_vma_manager.h static inline void drm_vma_node_reset(struct drm_vma_offset_node *node) node 149 include/drm/drm_vma_manager.h memset(node, 0, sizeof(*node)); node 150 include/drm/drm_vma_manager.h node->vm_files = RB_ROOT; node 151 include/drm/drm_vma_manager.h rwlock_init(&node->vm_lock); node 168 include/drm/drm_vma_manager.h static inline unsigned long drm_vma_node_start(const struct drm_vma_offset_node *node) node 170 include/drm/drm_vma_manager.h return node->vm_node.start; node 185 include/drm/drm_vma_manager.h static inline unsigned long drm_vma_node_size(struct drm_vma_offset_node *node) node 187 include/drm/drm_vma_manager.h return node->vm_node.size; node 202 include/drm/drm_vma_manager.h static inline __u64 drm_vma_node_offset_addr(struct drm_vma_offset_node *node) node 204 include/drm/drm_vma_manager.h return ((__u64)node->vm_node.start) << PAGE_SHIFT; node 219 include/drm/drm_vma_manager.h static inline void drm_vma_node_unmap(struct drm_vma_offset_node *node, node 222 include/drm/drm_vma_manager.h if (drm_mm_node_allocated(&node->vm_node)) node 224 include/drm/drm_vma_manager.h drm_vma_node_offset_addr(node), node 225 include/drm/drm_vma_manager.h drm_vma_node_size(node) << PAGE_SHIFT, 1); node 240 include/drm/drm_vma_manager.h static inline int drm_vma_node_verify_access(struct drm_vma_offset_node *node, node 243 include/drm/drm_vma_manager.h return drm_vma_node_is_allowed(node, tag) ? 0 : -EACCES; node 188 include/drm/gpu_scheduler.h struct list_head node; node 65 include/drm/spsc_queue.h static inline bool spsc_queue_push(struct spsc_queue *queue, struct spsc_node *node) node 69 include/drm/spsc_queue.h node->next = NULL; node 73 include/drm/spsc_queue.h tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next); node 74 include/drm/spsc_queue.h WRITE_ONCE(*tail, node); node 91 include/drm/spsc_queue.h struct spsc_node *next, *node; node 96 include/drm/spsc_queue.h node = READ_ONCE(queue->head); node 98 include/drm/spsc_queue.h if (!node) node 101 include/drm/spsc_queue.h next = READ_ONCE(node->next); node 108 include/drm/spsc_queue.h (long)&node->next, (long) &queue->head) != (long)&node->next) { node 112 include/drm/spsc_queue.h } while (unlikely(!(queue->head = READ_ONCE(node->next)))); node 117 include/drm/spsc_queue.h return node; node 43 include/linux/alarmtimer.h struct timerqueue_node node; node 39 include/linux/async.h int node); node 41 include/linux/async.h int node, node 31 include/linux/async_tx.h struct list_head node; node 17 include/linux/attribute_container.h struct list_head node; node 212 include/linux/backlight.h struct backlight_device *of_find_backlight_by_node(struct device_node *node); node 215 include/linux/backlight.h of_find_backlight_by_node(struct device_node *node) node 532 include/linux/blkdev.h int node; node 50 include/linux/bpf-cgroup.h struct rb_node node; node 55 include/linux/bpf-cgroup.h struct list_head node; node 36 include/linux/btree.h unsigned long *node; node 54 include/linux/ceph/mon_client.h struct rb_node node; node 267 include/linux/ceph/osd_client.h struct rb_node node; /* osd */ node 296 include/linux/ceph/osd_client.h struct rb_node node; node 45 include/linux/ceph/osdmap.h struct rb_node node; node 141 include/linux/ceph/osdmap.h struct rb_node node; node 13 include/linux/ceph/string_table.h struct rb_node node; node 571 include/linux/cgroup-defs.h struct list_head node; /* anchored at ss->cfts */ node 587 include/linux/clk-provider.h void of_fixed_factor_clk_setup(struct device_node *node); node 798 include/linux/clk-provider.h int __must_check of_clk_hw_register(struct device_node *node, struct clk_hw *hw); node 57 include/linux/clk.h struct list_head node; node 15 include/linux/clk/clk-conf.h int of_clk_set_defaults(struct device_node *node, bool clk_supplier); node 17 include/linux/clk/clk-conf.h static inline int of_clk_set_defaults(struct device_node *node, node 151 include/linux/clk/ti.h struct list_head node; node 19 include/linux/clkdev.h struct list_head node; node 241 include/linux/compaction.h struct node; node 243 include/linux/compaction.h extern int compaction_register_node(struct node *node); node 244 include/linux/compaction.h extern void compaction_unregister_node(struct node *node); node 248 include/linux/compaction.h static inline int compaction_register_node(struct node *node) node 253 include/linux/compaction.h static inline void compaction_unregister_node(struct node *node) node 432 include/linux/cper.h u16 node; node 452 include/linux/cper.h u16 node; node 472 include/linux/cper.h u16 node; node 269 include/linux/cpuhotplug.h struct hlist_node *node), node 271 include/linux/cpuhotplug.h struct hlist_node *node)) node 278 include/linux/cpuhotplug.h int __cpuhp_state_add_instance(enum cpuhp_state state, struct hlist_node *node, node 281 include/linux/cpuhotplug.h struct hlist_node *node, bool invoke); node 294 include/linux/cpuhotplug.h struct hlist_node *node) node 296 include/linux/cpuhotplug.h return __cpuhp_state_add_instance(state, node, true); node 309 include/linux/cpuhotplug.h struct hlist_node *node) node 311 include/linux/cpuhotplug.h return __cpuhp_state_add_instance(state, node, false); node 316 include/linux/cpuhotplug.h struct hlist_node *node) node 318 include/linux/cpuhotplug.h return __cpuhp_state_add_instance_cpuslocked(state, node, false); node 365 include/linux/cpuhotplug.h struct hlist_node *node, bool invoke); node 377 include/linux/cpuhotplug.h struct hlist_node *node) node 379 include/linux/cpuhotplug.h return __cpuhp_state_remove_instance(state, node, true); node 391 include/linux/cpuhotplug.h struct hlist_node *node) node 393 include/linux/cpuhotplug.h return __cpuhp_state_remove_instance(state, node, false); node 192 include/linux/cpumask.h static inline unsigned int cpumask_local_spread(unsigned int i, int node) node 247 include/linux/cpumask.h unsigned int cpumask_local_spread(unsigned int i, int node); node 737 include/linux/cpumask.h bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); node 739 include/linux/cpumask.h bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); node 762 include/linux/cpumask.h int node) node 774 include/linux/cpumask.h int node) node 67 include/linux/cpuset.h extern bool __cpuset_node_allowed(int node, gfp_t gfp_mask); node 69 include/linux/cpuset.h static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) node 72 include/linux/cpuset.h return __cpuset_node_allowed(node, gfp_mask); node 207 include/linux/cpuset.h static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) node 196 include/linux/crush/crush.h struct rb_node node; node 20 include/linux/dca.h struct list_head node; node 27 include/linux/dca.h struct list_head node; node 29 include/linux/debugobjects.h struct hlist_node node; node 27 include/linux/devfreq-event.h struct list_head node; node 149 include/linux/devfreq.h struct list_head node; node 531 include/linux/device.h struct list_head node; node 761 include/linux/device.h struct list_head node; node 951 include/linux/device.h struct device_node *node, int index, node 1356 include/linux/device.h static inline void set_dev_node(struct device *dev, int node) node 1358 include/linux/device.h dev->numa_node = node; node 1365 include/linux/device.h static inline void set_dev_node(struct device *dev, int node) node 87 include/linux/dio.h struct list_head node; node 343 include/linux/dma-buf.h struct list_head node; node 117 include/linux/dma-fence.h struct list_head node; node 280 include/linux/dmar.h extern int dmar_alloc_hwirq(int id, int node, void *arg); node 1324 include/linux/efi.h struct device *efi_get_device_by_path(struct efi_dev_path **node, size_t *len); node 164 include/linux/elevator.h #define rb_entry_rq(node) rb_entry((node), struct request, rb_node) node 100 include/linux/enclosure.h struct list_head node; node 225 include/linux/extcon.h extern struct extcon_dev *extcon_find_edev_by_node(struct device_node *node); node 279 include/linux/extcon.h static inline struct extcon_dev *extcon_find_edev_by_node(struct device_node *node) node 441 include/linux/fb.h int node; node 809 include/linux/fb.h pr_err("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__) node 811 include/linux/fb.h pr_notice("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__) node 813 include/linux/fb.h pr_warn("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__) node 815 include/linux/fb.h pr_info("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__) node 817 include/linux/fb.h pr_debug("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__) node 185 include/linux/firewire.h struct fw_node *node; node 288 include/linux/firmware/xlnx-zynqmp.h int (*request_node)(const u32 node, node 292 include/linux/firmware/xlnx-zynqmp.h int (*release_node)(const u32 node); node 293 include/linux/firmware/xlnx-zynqmp.h int (*set_requirement)(const u32 node, node 41 include/linux/fpga/fpga-bridge.h struct list_head node; node 47 include/linux/fpga/fpga-bridge.h struct fpga_bridge *of_fpga_bridge_get(struct device_node *node, node 188 include/linux/fpga/fpga-mgr.h struct fpga_manager *of_fpga_mgr_get(struct device_node *node); node 743 include/linux/fs.h static inline unsigned int i_blocksize(const struct inode *node) node 745 include/linux/fs.h return (1 << node->i_blkbits); node 87 include/linux/fsl/mc.h struct list_head node; node 47 include/linux/gameport.h struct list_head node; node 536 include/linux/gfp.h int node, bool hugepage); node 542 include/linux/gfp.h #define alloc_pages_vma(gfp_mask, order, vma, addr, node, false)\ node 550 include/linux/gfp.h #define alloc_page_vma_node(gfp_mask, vma, addr, node) \ node 551 include/linux/gfp.h alloc_pages_vma(gfp_mask, 0, vma, addr, node, false) node 567 include/linux/gpio/consumer.h struct gpio_desc *gpiod_get_from_of_node(struct device_node *node, node 577 include/linux/gpio/consumer.h struct gpio_desc *gpiod_get_from_of_node(struct device_node *node, node 591 include/linux/gpio/consumer.h struct device_node *node, node 602 include/linux/gpio/consumer.h struct device_node *node, node 671 include/linux/gpio/driver.h struct list_head node; node 60 include/linux/hashtable.h #define hash_add(hashtable, node, key) \ node 61 include/linux/hashtable.h hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))]) node 69 include/linux/hashtable.h #define hash_add_rcu(hashtable, node, key) \ node 70 include/linux/hashtable.h hlist_add_head_rcu(node, &hashtable[hash_min(key, HASH_BITS(hashtable))]) node 76 include/linux/hashtable.h static inline bool hash_hashed(struct hlist_node *node) node 78 include/linux/hashtable.h return !hlist_unhashed(node); node 105 include/linux/hashtable.h static inline void hash_del(struct hlist_node *node) node 107 include/linux/hashtable.h hlist_del_init(node); node 114 include/linux/hashtable.h static inline void hash_del_rcu(struct hlist_node *node) node 116 include/linux/hashtable.h hlist_del_init_rcu(node); node 34 include/linux/hid-debug.h struct list_head node; node 33 include/linux/hidraw.h struct list_head node; node 118 include/linux/hrtimer.h struct timerqueue_node node; node 241 include/linux/hrtimer.h timer->node.expires = time; node 248 include/linux/hrtimer.h timer->node.expires = ktime_add_safe(time, delta); node 254 include/linux/hrtimer.h timer->node.expires = ktime_add_safe(time, ns_to_ktime(delta)); node 259 include/linux/hrtimer.h timer->node.expires = tv64; node 265 include/linux/hrtimer.h timer->node.expires = ktime_add_safe(timer->node.expires, time); node 271 include/linux/hrtimer.h timer->node.expires = ktime_add_ns(timer->node.expires, ns); node 277 include/linux/hrtimer.h return timer->node.expires; node 287 include/linux/hrtimer.h return timer->node.expires; node 296 include/linux/hrtimer.h return ktime_to_ns(timer->node.expires); node 301 include/linux/hrtimer.h return ktime_sub(timer->node.expires, timer->base->get_time()); node 335 include/linux/hrtimer.h ktime_t rem = ktime_sub(timer->node.expires, now); node 929 include/linux/i2c.h extern struct i2c_client *of_find_i2c_device_by_node(struct device_node *node); node 932 include/linux/i2c.h extern struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node); node 935 include/linux/i2c.h struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node); node 941 include/linux/i2c.h int of_i2c_get_board_info(struct device *dev, struct device_node *node, node 946 include/linux/i2c.h static inline struct i2c_client *of_find_i2c_device_by_node(struct device_node *node) node 951 include/linux/i2c.h static inline struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node) node 956 include/linux/i2c.h static inline struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node) node 969 include/linux/i2c.h struct device_node *node, node 42 include/linux/i3c/master.h struct list_head node; node 64 include/linux/i3c/master.h struct list_head node; node 180 include/linux/i3c/master.h struct list_head node; node 507 include/linux/i3c/master.h list_for_each_entry(dev, &(bus)->devs.i2c, common.node) node 518 include/linux/i3c/master.h list_for_each_entry(dev, &(bus)->devs.i3c, common.node) node 195 include/linux/input.h struct list_head node; node 322 include/linux/input.h struct list_head node; node 557 include/linux/intel-iommu.h int node; node 654 include/linux/intel-iommu.h void *alloc_pgtable_page(int node); node 50 include/linux/interconnect-provider.h int (*aggregate)(struct icc_node *node, u32 tag, u32 avg_bw, node 52 include/linux/interconnect-provider.h void (*pre_aggregate)(struct icc_node *node); node 97 include/linux/interconnect-provider.h int icc_link_create(struct icc_node *node, const int dst_id); node 99 include/linux/interconnect-provider.h void icc_node_add(struct icc_node *node, struct icc_provider *provider); node 100 include/linux/interconnect-provider.h void icc_node_del(struct icc_node *node); node 115 include/linux/interconnect-provider.h static inline int icc_link_create(struct icc_node *node, const int dst_id) node 125 include/linux/interconnect-provider.h void icc_node_add(struct icc_node *node, struct icc_provider *provider) node 129 include/linux/interconnect-provider.h void icc_node_del(struct icc_node *node) node 15 include/linux/interval_tree.h interval_tree_insert(struct interval_tree_node *node, node 19 include/linux/interval_tree.h interval_tree_remove(struct interval_tree_node *node, node 27 include/linux/interval_tree.h interval_tree_iter_next(struct interval_tree_node *node, node 38 include/linux/interval_tree_generic.h ITSTATIC void ITPREFIX ## _insert(ITSTRUCT *node, \ node 42 include/linux/interval_tree_generic.h ITTYPE start = ITSTART(node), last = ITLAST(node); \ node 59 include/linux/interval_tree_generic.h node->ITSUBTREE = last; \ node 60 include/linux/interval_tree_generic.h rb_link_node(&node->ITRB, rb_parent, link); \ node 61 include/linux/interval_tree_generic.h rb_insert_augmented_cached(&node->ITRB, root, \ node 65 include/linux/interval_tree_generic.h ITSTATIC void ITPREFIX ## _remove(ITSTRUCT *node, \ node 68 include/linux/interval_tree_generic.h rb_erase_augmented_cached(&node->ITRB, root, &ITPREFIX ## _augment); \ node 81 include/linux/interval_tree_generic.h ITPREFIX ## _subtree_search(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ node 88 include/linux/interval_tree_generic.h if (node->ITRB.rb_left) { \ node 89 include/linux/interval_tree_generic.h ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \ node 100 include/linux/interval_tree_generic.h node = left; \ node 104 include/linux/interval_tree_generic.h if (ITSTART(node) <= last) { /* Cond1 */ \ node 105 include/linux/interval_tree_generic.h if (start <= ITLAST(node)) /* Cond2 */ \ node 106 include/linux/interval_tree_generic.h return node; /* node is leftmost match */ \ node 107 include/linux/interval_tree_generic.h if (node->ITRB.rb_right) { \ node 108 include/linux/interval_tree_generic.h node = rb_entry(node->ITRB.rb_right, \ node 110 include/linux/interval_tree_generic.h if (start <= node->ITSUBTREE) \ node 122 include/linux/interval_tree_generic.h ITSTRUCT *node, *leftmost; \ node 140 include/linux/interval_tree_generic.h node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \ node 141 include/linux/interval_tree_generic.h if (node->ITSUBTREE < start) \ node 148 include/linux/interval_tree_generic.h return ITPREFIX ## _subtree_search(node, start, last); \ node 152 include/linux/interval_tree_generic.h ITPREFIX ## _iter_next(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ node 154 include/linux/interval_tree_generic.h struct rb_node *rb = node->ITRB.rb_right, *prev; \ node 173 include/linux/interval_tree_generic.h rb = rb_parent(&node->ITRB); \ node 176 include/linux/interval_tree_generic.h prev = &node->ITRB; \ node 177 include/linux/interval_tree_generic.h node = rb_entry(rb, ITSTRUCT, ITRB); \ node 178 include/linux/interval_tree_generic.h rb = node->ITRB.rb_right; \ node 182 include/linux/interval_tree_generic.h if (last < ITSTART(node)) /* !Cond1 */ \ node 184 include/linux/interval_tree_generic.h else if (start <= ITLAST(node)) /* Cond2 */ \ node 185 include/linux/interval_tree_generic.h return node; \ node 152 include/linux/iocontext.h gfp_t gfp_flags, int node); node 20 include/linux/iova.h struct rb_node node; node 146 include/linux/irq.h unsigned int node; node 820 include/linux/irq.h return d->node; node 868 include/linux/irq.h int __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, node 873 include/linux/irq.h unsigned int cnt, int node, struct module *owner, node 877 include/linux/irq.h #define irq_alloc_descs(irq, from, cnt, node) \ node 878 include/linux/irq.h __irq_alloc_descs(irq, from, cnt, node, THIS_MODULE, NULL) node 880 include/linux/irq.h #define irq_alloc_desc(node) \ node 881 include/linux/irq.h irq_alloc_descs(-1, 0, 1, node) node 883 include/linux/irq.h #define irq_alloc_desc_at(at, node) \ node 884 include/linux/irq.h irq_alloc_descs(at, at, 1, node) node 886 include/linux/irq.h #define irq_alloc_desc_from(from, node) \ node 887 include/linux/irq.h irq_alloc_descs(-1, from, 1, node) node 889 include/linux/irq.h #define irq_alloc_descs_from(from, cnt, node) \ node 890 include/linux/irq.h irq_alloc_descs(-1, from, cnt, node) node 892 include/linux/irq.h #define devm_irq_alloc_descs(dev, irq, from, cnt, node) \ node 893 include/linux/irq.h __devm_irq_alloc_descs(dev, irq, from, cnt, node, THIS_MODULE, NULL) node 895 include/linux/irq.h #define devm_irq_alloc_desc(dev, node) \ node 896 include/linux/irq.h devm_irq_alloc_descs(dev, -1, 0, 1, node) node 898 include/linux/irq.h #define devm_irq_alloc_desc_at(dev, at, node) \ node 899 include/linux/irq.h devm_irq_alloc_descs(dev, at, at, 1, node) node 901 include/linux/irq.h #define devm_irq_alloc_desc_from(dev, from, node) \ node 902 include/linux/irq.h devm_irq_alloc_descs(dev, -1, from, 1, node) node 904 include/linux/irq.h #define devm_irq_alloc_descs_from(dev, from, cnt, node) \ node 905 include/linux/irq.h devm_irq_alloc_descs(dev, -1, from, cnt, node) node 914 include/linux/irq.h unsigned int irq_alloc_hwirqs(int cnt, int node); node 915 include/linux/irq.h static inline unsigned int irq_alloc_hwirq(int node) node 917 include/linux/irq.h return irq_alloc_hwirqs(1, node); node 924 include/linux/irq.h int arch_setup_hwirq(unsigned int irq, int node); node 46 include/linux/irqbypass.h struct list_head node; node 72 include/linux/irqbypass.h struct list_head node; node 146 include/linux/irqchip/arm-gic.h int gic_of_init(struct device_node *node, struct device_node *parent); node 20 include/linux/irqchip/arm-vic.h u32 vic_sources, u32 resume_sources, struct device_node *node); node 10 include/linux/irqchip/versatile-fpga.h struct device_node *node); node 11 include/linux/irqchip/versatile-fpga.h int fpga_irq_of_init(struct device_node *node, node 104 include/linux/irqdomain.h int (*match)(struct irq_domain *d, struct device_node *node, node 110 include/linux/irqdomain.h int (*xlate)(struct irq_domain *d, struct device_node *node, node 278 include/linux/irqdomain.h irq_hw_number_t hwirq, int node, node 281 include/linux/irqdomain.h static inline struct fwnode_handle *of_node_to_fwnode(struct device_node *node) node 283 include/linux/irqdomain.h return node ? &node->fwnode : NULL; node 307 include/linux/irqdomain.h static inline struct irq_domain *irq_find_matching_host(struct device_node *node, node 310 include/linux/irqdomain.h return irq_find_matching_fwnode(of_node_to_fwnode(node), bus_token); node 313 include/linux/irqdomain.h static inline struct irq_domain *irq_find_host(struct device_node *node) node 317 include/linux/irqdomain.h d = irq_find_matching_host(node, DOMAIN_BUS_WIRED); node 319 include/linux/irqdomain.h d = irq_find_matching_host(node, DOMAIN_BUS_ANY); node 456 include/linux/irqdomain.h struct device_node *node, node 461 include/linux/irqdomain.h of_node_to_fwnode(node), node 466 include/linux/irqdomain.h unsigned int nr_irqs, int node, void *arg, node 474 include/linux/irqdomain.h unsigned int nr_irqs, int node, void *arg) node 476 include/linux/irqdomain.h return __irq_domain_alloc_irqs(domain, -1, nr_irqs, node, arg, false, node 541 include/linux/irqdomain.h unsigned int nr_irqs, int node, void *arg) node 11 include/linux/kthread.h int node, node 96 include/linux/kthread.h struct list_head node; node 115 include/linux/kthread.h .node = LIST_HEAD_INIT((work).node), \ node 160 include/linux/kthread.h INIT_LIST_HEAD(&(work)->node); \ node 121 include/linux/leds.h struct list_head node; /* LED Device list */ node 690 include/linux/lightnvm.h static inline struct nvm_dev *nvm_alloc_dev(int node) node 53 include/linux/list_lru.h struct list_lru_node *node; node 73 include/linux/livepatch.h struct list_head node; node 127 include/linux/livepatch.h struct list_head node; node 166 include/linux/livepatch.h list_for_each_entry_safe(obj, tmp_obj, &patch->obj_list, node) node 169 include/linux/livepatch.h list_for_each_entry(obj, &patch->obj_list, node) node 177 include/linux/livepatch.h list_for_each_entry_safe(func, tmp_func, &obj->func_list, node) node 180 include/linux/livepatch.h list_for_each_entry(func, &obj->func_list, node) node 114 include/linux/llist.h #define llist_for_each(pos, node) \ node 115 include/linux/llist.h for ((pos) = (node); pos; (pos) = (pos)->next) node 133 include/linux/llist.h #define llist_for_each_safe(pos, n, node) \ node 134 include/linux/llist.h for ((pos) = (node); (pos) && ((n) = (pos)->next, true); (pos) = (n)) node 151 include/linux/llist.h #define llist_for_each_entry(pos, node, member) \ node 152 include/linux/llist.h for ((pos) = llist_entry((node), typeof(*(pos)), member); \ node 173 include/linux/llist.h #define llist_for_each_entry_safe(pos, n, node, member) \ node 174 include/linux/llist.h for (pos = llist_entry((node), typeof(*pos), member); \ node 192 include/linux/llist.h static inline struct llist_node *llist_next(struct llist_node *node) node 194 include/linux/llist.h return node->next; node 86 include/linux/mailbox_controller.h struct list_head node; node 90 include/linux/mfd/ipaq-micro.h struct list_head node; node 84 include/linux/mfd/mc13xxx.h struct device_node *node; node 136 include/linux/mfd/ucb1x00.h struct list_head node; node 152 include/linux/mfd/ucb1x00.h struct list_head node; node 127 include/linux/migrate.h struct vm_area_struct *vma, int node); node 134 include/linux/migrate.h struct vm_area_struct *vma, int node) node 145 include/linux/migrate.h struct page *page, int node); node 151 include/linux/migrate.h struct page *page, int node) node 1109 include/linux/mlx4/device.h int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf, int node); node 935 include/linux/mlx5/driver.h struct mlx5_frag_buf *buf, int node); node 940 include/linux/mlx5/driver.h struct mlx5_frag_buf *buf, int node); node 989 include/linux/mlx5/driver.h int node); node 8433 include/linux/mlx5/mlx5_ifc.h u8 node[0x8]; node 667 include/linux/mm.h extern void *kvmalloc_node(size_t size, gfp_t flags, int node); node 672 include/linux/mm.h static inline void *kvzalloc_node(size_t size, gfp_t flags, int node) node 674 include/linux/mm.h return kvmalloc_node(size, flags | __GFP_ZERO, node); node 1283 include/linux/mm.h static inline void set_page_node(struct page *page, unsigned long node) node 1286 include/linux/mm.h page->flags |= (node & NODES_MASK) << NODES_PGSHIFT; node 1290 include/linux/mm.h unsigned long node, unsigned long pfn) node 1293 include/linux/mm.h set_page_node(page, node); node 2242 include/linux/mm.h void vma_interval_tree_insert(struct vm_area_struct *node, node 2244 include/linux/mm.h void vma_interval_tree_insert_after(struct vm_area_struct *node, node 2247 include/linux/mm.h void vma_interval_tree_remove(struct vm_area_struct *node, node 2251 include/linux/mm.h struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node, node 2258 include/linux/mm.h void anon_vma_interval_tree_insert(struct anon_vma_chain *node, node 2260 include/linux/mm.h void anon_vma_interval_tree_remove(struct anon_vma_chain *node, node 2266 include/linux/mm.h struct anon_vma_chain *node, unsigned long start, unsigned long last); node 2268 include/linux/mm.h void anon_vma_interval_tree_verify(struct anon_vma_chain *node); node 2775 include/linux/mm.h pgd_t *vmemmap_pgd_populate(unsigned long addr, int node); node 2776 include/linux/mm.h p4d_t *vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node); node 2777 include/linux/mm.h pud_t *vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node); node 2778 include/linux/mm.h pmd_t *vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node); node 2779 include/linux/mm.h pte_t *vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node); node 2780 include/linux/mm.h void *vmemmap_alloc_block(unsigned long size, int node); node 2782 include/linux/mm.h void *vmemmap_alloc_block_buf(unsigned long size, int node); node 2786 include/linux/mm.h int node); node 2787 include/linux/mm.h int vmemmap_populate(unsigned long start, unsigned long end, int node, node 438 include/linux/mmzone.h int node; node 886 include/linux/mmzone.h return zone->node; node 891 include/linux/mmzone.h zone->node = nid; node 304 include/linux/module.h struct latch_tree_node node; node 652 include/linux/netdevice.h static inline void netdev_queue_numa_node_write(struct netdev_queue *q, int node) node 655 include/linux/netdevice.h q->numa_node = node; node 98 include/linux/node.h extern struct node *node_devices[]; node 99 include/linux/node.h typedef void (*node_registration_func_t)(struct node *); node 112 include/linux/node.h extern void unregister_node(struct node *node); node 181 include/linux/node.h #define to_node(device) container_of(device, struct node, dev) node 127 include/linux/nodemask.h #define node_set(node, dst) __node_set((node), &(dst)) node 128 include/linux/nodemask.h static __always_inline void __node_set(int node, volatile nodemask_t *dstp) node 130 include/linux/nodemask.h set_bit(node, dstp->bits); node 133 include/linux/nodemask.h #define node_clear(node, dst) __node_clear((node), &(dst)) node 134 include/linux/nodemask.h static inline void __node_clear(int node, volatile nodemask_t *dstp) node 136 include/linux/nodemask.h clear_bit(node, dstp->bits); node 152 include/linux/nodemask.h #define node_isset(node, nodemask) test_bit((node), (nodemask).bits) node 154 include/linux/nodemask.h #define node_test_and_set(node, nodemask) \ node 155 include/linux/nodemask.h __node_test_and_set((node), &(nodemask)) node 156 include/linux/nodemask.h static inline int __node_test_and_set(int node, nodemask_t *addr) node 158 include/linux/nodemask.h return test_and_set_bit(node, addr->bits); node 279 include/linux/nodemask.h int __next_node_in(int node, const nodemask_t *srcp); node 281 include/linux/nodemask.h static inline void init_nodemask_of_node(nodemask_t *mask, int node) node 284 include/linux/nodemask.h node_set(node, *mask); node 287 include/linux/nodemask.h #define nodemask_of_node(node) \ node 291 include/linux/nodemask.h m.bits[0] = 1UL << (node); \ node 293 include/linux/nodemask.h init_nodemask_of_node(&m, (node)); \ node 378 include/linux/nodemask.h #define for_each_node_mask(node, mask) \ node 379 include/linux/nodemask.h for ((node) = first_node(mask); \ node 380 include/linux/nodemask.h (node) < MAX_NUMNODES; \ node 381 include/linux/nodemask.h (node) = next_node((node), (mask))) node 383 include/linux/nodemask.h #define for_each_node_mask(node, mask) \ node 385 include/linux/nodemask.h for ((node) = 0; (node) < 1; (node)++) node 413 include/linux/nodemask.h static inline int node_state(int node, enum node_states state) node 415 include/linux/nodemask.h return node_isset(node, node_states[state]); node 418 include/linux/nodemask.h static inline void node_set_state(int node, enum node_states state) node 420 include/linux/nodemask.h __node_set(node, &node_states[state]); node 423 include/linux/nodemask.h static inline void node_clear_state(int node, enum node_states state) node 425 include/linux/nodemask.h __node_clear(node, &node_states[state]); node 464 include/linux/nodemask.h static inline int node_state(int node, enum node_states state) node 466 include/linux/nodemask.h return node == 0; node 469 include/linux/nodemask.h static inline void node_set_state(int node, enum node_states state) node 473 include/linux/nodemask.h static inline void node_clear_state(int node, enum node_states state) node 482 include/linux/nodemask.h #define for_each_node_state(node, __state) \ node 483 include/linux/nodemask.h for ( (node) = 0; (node) == 0; (node) = 1) node 491 include/linux/nodemask.h #define node_set_online(node) node_set_state((node), N_ONLINE) node 492 include/linux/nodemask.h #define node_set_offline(node) node_clear_state((node), N_ONLINE) node 510 include/linux/nodemask.h #define node_online(node) node_state((node), N_ONLINE) node 511 include/linux/nodemask.h #define node_possible(node) node_state((node), N_POSSIBLE) node 513 include/linux/nodemask.h #define for_each_node(node) for_each_node_state(node, N_POSSIBLE) node 514 include/linux/nodemask.h #define for_each_online_node(node) for_each_node_state(node, N_ONLINE) node 45 include/linux/nvmem-consumer.h struct list_head node; node 94 include/linux/nvmem-provider.h struct list_head node; node 94 include/linux/of.h struct device_node *node; node 106 include/linux/of.h static inline void of_node_init(struct device_node *node) node 109 include/linux/of.h kobject_init(&node->kobj, &of_node_ktype); node 111 include/linux/of.h node->fwnode.ops = &of_fwnode_ops; node 121 include/linux/of.h extern struct device_node *of_node_get(struct device_node *node); node 122 include/linux/of.h extern void of_node_put(struct device_node *node); node 125 include/linux/of.h static inline struct device_node *of_node_get(struct device_node *node) node 127 include/linux/of.h return node; node 129 include/linux/of.h static inline void of_node_put(struct device_node *node) { } node 170 include/linux/of.h #define of_fwnode_handle(node) \ node 172 include/linux/of.h typeof(node) __of_fwnode_handle_node = (node); \ node 183 include/linux/of.h static inline bool of_node_is_root(const struct device_node *node) node 185 include/linux/of.h return node && (node->parent == NULL); node 286 include/linux/of.h extern struct device_node *of_get_parent(const struct device_node *node); node 287 include/linux/of.h extern struct device_node *of_get_next_parent(struct device_node *node); node 288 include/linux/of.h extern struct device_node *of_get_next_child(const struct device_node *node, node 291 include/linux/of.h const struct device_node *node, struct device_node *prev); node 295 include/linux/of.h extern struct device_node *of_get_child_by_name(const struct device_node *node, node 349 include/linux/of.h extern const void *of_get_property(const struct device_node *node, node 361 include/linux/of.h const struct of_device_id *matches, const struct device_node *node); node 362 include/linux/of.h extern int of_modalias_node(struct device_node *node, char *modalias, int len); node 626 include/linux/of.h static inline struct device_node *of_get_parent(const struct device_node *node) node 632 include/linux/of.h const struct device_node *node, struct device_node *prev) node 638 include/linux/of.h const struct device_node *node, struct device_node *prev) node 649 include/linux/of.h #define of_fwnode_handle(node) NULL node 663 include/linux/of.h const struct device_node *node, node 750 include/linux/of.h static inline const void *of_get_property(const struct device_node *node, node 1321 include/linux/of.h struct list_head node; node 10 include/linux/of_address.h struct device_node *node; node 52 include/linux/of_address.h struct device_node *node); node 54 include/linux/of_address.h struct device_node *node); node 89 include/linux/of_address.h struct device_node *node) node 95 include/linux/of_address.h struct device_node *node) node 122 include/linux/of_address.h void __iomem *of_iomap(struct device_node *node, int index); node 39 include/linux/of_fdt.h extern u64 of_flat_dt_translate_address(unsigned long node); node 45 include/linux/of_fdt.h extern int of_scan_flat_dt(int (*it)(unsigned long node, const char *uname, node 48 include/linux/of_fdt.h extern int of_scan_flat_dt_subnodes(unsigned long node, node 49 include/linux/of_fdt.h int (*it)(unsigned long node, node 53 include/linux/of_fdt.h extern int of_get_flat_dt_subnode_by_name(unsigned long node, node 55 include/linux/of_fdt.h extern const void *of_get_flat_dt_prop(unsigned long node, const char *name, node 57 include/linux/of_fdt.h extern int of_flat_dt_is_compatible(unsigned long node, const char *name); node 59 include/linux/of_fdt.h extern uint32_t of_get_flat_dt_phandle(unsigned long node); node 61 include/linux/of_fdt.h extern int early_init_dt_scan_chosen(unsigned long node, const char *uname, node 63 include/linux/of_fdt.h extern int early_init_dt_scan_memory(unsigned long node, const char *uname, node 68 include/linux/of_fdt.h extern void __init early_init_dt_scan_chosen_arch(unsigned long node); node 76 include/linux/of_fdt.h extern int early_init_dt_scan_root(unsigned long node, const char *uname, node 41 include/linux/of_graph.h int of_graph_parse_endpoint(const struct device_node *node, node 44 include/linux/of_graph.h struct device_node *of_graph_get_port_by_id(struct device_node *node, u32 id); node 50 include/linux/of_graph.h const struct device_node *node); node 51 include/linux/of_graph.h struct device_node *of_graph_get_port_parent(struct device_node *node); node 53 include/linux/of_graph.h const struct device_node *node); node 54 include/linux/of_graph.h struct device_node *of_graph_get_remote_port(const struct device_node *node); node 55 include/linux/of_graph.h struct device_node *of_graph_get_remote_node(const struct device_node *node, node 59 include/linux/of_graph.h static inline int of_graph_parse_endpoint(const struct device_node *node, node 71 include/linux/of_graph.h struct device_node *node, u32 id) node 90 include/linux/of_graph.h const struct device_node *node) node 96 include/linux/of_graph.h struct device_node *node) node 102 include/linux/of_graph.h const struct device_node *node) node 108 include/linux/of_graph.h const struct device_node *node) node 113 include/linux/of_graph.h const struct device_node *node, node 108 include/linux/of_irq.h extern unsigned int irq_of_parse_and_map(struct device_node *node, int index); node 18 include/linux/of_pdt.h int (*nextprop)(phandle node, char *prev, char *buf); node 21 include/linux/of_pdt.h int (*getproplen)(phandle node, const char *prop); node 22 include/linux/of_pdt.h int (*getproperty)(phandle node, const char *prop, char *buf, node 27 include/linux/of_pdt.h phandle (*getsibling)(phandle node); node 30 include/linux/of_pdt.h int (*pkg2path)(phandle node, char *buf, const int buflen, int *len); node 39 include/linux/of_reserved_mem.h void fdt_reserved_mem_save_node(unsigned long node, const char *uname, node 51 include/linux/of_reserved_mem.h static inline void fdt_reserved_mem_save_node(unsigned long node, node 571 include/linux/pci.h struct list_head node; /* Node in list of buses */ node 826 include/linux/pci.h struct list_head node; node 1524 include/linux/pci.h struct device_node *node, node 1745 include/linux/pci.h struct device_node *node, node 107 include/linux/perf/arm_pmu.h struct hlist_node node; node 166 include/linux/phy/phy.h struct list_head node; node 241 include/linux/phy/phy.h struct phy *phy_create(struct device *dev, struct device_node *node, node 243 include/linux/phy/phy.h struct phy *devm_phy_create(struct device *dev, struct device_node *node, node 441 include/linux/phy/phy.h struct device_node *node, node 448 include/linux/phy/phy.h struct device_node *node, node 59 include/linux/pinctrl/pinctrl.h struct list_head node; node 110 include/linux/plist.h #define PLIST_NODE_INIT(node, __prio) \ node 113 include/linux/plist.h .prio_list = LIST_HEAD_INIT((node).prio_list), \ node 114 include/linux/plist.h .node_list = LIST_HEAD_INIT((node).node_list), \ node 132 include/linux/plist.h static inline void plist_node_init(struct plist_node *node, int prio) node 134 include/linux/plist.h node->prio = prio; node 135 include/linux/plist.h INIT_LIST_HEAD(&node->prio_list); node 136 include/linux/plist.h INIT_LIST_HEAD(&node->node_list); node 139 include/linux/plist.h extern void plist_add(struct plist_node *node, struct plist_head *head); node 140 include/linux/plist.h extern void plist_del(struct plist_node *node, struct plist_head *head); node 142 include/linux/plist.h extern void plist_requeue(struct plist_node *node, struct plist_head *head); node 219 include/linux/plist.h static inline int plist_node_empty(const struct plist_node *node) node 221 include/linux/plist.h return list_empty(&node->node_list); node 42 include/linux/pm_qos.h struct plist_node node; node 48 include/linux/pm_qos.h struct list_head node; node 114 include/linux/pm_qos.h int pm_qos_update_target(struct pm_qos_constraints *c, struct plist_node *node, node 70 include/linux/posix-timers.h struct timerqueue_node node; node 81 include/linux/posix-timers.h return timerqueue_add(head, &ctmr->node); node 87 include/linux/posix-timers.h timerqueue_del(ctmr->head, &ctmr->node); node 94 include/linux/posix-timers.h return ctmr->node.expires; node 99 include/linux/posix-timers.h ctmr->node.expires = exp; node 152 include/linux/power/smartreflex.h struct list_head node; node 72 include/linux/powercap.h struct list_head node; node 389 include/linux/property.h const struct software_node *node; node 422 include/linux/property.h struct fwnode_handle *software_node_fwnode(const struct software_node *node); node 431 include/linux/property.h int software_node_register(const struct software_node *node); node 104 include/linux/psi_types.h struct list_head node; node 210 include/linux/pxa2xx_ssp.h struct list_head node; node 98 include/linux/radix-tree.h struct radix_tree_node *node; node 43 include/linux/rbtree.h #define RB_EMPTY_NODE(node) \ node 44 include/linux/rbtree.h ((node)->__rb_parent_color == (unsigned long)(node)) node 45 include/linux/rbtree.h #define RB_CLEAR_NODE(node) \ node 46 include/linux/rbtree.h ((node)->__rb_parent_color = (unsigned long)(node)) node 69 include/linux/rbtree.h static inline void rb_link_node(struct rb_node *node, struct rb_node *parent, node 72 include/linux/rbtree.h node->__rb_parent_color = (unsigned long)parent; node 73 include/linux/rbtree.h node->rb_left = node->rb_right = NULL; node 75 include/linux/rbtree.h *rb_link = node; node 78 include/linux/rbtree.h static inline void rb_link_node_rcu(struct rb_node *node, struct rb_node *parent, node 81 include/linux/rbtree.h node->__rb_parent_color = (unsigned long)parent; node 82 include/linux/rbtree.h node->rb_left = node->rb_right = NULL; node 84 include/linux/rbtree.h rcu_assign_pointer(*rb_link, node); node 135 include/linux/rbtree.h static inline void rb_insert_color_cached(struct rb_node *node, node 140 include/linux/rbtree.h root->rb_leftmost = node; node 141 include/linux/rbtree.h rb_insert_color(node, &root->rb_root); node 144 include/linux/rbtree.h static inline void rb_erase_cached(struct rb_node *node, node 147 include/linux/rbtree.h if (root->rb_leftmost == node) node 148 include/linux/rbtree.h root->rb_leftmost = rb_next(node); node 149 include/linux/rbtree.h rb_erase(node, &root->rb_root); node 28 include/linux/rbtree_augmented.h void (*propagate)(struct rb_node *node, struct rb_node *stop); node 33 include/linux/rbtree_augmented.h extern void __rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 47 include/linux/rbtree_augmented.h rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 50 include/linux/rbtree_augmented.h __rb_insert_augmented(node, root, augment->rotate); node 54 include/linux/rbtree_augmented.h rb_insert_augmented_cached(struct rb_node *node, node 59 include/linux/rbtree_augmented.h root->rb_leftmost = node; node 60 include/linux/rbtree_augmented.h rb_insert_augmented(node, &root->rb_root, augment); node 80 include/linux/rbtree_augmented.h RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ node 81 include/linux/rbtree_augmented.h if (RBCOMPUTE(node, true)) \ node 83 include/linux/rbtree_augmented.h rb = rb_parent(&node->RBFIELD); \ node 122 include/linux/rbtree_augmented.h static inline bool RBNAME ## _compute_max(RBSTRUCT *node, bool exit) \ node 125 include/linux/rbtree_augmented.h RBTYPE max = RBCOMPUTE(node); \ node 126 include/linux/rbtree_augmented.h if (node->RBFIELD.rb_left) { \ node 127 include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ node 131 include/linux/rbtree_augmented.h if (node->RBFIELD.rb_right) { \ node 132 include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \ node 136 include/linux/rbtree_augmented.h if (exit && node->RBAUGMENTED == max) \ node 138 include/linux/rbtree_augmented.h node->RBAUGMENTED = max; \ node 198 include/linux/rbtree_augmented.h __rb_erase_augmented(struct rb_node *node, struct rb_root *root, node 201 include/linux/rbtree_augmented.h struct rb_node *child = node->rb_right; node 202 include/linux/rbtree_augmented.h struct rb_node *tmp = node->rb_left; node 214 include/linux/rbtree_augmented.h pc = node->__rb_parent_color; node 216 include/linux/rbtree_augmented.h __rb_change_child(node, child, parent, root); node 225 include/linux/rbtree_augmented.h tmp->__rb_parent_color = pc = node->__rb_parent_color; node 227 include/linux/rbtree_augmented.h __rb_change_child(node, tmp, parent, root); node 247 include/linux/rbtree_augmented.h augment->copy(node, successor); node 273 include/linux/rbtree_augmented.h augment->copy(node, successor); node 277 include/linux/rbtree_augmented.h tmp = node->rb_left; node 281 include/linux/rbtree_augmented.h pc = node->__rb_parent_color; node 283 include/linux/rbtree_augmented.h __rb_change_child(node, successor, tmp, root); node 302 include/linux/rbtree_augmented.h rb_erase_augmented(struct rb_node *node, struct rb_root *root, node 305 include/linux/rbtree_augmented.h struct rb_node *rebalance = __rb_erase_augmented(node, root, augment); node 311 include/linux/rbtree_augmented.h rb_erase_augmented_cached(struct rb_node *node, struct rb_root_cached *root, node 314 include/linux/rbtree_augmented.h if (root->rb_leftmost == node) node 315 include/linux/rbtree_augmented.h root->rb_leftmost = rb_next(node); node 316 include/linux/rbtree_augmented.h rb_erase_augmented(node, &root->rb_root, augment); node 41 include/linux/rbtree_latch.h struct rb_node node[2]; node 70 include/linux/rbtree_latch.h __lt_from_rb(struct rb_node *node, int idx) node 72 include/linux/rbtree_latch.h return container_of(node, struct latch_tree_node, node[idx]); node 81 include/linux/rbtree_latch.h struct rb_node *node = <n->node[idx]; node 95 include/linux/rbtree_latch.h rb_link_node_rcu(node, parent, link); node 96 include/linux/rbtree_latch.h rb_insert_color(node, root); node 102 include/linux/rbtree_latch.h rb_erase(<n->node[idx], <r->tree[idx]); node 107 include/linux/rbtree_latch.h int (*comp)(void *key, struct latch_tree_node *node)) node 109 include/linux/rbtree_latch.h struct rb_node *node = rcu_dereference_raw(ltr->tree[idx].rb_node); node 113 include/linux/rbtree_latch.h while (node) { node 114 include/linux/rbtree_latch.h ltn = __lt_from_rb(node, idx); node 118 include/linux/rbtree_latch.h node = rcu_dereference_raw(node->rb_left); node 120 include/linux/rbtree_latch.h node = rcu_dereference_raw(node->rb_right); node 144 include/linux/rbtree_latch.h latch_tree_insert(struct latch_tree_node *node, node 149 include/linux/rbtree_latch.h __lt_insert(node, root, 0, ops->less); node 151 include/linux/rbtree_latch.h __lt_insert(node, root, 1, ops->less); node 171 include/linux/rbtree_latch.h latch_tree_erase(struct latch_tree_node *node, node 176 include/linux/rbtree_latch.h __lt_erase(node, root, 0); node 178 include/linux/rbtree_latch.h __lt_erase(node, root, 1); node 203 include/linux/rbtree_latch.h struct latch_tree_node *node; node 208 include/linux/rbtree_latch.h node = __lt_find(key, root, seq & 1, ops->comp); node 211 include/linux/rbtree_latch.h return node; node 503 include/linux/rculist.h #define hlist_next_rcu(node) (*((struct hlist_node __rcu **)(&(node)->next))) node 504 include/linux/rculist.h #define hlist_pprev_rcu(node) (*((struct hlist_node __rcu **)((node)->pprev))) node 44 include/linux/rculist_nulls.h #define hlist_nulls_next_rcu(node) \ node 45 include/linux/rculist_nulls.h (*((struct hlist_nulls_node __rcu __force **)&(node)->next)) node 23 include/linux/regulator/of_regulator.h struct device_node *node, node 25 include/linux/regulator/of_regulator.h extern int of_regulator_match(struct device *dev, struct device_node *node, node 31 include/linux/regulator/of_regulator.h struct device_node *node, node 38 include/linux/regulator/of_regulator.h struct device_node *node, node 336 include/linux/remoteproc.h struct list_head node; node 440 include/linux/remoteproc.h struct list_head node; node 486 include/linux/remoteproc.h struct list_head node; node 529 include/linux/remoteproc.h struct list_head node; node 580 include/linux/remoteproc.h struct list_head node; node 22 include/linux/reset.h struct reset_control *__of_reset_control_get(struct device_node *node, node 83 include/linux/reset.h struct device_node *node, node 229 include/linux/reset.h struct device_node *node, const char *id) node 231 include/linux/reset.h return __of_reset_control_get(node, id, 0, false, false, true); node 254 include/linux/reset.h struct device_node *node, const char *id) node 256 include/linux/reset.h return __of_reset_control_get(node, id, 0, true, false, false); node 271 include/linux/reset.h struct device_node *node, int index) node 273 include/linux/reset.h return __of_reset_control_get(node, NULL, index, false, false, true); node 299 include/linux/reset.h struct device_node *node, int index) node 301 include/linux/reset.h return __of_reset_control_get(node, NULL, index, true, false, false); node 412 include/linux/reset.h struct device_node *node, const char *id) node 414 include/linux/reset.h return of_reset_control_get_exclusive(node, id); node 418 include/linux/reset.h struct device_node *node, int index) node 420 include/linux/reset.h return of_reset_control_get_exclusive_by_index(node, index); node 470 include/linux/reset.h of_reset_control_array_get_exclusive(struct device_node *node) node 472 include/linux/reset.h return of_reset_control_array_get(node, false, false, true); node 476 include/linux/reset.h of_reset_control_array_get_exclusive_released(struct device_node *node) node 478 include/linux/reset.h return of_reset_control_array_get(node, false, false, false); node 482 include/linux/reset.h of_reset_control_array_get_shared(struct device_node *node) node 484 include/linux/reset.h return of_reset_control_array_get(node, true, false, true); node 488 include/linux/reset.h of_reset_control_array_get_optional_exclusive(struct device_node *node) node 490 include/linux/reset.h return of_reset_control_array_get(node, false, true, true); node 494 include/linux/reset.h of_reset_control_array_get_optional_shared(struct device_node *node) node 496 include/linux/reset.h return of_reset_control_array_get(node, true, true, true); node 24 include/linux/resource_ext.h struct list_head node; node 37 include/linux/resource_ext.h list_add(&entry->node, head); node 43 include/linux/resource_ext.h list_add_tail(&entry->node, head); node 48 include/linux/resource_ext.h list_del(&entry->node); node 64 include/linux/resource_ext.h list_for_each_entry((entry), (list), node) node 67 include/linux/resource_ext.h list_for_each_entry_safe((entry), (tmp), (list), node) node 208 include/linux/ring_buffer.h int trace_rb_cpu_prepare(unsigned int cpu, struct hlist_node *node); node 98 include/linux/rio.h struct list_head node; node 232 include/linux/rio.h struct list_head node; node 268 include/linux/rio.h struct list_head node; /* node in global list of ports */ node 321 include/linux/rio.h struct list_head node; /* node in list of networks */ node 456 include/linux/rio.h struct list_head node; node 543 include/linux/rio.h struct list_head node; node 13 include/linux/rpmsg/qcom_glink.h struct device_node *node); node 20 include/linux/rpmsg/qcom_glink.h struct device_node *node) node 13 include/linux/rpmsg/qcom_smd.h struct device_node *node); node 20 include/linux/rpmsg/qcom_smd.h struct device_node *node) node 90 include/linux/rtc.h struct timerqueue_node node; node 156 include/linux/sbitmap.h gfp_t flags, int node); node 376 include/linux/sbitmap.h int shift, bool round_robin, gfp_t flags, int node); node 19 include/linux/sched/numa_balancing.h extern void task_numa_fault(int last_node, int node, int pages, int flags); node 26 include/linux/sched/numa_balancing.h static inline void task_numa_fault(int last_node, int node, int pages, node 70 include/linux/sched/signal.h struct hlist_node node; node 383 include/linux/serial_core.h unsigned long node, node 57 include/linux/serio.h struct list_head node; node 39 include/linux/sh_clk.h struct list_head node; node 47 include/linux/shdma-base.h struct list_head node; node 10 include/linux/siox.h struct list_head node; /* node in smaster->devices */ node 1035 include/linux/skbuff.h int node); node 302 include/linux/skmsg.h struct sk_psock *sk_psock_init(struct sock *sk, int node); node 417 include/linux/slab.h void *__kmalloc_node(size_t size, gfp_t flags, int node) __assume_kmalloc_alignment __malloc; node 418 include/linux/slab.h void *kmem_cache_alloc_node(struct kmem_cache *, gfp_t flags, int node) __assume_slab_alignment __malloc; node 420 include/linux/slab.h static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) node 425 include/linux/slab.h static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) node 437 include/linux/slab.h int node, size_t size) __assume_slab_alignment __malloc; node 442 include/linux/slab.h int node, size_t size) node 461 include/linux/slab.h int node, size_t size) node 463 include/linux/slab.h void *ret = kmem_cache_alloc_node(s, gfpflags, node); node 584 include/linux/slab.h static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node) node 596 include/linux/slab.h flags, node, size); node 599 include/linux/slab.h return __kmalloc_node(size, flags, node); node 645 include/linux/slab.h int node) node 652 include/linux/slab.h return kmalloc_node(bytes, flags, node); node 653 include/linux/slab.h return __kmalloc_node(bytes, flags, node); node 656 include/linux/slab.h static inline void *kcalloc_node(size_t n, size_t size, gfp_t flags, int node) node 658 include/linux/slab.h return kmalloc_array_node(n, size, flags | __GFP_ZERO, node); node 664 include/linux/slab.h #define kmalloc_node_track_caller(size, flags, node) \ node 665 include/linux/slab.h __kmalloc_node_track_caller(size, flags, node, \ node 670 include/linux/slab.h #define kmalloc_node_track_caller(size, flags, node) \ node 699 include/linux/slab.h static inline void *kzalloc_node(size_t size, gfp_t flags, int node) node 701 include/linux/slab.h return kmalloc_node(size, flags | __GFP_ZERO, node); node 89 include/linux/slab_def.h struct kmem_cache_node *node[MAX_NUMNODES]; node 142 include/linux/slub_def.h struct kmem_cache_node *node[MAX_NUMNODES]; node 89 include/linux/soc/qcom/apr.h struct list_head node; node 121 include/linux/soc/qcom/qmi.h unsigned int node; node 151 include/linux/soc/qcom/qmi.h void (*bye)(struct qmi_handle *qmi, unsigned int node); node 153 include/linux/soc/qcom/qmi.h unsigned int node, unsigned int port); node 562 include/linux/soundwire/sdw.h struct list_head node; node 1386 include/linux/spi/spi.h of_find_spi_device_by_node(struct device_node *node); node 1391 include/linux/spi/spi.h of_find_spi_device_by_node(struct device_node *node) node 65 include/linux/srcutree.h struct srcu_node node[NUM_RCU_NODES]; /* Combining tree. */ node 496 include/linux/sunrpc/svc.h struct svc_pool *pool, int node); node 498 include/linux/sunrpc/svc.h struct svc_pool *pool, int node); node 118 include/linux/svga.h int svga_compute_pll(const struct svga_pll *pll, u32 f_wanted, u16 *m, u16 *n, u16 *r, int node); node 119 include/linux/svga.h int svga_check_timings(const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, int node); node 120 include/linux/svga.h void svga_set_timings(void __iomem *regbase, const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, u32 hmul, u32 hdiv, u32 vmul, u32 vdiv, u32 hborder, int node); node 315 include/linux/swap.h void workingset_update_node(struct xa_node *node); node 649 include/linux/swap.h extern void mem_cgroup_throttle_swaprate(struct mem_cgroup *memcg, int node, node 653 include/linux/swap.h int node, gfp_t gfp_mask) node 732 include/linux/syscalls.h asmlinkage long sys_getcpu(unsigned __user *cpu, unsigned __user *node, struct getcpu_cache __user *cache); node 14 include/linux/syscore_ops.h struct list_head node; node 137 include/linux/sysctl.h struct rb_node node; node 159 include/linux/sysctl.h struct ctl_node *node; node 81 include/linux/tc.h struct list_head node; /* Node in list of all TC devices. */ node 106 include/linux/tc.h struct list_head node; node 141 include/linux/thermal.h struct list_head node; node 219 include/linux/thermal.h struct list_head node; node 10 include/linux/timerqueue.h struct rb_node node; node 20 include/linux/timerqueue.h struct timerqueue_node *node); node 22 include/linux/timerqueue.h struct timerqueue_node *node); node 24 include/linux/timerqueue.h struct timerqueue_node *node); node 38 include/linux/timerqueue.h return rb_entry(leftmost, struct timerqueue_node, node); node 41 include/linux/timerqueue.h static inline void timerqueue_init(struct timerqueue_node *node) node 43 include/linux/timerqueue.h RB_CLEAR_NODE(&node->node); node 46 include/linux/timerqueue.h static inline bool timerqueue_node_queued(struct timerqueue_node *node) node 48 include/linux/timerqueue.h return !RB_EMPTY_NODE(&node->node); node 51 include/linux/timerqueue.h static inline bool timerqueue_node_expires(struct timerqueue_node *node) node 53 include/linux/timerqueue.h return node->expires; node 39 include/linux/topology.h #define nr_cpus_node(node) cpumask_weight(cpumask_of_node(node)) node 42 include/linux/topology.h #define for_each_node_with_cpus(node) \ node 43 include/linux/topology.h for_each_online_node(node) \ node 44 include/linux/topology.h if (nr_cpus_node(node)) node 100 include/linux/topology.h static inline void set_numa_node(int node) node 102 include/linux/topology.h this_cpu_write(numa_node, node); node 107 include/linux/topology.h static inline void set_cpu_numa_node(int cpu, int node) node 109 include/linux/topology.h per_cpu(numa_node, cpu) = node; node 136 include/linux/topology.h static inline void set_numa_mem(int node) node 138 include/linux/topology.h this_cpu_write(_numa_mem_, node); node 139 include/linux/topology.h _node_numa_mem_[numa_node_id()] = node; node 144 include/linux/topology.h static inline int node_to_mem_node(int node) node 146 include/linux/topology.h return _node_numa_mem_[node]; node 166 include/linux/topology.h static inline void set_cpu_numa_mem(int cpu, int node) node 168 include/linux/topology.h per_cpu(_numa_mem_, cpu) = node; node 169 include/linux/topology.h _node_numa_mem_[cpu_to_node(cpu)] = node; node 184 include/linux/topology.h static inline int node_to_mem_node(int node) node 186 include/linux/topology.h return node; node 1096 include/linux/usb.h struct list_head node; node 79 include/linux/usb/isp1301.h struct i2c_client *isp1301_get_client(struct device_node *node); node 224 include/linux/usb/phy.h struct device_node *node, struct notifier_block *nb); node 253 include/linux/usb/phy.h struct device_node *node, struct notifier_block *nb) node 75 include/linux/usb/role.h fwnode_usb_role_switch_get(struct fwnode_handle *node) node 92 include/linux/user_namespace.h struct hlist_node node; node 26 include/linux/vexpress.h int vexpress_config_get_topo(struct device_node *node, u32 *site, node 78 include/linux/vmalloc.h int node, pgprot_t prot); node 94 include/linux/vmalloc.h extern void *vmalloc_node(unsigned long size, int node); node 95 include/linux/vmalloc.h extern void *vzalloc_node(unsigned long size, int node); node 102 include/linux/vmalloc.h pgprot_t prot, unsigned long vm_flags, int node, node 105 include/linux/vmalloc.h extern void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags); node 106 include/linux/vmalloc.h static inline void *__vmalloc_node_flags_caller(unsigned long size, int node, node 109 include/linux/vmalloc.h return __vmalloc_node_flags(size, node, flags); node 113 include/linux/vmalloc.h int node, gfp_t flags, void *caller); node 232 include/linux/vmstat.h extern unsigned long sum_zone_node_page_state(int node, node 234 include/linux/vmstat.h extern unsigned long sum_zone_numa_state(int node, enum numa_stat_item item); node 238 include/linux/vmstat.h #define sum_zone_node_page_state(node, item) global_zone_page_state(item) node 239 include/linux/vmstat.h #define node_page_state(node, item) global_node_page_state(item) node 446 include/linux/workqueue.h extern bool queue_work_node(int node, struct workqueue_struct *wq, node 1115 include/linux/xarray.h #define XA_NODE_BUG_ON(node, x) do { \ node 1117 include/linux/xarray.h if (node) xa_dump_node(node); \ node 1123 include/linux/xarray.h #define XA_NODE_BUG_ON(node, x) do { } while (0) node 1142 include/linux/xarray.h const struct xa_node *node, unsigned int offset) node 1144 include/linux/xarray.h XA_NODE_BUG_ON(node, offset >= XA_CHUNK_SIZE); node 1145 include/linux/xarray.h return rcu_dereference_check(node->slots[offset], node 1151 include/linux/xarray.h const struct xa_node *node, unsigned int offset) node 1153 include/linux/xarray.h XA_NODE_BUG_ON(node, offset >= XA_CHUNK_SIZE); node 1154 include/linux/xarray.h return rcu_dereference_protected(node->slots[offset], node 1160 include/linux/xarray.h const struct xa_node *node) node 1162 include/linux/xarray.h return rcu_dereference_check(node->parent, node 1168 include/linux/xarray.h const struct xa_node *node) node 1170 include/linux/xarray.h return rcu_dereference_protected(node->parent, node 1175 include/linux/xarray.h static inline void *xa_mk_node(const struct xa_node *node) node 1177 include/linux/xarray.h return (void *)((unsigned long)node | 2); node 1252 include/linux/xarray.h typedef void (*xa_update_node_t)(struct xa_node *node); node 1403 include/linux/xarray.h static inline bool xas_not_node(struct xa_node *node) node 1405 include/linux/xarray.h return ((unsigned long)node & 3) || !node; node 1409 include/linux/xarray.h static inline bool xas_frozen(struct xa_node *node) node 1411 include/linux/xarray.h return (unsigned long)node & 2; node 1415 include/linux/xarray.h static inline bool xas_top(struct xa_node *node) node 1417 include/linux/xarray.h return node <= XAS_RESTART; node 1489 include/linux/xarray.h struct xa_node *node = xas->xa_node; node 1491 include/linux/xarray.h if (node) node 1492 include/linux/xarray.h return xa_entry(xas->xa, node, xas->xa_offset); node 1557 include/linux/xarray.h struct xa_node *node = xas->xa_node; node 1560 include/linux/xarray.h if (unlikely(xas_not_node(node) || node->shift || node 1569 include/linux/xarray.h entry = xa_entry(xas->xa, node, xas->xa_offset + 1); node 1615 include/linux/xarray.h struct xa_node *node = xas->xa_node; node 1619 include/linux/xarray.h if (unlikely(xas_not_node(node) || node->shift)) node 1628 include/linux/xarray.h entry = xa_entry(xas->xa, node, offset); node 1714 include/linux/xarray.h struct xa_node *node = xas->xa_node; node 1716 include/linux/xarray.h if (unlikely(xas_not_node(node) || node->shift || node 1722 include/linux/xarray.h return xa_entry(xas->xa, node, xas->xa_offset); node 1743 include/linux/xarray.h struct xa_node *node = xas->xa_node; node 1745 include/linux/xarray.h if (unlikely(xas_not_node(node) || node->shift || node 1751 include/linux/xarray.h return xa_entry(xas->xa, node, xas->xa_offset); node 93 include/linux/xattr.h struct simple_xattr *xattr, *node; node 95 include/linux/xattr.h list_for_each_entry_safe(xattr, node, &xattrs->head, list) { node 55 include/linux/zorro.h struct list_head node; node 70 include/media/v4l2-clk.h #define v4l2_clk_name_of(name, size, node) snprintf(name, size, \ node 71 include/media/v4l2-clk.h "of-%pOF", node) node 212 include/media/v4l2-ctrls.h struct list_head node; node 288 include/media/v4l2-ctrls.h struct list_head node; node 76 include/media/v4l2-event.h struct list_head node; node 151 include/net/caif/caif_layer.h struct list_head node; node 16 include/net/datalink.h struct list_head node; node 80 include/net/garp.h struct rb_node node; node 78 include/net/inet_frag.h struct rhash_head node; node 89 include/net/inet_hashtables.h struct hlist_node node; node 99 include/net/inet_hashtables.h hlist_for_each_entry(tb, head, node) node 318 include/net/ip6_fib.h struct fib6_node *root, *node; node 22 include/net/ipx.h __u8 node[IPX_NODE_LEN]; node 71 include/net/ipx.h struct list_head node; /* node in ipx_interfaces list */ node 79 include/net/ipx.h struct list_head node; /* node in ipx_routes list */ node 102 include/net/ipx.h unsigned char node[IPX_NODE_LEN]; node 144 include/net/ipx.h int ipxitf_send(struct ipx_interface *intrfc, struct sk_buff *skb, char *node); node 147 include/net/ipx.h unsigned char *node); node 84 include/net/lapb.h struct list_head node; node 64 include/net/llc.h struct list_head node; node 91 include/net/mrp.h struct rb_node node; node 63 include/net/netfilter/nf_flow_table.h struct rhash_head node; node 19 include/net/ping.h #define ping_portaddr_for_each_entry(__sk, node, list) \ node 20 include/net/ping.h hlist_nulls_for_each_entry(__sk, node, list, sk_nulls_node) node 22 include/net/pkt_cls.h int (*fn)(struct tcf_proto *, void *node, struct tcf_walker *); node 510 include/net/sctp/sctp.h hlist_for_each_entry(epb, head, node) node 87 include/net/sctp/structs.h struct hlist_node node; node 776 include/net/sctp/structs.h struct rhlist_head node; node 1226 include/net/sctp/structs.h struct hlist_node node; node 26 include/net/seg6_hmac.h struct rhash_head node; node 567 include/net/sock.h static inline struct sock *sk_entry(const struct hlist_node *node) node 569 include/net/sock.h return hlist_entry(node, struct sock, sk_node); node 615 include/net/sock.h static inline void sk_node_init(struct hlist_node *node) node 617 include/net/sock.h node->pprev = NULL; node 620 include/net/sock.h static inline void sk_nulls_node_init(struct hlist_nulls_node *node) node 622 include/net/sock.h node->pprev = NULL; node 752 include/net/sock.h #define sk_nulls_for_each(__sk, node, list) \ node 753 include/net/sock.h hlist_nulls_for_each_entry(__sk, node, list, sk_nulls_node) node 754 include/net/sock.h #define sk_nulls_for_each_rcu(__sk, node, list) \ node 755 include/net/sock.h hlist_nulls_for_each_entry_rcu(__sk, node, list, sk_nulls_node) node 758 include/net/sock.h #define sk_nulls_for_each_from(__sk, node) \ node 759 include/net/sock.h if (__sk && ({ node = &(__sk)->sk_nulls_node; 1; })) \ node 760 include/net/sock.h hlist_nulls_for_each_entry_from(__sk, node, sk_nulls_node) node 1182 include/net/sock.h struct list_head node; node 1539 include/net/tcp.h struct hlist_node node; node 130 include/net/x25.h struct list_head node; node 138 include/net/x25.h struct list_head node; node 172 include/net/x25.h struct list_head node; node 15 include/net/xdp_priv.h struct rhash_head node; node 74 include/net/xdp_sock.h struct list_head node; node 166 include/rdma/ib_umem_odp.h struct interval_tree_node *node; node 168 include/rdma/ib_umem_odp.h node = interval_tree_iter_first(root, addr, addr + length - 1); node 169 include/rdma/ib_umem_odp.h if (!node) node 171 include/rdma/ib_umem_odp.h return container_of(node, struct ib_umem_odp, interval_tree); node 179 include/rdma/rdma_vt.h int node; node 82 include/scsi/scsi_device.h struct list_head node; node 71 include/soc/at91/atmel_tcb.h struct list_head node; node 90 include/soc/fsl/dpaa2-io.h struct list_head node; node 57 include/soc/fsl/qe/qe_ic.h void qe_ic_init(struct device_node *node, unsigned int flags, node 63 include/soc/fsl/qe/qe_ic.h static inline void qe_ic_init(struct device_node *node, unsigned int flags, node 73 include/sound/simple_card_utils.h struct device_node *node, node 84 include/sound/simple_card_utils.h #define asoc_simple_parse_clk_cpu(dev, node, dai_link, simple_dai) \ node 85 include/sound/simple_card_utils.h asoc_simple_parse_clk(dev, node, simple_dai, dai_link->cpus) node 86 include/sound/simple_card_utils.h #define asoc_simple_parse_clk_codec(dev, node, dai_link, simple_dai) \ node 87 include/sound/simple_card_utils.h asoc_simple_parse_clk(dev, node, simple_dai, dai_link->codecs) node 89 include/sound/simple_card_utils.h struct device_node *node, node 100 include/sound/simple_card_utils.h #define asoc_simple_parse_cpu(node, dai_link, is_single_link) \ node 101 include/sound/simple_card_utils.h asoc_simple_parse_dai(node, dai_link->cpus, is_single_link) node 102 include/sound/simple_card_utils.h #define asoc_simple_parse_codec(node, dai_link) \ node 103 include/sound/simple_card_utils.h asoc_simple_parse_dai(node, dai_link->codecs, NULL) node 104 include/sound/simple_card_utils.h #define asoc_simple_parse_platform(node, dai_link) \ node 105 include/sound/simple_card_utils.h asoc_simple_parse_dai(node, dai_link->platforms, NULL) node 576 include/sound/soc-dapm.h struct snd_soc_dapm_widget *node[2]; node 61 include/trace/events/alarmtimer.h __entry->expires = alarm->node.expires; node 167 include/trace/events/asoc.h __string( pnname, path->node[dir]->name ) node 176 include/trace/events/asoc.h __assign_str(pnname, path->node[dir]->name); node 178 include/trace/events/asoc.h __entry->path_node = (long)path->node[dir]; node 43 include/trace/events/cpuhp.h struct hlist_node *node), node 45 include/trace/events/cpuhp.h TP_ARGS(cpu, target, idx, fun, node), node 69 include/trace/events/kmem.h int node), node 71 include/trace/events/kmem.h TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node), node 79 include/trace/events/kmem.h __field( int, node ) node 88 include/trace/events/kmem.h __entry->node = node; node 97 include/trace/events/kmem.h __entry->node) node 104 include/trace/events/kmem.h gfp_t gfp_flags, int node), node 106 include/trace/events/kmem.h TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) node 113 include/trace/events/kmem.h gfp_t gfp_flags, int node), node 115 include/trace/events/kmem.h TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) node 45 include/trace/events/oom.h __field( int, node) node 56 include/trace/events/oom.h __entry->node = zone_to_nid(zoneref->zone); node 67 include/trace/events/oom.h __entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE), node 173 include/trace/events/rtc.h __entry->expires = timer->node.expires; node 38 include/uapi/linux/qrtr.h __le32 node; node 43 include/uapi/linux/qrtr.h __le32 node; node 70 include/uapi/linux/scif_ioctl.h __u16 node; node 50 include/uapi/linux/tipc.h __u32 node; node 269 include/uapi/linux/tipc.h unsigned int node) node 273 include/uapi/linux/tipc.h node; node 173 include/video/mmp_disp.h struct list_head node; node 211 include/video/mmp_disp.h struct list_head node; node 826 include/video/omapfb_dss.h omapdss_of_find_source_for_first_ep(struct device_node *node); node 60 include/xen/xenbus.h const char *node; node 137 include/xen/xenbus.h const char *dir, const char *node, unsigned int *num); node 139 include/xen/xenbus.h const char *dir, const char *node, unsigned int *len); node 141 include/xen/xenbus.h const char *dir, const char *node, const char *string); node 143 include/xen/xenbus.h const char *dir, const char *node); node 145 include/xen/xenbus.h const char *dir, const char *node); node 146 include/xen/xenbus.h int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node); node 153 include/xen/xenbus.h const char *dir, const char *node, const char *fmt, ...); node 156 include/xen/xenbus.h unsigned int xenbus_read_unsigned(const char *dir, const char *node, node 162 include/xen/xenbus.h const char *dir, const char *node, const char *fmt, ...); node 177 ipc/mqueue.c struct rb_node *node = &leaf->rb_node; node 179 ipc/mqueue.c if (info->msg_tree_rightmost == node) node 180 ipc/mqueue.c info->msg_tree_rightmost = rb_prev(node); node 182 ipc/mqueue.c rb_erase(node, &info->msg_tree); node 166 kernel/async.c int node, struct async_domain *domain) node 212 kernel/async.c queue_work_node(node, system_unbound_wq, &entry->work); node 231 kernel/async.c async_cookie_t async_schedule_node(async_func_t func, void *data, int node) node 233 kernel/async.c return async_schedule_node_domain(func, data, node, &async_dfl_domain); node 194 kernel/audit_tree.c size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); node 274 kernel/audit_tree.c static struct audit_chunk *find_chunk(struct node *p) node 327 kernel/audit_tree.c static void remove_chunk_node(struct audit_chunk *chunk, struct node *p) node 464 kernel/audit_tree.c struct node *p; node 575 kernel/audit_tree.c struct node *p; node 579 kernel/audit_tree.c p = list_first_entry(&victim->chunks, struct node, list); node 621 kernel/audit_tree.c struct node *node = list_entry(p, struct node, list); node 623 kernel/audit_tree.c if (node->index & (1U<<31)) { node 689 kernel/audit_tree.c struct node *node; node 708 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) { node 709 kernel/audit_tree.c struct audit_chunk *chunk = find_chunk(node); node 711 kernel/audit_tree.c node->index |= 1U<<31; node 715 kernel/audit_tree.c node->index &= ~(1U<<31); node 845 kernel/audit_tree.c struct node *node; node 847 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) node 848 kernel/audit_tree.c node->index &= ~(1U<<31); node 947 kernel/audit_tree.c struct node *node; node 949 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) node 950 kernel/audit_tree.c node->index &= ~(1U<<31); node 42 kernel/bpf/bpf_lru_list.c static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node) node 44 kernel/bpf/bpf_lru_list.c return node->ref; node 62 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, node 66 kernel/bpf/bpf_lru_list.c if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type))) node 72 kernel/bpf/bpf_lru_list.c if (&node->list == l->next_inactive_rotation) node 75 kernel/bpf/bpf_lru_list.c bpf_lru_list_count_dec(l, node->type); node 77 kernel/bpf/bpf_lru_list.c node->type = tgt_free_type; node 78 kernel/bpf/bpf_lru_list.c list_move(&node->list, free_list); node 83 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, node 86 kernel/bpf/bpf_lru_list.c if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) || node 91 kernel/bpf/bpf_lru_list.c node->type = tgt_type; node 92 kernel/bpf/bpf_lru_list.c node->ref = 0; node 93 kernel/bpf/bpf_lru_list.c list_move(&node->list, &l->lists[tgt_type]); node 101 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, node 104 kernel/bpf/bpf_lru_list.c if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) || node 108 kernel/bpf/bpf_lru_list.c if (node->type != tgt_type) { node 109 kernel/bpf/bpf_lru_list.c bpf_lru_list_count_dec(l, node->type); node 111 kernel/bpf/bpf_lru_list.c node->type = tgt_type; node 113 kernel/bpf/bpf_lru_list.c node->ref = 0; node 118 kernel/bpf/bpf_lru_list.c if (&node->list == l->next_inactive_rotation) node 121 kernel/bpf/bpf_lru_list.c list_move(&node->list, &l->lists[tgt_type]); node 143 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, *tmp_node, *first_node; node 147 kernel/bpf/bpf_lru_list.c list_for_each_entry_safe_reverse(node, tmp_node, active, list) { node 148 kernel/bpf/bpf_lru_list.c if (bpf_lru_node_is_ref(node)) node 149 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); node 151 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE); node 153 kernel/bpf/bpf_lru_list.c if (++i == lru->nr_scans || node == first_node) node 171 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 188 kernel/bpf/bpf_lru_list.c node = list_entry(cur, struct bpf_lru_node, list); node 190 kernel/bpf/bpf_lru_list.c if (bpf_lru_node_is_ref(node)) node 191 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); node 213 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, *tmp_node; node 217 kernel/bpf/bpf_lru_list.c list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) { node 218 kernel/bpf/bpf_lru_list.c if (bpf_lru_node_is_ref(node)) { node 219 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); node 220 kernel/bpf/bpf_lru_list.c } else if (lru->del_from_htab(lru->del_arg, node)) { node 221 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move_to_free(l, node, free_list, node 262 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, *tmp_node; node 277 kernel/bpf/bpf_lru_list.c list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list, node 279 kernel/bpf/bpf_lru_list.c if (lru->del_from_htab(lru->del_arg, node)) { node 280 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move_to_free(l, node, free_list, node 293 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, *tmp_node; node 295 kernel/bpf/bpf_lru_list.c list_for_each_entry_safe_reverse(node, tmp_node, node 297 kernel/bpf/bpf_lru_list.c if (bpf_lru_node_is_ref(node)) node 298 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE); node 300 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move_in(l, node, node 306 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node) node 310 kernel/bpf/bpf_lru_list.c if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type))) node 314 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE); node 322 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, *tmp_node; node 331 kernel/bpf/bpf_lru_list.c list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE], node 333 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move_to_free(l, node, local_free_list(loc_l), node 350 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node, node 353 kernel/bpf/bpf_lru_list.c *(u32 *)((void *)node + lru->hash_offset) = hash; node 354 kernel/bpf/bpf_lru_list.c node->cpu = cpu; node 355 kernel/bpf/bpf_lru_list.c node->type = BPF_LRU_LOCAL_LIST_T_PENDING; node 356 kernel/bpf/bpf_lru_list.c node->ref = 0; node 357 kernel/bpf/bpf_lru_list.c list_add(&node->list, local_pending_list(loc_l)); node 363 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 365 kernel/bpf/bpf_lru_list.c node = list_first_entry_or_null(local_free_list(loc_l), node 368 kernel/bpf/bpf_lru_list.c if (node) node 369 kernel/bpf/bpf_lru_list.c list_del(&node->list); node 371 kernel/bpf/bpf_lru_list.c return node; node 377 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 382 kernel/bpf/bpf_lru_list.c list_for_each_entry_reverse(node, local_pending_list(loc_l), node 384 kernel/bpf/bpf_lru_list.c if ((!bpf_lru_node_is_ref(node) || force) && node 385 kernel/bpf/bpf_lru_list.c lru->del_from_htab(lru->del_arg, node)) { node 386 kernel/bpf/bpf_lru_list.c list_del(&node->list); node 387 kernel/bpf/bpf_lru_list.c return node; node 403 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node = NULL; node 420 kernel/bpf/bpf_lru_list.c node = list_first_entry(free_list, struct bpf_lru_node, list); node 421 kernel/bpf/bpf_lru_list.c *(u32 *)((void *)node + lru->hash_offset) = hash; node 422 kernel/bpf/bpf_lru_list.c node->ref = 0; node 423 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE); node 428 kernel/bpf/bpf_lru_list.c return node; node 436 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 445 kernel/bpf/bpf_lru_list.c node = __local_list_pop_free(loc_l); node 446 kernel/bpf/bpf_lru_list.c if (!node) { node 448 kernel/bpf/bpf_lru_list.c node = __local_list_pop_free(loc_l); node 451 kernel/bpf/bpf_lru_list.c if (node) node 452 kernel/bpf/bpf_lru_list.c __local_list_add_pending(lru, loc_l, cpu, node, hash); node 456 kernel/bpf/bpf_lru_list.c if (node) node 457 kernel/bpf/bpf_lru_list.c return node; node 474 kernel/bpf/bpf_lru_list.c node = __local_list_pop_free(steal_loc_l); node 475 kernel/bpf/bpf_lru_list.c if (!node) node 476 kernel/bpf/bpf_lru_list.c node = __local_list_pop_pending(lru, steal_loc_l); node 481 kernel/bpf/bpf_lru_list.c } while (!node && steal != first_steal); node 485 kernel/bpf/bpf_lru_list.c if (node) { node 487 kernel/bpf/bpf_lru_list.c __local_list_add_pending(lru, loc_l, cpu, node, hash); node 491 kernel/bpf/bpf_lru_list.c return node; node 503 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node) node 507 kernel/bpf/bpf_lru_list.c if (WARN_ON_ONCE(node->type == BPF_LRU_LIST_T_FREE) || node 508 kernel/bpf/bpf_lru_list.c WARN_ON_ONCE(node->type == BPF_LRU_LOCAL_LIST_T_FREE)) node 511 kernel/bpf/bpf_lru_list.c if (node->type == BPF_LRU_LOCAL_LIST_T_PENDING) { node 514 kernel/bpf/bpf_lru_list.c loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu); node 518 kernel/bpf/bpf_lru_list.c if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) { node 523 kernel/bpf/bpf_lru_list.c node->type = BPF_LRU_LOCAL_LIST_T_FREE; node 524 kernel/bpf/bpf_lru_list.c node->ref = 0; node 525 kernel/bpf/bpf_lru_list.c list_move(&node->list, local_free_list(loc_l)); node 532 kernel/bpf/bpf_lru_list.c bpf_lru_list_push_free(&lru->common_lru.lru_list, node); node 536 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node) node 541 kernel/bpf/bpf_lru_list.c l = per_cpu_ptr(lru->percpu_lru, node->cpu); node 545 kernel/bpf/bpf_lru_list.c __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE); node 550 kernel/bpf/bpf_lru_list.c void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node) node 553 kernel/bpf/bpf_lru_list.c bpf_percpu_lru_push_free(lru, node); node 555 kernel/bpf/bpf_lru_list.c bpf_common_lru_push_free(lru, node); node 566 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 568 kernel/bpf/bpf_lru_list.c node = (struct bpf_lru_node *)(buf + node_offset); node 569 kernel/bpf/bpf_lru_list.c node->type = BPF_LRU_LIST_T_FREE; node 570 kernel/bpf/bpf_lru_list.c node->ref = 0; node 571 kernel/bpf/bpf_lru_list.c list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]); node 589 kernel/bpf/bpf_lru_list.c struct bpf_lru_node *node; node 593 kernel/bpf/bpf_lru_list.c node = (struct bpf_lru_node *)(buf + node_offset); node 594 kernel/bpf/bpf_lru_list.c node->cpu = cpu; node 595 kernel/bpf/bpf_lru_list.c node->type = BPF_LRU_LIST_T_FREE; node 596 kernel/bpf/bpf_lru_list.c node->ref = 0; node 597 kernel/bpf/bpf_lru_list.c list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]); node 50 kernel/bpf/bpf_lru_list.h typedef bool (*del_from_htab_func)(void *arg, struct bpf_lru_node *node); node 64 kernel/bpf/bpf_lru_list.h static inline void bpf_lru_node_set_ref(struct bpf_lru_node *node) node 69 kernel/bpf/bpf_lru_list.h if (!node->ref) node 70 kernel/bpf/bpf_lru_list.h node->ref = 1; node 79 kernel/bpf/bpf_lru_list.h void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node); node 80 kernel/bpf/bpf_lru_list.h void bpf_lru_promote(struct bpf_lru *lru, struct bpf_lru_node *node); node 50 kernel/bpf/cgroup.c list_for_each_entry_safe(pl, tmp, progs, node) { node 51 kernel/bpf/cgroup.c list_del(&pl->node); node 96 kernel/bpf/cgroup.c list_for_each_entry(pl, head, node) { node 166 kernel/bpf/cgroup.c list_for_each_entry(pl, &p->bpf.progs[type], node) { node 341 kernel/bpf/cgroup.c list_for_each_entry(pl, progs, node) { node 361 kernel/bpf/cgroup.c list_add_tail(&pl->node, progs); node 371 kernel/bpf/cgroup.c list_add_tail(&pl->node, progs); node 373 kernel/bpf/cgroup.c pl = list_first_entry(progs, typeof(*pl), node); node 415 kernel/bpf/cgroup.c list_del(&pl->node); node 454 kernel/bpf/cgroup.c list_for_each_entry(pl, progs, node) { node 470 kernel/bpf/cgroup.c pl = list_first_entry(progs, typeof(*pl), node); node 480 kernel/bpf/cgroup.c list_del(&pl->node); node 538 kernel/bpf/cgroup.c list_for_each_entry(pl, progs, node) { node 60 kernel/bpf/hashtab.c static bool htab_lru_map_delete_node(void *arg, struct bpf_lru_node *node); node 122 kernel/bpf/hashtab.c struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash); node 125 kernel/bpf/hashtab.c if (node) { node 126 kernel/bpf/hashtab.c l = container_of(node, struct htab_elem, lru_node); node 570 kernel/bpf/hashtab.c static bool htab_lru_map_delete_node(void *arg, struct bpf_lru_node *node) node 579 kernel/bpf/hashtab.c tgt_l = container_of(node, struct htab_elem, lru_node); node 53 kernel/bpf/local_storage.c struct rb_node *node; node 58 kernel/bpf/local_storage.c node = root->rb_node; node 59 kernel/bpf/local_storage.c while (node) { node 62 kernel/bpf/local_storage.c storage = container_of(node, struct bpf_cgroup_storage, node); node 66 kernel/bpf/local_storage.c node = node->rb_left; node 69 kernel/bpf/local_storage.c node = node->rb_right; node 93 kernel/bpf/local_storage.c this = container_of(*new, struct bpf_cgroup_storage, node); node 108 kernel/bpf/local_storage.c rb_link_node(&storage->node, parent, new); node 109 kernel/bpf/local_storage.c rb_insert_color(&storage->node, root); node 594 kernel/bpf/local_storage.c rb_erase(&storage->node, root); node 165 kernel/bpf/lpm_trie.c const struct lpm_trie_node *node, node 168 kernel/bpf/lpm_trie.c u32 limit = min(node->prefixlen, key->prefixlen); node 180 kernel/bpf/lpm_trie.c u64 diff = be64_to_cpu(*(__be64 *)node->data ^ node 193 kernel/bpf/lpm_trie.c u32 diff = be32_to_cpu(*(__be32 *)&node->data[i] ^ node 205 kernel/bpf/lpm_trie.c u16 diff = be16_to_cpu(*(__be16 *)&node->data[i] ^ node 217 kernel/bpf/lpm_trie.c prefixlen += 8 - fls(node->data[i] ^ key->data[i]); node 230 kernel/bpf/lpm_trie.c struct lpm_trie_node *node, *found = NULL; node 235 kernel/bpf/lpm_trie.c for (node = rcu_dereference(trie->root); node;) { node 243 kernel/bpf/lpm_trie.c matchlen = longest_prefix_match(trie, node, key); node 245 kernel/bpf/lpm_trie.c found = node; node 253 kernel/bpf/lpm_trie.c if (matchlen < node->prefixlen) node 259 kernel/bpf/lpm_trie.c if (!(node->flags & LPM_TREE_NODE_FLAG_IM)) node 260 kernel/bpf/lpm_trie.c found = node; node 266 kernel/bpf/lpm_trie.c next_bit = extract_bit(key->data, node->prefixlen); node 267 kernel/bpf/lpm_trie.c node = rcu_dereference(node->child[next_bit]); node 279 kernel/bpf/lpm_trie.c struct lpm_trie_node *node; node 285 kernel/bpf/lpm_trie.c node = kmalloc_node(size, GFP_ATOMIC | __GFP_NOWARN, node 287 kernel/bpf/lpm_trie.c if (!node) node 290 kernel/bpf/lpm_trie.c node->flags = 0; node 293 kernel/bpf/lpm_trie.c memcpy(node->data + trie->data_size, value, node 296 kernel/bpf/lpm_trie.c return node; node 304 kernel/bpf/lpm_trie.c struct lpm_trie_node *node, *im_node = NULL, *new_node = NULL; node 347 kernel/bpf/lpm_trie.c while ((node = rcu_dereference_protected(*slot, node 349 kernel/bpf/lpm_trie.c matchlen = longest_prefix_match(trie, node, key); node 351 kernel/bpf/lpm_trie.c if (node->prefixlen != matchlen || node 352 kernel/bpf/lpm_trie.c node->prefixlen == key->prefixlen || node 353 kernel/bpf/lpm_trie.c node->prefixlen == trie->max_prefixlen) node 356 kernel/bpf/lpm_trie.c next_bit = extract_bit(key->data, node->prefixlen); node 357 kernel/bpf/lpm_trie.c slot = &node->child[next_bit]; node 363 kernel/bpf/lpm_trie.c if (!node) { node 371 kernel/bpf/lpm_trie.c if (node->prefixlen == matchlen) { node 372 kernel/bpf/lpm_trie.c new_node->child[0] = node->child[0]; node 373 kernel/bpf/lpm_trie.c new_node->child[1] = node->child[1]; node 375 kernel/bpf/lpm_trie.c if (!(node->flags & LPM_TREE_NODE_FLAG_IM)) node 379 kernel/bpf/lpm_trie.c kfree_rcu(node, rcu); node 388 kernel/bpf/lpm_trie.c next_bit = extract_bit(node->data, matchlen); node 389 kernel/bpf/lpm_trie.c rcu_assign_pointer(new_node->child[next_bit], node); node 402 kernel/bpf/lpm_trie.c memcpy(im_node->data, node->data, trie->data_size); node 406 kernel/bpf/lpm_trie.c rcu_assign_pointer(im_node->child[0], node); node 410 kernel/bpf/lpm_trie.c rcu_assign_pointer(im_node->child[1], node); node 436 kernel/bpf/lpm_trie.c struct lpm_trie_node *node, *parent; node 456 kernel/bpf/lpm_trie.c while ((node = rcu_dereference_protected( node 458 kernel/bpf/lpm_trie.c matchlen = longest_prefix_match(trie, node, key); node 460 kernel/bpf/lpm_trie.c if (node->prefixlen != matchlen || node 461 kernel/bpf/lpm_trie.c node->prefixlen == key->prefixlen) node 464 kernel/bpf/lpm_trie.c parent = node; node 466 kernel/bpf/lpm_trie.c next_bit = extract_bit(key->data, node->prefixlen); node 467 kernel/bpf/lpm_trie.c trim = &node->child[next_bit]; node 470 kernel/bpf/lpm_trie.c if (!node || node->prefixlen != key->prefixlen || node 471 kernel/bpf/lpm_trie.c node->prefixlen != matchlen || node 472 kernel/bpf/lpm_trie.c (node->flags & LPM_TREE_NODE_FLAG_IM)) { node 482 kernel/bpf/lpm_trie.c if (rcu_access_pointer(node->child[0]) && node 483 kernel/bpf/lpm_trie.c rcu_access_pointer(node->child[1])) { node 484 kernel/bpf/lpm_trie.c node->flags |= LPM_TREE_NODE_FLAG_IM; node 496 kernel/bpf/lpm_trie.c !node->child[0] && !node->child[1]) { node 497 kernel/bpf/lpm_trie.c if (node == rcu_access_pointer(parent->child[0])) node 504 kernel/bpf/lpm_trie.c kfree_rcu(node, rcu); node 512 kernel/bpf/lpm_trie.c if (node->child[0]) node 513 kernel/bpf/lpm_trie.c rcu_assign_pointer(*trim, rcu_access_pointer(node->child[0])); node 514 kernel/bpf/lpm_trie.c else if (node->child[1]) node 515 kernel/bpf/lpm_trie.c rcu_assign_pointer(*trim, rcu_access_pointer(node->child[1])); node 518 kernel/bpf/lpm_trie.c kfree_rcu(node, rcu); node 590 kernel/bpf/lpm_trie.c struct lpm_trie_node *node; node 606 kernel/bpf/lpm_trie.c node = rcu_dereference_protected(*slot, 1); node 607 kernel/bpf/lpm_trie.c if (!node) node 610 kernel/bpf/lpm_trie.c if (rcu_access_pointer(node->child[0])) { node 611 kernel/bpf/lpm_trie.c slot = &node->child[0]; node 615 kernel/bpf/lpm_trie.c if (rcu_access_pointer(node->child[1])) { node 616 kernel/bpf/lpm_trie.c slot = &node->child[1]; node 620 kernel/bpf/lpm_trie.c kfree(node); node 632 kernel/bpf/lpm_trie.c struct lpm_trie_node *node, *next_node = NULL, *parent, *search_root; node 667 kernel/bpf/lpm_trie.c for (node = search_root; node;) { node 668 kernel/bpf/lpm_trie.c node_stack[++stack_ptr] = node; node 669 kernel/bpf/lpm_trie.c matchlen = longest_prefix_match(trie, node, key); node 670 kernel/bpf/lpm_trie.c if (node->prefixlen != matchlen || node 671 kernel/bpf/lpm_trie.c node->prefixlen == key->prefixlen) node 674 kernel/bpf/lpm_trie.c next_bit = extract_bit(key->data, node->prefixlen); node 675 kernel/bpf/lpm_trie.c node = rcu_dereference(node->child[next_bit]); node 677 kernel/bpf/lpm_trie.c if (!node || node->prefixlen != key->prefixlen || node 678 kernel/bpf/lpm_trie.c (node->flags & LPM_TREE_NODE_FLAG_IM)) node 684 kernel/bpf/lpm_trie.c node = node_stack[stack_ptr]; node 687 kernel/bpf/lpm_trie.c if (rcu_dereference(parent->child[0]) == node) { node 697 kernel/bpf/lpm_trie.c node = parent; node 709 kernel/bpf/lpm_trie.c for (node = search_root; node;) { node 710 kernel/bpf/lpm_trie.c if (node->flags & LPM_TREE_NODE_FLAG_IM) { node 711 kernel/bpf/lpm_trie.c node = rcu_dereference(node->child[0]); node 713 kernel/bpf/lpm_trie.c next_node = node; node 714 kernel/bpf/lpm_trie.c node = rcu_dereference(node->child[0]); node 715 kernel/bpf/lpm_trie.c if (!node) node 716 kernel/bpf/lpm_trie.c node = rcu_dereference(next_node->child[1]); node 29 kernel/bpf/percpu_freelist.c struct pcpu_freelist_node *node) node 32 kernel/bpf/percpu_freelist.c node->next = head->first; node 33 kernel/bpf/percpu_freelist.c head->first = node; node 38 kernel/bpf/percpu_freelist.c struct pcpu_freelist_node *node) node 42 kernel/bpf/percpu_freelist.c ___pcpu_freelist_push(head, node); node 46 kernel/bpf/percpu_freelist.c struct pcpu_freelist_node *node) node 51 kernel/bpf/percpu_freelist.c __pcpu_freelist_push(s, node); node 87 kernel/bpf/percpu_freelist.c struct pcpu_freelist_node *node; node 94 kernel/bpf/percpu_freelist.c node = head->first; node 95 kernel/bpf/percpu_freelist.c if (node) { node 96 kernel/bpf/percpu_freelist.c head->first = node->next; node 98 kernel/bpf/percpu_freelist.c return node; node 35 kernel/bpf/xskmap.c struct xsk_map_node *node; node 38 kernel/bpf/xskmap.c node = kzalloc(sizeof(*node), GFP_ATOMIC | __GFP_NOWARN); node 39 kernel/bpf/xskmap.c if (!node) node 44 kernel/bpf/xskmap.c kfree(node); node 48 kernel/bpf/xskmap.c node->map = map; node 49 kernel/bpf/xskmap.c node->map_entry = map_entry; node 50 kernel/bpf/xskmap.c return node; node 53 kernel/bpf/xskmap.c static void xsk_map_node_free(struct xsk_map_node *node) node 55 kernel/bpf/xskmap.c xsk_map_put(node->map); node 56 kernel/bpf/xskmap.c kfree(node); node 59 kernel/bpf/xskmap.c static void xsk_map_sock_add(struct xdp_sock *xs, struct xsk_map_node *node) node 62 kernel/bpf/xskmap.c list_add_tail(&node->node, &xs->map_list); node 72 kernel/bpf/xskmap.c list_for_each_entry_safe(n, tmp, &xs->map_list, node) { node 74 kernel/bpf/xskmap.c list_del(&n->node); node 221 kernel/bpf/xskmap.c struct xsk_map_node *node; node 247 kernel/bpf/xskmap.c node = xsk_map_node_alloc(m, map_entry); node 248 kernel/bpf/xskmap.c if (IS_ERR(node)) { node 250 kernel/bpf/xskmap.c return PTR_ERR(node); node 265 kernel/bpf/xskmap.c xsk_map_sock_add(xs, node); node 276 kernel/bpf/xskmap.c xsk_map_node_free(node); node 1668 kernel/cgroup/cgroup.c list_for_each_entry(cfts, &css->ss->cfts, node) node 1698 kernel/cgroup/cgroup.c list_for_each_entry(cfts, &css->ss->cfts, node) { node 1711 kernel/cgroup/cgroup.c list_for_each_entry(cfts, &css->ss->cfts, node) { node 4038 kernel/cgroup/cgroup.c list_del(&cfts->node); node 4095 kernel/cgroup/cgroup.c list_add_tail(&cfts->node, &ss->cfts); node 3417 kernel/cgroup/cpuset.c bool __cpuset_node_allowed(int node, gfp_t gfp_mask) node 3425 kernel/cgroup/cpuset.c if (node_isset(node, current->mems_allowed)) node 3444 kernel/cgroup/cpuset.c allowed = node_isset(node, cs->mems_allowed); node 65 kernel/cpu.c struct hlist_node *node; node 117 kernel/cpu.c struct hlist_node *node); node 122 kernel/cpu.c struct hlist_node *node); node 148 kernel/cpu.c bool bringup, struct hlist_node *node, node 153 kernel/cpu.c int (*cbm)(unsigned int cpu, struct hlist_node *node); node 181 kernel/cpu.c if (node) { node 183 kernel/cpu.c trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node); node 184 kernel/cpu.c ret = cbm(cpu, node); node 191 kernel/cpu.c hlist_for_each(node, &step->list) { node 192 kernel/cpu.c if (lastp && node == *lastp) node 195 kernel/cpu.c trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node); node 196 kernel/cpu.c ret = cbm(cpu, node); node 202 kernel/cpu.c *lastp = node; node 216 kernel/cpu.c hlist_for_each(node, &step->list) { node 220 kernel/cpu.c trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node); node 221 kernel/cpu.c ret = cbm(cpu, node); node 687 kernel/cpu.c st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); node 695 kernel/cpu.c st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); node 718 kernel/cpu.c struct hlist_node *node) node 737 kernel/cpu.c return cpuhp_invoke_callback(cpu, state, bringup, node, NULL); node 742 kernel/cpu.c st->node = node; node 763 kernel/cpu.c st->node = st->last = NULL; node 1622 kernel/cpu.c struct hlist_node *node) node 1640 kernel/cpu.c ret = cpuhp_invoke_ap_callback(cpu, state, bringup, node); node 1642 kernel/cpu.c ret = cpuhp_invoke_callback(cpu, state, bringup, node, NULL); node 1644 kernel/cpu.c ret = cpuhp_invoke_callback(cpu, state, bringup, node, NULL); node 1656 kernel/cpu.c struct hlist_node *node) node 1670 kernel/cpu.c cpuhp_issue_call(cpu, state, false, node); node 1675 kernel/cpu.c struct hlist_node *node, node 1704 kernel/cpu.c ret = cpuhp_issue_call(cpu, state, true, node); node 1707 kernel/cpu.c cpuhp_rollback_install(cpu, state, node); node 1713 kernel/cpu.c hlist_add_head(node, &sp->list); node 1719 kernel/cpu.c int __cpuhp_state_add_instance(enum cpuhp_state state, struct hlist_node *node, node 1725 kernel/cpu.c ret = __cpuhp_state_add_instance_cpuslocked(state, node, invoke); node 1824 kernel/cpu.c struct hlist_node *node, bool invoke) node 1849 kernel/cpu.c cpuhp_issue_call(cpu, state, false, node); node 1853 kernel/cpu.c hlist_del(node); node 337 kernel/dma/coherent.c unsigned long node = rmem->fdt_node; node 339 kernel/dma/coherent.c if (of_get_flat_dt_prop(node, "reusable", NULL)) node 343 kernel/dma/coherent.c if (!of_get_flat_dt_prop(node, "no-map", NULL)) { node 348 kernel/dma/coherent.c if (of_get_flat_dt_prop(node, "linux,dma-default", NULL)) { node 303 kernel/dma/contiguous.c unsigned long node = rmem->fdt_node; node 307 kernel/dma/contiguous.c if (!of_get_flat_dt_prop(node, "reusable", NULL) || node 308 kernel/dma/contiguous.c of_get_flat_dt_prop(node, "no-map", NULL)) node 324 kernel/dma/contiguous.c if (of_get_flat_dt_prop(node, "linux,cma-default", NULL)) node 90 kernel/dma/direct.c int node = dev_to_node(dev); node 108 kernel/dma/direct.c page = alloc_pages_node(node, gfp, get_order(alloc_size)); node 1557 kernel/events/core.c struct rb_node **node; node 1561 kernel/events/core.c node = &groups->tree.rb_node; node 1562 kernel/events/core.c parent = *node; node 1564 kernel/events/core.c while (*node) { node 1565 kernel/events/core.c parent = *node; node 1566 kernel/events/core.c node_event = container_of(*node, struct perf_event, group_node); node 1569 kernel/events/core.c node = &parent->rb_left; node 1571 kernel/events/core.c node = &parent->rb_right; node 1574 kernel/events/core.c rb_link_node(&event->group_node, parent, node); node 1623 kernel/events/core.c struct rb_node *node = groups->tree.rb_node; node 1625 kernel/events/core.c while (node) { node 1626 kernel/events/core.c node_event = container_of(node, struct perf_event, group_node); node 1629 kernel/events/core.c node = node->rb_left; node 1631 kernel/events/core.c node = node->rb_right; node 1634 kernel/events/core.c node = node->rb_left; node 9158 kernel/events/core.c int node = cpu_to_node(event->cpu == -1 ? 0 : event->cpu); node 9161 kernel/events/core.c filter = kzalloc_node(sizeof(*filter), GFP_KERNEL, node); node 567 kernel/events/ring_buffer.c static struct page *rb_alloc_aux_page(int node, int order) node 575 kernel/events/ring_buffer.c page = alloc_pages_node(node, PERF_AUX_GFP, order); node 633 kernel/events/ring_buffer.c int node = (event->cpu == -1) ? -1 : cpu_to_node(event->cpu); node 657 kernel/events/ring_buffer.c node); node 667 kernel/events/ring_buffer.c page = rb_alloc_aux_page(node, order); node 747 kernel/events/ring_buffer.c int node; node 749 kernel/events/ring_buffer.c node = (cpu == -1) ? cpu : cpu_to_node(cpu); node 750 kernel/events/ring_buffer.c page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); node 167 kernel/fork.c static inline struct task_struct *alloc_task_struct_node(int node) node 169 kernel/fork.c return kmem_cache_alloc_node(task_struct_cachep, GFP_KERNEL, node); node 213 kernel/fork.c static unsigned long *alloc_thread_stack_node(struct task_struct *tsk, int node) node 244 kernel/fork.c 0, node, __builtin_return_address(0)); node 257 kernel/fork.c struct page *page = alloc_pages_node(node, THREADINFO_GFP, node 303 kernel/fork.c int node) node 306 kernel/fork.c stack = kmem_cache_alloc_node(thread_stack_cache, THREADINFO_GFP, node); node 856 kernel/fork.c static struct task_struct *dup_task_struct(struct task_struct *orig, int node) node 863 kernel/fork.c if (node == NUMA_NO_NODE) node 864 kernel/fork.c node = tsk_fork_get_node(orig); node 865 kernel/fork.c tsk = alloc_task_struct_node(node); node 869 kernel/fork.c stack = alloc_thread_stack_node(tsk, node); node 1763 kernel/fork.c int node, node 1835 kernel/fork.c INIT_HLIST_NODE(&delayed.node); node 1839 kernel/fork.c hlist_add_head(&delayed.node, ¤t->signal->multiprocess); node 1847 kernel/fork.c p = dup_task_struct(current, node); node 2211 kernel/fork.c hlist_del_init(&delayed.node); node 2285 kernel/fork.c hlist_del_init(&delayed.node); node 145 kernel/gcov/fs.c static struct gcov_info *get_node_info(struct gcov_node *node) node 147 kernel/gcov/fs.c if (node->num_loaded > 0) node 148 kernel/gcov/fs.c return node->loaded_info[0]; node 150 kernel/gcov/fs.c return node->unloaded_info; node 157 kernel/gcov/fs.c static struct gcov_info *get_accumulated_info(struct gcov_node *node) node 162 kernel/gcov/fs.c if (node->unloaded_info) node 163 kernel/gcov/fs.c info = gcov_info_dup(node->unloaded_info); node 165 kernel/gcov/fs.c info = gcov_info_dup(node->loaded_info[i++]); node 168 kernel/gcov/fs.c for (; i < node->num_loaded; i++) node 169 kernel/gcov/fs.c gcov_info_add(info, node->loaded_info[i]); node 180 kernel/gcov/fs.c struct gcov_node *node = inode->i_private; node 192 kernel/gcov/fs.c info = get_accumulated_info(node); node 240 kernel/gcov/fs.c struct gcov_node *node; node 243 kernel/gcov/fs.c list_for_each_entry(node, &all_head, all) { node 244 kernel/gcov/fs.c info = get_node_info(node); node 246 kernel/gcov/fs.c return node; node 255 kernel/gcov/fs.c static void reset_node(struct gcov_node *node) node 259 kernel/gcov/fs.c if (node->unloaded_info) node 260 kernel/gcov/fs.c gcov_info_reset(node->unloaded_info); node 261 kernel/gcov/fs.c for (i = 0; i < node->num_loaded; i++) node 262 kernel/gcov/fs.c gcov_info_reset(node->loaded_info[i]); node 265 kernel/gcov/fs.c static void remove_node(struct gcov_node *node); node 277 kernel/gcov/fs.c struct gcov_node *node; node 282 kernel/gcov/fs.c node = get_node_by_name(gcov_info_filename(info)); node 283 kernel/gcov/fs.c if (node) { node 285 kernel/gcov/fs.c if (node->num_loaded == 0) node 286 kernel/gcov/fs.c remove_node(node); node 288 kernel/gcov/fs.c reset_node(node); node 366 kernel/gcov/fs.c static void add_links(struct gcov_node *node, struct dentry *parent) node 375 kernel/gcov/fs.c node->links = kcalloc(num, sizeof(struct dentry *), GFP_KERNEL); node 376 kernel/gcov/fs.c if (!node->links) node 380 kernel/gcov/fs.c gcov_info_filename(get_node_info(node)), node 387 kernel/gcov/fs.c node->links[i] = debugfs_create_symlink(deskew(basename), node 396 kernel/gcov/fs.c debugfs_remove(node->links[i]); node 397 kernel/gcov/fs.c kfree(node->links); node 398 kernel/gcov/fs.c node->links = NULL; node 410 kernel/gcov/fs.c static void init_node(struct gcov_node *node, struct gcov_info *info, node 413 kernel/gcov/fs.c INIT_LIST_HEAD(&node->list); node 414 kernel/gcov/fs.c INIT_LIST_HEAD(&node->children); node 415 kernel/gcov/fs.c INIT_LIST_HEAD(&node->all); node 416 kernel/gcov/fs.c if (node->loaded_info) { node 417 kernel/gcov/fs.c node->loaded_info[0] = info; node 418 kernel/gcov/fs.c node->num_loaded = 1; node 420 kernel/gcov/fs.c node->parent = parent; node 422 kernel/gcov/fs.c strcpy(node->name, name); node 432 kernel/gcov/fs.c struct gcov_node *node; node 434 kernel/gcov/fs.c node = kzalloc(sizeof(struct gcov_node) + strlen(name) + 1, GFP_KERNEL); node 435 kernel/gcov/fs.c if (!node) node 438 kernel/gcov/fs.c node->loaded_info = kcalloc(1, sizeof(struct gcov_info *), node 440 kernel/gcov/fs.c if (!node->loaded_info) node 443 kernel/gcov/fs.c init_node(node, info, name, parent); node 446 kernel/gcov/fs.c node->dentry = debugfs_create_file(deskew(node->name), 0600, node 447 kernel/gcov/fs.c parent->dentry, node, &gcov_data_fops); node 449 kernel/gcov/fs.c node->dentry = debugfs_create_dir(node->name, parent->dentry); node 451 kernel/gcov/fs.c add_links(node, parent->dentry); node 452 kernel/gcov/fs.c list_add(&node->list, &parent->children); node 453 kernel/gcov/fs.c list_add(&node->all, &all_head); node 455 kernel/gcov/fs.c return node; node 458 kernel/gcov/fs.c kfree(node); node 464 kernel/gcov/fs.c static void remove_links(struct gcov_node *node) node 468 kernel/gcov/fs.c if (!node->links) node 471 kernel/gcov/fs.c debugfs_remove(node->links[i]); node 472 kernel/gcov/fs.c kfree(node->links); node 473 kernel/gcov/fs.c node->links = NULL; node 480 kernel/gcov/fs.c static void release_node(struct gcov_node *node) node 482 kernel/gcov/fs.c list_del(&node->list); node 483 kernel/gcov/fs.c list_del(&node->all); node 484 kernel/gcov/fs.c debugfs_remove(node->dentry); node 485 kernel/gcov/fs.c remove_links(node); node 486 kernel/gcov/fs.c kfree(node->loaded_info); node 487 kernel/gcov/fs.c if (node->unloaded_info) node 488 kernel/gcov/fs.c gcov_info_free(node->unloaded_info); node 489 kernel/gcov/fs.c kfree(node); node 493 kernel/gcov/fs.c static void remove_node(struct gcov_node *node) node 497 kernel/gcov/fs.c while ((node != &root_node) && list_empty(&node->children)) { node 498 kernel/gcov/fs.c parent = node->parent; node 499 kernel/gcov/fs.c release_node(node); node 500 kernel/gcov/fs.c node = parent; node 511 kernel/gcov/fs.c struct gcov_node *node; node 513 kernel/gcov/fs.c list_for_each_entry(node, &parent->children, list) { node 514 kernel/gcov/fs.c if (strcmp(node->name, name) == 0) node 515 kernel/gcov/fs.c return node; node 528 kernel/gcov/fs.c struct gcov_node *node; node 532 kernel/gcov/fs.c list_for_each_entry(node, &all_head, all) { node 533 kernel/gcov/fs.c if (node->num_loaded > 0) node 534 kernel/gcov/fs.c reset_node(node); node 535 kernel/gcov/fs.c else if (list_empty(&node->children)) { node 536 kernel/gcov/fs.c remove_node(node); node 570 kernel/gcov/fs.c struct gcov_node *node; node 589 kernel/gcov/fs.c node = get_child_by_name(parent, curr); node 590 kernel/gcov/fs.c if (!node) { node 591 kernel/gcov/fs.c node = new_node(parent, NULL, curr); node 592 kernel/gcov/fs.c if (!node) node 595 kernel/gcov/fs.c parent = node; node 598 kernel/gcov/fs.c node = new_node(parent, info, curr); node 599 kernel/gcov/fs.c if (!node) node 614 kernel/gcov/fs.c static void add_info(struct gcov_node *node, struct gcov_info *info) node 617 kernel/gcov/fs.c int num = node->num_loaded; node 630 kernel/gcov/fs.c memcpy(loaded_info, node->loaded_info, node 639 kernel/gcov/fs.c if (!gcov_info_is_compatible(node->unloaded_info, info)) { node 643 kernel/gcov/fs.c gcov_info_free(node->unloaded_info); node 644 kernel/gcov/fs.c node->unloaded_info = NULL; node 651 kernel/gcov/fs.c if (!gcov_info_is_compatible(node->loaded_info[0], info)) { node 659 kernel/gcov/fs.c kfree(node->loaded_info); node 660 kernel/gcov/fs.c node->loaded_info = loaded_info; node 661 kernel/gcov/fs.c node->num_loaded = num + 1; node 667 kernel/gcov/fs.c static int get_info_index(struct gcov_node *node, struct gcov_info *info) node 671 kernel/gcov/fs.c for (i = 0; i < node->num_loaded; i++) { node 672 kernel/gcov/fs.c if (node->loaded_info[i] == info) node 681 kernel/gcov/fs.c static void save_info(struct gcov_node *node, struct gcov_info *info) node 683 kernel/gcov/fs.c if (node->unloaded_info) node 684 kernel/gcov/fs.c gcov_info_add(node->unloaded_info, info); node 686 kernel/gcov/fs.c node->unloaded_info = gcov_info_dup(info); node 687 kernel/gcov/fs.c if (!node->unloaded_info) { node 699 kernel/gcov/fs.c static void remove_info(struct gcov_node *node, struct gcov_info *info) node 703 kernel/gcov/fs.c i = get_info_index(node, info); node 710 kernel/gcov/fs.c save_info(node, info); node 712 kernel/gcov/fs.c node->loaded_info[i] = node->loaded_info[node->num_loaded - 1]; node 713 kernel/gcov/fs.c node->num_loaded--; node 714 kernel/gcov/fs.c if (node->num_loaded > 0) node 717 kernel/gcov/fs.c kfree(node->loaded_info); node 718 kernel/gcov/fs.c node->loaded_info = NULL; node 719 kernel/gcov/fs.c node->num_loaded = 0; node 720 kernel/gcov/fs.c if (!node->unloaded_info) node 721 kernel/gcov/fs.c remove_node(node); node 730 kernel/gcov/fs.c struct gcov_node *node; node 733 kernel/gcov/fs.c node = get_node_by_name(gcov_info_filename(info)); node 736 kernel/gcov/fs.c if (node) node 737 kernel/gcov/fs.c add_info(node, info); node 742 kernel/gcov/fs.c if (node) node 743 kernel/gcov/fs.c remove_info(node, info); node 46 kernel/irq/affinity.c int node; node 52 kernel/irq/affinity.c for (node = 0; node < nr_node_ids; node++) { node 53 kernel/irq/affinity.c if (!zalloc_cpumask_var(&masks[node], GFP_KERNEL)) node 60 kernel/irq/affinity.c while (--node >= 0) node 61 kernel/irq/affinity.c free_cpumask_var(masks[node]); node 68 kernel/irq/affinity.c int node; node 70 kernel/irq/affinity.c for (node = 0; node < nr_node_ids; node++) node 71 kernel/irq/affinity.c free_cpumask_var(masks[node]); node 179 kernel/irq/devres.c unsigned int cnt, int node, struct module *owner, node 189 kernel/irq/devres.c base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); node 104 kernel/irq/internals.h extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr); node 55 kernel/irq/irqdesc.c static int alloc_masks(struct irq_desc *desc, int node) node 58 kernel/irq/irqdesc.c GFP_KERNEL, node)) node 63 kernel/irq/irqdesc.c GFP_KERNEL, node)) { node 70 kernel/irq/irqdesc.c if (!zalloc_cpumask_var_node(&desc->pending_mask, GFP_KERNEL, node)) { node 81 kernel/irq/irqdesc.c static void desc_smp_init(struct irq_desc *desc, int node, node 92 kernel/irq/irqdesc.c desc->irq_common_data.node = node; node 98 kernel/irq/irqdesc.c alloc_masks(struct irq_desc *desc, int node) { return 0; } node 100 kernel/irq/irqdesc.c desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { } node 103 kernel/irq/irqdesc.c static void desc_set_defaults(unsigned int irq, struct irq_desc *desc, int node, node 127 kernel/irq/irqdesc.c desc_smp_init(desc, node, affinity); node 387 kernel/irq/irqdesc.c static struct irq_desc *alloc_desc(int irq, int node, unsigned int flags, node 393 kernel/irq/irqdesc.c desc = kzalloc_node(sizeof(*desc), GFP_KERNEL, node); node 401 kernel/irq/irqdesc.c if (alloc_masks(desc, node)) node 409 kernel/irq/irqdesc.c desc_set_defaults(irq, desc, node, affinity, owner); node 466 kernel/irq/irqdesc.c static int alloc_descs(unsigned int start, unsigned int cnt, int node, node 491 kernel/irq/irqdesc.c node = cpu_to_node(cpumask_first(mask)); node 495 kernel/irq/irqdesc.c desc = alloc_desc(start + i, node, flags, mask, owner); node 521 kernel/irq/irqdesc.c int i, initcnt, node = first_online_node; node 541 kernel/irq/irqdesc.c desc = alloc_desc(i, node, 0, NULL, NULL); node 560 kernel/irq/irqdesc.c int count, i, node = first_online_node; node 572 kernel/irq/irqdesc.c alloc_masks(&desc[i], node); node 576 kernel/irq/irqdesc.c desc_set_defaults(i, &desc[i], node, NULL, NULL); node 597 kernel/irq/irqdesc.c static inline int alloc_descs(unsigned int start, unsigned int cnt, int node, node 766 kernel/irq/irqdesc.c __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, node 800 kernel/irq/irqdesc.c ret = alloc_descs(start, cnt, node, affinity, owner); node 815 kernel/irq/irqdesc.c unsigned int irq_alloc_hwirqs(int cnt, int node) node 817 kernel/irq/irqdesc.c int i, irq = __irq_alloc_descs(-1, 0, cnt, node, NULL, NULL); node 823 kernel/irq/irqdesc.c if (arch_setup_hwirq(i, node)) node 1011 kernel/irq/irqdomain.c int node, const struct irq_affinity_desc *affinity) node 1016 kernel/irq/irqdomain.c virq = __irq_alloc_descs(virq, virq, cnt, node, THIS_MODULE, node 1022 kernel/irq/irqdomain.c virq = __irq_alloc_descs(-1, hint, cnt, node, THIS_MODULE, node 1025 kernel/irq/irqdomain.c virq = __irq_alloc_descs(-1, 1, cnt, node, THIS_MODULE, node 1327 kernel/irq/irqdomain.c unsigned int nr_irqs, int node, void *arg, node 1341 kernel/irq/irqdomain.c virq = irq_domain_alloc_descs(irq_base, nr_irqs, 0, node, node 409 kernel/irq/manage.c int ret, node = irq_desc_get_node(desc); node 435 kernel/irq/manage.c if (node != NUMA_NO_NODE) { node 436 kernel/irq/manage.c const struct cpumask *nodemask = cpumask_of_node(node); node 37 kernel/kthread.c int node; node 275 kernel/kthread.c current->pref_node_fork = create->node; node 294 kernel/kthread.c void *data, int node, node 307 kernel/kthread.c create->node = node; node 380 kernel/kthread.c void *data, int node, node 388 kernel/kthread.c task = __kthread_create_on_node(threadfn, data, node, namefmt, args); node 663 kernel/kthread.c struct kthread_work, node); node 664 kernel/kthread.c list_del_init(&work->node); node 687 kernel/kthread.c int node = NUMA_NO_NODE; node 696 kernel/kthread.c node = cpu_to_node(cpu); node 699 kernel/kthread.c node, namefmt, args); node 781 kernel/kthread.c return !list_empty(&work->node) || work->canceling; node 788 kernel/kthread.c WARN_ON_ONCE(!list_empty(&work->node)); node 800 kernel/kthread.c list_add_tail(&work->node, pos); node 861 kernel/kthread.c WARN_ON_ONCE(list_empty(&work->node)); node 862 kernel/kthread.c list_del_init(&work->node); node 892 kernel/kthread.c list_add(&work->node, &worker->delayed_work_list); node 968 kernel/kthread.c if (!list_empty(&work->node)) node 969 kernel/kthread.c kthread_insert_work(worker, &fwork.work, work->node.next); node 1019 kernel/kthread.c if (!list_empty(&work->node)) { node 1020 kernel/kthread.c list_del_init(&work->node); node 583 kernel/livepatch/core.c list_del(&func->node); node 613 kernel/livepatch/core.c list_del(&obj->node); node 802 kernel/livepatch/core.c list_add_tail(&func->node, &obj->func_list); node 810 kernel/livepatch/core.c list_add_tail(&obj->node, &patch->obj_list); node 30 kernel/livepatch/patch.c list_for_each_entry(ops, &klp_ops, node) { node 160 kernel/livepatch/patch.c list_del(&ops->node); node 201 kernel/livepatch/patch.c list_add(&ops->node, &klp_ops); node 232 kernel/livepatch/patch.c list_del(&ops->node); node 23 kernel/livepatch/patch.h struct list_head node; node 55 kernel/livepatch/shadow.c struct hlist_node node; node 89 kernel/livepatch/shadow.c hash_for_each_possible_rcu(klp_shadow_hash, shadow, node, node 157 kernel/livepatch/shadow.c hash_add_rcu(klp_shadow_hash, &new_shadow->node, node 237 kernel/livepatch/shadow.c hash_del_rcu(&shadow->node); node 261 kernel/livepatch/shadow.c hash_for_each_possible(klp_shadow_hash, shadow, node, node 292 kernel/livepatch/shadow.c hash_for_each(klp_shadow_hash, i, shadow, node) { node 65 kernel/locking/mcs_spinlock.h void mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node) node 70 kernel/locking/mcs_spinlock.h node->locked = 0; node 71 kernel/locking/mcs_spinlock.h node->next = NULL; node 79 kernel/locking/mcs_spinlock.h prev = xchg(lock, node); node 91 kernel/locking/mcs_spinlock.h WRITE_ONCE(prev->next, node); node 94 kernel/locking/mcs_spinlock.h arch_mcs_spin_lock_contended(&node->locked); node 102 kernel/locking/mcs_spinlock.h void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *node) node 104 kernel/locking/mcs_spinlock.h struct mcs_spinlock *next = READ_ONCE(node->next); node 110 kernel/locking/mcs_spinlock.h if (likely(cmpxchg_release(lock, node, NULL) == node)) node 113 kernel/locking/mcs_spinlock.h while (!(next = READ_ONCE(node->next))) node 25 kernel/locking/osq_lock.c static inline int node_cpu(struct optimistic_spin_node *node) node 27 kernel/locking/osq_lock.c return node->cpu - 1; node 43 kernel/locking/osq_lock.c struct optimistic_spin_node *node, node 78 kernel/locking/osq_lock.c if (node->next) { node 79 kernel/locking/osq_lock.c next = xchg(&node->next, NULL); node 92 kernel/locking/osq_lock.c struct optimistic_spin_node *node = this_cpu_ptr(&osq_node); node 97 kernel/locking/osq_lock.c node->locked = 0; node 98 kernel/locking/osq_lock.c node->next = NULL; node 99 kernel/locking/osq_lock.c node->cpu = curr; node 112 kernel/locking/osq_lock.c node->prev = prev; node 126 kernel/locking/osq_lock.c WRITE_ONCE(prev->next, node); node 137 kernel/locking/osq_lock.c while (!READ_ONCE(node->locked)) { node 143 kernel/locking/osq_lock.c if (need_resched() || vcpu_is_preempted(node_cpu(node->prev))) node 160 kernel/locking/osq_lock.c if (prev->next == node && node 161 kernel/locking/osq_lock.c cmpxchg(&prev->next, node, NULL) == node) node 169 kernel/locking/osq_lock.c if (smp_load_acquire(&node->locked)) node 178 kernel/locking/osq_lock.c prev = READ_ONCE(node->prev); node 188 kernel/locking/osq_lock.c next = osq_wait_next(lock, node, prev); node 208 kernel/locking/osq_lock.c struct optimistic_spin_node *node, *next; node 221 kernel/locking/osq_lock.c node = this_cpu_ptr(&osq_node); node 222 kernel/locking/osq_lock.c next = xchg(&node->next, NULL); node 228 kernel/locking/osq_lock.c next = osq_wait_next(lock, node, NULL); node 271 kernel/locking/qspinlock.c static __always_inline void __pv_init_node(struct mcs_spinlock *node) { } node 272 kernel/locking/qspinlock.c static __always_inline void __pv_wait_node(struct mcs_spinlock *node, node 275 kernel/locking/qspinlock.c struct mcs_spinlock *node) { } node 277 kernel/locking/qspinlock.c struct mcs_spinlock *node) node 316 kernel/locking/qspinlock.c struct mcs_spinlock *prev, *next, *node; node 399 kernel/locking/qspinlock.c node = this_cpu_ptr(&qnodes[0].mcs); node 400 kernel/locking/qspinlock.c idx = node->count++; node 419 kernel/locking/qspinlock.c node = grab_mcs_node(node, idx); node 433 kernel/locking/qspinlock.c node->locked = 0; node 434 kernel/locking/qspinlock.c node->next = NULL; node 435 kernel/locking/qspinlock.c pv_init_node(node); node 470 kernel/locking/qspinlock.c WRITE_ONCE(prev->next, node); node 472 kernel/locking/qspinlock.c pv_wait_node(node, prev); node 473 kernel/locking/qspinlock.c arch_mcs_spin_lock_contended(&node->locked); node 481 kernel/locking/qspinlock.c next = READ_ONCE(node->next); node 507 kernel/locking/qspinlock.c if ((val = pv_wait_head_or_lock(lock, node))) node 550 kernel/locking/qspinlock.c next = smp_cond_load_relaxed(&node->next, (VAL)); node 173 kernel/locking/qspinlock_paravirt.h struct pv_node *node; node 212 kernel/locking/qspinlock_paravirt.h static struct qspinlock **pv_hash(struct qspinlock *lock, struct pv_node *node) node 221 kernel/locking/qspinlock_paravirt.h WRITE_ONCE(he->node, node); node 243 kernel/locking/qspinlock_paravirt.h struct pv_node *node; node 247 kernel/locking/qspinlock_paravirt.h node = READ_ONCE(he->node); node 249 kernel/locking/qspinlock_paravirt.h return node; node 278 kernel/locking/qspinlock_paravirt.h static void pv_init_node(struct mcs_spinlock *node) node 280 kernel/locking/qspinlock_paravirt.h struct pv_node *pn = (struct pv_node *)node; node 293 kernel/locking/qspinlock_paravirt.h static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev) node 295 kernel/locking/qspinlock_paravirt.h struct pv_node *pn = (struct pv_node *)node; node 302 kernel/locking/qspinlock_paravirt.h if (READ_ONCE(node->locked)) node 322 kernel/locking/qspinlock_paravirt.h if (!READ_ONCE(node->locked)) { node 343 kernel/locking/qspinlock_paravirt.h !READ_ONCE(node->locked)); node 360 kernel/locking/qspinlock_paravirt.h static void pv_kick_node(struct qspinlock *lock, struct mcs_spinlock *node) node 362 kernel/locking/qspinlock_paravirt.h struct pv_node *pn = (struct pv_node *)node; node 403 kernel/locking/qspinlock_paravirt.h pv_wait_head_or_lock(struct qspinlock *lock, struct mcs_spinlock *node) node 405 kernel/locking/qspinlock_paravirt.h struct pv_node *pn = (struct pv_node *)node; node 495 kernel/locking/qspinlock_paravirt.h struct pv_node *node; node 517 kernel/locking/qspinlock_paravirt.h node = pv_unhash(lock); node 533 kernel/locking/qspinlock_paravirt.h pv_kick(node->cpu); node 107 kernel/module.c struct module_layout *layout = container_of(n, struct module_layout, mtn.node); node 114 kernel/module.c struct module_layout *layout = container_of(n, struct module_layout, mtn.node); node 158 kernel/module.c static noinline void __mod_tree_insert(struct mod_tree_node *node) node 160 kernel/module.c latch_tree_insert(&node->node, &mod_tree.root, &mod_tree_ops); node 163 kernel/module.c static void __mod_tree_remove(struct mod_tree_node *node) node 165 kernel/module.c latch_tree_erase(&node->node, &mod_tree.root, &mod_tree_ops); node 202 kernel/module.c return container_of(ltn, struct mod_tree_node, node)->mod; node 3527 kernel/module.c struct llist_node node; node 3541 kernel/module.c initfree = container_of(pos, struct mod_initfree, node); node 3650 kernel/module.c if (llist_add(&freeinit->node, &init_free_list)) node 780 kernel/padata.c static int padata_cpu_online(unsigned int cpu, struct hlist_node *node) node 785 kernel/padata.c pinst = hlist_entry_safe(node, struct padata_instance, cpu_online_node); node 795 kernel/padata.c static int padata_cpu_dead(unsigned int cpu, struct hlist_node *node) node 800 kernel/padata.c pinst = hlist_entry_safe(node, struct padata_instance, cpu_dead_node); node 104 kernel/power/qos.c struct plist_node *node; node 118 kernel/power/qos.c plist_for_each(node, &c->list) node 119 kernel/power/qos.c total_value += node->prio; node 181 kernel/power/qos.c plist_for_each_entry(req, &c->list, node) { node 184 kernel/power/qos.c if ((req->node).prio != c->default_value) { node 190 kernel/power/qos.c (req->node).prio, state); node 214 kernel/power/qos.c int pm_qos_update_target(struct pm_qos_constraints *c, struct plist_node *node, node 230 kernel/power/qos.c plist_del(node, &c->list); node 238 kernel/power/qos.c plist_del(node, &c->list); node 241 kernel/power/qos.c plist_node_init(node, new_value); node 242 kernel/power/qos.c plist_add(node, &c->list); node 277 kernel/power/qos.c list_del(&req->node); node 278 kernel/power/qos.c list_for_each_entry(req, &pqf->list, node) node 315 kernel/power/qos.c INIT_LIST_HEAD(&req->node); node 316 kernel/power/qos.c list_add_tail(&req->node, &pqf->list); node 355 kernel/power/qos.c if (new_value != req->node.prio) node 358 kernel/power/qos.c &req->node, PM_QOS_UPDATE_REQ, new_value); node 403 kernel/power/qos.c &req->node, PM_QOS_ADD_REQ, value); node 454 kernel/power/qos.c if (new_value != req->node.prio) node 457 kernel/power/qos.c &req->node, PM_QOS_UPDATE_REQ, new_value); node 485 kernel/power/qos.c &req->node, PM_QOS_REMOVE_REQ, node 372 kernel/power/snapshot.c struct rtree_node *node; node 406 kernel/power/snapshot.c struct rtree_node *node; node 408 kernel/power/snapshot.c node = chain_alloc(ca, sizeof(struct rtree_node)); node 409 kernel/power/snapshot.c if (!node) node 412 kernel/power/snapshot.c node->data = get_image_page(gfp_mask, safe_needed); node 413 kernel/power/snapshot.c if (!node->data) node 416 kernel/power/snapshot.c list_add_tail(&node->list, list); node 418 kernel/power/snapshot.c return node; node 431 kernel/power/snapshot.c struct rtree_node *node, *block, **dst; node 446 kernel/power/snapshot.c node = alloc_rtree_node(gfp_mask, safe_needed, ca, node 448 kernel/power/snapshot.c if (!node) node 451 kernel/power/snapshot.c node->data[0] = (unsigned long)zone->rtree; node 452 kernel/power/snapshot.c zone->rtree = node; node 462 kernel/power/snapshot.c node = zone->rtree; node 468 kernel/power/snapshot.c if (!node) { node 469 kernel/power/snapshot.c node = alloc_rtree_node(gfp_mask, safe_needed, ca, node 471 kernel/power/snapshot.c if (!node) node 473 kernel/power/snapshot.c *dst = node; node 479 kernel/power/snapshot.c node = *dst; node 539 kernel/power/snapshot.c struct rtree_node *node; node 541 kernel/power/snapshot.c list_for_each_entry(node, &zone->nodes, list) node 542 kernel/power/snapshot.c free_image_page(node->data, clear_nosave_free); node 544 kernel/power/snapshot.c list_for_each_entry(node, &zone->leaves, list) node 545 kernel/power/snapshot.c free_image_page(node->data, clear_nosave_free); node 552 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, node 711 kernel/power/snapshot.c struct rtree_node *node; node 743 kernel/power/snapshot.c node = bm->cur.node; node 748 kernel/power/snapshot.c node = zone->rtree; node 756 kernel/power/snapshot.c BUG_ON(node->data[index] == 0); node 757 kernel/power/snapshot.c node = (struct rtree_node *)node->data[index]; node 763 kernel/power/snapshot.c bm->cur.node = node; node 767 kernel/power/snapshot.c *addr = node->data; node 813 kernel/power/snapshot.c clear_bit(bit, bm->cur.node->data); node 847 kernel/power/snapshot.c if (!list_is_last(&bm->cur.node->list, &bm->cur.zone->leaves)) { node 848 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.node->list.next, node 860 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, node 890 kernel/power/snapshot.c bit = find_next_bit(bm->cur.node->data, bits, node 916 kernel/power/snapshot.c struct rtree_node *node; node 918 kernel/power/snapshot.c list_for_each_entry(node, &zone->nodes, list) node 919 kernel/power/snapshot.c recycle_safe_page(node->data); node 921 kernel/power/snapshot.c list_for_each_entry(node, &zone->leaves, list) node 922 kernel/power/snapshot.c recycle_safe_page(node->data); node 123 kernel/power/swap.c struct rb_node node; node 138 kernel/power/swap.c ext = rb_entry(*new, struct swsusp_extent, node); node 166 kernel/power/swap.c rb_link_node(&ext->node, parent, new); node 167 kernel/power/swap.c rb_insert_color(&ext->node, &swsusp_extents); node 198 kernel/power/swap.c struct rb_node *node; node 200 kernel/power/swap.c while ((node = swsusp_extents.rb_node)) { node 204 kernel/power/swap.c ext = rb_entry(node, struct swsusp_extent, node); node 205 kernel/power/swap.c rb_erase(node, &swsusp_extents); node 29 kernel/power/wakelock.c struct rb_node node; node 40 kernel/power/wakelock.c struct rb_node *node; node 47 kernel/power/wakelock.c for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) { node 48 kernel/power/wakelock.c wl = rb_entry(node, struct wakelock, node); node 125 kernel/power/wakelock.c rb_erase(&wl->node, &wakelocks_tree); node 153 kernel/power/wakelock.c struct rb_node **node = &wakelocks_tree.rb_node; node 154 kernel/power/wakelock.c struct rb_node *parent = *node; node 157 kernel/power/wakelock.c while (*node) { node 160 kernel/power/wakelock.c parent = *node; node 161 kernel/power/wakelock.c wl = rb_entry(*node, struct wakelock, node); node 170 kernel/power/wakelock.c node = &(*node)->rb_left; node 172 kernel/power/wakelock.c node = &(*node)->rb_right; node 199 kernel/power/wakelock.c rb_link_node(&wl->node, parent, node); node 200 kernel/power/wakelock.c rb_insert_color(&wl->node, &wakelocks_tree); node 354 kernel/profile.c int i, node = cpu_to_mem(cpu); node 363 kernel/profile.c page = __alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); node 326 kernel/rcu/rcu.h #define rcu_is_last_leaf_node(rnp) ((rnp) == &rcu_state.node[rcu_num_nodes - 1]) node 334 kernel/rcu/rcu.h for ((rnp) = &(sp)->node[0]; \ node 335 kernel/rcu/rcu.h (rnp) < &(sp)->node[rcu_num_nodes]; (rnp)++) node 347 kernel/rcu/rcu.h (rnp) < &rcu_state.node[rcu_num_nodes]; (rnp)++) node 94 kernel/rcu/srcutree.c ssp->level[0] = &ssp->node[0]; node 111 kernel/rcu/srcutree.c if (snp == &ssp->node[0]) { node 88 kernel/rcu/tree.c .level = { &rcu_state.node[0] }, node 527 kernel/rcu/tree.c return &rcu_state.node[0]; node 291 kernel/rcu/tree.h struct rcu_node node[NUM_RCU_NODES]; /* Hierarchy. */ node 1629 kernel/resource.c INIT_LIST_HEAD(&entry->node); node 1641 kernel/resource.c list_for_each_entry_safe(entry, tmp, head, node) node 415 kernel/sched/core.c struct wake_q_node *node = &task->wake_q; node 426 kernel/sched/core.c if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL))) node 432 kernel/sched/core.c *head->lastp = node; node 433 kernel/sched/core.c head->lastp = &node->next; node 480 kernel/sched/core.c struct wake_q_node *node = head->first; node 482 kernel/sched/core.c while (node != WAKE_Q_TAIL) { node 485 kernel/sched/core.c task = container_of(node, struct task_struct, wake_q); node 488 kernel/sched/core.c node = node->next; node 821 kernel/sched/debug.c void print_numa_stats(struct seq_file *m, int node, unsigned long tsf, node 824 kernel/sched/debug.c SEQ_printf(m, "numa_faults node=%d ", node); node 1265 kernel/sched/fair.c int node; node 1267 kernel/sched/fair.c for_each_online_node(node) { node 1268 kernel/sched/fair.c faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; node 1277 kernel/sched/fair.c int node; node 1279 kernel/sched/fair.c for_each_online_node(node) { node 1280 kernel/sched/fair.c faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; node 1303 kernel/sched/fair.c int node; node 1316 kernel/sched/fair.c for_each_online_node(node) { node 1318 kernel/sched/fair.c int dist = node_distance(nid, node); node 1324 kernel/sched/fair.c if (dist == sched_max_numa_distance || node == nid) node 1340 kernel/sched/fair.c faults = task_faults(p, node); node 1342 kernel/sched/fair.c faults = group_faults(p, node); node 2056 kernel/sched/fair.c int node, max_node = nid; node 2060 kernel/sched/fair.c for_each_online_node(node) { node 2061 kernel/sched/fair.c score = group_weight(p, node, dist); node 2064 kernel/sched/fair.c max_node = node; node 10524 kernel/sched/fair.c int node; node 10530 kernel/sched/fair.c for_each_online_node(node) { node 10532 kernel/sched/fair.c tsf = p->numa_faults[task_faults_idx(NUMA_MEM, node, 0)]; node 10533 kernel/sched/fair.c tpf = p->numa_faults[task_faults_idx(NUMA_MEM, node, 1)]; node 10536 kernel/sched/fair.c gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], node 10537 kernel/sched/fair.c gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; node 10539 kernel/sched/fair.c print_numa_stats(m, node, tsf, tpf, gsf, gpf); node 495 kernel/sched/psi.c list_for_each_entry(t, &group->triggers, node) node 513 kernel/sched/psi.c list_for_each_entry(t, &group->triggers, node) { node 1071 kernel/sched/psi.c list_add(&t->node, &group->triggers); node 1099 kernel/sched/psi.c if (!list_empty(&t->node)) { node 1103 kernel/sched/psi.c list_del(&t->node); node 1108 kernel/sched/psi.c list_for_each_entry(tmp, &group->triggers, node) node 1300 kernel/sched/sched.h extern void sched_setnuma(struct task_struct *p, int node); node 2194 kernel/sched/sched.h print_numa_stats(struct seq_file *m, int node, unsigned long tsf, node 1707 kernel/sched/topology.c int node = cpu_to_node(cpu); node 1712 kernel/sched/topology.c if (node_distance(j, node) <= sched_domains_numa_distance[i]) node 1169 kernel/signal.c hlist_for_each_entry(delayed, &t->signal->multiprocess, node) { node 1144 kernel/sysctl_binary.c unsigned long area, node; node 1162 kernel/sysctl_binary.c node = simple_strtoul(nodep, NULL, 10); node 1165 kernel/sysctl_binary.c if ((area > 63)||(node > 1023)) node 1168 kernel/sysctl_binary.c dnaddr = cpu_to_le16((area << 10) | node); node 172 kernel/time/alarmtimer.c timerqueue_del(&base->timerqueue, &alarm->node); node 174 kernel/time/alarmtimer.c timerqueue_add(&base->timerqueue, &alarm->node); node 192 kernel/time/alarmtimer.c timerqueue_del(&base->timerqueue, &alarm->node); node 223 kernel/time/alarmtimer.c hrtimer_set_expires(&alarm->timer, alarm->node.expires); node 237 kernel/time/alarmtimer.c return ktime_sub(alarm->node.expires, base->gettime()); node 338 kernel/time/alarmtimer.c timerqueue_init(&alarm->node); node 371 kernel/time/alarmtimer.c alarm->node.expires = start; node 373 kernel/time/alarmtimer.c hrtimer_start(&alarm->timer, alarm->node.expires, HRTIMER_MODE_ABS); node 400 kernel/time/alarmtimer.c hrtimer_set_expires(&alarm->timer, alarm->node.expires); node 455 kernel/time/alarmtimer.c delta = ktime_sub(now, alarm->node.expires); node 465 kernel/time/alarmtimer.c alarm->node.expires = ktime_add_ns(alarm->node.expires, node 468 kernel/time/alarmtimer.c if (alarm->node.expires > now) node 477 kernel/time/alarmtimer.c alarm->node.expires = ktime_add_safe(alarm->node.expires, interval); node 581 kernel/time/alarmtimer.c alarm_start(alarm, alarm->node.expires); node 605 kernel/time/alarmtimer.c return ktime_sub(alarm->node.expires, now); node 646 kernel/time/alarmtimer.c alarm->node.expires = expires; node 516 kernel/time/hrtimer.c timer = container_of(next, struct hrtimer, node); node 523 kernel/time/hrtimer.c timer = container_of(next, struct hrtimer, node); node 972 kernel/time/hrtimer.c return timerqueue_add(&base->active, &timer->node); node 997 kernel/time/hrtimer.c if (!timerqueue_del(&base->active, &timer->node)) node 1409 kernel/time/hrtimer.c timerqueue_init(&timer->node); node 1554 kernel/time/hrtimer.c struct timerqueue_node *node; node 1559 kernel/time/hrtimer.c while ((node = timerqueue_getnext(&base->active))) { node 1562 kernel/time/hrtimer.c timer = container_of(node, struct hrtimer, node); node 2016 kernel/time/hrtimer.c struct timerqueue_node *node; node 2018 kernel/time/hrtimer.c while ((node = timerqueue_getnext(&old_base->active))) { node 2019 kernel/time/hrtimer.c timer = container_of(node, struct hrtimer, node); node 127 kernel/time/posix-cpu-timers.c u64 delta, incr, expires = timer->it.cpu.node.expires; node 147 kernel/time/posix-cpu-timers.c timer->it.cpu.node.expires += incr; node 151 kernel/time/posix-cpu-timers.c return timer->it.cpu.node.expires; node 395 kernel/time/posix-cpu-timers.c timerqueue_init(&new_timer->it.cpu.node); node 427 kernel/time/posix-cpu-timers.c WARN_ON_ONCE(ctmr->head || timerqueue_node_queued(&ctmr->node)); node 445 kernel/time/posix-cpu-timers.c struct timerqueue_node *node; node 448 kernel/time/posix-cpu-timers.c while ((node = timerqueue_getnext(head))) { node 449 kernel/time/posix-cpu-timers.c timerqueue_del(head, node); node 450 kernel/time/posix-cpu-timers.c ctmr = container_of(node, struct cpu_timer, node); node 773 kernel/time/posix-cpu-timers.c ctmr = container_of(next, struct cpu_timer, node); node 100 kernel/time/timer_list.c timer = container_of(curr, struct hrtimer, node); node 390 kernel/trace/ftrace.c struct hlist_node node; node 706 kernel/trace/ftrace.c hlist_for_each_entry_rcu_notrace(rec, hhd, node) { node 720 kernel/trace/ftrace.c hlist_add_head_rcu(&rec->node, &stat->hash[key]); node 498 kernel/trace/ring_buffer.c struct hlist_node node; node 1421 kernel/trace/ring_buffer.c ret = cpuhp_state_add_instance(CPUHP_TRACE_RB_PREPARE, &buffer->node); node 1454 kernel/trace/ring_buffer.c cpuhp_state_remove_instance(CPUHP_TRACE_RB_PREPARE, &buffer->node); node 4869 kernel/trace/ring_buffer.c int trace_rb_cpu_prepare(unsigned int cpu, struct hlist_node *node) node 4876 kernel/trace/ring_buffer.c buffer = container_of(node, struct ring_buffer, node); node 1188 kernel/trace/trace_events.c struct list_head *node = v; node 1194 kernel/trace/trace_events.c node = common_head; node 1198 kernel/trace/trace_events.c node = head; node 1206 kernel/trace/trace_events.c node = node->prev; node 1207 kernel/trace/trace_events.c if (node == common_head) node 1209 kernel/trace/trace_events.c else if (node == head) node 1212 kernel/trace/trace_events.c return node; node 653 kernel/trace/trace_output.c hlist_for_each_entry(event, &event_hash[key], node) { node 772 kernel/trace/trace_output.c hlist_add_head(&event->node, &event_hash[key]); node 787 kernel/trace/trace_output.c hlist_del(&event->node); node 27 kernel/trace/trace_stat.c struct rb_node node; node 51 kernel/trace/trace_stat.c rbtree_postorder_for_each_entry_safe(snode, n, &session->stat_root, node) { node 95 kernel/trace/trace_stat.c this = container_of(*new, struct stat_node, node); node 105 kernel/trace/trace_stat.c rb_link_node(&data->node, parent, new); node 106 kernel/trace/trace_stat.c rb_insert_color(&data->node, root); node 176 kernel/trace/trace_stat.c struct rb_node *node; node 190 kernel/trace/trace_stat.c node = rb_first(&session->stat_root); node 191 kernel/trace/trace_stat.c for (i = 0; node && i < n; i++) node 192 kernel/trace/trace_stat.c node = rb_next(node); node 194 kernel/trace/trace_stat.c return node; node 200 kernel/trace/trace_stat.c struct rb_node *node = p; node 207 kernel/trace/trace_stat.c return rb_next(node); node 219 kernel/trace/trace_stat.c struct stat_node *l = container_of(v, struct stat_node, node); node 312 kernel/trace/trace_stat.c struct stat_session *session, *node; node 323 kernel/trace/trace_stat.c list_for_each_entry(node, &all_stat_sessions, session_list) { node 324 kernel/trace/trace_stat.c if (node->ts == trace) node 355 kernel/trace/trace_stat.c struct stat_session *node, *tmp; node 358 kernel/trace/trace_stat.c list_for_each_entry_safe(node, tmp, &all_stat_sessions, session_list) { node 359 kernel/trace/trace_stat.c if (node->ts == trace) { node 360 kernel/trace/trace_stat.c list_del(&node->session_list); node 361 kernel/trace/trace_stat.c destroy_session(node); node 118 kernel/ucount.c hlist_for_each_entry(ucounts, hashent, node) { node 148 kernel/ucount.c hlist_add_head(&new->node, hashent); node 167 kernel/ucount.c hlist_del_init(&ucounts->node); node 150 kernel/workqueue.c int node; /* I: the associated node ID */ node 411 kernel/workqueue.c list_for_each_entry((worker), &(pool)->workers, node) \ node 565 kernel/workqueue.c int node) node 575 kernel/workqueue.c if (unlikely(node == NUMA_NO_NODE)) node 578 kernel/workqueue.c return rcu_dereference_raw(wq->numa_pwq_tbl[node]); node 1539 kernel/workqueue.c static int workqueue_select_cpu_near(int node) node 1548 kernel/workqueue.c if (node < 0 || node >= MAX_NUMNODES || !node_online(node)) node 1553 kernel/workqueue.c if (node == cpu_to_node(cpu)) node 1557 kernel/workqueue.c cpu = cpumask_any_and(cpumask_of_node(node), cpu_online_mask); node 1583 kernel/workqueue.c bool queue_work_node(int node, struct workqueue_struct *wq, node 1603 kernel/workqueue.c int cpu = workqueue_select_cpu_near(node); node 1819 kernel/workqueue.c static struct worker *alloc_worker(int node) node 1823 kernel/workqueue.c worker = kzalloc_node(sizeof(*worker), GFP_KERNEL, node); node 1827 kernel/workqueue.c INIT_LIST_HEAD(&worker->node); node 1862 kernel/workqueue.c list_add_tail(&worker->node, &pool->workers); node 1883 kernel/workqueue.c list_del(&worker->node); node 1920 kernel/workqueue.c worker = alloc_worker(pool->node); node 1932 kernel/workqueue.c worker->task = kthread_create_on_node(worker_thread, worker, pool->node, node 3424 kernel/workqueue.c pool->node = NUMA_NO_NODE; node 3591 kernel/workqueue.c int node; node 3606 kernel/workqueue.c for_each_node(node) { node 3608 kernel/workqueue.c wq_numa_possible_cpumask[node])) { node 3609 kernel/workqueue.c target_node = node; node 3622 kernel/workqueue.c pool->node = target_node; node 3789 kernel/workqueue.c pwq = kmem_cache_alloc_node(pwq_cache, GFP_KERNEL, pool->node); node 3821 kernel/workqueue.c static bool wq_calc_node_cpumask(const struct workqueue_attrs *attrs, int node, node 3828 kernel/workqueue.c cpumask_and(cpumask, cpumask_of_node(node), attrs->cpumask); node 3836 kernel/workqueue.c cpumask_and(cpumask, attrs->cpumask, wq_numa_possible_cpumask[node]); node 3853 kernel/workqueue.c int node, node 3864 kernel/workqueue.c old_pwq = rcu_access_pointer(wq->numa_pwq_tbl[node]); node 3865 kernel/workqueue.c rcu_assign_pointer(wq->numa_pwq_tbl[node], pwq); node 3882 kernel/workqueue.c int node; node 3884 kernel/workqueue.c for_each_node(node) node 3885 kernel/workqueue.c put_pwq_unlocked(ctx->pwq_tbl[node]); node 3901 kernel/workqueue.c int node; node 3938 kernel/workqueue.c for_each_node(node) { node 3939 kernel/workqueue.c if (wq_calc_node_cpumask(new_attrs, node, -1, tmp_attrs->cpumask)) { node 3940 kernel/workqueue.c ctx->pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs); node 3941 kernel/workqueue.c if (!ctx->pwq_tbl[node]) node 3945 kernel/workqueue.c ctx->pwq_tbl[node] = ctx->dfl_pwq; node 3968 kernel/workqueue.c int node; node 3976 kernel/workqueue.c for_each_node(node) node 3977 kernel/workqueue.c ctx->pwq_tbl[node] = numa_pwq_tbl_install(ctx->wq, node, node 3978 kernel/workqueue.c ctx->pwq_tbl[node]); node 4085 kernel/workqueue.c int node = cpu_to_node(cpu); node 4106 kernel/workqueue.c pwq = unbound_pwq_by_node(wq, node); node 4114 kernel/workqueue.c if (wq_calc_node_cpumask(wq->dfl_pwq->pool->attrs, node, cpu_off, cpumask)) { node 4131 kernel/workqueue.c old_pwq = numa_pwq_tbl_install(wq, node, pwq); node 4139 kernel/workqueue.c old_pwq = numa_pwq_tbl_install(wq, node, wq->dfl_pwq); node 4335 kernel/workqueue.c int node; node 4404 kernel/workqueue.c for_each_node(node) { node 4405 kernel/workqueue.c pwq = rcu_access_pointer(wq->numa_pwq_tbl[node]); node 4406 kernel/workqueue.c RCU_INIT_POINTER(wq->numa_pwq_tbl[node], NULL); node 4635 kernel/workqueue.c if (pool->node != NUMA_NO_NODE) node 4636 kernel/workqueue.c pr_cont(" node=%d", pool->node); node 5393 kernel/workqueue.c int node, written = 0; node 5397 kernel/workqueue.c for_each_node(node) { node 5399 kernel/workqueue.c "%s%d:%d", delim, node, node 5400 kernel/workqueue.c unbound_pwq_by_node(wq, node)->pool->id); node 5842 kernel/workqueue.c int node, cpu; node 5863 kernel/workqueue.c for_each_node(node) node 5864 kernel/workqueue.c BUG_ON(!zalloc_cpumask_var_node(&tbl[node], GFP_KERNEL, node 5865 kernel/workqueue.c node_online(node) ? node : NUMA_NO_NODE)); node 5868 kernel/workqueue.c node = cpu_to_node(cpu); node 5869 kernel/workqueue.c if (WARN_ON(node == NUMA_NO_NODE)) { node 5874 kernel/workqueue.c cpumask_set_cpu(cpu, tbl[node]); node 5914 kernel/workqueue.c pool->node = cpu_to_node(cpu); node 5992 kernel/workqueue.c pool->node = cpu_to_node(cpu); node 41 kernel/workqueue_internal.h struct list_head node; /* A: anchored at pool->workers */ node 68 lib/842/842_compress.c struct hlist_node node; node 74 lib/842/842_compress.c struct hlist_node node; node 80 lib/842/842_compress.c struct hlist_node node; node 115 lib/842/842_compress.c for (_i = 0; _i < ARRAY_SIZE((p)->node##b); _i++) { \ node 116 lib/842/842_compress.c (p)->node##b[_i].index = _i; \ node 117 lib/842/842_compress.c (p)->node##b[_i].data = 0; \ node 118 lib/842/842_compress.c INIT_HLIST_NODE(&(p)->node##b[_i].node); \ node 125 lib/842/842_compress.c hash_for_each_possible(p->htable##b, _n, node, p->data##b[n]) { \ node 140 lib/842/842_compress.c struct sw842_hlist_node##b *_n = &(p)->node##b[(i)+(d)]; \ node 141 lib/842/842_compress.c hash_del(&_n->node); \ node 147 lib/842/842_compress.c hash_add((p)->htable##b, &_n->node, _n->data); \ node 26 lib/assoc_array.c const struct assoc_array_node *node; node 40 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 52 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ node 79 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 81 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ node 90 lib/assoc_array.c parent = READ_ONCE(node->back_pointer); /* Address dependency. */ node 91 lib/assoc_array.c slot = node->parent_slot; node 153 lib/assoc_array.c struct assoc_array_node *node; /* Node in which leaf might be found */ node 176 lib/assoc_array.c struct assoc_array_node *node; node 206 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 209 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ node 218 lib/assoc_array.c result->terminal_node.node = node; node 309 lib/assoc_array.c const struct assoc_array_node *node; node 318 lib/assoc_array.c node = result.terminal_node.node; node 324 lib/assoc_array.c ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ node 347 lib/assoc_array.c struct assoc_array_node *node; node 374 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 375 lib/assoc_array.c BUG_ON(node->back_pointer != parent); node 376 lib/assoc_array.c BUG_ON(slot != -1 && node->parent_slot != slot); node 380 lib/assoc_array.c pr_devel("Node %p [back=%p]\n", node, node->back_pointer); node 382 lib/assoc_array.c struct assoc_array_ptr *ptr = node->slots[slot]; node 397 lib/assoc_array.c parent = node->back_pointer; node 398 lib/assoc_array.c slot = node->parent_slot; node 400 lib/assoc_array.c kfree(node); node 423 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 480 lib/assoc_array.c struct assoc_array_node *node, *new_n0, *new_n1, *side; node 488 lib/assoc_array.c node = result->terminal_node.node; node 505 lib/assoc_array.c ptr = node->slots[i]; node 514 lib/assoc_array.c edit->leaf_p = &node->slots[i]; node 515 lib/assoc_array.c edit->dead_leaf = node->slots[i]; node 526 lib/assoc_array.c edit->leaf_p = &node->slots[free_slot]; node 527 lib/assoc_array.c edit->adjust_count_on = node; node 552 lib/assoc_array.c ptr = node->slots[i]; node 614 lib/assoc_array.c new_n0->back_pointer = node->back_pointer; node 615 lib/assoc_array.c new_n0->parent_slot = node->parent_slot; node 622 lib/assoc_array.c new_n0->nr_leaves_on_branch = node->nr_leaves_on_branch; node 648 lib/assoc_array.c if (assoc_array_ptr_is_meta(node->slots[i])) node 649 lib/assoc_array.c new_n0->slots[i] = node->slots[i]; node 659 lib/assoc_array.c if (assoc_array_ptr_is_meta(node->slots[i])) node 662 lib/assoc_array.c new_n1->slots[next_slot++] = node->slots[i]; node 668 lib/assoc_array.c new_n0->slots[free_slot] = node->slots[i]; node 690 lib/assoc_array.c ptr = node->slots[i]; node 702 lib/assoc_array.c ptr = node->back_pointer; node 706 lib/assoc_array.c edit->set[0].ptr = &assoc_array_ptr_to_node(ptr)->slots[node->parent_slot]; node 709 lib/assoc_array.c edit->excised_meta[0] = assoc_array_node_to_ptr(node); node 731 lib/assoc_array.c int x = ops->diff_objects(assoc_array_ptr_to_leaf(node->slots[i]), node 751 lib/assoc_array.c new_s0->back_pointer = node->back_pointer; node 752 lib/assoc_array.c new_s0->parent_slot = node->parent_slot; node 777 lib/assoc_array.c ptr = node->slots[i]; node 798 lib/assoc_array.c struct assoc_array_node *node, *new_n0, *side; node 827 lib/assoc_array.c node = assoc_array_ptr_to_node(shortcut->back_pointer); node 828 lib/assoc_array.c edit->set[0].ptr = &node->slots[shortcut->parent_slot]; node 1038 lib/assoc_array.c struct assoc_array_node *node; node 1056 lib/assoc_array.c collapse->node->slots[collapse->slot++] = assoc_array_leaf_to_ptr(leaf); node 1085 lib/assoc_array.c struct assoc_array_node *node, *new_n0; node 1106 lib/assoc_array.c node = result.terminal_node.node; node 1109 lib/assoc_array.c ptr = node->slots[slot]; node 1131 lib/assoc_array.c edit->dead_leaf = node->slots[slot]; node 1132 lib/assoc_array.c edit->set[0].ptr = &node->slots[slot]; node 1134 lib/assoc_array.c edit->adjust_count_on = node; node 1158 lib/assoc_array.c if (node->nr_leaves_on_branch <= ASSOC_ARRAY_FAN_OUT + 1) { node 1168 lib/assoc_array.c ptr = node->slots[i]; node 1176 lib/assoc_array.c node->nr_leaves_on_branch - 1, has_meta); node 1181 lib/assoc_array.c parent = node; node 1206 lib/assoc_array.c if (has_meta || parent != node) { node 1207 lib/assoc_array.c node = parent; node 1215 lib/assoc_array.c new_n0->back_pointer = node->back_pointer; node 1216 lib/assoc_array.c new_n0->parent_slot = node->parent_slot; node 1217 lib/assoc_array.c new_n0->nr_leaves_on_branch = node->nr_leaves_on_branch; node 1220 lib/assoc_array.c collapse.node = new_n0; node 1223 lib/assoc_array.c assoc_array_subtree_iterate(assoc_array_node_to_ptr(node), node 1224 lib/assoc_array.c node->back_pointer, node 1230 lib/assoc_array.c if (!node->back_pointer) { node 1232 lib/assoc_array.c } else if (assoc_array_ptr_is_leaf(node->back_pointer)) { node 1234 lib/assoc_array.c } else if (assoc_array_ptr_is_node(node->back_pointer)) { node 1236 lib/assoc_array.c assoc_array_ptr_to_node(node->back_pointer); node 1237 lib/assoc_array.c edit->set[1].ptr = &p->slots[node->parent_slot]; node 1238 lib/assoc_array.c } else if (assoc_array_ptr_is_shortcut(node->back_pointer)) { node 1240 lib/assoc_array.c assoc_array_ptr_to_shortcut(node->back_pointer); node 1244 lib/assoc_array.c edit->excised_subtree = assoc_array_node_to_ptr(node); node 1349 lib/assoc_array.c struct assoc_array_node *node; node 1377 lib/assoc_array.c node = edit->adjust_count_on; node 1379 lib/assoc_array.c node->nr_leaves_on_branch += edit->adjust_count_by; node 1381 lib/assoc_array.c ptr = node->back_pointer; node 1391 lib/assoc_array.c node = assoc_array_ptr_to_node(ptr); node 1460 lib/assoc_array.c struct assoc_array_node *node, *new_n; node 1508 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 1512 lib/assoc_array.c pr_devel("dup node %p -> %p\n", node, new_n); node 1514 lib/assoc_array.c new_n->parent_slot = node->parent_slot; node 1522 lib/assoc_array.c ptr = node->slots[slot]; node 1693 lib/assoc_array.c ptr = node->back_pointer; node 1701 lib/assoc_array.c slot = node->parent_slot; node 1705 lib/assoc_array.c node = assoc_array_ptr_to_node(cursor); node 95 lib/btree.c unsigned long *node; node 97 lib/btree.c node = mempool_alloc(head->mempool, gfp); node 98 lib/btree.c if (likely(node)) node 99 lib/btree.c memset(node, 0, NODESIZE); node 100 lib/btree.c return node; node 148 lib/btree.c static unsigned long *bkey(struct btree_geo *geo, unsigned long *node, int n) node 150 lib/btree.c return &node[n * geo->keylen]; node 153 lib/btree.c static void *bval(struct btree_geo *geo, unsigned long *node, int n) node 155 lib/btree.c return (void *)node[geo->no_longs + n]; node 158 lib/btree.c static void setkey(struct btree_geo *geo, unsigned long *node, int n, node 161 lib/btree.c longcpy(bkey(geo, node, n), key, geo->keylen); node 164 lib/btree.c static void setval(struct btree_geo *geo, unsigned long *node, int n, node 167 lib/btree.c node[geo->no_longs + n] = (unsigned long) val; node 170 lib/btree.c static void clearpair(struct btree_geo *geo, unsigned long *node, int n) node 172 lib/btree.c longset(bkey(geo, node, n), 0, geo->keylen); node 173 lib/btree.c node[geo->no_longs + n] = 0; node 178 lib/btree.c head->node = NULL; node 201 lib/btree.c mempool_free(head->node, head->mempool); node 211 lib/btree.c unsigned long *node = head->node; node 217 lib/btree.c node = bval(geo, node, 0); node 219 lib/btree.c longcpy(key, bkey(geo, node, 0), geo->keylen); node 220 lib/btree.c return bval(geo, node, 0); node 224 lib/btree.c static int keycmp(struct btree_geo *geo, unsigned long *node, int pos, node 227 lib/btree.c return longcmp(bkey(geo, node, pos), key, geo->keylen); node 245 lib/btree.c unsigned long *node = head->node; node 252 lib/btree.c if (keycmp(geo, node, i, key) <= 0) node 256 lib/btree.c node = bval(geo, node, i); node 257 lib/btree.c if (!node) node 261 lib/btree.c if (!node) node 265 lib/btree.c if (keycmp(geo, node, i, key) == 0) node 266 lib/btree.c return bval(geo, node, i); node 275 lib/btree.c unsigned long *node = head->node; node 282 lib/btree.c if (keycmp(geo, node, i, key) <= 0) node 286 lib/btree.c node = bval(geo, node, i); node 287 lib/btree.c if (!node) node 291 lib/btree.c if (!node) node 295 lib/btree.c if (keycmp(geo, node, i, key) == 0) { node 296 lib/btree.c setval(geo, node, i, val); node 315 lib/btree.c unsigned long *node, *oldnode; node 327 lib/btree.c node = head->node; node 330 lib/btree.c if (keycmp(geo, node, i, key) <= 0) node 334 lib/btree.c oldnode = node; node 335 lib/btree.c node = bval(geo, node, i); node 336 lib/btree.c if (!node) node 341 lib/btree.c if (!node) node 345 lib/btree.c if (keycmp(geo, node, i, key) <= 0) { node 346 lib/btree.c if (bval(geo, node, i)) { node 347 lib/btree.c longcpy(__key, bkey(geo, node, i), geo->keylen); node 348 lib/btree.c return bval(geo, node, i); node 363 lib/btree.c static int getpos(struct btree_geo *geo, unsigned long *node, node 369 lib/btree.c if (keycmp(geo, node, i, key) <= 0) node 375 lib/btree.c static int getfill(struct btree_geo *geo, unsigned long *node, int start) node 380 lib/btree.c if (!bval(geo, node, i)) node 391 lib/btree.c unsigned long *node = head->node; node 396 lib/btree.c if (keycmp(geo, node, i, key) <= 0) node 399 lib/btree.c if ((i == geo->no_pairs) || !bval(geo, node, i)) { node 404 lib/btree.c setkey(geo, node, i, key); node 407 lib/btree.c node = bval(geo, node, i); node 409 lib/btree.c BUG_ON(!node); node 410 lib/btree.c return node; node 416 lib/btree.c unsigned long *node; node 419 lib/btree.c node = btree_node_alloc(head, gfp); node 420 lib/btree.c if (!node) node 422 lib/btree.c if (head->node) { node 423 lib/btree.c fill = getfill(geo, head->node, 0); node 424 lib/btree.c setkey(geo, node, 0, bkey(geo, head->node, fill - 1)); node 425 lib/btree.c setval(geo, node, 0, head->node); node 427 lib/btree.c head->node = node; node 434 lib/btree.c unsigned long *node; node 440 lib/btree.c node = head->node; node 441 lib/btree.c fill = getfill(geo, node, 0); node 443 lib/btree.c head->node = bval(geo, node, 0); node 445 lib/btree.c mempool_free(node, head->mempool); node 452 lib/btree.c unsigned long *node; node 463 lib/btree.c node = find_level(head, geo, key, level); node 464 lib/btree.c pos = getpos(geo, node, key); node 465 lib/btree.c fill = getfill(geo, node, pos); node 467 lib/btree.c BUG_ON(pos < fill && keycmp(geo, node, pos, key) == 0); node 477 lib/btree.c bkey(geo, node, fill / 2 - 1), node 484 lib/btree.c setkey(geo, new, i, bkey(geo, node, i)); node 485 lib/btree.c setval(geo, new, i, bval(geo, node, i)); node 486 lib/btree.c setkey(geo, node, i, bkey(geo, node, i + fill / 2)); node 487 lib/btree.c setval(geo, node, i, bval(geo, node, i + fill / 2)); node 488 lib/btree.c clearpair(geo, node, i + fill / 2); node 491 lib/btree.c setkey(geo, node, i, bkey(geo, node, fill - 1)); node 492 lib/btree.c setval(geo, node, i, bval(geo, node, fill - 1)); node 493 lib/btree.c clearpair(geo, node, fill - 1); node 501 lib/btree.c setkey(geo, node, i, bkey(geo, node, i - 1)); node 502 lib/btree.c setval(geo, node, i, bval(geo, node, i - 1)); node 504 lib/btree.c setkey(geo, node, pos, key); node 505 lib/btree.c setval(geo, node, pos, val); node 594 lib/btree.c unsigned long *node; node 601 lib/btree.c head->node = NULL; node 605 lib/btree.c node = find_level(head, geo, key, level); node 606 lib/btree.c pos = getpos(geo, node, key); node 607 lib/btree.c fill = getfill(geo, node, pos); node 608 lib/btree.c if ((level == 1) && (keycmp(geo, node, pos, key) != 0)) node 610 lib/btree.c ret = bval(geo, node, pos); node 614 lib/btree.c setkey(geo, node, i, bkey(geo, node, i + 1)); node 615 lib/btree.c setval(geo, node, i, bval(geo, node, i + 1)); node 617 lib/btree.c clearpair(geo, node, fill - 1); node 621 lib/btree.c rebalance(head, geo, key, level, node, fill - 1); node 649 lib/btree.c if (!(target->node)) { node 651 lib/btree.c target->node = victim->node; node 677 lib/btree.c unsigned long *node, unsigned long opaque, node 687 lib/btree.c child = bval(geo, node, i); node 694 lib/btree.c func(child, opaque, bkey(geo, node, i), count++, node 698 lib/btree.c mempool_free(node, head->mempool); node 757 lib/btree.c if (head->node) node 758 lib/btree.c count = __btree_for_each(head, geo, head->node, opaque, func, node 775 lib/btree.c if (head->node) node 776 lib/btree.c count = __btree_for_each(head, geo, head->node, opaque, func, node 113 lib/cpumask.c bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node) node 115 lib/cpumask.c *mask = kmalloc_node(cpumask_size(), flags, node); node 128 lib/cpumask.c bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node) node 130 lib/cpumask.c return alloc_cpumask_var_node(mask, flags | __GFP_ZERO, node); node 206 lib/cpumask.c unsigned int cpumask_local_spread(unsigned int i, int node) node 213 lib/cpumask.c if (node == NUMA_NO_NODE) { node 219 lib/cpumask.c for_each_cpu_and(cpu, cpumask_of_node(node), cpu_online_mask) node 225 lib/cpumask.c if (cpumask_test_cpu(cpu, cpumask_of_node(node))) node 153 lib/debugobjects.c obj = hlist_entry(obj_to_free.first, typeof(*obj), node); node 154 lib/debugobjects.c hlist_del(&obj->node); node 156 lib/debugobjects.c hlist_add_head(&obj->node, &obj_pool); node 179 lib/debugobjects.c hlist_add_head(&new[--cnt]->node, &obj_pool); node 195 lib/debugobjects.c hlist_for_each_entry(obj, &b->list, node) { node 214 lib/debugobjects.c obj = hlist_entry(list->first, typeof(*obj), node); node 215 lib/debugobjects.c hlist_del(&obj->node); node 258 lib/debugobjects.c hlist_add_head(&obj2->node, node 280 lib/debugobjects.c hlist_add_head(&obj->node, &b->list); node 313 lib/debugobjects.c obj = hlist_entry(obj_to_free.first, typeof(*obj), node); node 314 lib/debugobjects.c hlist_del(&obj->node); node 315 lib/debugobjects.c hlist_add_head(&obj->node, &obj_pool); node 335 lib/debugobjects.c hlist_for_each_entry_safe(obj, tmp, &tofree, node) { node 336 lib/debugobjects.c hlist_del(&obj->node); node 358 lib/debugobjects.c hlist_add_head(&obj->node, &percpu_pool->free_objs); node 383 lib/debugobjects.c hlist_add_head(&obj->node, &obj_to_free); node 388 lib/debugobjects.c hlist_add_head(&objs[--lookahead_count]->node, node 402 lib/debugobjects.c hlist_add_head(&obj->node, &obj_to_free); node 409 lib/debugobjects.c hlist_add_head(&obj->node, &obj_pool); node 414 lib/debugobjects.c hlist_add_head(&objs[--lookahead_count]->node, node 457 lib/debugobjects.c hlist_for_each_entry_safe(obj, tmp, &freelist, node) { node 458 lib/debugobjects.c hlist_del(&obj->node); node 826 lib/debugobjects.c hlist_del(&obj->node); node 956 lib/debugobjects.c hlist_for_each_entry_safe(obj, tmp, &db->list, node) { node 972 lib/debugobjects.c hlist_del(&obj->node); node 1289 lib/debugobjects.c hlist_add_head(&obj_static_pool[i].node, &obj_pool); node 1307 lib/debugobjects.c hlist_add_head(&obj->node, &objects); node 1317 lib/debugobjects.c hlist_for_each_entry_safe(obj, tmp, &obj_pool, node) node 1318 lib/debugobjects.c hlist_del(&obj->node); node 1326 lib/debugobjects.c hlist_for_each_entry(obj, &objects, node) { node 1327 lib/debugobjects.c new = hlist_entry(obj_pool.first, typeof(*obj), node); node 1328 lib/debugobjects.c hlist_del(&new->node); node 1331 lib/debugobjects.c hlist_add_head(&new->node, &db->list); node 1340 lib/debugobjects.c hlist_for_each_entry_safe(obj, tmp, &objects, node) { node 1341 lib/debugobjects.c hlist_del(&obj->node); node 206 lib/devres.c void __iomem *devm_of_iomap(struct device *dev, struct device_node *node, int index, node 211 lib/devres.c if (of_address_to_resource(node, index, &res)) node 81 lib/generic-radix-tree.c struct genradix_node *node; node 83 lib/generic-radix-tree.c node = (struct genradix_node *)__get_free_page(gfp_mask|__GFP_ZERO); node 90 lib/generic-radix-tree.c kmemleak_alloc(node, PAGE_SIZE, 1, gfp_mask); node 91 lib/generic-radix-tree.c return node; node 94 lib/generic-radix-tree.c static inline void genradix_free_node(struct genradix_node *node) node 96 lib/generic-radix-tree.c kmemleak_free(node); node 97 lib/generic-radix-tree.c free_page((unsigned long)node); node 292 lib/idr.c struct radix_tree_node *node; node 298 lib/idr.c entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot); node 302 lib/idr.c __radix_tree_replace(&idr->idr_rt, node, slot, ptr); node 564 lib/idr.c struct xa_node *node = xa_to_node(entry); node 565 lib/idr.c unsigned int shift = node->shift + IDA_CHUNK_SHIFT + node 569 lib/idr.c xa_dump_node(node); node 571 lib/idr.c ida_dump_entry(node->slots[i], node 572 lib/idr.c index | (i << node->shift)); node 7 lib/interval_tree.c #define START(node) ((node)->start) node 8 lib/interval_tree.c #define LAST(node) ((node)->last) node 32 lib/interval_tree_test.c struct interval_tree_node *node; node 35 lib/interval_tree_test.c for (node = interval_tree_iter_first(root, start, last); node; node 36 lib/interval_tree_test.c node = interval_tree_iter_next(node, start, last)) node 175 lib/klist.c struct klist_node *node; node 192 lib/klist.c if (waiter->node != n) node 242 lib/klist.c waiter.node = n; node 6 lib/nodemask.c int __next_node_in(int node, const nodemask_t *srcp) node 8 lib/nodemask.c int ret = __next_node(node, srcp); node 737 lib/objagg.c struct objagg_tmp_node *node = &graph->nodes[index]; node 738 lib/objagg.c unsigned int weight = node->objagg_obj->stats.user_count; node 748 lib/objagg.c node = &graph->nodes[j]; node 749 lib/objagg.c if (node->crossed_out) node 751 lib/objagg.c weight += node->objagg_obj->stats.user_count; node 758 lib/objagg.c struct objagg_tmp_node *node; node 765 lib/objagg.c node = &graph->nodes[i]; node 766 lib/objagg.c if (node->crossed_out) node 781 lib/objagg.c struct objagg_tmp_node *node; node 804 lib/objagg.c node = &graph->nodes[i++]; node 805 lib/objagg.c node->objagg_obj = objagg_obj; node 816 lib/objagg.c node = &graph->nodes[j]; node 819 lib/objagg.c node->objagg_obj->obj)) { node 847 lib/objagg.c struct objagg_tmp_node *node; node 860 lib/objagg.c node = &graph->nodes[index]; node 861 lib/objagg.c node->crossed_out = true; node 863 lib/objagg.c node->objagg_obj, node 874 lib/objagg.c node = &graph->nodes[j]; node 875 lib/objagg.c if (node->crossed_out) node 877 lib/objagg.c node->crossed_out = true; node 879 lib/objagg.c node->objagg_obj, node 73 lib/plist.c void plist_add(struct plist_node *node, struct plist_head *head) node 79 lib/plist.c WARN_ON(!plist_node_empty(node)); node 80 lib/plist.c WARN_ON(!list_empty(&node->prio_list)); node 88 lib/plist.c if (node->prio < iter->prio) { node 98 lib/plist.c if (!prev || prev->prio != node->prio) node 99 lib/plist.c list_add_tail(&node->prio_list, &iter->prio_list); node 101 lib/plist.c list_add_tail(&node->node_list, node_next); node 112 lib/plist.c void plist_del(struct plist_node *node, struct plist_head *head) node 116 lib/plist.c if (!list_empty(&node->prio_list)) { node 117 lib/plist.c if (node->node_list.next != &head->node_list) { node 120 lib/plist.c next = list_entry(node->node_list.next, node 125 lib/plist.c list_add(&next->prio_list, &node->prio_list); node 127 lib/plist.c list_del_init(&node->prio_list); node 130 lib/plist.c list_del_init(&node->node_list); node 145 lib/plist.c void plist_requeue(struct plist_node *node, struct plist_head *head) node 152 lib/plist.c BUG_ON(plist_node_empty(node)); node 154 lib/plist.c if (node == plist_last(head)) node 157 lib/plist.c iter = plist_next(node); node 159 lib/plist.c if (node->prio != iter->prio) node 162 lib/plist.c plist_del(node, head); node 165 lib/plist.c if (node->prio != iter->prio) { node 170 lib/plist.c list_add_tail(&node->node_list, node_next); node 212 lib/plist.c static void __init plist_test_requeue(struct plist_node *node) node 214 lib/plist.c plist_requeue(node, &test_head); node 216 lib/plist.c if (node != plist_last(&test_head)) node 217 lib/plist.c BUG_ON(node->prio == plist_next(node)->prio); node 109 lib/radix-tree.c static inline void tag_set(struct radix_tree_node *node, unsigned int tag, node 112 lib/radix-tree.c __set_bit(offset, node->tags[tag]); node 115 lib/radix-tree.c static inline void tag_clear(struct radix_tree_node *node, unsigned int tag, node 118 lib/radix-tree.c __clear_bit(offset, node->tags[tag]); node 121 lib/radix-tree.c static inline int tag_get(const struct radix_tree_node *node, unsigned int tag, node 124 lib/radix-tree.c return test_bit(offset, node->tags[tag]); node 161 lib/radix-tree.c static inline int any_tag_set(const struct radix_tree_node *node, node 166 lib/radix-tree.c if (node->tags[tag][idx]) node 172 lib/radix-tree.c static inline void all_tag_set(struct radix_tree_node *node, unsigned int tag) node 174 lib/radix-tree.c bitmap_fill(node->tags[tag], RADIX_TREE_MAP_SIZE); node 189 lib/radix-tree.c radix_tree_find_next_bit(struct radix_tree_node *node, unsigned int tag, node 192 lib/radix-tree.c const unsigned long *addr = node->tags[tag]; node 225 lib/radix-tree.c static inline unsigned long node_maxindex(const struct radix_tree_node *node) node 227 lib/radix-tree.c return shift_maxindex(node->shift); node 231 lib/radix-tree.c const struct radix_tree_node *node, node 234 lib/radix-tree.c return (index & ~node_maxindex(node)) + (offset << node->shift); node 301 lib/radix-tree.c struct radix_tree_node *node = node 309 lib/radix-tree.c memset(node->slots, 0, sizeof(node->slots)); node 310 lib/radix-tree.c memset(node->tags, 0, sizeof(node->tags)); node 311 lib/radix-tree.c INIT_LIST_HEAD(&node->private_list); node 313 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node); node 317 lib/radix-tree.c radix_tree_node_free(struct radix_tree_node *node) node 319 lib/radix-tree.c call_rcu(&node->rcu_head, radix_tree_node_rcu_free); node 334 lib/radix-tree.c struct radix_tree_node *node; node 347 lib/radix-tree.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); node 348 lib/radix-tree.c if (node == NULL) node 353 lib/radix-tree.c node->parent = rtp->nodes; node 354 lib/radix-tree.c rtp->nodes = node; node 357 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node); node 400 lib/radix-tree.c struct radix_tree_node *node = rcu_dereference_raw(root->xa_head); node 402 lib/radix-tree.c *nodep = node; node 404 lib/radix-tree.c if (likely(radix_tree_is_internal_node(node))) { node 405 lib/radix-tree.c node = entry_to_node(node); node 406 lib/radix-tree.c *maxindex = node_maxindex(node); node 407 lib/radix-tree.c return node->shift + RADIX_TREE_MAP_SHIFT; node 434 lib/radix-tree.c struct radix_tree_node *node = radix_tree_node_alloc(gfp, NULL, node 436 lib/radix-tree.c if (!node) node 440 lib/radix-tree.c all_tag_set(node, IDR_FREE); node 442 lib/radix-tree.c tag_clear(node, IDR_FREE, 0); node 449 lib/radix-tree.c tag_set(node, tag, 0); node 455 lib/radix-tree.c entry_to_node(entry)->parent = node; node 458 lib/radix-tree.c node->nr_values = 1; node 464 lib/radix-tree.c node->slots[0] = (void __rcu *)entry; node 465 lib/radix-tree.c entry = node_to_entry(node); node 482 lib/radix-tree.c struct radix_tree_node *node = rcu_dereference_raw(root->xa_head); node 485 lib/radix-tree.c if (!radix_tree_is_internal_node(node)) node 487 lib/radix-tree.c node = entry_to_node(node); node 493 lib/radix-tree.c if (node->count != 1) node 495 lib/radix-tree.c child = rcu_dereference_raw(node->slots[0]); node 504 lib/radix-tree.c if (!node->shift && is_idr(root)) node 518 lib/radix-tree.c if (is_idr(root) && !tag_get(node, IDR_FREE, 0)) node 539 lib/radix-tree.c node->count = 0; node 541 lib/radix-tree.c node->slots[0] = (void __rcu *)RADIX_TREE_RETRY; node 544 lib/radix-tree.c WARN_ON_ONCE(!list_empty(&node->private_list)); node 545 lib/radix-tree.c radix_tree_node_free(node); node 553 lib/radix-tree.c struct radix_tree_node *node) node 560 lib/radix-tree.c if (node->count) { node 561 lib/radix-tree.c if (node_to_entry(node) == node 567 lib/radix-tree.c parent = node->parent; node 569 lib/radix-tree.c parent->slots[node->offset] = NULL; node 581 lib/radix-tree.c WARN_ON_ONCE(!list_empty(&node->private_list)); node 582 lib/radix-tree.c radix_tree_node_free(node); node 585 lib/radix-tree.c node = parent; node 586 lib/radix-tree.c } while (node); node 611 lib/radix-tree.c struct radix_tree_node *node = NULL, *child; node 633 lib/radix-tree.c child = radix_tree_node_alloc(gfp, node, root, shift, node 638 lib/radix-tree.c if (node) node 639 lib/radix-tree.c node->count++; node 644 lib/radix-tree.c node = entry_to_node(child); node 645 lib/radix-tree.c offset = radix_tree_descend(node, &child, index); node 646 lib/radix-tree.c slot = &node->slots[offset]; node 650 lib/radix-tree.c *nodep = node; node 665 lib/radix-tree.c static void radix_tree_free_nodes(struct radix_tree_node *node) node 668 lib/radix-tree.c struct radix_tree_node *child = entry_to_node(node); node 684 lib/radix-tree.c if (old == entry_to_node(node)) node 690 lib/radix-tree.c static inline int insert_entries(struct radix_tree_node *node, node 696 lib/radix-tree.c if (node) { node 697 lib/radix-tree.c node->count++; node 699 lib/radix-tree.c node->nr_values++; node 715 lib/radix-tree.c struct radix_tree_node *node; node 721 lib/radix-tree.c error = __radix_tree_create(root, index, &node, &slot); node 725 lib/radix-tree.c error = insert_entries(node, slot, item, false); node 729 lib/radix-tree.c if (node) { node 730 lib/radix-tree.c unsigned offset = get_slot_offset(node, slot); node 731 lib/radix-tree.c BUG_ON(tag_get(node, 0, offset)); node 732 lib/radix-tree.c BUG_ON(tag_get(node, 1, offset)); node 733 lib/radix-tree.c BUG_ON(tag_get(node, 2, offset)); node 760 lib/radix-tree.c struct radix_tree_node *node, *parent; node 767 lib/radix-tree.c radix_tree_load_root(root, &node, &maxindex); node 771 lib/radix-tree.c while (radix_tree_is_internal_node(node)) { node 774 lib/radix-tree.c parent = entry_to_node(node); node 775 lib/radix-tree.c offset = radix_tree_descend(parent, &node, index); node 777 lib/radix-tree.c if (node == RADIX_TREE_RETRY) node 787 lib/radix-tree.c return node; node 833 lib/radix-tree.c struct radix_tree_node *node, int count, int values) node 835 lib/radix-tree.c if (node && (count || values)) { node 836 lib/radix-tree.c node->count += count; node 837 lib/radix-tree.c node->nr_values += values; node 844 lib/radix-tree.c const struct radix_tree_node *node, node 847 lib/radix-tree.c if (node) node 848 lib/radix-tree.c return tag_get(node, tag, offset); node 860 lib/radix-tree.c struct radix_tree_node *node, void __rcu **slot, node 864 lib/radix-tree.c unsigned offset = get_slot_offset(node, slot); node 865 lib/radix-tree.c bool free = node_tag_get(root, node, IDR_FREE, offset); node 885 lib/radix-tree.c struct radix_tree_node *node, node 890 lib/radix-tree.c int count = calculate_count(root, node, slot, item, old); node 897 lib/radix-tree.c WARN_ON_ONCE(!node && (slot != (void __rcu **)&root->xa_head) && node 899 lib/radix-tree.c replace_slot(slot, item, node, count, values); node 901 lib/radix-tree.c if (!node) node 904 lib/radix-tree.c delete_node(root, node); node 943 lib/radix-tree.c __radix_tree_replace(root, iter->node, slot, item); node 947 lib/radix-tree.c struct radix_tree_node *node, node 950 lib/radix-tree.c while (node) { node 951 lib/radix-tree.c if (tag_get(node, tag, offset)) node 953 lib/radix-tree.c tag_set(node, tag, offset); node 954 lib/radix-tree.c offset = node->offset; node 955 lib/radix-tree.c node = node->parent; node 978 lib/radix-tree.c struct radix_tree_node *node, *parent; node 981 lib/radix-tree.c radix_tree_load_root(root, &node, &maxindex); node 984 lib/radix-tree.c while (radix_tree_is_internal_node(node)) { node 987 lib/radix-tree.c parent = entry_to_node(node); node 988 lib/radix-tree.c offset = radix_tree_descend(parent, &node, index); node 989 lib/radix-tree.c BUG_ON(!node); node 999 lib/radix-tree.c return node; node 1004 lib/radix-tree.c struct radix_tree_node *node, node 1007 lib/radix-tree.c while (node) { node 1008 lib/radix-tree.c if (!tag_get(node, tag, offset)) node 1010 lib/radix-tree.c tag_clear(node, tag, offset); node 1011 lib/radix-tree.c if (any_tag_set(node, tag)) node 1014 lib/radix-tree.c offset = node->offset; node 1015 lib/radix-tree.c node = node->parent; node 1040 lib/radix-tree.c struct radix_tree_node *node, *parent; node 1044 lib/radix-tree.c radix_tree_load_root(root, &node, &maxindex); node 1050 lib/radix-tree.c while (radix_tree_is_internal_node(node)) { node 1051 lib/radix-tree.c parent = entry_to_node(node); node 1052 lib/radix-tree.c offset = radix_tree_descend(parent, &node, index); node 1055 lib/radix-tree.c if (node) node 1058 lib/radix-tree.c return node; node 1071 lib/radix-tree.c node_tag_clear(root, iter->node, tag, iter_offset(iter)); node 1092 lib/radix-tree.c struct radix_tree_node *node, *parent; node 1098 lib/radix-tree.c radix_tree_load_root(root, &node, &maxindex); node 1102 lib/radix-tree.c while (radix_tree_is_internal_node(node)) { node 1105 lib/radix-tree.c parent = entry_to_node(node); node 1106 lib/radix-tree.c offset = radix_tree_descend(parent, &node, index); node 1110 lib/radix-tree.c if (node == RADIX_TREE_RETRY) node 1120 lib/radix-tree.c struct radix_tree_node *node, unsigned offset, node 1126 lib/radix-tree.c if (!node) { node 1131 lib/radix-tree.c iter->tags = node->tags[tag][tag_long] >> tag_bit; node 1137 lib/radix-tree.c iter->tags |= node->tags[tag][tag_long + 1] << node 1167 lib/radix-tree.c struct radix_tree_node *node, *child; node 1198 lib/radix-tree.c iter->node = NULL; node 1203 lib/radix-tree.c node = entry_to_node(child); node 1204 lib/radix-tree.c offset = radix_tree_descend(node, &child, index); node 1207 lib/radix-tree.c !tag_get(node, tag, offset) : !child) { node 1213 lib/radix-tree.c offset = radix_tree_find_next_bit(node, tag, node 1218 lib/radix-tree.c node->slots[offset]); node 1222 lib/radix-tree.c index &= ~node_maxindex(node); node 1223 lib/radix-tree.c index += offset << node->shift; node 1229 lib/radix-tree.c child = rcu_dereference_raw(node->slots[offset]); node 1236 lib/radix-tree.c } while (node->shift && radix_tree_is_internal_node(child)); node 1239 lib/radix-tree.c iter->index = (index &~ node_maxindex(node)) | offset; node 1240 lib/radix-tree.c iter->next_index = (index | node_maxindex(node)) + 1; node 1241 lib/radix-tree.c iter->node = node; node 1244 lib/radix-tree.c set_iter_tags(iter, node, offset, tag); node 1246 lib/radix-tree.c return node->slots + offset; node 1374 lib/radix-tree.c struct radix_tree_node *node, void __rcu **slot) node 1378 lib/radix-tree.c unsigned offset = get_slot_offset(node, slot); node 1382 lib/radix-tree.c node_tag_set(root, node, IDR_FREE, offset); node 1385 lib/radix-tree.c node_tag_clear(root, node, tag, offset); node 1387 lib/radix-tree.c replace_slot(slot, NULL, node, -1, values); node 1388 lib/radix-tree.c return node && delete_node(root, node); node 1406 lib/radix-tree.c if (__radix_tree_delete(root, iter->node, slot)) node 1425 lib/radix-tree.c struct radix_tree_node *node = NULL; node 1429 lib/radix-tree.c entry = __radix_tree_lookup(root, index, &node, &slot); node 1432 lib/radix-tree.c if (!entry && (!is_idr(root) || node_tag_get(root, node, IDR_FREE, node 1433 lib/radix-tree.c get_slot_offset(node, slot)))) node 1439 lib/radix-tree.c __radix_tree_delete(root, node, slot); node 1489 lib/radix-tree.c struct radix_tree_node *node = NULL, *child; node 1515 lib/radix-tree.c child = radix_tree_node_alloc(gfp, node, root, shift, node 1521 lib/radix-tree.c if (node) node 1522 lib/radix-tree.c node->count++; node 1526 lib/radix-tree.c node = entry_to_node(child); node 1527 lib/radix-tree.c offset = radix_tree_descend(node, &child, start); node 1528 lib/radix-tree.c if (!tag_get(node, IDR_FREE, offset)) { node 1529 lib/radix-tree.c offset = radix_tree_find_next_bit(node, IDR_FREE, node 1531 lib/radix-tree.c start = next_index(start, node, offset); node 1535 lib/radix-tree.c offset = node->offset + 1; node 1536 lib/radix-tree.c node = node->parent; node 1537 lib/radix-tree.c if (!node) node 1539 lib/radix-tree.c shift = node->shift; node 1541 lib/radix-tree.c child = rcu_dereference_raw(node->slots[offset]); node 1543 lib/radix-tree.c slot = &node->slots[offset]; node 1547 lib/radix-tree.c if (node) node 1548 lib/radix-tree.c iter->next_index = 1 + min(max, (start | node_maxindex(node))); node 1551 lib/radix-tree.c iter->node = node; node 1552 lib/radix-tree.c set_iter_tags(iter, node, offset, IDR_FREE); node 1570 lib/radix-tree.c struct radix_tree_node *node = rcu_dereference_raw(idr->idr_rt.xa_head); node 1571 lib/radix-tree.c if (radix_tree_is_internal_node(node)) node 1572 lib/radix-tree.c radix_tree_free_nodes(node); node 1581 lib/radix-tree.c struct radix_tree_node *node = arg; node 1583 lib/radix-tree.c memset(node, 0, sizeof(*node)); node 1584 lib/radix-tree.c INIT_LIST_HEAD(&node->private_list); node 1590 lib/radix-tree.c struct radix_tree_node *node; node 1595 lib/radix-tree.c node = rtp->nodes; node 1596 lib/radix-tree.c rtp->nodes = node->parent; node 1597 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node); node 85 lib/rbtree.c __rb_insert(struct rb_node *node, struct rb_root *root, node 88 lib/rbtree.c struct rb_node *parent = rb_red_parent(node), *gparent, *tmp; node 100 lib/rbtree.c rb_set_parent_color(node, NULL, RB_BLACK); node 133 lib/rbtree.c node = gparent; node 134 lib/rbtree.c parent = rb_parent(node); node 135 lib/rbtree.c rb_set_parent_color(node, parent, RB_RED); node 140 lib/rbtree.c if (node == tmp) { node 154 lib/rbtree.c tmp = node->rb_left; node 156 lib/rbtree.c WRITE_ONCE(node->rb_left, parent); node 160 lib/rbtree.c rb_set_parent_color(parent, node, RB_RED); node 161 lib/rbtree.c augment_rotate(parent, node); node 162 lib/rbtree.c parent = node; node 163 lib/rbtree.c tmp = node->rb_right; node 189 lib/rbtree.c node = gparent; node 190 lib/rbtree.c parent = rb_parent(node); node 191 lib/rbtree.c rb_set_parent_color(node, parent, RB_RED); node 196 lib/rbtree.c if (node == tmp) { node 198 lib/rbtree.c tmp = node->rb_right; node 200 lib/rbtree.c WRITE_ONCE(node->rb_right, parent); node 204 lib/rbtree.c rb_set_parent_color(parent, node, RB_RED); node 205 lib/rbtree.c augment_rotate(parent, node); node 206 lib/rbtree.c parent = node; node 207 lib/rbtree.c tmp = node->rb_left; node 230 lib/rbtree.c struct rb_node *node = NULL, *sibling, *tmp1, *tmp2; node 241 lib/rbtree.c if (node != sibling) { /* node == parent->rb_left */ node 285 lib/rbtree.c node = parent; node 286 lib/rbtree.c parent = rb_parent(node); node 375 lib/rbtree.c node = parent; node 376 lib/rbtree.c parent = rb_parent(node); node 424 lib/rbtree.c static inline void dummy_propagate(struct rb_node *node, struct rb_node *stop) {} node 434 lib/rbtree.c void rb_insert_color(struct rb_node *node, struct rb_root *root) node 436 lib/rbtree.c __rb_insert(node, root, dummy_rotate); node 440 lib/rbtree.c void rb_erase(struct rb_node *node, struct rb_root *root) node 443 lib/rbtree.c rebalance = __rb_erase_augmented(node, root, &dummy_callbacks); node 456 lib/rbtree.c void __rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 459 lib/rbtree.c __rb_insert(node, root, augment_rotate); node 492 lib/rbtree.c struct rb_node *rb_next(const struct rb_node *node) node 496 lib/rbtree.c if (RB_EMPTY_NODE(node)) node 503 lib/rbtree.c if (node->rb_right) { node 504 lib/rbtree.c node = node->rb_right; node 505 lib/rbtree.c while (node->rb_left) node 506 lib/rbtree.c node=node->rb_left; node 507 lib/rbtree.c return (struct rb_node *)node; node 517 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_right) node 518 lib/rbtree.c node = parent; node 524 lib/rbtree.c struct rb_node *rb_prev(const struct rb_node *node) node 528 lib/rbtree.c if (RB_EMPTY_NODE(node)) node 535 lib/rbtree.c if (node->rb_left) { node 536 lib/rbtree.c node = node->rb_left; node 537 lib/rbtree.c while (node->rb_right) node 538 lib/rbtree.c node=node->rb_right; node 539 lib/rbtree.c return (struct rb_node *)node; node 546 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_left) node 547 lib/rbtree.c node = parent; node 592 lib/rbtree.c static struct rb_node *rb_left_deepest_node(const struct rb_node *node) node 595 lib/rbtree.c if (node->rb_left) node 596 lib/rbtree.c node = node->rb_left; node 597 lib/rbtree.c else if (node->rb_right) node 598 lib/rbtree.c node = node->rb_right; node 600 lib/rbtree.c return (struct rb_node *)node; node 604 lib/rbtree.c struct rb_node *rb_next_postorder(const struct rb_node *node) node 607 lib/rbtree.c if (!node) node 609 lib/rbtree.c parent = rb_parent(node); node 612 lib/rbtree.c if (parent && node == parent->rb_left && parent->rb_right) { node 32 lib/rbtree_test.c static void insert(struct test_node *node, struct rb_root_cached *root) node 35 lib/rbtree_test.c u32 key = node->key; node 45 lib/rbtree_test.c rb_link_node(&node->rb, parent, new); node 46 lib/rbtree_test.c rb_insert_color(&node->rb, &root->rb_root); node 49 lib/rbtree_test.c static void insert_cached(struct test_node *node, struct rb_root_cached *root) node 52 lib/rbtree_test.c u32 key = node->key; node 65 lib/rbtree_test.c rb_link_node(&node->rb, parent, new); node 66 lib/rbtree_test.c rb_insert_color_cached(&node->rb, root, leftmost); node 69 lib/rbtree_test.c static inline void erase(struct test_node *node, struct rb_root_cached *root) node 71 lib/rbtree_test.c rb_erase(&node->rb, &root->rb_root); node 74 lib/rbtree_test.c static inline void erase_cached(struct test_node *node, struct rb_root_cached *root) node 76 lib/rbtree_test.c rb_erase_cached(&node->rb, root); node 80 lib/rbtree_test.c #define NODE_VAL(node) ((node)->val) node 85 lib/rbtree_test.c static void insert_augmented(struct test_node *node, node 89 lib/rbtree_test.c u32 key = node->key; node 90 lib/rbtree_test.c u32 val = node->val; node 104 lib/rbtree_test.c node->augmented = val; node 105 lib/rbtree_test.c rb_link_node(&node->rb, rb_parent, new); node 106 lib/rbtree_test.c rb_insert_augmented(&node->rb, &root->rb_root, &augment_callbacks); node 109 lib/rbtree_test.c static void insert_augmented_cached(struct test_node *node, node 113 lib/rbtree_test.c u32 key = node->key; node 114 lib/rbtree_test.c u32 val = node->val; node 131 lib/rbtree_test.c node->augmented = val; node 132 lib/rbtree_test.c rb_link_node(&node->rb, rb_parent, new); node 133 lib/rbtree_test.c rb_insert_augmented_cached(&node->rb, root, node 138 lib/rbtree_test.c static void erase_augmented(struct test_node *node, struct rb_root_cached *root) node 140 lib/rbtree_test.c rb_erase_augmented(&node->rb, &root->rb_root, &augment_callbacks); node 143 lib/rbtree_test.c static void erase_augmented_cached(struct test_node *node, node 146 lib/rbtree_test.c rb_erase_augmented_cached(&node->rb, root, &augment_callbacks); node 198 lib/rbtree_test.c struct test_node *node = rb_entry(rb, struct test_node, rb); node 199 lib/rbtree_test.c WARN_ON_ONCE(node->key < prev_key); node 207 lib/rbtree_test.c prev_key = node->key; node 224 lib/rbtree_test.c struct test_node *node = rb_entry(rb, struct test_node, rb); node 225 lib/rbtree_test.c u32 subtree, max = node->val; node 226 lib/rbtree_test.c if (node->rb.rb_left) { node 227 lib/rbtree_test.c subtree = rb_entry(node->rb.rb_left, struct test_node, node 232 lib/rbtree_test.c if (node->rb.rb_right) { node 233 lib/rbtree_test.c subtree = rb_entry(node->rb.rb_right, struct test_node, node 238 lib/rbtree_test.c WARN_ON_ONCE(node->augmented != max); node 246 lib/rbtree_test.c struct rb_node *node; node 295 lib/rbtree_test.c for (node = rb_first(&root.rb_root); node; node = rb_next(node)) node 309 lib/rbtree_test.c node = rb_first(&root.rb_root); node 321 lib/rbtree_test.c node = rb_first_cached(&root); node 45 lib/sbitmap.c gfp_t flags, int node) node 76 lib/sbitmap.c sb->map = kcalloc_node(sb->map_nr, sizeof(*sb->map), flags, node); node 370 lib/sbitmap.c int shift, bool round_robin, gfp_t flags, int node) node 375 lib/sbitmap.c ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node); node 395 lib/sbitmap.c sbq->ws = kzalloc_node(SBQ_WAIT_QUEUES * sizeof(*sbq->ws), flags, node); node 64 lib/test_rhashtable.c struct rhash_head node; node 95 lib/test_rhashtable.c .head_offset = offsetof(struct test_obj, node), node 123 lib/test_rhashtable.c err = rhashtable_insert_fast(ht, &obj->node, params); node 254 lib/test_rhashtable.c rhashtable_remove_fast(ht, &obj->node, test_rht_params); node 670 lib/test_rhashtable.c err = rhashtable_remove_fast(&ht, &tdata->objs[i].node, node 305 lib/test_xarray.c struct xa_node *node; node 319 lib/test_xarray.c node = xas.xa_node; node 320 lib/test_xarray.c XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != xa_mk_value(0)); node 324 lib/test_xarray.c XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != XA_RETRY_ENTRY); node 337 lib/test_xarray.c node = xa_head(xa); node 342 lib/test_xarray.c XA_BUG_ON(xa, xa_head(xa) == node); node 346 lib/test_xarray.c XA_BUG_ON(xa, xa->xa_head != node); node 1560 lib/test_xarray.c static void test_update_node(struct xa_node *node) node 1562 lib/test_xarray.c if (node->count && node->count == node->nr_values) { node 1563 lib/test_xarray.c if (list_empty(&node->private_list)) node 1564 lib/test_xarray.c list_add(&shadow_nodes, &node->private_list); node 1566 lib/test_xarray.c if (!list_empty(&node->private_list)) node 1567 lib/test_xarray.c list_del_init(&node->private_list); node 1573 lib/test_xarray.c struct xa_node *node; node 1576 lib/test_xarray.c while ((node = list_first_entry_or_null(&shadow_nodes, node 1578 lib/test_xarray.c XA_STATE(xas, node->array, 0); node 1579 lib/test_xarray.c XA_BUG_ON(xa, node->array != xa); node 1580 lib/test_xarray.c list_del_init(&node->private_list); node 1581 lib/test_xarray.c xas.xa_node = xa_parent_locked(node->array, node); node 1582 lib/test_xarray.c xas.xa_offset = node->offset; node 1583 lib/test_xarray.c xas.xa_shift = node->shift + XA_CHUNK_SHIFT; node 27 lib/timerqueue.c bool timerqueue_add(struct timerqueue_head *head, struct timerqueue_node *node) node 35 lib/timerqueue.c WARN_ON_ONCE(!RB_EMPTY_NODE(&node->node)); node 39 lib/timerqueue.c ptr = rb_entry(parent, struct timerqueue_node, node); node 40 lib/timerqueue.c if (node->expires < ptr->expires) { node 47 lib/timerqueue.c rb_link_node(&node->node, parent, p); node 48 lib/timerqueue.c rb_insert_color_cached(&node->node, &head->rb_root, leftmost); node 63 lib/timerqueue.c bool timerqueue_del(struct timerqueue_head *head, struct timerqueue_node *node) node 65 lib/timerqueue.c WARN_ON_ONCE(RB_EMPTY_NODE(&node->node)); node 67 lib/timerqueue.c rb_erase_cached(&node->node, &head->rb_root); node 68 lib/timerqueue.c RB_CLEAR_NODE(&node->node); node 83 lib/timerqueue.c struct timerqueue_node *timerqueue_iterate_next(struct timerqueue_node *node) node 87 lib/timerqueue.c if (!node) node 89 lib/timerqueue.c next = rb_next(&node->node); node 92 lib/timerqueue.c return container_of(next, struct timerqueue_node, node); node 78 lib/xarray.c static inline unsigned long *node_marks(struct xa_node *node, xa_mark_t mark) node 80 lib/xarray.c return node->marks[(__force unsigned)mark]; node 83 lib/xarray.c static inline bool node_get_mark(struct xa_node *node, node 86 lib/xarray.c return test_bit(offset, node_marks(node, mark)); node 90 lib/xarray.c static inline bool node_set_mark(struct xa_node *node, unsigned int offset, node 93 lib/xarray.c return __test_and_set_bit(offset, node_marks(node, mark)); node 97 lib/xarray.c static inline bool node_clear_mark(struct xa_node *node, unsigned int offset, node 100 lib/xarray.c return __test_and_clear_bit(offset, node_marks(node, mark)); node 103 lib/xarray.c static inline bool node_any_mark(struct xa_node *node, xa_mark_t mark) node 105 lib/xarray.c return !bitmap_empty(node_marks(node, mark), XA_CHUNK_SIZE); node 108 lib/xarray.c static inline void node_mark_all(struct xa_node *node, xa_mark_t mark) node 110 lib/xarray.c bitmap_fill(node_marks(node, mark), XA_CHUNK_SIZE); node 142 lib/xarray.c static unsigned int get_offset(unsigned long index, struct xa_node *node) node 144 lib/xarray.c return (index >> node->shift) & XA_CHUNK_MASK; node 201 lib/xarray.c static void *xas_descend(struct xa_state *xas, struct xa_node *node) node 203 lib/xarray.c unsigned int offset = get_offset(xas->xa_index, node); node 204 lib/xarray.c void *entry = xa_entry(xas->xa, node, offset); node 206 lib/xarray.c xas->xa_node = node; node 209 lib/xarray.c entry = xa_entry(xas->xa, node, offset); node 236 lib/xarray.c struct xa_node *node = xa_to_node(entry); node 238 lib/xarray.c if (xas->xa_shift > node->shift) node 240 lib/xarray.c entry = xas_descend(xas, node); node 241 lib/xarray.c if (node->shift == 0) node 254 lib/xarray.c static void xa_node_free(struct xa_node *node) node 256 lib/xarray.c XA_NODE_BUG_ON(node, !list_empty(&node->private_list)); node 257 lib/xarray.c node->array = XA_RCU_FREE; node 258 lib/xarray.c call_rcu(&node->rcu_head, radix_tree_node_rcu_free); node 269 lib/xarray.c struct xa_node *node = xas->xa_alloc; node 271 lib/xarray.c if (!node) node 273 lib/xarray.c XA_NODE_BUG_ON(node, !list_empty(&node->private_list)); node 274 lib/xarray.c kmem_cache_free(radix_tree_node_cachep, node); node 347 lib/xarray.c static void xas_update(struct xa_state *xas, struct xa_node *node) node 350 lib/xarray.c xas->xa_update(node); node 352 lib/xarray.c XA_NODE_BUG_ON(node, !list_empty(&node->private_list)); node 358 lib/xarray.c struct xa_node *node = xas->xa_alloc; node 363 lib/xarray.c if (node) { node 371 lib/xarray.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp); node 372 lib/xarray.c if (!node) { node 379 lib/xarray.c node->offset = xas->xa_offset; node 381 lib/xarray.c XA_NODE_BUG_ON(node, parent->count > XA_CHUNK_SIZE); node 384 lib/xarray.c XA_NODE_BUG_ON(node, shift > BITS_PER_LONG); node 385 lib/xarray.c XA_NODE_BUG_ON(node, !list_empty(&node->private_list)); node 386 lib/xarray.c node->shift = shift; node 387 lib/xarray.c node->count = 0; node 388 lib/xarray.c node->nr_values = 0; node 389 lib/xarray.c RCU_INIT_POINTER(node->parent, xas->xa_node); node 390 lib/xarray.c node->array = xas->xa; node 392 lib/xarray.c return node; node 436 lib/xarray.c struct xa_node *node = xas->xa_node; node 441 lib/xarray.c XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE); node 442 lib/xarray.c if (node->count != 1) node 444 lib/xarray.c entry = xa_entry_locked(xa, node, 0); node 447 lib/xarray.c if (!xa_is_node(entry) && node->shift) node 454 lib/xarray.c if (xa_track_free(xa) && !node_get_mark(node, 0, XA_FREE_MARK)) node 457 lib/xarray.c node->count = 0; node 458 lib/xarray.c node->nr_values = 0; node 460 lib/xarray.c RCU_INIT_POINTER(node->slots[0], XA_RETRY_ENTRY); node 461 lib/xarray.c xas_update(xas, node); node 462 lib/xarray.c xa_node_free(node); node 465 lib/xarray.c node = xa_to_node(entry); node 466 lib/xarray.c node->parent = NULL; node 479 lib/xarray.c struct xa_node *node = xas->xa_node; node 484 lib/xarray.c XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE); node 485 lib/xarray.c if (node->count) node 488 lib/xarray.c parent = xa_parent_locked(xas->xa, node); node 490 lib/xarray.c xas->xa_offset = node->offset; node 491 lib/xarray.c xa_node_free(node); node 502 lib/xarray.c node = parent; node 503 lib/xarray.c xas_update(xas, node); node 506 lib/xarray.c if (!node->parent) node 522 lib/xarray.c struct xa_node *node = top; node 525 lib/xarray.c void *entry = xa_entry_locked(xas->xa, node, offset); node 527 lib/xarray.c if (node->shift && xa_is_node(entry)) { node 528 lib/xarray.c node = xa_to_node(entry); node 533 lib/xarray.c RCU_INIT_POINTER(node->slots[offset], XA_RETRY_ENTRY); node 538 lib/xarray.c parent = xa_parent_locked(xas->xa, node); node 539 lib/xarray.c offset = node->offset + 1; node 540 lib/xarray.c node->count = 0; node 541 lib/xarray.c node->nr_values = 0; node 542 lib/xarray.c xas_update(xas, node); node 543 lib/xarray.c xa_node_free(node); node 544 lib/xarray.c if (node == top) node 546 lib/xarray.c node = parent; node 558 lib/xarray.c struct xa_node *node = NULL; node 569 lib/xarray.c node = xa_to_node(head); node 570 lib/xarray.c shift = node->shift + XA_CHUNK_SHIFT; node 577 lib/xarray.c XA_NODE_BUG_ON(node, shift > BITS_PER_LONG); node 578 lib/xarray.c node = xas_alloc(xas, shift); node 579 lib/xarray.c if (!node) node 582 lib/xarray.c node->count = 1; node 584 lib/xarray.c node->nr_values = 1; node 585 lib/xarray.c RCU_INIT_POINTER(node->slots[0], head); node 590 lib/xarray.c node_mark_all(node, XA_FREE_MARK); node 592 lib/xarray.c node_clear_mark(node, 0, XA_FREE_MARK); node 596 lib/xarray.c node_set_mark(node, 0, mark); node 609 lib/xarray.c rcu_assign_pointer(xa_to_node(head)->parent, node); node 611 lib/xarray.c head = xa_mk_node(node); node 613 lib/xarray.c xas_update(xas, node); node 618 lib/xarray.c xas->xa_node = node; node 640 lib/xarray.c struct xa_node *node = xas->xa_node; node 644 lib/xarray.c if (xas_top(node)) { node 658 lib/xarray.c } else if (node) { node 661 lib/xarray.c shift = node->shift; node 662 lib/xarray.c entry = xa_entry_locked(xa, node, offset); node 663 lib/xarray.c slot = &node->slots[offset]; node 673 lib/xarray.c node = xas_alloc(xas, shift); node 674 lib/xarray.c if (!node) node 677 lib/xarray.c node_mark_all(node, XA_FREE_MARK); node 678 lib/xarray.c rcu_assign_pointer(*slot, xa_mk_node(node)); node 680 lib/xarray.c node = xa_to_node(entry); node 684 lib/xarray.c entry = xas_descend(xas, node); node 685 lib/xarray.c slot = &node->slots[xas->xa_offset]; node 721 lib/xarray.c struct xa_node *node = xas->xa_node; node 722 lib/xarray.c xas->xa_node = xa_parent_locked(xas->xa, node); node 723 lib/xarray.c xas->xa_offset = node->offset - 1; node 724 lib/xarray.c if (node->offset != 0) node 741 lib/xarray.c static void update_node(struct xa_state *xas, struct xa_node *node, node 744 lib/xarray.c if (!node || (!count && !values)) node 747 lib/xarray.c node->count += count; node 748 lib/xarray.c node->nr_values += values; node 749 lib/xarray.c XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE); node 750 lib/xarray.c XA_NODE_BUG_ON(node, node->nr_values > XA_CHUNK_SIZE); node 751 lib/xarray.c xas_update(xas, node); node 771 lib/xarray.c struct xa_node *node; node 788 lib/xarray.c node = xas->xa_node; node 789 lib/xarray.c if (node && (xas->xa_shift < node->shift)) node 797 lib/xarray.c if (node) { node 798 lib/xarray.c slot = &node->slots[offset]; node 814 lib/xarray.c if (xa_is_node(next) && (!node || node->shift)) node 816 lib/xarray.c if (!node) node 829 lib/xarray.c next = xa_entry_locked(xas->xa, node, ++offset); node 838 lib/xarray.c update_node(xas, node, count, values); node 872 lib/xarray.c struct xa_node *node = xas->xa_node; node 878 lib/xarray.c while (node) { node 879 lib/xarray.c if (node_set_mark(node, offset, mark)) node 881 lib/xarray.c offset = node->offset; node 882 lib/xarray.c node = xa_parent_locked(xas->xa, node); node 901 lib/xarray.c struct xa_node *node = xas->xa_node; node 907 lib/xarray.c while (node) { node 908 lib/xarray.c if (!node_clear_mark(node, offset, mark)) node 910 lib/xarray.c if (node_any_mark(node, mark)) node 913 lib/xarray.c offset = node->offset; node 914 lib/xarray.c node = xa_parent_locked(xas->xa, node); node 966 lib/xarray.c struct xa_node *node = xas->xa_node; node 972 lib/xarray.c if (node) { node 975 lib/xarray.c if (!xa_is_sibling(xa_entry(xas->xa, node, offset))) node 978 lib/xarray.c xas->xa_index += (offset - xas->xa_offset) << node->shift; node 1253 lib/xarray.c struct xa_node *node = xa_to_node(curr); node 1254 lib/xarray.c curr = xas_descend(xas, node); node 1838 lib/xarray.c struct xa_node *node = xas->xa_node; node 1841 lib/xarray.c if (!node) node 1843 lib/xarray.c mask = (XA_CHUNK_SIZE << node->shift) - 1; node 1845 lib/xarray.c ((unsigned long)xas->xa_offset << node->shift); node 2007 lib/xarray.c void xa_dump_node(const struct xa_node *node) node 2011 lib/xarray.c if (!node) node 2013 lib/xarray.c if ((unsigned long)node & 3) { node 2014 lib/xarray.c pr_cont("node %px\n", node); node 2020 lib/xarray.c node, node->parent ? "offset" : "max", node->offset, node 2021 lib/xarray.c node->parent, node->shift, node->count, node->nr_values, node 2022 lib/xarray.c node->array, node->private_list.prev, node->private_list.next); node 2025 lib/xarray.c pr_cont(" %lx", node->marks[i][j]); node 2051 lib/xarray.c struct xa_node *node = xa_to_node(entry); node 2052 lib/xarray.c xa_dump_node(node); node 2054 lib/xarray.c xa_dump_entry(node->slots[i], node 2055 lib/xarray.c index + (i << node->shift), node->shift); node 548 lib/zlib_deflate/deftree.c int node; /* new node being created */ node 571 lib/zlib_deflate/deftree.c node = s->heap[++(s->heap_len)] = (max_code < 2 ? ++max_code : 0); node 572 lib/zlib_deflate/deftree.c tree[node].Freq = 1; node 573 lib/zlib_deflate/deftree.c s->depth[node] = 0; node 574 lib/zlib_deflate/deftree.c s->opt_len--; if (stree) s->static_len -= stree[node].Len; node 587 lib/zlib_deflate/deftree.c node = elems; /* next internal node of the tree */ node 596 lib/zlib_deflate/deftree.c tree[node].Freq = tree[n].Freq + tree[m].Freq; node 597 lib/zlib_deflate/deftree.c s->depth[node] = (uch) (max(s->depth[n], s->depth[m]) + 1); node 598 lib/zlib_deflate/deftree.c tree[n].Dad = tree[m].Dad = (ush)node; node 602 lib/zlib_deflate/deftree.c node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq); node 606 lib/zlib_deflate/deftree.c s->heap[SMALLEST] = node++; node 409 mm/backing-dev.c struct rb_node **node, *parent; node 414 mm/backing-dev.c node = &bdi->cgwb_congested_tree.rb_node; node 417 mm/backing-dev.c while (*node != NULL) { node 418 mm/backing-dev.c parent = *node; node 422 mm/backing-dev.c node = &parent->rb_left; node 424 mm/backing-dev.c node = &parent->rb_right; node 432 mm/backing-dev.c rb_link_node(&congested->rb_node, parent, node); node 19 mm/cma_debug.c struct hlist_node node; node 74 mm/cma_debug.c hlist_add_head(&mem->node, &cma->mem_head); node 84 mm/cma_debug.c mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); node 85 mm/cma_debug.c hlist_del_init(&mem->node); node 2484 mm/compaction.c int compaction_register_node(struct node *node) node 2486 mm/compaction.c return device_create_file(&node->dev, &dev_attr_compact); node 2489 mm/compaction.c void compaction_unregister_node(struct node *node) node 2491 mm/compaction.c return device_remove_file(&node->dev, &dev_attr_compact); node 901 mm/hugetlb.c int node = NUMA_NO_NODE; node 916 mm/hugetlb.c if (zone_to_nid(zone) == node) node 918 mm/hugetlb.c node = zone_to_nid(zone); node 920 mm/hugetlb.c page = dequeue_huge_page_node_exact(h, node); node 1037 mm/hugetlb.c #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ node 1040 mm/hugetlb.c ((node = hstate_next_node_to_alloc(hs, mask)) || 1); \ node 1043 mm/hugetlb.c #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ node 1046 mm/hugetlb.c ((node = hstate_next_node_to_free(hs, mask)) || 1); \ node 1336 mm/hugetlb.c struct llist_node *node; node 1339 mm/hugetlb.c node = llist_del_all(&hpage_freelist); node 1341 mm/hugetlb.c while (node) { node 1342 mm/hugetlb.c page = container_of((struct address_space **)node, node 1344 mm/hugetlb.c node = node->next; node 1535 mm/hugetlb.c int nr_nodes, node; node 1538 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { node 1539 mm/hugetlb.c page = alloc_fresh_huge_page(h, gfp_mask, node, nodes_allowed, node 1562 mm/hugetlb.c int nr_nodes, node; node 1565 mm/hugetlb.c for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { node 1570 mm/hugetlb.c if ((!acct_surplus || h->surplus_huge_pages_node[node]) && node 1571 mm/hugetlb.c !list_empty(&h->hugepage_freelists[node])) { node 1573 mm/hugetlb.c list_entry(h->hugepage_freelists[node].next, node 1577 mm/hugetlb.c h->free_huge_pages_node[node]--; node 1580 mm/hugetlb.c h->surplus_huge_pages_node[node]--; node 1802 mm/hugetlb.c int node; node 1805 mm/hugetlb.c node = huge_node(vma, address, gfp_mask, &mpol, &nodemask); node 1806 mm/hugetlb.c page = alloc_huge_page_nodemask(h, node, nodemask); node 2222 mm/hugetlb.c int nr_nodes, node; node 2224 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { node 2229 mm/hugetlb.c 0, MEMBLOCK_ALLOC_ACCESSIBLE, node); node 2399 mm/hugetlb.c int nr_nodes, node; node 2404 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { node 2405 mm/hugetlb.c if (h->surplus_huge_pages_node[node]) node 2409 mm/hugetlb.c for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { node 2410 mm/hugetlb.c if (h->surplus_huge_pages_node[node] < node 2411 mm/hugetlb.c h->nr_huge_pages_node[node]) node 2419 mm/hugetlb.c h->surplus_huge_pages_node[node] += delta; node 2856 mm/hugetlb.c static void hugetlb_unregister_node(struct node *node) node 2859 mm/hugetlb.c struct node_hstate *nhs = &node_hstates[node->dev.id]; node 2881 mm/hugetlb.c static void hugetlb_register_node(struct node *node) node 2884 mm/hugetlb.c struct node_hstate *nhs = &node_hstates[node->dev.id]; node 2891 mm/hugetlb.c &node->dev.kobj); node 2901 mm/hugetlb.c h->name, node->dev.id); node 2902 mm/hugetlb.c hugetlb_unregister_node(node); node 2918 mm/hugetlb.c struct node *node = node_devices[nid]; node 2919 mm/hugetlb.c if (node->dev.id == nid) node 2920 mm/hugetlb.c hugetlb_register_node(node); node 3075 mm/hugetlb.c int node; node 3078 mm/hugetlb.c for_each_node_mask(node, cpuset_current_mems_allowed) node 3079 mm/hugetlb.c nr += array[node]; node 574 mm/internal.h extern struct page *alloc_new_node_page(struct page *page, unsigned long node); node 28 mm/interval_tree.c void vma_interval_tree_insert_after(struct vm_area_struct *node, node 34 mm/interval_tree.c unsigned long last = vma_last_pgoff(node); node 36 mm/interval_tree.c VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); node 55 mm/interval_tree.c node->shared.rb_subtree_last = last; node 56 mm/interval_tree.c rb_link_node(&node->shared.rb, &parent->shared.rb, link); node 57 mm/interval_tree.c rb_insert_augmented(&node->shared.rb, &root->rb_root, node 75 mm/interval_tree.c void anon_vma_interval_tree_insert(struct anon_vma_chain *node, node 79 mm/interval_tree.c node->cached_vma_start = avc_start_pgoff(node); node 80 mm/interval_tree.c node->cached_vma_last = avc_last_pgoff(node); node 82 mm/interval_tree.c __anon_vma_interval_tree_insert(node, root); node 85 mm/interval_tree.c void anon_vma_interval_tree_remove(struct anon_vma_chain *node, node 88 mm/interval_tree.c __anon_vma_interval_tree_remove(node, root); node 99 mm/interval_tree.c anon_vma_interval_tree_iter_next(struct anon_vma_chain *node, node 102 mm/interval_tree.c return __anon_vma_interval_tree_iter_next(node, first, last); node 106 mm/interval_tree.c void anon_vma_interval_tree_verify(struct anon_vma_chain *node) node 108 mm/interval_tree.c WARN_ON_ONCE(node->cached_vma_start != avc_start_pgoff(node)); node 109 mm/interval_tree.c WARN_ON_ONCE(node->cached_vma_last != avc_last_pgoff(node)); node 84 mm/kasan/init.c static __init void *early_alloc(size_t size, int node) node 87 mm/kasan/init.c MEMBLOCK_ALLOC_ACCESSIBLE, node); node 91 mm/kasan/init.c __func__, size, size, node, (u64)__pa(MAX_DMA_ADDRESS)); node 781 mm/khugepaged.c khugepaged_alloc_page(struct page **hpage, gfp_t gfp, int node) node 785 mm/khugepaged.c *hpage = __alloc_pages_node(node, gfp, HPAGE_PMD_ORDER); node 845 mm/khugepaged.c khugepaged_alloc_page(struct page **hpage, gfp_t gfp, int node) node 949 mm/khugepaged.c int node, int referenced) node 974 mm/khugepaged.c new_page = khugepaged_alloc_page(hpage, gfp, node); node 1126 mm/khugepaged.c int node = NUMA_NO_NODE, unmapped = 0; node 1184 mm/khugepaged.c node = page_to_nid(page); node 1185 mm/khugepaged.c if (khugepaged_scan_abort(node)) { node 1189 mm/khugepaged.c khugepaged_node_load[node]++; node 1230 mm/khugepaged.c node = khugepaged_find_target_node(); node 1232 mm/khugepaged.c collapse_huge_page(mm, address, hpage, node, referenced); node 1494 mm/khugepaged.c struct page **hpage, int node) node 1512 mm/khugepaged.c new_page = khugepaged_alloc_page(hpage, gfp, node); node 1822 mm/khugepaged.c int node = NUMA_NO_NODE; node 1846 mm/khugepaged.c node = page_to_nid(page); node 1847 mm/khugepaged.c if (khugepaged_scan_abort(node)) { node 1851 mm/khugepaged.c khugepaged_node_load[node]++; node 1883 mm/khugepaged.c node = khugepaged_find_target_node(); node 1884 mm/khugepaged.c collapse_file(mm, file, start, hpage, node); node 121 mm/kmemleak.c struct hlist_node node; node 475 mm/kmemleak.c hlist_for_each_entry_safe(area, tmp, &object->area_list, node) { node 476 mm/kmemleak.c hlist_del(&area->node); node 817 mm/kmemleak.c INIT_HLIST_NODE(&area->node); node 821 mm/kmemleak.c hlist_add_head(&area->node, &object->area_list); node 1352 mm/kmemleak.c hlist_for_each_entry(area, &object->area_list, node) node 157 mm/ksm.c struct rb_node node; /* when node of stable tree */ node 207 mm/ksm.c struct rb_node node; /* when node of unstable tree */ node 371 mm/ksm.c rb_erase(&dup->node, root_stable_tree + NUMA(dup->nid)); node 609 mm/ksm.c rb_replace_node(&dup->node, &chain->node, root); node 626 mm/ksm.c rb_erase(&chain->node, root); node 810 mm/ksm.c rb_erase(&rmap_item->node, node 947 mm/ksm.c struct stable_node, node); node 1428 mm/ksm.c rb_replace_node(&stable_node->node, &found->node, node 1578 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); node 1703 mm/ksm.c rb_link_node(&page_node->node, parent, new); node 1704 mm/ksm.c rb_insert_color(&page_node->node, root); node 1729 mm/ksm.c rb_replace_node(&stable_node_dup->node, node 1730 mm/ksm.c &page_node->node, node 1737 mm/ksm.c rb_erase(&stable_node_dup->node, root); node 1825 mm/ksm.c stable_node = rb_entry(*new, struct stable_node, node); node 1890 mm/ksm.c rb_link_node(&stable_node_dup->node, parent, new); node 1891 mm/ksm.c rb_insert_color(&stable_node_dup->node, root); node 1942 mm/ksm.c tree_rmap_item = rb_entry(*new, struct rmap_item, node); node 1982 mm/ksm.c rb_link_node(&rmap_item->node, parent, new); node 1983 mm/ksm.c rb_insert_color(&rmap_item->node, root); node 2769 mm/ksm.c struct rb_node *node; node 2773 mm/ksm.c node = rb_first(root_stable_tree + nid); node 2774 mm/ksm.c while (node) { node 2775 mm/ksm.c stable_node = rb_entry(node, struct stable_node, node); node 2780 mm/ksm.c node = rb_first(root_stable_tree + nid); node 2782 mm/ksm.c node = rb_next(node); node 128 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 152 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 187 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 204 mm/list_lru.c nlru = &lru->node[nid]; node 272 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 288 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 309 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 458 mm/list_lru.c if (memcg_init_list_lru_node(&lru->node[i])) node 464 mm/list_lru.c if (!lru->node[i].memcg_lrus) node 466 mm/list_lru.c memcg_destroy_list_lru_node(&lru->node[i]); node 479 mm/list_lru.c memcg_destroy_list_lru_node(&lru->node[i]); node 491 mm/list_lru.c if (memcg_update_list_lru_node(&lru->node[i], node 498 mm/list_lru.c if (!lru->node[i].memcg_lrus) node 501 mm/list_lru.c memcg_cancel_update_list_lru_node(&lru->node[i], node 516 mm/list_lru.c memcg_cancel_update_list_lru_node(&lru->node[i], node 544 mm/list_lru.c struct list_lru_node *nlru = &lru->node[nid]; node 614 mm/list_lru.c lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL); node 615 mm/list_lru.c if (!lru->node) node 619 mm/list_lru.c spin_lock_init(&lru->node[i].lock); node 621 mm/list_lru.c lockdep_set_class(&lru->node[i].lock, key); node 622 mm/list_lru.c init_one_lru(&lru->node[i].lru); node 627 mm/list_lru.c kfree(lru->node); node 629 mm/list_lru.c lru->node = NULL; node 643 mm/list_lru.c if (!lru->node) node 651 mm/list_lru.c kfree(lru->node); node 652 mm/list_lru.c lru->node = NULL; node 1685 mm/memcontrol.c int node; node 1688 mm/memcontrol.c node = memcg->last_scanned_node; node 1690 mm/memcontrol.c node = next_node_in(node, memcg->scan_nodes); node 1696 mm/memcontrol.c if (unlikely(node == MAX_NUMNODES)) node 1697 mm/memcontrol.c node = numa_node_id(); node 1699 mm/memcontrol.c memcg->last_scanned_node = node; node 1700 mm/memcontrol.c return node; node 3477 mm/memcontrol.c int node, cpu, i; node 3487 mm/memcontrol.c for_each_node(node) { node 3488 mm/memcontrol.c struct mem_cgroup_per_node *pn = memcg->nodeinfo[node]; node 3499 mm/memcontrol.c for (pi = pn; pi; pi = parent_nodeinfo(pi, node)) node 5023 mm/memcontrol.c static int alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) node 5026 mm/memcontrol.c int tmp = node; node 5035 mm/memcontrol.c if (!node_state(node, N_NORMAL_MEMORY)) node 5059 mm/memcontrol.c memcg->nodeinfo[node] = pn; node 5063 mm/memcontrol.c static void free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) node 5065 mm/memcontrol.c struct mem_cgroup_per_node *pn = memcg->nodeinfo[node]; node 5077 mm/memcontrol.c int node; node 5079 mm/memcontrol.c for_each_node(node) node 5080 mm/memcontrol.c free_mem_cgroup_per_node_info(memcg, node); node 5102 mm/memcontrol.c int node; node 5129 mm/memcontrol.c for_each_node(node) node 5130 mm/memcontrol.c if (alloc_mem_cgroup_per_node_info(memcg, node)) node 6970 mm/memcontrol.c int cpu, node; node 6990 mm/memcontrol.c for_each_node(node) { node 6994 mm/memcontrol.c node_online(node) ? node : NUMA_NO_NODE); node 6999 mm/memcontrol.c soft_limit_tree.rb_tree_per_node[node] = rtpn; node 228 mm/memory_hotplug.c int node = pgdat->node_id; node 235 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO); node 248 mm/memory_hotplug.c if (pfn_valid(pfn) && (early_pfn_to_nid(pfn) == node)) node 663 mm/memory_hotplug.c static void node_states_set_node(int node, struct memory_notify *arg) node 666 mm/memory_hotplug.c node_set_state(node, N_NORMAL_MEMORY); node 669 mm/memory_hotplug.c node_set_state(node, N_HIGH_MEMORY); node 672 mm/memory_hotplug.c node_set_state(node, N_MEMORY); node 1475 mm/memory_hotplug.c static void node_states_clear_node(int node, struct memory_notify *arg) node 1478 mm/memory_hotplug.c node_clear_state(node, N_NORMAL_MEMORY); node 1481 mm/memory_hotplug.c node_clear_state(node, N_HIGH_MEMORY); node 1484 mm/memory_hotplug.c node_clear_state(node, N_MEMORY); node 1492 mm/memory_hotplug.c int ret, node, nr_isolate_pageblock; node 1511 mm/memory_hotplug.c node = zone_to_nid(zone); node 1601 mm/memory_hotplug.c node_states_clear_node(node, &arg); node 1603 mm/memory_hotplug.c kswapd_stop(node); node 1604 mm/memory_hotplug.c kcompactd_stop(node); node 133 mm/mempolicy.c int node; node 138 mm/mempolicy.c node = numa_node_id(); node 139 mm/mempolicy.c if (node != NUMA_NO_NODE) { node 140 mm/mempolicy.c pol = &preferred_node_policy[node]; node 324 mm/mempolicy.c int node = first_node(pol->w.user_nodemask); node 326 mm/mempolicy.c if (node_isset(node, *nodes)) { node 327 mm/mempolicy.c pol->v.preferred_node = node; node 1000 mm/mempolicy.c struct page *alloc_new_node_page(struct page *page, unsigned long node) node 1004 mm/mempolicy.c node); node 1008 mm/mempolicy.c thp = alloc_pages_node(node, node 1016 mm/mempolicy.c return __alloc_pages_node(node, GFP_HIGHUSER_MOVABLE | node 1844 mm/mempolicy.c int node = numa_mem_id(); node 1847 mm/mempolicy.c return node; node 1851 mm/mempolicy.c return node; node 1872 mm/mempolicy.c zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; node 1875 mm/mempolicy.c return z->zone ? zone_to_nid(z->zone) : node; node 2099 mm/mempolicy.c unsigned long addr, int node, bool hugepage) node 2118 mm/mempolicy.c int hpage_node = node; node 2154 mm/mempolicy.c preferred_nid = policy_node(gfp, pol, node); node 2475 mm/mempolicy.c static void sp_node_init(struct sp_node *node, unsigned long start, node 2478 mm/mempolicy.c node->start = start; node 2479 mm/mempolicy.c node->end = end; node 2480 mm/mempolicy.c node->policy = pol; node 1502 mm/migrate.c struct list_head *pagelist, int node) node 1509 mm/migrate.c err = migrate_pages(pagelist, alloc_new_node_page, NULL, node, node 1526 mm/migrate.c int node, struct list_head *pagelist, bool migrate_all) node 1552 mm/migrate.c if (page_to_nid(page) == node) node 1610 mm/migrate.c int node; node 1615 mm/migrate.c if (get_user(node, nodes + i)) node 1620 mm/migrate.c if (node < 0 || node >= MAX_NUMNODES) node 1622 mm/migrate.c if (!node_state(node, N_MEMORY)) node 1626 mm/migrate.c if (!node_isset(node, task_nodes)) node 1630 mm/migrate.c current_node = node; node 1632 mm/migrate.c } else if (node != current_node) { node 1651 mm/migrate.c current_node = node; node 1970 mm/migrate.c int node) node 1972 mm/migrate.c pg_data_t *pgdat = NODE_DATA(node); node 1998 mm/migrate.c NULL, node, MIGRATE_ASYNC, node 2028 mm/migrate.c struct page *page, int node) node 2031 mm/migrate.c pg_data_t *pgdat = NODE_DATA(node); node 2037 mm/migrate.c new_page = alloc_pages_node(node, node 153 mm/nommu.c void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags) node 256 mm/nommu.c void *vmalloc_node(unsigned long size, int node) node 274 mm/nommu.c void *vzalloc_node(unsigned long size, int node) node 343 mm/nommu.c void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) node 307 mm/page-writeback.c int node; node 311 mm/page-writeback.c for_each_node_state(node, N_HIGH_MEMORY) { node 319 mm/page-writeback.c z = &NODE_DATA(node)->node_zones[i]; node 1472 mm/page_alloc.c static inline bool __meminit early_pfn_in_nid(unsigned long pfn, int node) node 1477 mm/page_alloc.c if (nid >= 0 && nid != node) node 1483 mm/page_alloc.c static inline bool __meminit early_pfn_in_nid(unsigned long pfn, int node) node 5535 mm/page_alloc.c static int find_next_best_node(int node, nodemask_t *used_node_mask) node 5543 mm/page_alloc.c if (!node_isset(node, *used_node_mask)) { node 5544 mm/page_alloc.c node_set(node, *used_node_mask); node 5545 mm/page_alloc.c return node; node 5555 mm/page_alloc.c val = node_distance(node, n); node 5558 mm/page_alloc.c val += (n < node); node 5598 mm/page_alloc.c pg_data_t *node = NODE_DATA(node_order[i]); node 5600 mm/page_alloc.c nr_zones = build_zonerefs_node(node, zonerefs); node 5632 mm/page_alloc.c int node, load, nr_nodes = 0; node 5643 mm/page_alloc.c while ((node = find_next_best_node(local_node, &used_mask)) >= 0) { node 5649 mm/page_alloc.c if (node_distance(local_node, node) != node 5651 mm/page_alloc.c node_load[node] = load; node 5653 mm/page_alloc.c node_order[nr_nodes++] = node; node 5654 mm/page_alloc.c prev_node = node; node 5669 mm/page_alloc.c int local_memory_node(int node) node 5673 mm/page_alloc.c z = first_zones_zonelist(node_zonelist(node, GFP_KERNEL), node 5686 mm/page_alloc.c int node, local_node; node 5704 mm/page_alloc.c for (node = local_node + 1; node < MAX_NUMNODES; node++) { node 5705 mm/page_alloc.c if (!node_online(node)) node 5707 mm/page_alloc.c nr_zones = build_zonerefs_node(NODE_DATA(node), zonerefs); node 5710 mm/page_alloc.c for (node = 0; node < local_node; node++) { node 5711 mm/page_alloc.c if (!node_online(node)) node 5713 mm/page_alloc.c nr_zones = build_zonerefs_node(NODE_DATA(node), zonerefs); node 212 mm/slab.c int node, struct list_head *list); node 492 mm/slab.c int node = __this_cpu_read(slab_reap_node); node 494 mm/slab.c node = next_node_in(node, node_online_map); node 495 mm/slab.c __this_cpu_write(slab_reap_node, node); node 532 mm/slab.c static struct array_cache *alloc_arraycache(int node, int entries, node 538 mm/slab.c ac = kmalloc_node(memsize, gfp, node); node 596 mm/slab.c static inline struct alien_cache **alloc_alien_cache(int node, node 633 mm/slab.c static struct alien_cache *__alloc_alien_cache(int node, int entries, node 639 mm/slab.c alc = kmalloc_node(memsize, gfp, node); node 648 mm/slab.c static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp) node 655 mm/slab.c alc_ptr = kcalloc_node(nr_node_ids, sizeof(void *), gfp, node); node 660 mm/slab.c if (i == node || !node_online(i)) node 662 mm/slab.c alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp); node 685 mm/slab.c struct array_cache *ac, int node, node 688 mm/slab.c struct kmem_cache_node *n = get_node(cachep, node); node 700 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, list); node 711 mm/slab.c int node = __this_cpu_read(slab_reap_node); node 714 mm/slab.c struct alien_cache *alc = n->alien[node]; node 722 mm/slab.c __drain_alien_cache(cachep, ac, node, &list); node 753 mm/slab.c int node, int page_node) node 760 mm/slab.c n = get_node(cachep, node); node 786 mm/slab.c int node = numa_mem_id(); node 791 mm/slab.c if (likely(node == page_node)) node 794 mm/slab.c return __cache_free_alien(cachep, objp, node, page_node); node 807 mm/slab.c static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp) node 816 mm/slab.c n = get_node(cachep, node); node 819 mm/slab.c n->free_limit = (1 + nr_cpus_node(node)) * cachep->batchcount + node 826 mm/slab.c n = kmalloc_node(sizeof(struct kmem_cache_node), gfp, node); node 835 mm/slab.c (1 + nr_cpus_node(node)) * cachep->batchcount + cachep->num; node 842 mm/slab.c cachep->node[node] = n; node 857 mm/slab.c static int init_cache_node_node(int node) node 863 mm/slab.c ret = init_cache_node(cachep, node, GFP_KERNEL); node 873 mm/slab.c int node, gfp_t gfp, bool force_change) node 883 mm/slab.c new_alien = alloc_alien_cache(node, cachep->limit, gfp); node 889 mm/slab.c new_shared = alloc_arraycache(node, node 895 mm/slab.c ret = init_cache_node(cachep, node, gfp); node 899 mm/slab.c n = get_node(cachep, node); node 903 mm/slab.c n->shared->avail, node, &list); node 944 mm/slab.c int node = cpu_to_mem(cpu); node 945 mm/slab.c const struct cpumask *mask = cpumask_of_node(node); node 953 mm/slab.c n = get_node(cachep, node); node 964 mm/slab.c free_block(cachep, nc->entry, nc->avail, node, &list); node 975 mm/slab.c shared->avail, node, &list); node 999 mm/slab.c n = get_node(cachep, node); node 1009 mm/slab.c int node = cpu_to_mem(cpu); node 1018 mm/slab.c err = init_cache_node_node(node); node 1027 mm/slab.c err = setup_kmem_cache_node(cachep, node, GFP_KERNEL, false); node 1095 mm/slab.c static int __meminit drain_cache_node_node(int node) node 1103 mm/slab.c n = get_node(cachep, node); node 1169 mm/slab.c cachep->node[nodeid] = ptr; node 1178 mm/slab.c int node; node 1180 mm/slab.c for_each_online_node(node) { node 1181 mm/slab.c cachep->node[node] = &init_kmem_cache_node[index + node]; node 1182 mm/slab.c cachep->node[node]->next_reap = jiffies + node 1238 mm/slab.c offsetof(struct kmem_cache, node) + node 1322 mm/slab.c int node; node 1334 mm/slab.c for_each_kmem_cache_node(cachep, node, n) { node 1344 mm/slab.c node, total_slabs - free_slabs, total_slabs, node 1758 mm/slab.c int node; node 1760 mm/slab.c for_each_online_node(node) { node 1761 mm/slab.c cachep->node[node] = kmalloc_node( node 1762 mm/slab.c sizeof(struct kmem_cache_node), gfp, node); node 1763 mm/slab.c BUG_ON(!cachep->node[node]); node 1764 mm/slab.c kmem_cache_node_init(cachep->node[node]); node 1768 mm/slab.c cachep->node[numa_mem_id()]->next_reap = node 2103 mm/slab.c static void check_spinlock_acquired_node(struct kmem_cache *cachep, int node) node 2107 mm/slab.c assert_spin_locked(&get_node(cachep, node)->list_lock); node 2120 mm/slab.c int node, bool free_all, struct list_head *list) node 2131 mm/slab.c free_block(cachep, ac->entry, tofree, node, list); node 2140 mm/slab.c int node = numa_mem_id(); node 2146 mm/slab.c n = get_node(cachep, node); node 2148 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, &list); node 2157 mm/slab.c int node; node 2162 mm/slab.c for_each_kmem_cache_node(cachep, node, n) node 2166 mm/slab.c for_each_kmem_cache_node(cachep, node, n) { node 2168 mm/slab.c drain_array_locked(cachep, n->shared, node, true, &list); node 2217 mm/slab.c int node; node 2220 mm/slab.c for_each_kmem_cache_node(s, node, n) node 2230 mm/slab.c int node; node 2236 mm/slab.c for_each_kmem_cache_node(cachep, node, n) { node 2275 mm/slab.c cachep->node[i] = NULL; node 2910 mm/slab.c int node; node 2915 mm/slab.c node = numa_mem_id(); node 2927 mm/slab.c n = get_node(cachep, node); node 2971 mm/slab.c page = cache_grow_begin(cachep, gfp_exact_node(flags), node); node 3328 mm/slab.c int nr_objects, int node, struct list_head *list) node 3331 mm/slab.c struct kmem_cache_node *n = get_node(cachep, node); node 3344 mm/slab.c check_spinlock_acquired_node(cachep, node); node 3375 mm/slab.c int node = numa_mem_id(); node 3381 mm/slab.c n = get_node(cachep, node); node 3396 mm/slab.c free_block(cachep, ac->entry, batchcount, node, &list); node 3604 mm/slab.c __do_kmalloc_node(size_t size, gfp_t flags, int node, unsigned long caller) node 3614 mm/slab.c ret = kmem_cache_alloc_node_trace(cachep, flags, node, size); node 3620 mm/slab.c void *__kmalloc_node(size_t size, gfp_t flags, int node) node 3622 mm/slab.c return __do_kmalloc_node(size, flags, node, _RET_IP_); node 3627 mm/slab.c int node, unsigned long caller) node 3629 mm/slab.c return __do_kmalloc_node(size, flags, node, caller); node 3767 mm/slab.c int node; node 3770 mm/slab.c for_each_online_node(node) { node 3771 mm/slab.c ret = setup_kmem_cache_node(cachep, node, gfp, true); node 3782 mm/slab.c node--; node 3783 mm/slab.c while (node >= 0) { node 3784 mm/slab.c n = get_node(cachep, node); node 3789 mm/slab.c cachep->node[node] = NULL; node 3791 mm/slab.c node--; node 3827 mm/slab.c int node; node 3831 mm/slab.c node = cpu_to_mem(cpu); node 3832 mm/slab.c n = get_node(cachep, node); node 3834 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, &list); node 3945 mm/slab.c struct array_cache *ac, int node) node 3961 mm/slab.c drain_array_locked(cachep, ac, node, false, &list); node 3983 mm/slab.c int node = numa_mem_id(); node 3998 mm/slab.c n = get_node(searchp, node); node 4002 mm/slab.c drain_array(searchp, n, cpu_cache_get(searchp), node); node 4013 mm/slab.c drain_array(searchp, n, n->shared, node); node 4041 mm/slab.c int node; node 4044 mm/slab.c for_each_kmem_cache_node(cachep, node, n) { node 628 mm/slab.h static inline struct kmem_cache_node *get_node(struct kmem_cache *s, int node) node 630 mm/slab.h return s->node[node]; node 191 mm/slob.c static void *slob_new_pages(gfp_t gfp, int order, int node) node 196 mm/slob.c if (node != NUMA_NO_NODE) node 197 mm/slob.c page = __alloc_pages_node(node, gfp, order); node 301 mm/slob.c static void *slob_alloc(size_t size, gfp_t gfp, int align, int node, node 326 mm/slob.c if (node != NUMA_NO_NODE && page_to_nid(sp) != node) node 358 mm/slob.c b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node); node 469 mm/slob.c __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller) node 493 mm/slob.c m = slob_alloc(size + minalign, gfp, align, node, minalign); node 501 mm/slob.c size, size + minalign, gfp, node); node 507 mm/slob.c ret = slob_new_pages(gfp, order, node); node 510 mm/slob.c size, PAGE_SIZE << order, gfp, node); node 530 mm/slob.c int node, unsigned long caller) node 532 mm/slob.c return __do_kmalloc_node(size, gfp, node, caller); node 592 mm/slob.c static void *slob_alloc_node(struct kmem_cache *c, gfp_t flags, int node) node 602 mm/slob.c b = slob_alloc(c->size, flags, c->align, node, 0); node 605 mm/slob.c flags, node); node 607 mm/slob.c b = slob_new_pages(flags, get_order(c->size), node); node 610 mm/slob.c flags, node); node 629 mm/slob.c void *__kmalloc_node(size_t size, gfp_t gfp, int node) node 631 mm/slob.c return __do_kmalloc_node(size, gfp, node, _RET_IP_); node 635 mm/slob.c void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node) node 637 mm/slob.c return slob_alloc_node(cachep, gfp, node); node 1031 mm/slub.c static inline unsigned long slabs_node(struct kmem_cache *s, int node) node 1033 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 1043 mm/slub.c static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects) node 1045 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 1058 mm/slub.c static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects) node 1060 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 1373 mm/slub.c static inline unsigned long slabs_node(struct kmem_cache *s, int node) node 1377 mm/slub.c static inline void inc_slabs_node(struct kmem_cache *s, int node, node 1379 mm/slub.c static inline void dec_slabs_node(struct kmem_cache *s, int node, node 1489 mm/slub.c gfp_t flags, int node, struct kmem_cache_order_objects oo) node 1494 mm/slub.c if (node == NUMA_NO_NODE) node 1497 mm/slub.c page = __alloc_pages_node(node, flags, order); node 1616 mm/slub.c static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node) node 1640 mm/slub.c page = alloc_slab_page(s, alloc_gfp, node, oo); node 1648 mm/slub.c page = alloc_slab_page(s, alloc_gfp, node, oo); node 1696 mm/slub.c static struct page *new_slab(struct kmem_cache *s, gfp_t flags, int node) node 1707 mm/slub.c flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node); node 1948 mm/slub.c static void *get_partial(struct kmem_cache *s, gfp_t flags, int node, node 1952 mm/slub.c int searchnode = node; node 1954 mm/slub.c if (node == NUMA_NO_NODE) node 1958 mm/slub.c if (object || node != NUMA_NO_NODE) node 2367 mm/slub.c static inline int node_match(struct page *page, int node) node 2370 mm/slub.c if (node != NUMA_NO_NODE && page_to_nid(page) != node) node 2410 mm/slub.c int node; node 2426 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 2436 mm/slub.c node, nr_slabs, nr_objs, nr_free); node 2442 mm/slub.c int node, struct kmem_cache_cpu **pc) node 2450 mm/slub.c freelist = get_partial(s, flags, node, c); node 2455 mm/slub.c page = new_slab(s, flags, node); node 2537 mm/slub.c static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, node 2549 mm/slub.c if (unlikely(node != NUMA_NO_NODE && node 2550 mm/slub.c !node_state(node, N_NORMAL_MEMORY))) node 2551 mm/slub.c node = NUMA_NO_NODE; node 2556 mm/slub.c if (unlikely(!node_match(page, node))) { node 2561 mm/slub.c if (!node_state(node, N_NORMAL_MEMORY)) { node 2562 mm/slub.c node = NUMA_NO_NODE; node 2616 mm/slub.c freelist = new_slab_objects(s, gfpflags, node, &c); node 2619 mm/slub.c slab_out_of_memory(s, gfpflags, node); node 2640 mm/slub.c static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, node 2656 mm/slub.c p = ___slab_alloc(s, gfpflags, node, addr, c); node 2683 mm/slub.c gfp_t gfpflags, int node, unsigned long addr) node 2729 mm/slub.c if (unlikely(!object || !node_match(page, node))) { node 2730 mm/slub.c object = __slab_alloc(s, gfpflags, node, addr, c); node 2800 mm/slub.c void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node) node 2802 mm/slub.c void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_); node 2805 mm/slub.c s->object_size, s->size, gfpflags, node); node 2814 mm/slub.c int node, size_t size) node 2816 mm/slub.c void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_); node 2819 mm/slub.c size, s->size, gfpflags, node); node 3385 mm/slub.c static void early_kmem_cache_node_alloc(int node) node 3392 mm/slub.c page = new_slab(kmem_cache_node, GFP_NOWAIT, node); node 3395 mm/slub.c if (page_to_nid(page) != node) { node 3396 mm/slub.c pr_err("SLUB: Unable to allocate memory from node %d\n", node); node 3411 mm/slub.c kmem_cache_node->node[node] = n; node 3413 mm/slub.c inc_slabs_node(kmem_cache_node, node, page->objects); node 3424 mm/slub.c int node; node 3427 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 3428 mm/slub.c s->node[node] = NULL; node 3442 mm/slub.c int node; node 3444 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 3448 mm/slub.c early_kmem_cache_node_alloc(node); node 3452 mm/slub.c GFP_KERNEL, node); node 3460 mm/slub.c s->node[node] = n; node 3744 mm/slub.c int node; node 3747 mm/slub.c for_each_kmem_cache_node(s, node, n) node 3748 mm/slub.c if (n->nr_partial || slabs_node(s, node)) node 3758 mm/slub.c int node; node 3763 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 3765 mm/slub.c if (n->nr_partial || slabs_node(s, node)) node 3828 mm/slub.c static void *kmalloc_large_node(size_t size, gfp_t flags, int node) node 3835 mm/slub.c page = alloc_pages_node(node, flags, order); node 3845 mm/slub.c void *__kmalloc_node(size_t size, gfp_t flags, int node) node 3851 mm/slub.c ret = kmalloc_large_node(size, flags, node); node 3855 mm/slub.c flags, node); node 3865 mm/slub.c ret = slab_alloc_node(s, flags, node, _RET_IP_); node 3867 mm/slub.c trace_kmalloc_node(_RET_IP_, ret, size, s->size, flags, node); node 3992 mm/slub.c int node; node 4003 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 4045 mm/slub.c if (slabs_node(s, node)) node 4122 mm/slub.c s->node[offline_node] = NULL; node 4162 mm/slub.c s->node[nid] = n; node 4213 mm/slub.c int node; node 4225 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 4262 mm/slub.c offsetof(struct kmem_cache, node) + node 4363 mm/slub.c int node, unsigned long caller) node 4369 mm/slub.c ret = kmalloc_large_node(size, gfpflags, node); node 4373 mm/slub.c gfpflags, node); node 4383 mm/slub.c ret = slab_alloc_node(s, gfpflags, node, caller); node 4386 mm/slub.c trace_kmalloc_node(caller, ret, size, s->size, gfpflags, node); node 4475 mm/slub.c int node; node 4484 mm/slub.c for_each_kmem_cache_node(s, node, n) node 4636 mm/slub.c int node; node 4648 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 4810 mm/slub.c int node; node 4824 mm/slub.c int node; node 4831 mm/slub.c node = page_to_nid(page); node 4840 mm/slub.c nodes[node] += x; node 4844 mm/slub.c node = page_to_nid(page); node 4852 mm/slub.c nodes[node] += x; node 4872 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 4882 mm/slub.c nodes[node] += x; node 4890 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 4898 mm/slub.c nodes[node] += x; node 4903 mm/slub.c for (node = 0; node < nr_node_ids; node++) node 4904 mm/slub.c if (nodes[node]) node 4906 mm/slub.c node, nodes[node]); node 4915 mm/slub.c int node; node 4918 mm/slub.c for_each_kmem_cache_node(s, node, n) node 5923 mm/slub.c int node; node 5926 mm/slub.c for_each_kmem_cache_node(s, node, n) { node 40 mm/sparse-vmemmap.c static void * __ref __earlyonly_bootmem_alloc(int node, node 46 mm/sparse-vmemmap.c MEMBLOCK_ALLOC_ACCESSIBLE, node); node 49 mm/sparse-vmemmap.c void * __meminit vmemmap_alloc_block(unsigned long size, int node) node 58 mm/sparse-vmemmap.c page = alloc_pages_node(node, gfp_mask, order); node 69 mm/sparse-vmemmap.c return __earlyonly_bootmem_alloc(node, size, size, node 74 mm/sparse-vmemmap.c void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node) node 79 mm/sparse-vmemmap.c ptr = vmemmap_alloc_block(size, node); node 132 mm/sparse-vmemmap.c void __meminit vmemmap_verify(pte_t *pte, int node, node 138 mm/sparse-vmemmap.c if (node_distance(actual_node, node) > LOCAL_DISTANCE) node 143 mm/sparse-vmemmap.c pte_t * __meminit vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node) node 148 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block_buf(PAGE_SIZE, node); node 157 mm/sparse-vmemmap.c static void * __meminit vmemmap_alloc_block_zero(unsigned long size, int node) node 159 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(size, node); node 168 mm/sparse-vmemmap.c pmd_t * __meminit vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node) node 172 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node); node 180 mm/sparse-vmemmap.c pud_t * __meminit vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node) node 184 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node); node 192 mm/sparse-vmemmap.c p4d_t * __meminit vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node) node 196 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node); node 204 mm/sparse-vmemmap.c pgd_t * __meminit vmemmap_pgd_populate(unsigned long addr, int node) node 208 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node); node 217 mm/sparse-vmemmap.c unsigned long end, int node) node 227 mm/sparse-vmemmap.c pgd = vmemmap_pgd_populate(addr, node); node 230 mm/sparse-vmemmap.c p4d = vmemmap_p4d_populate(pgd, addr, node); node 233 mm/sparse-vmemmap.c pud = vmemmap_pud_populate(p4d, addr, node); node 236 mm/sparse-vmemmap.c pmd = vmemmap_pmd_populate(pud, addr, node); node 239 mm/sparse-vmemmap.c pte = vmemmap_pte_populate(pmd, addr, node); node 242 mm/sparse-vmemmap.c vmemmap_verify(pte, node, addr, addr + PAGE_SIZE); node 998 mm/swapfile.c int node; node 1018 mm/swapfile.c node = numa_node_id(); node 1019 mm/swapfile.c plist_for_each_entry_safe(si, next, &swap_avail_heads[node], avail_lists[node]) { node 1021 mm/swapfile.c plist_requeue(&si->avail_lists[node], &swap_avail_heads[node]); node 1026 mm/swapfile.c if (plist_node_empty(&si->avail_lists[node])) { node 1065 mm/swapfile.c if (plist_node_empty(&next->avail_lists[node])) node 3740 mm/swapfile.c void mem_cgroup_throttle_swaprate(struct mem_cgroup *memcg, int node, node 3758 mm/swapfile.c plist_for_each_entry_safe(si, next, &swap_avail_heads[node], node 3759 mm/swapfile.c avail_lists[node]) { node 538 mm/util.c void *kvmalloc_node(size_t size, gfp_t flags, int node) node 548 mm/util.c return kmalloc_node(size, flags, node); node 564 mm/util.c ret = kmalloc_node(size, kmalloc_flags, node); node 573 mm/util.c return __vmalloc_node_flags_caller(size, node, flags, node 381 mm/vmalloc.c get_subtree_max_size(struct rb_node *node) node 385 mm/vmalloc.c va = rb_entry_safe(node, struct vmap_area, rb_node); node 560 mm/vmalloc.c struct rb_node *node; node 569 mm/vmalloc.c node = n; node 571 mm/vmalloc.c while (node) { node 572 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); node 574 mm/vmalloc.c if (get_subtree_max_size(node->rb_left) == size) { node 575 mm/vmalloc.c node = node->rb_left; node 582 mm/vmalloc.c node = node->rb_right; node 627 mm/vmalloc.c struct rb_node *node = &va->rb_node; node 630 mm/vmalloc.c while (node) { node 631 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); node 644 mm/vmalloc.c node = rb_parent(&va->rb_node); node 793 mm/vmalloc.c struct rb_node *node; node 797 mm/vmalloc.c node = free_vmap_area_root.rb_node; node 802 mm/vmalloc.c while (node) { node 803 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); node 805 mm/vmalloc.c if (get_subtree_max_size(node->rb_left) >= length && node 807 mm/vmalloc.c node = node->rb_left; node 817 mm/vmalloc.c if (get_subtree_max_size(node->rb_right) >= length) { node 818 mm/vmalloc.c node = node->rb_right; node 827 mm/vmalloc.c while ((node = rb_parent(node))) { node 828 mm/vmalloc.c va = rb_entry(node, struct vmap_area, rb_node); node 832 mm/vmalloc.c if (get_subtree_max_size(node->rb_right) >= length && node 834 mm/vmalloc.c node = node->rb_right; node 1053 mm/vmalloc.c int node, gfp_t gfp_mask) node 1069 mm/vmalloc.c gfp_mask & GFP_RECLAIM_MASK, node); node 1095 mm/vmalloc.c pva = kmem_cache_alloc_node(vmap_area_cachep, GFP_KERNEL, node); node 1463 mm/vmalloc.c int node, err; node 1466 mm/vmalloc.c node = numa_node_id(); node 1469 mm/vmalloc.c gfp_mask & GFP_RECLAIM_MASK, node); node 1475 mm/vmalloc.c node, gfp_mask); node 1778 mm/vmalloc.c void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) node 1792 mm/vmalloc.c VMALLOC_START, VMALLOC_END, node, GFP_KERNEL); node 2043 mm/vmalloc.c unsigned long end, int node, gfp_t gfp_mask, const void *caller) node 2057 mm/vmalloc.c area = kzalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node); node 2064 mm/vmalloc.c va = alloc_vmap_area(size, align, start, end, node, gfp_mask); node 2397 mm/vmalloc.c int node, const void *caller); node 2399 mm/vmalloc.c pgprot_t prot, int node) node 2415 mm/vmalloc.c PAGE_KERNEL, node, area->caller); node 2417 mm/vmalloc.c pages = kmalloc_node(array_size, nested_gfp, node); node 2432 mm/vmalloc.c if (node == NUMA_NO_NODE) node 2435 mm/vmalloc.c page = alloc_pages_node(node, alloc_mask|highmem_mask, 0); node 2481 mm/vmalloc.c pgprot_t prot, unsigned long vm_flags, int node, node 2493 mm/vmalloc.c vm_flags, start, end, node, gfp_mask, caller); node 2497 mm/vmalloc.c addr = __vmalloc_area_node(area, gfp_mask, prot, node); node 2550 mm/vmalloc.c int node, const void *caller) node 2553 mm/vmalloc.c gfp_mask, prot, 0, node, caller); node 2564 mm/vmalloc.c int node, gfp_t flags) node 2567 mm/vmalloc.c node, __builtin_return_address(0)); node 2571 mm/vmalloc.c void *__vmalloc_node_flags_caller(unsigned long size, int node, gfp_t flags, node 2574 mm/vmalloc.c return __vmalloc_node(size, 1, flags, PAGE_KERNEL, node, caller); node 2647 mm/vmalloc.c void *vmalloc_node(unsigned long size, int node) node 2650 mm/vmalloc.c node, __builtin_return_address(0)); node 2668 mm/vmalloc.c void *vzalloc_node(unsigned long size, int node) node 2670 mm/vmalloc.c return __vmalloc_node_flags(size, node, node 155 mm/vmpressure.c struct list_head node; node 166 mm/vmpressure.c list_for_each_entry(ev, &vmpr->events, node) { node 406 mm/vmpressure.c list_add(&ev->node, &vmpr->events); node 432 mm/vmpressure.c list_for_each_entry(ev, &vmpr->events, node) { node 435 mm/vmpressure.c list_del(&ev->node); node 959 mm/vmstat.c unsigned long sum_zone_node_page_state(int node, node 962 mm/vmstat.c struct zone *zones = NODE_DATA(node)->node_zones; node 976 mm/vmstat.c unsigned long sum_zone_numa_state(int node, node 979 mm/vmstat.c struct zone *zones = NODE_DATA(node)->node_zones; node 1303 mm/vmstat.c loff_t node = *pos; node 1306 mm/vmstat.c pgdat && node; node 1308 mm/vmstat.c --node; node 1925 mm/vmstat.c int node; node 1927 mm/vmstat.c for_each_online_node(node) { node 1928 mm/vmstat.c if (cpumask_weight(cpumask_of_node(node)) > 0) node 1929 mm/vmstat.c node_set_state(node, N_CPU); node 1949 mm/vmstat.c int node; node 1951 mm/vmstat.c node = cpu_to_node(cpu); node 1954 mm/vmstat.c node_cpus = cpumask_of_node(node); node 1958 mm/vmstat.c node_clear_state(node, N_CPU); node 368 mm/workingset.c void workingset_update_node(struct xa_node *node) node 380 mm/workingset.c if (node->count && node->count == node->nr_values) { node 381 mm/workingset.c if (list_empty(&node->private_list)) { node 382 mm/workingset.c list_lru_add(&shadow_nodes, &node->private_list); node 383 mm/workingset.c __inc_lruvec_slab_state(node, WORKINGSET_NODES); node 386 mm/workingset.c if (!list_empty(&node->private_list)) { node 387 mm/workingset.c list_lru_del(&shadow_nodes, &node->private_list); node 388 mm/workingset.c __dec_lruvec_slab_state(node, WORKINGSET_NODES); node 454 mm/workingset.c struct xa_node *node = container_of(item, struct xa_node, private_list); node 455 mm/workingset.c XA_STATE(xas, node->array, 0); node 471 mm/workingset.c mapping = container_of(node->array, struct address_space, i_pages); node 481 mm/workingset.c __dec_lruvec_slab_state(node, WORKINGSET_NODES); node 490 mm/workingset.c if (WARN_ON_ONCE(!node->nr_values)) node 492 mm/workingset.c if (WARN_ON_ONCE(node->count != node->nr_values)) node 494 mm/workingset.c mapping->nrexceptional -= node->nr_values; node 495 mm/workingset.c xas.xa_node = xa_parent_locked(&mapping->i_pages, node); node 496 mm/workingset.c xas.xa_offset = node->offset; node 497 mm/workingset.c xas.xa_shift = node->shift + XA_CHUNK_SHIFT; node 504 mm/workingset.c __inc_lruvec_slab_state(node, WORKINGSET_NODERECLAIM); node 127 mm/zswap.c struct hlist_node node; node 269 mm/zswap.c struct rb_node *node = root->rb_node; node 272 mm/zswap.c while (node) { node 273 mm/zswap.c entry = rb_entry(node, struct zswap_entry, rbnode); node 275 mm/zswap.c node = node->rb_left; node 277 mm/zswap.c node = node->rb_right; node 398 mm/zswap.c static int zswap_cpu_comp_prepare(unsigned int cpu, struct hlist_node *node) node 400 mm/zswap.c struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node); node 416 mm/zswap.c static int zswap_cpu_comp_dead(unsigned int cpu, struct hlist_node *node) node 418 mm/zswap.c struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node); node 544 mm/zswap.c &pool->node); node 611 mm/zswap.c cpuhp_state_remove_instance(CPUHP_MM_ZSWP_POOL_PREPARE, &pool->node); node 26 net/6lowpan/nhc.c node); node 50 net/6lowpan/nhc.c rb_link_node(&nhc->node, parent, new); node 51 net/6lowpan/nhc.c rb_insert_color(&nhc->node, &rb_root); node 58 net/6lowpan/nhc.c rb_erase(&nhc->node, &rb_root); node 63 net/6lowpan/nhc.c struct rb_node *node = rb_root.rb_node; node 66 net/6lowpan/nhc.c while (node) { node 67 net/6lowpan/nhc.c struct lowpan_nhc *nhc = rb_entry(node, struct lowpan_nhc, node 68 net/6lowpan/nhc.c node); node 82 net/6lowpan/nhc.c node = node->rb_left; node 84 net/6lowpan/nhc.c node = node->rb_right; node 68 net/6lowpan/nhc.h struct rb_node node; node 155 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); node 176 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); node 195 net/802/garp.c rb_link_node(&attr->node, parent, p); node 196 net/802/garp.c rb_insert_color(&attr->node, &app->gid); node 202 net/802/garp.c rb_erase(&attr->node, &app->gid); node 382 net/802/garp.c struct rb_node *node, *next; node 385 net/802/garp.c for (node = rb_first(&app->gid); node 386 net/802/garp.c next = node ? rb_next(node) : NULL, node != NULL; node 387 net/802/garp.c node = next) { node 388 net/802/garp.c attr = rb_entry(node, struct garp_attr, node); node 244 net/802/mrp.c attr = rb_entry(parent, struct mrp_attr, node); node 265 net/802/mrp.c attr = rb_entry(parent, struct mrp_attr, node); node 284 net/802/mrp.c rb_link_node(&attr->node, parent, p); node 285 net/802/mrp.c rb_insert_color(&attr->node, &app->mad); node 291 net/802/mrp.c rb_erase(&attr->node, &app->mad); node 567 net/802/mrp.c struct rb_node *node, *next; node 570 net/802/mrp.c for (node = rb_first(&app->mad); node 571 net/802/mrp.c next = node ? rb_next(node) : NULL, node != NULL; node 572 net/802/mrp.c node = next) { node 573 net/802/mrp.c attr = rb_entry(node, struct mrp_attr, node); node 33 net/802/psnap.c list_for_each_entry_rcu(p, &snap_list, node) { node 141 net/802/psnap.c list_add_rcu(&proto->node, &snap_list); node 155 net/802/psnap.c list_del_rcu(&proto->node); node 375 net/appletalk/ddp.c static struct atalk_iface *atalk_find_anynet(int node, struct net_device *dev) node 382 net/appletalk/ddp.c if (node != ATADDR_BCAST && node 383 net/appletalk/ddp.c iface->address.s_node != node && node 384 net/appletalk/ddp.c node != ATADDR_ANYNODE) node 394 net/appletalk/ddp.c static struct atalk_iface *atalk_find_interface(__be16 net, int node) node 400 net/appletalk/ddp.c if ((node == ATADDR_BCAST || node 401 net/appletalk/ddp.c node == ATADDR_ANYNODE || node 402 net/appletalk/ddp.c iface->address.s_node == node) && node 408 net/appletalk/ddp.c if (node == ATADDR_ANYNODE && net != ATADDR_ANYNET && node 825 net/atm/lec.c struct hlist_node *node; node 835 net/atm/lec.c struct hlist_node *e = state->node; node 848 net/atm/lec.c state->node = e; node 942 net/atm/lec.c state->node = SEQ_START_TOKEN; node 979 net/atm/lec.c struct lec_arp_table *entry = hlist_entry(state->node, node 102 net/batman-adv/bridge_loop_avoidance.c static bool batadv_compare_backbone_gw(const struct hlist_node *node, node 105 net/batman-adv/bridge_loop_avoidance.c const void *data1 = container_of(node, struct batadv_bla_backbone_gw, node 126 net/batman-adv/bridge_loop_avoidance.c static bool batadv_compare_claim(const struct hlist_node *node, node 129 net/batman-adv/bridge_loop_avoidance.c const void *data1 = container_of(node, struct batadv_bla_claim, node 215 net/batman-adv/distributed-arp-table.c static bool batadv_compare_dat(const struct hlist_node *node, const void *data2) node 217 net/batman-adv/distributed-arp-table.c const void *data1 = container_of(node, struct batadv_dat_entry, node 42 net/batman-adv/fragmentation.c struct hlist_node *node; node 44 net/batman-adv/fragmentation.c hlist_for_each_entry_safe(entry, node, head, list) { node 84 net/batman-adv/hash.h struct hlist_node *node; node 96 net/batman-adv/hash.h hlist_for_each(node, head) { node 97 net/batman-adv/hash.h if (!compare(node, data)) node 135 net/batman-adv/hash.h struct hlist_node *node; node 143 net/batman-adv/hash.h hlist_for_each(node, head) { node 144 net/batman-adv/hash.h if (!compare(node, data)) node 147 net/batman-adv/hash.h data_save = node; node 148 net/batman-adv/hash.h hlist_del_rcu(node); node 1753 net/batman-adv/multicast.c struct hlist_node *node = &orig->mcast_want_all_unsnoopables_node; node 1765 net/batman-adv/multicast.c WARN_ON(!hlist_unhashed(node)); node 1767 net/batman-adv/multicast.c hlist_add_head_rcu(node, head); node 1776 net/batman-adv/multicast.c WARN_ON(hlist_unhashed(node)); node 1778 net/batman-adv/multicast.c hlist_del_init_rcu(node); node 1798 net/batman-adv/multicast.c struct hlist_node *node = &orig->mcast_want_all_ipv4_node; node 1810 net/batman-adv/multicast.c WARN_ON(!hlist_unhashed(node)); node 1812 net/batman-adv/multicast.c hlist_add_head_rcu(node, head); node 1821 net/batman-adv/multicast.c WARN_ON(hlist_unhashed(node)); node 1823 net/batman-adv/multicast.c hlist_del_init_rcu(node); node 1843 net/batman-adv/multicast.c struct hlist_node *node = &orig->mcast_want_all_ipv6_node; node 1855 net/batman-adv/multicast.c WARN_ON(!hlist_unhashed(node)); node 1857 net/batman-adv/multicast.c hlist_add_head_rcu(node, head); node 1866 net/batman-adv/multicast.c WARN_ON(hlist_unhashed(node)); node 1868 net/batman-adv/multicast.c hlist_del_init_rcu(node); node 1888 net/batman-adv/multicast.c struct hlist_node *node = &orig->mcast_want_all_rtr4_node; node 1900 net/batman-adv/multicast.c WARN_ON(!hlist_unhashed(node)); node 1902 net/batman-adv/multicast.c hlist_add_head_rcu(node, head); node 1911 net/batman-adv/multicast.c WARN_ON(hlist_unhashed(node)); node 1913 net/batman-adv/multicast.c hlist_del_init_rcu(node); node 1933 net/batman-adv/multicast.c struct hlist_node *node = &orig->mcast_want_all_rtr6_node; node 1945 net/batman-adv/multicast.c WARN_ON(!hlist_unhashed(node)); node 1947 net/batman-adv/multicast.c hlist_add_head_rcu(node, head); node 1956 net/batman-adv/multicast.c WARN_ON(hlist_unhashed(node)); node 1958 net/batman-adv/multicast.c hlist_del_init_rcu(node); node 513 net/batman-adv/network-coding.c static bool batadv_nc_hash_compare(const struct hlist_node *node, node 518 net/batman-adv/network-coding.c nc_path1 = container_of(node, struct batadv_nc_path, hash_entry); node 96 net/batman-adv/originator.c bool batadv_compare_orig(const struct hlist_node *node, const void *data2) node 98 net/batman-adv/originator.c const void *data1 = container_of(node, struct batadv_orig_node, node 20 net/batman-adv/originator.h bool batadv_compare_orig(const struct hlist_node *node, const void *data2); node 87 net/batman-adv/translation-table.c static bool batadv_compare_tt(const struct hlist_node *node, const void *data2) node 89 net/batman-adv/translation-table.c const void *data1 = container_of(node, struct batadv_tt_common_entry, node 2824 net/batman-adv/translation-table.c struct batadv_tt_req_node *node; node 2829 net/batman-adv/translation-table.c hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) { node 2830 net/batman-adv/translation-table.c hlist_del_init(&node->list); node 2831 net/batman-adv/translation-table.c batadv_tt_req_node_put(node); node 2860 net/batman-adv/translation-table.c struct batadv_tt_req_node *node; node 2864 net/batman-adv/translation-table.c hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) { node 2865 net/batman-adv/translation-table.c if (batadv_has_timed_out(node->issued_at, node 2867 net/batman-adv/translation-table.c hlist_del_init(&node->list); node 2868 net/batman-adv/translation-table.c batadv_tt_req_node_put(node); node 3607 net/batman-adv/translation-table.c struct batadv_tt_req_node *node; node 3646 net/batman-adv/translation-table.c hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) { node 3647 net/batman-adv/translation-table.c if (!batadv_compare_eth(node->addr, resp_src)) node 3649 net/batman-adv/translation-table.c hlist_del_init(&node->list); node 3650 net/batman-adv/translation-table.c batadv_tt_req_node_put(node); node 3661 net/batman-adv/translation-table.c struct batadv_tt_roam_node *node, *safe; node 3665 net/batman-adv/translation-table.c list_for_each_entry_safe(node, safe, &bat_priv->tt.roam_list, list) { node 3666 net/batman-adv/translation-table.c list_del(&node->list); node 3667 net/batman-adv/translation-table.c kmem_cache_free(batadv_tt_roam_cache, node); node 3675 net/batman-adv/translation-table.c struct batadv_tt_roam_node *node, *safe; node 3678 net/batman-adv/translation-table.c list_for_each_entry_safe(node, safe, &bat_priv->tt.roam_list, list) { node 3679 net/batman-adv/translation-table.c if (!batadv_has_timed_out(node->first_time, node 3683 net/batman-adv/translation-table.c list_del(&node->list); node 3684 net/batman-adv/translation-table.c kmem_cache_free(batadv_tt_roam_cache, node); node 30 net/caif/cfcnfg.c struct list_head node; node 138 net/caif/cfcnfg.c list_for_each_entry_rcu(phy, &cnfg->phys, node) node 154 net/caif/cfcnfg.c list_for_each_entry_rcu(phy, &cnfg->phys, node) { node 162 net/caif/cfcnfg.c list_for_each_entry_rcu(phy, &cnfg->phys, node) node 173 net/caif/cfcnfg.c list_for_each_entry_rcu(phy, &cnfg->phys, node) node 511 net/caif/cfcnfg.c list_add_rcu(&phyinfo->node, &cnfg->phys); node 576 net/caif/cfcnfg.c list_del_rcu(&phyinfo->node); node 582 net/caif/cfcnfg.c list_add_rcu(&phyinfo->node, &cnfg->phys); node 70 net/caif/cfmuxl.c list_add_rcu(&dn->node, &muxl->frml_list); node 78 net/caif/cfmuxl.c list_for_each_entry_rcu(lyr, list, node) { node 96 net/caif/cfmuxl.c list_del_rcu(&old->node); node 98 net/caif/cfmuxl.c list_add_rcu(&up->node, &muxl->srvl_list); node 116 net/caif/cfmuxl.c list_del_rcu(&dn->node); node 168 net/caif/cfmuxl.c list_del_rcu(&up->node); node 253 net/caif/cfmuxl.c list_for_each_entry_rcu(layer, &muxl->srvl_list, node) { node 125 net/ceph/auth_x.c th = rb_entry(parent, struct ceph_x_ticket_handler, node); node 139 net/ceph/auth_x.c rb_link_node(&th->node, parent, p); node 140 net/ceph/auth_x.c rb_insert_color(&th->node, &xi->ticket_handlers); node 150 net/ceph/auth_x.c rb_erase(&th->node, &xi->ticket_handlers); node 773 net/ceph/auth_x.c rb_entry(p, struct ceph_x_ticket_handler, node); node 16 net/ceph/auth_x.h struct rb_node node; node 71 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_pool_info, node); node 92 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); node 103 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); node 110 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); node 121 net/ceph/debugfs.c rb_entry(n, struct ceph_pg_mapping, node); node 160 net/ceph/debugfs.c req = rb_entry(rp, struct ceph_mon_generic_request, node); node 262 net/ceph/debugfs.c rb_entry(n, struct ceph_osd_linger_request, node); node 506 net/ceph/mon_client.c DEFINE_RB_FUNCS(generic_request, struct ceph_mon_generic_request, tid, node) node 515 net/ceph/mon_client.c WARN_ON(!RB_EMPTY_NODE(&req->node)); node 547 net/ceph/mon_client.c RB_CLEAR_NODE(&req->node); node 961 net/ceph/mon_client.c req = rb_entry(p, struct ceph_mon_generic_request, node); node 1328 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); node 1461 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); node 1622 net/ceph/osd_client.c RB_CLEAR_NODE(&spg->node); node 1629 net/ceph/osd_client.c WARN_ON(!RB_EMPTY_NODE(&spg->node)); node 1642 net/ceph/osd_client.c RB_BYPTR, const struct ceph_spg *, node) node 1874 net/ceph/osd_client.c struct ceph_spg_mapping, node); node 2660 net/ceph/osd_client.c WARN_ON(!RB_EMPTY_NODE(&lreq->node)); node 2699 net/ceph/osd_client.c RB_CLEAR_NODE(&lreq->node); node 2714 net/ceph/osd_client.c DEFINE_RB_INSDEL_FUNCS(linger, struct ceph_osd_linger_request, linger_id, node) node 3312 net/ceph/osd_client.c rb_entry(p, struct ceph_osd_linger_request, node); node 3710 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); node 3763 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); node 3857 net/ceph/osd_client.c rb_entry(n, struct ceph_pg_pool_info, node); node 4089 net/ceph/osd_client.c rb_entry(n, struct ceph_osd_linger_request, node); node 149 net/ceph/osdmap.c RB_CLEAR_NODE(&arg_map->node); node 158 net/ceph/osdmap.c WARN_ON(!RB_EMPTY_NODE(&arg_map->node)); node 174 net/ceph/osdmap.c node); node 181 net/ceph/osdmap.c struct crush_choose_arg_map, node); node 618 net/ceph/osdmap.c RB_CLEAR_NODE(&pg->node); node 624 net/ceph/osdmap.c WARN_ON(!RB_EMPTY_NODE(&pg->node)); node 634 net/ceph/osdmap.c RB_BYPTR, const struct ceph_pg *, node) node 647 net/ceph/osdmap.c pi = rb_entry(parent, struct ceph_pg_pool_info, node); node 656 net/ceph/osdmap.c rb_link_node(&new->node, parent, p); node 657 net/ceph/osdmap.c rb_insert_color(&new->node, root); node 667 net/ceph/osdmap.c pi = rb_entry(n, struct ceph_pg_pool_info, node); node 705 net/ceph/osdmap.c rb_entry(rbp, struct ceph_pg_pool_info, node); node 724 net/ceph/osdmap.c rb_erase(&pi->node, root); node 954 net/ceph/osdmap.c struct ceph_pg_mapping, node); node 961 net/ceph/osdmap.c struct ceph_pg_mapping, node); node 968 net/ceph/osdmap.c struct ceph_pg_mapping, node); node 969 net/ceph/osdmap.c rb_erase(&pg->node, &map->pg_upmap); node 975 net/ceph/osdmap.c struct ceph_pg_mapping, node); node 976 net/ceph/osdmap.c rb_erase(&pg->node, &map->pg_upmap_items); node 982 net/ceph/osdmap.c struct ceph_pg_pool_info, node); node 21 net/ceph/string_table.c exist = rb_entry(*p, struct ceph_string, node); node 32 net/ceph/string_table.c rb_erase(&exist->node, &string_tree); node 33 net/ceph/string_table.c RB_CLEAR_NODE(&exist->node); node 56 net/ceph/string_table.c exist = rb_entry(*p, struct ceph_string, node); node 68 net/ceph/string_table.c rb_link_node(&cs->node, parent, p); node 69 net/ceph/string_table.c rb_insert_color(&cs->node, &string_tree); node 71 net/ceph/string_table.c rb_erase(&exist->node, &string_tree); node 72 net/ceph/string_table.c RB_CLEAR_NODE(&exist->node); node 93 net/ceph/string_table.c if (!RB_EMPTY_NODE(&cs->node)) { node 94 net/ceph/string_table.c rb_erase(&cs->node, &string_tree); node 95 net/ceph/string_table.c RB_CLEAR_NODE(&cs->node); node 409 net/core/pktgen.c int node; /* Memory node */ node 637 net/core/pktgen.c if (pkt_dev->node >= 0) node 638 net/core/pktgen.c seq_printf(seq, " node: %d\n", pkt_dev->node); node 1160 net/core/pktgen.c pkt_dev->node = value; node 1161 net/core/pktgen.c sprintf(pg_result, "OK: node=%d", pkt_dev->node); node 2643 net/core/pktgen.c int node = numa_node_id(); node 2645 net/core/pktgen.c if (pkt_dev->node >= 0 && (pkt_dev->flags & F_NODE)) node 2646 net/core/pktgen.c node = pkt_dev->node; node 2647 net/core/pktgen.c pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); node 2701 net/core/pktgen.c int node = pkt_dev->node >= 0 ? pkt_dev->node : numa_node_id(); node 2703 net/core/pktgen.c skb = __alloc_skb(NET_SKB_PAD + size, GFP_NOWAIT, 0, node); node 3594 net/core/pktgen.c int node = cpu_to_node(t->cpu); node 3604 net/core/pktgen.c pkt_dev = kzalloc_node(sizeof(struct pktgen_dev), GFP_KERNEL, node); node 3611 net/core/pktgen.c node); node 3633 net/core/pktgen.c pkt_dev->node = NUMA_NO_NODE; node 128 net/core/skbuff.c #define kmalloc_reserve(size, gfp, node, pfmemalloc) \ node 129 net/core/skbuff.c __kmalloc_reserve(size, gfp, node, _RET_IP_, pfmemalloc) node 131 net/core/skbuff.c static void *__kmalloc_reserve(size_t size, gfp_t flags, int node, node 143 net/core/skbuff.c node); node 149 net/core/skbuff.c obj = kmalloc_node_track_caller(size, flags, node); node 182 net/core/skbuff.c int flags, int node) node 197 net/core/skbuff.c skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node); node 209 net/core/skbuff.c data = kmalloc_reserve(size, gfp_mask, node, &pfmemalloc); node 494 net/core/skmsg.c struct sk_psock *sk_psock_init(struct sock *sk, int node) node 498 net/core/skmsg.c node); node 3418 net/core/sock.c list_add(&prot->node, &proto_list); node 3441 net/core/sock.c list_del(&prot->node); node 3561 net/core/sock.c proto_seq_printf(seq, list_entry(v, struct proto, node)); node 522 net/core/sock_map.c struct hlist_node node; node 557 net/core/sock_map.c hlist_for_each_entry_rcu(elem, head, node) { node 607 net/core/sock_map.c hlist_del_rcu(&elem->node); node 628 net/core/sock_map.c hlist_del_rcu(&elem->node); node 715 net/core/sock_map.c hlist_add_head_rcu(&elem_new->node, &bucket->head); node 717 net/core/sock_map.c hlist_del_rcu(&elem->node); node 780 net/core/sock_map.c elem_next = hlist_entry_safe(rcu_dereference_raw(hlist_next_rcu(&elem->node)), node 781 net/core/sock_map.c struct bpf_htab_elem, node); node 793 net/core/sock_map.c struct bpf_htab_elem, node); node 865 net/core/sock_map.c struct hlist_node *node; node 875 net/core/sock_map.c hlist_for_each_entry_safe(elem, node, &bucket->head, node) { node 876 net/core/sock_map.c hlist_del_rcu(&elem->node); node 57 net/core/xdp.c .head_offset = offsetof(struct xdp_mem_allocator, node), node 88 net/core/xdp.c if (!rhashtable_remove_fast(mem_id_ht, &xa->node, mem_id_rht_params)) node 131 net/core/xdp.c if (!rhashtable_remove_fast(mem_id_ht, &xa->node, mem_id_rht_params)) node 339 net/core/xdp.c ptr = rhashtable_insert_slow(mem_id_ht, &id, &xdp_alloc->node); node 357 net/dccp/ackvec.c list_add_tail(&new->node, head); node 366 net/dccp/ackvec.c list_for_each_entry_safe(cur, next, parsed_chunks, node) node 131 net/dccp/ackvec.h struct list_head node; node 584 net/dccp/ccids/ccid2.c list_for_each_entry(avp, &hc->tx_av_chunks, node) { node 295 net/dccp/feat.c list_for_each_entry(___entry, fn_list, node) \ node 431 net/dccp/feat.c list_for_each_entry(entry, fn_list, node) { node 453 net/dccp/feat.c list_for_each_entry(entry, head, node) node 458 net/dccp/feat.c head = &entry->node; node 466 net/dccp/feat.c list_add_tail(&entry->node, head); node 535 net/dccp/feat.c list_del(&entry->node); node 543 net/dccp/feat.c list_for_each_entry_safe(entry, next, fn_list, node) node 555 net/dccp/feat.c list_for_each_entry(entry, from, node) { node 559 net/dccp/feat.c list_add_tail(&new->node, to); node 638 net/dccp/feat.c list_for_each_entry_safe_reverse(pos, next, fn, node) { node 983 net/dccp/feat.c list_for_each_entry(entry, fn, node) node 1506 net/dccp/feat.c list_for_each_entry(cur, fn_list, node) { node 1544 net/dccp/feat.c list_for_each_entry_safe(cur, next, fn_list, node) node 76 net/dccp/feat.h struct list_head node; node 657 net/decnet/af_decnet.c unsigned short node, area; node 659 net/decnet/af_decnet.c node = addr & 0x03ff; node 661 net/decnet/af_decnet.c sprintf(buf, "%hd.%hd", area, node); node 92 net/decnet/sysctl_net_decnet.c __u16 area, node; node 111 net/decnet/sysctl_net_decnet.c node = *str++ - '0'; node 113 net/decnet/sysctl_net_decnet.c node *= 10; node 114 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 117 net/decnet/sysctl_net_decnet.c node *= 10; node 118 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 121 net/decnet/sysctl_net_decnet.c node *= 10; node 122 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 125 net/decnet/sysctl_net_decnet.c if ((node > 1023) || (area > 63)) node 131 net/decnet/sysctl_net_decnet.c *addr = cpu_to_le16((area << 10) | node); node 144 net/dsa/dsa2.c link_dp = dsa_tree_find_port_by_node(dst, it.node); node 146 net/dsa/dsa2.c of_node_put(it.node); node 36 net/hsr/hsr_debugfs.c struct hsr_node *node; node 42 net/hsr/hsr_debugfs.c list_for_each_entry_rcu(node, &priv->node_db, mac_list) { node 44 net/hsr/hsr_debugfs.c if (hsr_addr_is_self(priv, node->macaddress_A)) node 46 net/hsr/hsr_debugfs.c print_mac_address(sfp, &node->macaddress_A[0]); node 48 net/hsr/hsr_debugfs.c print_mac_address(sfp, &node->macaddress_B[0]); node 49 net/hsr/hsr_debugfs.c seq_printf(sfp, "0x%lx, ", node->time_in[HSR_PT_SLAVE_A]); node 50 net/hsr/hsr_debugfs.c seq_printf(sfp, "0x%lx ", node->time_in[HSR_PT_SLAVE_B]); node 51 net/hsr/hsr_debugfs.c seq_printf(sfp, "0x%x\n", node->addr_B_port); node 43 net/hsr/hsr_framereg.c struct hsr_node *node; node 45 net/hsr/hsr_framereg.c node = list_first_or_null_rcu(&hsr->self_node_db, struct hsr_node, node 47 net/hsr/hsr_framereg.c if (!node) { node 52 net/hsr/hsr_framereg.c if (ether_addr_equal(addr, node->macaddress_A)) node 54 net/hsr/hsr_framereg.c if (ether_addr_equal(addr, node->macaddress_B)) node 65 net/hsr/hsr_framereg.c struct hsr_node *node; node 67 net/hsr/hsr_framereg.c list_for_each_entry_rcu(node, node_db, mac_list) { node 68 net/hsr/hsr_framereg.c if (ether_addr_equal(node->macaddress_A, addr)) node 69 net/hsr/hsr_framereg.c return node; node 83 net/hsr/hsr_framereg.c struct hsr_node *node, *oldnode; node 85 net/hsr/hsr_framereg.c node = kmalloc(sizeof(*node), GFP_KERNEL); node 86 net/hsr/hsr_framereg.c if (!node) node 89 net/hsr/hsr_framereg.c ether_addr_copy(node->macaddress_A, addr_a); node 90 net/hsr/hsr_framereg.c ether_addr_copy(node->macaddress_B, addr_b); node 96 net/hsr/hsr_framereg.c list_replace_rcu(&oldnode->mac_list, &node->mac_list); node 100 net/hsr/hsr_framereg.c list_add_tail_rcu(&node->mac_list, self_node_db); node 110 net/hsr/hsr_framereg.c struct hsr_node *node; node 113 net/hsr/hsr_framereg.c node = list_first_or_null_rcu(self_node_db, struct hsr_node, mac_list); node 114 net/hsr/hsr_framereg.c if (node) { node 115 net/hsr/hsr_framereg.c list_del_rcu(&node->mac_list); node 116 net/hsr/hsr_framereg.c kfree_rcu(node, rcu_head); node 123 net/hsr/hsr_framereg.c struct hsr_node *node; node 126 net/hsr/hsr_framereg.c list_for_each_entry_safe(node, tmp, node_db, mac_list) node 127 net/hsr/hsr_framereg.c kfree(node); node 139 net/hsr/hsr_framereg.c struct hsr_node *new_node, *node; node 159 net/hsr/hsr_framereg.c list_for_each_entry_rcu(node, node_db, mac_list) { node 160 net/hsr/hsr_framereg.c if (ether_addr_equal(node->macaddress_A, addr)) node 162 net/hsr/hsr_framereg.c if (ether_addr_equal(node->macaddress_B, addr)) node 171 net/hsr/hsr_framereg.c return node; node 181 net/hsr/hsr_framereg.c struct hsr_node *node; node 190 net/hsr/hsr_framereg.c list_for_each_entry_rcu(node, node_db, mac_list) { node 191 net/hsr/hsr_framereg.c if (ether_addr_equal(node->macaddress_A, ethhdr->h_source)) node 192 net/hsr/hsr_framereg.c return node; node 193 net/hsr/hsr_framereg.c if (ether_addr_equal(node->macaddress_B, ethhdr->h_source)) node 194 net/hsr/hsr_framereg.c return node; node 286 net/hsr/hsr_framereg.c void hsr_addr_subst_source(struct hsr_node *node, struct sk_buff *skb) node 293 net/hsr/hsr_framereg.c memcpy(ð_hdr(skb)->h_source, node->macaddress_A, ETH_ALEN); node 330 net/hsr/hsr_framereg.c void hsr_register_frame_in(struct hsr_node *node, struct hsr_port *port, node 337 net/hsr/hsr_framereg.c if (seq_nr_before(sequence_nr, node->seq_out[port->type])) node 340 net/hsr/hsr_framereg.c node->time_in[port->type] = jiffies; node 341 net/hsr/hsr_framereg.c node->time_in_stale[port->type] = false; node 352 net/hsr/hsr_framereg.c int hsr_register_frame_out(struct hsr_port *port, struct hsr_node *node, node 355 net/hsr/hsr_framereg.c if (seq_nr_before_or_eq(sequence_nr, node->seq_out[port->type])) node 358 net/hsr/hsr_framereg.c node->seq_out[port->type] = sequence_nr; node 363 net/hsr/hsr_framereg.c struct hsr_node *node) node 365 net/hsr/hsr_framereg.c if (node->time_in_stale[HSR_PT_SLAVE_A]) node 367 net/hsr/hsr_framereg.c if (node->time_in_stale[HSR_PT_SLAVE_B]) node 370 net/hsr/hsr_framereg.c if (time_after(node->time_in[HSR_PT_SLAVE_B], node 371 net/hsr/hsr_framereg.c node->time_in[HSR_PT_SLAVE_A] + node 374 net/hsr/hsr_framereg.c if (time_after(node->time_in[HSR_PT_SLAVE_A], node 375 net/hsr/hsr_framereg.c node->time_in[HSR_PT_SLAVE_B] + node 388 net/hsr/hsr_framereg.c struct hsr_node *node; node 395 net/hsr/hsr_framereg.c list_for_each_entry_safe(node, tmp, &hsr->node_db, mac_list) { node 401 net/hsr/hsr_framereg.c if (hsr_addr_is_self(hsr, node->macaddress_A)) node 405 net/hsr/hsr_framereg.c time_a = node->time_in[HSR_PT_SLAVE_A]; node 406 net/hsr/hsr_framereg.c time_b = node->time_in[HSR_PT_SLAVE_B]; node 410 net/hsr/hsr_framereg.c node->time_in_stale[HSR_PT_SLAVE_A] = true; node 412 net/hsr/hsr_framereg.c node->time_in_stale[HSR_PT_SLAVE_B] = true; node 419 net/hsr/hsr_framereg.c if (node->time_in_stale[HSR_PT_SLAVE_A] || node 420 net/hsr/hsr_framereg.c (!node->time_in_stale[HSR_PT_SLAVE_B] && node 428 net/hsr/hsr_framereg.c port = get_late_port(hsr, node); node 430 net/hsr/hsr_framereg.c hsr_nl_ringerror(hsr, node->macaddress_A, port); node 437 net/hsr/hsr_framereg.c hsr_nl_nodedown(hsr, node->macaddress_A); node 438 net/hsr/hsr_framereg.c list_del_rcu(&node->mac_list); node 440 net/hsr/hsr_framereg.c kfree_rcu(node, rcu_head); node 453 net/hsr/hsr_framereg.c struct hsr_node *node; node 456 net/hsr/hsr_framereg.c node = list_first_or_null_rcu(&hsr->node_db, node 458 net/hsr/hsr_framereg.c if (node) node 459 net/hsr/hsr_framereg.c ether_addr_copy(addr, node->macaddress_A); node 460 net/hsr/hsr_framereg.c return node; node 463 net/hsr/hsr_framereg.c node = _pos; node 464 net/hsr/hsr_framereg.c list_for_each_entry_continue_rcu(node, &hsr->node_db, mac_list) { node 465 net/hsr/hsr_framereg.c ether_addr_copy(addr, node->macaddress_A); node 466 net/hsr/hsr_framereg.c return node; node 481 net/hsr/hsr_framereg.c struct hsr_node *node; node 485 net/hsr/hsr_framereg.c node = find_node_by_addr_A(&hsr->node_db, addr); node 486 net/hsr/hsr_framereg.c if (!node) node 489 net/hsr/hsr_framereg.c ether_addr_copy(addr_b, node->macaddress_B); node 491 net/hsr/hsr_framereg.c tdiff = jiffies - node->time_in[HSR_PT_SLAVE_A]; node 492 net/hsr/hsr_framereg.c if (node->time_in_stale[HSR_PT_SLAVE_A]) node 501 net/hsr/hsr_framereg.c tdiff = jiffies - node->time_in[HSR_PT_SLAVE_B]; node 502 net/hsr/hsr_framereg.c if (node->time_in_stale[HSR_PT_SLAVE_B]) node 512 net/hsr/hsr_framereg.c *if1_seq = node->seq_out[HSR_PT_SLAVE_B]; node 513 net/hsr/hsr_framereg.c *if2_seq = node->seq_out[HSR_PT_SLAVE_A]; node 515 net/hsr/hsr_framereg.c if (node->addr_B_port != HSR_PT_NONE) { node 516 net/hsr/hsr_framereg.c port = hsr_port_get_hsr(hsr, node->addr_B_port); node 23 net/hsr/hsr_framereg.h void hsr_addr_subst_source(struct hsr_node *node, struct sk_buff *skb); node 27 net/hsr/hsr_framereg.h void hsr_register_frame_in(struct hsr_node *node, struct hsr_port *port, node 29 net/hsr/hsr_framereg.h int hsr_register_frame_out(struct hsr_port *port, struct hsr_node *node, node 508 net/ieee802154/6lowpan/reassembly.c .head_offset = offsetof(struct inet_frag_queue, node), node 913 net/ipv4/inet_diag.c struct hlist_nulls_node *node; node 918 net/ipv4/inet_diag.c sk_nulls_for_each(sk, node, &ilb->nulls_head) { node 962 net/ipv4/inet_diag.c struct hlist_nulls_node *node; node 977 net/ipv4/inet_diag.c sk_nulls_for_each(sk, node, &head->chain) { node 210 net/ipv4/inet_fragment.c rhashtable_remove_fast(&fqdir->rhashtable, &fq->node, node 311 net/ipv4/inet_fragment.c &q->node, f->rhash_params); node 76 net/ipv4/inet_hashtables.c hlist_add_head(&tb->node, &head->chain); node 87 net/ipv4/inet_hashtables.c __hlist_del(&tb->node); node 355 net/ipv4/inet_hashtables.c const struct hlist_nulls_node *node; node 364 net/ipv4/inet_hashtables.c sk_nulls_for_each_rcu(sk, node, &head->chain) { node 385 net/ipv4/inet_hashtables.c if (get_nulls_value(node) != slot) node 413 net/ipv4/inet_hashtables.c const struct hlist_nulls_node *node; node 418 net/ipv4/inet_hashtables.c sk_nulls_for_each(sk2, node, &head->chain) { node 519 net/ipv4/inet_hashtables.c const struct hlist_nulls_node *node; node 523 net/ipv4/inet_hashtables.c sk_nulls_for_each_rcu(sk2, node, &ilb->nulls_head) { node 261 net/ipv4/inet_timewait_sock.c struct hlist_nulls_node *node; node 270 net/ipv4/inet_timewait_sock.c sk_nulls_for_each_rcu(sk, node, &head->chain) { node 297 net/ipv4/inet_timewait_sock.c if (get_nulls_value(node) != slot) node 730 net/ipv4/ip_fragment.c .head_offset = offsetof(struct inet_frag_queue, node), node 1088 net/ipv4/nexthop.c struct rb_node *node; node 1091 net/ipv4/nexthop.c while ((node = rb_first(root))) { node 1092 net/ipv4/nexthop.c nh = rb_entry(node, struct nexthop, rb_node); node 1721 net/ipv4/nexthop.c struct rb_node *node; node 1731 net/ipv4/nexthop.c for (node = rb_first(root); node; node = rb_next(node)) { node 1737 net/ipv4/nexthop.c nh = rb_entry(node, struct nexthop, rb_node); node 79 net/ipv4/ping.c struct hlist_nulls_node *node; node 95 net/ipv4/ping.c ping_portaddr_for_each_entry(sk2, node, hlist) { node 112 net/ipv4/ping.c ping_portaddr_for_each_entry(sk2, node, hlist) { node 1019 net/ipv4/ping.c struct hlist_nulls_node *node; node 1027 net/ipv4/ping.c sk_nulls_for_each(sk, node, hslot) { node 63 net/ipv4/tcp_diag.c hlist_for_each_entry_rcu(key, &md5sig->head, node) node 75 net/ipv4/tcp_diag.c hlist_for_each_entry_rcu(key, &md5sig->head, node) { node 158 net/ipv4/tcp_diag.c hlist_for_each_entry_rcu(key, &md5sig->head, node) node 5077 net/ipv4/tcp_input.c struct rb_node *node, *prev; node 5085 net/ipv4/tcp_input.c node = &tp->ooo_last_skb->rbnode; node 5087 net/ipv4/tcp_input.c prev = rb_prev(node); node 5088 net/ipv4/tcp_input.c rb_erase(node, &tp->out_of_order_queue); node 5089 net/ipv4/tcp_input.c goal -= rb_to_skb(node)->truesize; node 5090 net/ipv4/tcp_input.c tcp_drop(sk, rb_to_skb(node)); node 5098 net/ipv4/tcp_input.c node = prev; node 5099 net/ipv4/tcp_input.c } while (node); node 1005 net/ipv4/tcp_ipv4.c hlist_for_each_entry_rcu(key, &md5sig->head, node) { node 1048 net/ipv4/tcp_ipv4.c hlist_for_each_entry_rcu(key, &md5sig->head, node) { node 1113 net/ipv4/tcp_ipv4.c hlist_add_head_rcu(&key->node, &md5sig->head); node 1126 net/ipv4/tcp_ipv4.c hlist_del_rcu(&key->node); node 1142 net/ipv4/tcp_ipv4.c hlist_for_each_entry_safe(key, n, &md5sig->head, node) { node 1143 net/ipv4/tcp_ipv4.c hlist_del_rcu(&key->node); node 2152 net/ipv4/tcp_ipv4.c struct hlist_nulls_node *node; node 2169 net/ipv4/tcp_ipv4.c sk_nulls_for_each_from(sk, node) { node 2217 net/ipv4/tcp_ipv4.c struct hlist_nulls_node *node; node 2225 net/ipv4/tcp_ipv4.c sk_nulls_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) { node 2243 net/ipv4/tcp_ipv4.c struct hlist_nulls_node *node; node 2252 net/ipv4/tcp_ipv4.c sk_nulls_for_each_from(sk, node) { node 2156 net/ipv4/udp.c struct hlist_node *node; node 2168 net/ipv4/udp.c sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) { node 20 net/ipv6/ila/ila_xlat.c struct rhash_head node; node 87 net/ipv6/ila/ila_xlat.c .head_offset = offsetof(struct ila_map, node), node 241 net/ipv6/ila/ila_xlat.c &ila->node, rht_params); node 267 net/ipv6/ila/ila_xlat.c &head->node, node 268 net/ipv6/ila/ila_xlat.c &ila->node, rht_params); node 322 net/ipv6/ila/ila_xlat.c &ilan->xlat.rhash_table, &ila->node, node 323 net/ipv6/ila/ila_xlat.c &head->node, rht_params); node 330 net/ipv6/ila/ila_xlat.c &ila->node, rht_params); node 408 net/ipv6/ila/ila_xlat.c &ila->node, rht_params); node 58 net/ipv6/inet6_hashtables.c const struct hlist_nulls_node *node; node 69 net/ipv6/inet6_hashtables.c sk_nulls_for_each_rcu(sk, node, &head->chain) { node 83 net/ipv6/inet6_hashtables.c if (get_nulls_value(node) != slot) node 217 net/ipv6/inet6_hashtables.c const struct hlist_nulls_node *node; node 222 net/ipv6/inet6_hashtables.c sk_nulls_for_each(sk2, node, &head->chain) { node 545 net/ipv6/ip6_fib.c w->node = w->root; node 1796 net/ipv6/ip6_fib.c if (w->node == fn) { node 1798 net/ipv6/ip6_fib.c w->node = pn; node 1802 net/ipv6/ip6_fib.c if (w->node == fn) { node 1803 net/ipv6/ip6_fib.c w->node = child; node 1961 net/ipv6/ip6_fib.c fn = w->node; node 1969 net/ipv6/ip6_fib.c w->node = FIB6_SUBTREE(fn); node 1978 net/ipv6/ip6_fib.c w->node = left; node 1987 net/ipv6/ip6_fib.c w->node = right; node 2019 net/ipv6/ip6_fib.c w->node = pn; node 2033 net/ipv6/ip6_fib.c w->leaf = rcu_dereference_protected(w->node->leaf, 1); node 2048 net/ipv6/ip6_fib.c w->node = w->root; node 2068 net/ipv6/ip6_fib.c w->node->fn_sernum != c->sernum) node 2069 net/ipv6/ip6_fib.c w->node->fn_sernum = c->sernum; node 2431 net/ipv6/ip6_fib.c iter->w.node = iter->w.root; node 2442 net/ipv6/ip6_fib.c struct hlist_node *node; node 2446 net/ipv6/ip6_fib.c node = rcu_dereference_bh(hlist_next_rcu(&tbl->tb6_hlist)); node 2449 net/ipv6/ip6_fib.c node = NULL; node 2452 net/ipv6/ip6_fib.c while (!node && h < FIB6_TABLE_HASHSZ) { node 2453 net/ipv6/ip6_fib.c node = rcu_dereference_bh( node 2456 net/ipv6/ip6_fib.c return hlist_entry_safe(node, struct fib6_table, tb6_hlist); node 2464 net/ipv6/ip6_fib.c iter->w.node = iter->w.root; node 2530 net/ipv6/ip6_fib.c return w->node && !(w->state == FWS_U && w->node == w->root); node 522 net/ipv6/netfilter/nf_conntrack_reasm.c .head_offset = offsetof(struct inet_frag_queue, node), node 541 net/ipv6/reassembly.c .head_offset = offsetof(struct inet_frag_queue, node), node 68 net/ipv6/seg6_hmac.c .head_offset = offsetof(struct seg6_hmac_info, node), node 298 net/ipv6/seg6_hmac.c err = rhashtable_lookup_insert_fast(&sdata->hmac_infos, &hinfo->node, node 315 net/ipv6/seg6_hmac.c err = rhashtable_remove_fast(&sdata->hmac_infos, &hinfo->node, node 755 net/ipv6/udp.c struct hlist_node *node; node 767 net/ipv6/udp.c sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) { node 66 net/lapb/lapb_iface.c if (lapb->node.next) { node 67 net/lapb/lapb_iface.c list_del(&lapb->node); node 77 net/lapb/lapb_iface.c list_add(&lapb->node, &lapb_list); node 87 net/lapb/lapb_iface.c lapb = list_entry(entry, struct lapb_cb, node); node 482 net/llc/llc_conn.c struct hlist_nulls_node *node; node 488 net/llc/llc_conn.c sk_nulls_for_each_rcu(rc, node, laddr_hb) { node 507 net/llc/llc_conn.c if (unlikely(get_nulls_value(node) != slot)) node 541 net/llc/llc_conn.c struct hlist_nulls_node *node; node 547 net/llc/llc_conn.c sk_nulls_for_each_rcu(rc, node, laddr_hb) { node 566 net/llc/llc_conn.c if (unlikely(get_nulls_value(node) != slot)) node 53 net/llc/llc_core.c list_for_each_entry(sap, &llc_sap_list, node) node 107 net/llc/llc_core.c list_add_tail_rcu(&sap->node, &llc_sap_list); node 127 net/llc/llc_core.c list_del_rcu(&sap->node); node 40 net/llc/llc_proc.c list_for_each_entry_rcu(sap, &llc_sap_list, node) { node 44 net/llc/llc_proc.c struct hlist_nulls_node *node; node 46 net/llc/llc_proc.c sk_nulls_for_each(sk, node, head) { node 69 net/llc/llc_proc.c struct hlist_nulls_node *node; node 73 net/llc/llc_proc.c sk_nulls_for_each(sk, node, &sap->sk_laddr_hash[bucket]) node 103 net/llc/llc_proc.c list_for_each_entry_continue_rcu(sap, &llc_sap_list, node) { node 316 net/llc/llc_sap.c struct hlist_nulls_node *node; node 322 net/llc/llc_sap.c sk_nulls_for_each_rcu(rc, node, laddr_hb) { node 341 net/llc/llc_sap.c if (unlikely(get_nulls_value(node) != slot)) node 214 net/ncsi/internal.h struct list_head node; node 225 net/ncsi/internal.h struct list_head node; /* Form list of packages */ node 309 net/ncsi/internal.h struct list_head node; /* Form NCSI device list */ node 340 net/ncsi/internal.h list_for_each_entry_rcu(ndp, &ncsi_dev_list, node) node 342 net/ncsi/internal.h list_for_each_entry_rcu(np, &ndp->packages, node) node 344 net/ncsi/internal.h list_for_each_entry_rcu(nc, &np->channels, node) node 230 net/ncsi/ncsi-manage.c list_add_tail_rcu(&nc->node, &np->channels); node 254 net/ncsi/ncsi-manage.c list_del_rcu(&nc->node); node 298 net/ncsi/ncsi-manage.c list_add_tail_rcu(&np->node, &ndp->packages); node 312 net/ncsi/ncsi-manage.c list_for_each_entry_safe(nc, tmp, &np->channels, node) node 317 net/ncsi/ncsi-manage.c list_del_rcu(&np->node); node 1661 net/ncsi/ncsi-manage.c list_add_tail_rcu(&ndp->node, &ncsi_dev_list); node 1822 net/ncsi/ncsi-manage.c list_for_each_entry_safe(np, tmp, &ndp->packages, node) node 1826 net/ncsi/ncsi-manage.c list_del_rcu(&ndp->node); node 42 net/netfilter/nf_conncount.c struct list_head node; node 50 net/netfilter/nf_conncount.c struct rb_node node; node 91 net/netfilter/nf_conncount.c list_del(&conn->node); node 136 net/netfilter/nf_conncount.c list_for_each_entry_safe(conn, conn_n, &list->head, node) { node 191 net/netfilter/nf_conncount.c list_add_tail(&conn->node, &list->head); node 234 net/netfilter/nf_conncount.c list_for_each_entry_safe(conn, conn_n, &list->head, node) { node 286 net/netfilter/nf_conncount.c rb_erase(&rbconn->node, root); node 322 net/netfilter/nf_conncount.c rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node); node 373 net/netfilter/nf_conncount.c list_add(&conn->node, &rbconn->list.head); node 377 net/netfilter/nf_conncount.c rb_link_node_rcu(&rbconn->node, parent, rbnode); node 378 net/netfilter/nf_conncount.c rb_insert_color(&rbconn->node, root); node 404 net/netfilter/nf_conncount.c rbconn = rb_entry(parent, struct nf_conncount_rb, node); node 449 net/netfilter/nf_conncount.c struct rb_node *node; node 457 net/netfilter/nf_conncount.c for (node = rb_first(root); node != NULL; node = rb_next(node)) { node 458 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); node 472 net/netfilter/nf_conncount.c node = rb_first(root); node 473 net/netfilter/nf_conncount.c while (node != NULL) { node 474 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); node 475 net/netfilter/nf_conncount.c node = rb_next(node); node 554 net/netfilter/nf_conncount.c list_for_each_entry_safe(conn, conn_n, &list->head, node) node 562 net/netfilter/nf_conncount.c struct rb_node *node; node 564 net/netfilter/nf_conncount.c while ((node = rb_first(r)) != NULL) { node 565 net/netfilter/nf_conncount.c rbconn = rb_entry(node, struct nf_conncount_rb, node); node 567 net/netfilter/nf_conncount.c rb_erase(node, r); node 194 net/netfilter/nf_flow_table_core.c .head_offset = offsetof(struct flow_offload_tuple_rhash, node), node 208 net/netfilter/nf_flow_table_core.c &flow->tuplehash[0].node, node 214 net/netfilter/nf_flow_table_core.c &flow->tuplehash[1].node, node 218 net/netfilter/nf_flow_table_core.c &flow->tuplehash[0].node, node 238 net/netfilter/nf_flow_table_core.c &flow->tuplehash[FLOW_OFFLOAD_DIR_ORIGINAL].node, node 241 net/netfilter/nf_flow_table_core.c &flow->tuplehash[FLOW_OFFLOAD_DIR_REPLY].node, node 30 net/netfilter/nft_set_hash.c struct rhash_head node; node 70 net/netfilter/nft_set_hash.c .head_offset = offsetof(struct nft_rhash_elem, node), node 137 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, node 171 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, node 234 net/netfilter/nft_set_hash.c rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params); node 252 net/netfilter/nft_set_hash.c return rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params) == 0; node 333 net/netfilter/nft_set_hash.c rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params); node 418 net/netfilter/nft_set_hash.c struct hlist_node node; node 432 net/netfilter/nft_set_hash.c hlist_for_each_entry_rcu(he, &priv->table[hash], node) { node 452 net/netfilter/nft_set_hash.c hlist_for_each_entry_rcu(he, &priv->table[hash], node) { node 472 net/netfilter/nft_set_hash.c hlist_for_each_entry_rcu(he, &priv->table[hash], node) { node 510 net/netfilter/nft_set_hash.c hlist_for_each_entry(he, &priv->table[hash], node) { node 518 net/netfilter/nft_set_hash.c hlist_add_head_rcu(&this->node, &priv->table[hash]); node 549 net/netfilter/nft_set_hash.c hlist_for_each_entry(he, &priv->table[hash], node) { node 566 net/netfilter/nft_set_hash.c hlist_del_rcu(&he->node); node 578 net/netfilter/nft_set_hash.c hlist_for_each_entry_rcu(he, &priv->table[i], node) { node 622 net/netfilter/nft_set_hash.c hlist_for_each_entry_safe(he, next, &priv->table[i], node) { node 623 net/netfilter/nft_set_hash.c hlist_del_rcu(&he->node); node 26 net/netfilter/nft_set_rbtree.c struct rb_node node; node 63 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); node 144 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); node 231 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); node 254 net/netfilter/nft_set_rbtree.c rb_link_node_rcu(&new->node, parent, p); node 255 net/netfilter/nft_set_rbtree.c rb_insert_color(&new->node, &priv->root); node 285 net/netfilter/nft_set_rbtree.c rb_erase(&rbe->node, &priv->root); node 324 net/netfilter/nft_set_rbtree.c rbe = rb_entry(parent, struct nft_rbtree_elem, node); node 359 net/netfilter/nft_set_rbtree.c struct rb_node *node; node 362 net/netfilter/nft_set_rbtree.c for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) { node 363 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); node 390 net/netfilter/nft_set_rbtree.c struct rb_node *node; node 398 net/netfilter/nft_set_rbtree.c for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) { node 399 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); node 411 net/netfilter/nft_set_rbtree.c rb_erase(&rbe_prev->node, &priv->root); node 425 net/netfilter/nft_set_rbtree.c rb_erase(&rbe_end->node, &priv->root); node 428 net/netfilter/nft_set_rbtree.c node = rb_next(node); node 429 net/netfilter/nft_set_rbtree.c if (!node) node 433 net/netfilter/nft_set_rbtree.c rb_erase(&rbe_prev->node, &priv->root); node 471 net/netfilter/nft_set_rbtree.c struct rb_node *node; node 475 net/netfilter/nft_set_rbtree.c while ((node = priv->root.rb_node) != NULL) { node 476 net/netfilter/nft_set_rbtree.c rb_erase(node, &priv->root); node 477 net/netfilter/nft_set_rbtree.c rbe = rb_entry(node, struct nft_rbtree_elem, node); node 91 net/netfilter/xt_hashlimit.c struct hlist_node node; node 117 net/netfilter/xt_hashlimit.c struct hlist_node node; /* global list of all htables */ node 206 net/netfilter/xt_hashlimit.c hlist_for_each_entry_rcu(ent, &ht->hash[hash], node) node 252 net/netfilter/xt_hashlimit.c hlist_add_head_rcu(&ent->node, &ht->hash[hash_dst(ht, dst)]); node 269 net/netfilter/xt_hashlimit.c hlist_del_rcu(&ent->node); node 356 net/netfilter/xt_hashlimit.c hlist_add_head(&hinfo->node, &hashlimit_net->htables); node 370 net/netfilter/xt_hashlimit.c hlist_for_each_entry_safe(dh, n, &ht->hash[i], node) { node 412 net/netfilter/xt_hashlimit.c hlist_for_each_entry(hinfo, &hashlimit_net->htables, node) { node 425 net/netfilter/xt_hashlimit.c hlist_del(&hinfo->node); node 1177 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, &htable->hash[*bucket], node) node 1191 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, &htable->hash[*bucket], node) node 1205 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, &htable->hash[*bucket], node) node 1260 net/netfilter/xt_hashlimit.c hlist_for_each_entry(hinfo, &hashlimit_net->htables, node) node 513 net/netlink/af_netlink.c &nlk_sk(sk)->node, node 598 net/netlink/af_netlink.c if (!rhashtable_remove_fast(&table->hash, &nlk_sk(sk)->node, node 2744 net/netlink/af_netlink.c .head_offset = offsetof(struct netlink_sock, node), node 47 net/netlink/af_netlink.h struct rhash_head node; node 48 net/nfc/llcp.h struct hlist_node node; node 128 net/nfc/llcp_commands.c INIT_HLIST_NODE(&sdres->node); node 169 net/nfc/llcp_commands.c INIT_HLIST_NODE(&sdreq->node); node 185 net/nfc/llcp_commands.c hlist_for_each_entry_safe(sdp, n, head, node) { node 186 net/nfc/llcp_commands.c hlist_del(&sdp->node); node 563 net/nfc/llcp_commands.c hlist_for_each_entry_safe(sdp, n, tlv_list, node) { node 566 net/nfc/llcp_commands.c hlist_del(&sdp->node); node 593 net/nfc/llcp_commands.c hlist_for_each_entry_safe(sdreq, n, tlv_list, node) { node 598 net/nfc/llcp_commands.c hlist_del(&sdreq->node); node 600 net/nfc/llcp_commands.c hlist_add_head(&sdreq->node, &local->pending_sdreqs); node 255 net/nfc/llcp_core.c hlist_for_each_entry_safe(sdp, n, &local->pending_sdreqs, node) { node 261 net/nfc/llcp_core.c hlist_del(&sdp->node); node 263 net/nfc/llcp_core.c hlist_add_head(&sdp->node, &nl_sdres_list); node 1319 net/nfc/llcp_core.c hlist_add_head(&sdp->node, &llc_sdres_list); node 1327 net/nfc/llcp_core.c hlist_for_each_entry(sdp, &local->pending_sdreqs, node) { node 1336 net/nfc/llcp_core.c hlist_del(&sdp->node); node 1338 net/nfc/llcp_core.c hlist_add_head(&sdp->node, &nl_sdres_list); node 395 net/nfc/netlink.c hlist_for_each_entry_safe(sdres, n, sdres_list, node) { node 412 net/nfc/netlink.c hlist_del(&sdres->node); node 1204 net/nfc/netlink.c hlist_add_head(&sdreq->node, &sdreq_list); node 208 net/openvswitch/flow.h struct hlist_node node[2]; node 220 net/openvswitch/flow_table.c hlist_for_each_entry_safe(flow, n, head, flow_table.node[ver]) { node 221 net/openvswitch/flow_table.c hlist_del_rcu(&flow->flow_table.node[ver]); node 223 net/openvswitch/flow_table.c hlist_del_rcu(&flow->ufid_table.node[ufid_ver]); node 261 net/openvswitch/flow_table.c hlist_for_each_entry_rcu(flow, head, flow_table.node[ver]) { node 288 net/openvswitch/flow_table.c hlist_add_head_rcu(&flow->flow_table.node[ti->node_ver], head); node 297 net/openvswitch/flow_table.c hlist_add_head_rcu(&flow->ufid_table.node[ti->node_ver], head); node 316 net/openvswitch/flow_table.c ufid_table.node[old_ver]) node 320 net/openvswitch/flow_table.c flow_table.node[old_ver]) node 438 net/openvswitch/flow_table.c hlist_for_each_entry_rcu(flow, head, flow_table.node[ti->node_ver]) { node 521 net/openvswitch/flow_table.c hlist_for_each_entry_rcu(flow, head, ufid_table.node[ti->node_ver]) { node 571 net/openvswitch/flow_table.c hlist_del_rcu(&flow->flow_table.node[ti->node_ver]); node 574 net/openvswitch/flow_table.c hlist_del_rcu(&flow->ufid_table.node[ufid_ti->node_ver]); node 131 net/qrtr/qrtr.c static int qrtr_local_enqueue(struct qrtr_node *node, struct sk_buff *skb, node 134 net/qrtr/qrtr.c static int qrtr_bcast_enqueue(struct qrtr_node *node, struct sk_buff *skb, node 145 net/qrtr/qrtr.c struct qrtr_node *node = container_of(kref, struct qrtr_node, ref); node 147 net/qrtr/qrtr.c if (node->nid != QRTR_EP_NID_AUTO) node 148 net/qrtr/qrtr.c radix_tree_delete(&qrtr_nodes, node->nid); node 150 net/qrtr/qrtr.c list_del(&node->item); node 153 net/qrtr/qrtr.c cancel_work_sync(&node->work); node 154 net/qrtr/qrtr.c skb_queue_purge(&node->rx_queue); node 155 net/qrtr/qrtr.c kfree(node); node 159 net/qrtr/qrtr.c static struct qrtr_node *qrtr_node_acquire(struct qrtr_node *node) node 161 net/qrtr/qrtr.c if (node) node 162 net/qrtr/qrtr.c kref_get(&node->ref); node 163 net/qrtr/qrtr.c return node; node 167 net/qrtr/qrtr.c static void qrtr_node_release(struct qrtr_node *node) node 169 net/qrtr/qrtr.c if (!node) node 171 net/qrtr/qrtr.c kref_put_mutex(&node->ref, __qrtr_node_release, &qrtr_node_lock); node 175 net/qrtr/qrtr.c static int qrtr_node_enqueue(struct qrtr_node *node, struct sk_buff *skb, node 189 net/qrtr/qrtr.c hdr->dst_node_id = cpu_to_le32(node->nid); node 201 net/qrtr/qrtr.c mutex_lock(&node->ep_lock); node 202 net/qrtr/qrtr.c if (node->ep) node 203 net/qrtr/qrtr.c rc = node->ep->xmit(node->ep, skb); node 206 net/qrtr/qrtr.c mutex_unlock(&node->ep_lock); node 217 net/qrtr/qrtr.c struct qrtr_node *node; node 220 net/qrtr/qrtr.c node = radix_tree_lookup(&qrtr_nodes, nid); node 221 net/qrtr/qrtr.c node = qrtr_node_acquire(node); node 224 net/qrtr/qrtr.c return node; node 232 net/qrtr/qrtr.c static void qrtr_node_assign(struct qrtr_node *node, unsigned int nid) node 234 net/qrtr/qrtr.c if (node->nid != QRTR_EP_NID_AUTO || nid == QRTR_EP_NID_AUTO) node 238 net/qrtr/qrtr.c radix_tree_insert(&qrtr_nodes, nid, node); node 239 net/qrtr/qrtr.c node->nid = nid; node 253 net/qrtr/qrtr.c struct qrtr_node *node = ep->node; node 319 net/qrtr/qrtr.c skb_queue_tail(&node->rx_queue, skb); node 320 net/qrtr/qrtr.c schedule_work(&node->work); node 364 net/qrtr/qrtr.c struct qrtr_node *node = container_of(work, struct qrtr_node, work); node 370 net/qrtr/qrtr.c while ((skb = skb_dequeue(&node->rx_queue)) != NULL) { node 382 net/qrtr/qrtr.c qrtr_node_assign(node, cb->src_node); node 400 net/qrtr/qrtr.c pkt->client.node = cpu_to_le32(dst.sq_node); node 403 net/qrtr/qrtr.c if (qrtr_node_enqueue(node, skb, QRTR_TYPE_RESUME_TX, node 420 net/qrtr/qrtr.c struct qrtr_node *node; node 425 net/qrtr/qrtr.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 426 net/qrtr/qrtr.c if (!node) node 429 net/qrtr/qrtr.c INIT_WORK(&node->work, qrtr_node_rx_work); node 430 net/qrtr/qrtr.c kref_init(&node->ref); node 431 net/qrtr/qrtr.c mutex_init(&node->ep_lock); node 432 net/qrtr/qrtr.c skb_queue_head_init(&node->rx_queue); node 433 net/qrtr/qrtr.c node->nid = QRTR_EP_NID_AUTO; node 434 net/qrtr/qrtr.c node->ep = ep; node 436 net/qrtr/qrtr.c qrtr_node_assign(node, nid); node 439 net/qrtr/qrtr.c list_add(&node->item, &qrtr_all_nodes); node 441 net/qrtr/qrtr.c ep->node = node; node 453 net/qrtr/qrtr.c struct qrtr_node *node = ep->node; node 454 net/qrtr/qrtr.c struct sockaddr_qrtr src = {AF_QIPCRTR, node->nid, QRTR_PORT_CTRL}; node 459 net/qrtr/qrtr.c mutex_lock(&node->ep_lock); node 460 net/qrtr/qrtr.c node->ep = NULL; node 461 net/qrtr/qrtr.c mutex_unlock(&node->ep_lock); node 470 net/qrtr/qrtr.c qrtr_node_release(node); node 471 net/qrtr/qrtr.c ep->node = NULL; node 516 net/qrtr/qrtr.c pkt->client.node = cpu_to_le32(ipc->us.sq_node); node 669 net/qrtr/qrtr.c static int qrtr_local_enqueue(struct qrtr_node *node, struct sk_buff *skb, node 698 net/qrtr/qrtr.c static int qrtr_bcast_enqueue(struct qrtr_node *node, struct sk_buff *skb, node 705 net/qrtr/qrtr.c list_for_each_entry(node, &qrtr_all_nodes, item) { node 710 net/qrtr/qrtr.c qrtr_node_enqueue(node, skbn, type, from, to); node 727 net/qrtr/qrtr.c struct qrtr_node *node; node 764 net/qrtr/qrtr.c node = NULL; node 775 net/qrtr/qrtr.c node = qrtr_node_lookup(addr->sq_node); node 776 net/qrtr/qrtr.c if (!node) { node 809 net/qrtr/qrtr.c rc = enqueue_fn(node, skb, type, &ipc->us, addr); node 814 net/qrtr/qrtr.c qrtr_node_release(node); node 23 net/qrtr/qrtr.h struct qrtr_node *node; node 401 net/rds/cong.c struct rb_node *node; node 405 net/rds/cong.c while ((node = rb_first(&rds_cong_tree))) { node 406 net/rds/cong.c map = rb_entry(node, struct rds_cong_map, m_rb_node); node 289 net/rds/ib_rdma.c struct llist_node *node; node 293 net/rds/ib_rdma.c node = llist_del_all(llist); node 294 net/rds/ib_rdma.c while (node) { node 295 net/rds/ib_rdma.c next = node->next; node 296 net/rds/ib_rdma.c ibmr = llist_entry(node, struct rds_ib_mr, llnode); node 298 net/rds/ib_rdma.c node = next; node 131 net/rds/rdma.c struct rb_node *node; node 136 net/rds/rdma.c while ((node = rb_first(&rs->rs_rdma_keys))) { node 137 net/rds/rdma.c mr = rb_entry(node, struct rds_mr, r_rb_node); node 60 net/rfkill/core.c struct list_head node; node 180 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) { node 398 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) { node 449 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) node 945 net/rfkill/core.c INIT_LIST_HEAD(&rfkill->node); node 1024 net/rfkill/core.c list_add_tail(&rfkill->node, &rfkill_list); node 1065 net/rfkill/core.c list_del_init(&rfkill->node); node 1088 net/rfkill/core.c list_del_init(&rfkill->node); node 1124 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) { node 1234 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) node 1241 net/rfkill/core.c list_for_each_entry(rfkill, &rfkill_list, node) node 678 net/rose/rose_route.c struct rose_node *node; node 683 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) { node 684 net/rose/rose_route.c if (rosecmpm(addr, &node->address, node->mask) == 0) { node 685 net/rose/rose_route.c for (i = 0; i < node->count; i++) { node 686 net/rose/rose_route.c if (node->neighbour[i]->restarted) { node 687 net/rose/rose_route.c res = node->neighbour[i]; node 694 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) { node 695 net/rose/rose_route.c if (rosecmpm(addr, &node->address, node->mask) == 0) { node 696 net/rose/rose_route.c for (i = 0; i < node->count; i++) { node 697 net/rose/rose_route.c if (!rose_ftimer_running(node->neighbour[i])) { node 698 net/rose/rose_route.c res = node->neighbour[i]; node 137 net/sched/sch_htb.c struct rb_node node[TC_HTB_NUMPRIO]; /* node for self or feed tree */ node 282 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, node[prio]); node 289 net/sched/sch_htb.c rb_link_node(&cl->node[prio], parent, p); node 290 net/sched/sch_htb.c rb_insert_color(&cl->node[prio], root); node 383 net/sched/sch_htb.c if (hprio->ptr == cl->node + prio) node 386 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, &hprio->row); node 447 net/sched/sch_htb.c if (p->inner.clprio[prio].ptr == cl->node + prio) { node 456 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, node 747 net/sched/sch_htb.c rb_entry(n, struct htb_class, node[prio]); node 807 net/sched/sch_htb.c cl = rb_entry(*sp->pptr, struct htb_class, node[prio]); node 1393 net/sched/sch_htb.c RB_CLEAR_NODE(&cl->node[prio]); node 766 net/sctp/input.c hlist_add_head(&epb->node, &head->chain); node 800 net/sctp/input.c hlist_del_init(&epb->node); node 913 net/sctp/input.c .head_offset = offsetof(struct sctp_transport, node), node 948 net/sctp/input.c rhl_for_each_entry_rcu(transport, tmp, list, node) node 956 net/sctp/input.c &t->node, sctp_hash_params); node 968 net/sctp/input.c rhltable_remove(&sctp_transport_hashtable, &t->node, node 989 net/sctp/input.c rhl_for_each_entry_rcu(t, tmp, list, node) { node 1019 net/sctp/input.c rhl_for_each_entry_rcu(t, tmp, list, node) node 8545 net/sctp/socket.c hlist_add_head(&pp->node, &head->chain); node 8554 net/sctp/socket.c __hlist_del(&pp->node); node 462 net/smc/smc_core.c struct rb_node *node; node 473 net/smc/smc_core.c node = rb_first(&lgr->conns_all); node 474 net/smc/smc_core.c while (node) { node 475 net/smc/smc_core.c conn = rb_entry(node, struct smc_connection, alert_node); node 485 net/smc/smc_core.c node = rb_first(&lgr->conns_all); node 261 net/smc/smc_core.h struct rb_node *node; node 263 net/smc/smc_core.h node = lgr->conns_all.rb_node; node 264 net/smc/smc_core.h while (node) { node 265 net/smc/smc_core.h struct smc_connection *cur = rb_entry(node, node 269 net/smc/smc_core.h node = node->rb_left; node 272 net/smc/smc_core.h node = node->rb_right; node 113 net/sunrpc/svc.c unsigned int node; node 123 net/sunrpc/svc.c node = first_online_node; node 124 net/sunrpc/svc.c if (nr_cpus_node(node) > 2) { node 198 net/sunrpc/svc.c unsigned int node; node 205 net/sunrpc/svc.c for_each_node_with_cpus(node) { node 208 net/sunrpc/svc.c m->to_pool[node] = pidx; node 209 net/sunrpc/svc.c m->pool_to[pidx] = node; node 306 net/sunrpc/svc.c unsigned int node = m->pool_to[pidx]; node 319 net/sunrpc/svc.c set_cpus_allowed_ptr(task, cpumask_of(node)); node 324 net/sunrpc/svc.c set_cpus_allowed_ptr(task, cpumask_of_node(node)); node 565 net/sunrpc/svc.c svc_init_buffer(struct svc_rqst *rqstp, unsigned int size, int node) node 581 net/sunrpc/svc.c struct page *p = alloc_pages_node(node, GFP_KERNEL, 0); node 604 net/sunrpc/svc.c svc_rqst_alloc(struct svc_serv *serv, struct svc_pool *pool, int node) node 608 net/sunrpc/svc.c rqstp = kzalloc_node(sizeof(*rqstp), GFP_KERNEL, node); node 617 net/sunrpc/svc.c rqstp->rq_argp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node); node 621 net/sunrpc/svc.c rqstp->rq_resp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node); node 625 net/sunrpc/svc.c if (!svc_init_buffer(rqstp, serv->sv_max_mesg, node)) node 636 net/sunrpc/svc.c svc_prepare_thread(struct svc_serv *serv, struct svc_pool *pool, int node) node 640 net/sunrpc/svc.c rqstp = svc_rqst_alloc(serv, pool, node); node 714 net/sunrpc/svc.c int node; node 720 net/sunrpc/svc.c node = svc_pool_map_get_node(chosen_pool->sp_id); node 721 net/sunrpc/svc.c rqstp = svc_prepare_thread(serv, chosen_pool, node); node 727 net/sunrpc/svc.c node, "%s", serv->sv_name); node 175 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c struct llist_node *node; node 177 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c while ((node = llist_del_first(&rdma->sc_recv_ctxts))) { node 178 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c ctxt = llist_entry(node, struct svc_rdma_recv_ctxt, rc_node); node 187 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c struct llist_node *node; node 189 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c node = llist_del_first(&rdma->sc_recv_ctxts); node 190 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c if (!node) node 192 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c ctxt = llist_entry(node, struct svc_rdma_recv_ctxt, rc_node); node 1107 net/sunrpc/xprtrdma/verbs.c struct llist_node *node; node 1110 net/sunrpc/xprtrdma/verbs.c node = llist_del_first(&buf->rb_free_reps); node 1111 net/sunrpc/xprtrdma/verbs.c if (!node) node 1113 net/sunrpc/xprtrdma/verbs.c return llist_entry(node, struct rpcrdma_rep, rr_node); node 70 net/tipc/addr.h static inline int tipc_node2scope(u32 node) node 72 net/tipc/addr.h return node ? TIPC_NODE_SCOPE : TIPC_CLUSTER_SCOPE; node 292 net/tipc/bcast.c dnode = dst->node; node 725 net/tipc/bcast.c void tipc_nlist_add(struct tipc_nlist *nl, u32 node) node 727 net/tipc/bcast.c if (node == nl->self) node 729 net/tipc/bcast.c else if (tipc_dest_push(&nl->list, node, 0)) node 733 net/tipc/bcast.c void tipc_nlist_del(struct tipc_nlist *nl, u32 node) node 735 net/tipc/bcast.c if (node == nl->self) node 737 net/tipc/bcast.c else if (tipc_dest_del(&nl->list, node, 0)) node 777 net/tipc/bcast.c u32 node, port; node 788 net/tipc/bcast.c node = msg_orignode(hdr); node 789 net/tipc/bcast.c if (node == tipc_own_addr(net)) node 797 net/tipc/bcast.c if (msg_orignode(_hdr) != node) node 842 net/tipc/bcast.c if (msg_orignode(_hdr) != node) node 64 net/tipc/bcast.h void tipc_nlist_add(struct tipc_nlist *nl, u32 node); node 65 net/tipc/bcast.h void tipc_nlist_del(struct tipc_nlist *nl, u32 node); node 68 net/tipc/group.c u32 node; node 232 net/tipc/group.c u32 node, u32 port) node 235 net/tipc/group.c u64 nkey, key = (u64)node << 32 | port; node 240 net/tipc/group.c nkey = (u64)m->node << 32 | m->port; node 252 net/tipc/group.c u32 node, u32 port) node 256 net/tipc/group.c m = tipc_group_find_member(grp, node, port); node 263 net/tipc/group.c u32 node) node 270 net/tipc/group.c if (m->node == node) node 279 net/tipc/group.c u64 nkey, key = (u64)m->node << 32 | m->port; node 288 net/tipc/group.c nkey = (u64)tmp->node << 32 | tmp->port; node 301 net/tipc/group.c u32 node, u32 port, node 313 net/tipc/group.c m->node = node; node 319 net/tipc/group.c tipc_nlist_add(&grp->dests, m->node); node 324 net/tipc/group.c void tipc_group_add_member(struct tipc_group *grp, u32 node, node 327 net/tipc/group.c tipc_group_create_member(grp, node, port, instance, MBR_PUBLISHED); node 345 net/tipc/group.c if (!tipc_group_find_node(grp, m->node)) node 346 net/tipc/group.c tipc_nlist_del(&grp->dests, m->node); node 460 net/tipc/group.c return tipc_group_cong(grp, m->node, m->port, len, &m); node 497 net/tipc/group.c u32 node, port; node 504 net/tipc/group.c node = msg_orignode(hdr); node 510 net/tipc/group.c m = tipc_group_find_member(grp, node, port); node 574 net/tipc/group.c tipc_group_update_rcv_win(grp, blks, node, port, xmitq); node 581 net/tipc/group.c void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node, node 590 net/tipc/group.c m = tipc_group_find_member(grp, node, port); node 675 net/tipc/group.c evt.port.node = m->node; node 681 net/tipc/group.c GROUP_H_SIZE, sizeof(evt), dnode, m->node, node 704 net/tipc/group.c m->node, tipc_own_addr(grp->net), node 738 net/tipc/group.c u32 node = msg_orignode(hdr); node 746 net/tipc/group.c if (grp->scope == TIPC_NODE_SCOPE && node != tipc_own_addr(grp->net)) node 749 net/tipc/group.c m = tipc_group_find_member(grp, node, port); node 754 net/tipc/group.c m = tipc_group_create_member(grp, node, port, node 857 net/tipc/group.c u32 node = evt->port.node; node 869 net/tipc/group.c if (!grp->loopback && node == self && port == grp->portid) node 872 net/tipc/group.c m = tipc_group_find_member(grp, node, port); node 878 net/tipc/group.c m = tipc_group_create_member(grp, node, port, instance, node 909 net/tipc/group.c if (!tipc_node_is_up(net, node)) node 50 net/tipc/group.h void tipc_group_add_member(struct tipc_group *grp, u32 node, node 71 net/tipc/group.h void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node, node 46 net/tipc/name_distr.c u32 node; node 214 net/tipc/name_distr.c publ->node, publ->key); node 222 net/tipc/name_distr.c publ->type, publ->lower, publ->node, publ->port, node 240 net/tipc/name_distr.c if (e->node != addr) node 265 net/tipc/name_distr.c u32 node, u32 dtype) node 276 net/tipc/name_distr.c TIPC_CLUSTER_SCOPE, node, node 279 net/tipc/name_distr.c tipc_node_subscribe(net, &p->binding_node, node); node 284 net/tipc/name_distr.c upper, node, key); node 286 net/tipc/name_distr.c tipc_node_unsubscribe(net, &p->binding_node, node); node 291 net/tipc/name_distr.c type, lower, node); node 307 net/tipc/name_distr.c u32 node; node 318 net/tipc/name_distr.c node = msg_orignode(msg); node 320 net/tipc/name_distr.c tipc_update_nametbl(net, item, node, mtype); node 345 net/tipc/name_distr.c publ->node = self; node 347 net/tipc/name_distr.c publ->node = self; node 93 net/tipc/name_table.c u32 scope, u32 node, u32 port, node 105 net/tipc/name_table.c publ->node = node; node 223 net/tipc/name_table.c u32 node, u32 port, node 239 net/tipc/name_table.c if (p->key == key && (!p->node || p->node == node)) node 244 net/tipc/name_table.c p = tipc_publ_create(type, lower, upper, scope, node, port, key); node 247 net/tipc/name_table.c if (in_own_node(net, node)) node 254 net/tipc/name_table.c p->port, p->node, p->scope, first); node 266 net/tipc/name_table.c u32 node, u32 key) node 271 net/tipc/name_table.c if (p->key != key || (node && node != p->node)) node 316 net/tipc/name_table.c p->node, p->scope, first); node 338 net/tipc/name_table.c u32 scope, u32 node, node 358 net/tipc/name_table.c scope, node, port, key); node 365 net/tipc/name_table.c u32 node, u32 key) node 380 net/tipc/name_table.c p = tipc_service_remove_publ(sr, node, key); node 388 net/tipc/name_table.c p->port, node, p->scope, last); node 433 net/tipc/name_table.c u32 node = 0; node 465 net/tipc/name_table.c node = p->node; node 470 net/tipc/name_table.c *dnode = node; node 498 net/tipc/name_table.c if (p->port == exclude && p->node == self) node 500 net/tipc/name_table.c tipc_dest_push(dsts, p->node, p->port); node 571 net/tipc/name_table.c tipc_nlist_add(nodes, p->node); node 600 net/tipc/name_table.c tipc_group_add_member(grp, p->node, p->port, p->lower); node 764 net/tipc/name_table.c tipc_service_remove_publ(sr, p->node, p->key); node 848 net/tipc/name_table.c if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_NODE, p->node)) node 983 net/tipc/name_table.c struct tipc_dest *tipc_dest_find(struct list_head *l, u32 node, u32 port) node 988 net/tipc/name_table.c if (dst->node == node && dst->port == port) node 994 net/tipc/name_table.c bool tipc_dest_push(struct list_head *l, u32 node, u32 port) node 998 net/tipc/name_table.c if (tipc_dest_find(l, node, port)) node 1004 net/tipc/name_table.c dst->node = node; node 1010 net/tipc/name_table.c bool tipc_dest_pop(struct list_head *l, u32 *node, u32 *port) node 1019 net/tipc/name_table.c if (node) node 1020 net/tipc/name_table.c *node = dst->node; node 1026 net/tipc/name_table.c bool tipc_dest_del(struct list_head *l, u32 node, u32 port) node 1030 net/tipc/name_table.c dst = tipc_dest_find(l, node, port); node 79 net/tipc/name_table.h u32 node; node 109 net/tipc/name_table.h u32 tipc_nametbl_translate(struct net *net, u32 type, u32 instance, u32 *node); node 126 net/tipc/name_table.h u32 node, u32 ref, u32 key); node 129 net/tipc/name_table.h u32 node, u32 key); node 138 net/tipc/name_table.h u32 node; node 141 net/tipc/name_table.h struct tipc_dest *tipc_dest_find(struct list_head *l, u32 node, u32 port); node 142 net/tipc/name_table.h bool tipc_dest_push(struct list_head *l, u32 node, u32 port); node 143 net/tipc/name_table.h bool tipc_dest_pop(struct list_head *l, u32 *node, u32 *port); node 144 net/tipc/name_table.h bool tipc_dest_del(struct list_head *l, u32 node, u32 port); node 871 net/tipc/netlink_compat.c u32 node, depth, type, lowbound, upbound; node 922 net/tipc/netlink_compat.c node = nla_get_u32(publ[TIPC_NLA_PUBL_NODE]); node 923 net/tipc/netlink_compat.c sprintf(port_str, "<%u.%u.%u:%u>", tipc_zone(node), tipc_cluster(node), node 924 net/tipc/netlink_compat.c tipc_node(node), nla_get_u32(publ[TIPC_NLA_PUBL_REF])); node 1028 net/tipc/netlink_compat.c u32 node; node 1038 net/tipc/netlink_compat.c node = nla_get_u32(con[TIPC_NLA_CON_NODE]); node 1040 net/tipc/netlink_compat.c tipc_zone(node), node 1041 net/tipc/netlink_compat.c tipc_cluster(node), node 1042 net/tipc/netlink_compat.c tipc_node(node), node 1086 net/tipc/netlink_compat.c struct nlattr *node[TIPC_NLA_NODE_MAX + 1]; node 1092 net/tipc/netlink_compat.c err = nla_parse_nested_deprecated(node, TIPC_NLA_NODE_MAX, node 1097 net/tipc/netlink_compat.c node_info.addr = htonl(nla_get_u32(node[TIPC_NLA_NODE_ADDR])); node 1098 net/tipc/netlink_compat.c node_info.up = htonl(nla_get_flag(node[TIPC_NLA_NODE_UP])); node 161 net/tipc/node.c static void tipc_node_delete(struct tipc_node *node); node 166 net/tipc/node.c static void tipc_node_put(struct tipc_node *node); node 168 net/tipc/node.c static void tipc_node_delete_from_list(struct tipc_node *node); node 246 net/tipc/node.c static void tipc_node_put(struct tipc_node *node) node 248 net/tipc/node.c kref_put(&node->kref, tipc_node_kref_release); node 251 net/tipc/node.c static void tipc_node_get(struct tipc_node *node) node 253 net/tipc/node.c kref_get(&node->kref); node 262 net/tipc/node.c struct tipc_node *node; node 266 net/tipc/node.c hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) { node 267 net/tipc/node.c if (node->addr != addr) node 269 net/tipc/node.c if (!kref_get_unless_zero(&node->kref)) node 270 net/tipc/node.c node = NULL; node 274 net/tipc/node.c return node; node 466 net/tipc/node.c static void tipc_node_delete_from_list(struct tipc_node *node) node 468 net/tipc/node.c list_del_rcu(&node->list); node 469 net/tipc/node.c hlist_del_rcu(&node->hash); node 470 net/tipc/node.c tipc_node_put(node); node 473 net/tipc/node.c static void tipc_node_delete(struct tipc_node *node) node 475 net/tipc/node.c trace_tipc_node_delete(node, true, " "); node 476 net/tipc/node.c tipc_node_delete_from_list(node); node 478 net/tipc/node.c del_timer_sync(&node->timer); node 479 net/tipc/node.c tipc_node_put(node); node 485 net/tipc/node.c struct tipc_node *node, *t_node; node 488 net/tipc/node.c list_for_each_entry_safe(node, t_node, &tn->node_list, list) node 489 net/tipc/node.c tipc_node_delete(node); node 531 net/tipc/node.c struct tipc_node *node; node 538 net/tipc/node.c node = tipc_node_find(net, dnode); node 539 net/tipc/node.c if (!node) { node 552 net/tipc/node.c tipc_node_write_lock(node); node 553 net/tipc/node.c list_add_tail(&conn->list, &node->conn_sks); node 554 net/tipc/node.c tipc_node_write_unlock(node); node 556 net/tipc/node.c tipc_node_put(node); node 562 net/tipc/node.c struct tipc_node *node; node 568 net/tipc/node.c node = tipc_node_find(net, dnode); node 569 net/tipc/node.c if (!node) node 572 net/tipc/node.c tipc_node_write_lock(node); node 573 net/tipc/node.c list_for_each_entry_safe(conn, safe, &node->conn_sks, list) { node 579 net/tipc/node.c tipc_node_write_unlock(node); node 580 net/tipc/node.c tipc_node_put(node); node 583 net/tipc/node.c static void tipc_node_clear_links(struct tipc_node *node) node 588 net/tipc/node.c struct tipc_link_entry *le = &node->links[i]; node 593 net/tipc/node.c node->link_cnt--; node 1373 net/tipc/node.c struct tipc_node *node = tipc_node_find(net, addr); node 1375 net/tipc/node.c if (!node) node 1381 net/tipc/node.c tipc_node_read_lock(node); node 1382 net/tipc/node.c link = node->links[bearer_id].link; node 1387 net/tipc/node.c tipc_node_read_unlock(node); node 1389 net/tipc/node.c tipc_node_put(node); node 1394 net/tipc/node.c static int __tipc_nl_add_node(struct tipc_nl_msg *msg, struct tipc_node *node) node 1408 net/tipc/node.c if (nla_put_u32(msg->skb, TIPC_NLA_NODE_ADDR, node->addr)) node 1410 net/tipc/node.c if (node_is_up(node)) node 1985 net/tipc/node.c struct tipc_node *node; node 1997 net/tipc/node.c node = tipc_node_find(net, last_addr); node 1998 net/tipc/node.c if (!node) { node 2010 net/tipc/node.c tipc_node_put(node); node 2013 net/tipc/node.c list_for_each_entry_rcu(node, &tn->node_list, list) { node 2015 net/tipc/node.c if (node->addr == last_addr) node 2021 net/tipc/node.c tipc_node_read_lock(node); node 2022 net/tipc/node.c err = __tipc_nl_add_node(&msg, node); node 2024 net/tipc/node.c last_addr = node->addr; node 2025 net/tipc/node.c tipc_node_read_unlock(node); node 2029 net/tipc/node.c tipc_node_read_unlock(node); node 2085 net/tipc/node.c struct tipc_node *node; node 2109 net/tipc/node.c node = tipc_node_find_by_name(net, name, &bearer_id); node 2110 net/tipc/node.c if (!node) node 2113 net/tipc/node.c tipc_node_read_lock(node); node 2115 net/tipc/node.c link = node->links[bearer_id].link; node 2152 net/tipc/node.c tipc_node_read_unlock(node); node 2153 net/tipc/node.c tipc_bearer_xmit(net, bearer_id, &xmitq, &node->links[bearer_id].maddr); node 2192 net/tipc/node.c struct tipc_node *node; node 2195 net/tipc/node.c node = tipc_node_find_by_name(net, name, &bearer_id); node 2196 net/tipc/node.c if (!node) { node 2201 net/tipc/node.c tipc_node_read_lock(node); node 2202 net/tipc/node.c link = node->links[bearer_id].link; node 2204 net/tipc/node.c tipc_node_read_unlock(node); node 2210 net/tipc/node.c tipc_node_read_unlock(node); node 2228 net/tipc/node.c struct tipc_node *node; node 2254 net/tipc/node.c node = tipc_node_find_by_name(net, link_name, &bearer_id); node 2255 net/tipc/node.c if (!node) node 2258 net/tipc/node.c le = &node->links[bearer_id]; node 2259 net/tipc/node.c tipc_node_read_lock(node); node 2261 net/tipc/node.c link = node->links[bearer_id].link; node 2264 net/tipc/node.c tipc_node_read_unlock(node); node 2269 net/tipc/node.c tipc_node_read_unlock(node); node 2275 net/tipc/node.c struct tipc_node *node, u32 *prev_link) node 2283 net/tipc/node.c if (!node->links[i].link) node 2287 net/tipc/node.c node->links[i].link, NLM_F_MULTI); node 2300 net/tipc/node.c struct tipc_node *node; node 2316 net/tipc/node.c node = tipc_node_find(net, prev_node); node 2317 net/tipc/node.c if (!node) { node 2327 net/tipc/node.c tipc_node_put(node); node 2329 net/tipc/node.c list_for_each_entry_continue_rcu(node, &tn->node_list, node 2331 net/tipc/node.c tipc_node_read_lock(node); node 2332 net/tipc/node.c err = __tipc_nl_add_node_links(net, &msg, node, node 2334 net/tipc/node.c tipc_node_read_unlock(node); node 2338 net/tipc/node.c prev_node = node->addr; node 2345 net/tipc/node.c list_for_each_entry_rcu(node, &tn->node_list, list) { node 2346 net/tipc/node.c tipc_node_read_lock(node); node 2347 net/tipc/node.c err = __tipc_nl_add_node_links(net, &msg, node, node 2349 net/tipc/node.c tipc_node_read_unlock(node); node 2353 net/tipc/node.c prev_node = node->addr; node 2533 net/tipc/node.c u32 tipc_node_get_addr(struct tipc_node *node) node 2535 net/tipc/node.c return (node) ? node->addr : 0; node 74 net/tipc/node.h u32 tipc_node_get_addr(struct tipc_node *node); node 83 net/tipc/node.h int tipc_node_get_linkname(struct net *net, u32 bearer_id, u32 node, node 115 net/tipc/socket.c struct rhash_head node; node 689 net/tipc/socket.c addr->addr.id.node = tsk_peer_node(tsk); node 692 net/tipc/socket.c addr->addr.id.node = tipc_own_addr(sock_net(sk)); node 897 net/tipc/socket.c u32 node, port; node 900 net/tipc/socket.c node = dest->addr.id.node; node 902 net/tipc/socket.c if (!port && !node) node 907 net/tipc/socket.c !tipc_dest_find(&tsk->cong_links, node, 0) && node 909 net/tipc/socket.c !tipc_group_cong(tsk->group, node, port, blks, node 917 net/tipc/socket.c rc = tipc_send_group_msg(net, tsk, m, mb, node, port, dlen); node 944 net/tipc/socket.c u32 node, port, exclude; node 967 net/tipc/socket.c tipc_dest_pop(&dsts, &node, &port); node 968 net/tipc/socket.c cong = tipc_group_cong(tsk->group, node, port, blks, node 982 net/tipc/socket.c if (likely(!cong && !tipc_dest_find(cong_links, node, 0))) node 987 net/tipc/socket.c !tipc_dest_find(cong_links, node, 0) && node 989 net/tipc/socket.c !tipc_group_cong(tsk->group, node, port, node 1002 net/tipc/socket.c rc = tipc_send_group_msg(net, tsk, m, mbr, node, port, dlen); node 1116 net/tipc/socket.c tipc_dest_pop(&dsts, &dest->addr.id.node, &dest->addr.id.ref); node 1367 net/tipc/socket.c dnode = dest->addr.id.node; node 1565 net/tipc/socket.c srcaddr->sock.addr.id.node = msg_orignode(hdr); node 2836 net/tipc/socket.c if (!rhashtable_lookup_insert_fast(&tn->sk_rht, &tsk->node, node 2850 net/tipc/socket.c if (!rhashtable_remove_fast(&tn->sk_rht, &tsk->node, tsk_rht_params)) { node 2858 net/tipc/socket.c .head_offset = offsetof(struct tipc_sock, node), node 3148 net/tipc/socket.c tsk1->peer.addr.id.node = onode; node 3153 net/tipc/socket.c tsk2->peer.addr.id.node = onode; node 43 net/tipc/subscr.c u32 event, u32 port, u32 node) node 53 net/tipc/subscr.c tipc_evt_write(evt, port.node, node); node 77 net/tipc/subscr.c u32 event, u32 port, u32 node, node 99 net/tipc/subscr.c event, port, node); node 78 net/tipc/subscr.h u32 event, u32 port, u32 node, node 50 net/x25/x25_forward.c x25_frwd = list_entry(entry, struct x25_forward, node); node 69 net/x25/x25_forward.c list_add(&new_frwd->node, &x25_forward_list); node 103 net/x25/x25_forward.c frwd = list_entry(entry, struct x25_forward, node); node 140 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); node 142 net/x25/x25_forward.c list_del(&fwd->node); node 158 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); node 160 net/x25/x25_forward.c list_del(&fwd->node); node 267 net/x25/x25_link.c list_add(&nb->node, &x25_neigh_list); node 283 net/x25/x25_link.c if (nb->node.next) { node 284 net/x25/x25_link.c list_del(&nb->node); node 300 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 321 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 400 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 45 net/x25/x25_proc.c struct x25_route *rt = list_entry(v, struct x25_route, node); node 133 net/x25/x25_proc.c struct x25_forward *f = list_entry(v, struct x25_forward, node); node 36 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 55 net/x25/x25_route.c list_add(&rt->node, &x25_route_list); node 71 net/x25/x25_route.c if (rt->node.next) { node 72 net/x25/x25_route.c list_del(&rt->node); node 87 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 112 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 157 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 217 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 497 net/xdp/xsk.c struct xsk_map_node *node; node 502 net/xdp/xsk.c node = list_first_entry_or_null(&xs->map_list, struct xsk_map_node, node 503 net/xdp/xsk.c node); node 504 net/xdp/xsk.c if (node) { node 505 net/xdp/xsk.c WARN_ON(xsk_map_inc(node->map)); node 506 net/xdp/xsk.c map = node->map; node 507 net/xdp/xsk.c *map_entry = node->map_entry; node 59 net/xfrm/xfrm_policy.c struct rb_node node; node 770 net/xfrm/xfrm_policy.c static void xfrm_pol_inexact_node_init(struct xfrm_pol_inexact_node *node, node 773 net/xfrm/xfrm_policy.c node->addr = *addr; node 774 net/xfrm/xfrm_policy.c node->prefixlen = prefixlen; node 780 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 782 net/xfrm/xfrm_policy.c node = kzalloc(sizeof(*node), GFP_ATOMIC); node 783 net/xfrm/xfrm_policy.c if (node) node 784 net/xfrm/xfrm_policy.c xfrm_pol_inexact_node_init(node, addr, prefixlen); node 786 net/xfrm/xfrm_policy.c return node; node 893 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 906 net/xfrm/xfrm_policy.c node = rb_entry(*p, struct xfrm_pol_inexact_node, node); node 908 net/xfrm/xfrm_policy.c prefixlen = min(node->prefixlen, n->prefixlen); node 910 net/xfrm/xfrm_policy.c delta = xfrm_policy_addr_delta(&n->addr, &node->addr, node 917 net/xfrm/xfrm_policy.c bool same_prefixlen = node->prefixlen == n->prefixlen; node 925 net/xfrm/xfrm_policy.c node->prefixlen = prefixlen; node 927 net/xfrm/xfrm_policy.c xfrm_policy_inexact_list_reinsert(net, node, family); node 936 net/xfrm/xfrm_policy.c n = node; node 941 net/xfrm/xfrm_policy.c rb_link_node_rcu(&n->node, parent, p); node 942 net/xfrm/xfrm_policy.c rb_insert_color(&n->node, new); node 951 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 960 net/xfrm/xfrm_policy.c node = rb_entry(rnode, struct xfrm_pol_inexact_node, node); node 961 net/xfrm/xfrm_policy.c rb_erase(&node->node, &v->root); node 962 net/xfrm/xfrm_policy.c xfrm_policy_inexact_node_reinsert(net, node, &n->root, node 982 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 989 net/xfrm/xfrm_policy.c node = rb_entry(*p, struct xfrm_pol_inexact_node, node); node 991 net/xfrm/xfrm_policy.c delta = xfrm_policy_addr_delta(addr, &node->addr, node 992 net/xfrm/xfrm_policy.c node->prefixlen, node 994 net/xfrm/xfrm_policy.c if (delta == 0 && prefixlen >= node->prefixlen) { node 996 net/xfrm/xfrm_policy.c return node; node 1004 net/xfrm/xfrm_policy.c if (prefixlen < node->prefixlen) { node 1005 net/xfrm/xfrm_policy.c delta = xfrm_policy_addr_delta(addr, &node->addr, node 1016 net/xfrm/xfrm_policy.c rb_erase(&node->node, root); node 1019 net/xfrm/xfrm_policy.c xfrm_pol_inexact_node_init(node, addr, node 1021 net/xfrm/xfrm_policy.c cached = node; node 1027 net/xfrm/xfrm_policy.c xfrm_policy_inexact_node_merge(net, node, node 1029 net/xfrm/xfrm_policy.c kfree_rcu(node, rcu); node 1038 net/xfrm/xfrm_policy.c node = cached; node 1039 net/xfrm/xfrm_policy.c if (!node) { node 1040 net/xfrm/xfrm_policy.c node = xfrm_pol_inexact_node_alloc(addr, prefixlen); node 1041 net/xfrm/xfrm_policy.c if (!node) node 1045 net/xfrm/xfrm_policy.c rb_link_node_rcu(&node->node, parent, p); node 1046 net/xfrm/xfrm_policy.c rb_insert_color(&node->node, root); node 1048 net/xfrm/xfrm_policy.c return node; node 1053 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 1057 net/xfrm/xfrm_policy.c node = rb_entry(rn, struct xfrm_pol_inexact_node, node); node 1059 net/xfrm/xfrm_policy.c xfrm_policy_inexact_gc_tree(&node->root, rm); node 1062 net/xfrm/xfrm_policy.c if (!hlist_empty(&node->hhead) || !RB_EMPTY_ROOT(&node->root)) { node 1067 net/xfrm/xfrm_policy.c rb_erase(&node->node, r); node 1068 net/xfrm/xfrm_policy.c kfree_rcu(node, rcu); node 1917 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_node *node; node 1920 net/xfrm/xfrm_policy.c node = rb_entry(parent, struct xfrm_pol_inexact_node, node); node 1922 net/xfrm/xfrm_policy.c delta = xfrm_policy_addr_delta(addr, &node->addr, node 1923 net/xfrm/xfrm_policy.c node->prefixlen, family); node 1932 net/xfrm/xfrm_policy.c return node; node 132 samples/bpf/test_lru_dist.c struct pfect_lru_node *node = NULL; node 136 samples/bpf/test_lru_dist.c if (!bpf_map_lookup_elem(lru->map_fd, &key, &node)) { node 137 samples/bpf/test_lru_dist.c if (node) { node 138 samples/bpf/test_lru_dist.c list_move(&node->list, &lru->list); node 145 samples/bpf/test_lru_dist.c node = &lru->free_nodes[lru->cur_size++]; node 146 samples/bpf/test_lru_dist.c INIT_LIST_HEAD(&node->list); node 150 samples/bpf/test_lru_dist.c node = list_last_entry(&lru->list, node 153 samples/bpf/test_lru_dist.c bpf_map_update_elem(lru->map_fd, &node->key, &null_node, BPF_EXIST); node 156 samples/bpf/test_lru_dist.c node->key = key; node 157 samples/bpf/test_lru_dist.c list_move(&node->list, &lru->list); node 161 samples/bpf/test_lru_dist.c assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_EXIST)); node 164 samples/bpf/test_lru_dist.c assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_NOEXIST)); node 539 samples/qmi/qmi_sample_client.c struct sockaddr_qrtr sq = { AF_QIPCRTR, service->node, service->port }; node 232 samples/v4l/v4l2-pci-skeleton.c struct skel_buffer *buf, *node; node 236 samples/v4l/v4l2-pci-skeleton.c list_for_each_entry_safe(buf, node, &skel->buf_list, list) { node 177 samples/vfio-mdev/mdpy-fb.c pci_info(pdev, "fb%d registered\n", info->node); node 29 scripts/dtc/checks.c typedef void (*check_fn)(struct check *c, struct dt_info *dti, struct node *node); node 62 scripts/dtc/checks.c struct node *node, node 76 scripts/dtc/checks.c else if (node && node->srcpos) node 77 scripts/dtc/checks.c pos = node->srcpos; node 92 scripts/dtc/checks.c if (node) { node 94 scripts/dtc/checks.c xasprintf_append(&str, "%s:%s: ", node->fullpath, prop->name); node 96 scripts/dtc/checks.c xasprintf_append(&str, "%s: ", node->fullpath); node 106 scripts/dtc/checks.c pos = node->srcpos; node 119 scripts/dtc/checks.c #define FAIL(c, dti, node, ...) \ node 123 scripts/dtc/checks.c check_msg((c), dti, node, NULL, __VA_ARGS__); \ node 126 scripts/dtc/checks.c #define FAIL_PROP(c, dti, node, prop, ...) \ node 130 scripts/dtc/checks.c check_msg((c), dti, node, prop, __VA_ARGS__); \ node 134 scripts/dtc/checks.c static void check_nodes_props(struct check *c, struct dt_info *dti, struct node *node) node 136 scripts/dtc/checks.c struct node *child; node 138 scripts/dtc/checks.c TRACE(c, "%s", node->fullpath); node 140 scripts/dtc/checks.c c->fn(c, dti, node); node 142 scripts/dtc/checks.c for_each_child(node, child) node 148 scripts/dtc/checks.c struct node *dt = dti->dt; node 192 scripts/dtc/checks.c struct node *node) node 194 scripts/dtc/checks.c FAIL(c, dti, node, "always_fail check"); node 199 scripts/dtc/checks.c struct node *node) node 204 scripts/dtc/checks.c prop = get_property(node, propname); node 209 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "property is not a string"); node 217 scripts/dtc/checks.c struct node *node) node 224 scripts/dtc/checks.c prop = get_property(node, propname); node 233 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "property is not a string list"); node 246 scripts/dtc/checks.c struct node *node) node 251 scripts/dtc/checks.c prop = get_property(node, propname); node 256 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "property is not a single cell"); node 268 scripts/dtc/checks.c struct node *node) node 270 scripts/dtc/checks.c struct node *child, *child2; node 272 scripts/dtc/checks.c for_each_child(node, child) node 282 scripts/dtc/checks.c struct node *node) node 286 scripts/dtc/checks.c for_each_property(node, prop) { node 291 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "Duplicate property name"); node 304 scripts/dtc/checks.c struct node *node) node 306 scripts/dtc/checks.c int n = strspn(node->name, c->data); node 308 scripts/dtc/checks.c if (n < strlen(node->name)) node 309 scripts/dtc/checks.c FAIL(c, dti, node, "Bad character '%c' in node name", node 310 scripts/dtc/checks.c node->name[n]); node 315 scripts/dtc/checks.c struct node *node) node 317 scripts/dtc/checks.c int n = strspn(node->name, c->data); node 319 scripts/dtc/checks.c if (n < node->basenamelen) node 320 scripts/dtc/checks.c FAIL(c, dti, node, "Character '%c' not recommended in node name", node 321 scripts/dtc/checks.c node->name[n]); node 326 scripts/dtc/checks.c struct node *node) node 328 scripts/dtc/checks.c if (strchr(get_unitname(node), '@')) node 329 scripts/dtc/checks.c FAIL(c, dti, node, "multiple '@' characters in node name"); node 334 scripts/dtc/checks.c struct node *node) node 336 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 337 scripts/dtc/checks.c struct property *prop = get_property(node, "reg"); node 339 scripts/dtc/checks.c if (get_subnode(node, "__overlay__")) { node 345 scripts/dtc/checks.c prop = get_property(node, "ranges"); node 352 scripts/dtc/checks.c FAIL(c, dti, node, "node has a reg or ranges property, but no unit name"); node 355 scripts/dtc/checks.c FAIL(c, dti, node, "node has a unit name, but no reg property"); node 361 scripts/dtc/checks.c struct node *node) node 365 scripts/dtc/checks.c for_each_property(node, prop) { node 369 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "Bad character '%c' in property name", node 377 scripts/dtc/checks.c struct node *node) node 381 scripts/dtc/checks.c for_each_property(node, prop) { node 401 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "Character '%c' not recommended in property name", node 408 scripts/dtc/checks.c #define DESCLABEL_ARGS(node,prop,mark) \ node 412 scripts/dtc/checks.c ((prop) ? "' in " : ""), (node)->fullpath node 415 scripts/dtc/checks.c const char *label, struct node *node, node 418 scripts/dtc/checks.c struct node *dt = dti->dt; node 419 scripts/dtc/checks.c struct node *othernode = NULL; node 434 scripts/dtc/checks.c if ((othernode != node) || (otherprop != prop) || (othermark != mark)) node 435 scripts/dtc/checks.c FAIL(c, dti, node, "Duplicate label '%s' on " DESCLABEL_FMT node 437 scripts/dtc/checks.c label, DESCLABEL_ARGS(node, prop, mark), node 442 scripts/dtc/checks.c struct node *node) node 447 scripts/dtc/checks.c for_each_label(node->labels, l) node 448 scripts/dtc/checks.c check_duplicate_label(c, dti, l->label, node, NULL, NULL); node 450 scripts/dtc/checks.c for_each_property(node, prop) { node 454 scripts/dtc/checks.c check_duplicate_label(c, dti, l->label, node, prop, NULL); node 457 scripts/dtc/checks.c check_duplicate_label(c, dti, m->ref, node, prop, m); node 463 scripts/dtc/checks.c struct node *node, const char *propname) node 465 scripts/dtc/checks.c struct node *root = dti->dt; node 470 scripts/dtc/checks.c prop = get_property(node, propname); node 475 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "bad length (%d) %s property", node 483 scripts/dtc/checks.c if (node != get_node_by_ref(root, m->ref)) node 487 scripts/dtc/checks.c FAIL(c, dti, node, "%s is a reference to another node", node 501 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "bad value (0x%x) in %s property", node 510 scripts/dtc/checks.c struct node *node) node 512 scripts/dtc/checks.c struct node *root = dti->dt; node 513 scripts/dtc/checks.c struct node *other; node 517 scripts/dtc/checks.c assert(!node->phandle); node 519 scripts/dtc/checks.c phandle = check_phandle_prop(c, dti, node, "phandle"); node 521 scripts/dtc/checks.c linux_phandle = check_phandle_prop(c, dti, node, "linux,phandle"); node 528 scripts/dtc/checks.c FAIL(c, dti, node, "mismatching 'phandle' and 'linux,phandle'" node 535 scripts/dtc/checks.c if (other && (other != node)) { node 536 scripts/dtc/checks.c FAIL(c, dti, node, "duplicated phandle 0x%x (seen before at %s)", node 541 scripts/dtc/checks.c node->phandle = phandle; node 546 scripts/dtc/checks.c struct node *node) node 550 scripts/dtc/checks.c for (pp = &node->proplist; *pp; pp = &((*pp)->next)) node 559 scripts/dtc/checks.c if ((prop->val.len != node->basenamelen+1) node 560 scripts/dtc/checks.c || (memcmp(prop->val.val, node->name, node->basenamelen) != 0)) { node 561 scripts/dtc/checks.c FAIL(c, dti, node, "\"name\" property is incorrect (\"%s\" instead" node 580 scripts/dtc/checks.c struct node *node) node 582 scripts/dtc/checks.c struct node *dt = dti->dt; node 585 scripts/dtc/checks.c for_each_property(node, prop) { node 587 scripts/dtc/checks.c struct node *refnode; node 596 scripts/dtc/checks.c FAIL(c, dti, node, "Reference to non-existent node or " node 615 scripts/dtc/checks.c struct node *node) node 617 scripts/dtc/checks.c struct node *dt = dti->dt; node 620 scripts/dtc/checks.c for_each_property(node, prop) { node 622 scripts/dtc/checks.c struct node *refnode; node 630 scripts/dtc/checks.c FAIL(c, dti, node, "Reference to non-existent node or label \"%s\"\n", node 646 scripts/dtc/checks.c struct node *node) node 648 scripts/dtc/checks.c if (generate_symbols && node->labels) node 650 scripts/dtc/checks.c if (node->omit_if_unused && !node->is_referenced) node 651 scripts/dtc/checks.c delete_node(node); node 670 scripts/dtc/checks.c struct node *node) node 674 scripts/dtc/checks.c for_each_property(node, prop) { node 680 scripts/dtc/checks.c check_is_string_list(c, dti, node); node 686 scripts/dtc/checks.c struct node *node) node 690 scripts/dtc/checks.c if (!streq(node->name, "aliases")) node 693 scripts/dtc/checks.c for_each_property(node, prop) { node 695 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "aliases property is not a valid node (%s)", node 700 scripts/dtc/checks.c FAIL(c, dti, node, "aliases property name must include only lowercase and '-'"); node 706 scripts/dtc/checks.c struct node *node) node 710 scripts/dtc/checks.c node->addr_cells = -1; node 711 scripts/dtc/checks.c node->size_cells = -1; node 713 scripts/dtc/checks.c prop = get_property(node, "#address-cells"); node 715 scripts/dtc/checks.c node->addr_cells = propval_cell(prop); node 717 scripts/dtc/checks.c prop = get_property(node, "#size-cells"); node 719 scripts/dtc/checks.c node->size_cells = propval_cell(prop); node 730 scripts/dtc/checks.c struct node *node) node 735 scripts/dtc/checks.c prop = get_property(node, "reg"); node 739 scripts/dtc/checks.c if (!node->parent) { node 740 scripts/dtc/checks.c FAIL(c, dti, node, "Root node has a \"reg\" property"); node 745 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "property is empty"); node 747 scripts/dtc/checks.c addr_cells = node_addr_cells(node->parent); node 748 scripts/dtc/checks.c size_cells = node_size_cells(node->parent); node 752 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "property has invalid length (%d bytes) " node 759 scripts/dtc/checks.c struct node *node) node 764 scripts/dtc/checks.c prop = get_property(node, "ranges"); node 768 scripts/dtc/checks.c if (!node->parent) { node 769 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "Root node has a \"ranges\" property"); node 773 scripts/dtc/checks.c p_addr_cells = node_addr_cells(node->parent); node 774 scripts/dtc/checks.c p_size_cells = node_size_cells(node->parent); node 775 scripts/dtc/checks.c c_addr_cells = node_addr_cells(node); node 776 scripts/dtc/checks.c c_size_cells = node_size_cells(node); node 781 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "empty \"ranges\" property but its " node 783 scripts/dtc/checks.c c_addr_cells, node->parent->fullpath, node 786 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "empty \"ranges\" property but its " node 788 scripts/dtc/checks.c c_size_cells, node->parent->fullpath, node 791 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "\"ranges\" property has invalid length (%d bytes) " node 803 scripts/dtc/checks.c static void check_pci_bridge(struct check *c, struct dt_info *dti, struct node *node) node 808 scripts/dtc/checks.c prop = get_property(node, "device_type"); node 812 scripts/dtc/checks.c node->bus = &pci_bus; node 814 scripts/dtc/checks.c if (!strprefixeq(node->name, node->basenamelen, "pci") && node 815 scripts/dtc/checks.c !strprefixeq(node->name, node->basenamelen, "pcie")) node 816 scripts/dtc/checks.c FAIL(c, dti, node, "node name is not \"pci\" or \"pcie\""); node 818 scripts/dtc/checks.c prop = get_property(node, "ranges"); node 820 scripts/dtc/checks.c FAIL(c, dti, node, "missing ranges for PCI bridge (or not a bridge)"); node 822 scripts/dtc/checks.c if (node_addr_cells(node) != 3) node 823 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #address-cells for PCI bridge"); node 824 scripts/dtc/checks.c if (node_size_cells(node) != 2) node 825 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #size-cells for PCI bridge"); node 827 scripts/dtc/checks.c prop = get_property(node, "bus-range"); node 832 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "value must be 2 cells"); node 837 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "1st cell must be less than or equal to 2nd cell"); node 839 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "maximum bus number must be less than 256"); node 844 scripts/dtc/checks.c static void check_pci_device_bus_num(struct check *c, struct dt_info *dti, struct node *node) node 850 scripts/dtc/checks.c if (!node->parent || (node->parent->bus != &pci_bus)) node 853 scripts/dtc/checks.c prop = get_property(node, "reg"); node 860 scripts/dtc/checks.c prop = get_property(node->parent, "bus-range"); node 869 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "PCI bus number %d out of range, expected (%d - %d)", node 874 scripts/dtc/checks.c static void check_pci_device_reg(struct check *c, struct dt_info *dti, struct node *node) node 877 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 882 scripts/dtc/checks.c if (!node->parent || (node->parent->bus != &pci_bus)) node 885 scripts/dtc/checks.c prop = get_property(node, "reg"); node 887 scripts/dtc/checks.c FAIL(c, dti, node, "missing PCI reg property"); node 893 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "PCI reg config space address cells 2 and 3 must be 0"); node 900 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "PCI reg address is not configuration space"); node 902 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "PCI reg config space address register number must be 0"); node 914 scripts/dtc/checks.c FAIL(c, dti, node, "PCI unit address format error, expected \"%s\"", node 923 scripts/dtc/checks.c static bool node_is_compatible(struct node *node, const char *compat) node 928 scripts/dtc/checks.c prop = get_property(node, "compatible"); node 940 scripts/dtc/checks.c static void check_simple_bus_bridge(struct check *c, struct dt_info *dti, struct node *node) node 942 scripts/dtc/checks.c if (node_is_compatible(node, "simple-bus")) node 943 scripts/dtc/checks.c node->bus = &simple_bus; node 948 scripts/dtc/checks.c static void check_simple_bus_reg(struct check *c, struct dt_info *dti, struct node *node) node 951 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 957 scripts/dtc/checks.c if (!node->parent || (node->parent->bus != &simple_bus)) node 960 scripts/dtc/checks.c prop = get_property(node, "reg"); node 964 scripts/dtc/checks.c prop = get_property(node, "ranges"); node 967 scripts/dtc/checks.c cells = ((cell_t *)prop->val.val) + node_addr_cells(node); node 971 scripts/dtc/checks.c if (node->parent->parent && !(node->bus == &simple_bus)) node 972 scripts/dtc/checks.c FAIL(c, dti, node, "missing or empty reg/ranges property"); node 976 scripts/dtc/checks.c size = node_addr_cells(node->parent); node 982 scripts/dtc/checks.c FAIL(c, dti, node, "simple-bus unit address format error, expected \"%s\"", node 991 scripts/dtc/checks.c static void check_i2c_bus_bridge(struct check *c, struct dt_info *dti, struct node *node) node 993 scripts/dtc/checks.c if (strprefixeq(node->name, node->basenamelen, "i2c-bus") || node 994 scripts/dtc/checks.c strprefixeq(node->name, node->basenamelen, "i2c-arb")) { node 995 scripts/dtc/checks.c node->bus = &i2c_bus; node 996 scripts/dtc/checks.c } else if (strprefixeq(node->name, node->basenamelen, "i2c")) { node 997 scripts/dtc/checks.c struct node *child; node 998 scripts/dtc/checks.c for_each_child(node, child) { node 999 scripts/dtc/checks.c if (strprefixeq(child->name, node->basenamelen, "i2c-bus")) node 1002 scripts/dtc/checks.c node->bus = &i2c_bus; node 1006 scripts/dtc/checks.c if (!node->children) node 1009 scripts/dtc/checks.c if (node_addr_cells(node) != 1) node 1010 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #address-cells for I2C bus"); node 1011 scripts/dtc/checks.c if (node_size_cells(node) != 0) node 1012 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #size-cells for I2C bus"); node 1017 scripts/dtc/checks.c static void check_i2c_bus_reg(struct check *c, struct dt_info *dti, struct node *node) node 1020 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 1026 scripts/dtc/checks.c if (!node->parent || (node->parent->bus != &i2c_bus)) node 1029 scripts/dtc/checks.c prop = get_property(node, "reg"); node 1034 scripts/dtc/checks.c FAIL(c, dti, node, "missing or empty reg property"); node 1041 scripts/dtc/checks.c FAIL(c, dti, node, "I2C bus unit address format error, expected \"%s\"", node 1047 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "I2C address must be less than 10-bits, got \"0x%x\"", node 1058 scripts/dtc/checks.c static void check_spi_bus_bridge(struct check *c, struct dt_info *dti, struct node *node) node 1062 scripts/dtc/checks.c if (strprefixeq(node->name, node->basenamelen, "spi")) { node 1063 scripts/dtc/checks.c node->bus = &spi_bus; node 1066 scripts/dtc/checks.c struct node *child; node 1068 scripts/dtc/checks.c if (node_addr_cells(node) != 1 || node_size_cells(node) != 0) node 1071 scripts/dtc/checks.c for_each_child(node, child) { node 1075 scripts/dtc/checks.c node->bus = &spi_bus; node 1079 scripts/dtc/checks.c if (node->bus == &spi_bus) node 1083 scripts/dtc/checks.c if (node->bus == &spi_bus && get_property(node, "reg")) node 1084 scripts/dtc/checks.c FAIL(c, dti, node, "node name for SPI buses should be 'spi'"); node 1086 scripts/dtc/checks.c if (node->bus != &spi_bus || !node->children) node 1089 scripts/dtc/checks.c if (get_property(node, "spi-slave")) node 1091 scripts/dtc/checks.c if (node_addr_cells(node) != spi_addr_cells) node 1092 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #address-cells for SPI bus"); node 1093 scripts/dtc/checks.c if (node_size_cells(node) != 0) node 1094 scripts/dtc/checks.c FAIL(c, dti, node, "incorrect #size-cells for SPI bus"); node 1099 scripts/dtc/checks.c static void check_spi_bus_reg(struct check *c, struct dt_info *dti, struct node *node) node 1102 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 1107 scripts/dtc/checks.c if (!node->parent || (node->parent->bus != &spi_bus)) node 1110 scripts/dtc/checks.c if (get_property(node->parent, "spi-slave")) node 1113 scripts/dtc/checks.c prop = get_property(node, "reg"); node 1118 scripts/dtc/checks.c FAIL(c, dti, node, "missing or empty reg property"); node 1125 scripts/dtc/checks.c FAIL(c, dti, node, "SPI bus unit address format error, expected \"%s\"", node 1131 scripts/dtc/checks.c struct node *node) node 1133 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 1135 scripts/dtc/checks.c if (node->parent && node->parent->bus) node 1142 scripts/dtc/checks.c FAIL(c, dti, node, "unit name should not have leading \"0x\""); node 1147 scripts/dtc/checks.c FAIL(c, dti, node, "unit name should not have leading 0s"); node 1156 scripts/dtc/checks.c struct node *node) node 1160 scripts/dtc/checks.c if (!node->parent) node 1163 scripts/dtc/checks.c reg = get_property(node, "reg"); node 1164 scripts/dtc/checks.c ranges = get_property(node, "ranges"); node 1169 scripts/dtc/checks.c if (node->parent->addr_cells == -1) node 1170 scripts/dtc/checks.c FAIL(c, dti, node, "Relying on default #address-cells value"); node 1172 scripts/dtc/checks.c if (node->parent->size_cells == -1) node 1173 scripts/dtc/checks.c FAIL(c, dti, node, "Relying on default #size-cells value"); node 1179 scripts/dtc/checks.c struct node *node) node 1182 scripts/dtc/checks.c struct node *child; node 1185 scripts/dtc/checks.c if (!node->parent || node->addr_cells < 0 || node->size_cells < 0) node 1188 scripts/dtc/checks.c if (get_property(node, "ranges") || !node->children) node 1191 scripts/dtc/checks.c for_each_child(node, child) { node 1198 scripts/dtc/checks.c FAIL(c, dti, node, "unnecessary #address-cells/#size-cells without \"ranges\" or child \"reg\" property"); node 1202 scripts/dtc/checks.c static bool node_is_disabled(struct node *node) node 1206 scripts/dtc/checks.c prop = get_property(node, "status"); node 1218 scripts/dtc/checks.c struct node *node, node 1221 scripts/dtc/checks.c struct node *childa; node 1223 scripts/dtc/checks.c if (node->addr_cells < 0 || node->size_cells < 0) node 1226 scripts/dtc/checks.c if (!node->children) node 1229 scripts/dtc/checks.c for_each_child(node, childa) { node 1230 scripts/dtc/checks.c struct node *childb; node 1239 scripts/dtc/checks.c for_each_child(node, childb) { node 1254 scripts/dtc/checks.c struct node *node) node 1256 scripts/dtc/checks.c check_unique_unit_address_common(c, dti, node, false); node 1261 scripts/dtc/checks.c struct node *node) node 1263 scripts/dtc/checks.c check_unique_unit_address_common(c, dti, node, true); node 1270 scripts/dtc/checks.c struct node *node) node 1272 scripts/dtc/checks.c struct node *dt = dti->dt; node 1273 scripts/dtc/checks.c struct node *chosen; node 1276 scripts/dtc/checks.c if (node != dt) node 1286 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1293 scripts/dtc/checks.c struct node *node) node 1295 scripts/dtc/checks.c if (!streq(node->name, "chosen")) node 1298 scripts/dtc/checks.c if (node->parent != dti->dt) node 1299 scripts/dtc/checks.c FAIL(c, dti, node, "chosen node must be at root node"); node 1304 scripts/dtc/checks.c struct node *node) node 1308 scripts/dtc/checks.c if (!streq(node->name, "chosen")) node 1311 scripts/dtc/checks.c prop = get_property(node, "bootargs"); node 1316 scripts/dtc/checks.c check_is_string(c, dti, node); node 1321 scripts/dtc/checks.c struct node *node) node 1325 scripts/dtc/checks.c if (!streq(node->name, "chosen")) node 1328 scripts/dtc/checks.c prop = get_property(node, "stdout-path"); node 1330 scripts/dtc/checks.c prop = get_property(node, "linux,stdout-path"); node 1333 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, "Use 'stdout-path' instead"); node 1337 scripts/dtc/checks.c check_is_string(c, dti, node); node 1349 scripts/dtc/checks.c struct node *node, node 1353 scripts/dtc/checks.c struct node *root = dti->dt; node 1357 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1364 scripts/dtc/checks.c struct node *provider_node; node 1390 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1397 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1409 scripts/dtc/checks.c FAIL(c, dti, node, "Missing property '%s' in node %s or bad phandle (referred from %s[%d])", node 1417 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1426 scripts/dtc/checks.c struct node *node) node 1431 scripts/dtc/checks.c prop = get_property(node, provider->prop_name); node 1435 scripts/dtc/checks.c check_property_phandle_args(c, dti, node, prop, provider); node 1482 scripts/dtc/checks.c struct node *node) node 1487 scripts/dtc/checks.c if (get_property(node, "gpio-hog")) node 1490 scripts/dtc/checks.c for_each_property(node, prop) { node 1499 scripts/dtc/checks.c check_property_phandle_args(c, dti, node, prop, &provider); node 1507 scripts/dtc/checks.c struct node *node) node 1511 scripts/dtc/checks.c for_each_property(node, prop) { node 1521 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1528 scripts/dtc/checks.c static bool node_is_interrupt_provider(struct node *node) node 1532 scripts/dtc/checks.c prop = get_property(node, "interrupt-controller"); node 1536 scripts/dtc/checks.c prop = get_property(node, "interrupt-map"); node 1544 scripts/dtc/checks.c struct node *node) node 1546 scripts/dtc/checks.c struct node *root = dti->dt; node 1547 scripts/dtc/checks.c struct node *irq_node = NULL, *parent = node; node 1551 scripts/dtc/checks.c irq_prop = get_property(node, "interrupts"); node 1556 scripts/dtc/checks.c FAIL_PROP(c, dti, node, irq_prop, "size (%d) is invalid, expected multiple of %zu", node 1560 scripts/dtc/checks.c if (parent != node && node_is_interrupt_provider(parent)) { node 1593 scripts/dtc/checks.c FAIL(c, dti, node, "Missing interrupt-parent"); node 1605 scripts/dtc/checks.c FAIL_PROP(c, dti, node, prop, node 1621 scripts/dtc/checks.c struct node *node) node 1623 scripts/dtc/checks.c struct node *child; node 1625 scripts/dtc/checks.c for_each_child(node, child) { node 1630 scripts/dtc/checks.c node->bus = &graph_port_bus; node 1633 scripts/dtc/checks.c if (!node->parent->bus && node 1634 scripts/dtc/checks.c (streq(node->parent->name, "ports") || get_property(node, "reg"))) node 1635 scripts/dtc/checks.c node->parent->bus = &graph_ports_bus; node 1644 scripts/dtc/checks.c struct node *node) node 1647 scripts/dtc/checks.c struct node *child; node 1649 scripts/dtc/checks.c if (node->bus != &graph_ports_bus && node->bus != &graph_port_bus) node 1652 scripts/dtc/checks.c for_each_child(node, child) { node 1662 scripts/dtc/checks.c if (cnt == 1 && node->addr_cells != -1) node 1663 scripts/dtc/checks.c FAIL(c, dti, node, "graph node has single child node '%s', #address-cells/#size-cells are not necessary", node 1664 scripts/dtc/checks.c node->children->name); node 1669 scripts/dtc/checks.c struct node *node) node 1672 scripts/dtc/checks.c const char *unitname = get_unitname(node); node 1675 scripts/dtc/checks.c prop = get_property(node, "reg"); node 1680 scripts/dtc/checks.c FAIL(c, dti, node, "graph node malformed 'reg' property"); node 1686 scripts/dtc/checks.c FAIL(c, dti, node, "graph node unit address error, expected \"%s\"", node 1689 scripts/dtc/checks.c if (node->parent->addr_cells != 1) node 1690 scripts/dtc/checks.c FAIL_PROP(c, dti, node, get_property(node, "#address-cells"), node 1692 scripts/dtc/checks.c node->parent->addr_cells); node 1693 scripts/dtc/checks.c if (node->parent->size_cells != 0) node 1694 scripts/dtc/checks.c FAIL_PROP(c, dti, node, get_property(node, "#size-cells"), node 1696 scripts/dtc/checks.c node->parent->size_cells); node 1700 scripts/dtc/checks.c struct node *node) node 1702 scripts/dtc/checks.c if (node->bus != &graph_port_bus) node 1705 scripts/dtc/checks.c if (!strprefixeq(node->name, node->basenamelen, "port")) node 1706 scripts/dtc/checks.c FAIL(c, dti, node, "graph port node name should be 'port'"); node 1708 scripts/dtc/checks.c check_graph_reg(c, dti, node); node 1712 scripts/dtc/checks.c static struct node *get_remote_endpoint(struct check *c, struct dt_info *dti, node 1713 scripts/dtc/checks.c struct node *endpoint) node 1716 scripts/dtc/checks.c struct node *node; node 1728 scripts/dtc/checks.c node = get_node_by_phandle(dti->dt, phandle); node 1729 scripts/dtc/checks.c if (!node) node 1732 scripts/dtc/checks.c return node; node 1736 scripts/dtc/checks.c struct node *node) node 1738 scripts/dtc/checks.c struct node *remote_node; node 1740 scripts/dtc/checks.c if (!node->parent || node->parent->bus != &graph_port_bus) node 1743 scripts/dtc/checks.c if (!strprefixeq(node->name, node->basenamelen, "endpoint")) node 1744 scripts/dtc/checks.c FAIL(c, dti, node, "graph endpoint node name should be 'endpoint'"); node 1746 scripts/dtc/checks.c check_graph_reg(c, dti, node); node 1748 scripts/dtc/checks.c remote_node = get_remote_endpoint(c, dti, node); node 1752 scripts/dtc/checks.c if (get_remote_endpoint(c, dti, remote_node) != node) node 1753 scripts/dtc/checks.c FAIL(c, dti, node, "graph connection to node '%s' is not bidirectional", node 37 scripts/dtc/dtc-parser.y struct node *node; node 38 scripts/dtc/dtc-parser.y struct node *nodelist; node 74 scripts/dtc/dtc-parser.y %type <node> devicetree node 75 scripts/dtc/dtc-parser.y %type <node> nodedef node 76 scripts/dtc/dtc-parser.y %type <node> subnode node 175 scripts/dtc/dtc-parser.y struct node *target = get_node_by_ref($1, $3); node 194 scripts/dtc/dtc-parser.y struct node *target = get_node_by_ref($1, $2); node 205 scripts/dtc/dtc-parser.y struct node *target = get_node_by_ref($1, $2); node 224 scripts/dtc/dtc-parser.y struct node *target = get_node_by_ref($1, $3); node 236 scripts/dtc/dtc-parser.y struct node *target = get_node_by_ref($1, $3); node 31 scripts/dtc/dtc.c static void fill_fullpaths(struct node *tree, const char *prefix) node 33 scripts/dtc/dtc.c struct node *child; node 154 scripts/dtc/dtc.h struct node *children; node 156 scripts/dtc/dtc.h struct node *parent; node 157 scripts/dtc/dtc.h struct node *next_sibling; node 202 scripts/dtc/dtc.h struct node *build_node(struct property *proplist, struct node *children, node 204 scripts/dtc/dtc.h struct node *build_node_delete(struct srcpos *srcpos); node 205 scripts/dtc/dtc.h struct node *name_node(struct node *node, char *name); node 206 scripts/dtc/dtc.h struct node *omit_node_if_unused(struct node *node); node 207 scripts/dtc/dtc.h struct node *reference_node(struct node *node); node 208 scripts/dtc/dtc.h struct node *chain_node(struct node *first, struct node *list); node 209 scripts/dtc/dtc.h struct node *merge_nodes(struct node *old_node, struct node *new_node); node 210 scripts/dtc/dtc.h struct node *add_orphan_node(struct node *old_node, struct node *new_node, char *ref); node 212 scripts/dtc/dtc.h void add_property(struct node *node, struct property *prop); node 213 scripts/dtc/dtc.h void delete_property_by_name(struct node *node, char *name); node 215 scripts/dtc/dtc.h void add_child(struct node *parent, struct node *child); node 216 scripts/dtc/dtc.h void delete_node_by_name(struct node *parent, char *name); node 217 scripts/dtc/dtc.h void delete_node(struct node *node); node 218 scripts/dtc/dtc.h void append_to_property(struct node *node, node 222 scripts/dtc/dtc.h const char *get_unitname(struct node *node); node 223 scripts/dtc/dtc.h struct property *get_property(struct node *node, const char *propname); node 226 scripts/dtc/dtc.h struct property *get_property_by_label(struct node *tree, const char *label, node 227 scripts/dtc/dtc.h struct node **node); node 228 scripts/dtc/dtc.h struct marker *get_marker_label(struct node *tree, const char *label, node 229 scripts/dtc/dtc.h struct node **node, struct property **prop); node 230 scripts/dtc/dtc.h struct node *get_subnode(struct node *node, const char *nodename); node 231 scripts/dtc/dtc.h struct node *get_node_by_path(struct node *tree, const char *path); node 232 scripts/dtc/dtc.h struct node *get_node_by_label(struct node *tree, const char *label); node 233 scripts/dtc/dtc.h struct node *get_node_by_phandle(struct node *tree, cell_t phandle); node 234 scripts/dtc/dtc.h struct node *get_node_by_ref(struct node *tree, const char *ref); node 235 scripts/dtc/dtc.h cell_t get_node_phandle(struct node *root, struct node *node); node 237 scripts/dtc/dtc.h uint32_t guess_boot_cpuid(struct node *tree); node 260 scripts/dtc/dtc.h struct node *dt; /* the device tree */ node 270 scripts/dtc/dtc.h struct node *tree, uint32_t boot_cpuid_phys); node 108 scripts/dtc/fdtget.c static int list_properties(const void *blob, int node) node 114 scripts/dtc/fdtget.c prop = fdt_first_property_offset(blob, node); node 136 scripts/dtc/fdtget.c static int list_subnodes(const void *blob, int node) node 145 scripts/dtc/fdtget.c tag = fdt_next_tag(blob, node, &nextoffset); node 148 scripts/dtc/fdtget.c pathp = fdt_get_name(blob, node, NULL); node 177 scripts/dtc/fdtget.c node = nextoffset; node 193 scripts/dtc/fdtget.c int node, const char *property) node 200 scripts/dtc/fdtget.c err = list_properties(blob, node); node 204 scripts/dtc/fdtget.c err = list_subnodes(blob, node); node 209 scripts/dtc/fdtget.c value = fdt_getprop(blob, node, property, &len); node 241 scripts/dtc/fdtget.c int i, node; node 248 scripts/dtc/fdtget.c node = fdt_path_offset(blob, arg[i]); node 249 scripts/dtc/fdtget.c if (node < 0) { node 254 scripts/dtc/fdtget.c report_error(arg[i], node); node 260 scripts/dtc/fdtget.c if (show_data_for_item(blob, disp, node, prop)) node 123 scripts/dtc/fdtput.c int node; node 126 scripts/dtc/fdtput.c node = fdt_path_offset(blob, node_name); node 127 scripts/dtc/fdtput.c if (node < 0) { node 128 scripts/dtc/fdtput.c report_error(node_name, -1, node); node 132 scripts/dtc/fdtput.c err = fdt_setprop(blob, node, property, buf, len); node 154 scripts/dtc/fdtput.c int node, offset = 0; node 160 scripts/dtc/fdtput.c for (sep = path; *sep; path = sep + 1, offset = node) { node 166 scripts/dtc/fdtput.c node = fdt_subnode_offset_namelen(blob, offset, path, node 168 scripts/dtc/fdtput.c if (node == -FDT_ERR_NOTFOUND) { node 169 scripts/dtc/fdtput.c node = fdt_add_subnode_namelen(blob, offset, path, node 172 scripts/dtc/fdtput.c if (node < 0) { node 173 scripts/dtc/fdtput.c report_error(path, sep - path, node); node 194 scripts/dtc/fdtput.c int node = 0; node 205 scripts/dtc/fdtput.c node = fdt_path_offset(blob, node_name); node 206 scripts/dtc/fdtput.c if (node < 0) { node 207 scripts/dtc/fdtput.c report_error(node_name, -1, node); node 212 scripts/dtc/fdtput.c node = fdt_add_subnode(blob, node, p + 1); node 213 scripts/dtc/fdtput.c if (node < 0) { node 214 scripts/dtc/fdtput.c report_error(p + 1, -1, node); node 235 scripts/dtc/flattree.c static void flatten_tree(struct node *tree, struct emitter *emit, node 240 scripts/dtc/flattree.c struct node *child; node 730 scripts/dtc/flattree.c static struct node *unflatten_tree(struct inbuf *dtbuf, node 734 scripts/dtc/flattree.c struct node *node; node 738 scripts/dtc/flattree.c node = build_node(NULL, NULL, NULL); node 743 scripts/dtc/flattree.c node->name = nodename_from_path(parent_flatname, flatname); node 745 scripts/dtc/flattree.c node->name = flatname; node 749 scripts/dtc/flattree.c struct node *child; node 754 scripts/dtc/flattree.c if (node->children) node 758 scripts/dtc/flattree.c add_property(node, prop); node 763 scripts/dtc/flattree.c add_child(node, child); node 787 scripts/dtc/flattree.c if (node->name != flatname) { node 791 scripts/dtc/flattree.c return node; node 809 scripts/dtc/flattree.c struct node *tree; node 11 scripts/dtc/fstree.c static struct node *read_fstree(const char *dirname) node 16 scripts/dtc/fstree.c struct node *tree; node 54 scripts/dtc/fstree.c struct node *newchild; node 70 scripts/dtc/fstree.c struct node *tree; node 116 scripts/dtc/libfdt/fdt_overlay.c static int overlay_phandle_add_offset(void *fdt, int node, node 123 scripts/dtc/libfdt/fdt_overlay.c val = fdt_getprop(fdt, node, name, &len); node 138 scripts/dtc/libfdt/fdt_overlay.c return fdt_setprop_inplace_u32(fdt, node, name, adj_val); node 156 scripts/dtc/libfdt/fdt_overlay.c static int overlay_adjust_node_phandles(void *fdto, int node, node 162 scripts/dtc/libfdt/fdt_overlay.c ret = overlay_phandle_add_offset(fdto, node, "phandle", delta); node 166 scripts/dtc/libfdt/fdt_overlay.c ret = overlay_phandle_add_offset(fdto, node, "linux,phandle", delta); node 170 scripts/dtc/libfdt/fdt_overlay.c fdt_for_each_subnode(child, fdto, node) { node 557 scripts/dtc/libfdt/fdt_overlay.c void *fdto, int node) node 562 scripts/dtc/libfdt/fdt_overlay.c fdt_for_each_property_offset(property, fdto, node) { node 580 scripts/dtc/libfdt/fdt_overlay.c fdt_for_each_subnode(subnode, fdto, node) { node 226 scripts/dtc/libfdt/libfdt.h #define fdt_for_each_subnode(node, fdt, parent) \ node 227 scripts/dtc/libfdt/libfdt.h for (node = fdt_first_subnode(fdt, parent); \ node 228 scripts/dtc/libfdt/libfdt.h node >= 0; \ node 229 scripts/dtc/libfdt/libfdt.h node = fdt_next_subnode(fdt, node)) node 605 scripts/dtc/libfdt/libfdt.h #define fdt_for_each_property_offset(property, fdt, node) \ node 606 scripts/dtc/libfdt/libfdt.h for (property = fdt_first_property_offset(fdt, node); \ node 88 scripts/dtc/livetree.c struct node *build_node(struct property *proplist, struct node *children, node 91 scripts/dtc/livetree.c struct node *new = xmalloc(sizeof(*new)); node 92 scripts/dtc/livetree.c struct node *child; node 107 scripts/dtc/livetree.c struct node *build_node_delete(struct srcpos *srcpos) node 109 scripts/dtc/livetree.c struct node *new = xmalloc(sizeof(*new)); node 119 scripts/dtc/livetree.c struct node *name_node(struct node *node, char *name) node 121 scripts/dtc/livetree.c assert(node->name == NULL); node 123 scripts/dtc/livetree.c node->name = name; node 125 scripts/dtc/livetree.c return node; node 128 scripts/dtc/livetree.c struct node *omit_node_if_unused(struct node *node) node 130 scripts/dtc/livetree.c node->omit_if_unused = 1; node 132 scripts/dtc/livetree.c return node; node 135 scripts/dtc/livetree.c struct node *reference_node(struct node *node) node 137 scripts/dtc/livetree.c node->is_referenced = 1; node 139 scripts/dtc/livetree.c return node; node 142 scripts/dtc/livetree.c struct node *merge_nodes(struct node *old_node, struct node *new_node) node 145 scripts/dtc/livetree.c struct node *new_child, *old_child; node 228 scripts/dtc/livetree.c struct node * add_orphan_node(struct node *dt, struct node *new_node, char *ref) node 231 scripts/dtc/livetree.c struct node *node; node 251 scripts/dtc/livetree.c node = build_node(p, new_node, NULL); node 252 scripts/dtc/livetree.c name_node(node, name); node 254 scripts/dtc/livetree.c add_child(dt, node); node 258 scripts/dtc/livetree.c struct node *chain_node(struct node *first, struct node *list) node 266 scripts/dtc/livetree.c void add_property(struct node *node, struct property *prop) node 272 scripts/dtc/livetree.c p = &node->proplist; node 279 scripts/dtc/livetree.c void delete_property_by_name(struct node *node, char *name) node 281 scripts/dtc/livetree.c struct property *prop = node->proplist; node 298 scripts/dtc/livetree.c void add_child(struct node *parent, struct node *child) node 300 scripts/dtc/livetree.c struct node **p; node 312 scripts/dtc/livetree.c void delete_node_by_name(struct node *parent, char *name) node 314 scripts/dtc/livetree.c struct node *node = parent->children; node 316 scripts/dtc/livetree.c while (node) { node 317 scripts/dtc/livetree.c if (streq(node->name, name)) { node 318 scripts/dtc/livetree.c delete_node(node); node 321 scripts/dtc/livetree.c node = node->next_sibling; node 325 scripts/dtc/livetree.c void delete_node(struct node *node) node 328 scripts/dtc/livetree.c struct node *child; node 330 scripts/dtc/livetree.c node->deleted = 1; node 331 scripts/dtc/livetree.c for_each_child(node, child) node 333 scripts/dtc/livetree.c for_each_property(node, prop) node 335 scripts/dtc/livetree.c delete_labels(&node->labels); node 338 scripts/dtc/livetree.c void append_to_property(struct node *node, node 345 scripts/dtc/livetree.c p = get_property(node, name); node 354 scripts/dtc/livetree.c add_property(node, p); node 399 scripts/dtc/livetree.c struct node *tree, uint32_t boot_cpuid_phys) node 416 scripts/dtc/livetree.c const char *get_unitname(struct node *node) node 418 scripts/dtc/livetree.c if (node->name[node->basenamelen] == '\0') node 421 scripts/dtc/livetree.c return node->name + node->basenamelen + 1; node 424 scripts/dtc/livetree.c struct property *get_property(struct node *node, const char *propname) node 428 scripts/dtc/livetree.c for_each_property(node, prop) node 447 scripts/dtc/livetree.c struct property *get_property_by_label(struct node *tree, const char *label, node 448 scripts/dtc/livetree.c struct node **node) node 451 scripts/dtc/livetree.c struct node *c; node 453 scripts/dtc/livetree.c *node = tree; node 464 scripts/dtc/livetree.c prop = get_property_by_label(c, label, node); node 469 scripts/dtc/livetree.c *node = NULL; node 473 scripts/dtc/livetree.c struct marker *get_marker_label(struct node *tree, const char *label, node 474 scripts/dtc/livetree.c struct node **node, struct property **prop) node 478 scripts/dtc/livetree.c struct node *c; node 480 scripts/dtc/livetree.c *node = tree; node 491 scripts/dtc/livetree.c m = get_marker_label(c, label, node, prop); node 497 scripts/dtc/livetree.c *node = NULL; node 501 scripts/dtc/livetree.c struct node *get_subnode(struct node *node, const char *nodename) node 503 scripts/dtc/livetree.c struct node *child; node 505 scripts/dtc/livetree.c for_each_child(node, child) node 512 scripts/dtc/livetree.c struct node *get_node_by_path(struct node *tree, const char *path) node 515 scripts/dtc/livetree.c struct node *child; node 539 scripts/dtc/livetree.c struct node *get_node_by_label(struct node *tree, const char *label) node 541 scripts/dtc/livetree.c struct node *child, *node; node 551 scripts/dtc/livetree.c node = get_node_by_label(child, label); node 552 scripts/dtc/livetree.c if (node) node 553 scripts/dtc/livetree.c return node; node 559 scripts/dtc/livetree.c struct node *get_node_by_phandle(struct node *tree, cell_t phandle) node 561 scripts/dtc/livetree.c struct node *child, *node; node 575 scripts/dtc/livetree.c node = get_node_by_phandle(child, phandle); node 576 scripts/dtc/livetree.c if (node) node 577 scripts/dtc/livetree.c return node; node 583 scripts/dtc/livetree.c struct node *get_node_by_ref(struct node *tree, const char *ref) node 593 scripts/dtc/livetree.c cell_t get_node_phandle(struct node *root, struct node *node) node 598 scripts/dtc/livetree.c if ((node->phandle != 0) && (node->phandle != -1)) node 599 scripts/dtc/livetree.c return node->phandle; node 604 scripts/dtc/livetree.c node->phandle = phandle; node 609 scripts/dtc/livetree.c if (!get_property(node, "linux,phandle") node 611 scripts/dtc/livetree.c add_property(node, build_property("linux,phandle", d, NULL)); node 613 scripts/dtc/livetree.c if (!get_property(node, "phandle") node 615 scripts/dtc/livetree.c add_property(node, build_property("phandle", d, NULL)); node 621 scripts/dtc/livetree.c return node->phandle; node 624 scripts/dtc/livetree.c uint32_t guess_boot_cpuid(struct node *tree) node 626 scripts/dtc/livetree.c struct node *cpus, *bootcpu; node 706 scripts/dtc/livetree.c static void sort_properties(struct node *node) node 711 scripts/dtc/livetree.c for_each_property_withdel(node, prop) node 719 scripts/dtc/livetree.c for_each_property_withdel(node, prop) node 724 scripts/dtc/livetree.c node->proplist = tbl[0]; node 734 scripts/dtc/livetree.c const struct node *a, *b; node 736 scripts/dtc/livetree.c a = *((const struct node * const *)ax); node 737 scripts/dtc/livetree.c b = *((const struct node * const *)bx); node 742 scripts/dtc/livetree.c static void sort_subnodes(struct node *node) node 745 scripts/dtc/livetree.c struct node *subnode, **tbl; node 747 scripts/dtc/livetree.c for_each_child_withdel(node, subnode) node 755 scripts/dtc/livetree.c for_each_child_withdel(node, subnode) node 760 scripts/dtc/livetree.c node->children = tbl[0]; node 768 scripts/dtc/livetree.c static void sort_node(struct node *node) node 770 scripts/dtc/livetree.c struct node *c; node 772 scripts/dtc/livetree.c sort_properties(node); node 773 scripts/dtc/livetree.c sort_subnodes(node); node 774 scripts/dtc/livetree.c for_each_child_withdel(node, c) node 785 scripts/dtc/livetree.c static struct node *build_and_name_child_node(struct node *parent, char *name) node 787 scripts/dtc/livetree.c struct node *node; node 789 scripts/dtc/livetree.c node = build_node(NULL, NULL, NULL); node 790 scripts/dtc/livetree.c name_node(node, xstrdup(name)); node 791 scripts/dtc/livetree.c add_child(parent, node); node 793 scripts/dtc/livetree.c return node; node 796 scripts/dtc/livetree.c static struct node *build_root_node(struct node *dt, char *name) node 798 scripts/dtc/livetree.c struct node *an; node 810 scripts/dtc/livetree.c static bool any_label_tree(struct dt_info *dti, struct node *node) node 812 scripts/dtc/livetree.c struct node *c; node 814 scripts/dtc/livetree.c if (node->labels) node 817 scripts/dtc/livetree.c for_each_child(node, c) node 825 scripts/dtc/livetree.c struct node *an, struct node *node, node 828 scripts/dtc/livetree.c struct node *dt = dti->dt; node 829 scripts/dtc/livetree.c struct node *c; node 834 scripts/dtc/livetree.c if (node->labels) { node 837 scripts/dtc/livetree.c for_each_label(node->labels, l) { node 850 scripts/dtc/livetree.c data_copy_escape_string(node->fullpath, node 851 scripts/dtc/livetree.c strlen(node->fullpath)), node 858 scripts/dtc/livetree.c (void)get_node_phandle(dt, node); node 861 scripts/dtc/livetree.c for_each_child(node, c) node 865 scripts/dtc/livetree.c static bool any_fixup_tree(struct dt_info *dti, struct node *node) node 867 scripts/dtc/livetree.c struct node *c; node 871 scripts/dtc/livetree.c for_each_property(node, prop) { node 879 scripts/dtc/livetree.c for_each_child(node, c) { node 887 scripts/dtc/livetree.c static void add_fixup_entry(struct dt_info *dti, struct node *fn, node 888 scripts/dtc/livetree.c struct node *node, struct property *prop, node 897 scripts/dtc/livetree.c if (strchr(node->fullpath, ':') || strchr(prop->name, ':')) node 901 scripts/dtc/livetree.c node->fullpath, prop->name, m->offset); node 908 scripts/dtc/livetree.c struct node *fn, node 909 scripts/dtc/livetree.c struct node *node) node 911 scripts/dtc/livetree.c struct node *dt = dti->dt; node 912 scripts/dtc/livetree.c struct node *c; node 915 scripts/dtc/livetree.c struct node *refnode; node 917 scripts/dtc/livetree.c for_each_property(node, prop) { node 922 scripts/dtc/livetree.c add_fixup_entry(dti, fn, node, prop, m); node 926 scripts/dtc/livetree.c for_each_child(node, c) node 930 scripts/dtc/livetree.c static bool any_local_fixup_tree(struct dt_info *dti, struct node *node) node 932 scripts/dtc/livetree.c struct node *c; node 936 scripts/dtc/livetree.c for_each_property(node, prop) { node 944 scripts/dtc/livetree.c for_each_child(node, c) { node 953 scripts/dtc/livetree.c struct node *lfn, struct node *node, node 955 scripts/dtc/livetree.c struct node *refnode) node 957 scripts/dtc/livetree.c struct node *wn, *nwn; /* local fixup node, walk node, new */ node 964 scripts/dtc/livetree.c for (wn = node; wn; wn = wn->parent) node 971 scripts/dtc/livetree.c for (wn = node, i = depth - 1; wn; wn = wn->parent, i--) node 989 scripts/dtc/livetree.c struct node *lfn, node 990 scripts/dtc/livetree.c struct node *node) node 992 scripts/dtc/livetree.c struct node *dt = dti->dt; node 993 scripts/dtc/livetree.c struct node *c; node 996 scripts/dtc/livetree.c struct node *refnode; node 998 scripts/dtc/livetree.c for_each_property(node, prop) { node 1003 scripts/dtc/livetree.c add_local_fixup_entry(dti, lfn, node, prop, m, refnode); node 1007 scripts/dtc/livetree.c for_each_child(node, c) node 130 scripts/dtc/srcpos.c struct search_path *node; node 140 scripts/dtc/srcpos.c for (node = search_path_head; !*fp && node; node = node->next) node 141 scripts/dtc/srcpos.c fullname = try_open(node->dirname, fname, fp); node 217 scripts/dtc/srcpos.c struct search_path *node; node 220 scripts/dtc/srcpos.c node = xmalloc(sizeof(*node)); node 221 scripts/dtc/srcpos.c node->next = NULL; node 222 scripts/dtc/srcpos.c node->dirname = xstrdup(dirname); node 226 scripts/dtc/srcpos.c *search_path_tail = node; node 228 scripts/dtc/srcpos.c search_path_head = node; node 229 scripts/dtc/srcpos.c search_path_tail = &node->next; node 281 scripts/dtc/treesource.c static void write_tree_source_node(FILE *f, struct node *tree, int level) node 284 scripts/dtc/treesource.c struct node *child; node 176 scripts/dtc/yamltree.c static void yaml_tree(struct node *tree, yaml_emitter_t *emitter) node 179 scripts/dtc/yamltree.c struct node *child; node 162 scripts/gcc-plugins/gcc-common.h #define DECL_NAME_POINTER(node) IDENTIFIER_POINTER(DECL_NAME(node)) node 163 scripts/gcc-plugins/gcc-common.h #define DECL_NAME_LENGTH(node) IDENTIFIER_LENGTH(DECL_NAME(node)) node 164 scripts/gcc-plugins/gcc-common.h #define TYPE_NAME_POINTER(node) IDENTIFIER_POINTER(TYPE_NAME(node)) node 165 scripts/gcc-plugins/gcc-common.h #define TYPE_NAME_LENGTH(node) IDENTIFIER_LENGTH(TYPE_NAME(node)) node 312 scripts/gcc-plugins/gcc-common.h struct cgraph_node *node = cgraph_get_node(decl); node 314 scripts/gcc-plugins/gcc-common.h return node ? node : cgraph_node(decl); node 317 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_function_with_gimple_body_p(struct cgraph_node *node) node 319 scripts/gcc-plugins/gcc-common.h return node->analyzed && !node->thunk.thunk_p && !node->alias; node 324 scripts/gcc-plugins/gcc-common.h struct cgraph_node *node; node 326 scripts/gcc-plugins/gcc-common.h for (node = cgraph_nodes; node; node = node->next) node 327 scripts/gcc-plugins/gcc-common.h if (cgraph_function_with_gimple_body_p(node)) node 328 scripts/gcc-plugins/gcc-common.h return node; node 332 scripts/gcc-plugins/gcc-common.h static inline struct cgraph_node *cgraph_next_function_with_gimple_body(struct cgraph_node *node) node 334 scripts/gcc-plugins/gcc-common.h for (node = node->next; node; node = node->next) node 335 scripts/gcc-plugins/gcc-common.h if (cgraph_function_with_gimple_body_p(node)) node 336 scripts/gcc-plugins/gcc-common.h return node; node 340 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_for_node_and_aliases(cgraph_node_ptr node, bool (*callback)(cgraph_node_ptr, void *), void *data, bool include_overwritable) node 344 scripts/gcc-plugins/gcc-common.h if (callback(node, data)) node 347 scripts/gcc-plugins/gcc-common.h for (alias = node->same_body; alias; alias = alias->next) { node 356 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_FUNCTION_WITH_GIMPLE_BODY(node) \ node 357 scripts/gcc-plugins/gcc-common.h for ((node) = cgraph_first_function_with_gimple_body(); (node); \ node 358 scripts/gcc-plugins/gcc-common.h (node) = cgraph_next_function_with_gimple_body(node)) node 367 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_FUNCTION(node) \ node 368 scripts/gcc-plugins/gcc-common.h for (node = cgraph_nodes; node; node = node->next) node 369 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_VARIABLE(node) \ node 370 scripts/gcc-plugins/gcc-common.h for (node = varpool_nodes; node; node = node->next) node 372 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (node) node 373 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->decl node 411 scripts/gcc-plugins/gcc-common.h #define NODE_IMPLICIT_ALIAS(node) (node)->same_body_alias node 547 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (&(node)->symbol) node 548 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->symbol.decl node 554 scripts/gcc-plugins/gcc-common.h #define varpool_mark_needed_node(node) node 579 scripts/gcc-plugins/gcc-common.h static inline void set_decl_section_name(tree node, const char *value) node 582 scripts/gcc-plugins/gcc-common.h DECL_SECTION_NAME(node) = build_string(strlen(value) + 1, value); node 584 scripts/gcc-plugins/gcc-common.h DECL_SECTION_NAME(node) = NULL; node 681 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (node) node 682 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->decl node 683 scripts/gcc-plugins/gcc-common.h #define cgraph_node_name(node) (node)->name() node 684 scripts/gcc-plugins/gcc-common.h #define NODE_IMPLICIT_ALIAS(node) (node)->cpp_implicit_alias node 716 scripts/gcc-plugins/gcc-common.h #define debug_cgraph_node(node) (node)->debug() node 723 scripts/gcc-plugins/gcc-common.h #define dump_varpool_node(file, node) (node)->dump(file) node 767 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_function_node(cgraph_node_ptr node, enum availability *availability) node 769 scripts/gcc-plugins/gcc-common.h return node->function_symbol(availability); node 772 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_function_or_thunk_node(cgraph_node_ptr node, enum availability *availability = NULL) node 774 scripts/gcc-plugins/gcc-common.h return node->ultimate_alias_target(availability); node 777 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_only_called_directly_p(cgraph_node_ptr node) node 779 scripts/gcc-plugins/gcc-common.h return node->only_called_directly_p(); node 782 scripts/gcc-plugins/gcc-common.h static inline enum availability cgraph_function_body_availability(cgraph_node_ptr node) node 784 scripts/gcc-plugins/gcc-common.h return node->get_availability(); node 787 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_alias_target(cgraph_node_ptr node) node 789 scripts/gcc-plugins/gcc-common.h return node->get_alias_target(); node 792 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_for_node_and_aliases(cgraph_node_ptr node, bool (*callback)(cgraph_node_ptr, void *), void *data, bool include_overwritable) node 794 scripts/gcc-plugins/gcc-common.h return node->call_for_symbol_thunks_and_aliases(callback, data, include_overwritable); node 827 scripts/gcc-plugins/gcc-common.h static inline void cgraph_call_node_duplication_hooks(cgraph_node_ptr node, cgraph_node_ptr node2) node 829 scripts/gcc-plugins/gcc-common.h symtab->call_cgraph_duplication_hooks(node, node2); node 122 scripts/gcc-plugins/latent_entropy_plugin.c static tree handle_latent_entropy_attribute(tree *node, tree name, node 134 scripts/gcc-plugins/latent_entropy_plugin.c switch (TREE_CODE(*node)) { node 142 scripts/gcc-plugins/latent_entropy_plugin.c if (DECL_INITIAL(*node)) { node 145 scripts/gcc-plugins/latent_entropy_plugin.c *node, name); node 149 scripts/gcc-plugins/latent_entropy_plugin.c if (!TREE_STATIC(*node)) { node 152 scripts/gcc-plugins/latent_entropy_plugin.c *node, name); node 156 scripts/gcc-plugins/latent_entropy_plugin.c type = TREE_TYPE(*node); node 161 scripts/gcc-plugins/latent_entropy_plugin.c *node, name); node 177 scripts/gcc-plugins/latent_entropy_plugin.c *node, name, fld); node 198 scripts/gcc-plugins/latent_entropy_plugin.c DECL_INITIAL(*node) = build_constructor(type, vals); node 204 scripts/gcc-plugins/latent_entropy_plugin.c DECL_INITIAL(*node) = tree_get_random_const(type); node 220 scripts/gcc-plugins/latent_entropy_plugin.c *node, name); node 245 scripts/gcc-plugins/latent_entropy_plugin.c DECL_INITIAL(*node) = build_constructor(type, vals); node 485 scripts/gcc-plugins/latent_entropy_plugin.c varpool_node_ptr node; node 490 scripts/gcc-plugins/latent_entropy_plugin.c FOR_EACH_VARIABLE(node) { node 491 scripts/gcc-plugins/latent_entropy_plugin.c tree name, var = NODE_DECL(node); node 26 scripts/gcc-plugins/randomize_layout_plugin.c #define ORIG_TYPE_NAME(node) \ node 27 scripts/gcc-plugins/randomize_layout_plugin.c (TYPE_NAME(TYPE_MAIN_VARIANT(node)) != NULL_TREE ? ((const unsigned char *)IDENTIFIER_POINTER(TYPE_NAME(TYPE_MAIN_VARIANT(node)))) : (const unsigned char *)"anonymous") node 76 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_layout_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs) node 81 scripts/gcc-plugins/randomize_layout_plugin.c if (TREE_CODE(*node) == FUNCTION_DECL) { node 82 scripts/gcc-plugins/randomize_layout_plugin.c error("%qE attribute does not apply to functions (%qF)", name, *node); node 86 scripts/gcc-plugins/randomize_layout_plugin.c if (TREE_CODE(*node) == PARM_DECL) { node 87 scripts/gcc-plugins/randomize_layout_plugin.c error("%qE attribute does not apply to function parameters (%qD)", name, *node); node 91 scripts/gcc-plugins/randomize_layout_plugin.c if (TREE_CODE(*node) == VAR_DECL) { node 92 scripts/gcc-plugins/randomize_layout_plugin.c error("%qE attribute does not apply to variables (%qD)", name, *node); node 96 scripts/gcc-plugins/randomize_layout_plugin.c if (TYPE_P(*node)) { node 97 scripts/gcc-plugins/randomize_layout_plugin.c type = *node; node 99 scripts/gcc-plugins/randomize_layout_plugin.c gcc_assert(TREE_CODE(*node) == TYPE_DECL); node 100 scripts/gcc-plugins/randomize_layout_plugin.c type = TREE_TYPE(*node); node 119 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_considered_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs) node 129 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_performed_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs) node 433 scripts/gcc-plugins/randomize_layout_plugin.c static int is_pure_ops_struct(const_tree node) node 437 scripts/gcc-plugins/randomize_layout_plugin.c gcc_assert(TREE_CODE(node) == RECORD_TYPE || TREE_CODE(node) == UNION_TYPE); node 439 scripts/gcc-plugins/randomize_layout_plugin.c for (field = TYPE_FIELDS(node); field; field = TREE_CHAIN(field)) { node 443 scripts/gcc-plugins/randomize_layout_plugin.c if (node == fieldtype) node 662 scripts/gcc-plugins/randomize_layout_plugin.c struct varpool_node *node; node 665 scripts/gcc-plugins/randomize_layout_plugin.c FOR_EACH_VARIABLE(node) { node 666 scripts/gcc-plugins/randomize_layout_plugin.c tree var = NODE_DECL(node); node 53 scripts/gcc-plugins/stackleak_plugin.c cgraph_node_ptr node; node 69 scripts/gcc-plugins/stackleak_plugin.c node = cgraph_get_create_node(track_function_decl); node 70 scripts/gcc-plugins/stackleak_plugin.c gcc_assert(node); node 72 scripts/gcc-plugins/stackleak_plugin.c cgraph_create_edge(cgraph_get_node(current_function_decl), node, node 53 scripts/gcc-plugins/structleak_plugin.c static tree handle_user_attribute(tree *node, tree name, tree args, int flags, bool *no_add_attrs) node 58 scripts/gcc-plugins/structleak_plugin.c if (TREE_CODE(*node) != FIELD_DECL) node 323 scripts/genksyms/genksyms.c void free_node(struct string_list *node) node 325 scripts/genksyms/genksyms.c free(node->string); node 326 scripts/genksyms/genksyms.c free(node); node 367 scripts/genksyms/genksyms.c struct string_list *copy_node(struct string_list *node) node 372 scripts/genksyms/genksyms.c newnode->string = xstrdup(node->string); node 373 scripts/genksyms/genksyms.c newnode->tag = node->tag; node 411 scripts/genksyms/genksyms.c struct string_list node = { node 418 scripts/genksyms/genksyms.c if (node.string == buffer) node 424 scripts/genksyms/genksyms.c if (node.string == buffer) node 429 scripts/genksyms/genksyms.c if (node.string >= buffer + sizeof(buffer) - 1) { node 433 scripts/genksyms/genksyms.c *node.string++ = c; node 435 scripts/genksyms/genksyms.c if (node.string == buffer) node 437 scripts/genksyms/genksyms.c *node.string = 0; node 438 scripts/genksyms/genksyms.c node.string = buffer; node 440 scripts/genksyms/genksyms.c if (node.string[1] == '#') { node 444 scripts/genksyms/genksyms.c if (node.string[0] == symbol_types[n].n) { node 445 scripts/genksyms/genksyms.c node.tag = n; node 446 scripts/genksyms/genksyms.c node.string += 2; node 447 scripts/genksyms/genksyms.c return copy_node(&node); node 450 scripts/genksyms/genksyms.c fprintf(stderr, "Unknown type %c\n", node.string[0]); node 453 scripts/genksyms/genksyms.c return copy_node(&node); node 29 scripts/genksyms/parse.y struct string_list *node = *p; node 30 scripts/genksyms/parse.y *p = node->next; node 31 scripts/genksyms/parse.y free_node(node); node 73 scripts/kconfig/gconf.c static void set_node(GtkTreeIter * node, struct menu *menu, gchar ** row); node 1161 scripts/kconfig/gconf.c static void set_node(GtkTreeIter * node, struct menu *menu, gchar ** row) node 1174 scripts/kconfig/gconf.c gtk_tree_store_set(tree, node, node 1200 scripts/kconfig/gconf.c GtkTreeIter *node = parents[indent]; node 1202 scripts/kconfig/gconf.c gtk_tree_store_append(tree, node, parent); node 1203 scripts/kconfig/gconf.c set_node(node, menu, row); node 162 scripts/kconfig/lxdialog/dialog.h struct dialog_item node; node 606 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str, sizeof(item_cur->node.str), fmt, ap); node 615 scripts/kconfig/lxdialog/util.c avail = sizeof(item_cur->node.str) - strlen(item_cur->node.str); node 618 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str + strlen(item_cur->node.str), node 620 scripts/kconfig/lxdialog/util.c item_cur->node.str[sizeof(item_cur->node.str) - 1] = '\0'; node 626 scripts/kconfig/lxdialog/util.c item_cur->node.tag = tag; node 630 scripts/kconfig/lxdialog/util.c item_cur->node.data = ptr; node 635 scripts/kconfig/lxdialog/util.c item_cur->node.selected = val; node 648 scripts/kconfig/lxdialog/util.c return item_cur->node.data; node 653 scripts/kconfig/lxdialog/util.c return item_cur->node.tag; node 689 scripts/kconfig/lxdialog/util.c return item_cur->node.str; node 694 scripts/kconfig/lxdialog/util.c return (item_cur->node.selected != 0); node 699 scripts/kconfig/lxdialog/util.c return (item_cur->node.tag == tag); node 41 scripts/kconfig/preprocess.c struct list_head node; node 52 scripts/kconfig/preprocess.c list_add_tail(&e->node, &env_list); node 57 scripts/kconfig/preprocess.c list_del(&e->node); node 72 scripts/kconfig/preprocess.c list_for_each_entry(e, &env_list, node) { node 94 scripts/kconfig/preprocess.c list_for_each_entry_safe(e, tmp, &env_list, node) { node 236 scripts/kconfig/preprocess.c struct list_head node; node 243 scripts/kconfig/preprocess.c list_for_each_entry(v, &variable_list, node) { node 303 scripts/kconfig/preprocess.c list_add_tail(&v->node, &variable_list); node 326 scripts/kconfig/preprocess.c list_del(&v->node); node 336 scripts/kconfig/preprocess.c list_for_each_entry_safe(v, tmp, &variable_list, node) node 127 security/apparmor/include/label.h struct rb_node node; node 412 security/apparmor/label.c RB_CLEAR_NODE(&label->node); node 578 security/apparmor/label.c rb_erase(&label->node, &ls->root); node 613 security/apparmor/label.c rb_replace_node(&old->node, &new->node, &ls->root); node 649 security/apparmor/label.c struct aa_label *this = rb_entry(*new, struct aa_label, node); node 673 security/apparmor/label.c rb_link_node(&label->node, parent, new); node 674 security/apparmor/label.c rb_insert_color(&label->node, &ls->root); node 694 security/apparmor/label.c struct rb_node *node; node 700 security/apparmor/label.c node = vec_labelset(vec, n)->root.rb_node; node 701 security/apparmor/label.c while (node) { node 702 security/apparmor/label.c struct aa_label *this = rb_entry(node, struct aa_label, node); node 706 security/apparmor/label.c node = node->rb_left; node 708 security/apparmor/label.c node = node->rb_right; node 1116 security/apparmor/label.c struct rb_node *node; node 1125 security/apparmor/label.c node = ls->root.rb_node; node 1126 security/apparmor/label.c while (node) { node 1127 security/apparmor/label.c struct aa_label *this = container_of(node, struct aa_label, node 1128 security/apparmor/label.c node); node 1132 security/apparmor/label.c node = node->rb_left; node 1134 security/apparmor/label.c node = node->rb_right; node 1954 security/apparmor/label.c struct rb_node *node; node 1960 security/apparmor/label.c for (node = rb_first(&ls->root); node; node = rb_first(&ls->root)) { node 1961 security/apparmor/label.c struct aa_label *this = rb_entry(node, struct aa_label, node); node 1986 security/apparmor/label.c struct rb_node *node; node 1993 security/apparmor/label.c __labelset_for_each(ls, node) { node 1994 security/apparmor/label.c label = rb_entry(node, struct aa_label, node); node 98 security/integrity/iint.c struct rb_node *node, *parent = NULL; node 123 security/integrity/iint.c node = &iint->rb_node; node 125 security/integrity/iint.c rb_link_node(node, parent, p); node 126 security/integrity/iint.c rb_insert_color(node, &integrity_iint_tree); node 54 security/keys/internal.h struct rb_node node; node 63 security/keys/key.c user = rb_entry(parent, struct key_user, node); node 101 security/keys/key.c rb_link_node(&candidate->node, parent, p); node 102 security/keys/key.c rb_insert_color(&candidate->node, &key_user_tree); node 122 security/keys/key.c rb_erase(&user->node, &key_user_tree); node 1190 security/keys/key.c rb_link_node(&root_key_user.node, node 1194 security/keys/key.c rb_insert_color(&root_key_user.node, node 668 security/keys/keyring.c struct assoc_array_node *node; node 673 security/keys/keyring.c struct assoc_array_node *node; node 746 security/keys/keyring.c node = assoc_array_ptr_to_node(ptr); node 750 security/keys/keyring.c node = assoc_array_ptr_to_node(ptr); node 751 security/keys/keyring.c ptr = node->slots[0]; node 765 security/keys/keyring.c node = assoc_array_ptr_to_node(ptr); node 773 security/keys/keyring.c ptr = READ_ONCE(node->slots[slot]); node 775 security/keys/keyring.c if (assoc_array_ptr_is_meta(ptr) && node->back_pointer) node 799 security/keys/keyring.c stack[sp].node = node; node 811 security/keys/keyring.c ptr = READ_ONCE(node->back_pointer); node 812 security/keys/keyring.c slot = node->parent_slot; node 821 security/keys/keyring.c node = assoc_array_ptr_to_node(ptr); node 828 security/keys/keyring.c if (node->back_pointer) { node 846 security/keys/keyring.c node = stack[sp].node; node 255 security/keys/proc.c struct key_user *user = rb_entry(n, struct key_user, node); node 306 security/keys/proc.c struct key_user *user = rb_entry(_p, struct key_user, node); node 153 security/selinux/avc.c struct avc_node *node; node 165 security/selinux/avc.c hlist_for_each_entry_rcu(node, head, list) node 326 security/selinux/avc.c static int avc_add_xperms_decision(struct avc_node *node, node 331 security/selinux/avc.c node->ae.xp_node->xp.len++; node 336 security/selinux/avc.c list_add(&dest_xpd->xpd_list, &node->ae.xp_node->xpd_head); node 351 security/selinux/avc.c static int avc_xperms_populate(struct avc_node *node, node 375 security/selinux/avc.c node->ae.xp_node = dest; node 432 security/selinux/avc.c struct avc_node *node = container_of(rhead, struct avc_node, rhead); node 433 security/selinux/avc.c avc_xperms_free(node->ae.xp_node); node 434 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node); node 438 security/selinux/avc.c static void avc_node_delete(struct selinux_avc *avc, struct avc_node *node) node 440 security/selinux/avc.c hlist_del_rcu(&node->list); node 441 security/selinux/avc.c call_rcu(&node->rhead, avc_node_free); node 445 security/selinux/avc.c static void avc_node_kill(struct selinux_avc *avc, struct avc_node *node) node 447 security/selinux/avc.c avc_xperms_free(node->ae.xp_node); node 448 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node); node 463 security/selinux/avc.c struct avc_node *node; node 479 security/selinux/avc.c hlist_for_each_entry(node, head, list) { node 480 security/selinux/avc.c avc_node_delete(avc, node); node 498 security/selinux/avc.c struct avc_node *node; node 500 security/selinux/avc.c node = kmem_cache_zalloc(avc_node_cachep, GFP_NOWAIT); node 501 security/selinux/avc.c if (!node) node 504 security/selinux/avc.c INIT_HLIST_NODE(&node->list); node 512 security/selinux/avc.c return node; node 515 security/selinux/avc.c static void avc_node_populate(struct avc_node *node, u32 ssid, u32 tsid, u16 tclass, struct av_decision *avd) node 517 security/selinux/avc.c node->ae.ssid = ssid; node 518 security/selinux/avc.c node->ae.tsid = tsid; node 519 security/selinux/avc.c node->ae.tclass = tclass; node 520 security/selinux/avc.c memcpy(&node->ae.avd, avd, sizeof(node->ae.avd)); node 526 security/selinux/avc.c struct avc_node *node, *ret = NULL; node 532 security/selinux/avc.c hlist_for_each_entry_rcu(node, head, list) { node 533 security/selinux/avc.c if (ssid == node->ae.ssid && node 534 security/selinux/avc.c tclass == node->ae.tclass && node 535 security/selinux/avc.c tsid == node->ae.tsid) { node 536 security/selinux/avc.c ret = node; node 559 security/selinux/avc.c struct avc_node *node; node 562 security/selinux/avc.c node = avc_search_node(avc, ssid, tsid, tclass); node 564 security/selinux/avc.c if (node) node 565 security/selinux/avc.c return node; node 617 security/selinux/avc.c struct avc_node *pos, *node = NULL; node 626 security/selinux/avc.c node = avc_alloc_node(avc); node 627 security/selinux/avc.c if (!node) node 630 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, avd); node 631 security/selinux/avc.c if (avc_xperms_populate(node, xp_node)) { node 632 security/selinux/avc.c avc_node_kill(avc, node); node 644 security/selinux/avc.c avc_node_replace(avc, node, pos); node 648 security/selinux/avc.c hlist_add_head_rcu(&node->list, head); node 651 security/selinux/avc.c return node; node 836 security/selinux/avc.c struct avc_node *pos, *node, *orig = NULL; node 855 security/selinux/avc.c node = avc_alloc_node(avc); node 856 security/selinux/avc.c if (!node) { node 881 security/selinux/avc.c avc_node_kill(avc, node); node 889 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, &orig->ae.avd); node 892 security/selinux/avc.c rc = avc_xperms_populate(node, orig->ae.xp_node); node 894 security/selinux/avc.c avc_node_kill(avc, node); node 901 security/selinux/avc.c node->ae.avd.allowed |= perms; node 902 security/selinux/avc.c if (node->ae.xp_node && (flags & AVC_EXTENDED_PERMS)) node 903 security/selinux/avc.c avc_xperms_allow_perm(node->ae.xp_node, driver, xperm); node 907 security/selinux/avc.c node->ae.avd.allowed &= ~perms; node 910 security/selinux/avc.c node->ae.avd.auditallow |= perms; node 913 security/selinux/avc.c node->ae.avd.auditallow &= ~perms; node 916 security/selinux/avc.c node->ae.avd.auditdeny |= perms; node 919 security/selinux/avc.c node->ae.avd.auditdeny &= ~perms; node 922 security/selinux/avc.c avc_add_xperms_decision(node, xpd); node 925 security/selinux/avc.c avc_node_replace(avc, node, orig); node 938 security/selinux/avc.c struct avc_node *node; node 953 security/selinux/avc.c hlist_for_each_entry(node, head, list) node 954 security/selinux/avc.c avc_node_delete(avc, node); node 1036 security/selinux/avc.c struct avc_node *node; node 1054 security/selinux/avc.c node = avc_lookup(state->avc, ssid, tsid, tclass); node 1055 security/selinux/avc.c if (unlikely(!node)) { node 1056 security/selinux/avc.c node = avc_compute_av(state, ssid, tsid, tclass, &avd, xp_node); node 1058 security/selinux/avc.c memcpy(&avd, &node->ae.avd, sizeof(avd)); node 1059 security/selinux/avc.c xp_node = node->ae.xp_node; node 1135 security/selinux/avc.c struct avc_node *node; node 1145 security/selinux/avc.c node = avc_lookup(state->avc, ssid, tsid, tclass); node 1146 security/selinux/avc.c if (unlikely(!node)) node 1147 security/selinux/avc.c node = avc_compute_av(state, ssid, tsid, tclass, avd, &xp_node); node 1149 security/selinux/avc.c memcpy(avd, &node->ae.avd, sizeof(*avd)); node 107 security/selinux/netnode.c struct sel_netnode *node; node 121 security/selinux/netnode.c list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list) node 122 security/selinux/netnode.c if (node->nsec.family == family) node 125 security/selinux/netnode.c if (node->nsec.addr.ipv4 == *(__be32 *)addr) node 126 security/selinux/netnode.c return node; node 129 security/selinux/netnode.c if (ipv6_addr_equal(&node->nsec.addr.ipv6, node 131 security/selinux/netnode.c return node; node 146 security/selinux/netnode.c static void sel_netnode_insert(struct sel_netnode *node) node 150 security/selinux/netnode.c switch (node->nsec.family) { node 152 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4); node 155 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6); node 164 security/selinux/netnode.c list_add_rcu(&node->list, &sel_netnode_hash[idx].list); node 193 security/selinux/netnode.c struct sel_netnode *node; node 197 security/selinux/netnode.c node = sel_netnode_find(addr, family); node 198 security/selinux/netnode.c if (node != NULL) { node 199 security/selinux/netnode.c *sid = node->nsec.sid; node 252 security/selinux/netnode.c struct sel_netnode *node; node 255 security/selinux/netnode.c node = sel_netnode_find(addr, family); node 256 security/selinux/netnode.c if (node != NULL) { node 257 security/selinux/netnode.c *sid = node->nsec.sid; node 276 security/selinux/netnode.c struct sel_netnode *node, *node_tmp; node 280 security/selinux/netnode.c list_for_each_entry_safe(node, node_tmp, node 282 security/selinux/netnode.c list_del_rcu(&node->list); node 283 security/selinux/netnode.c kfree_rcu(node, rcu); node 249 security/selinux/ss/avtab.c avtab_search_node_next(struct avtab_node *node, int specified) node 253 security/selinux/ss/avtab.c if (!node) node 257 security/selinux/ss/avtab.c for (cur = node->next; cur; cur = cur->next) { node 258 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 259 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && node 260 security/selinux/ss/avtab.c node->key.target_class == cur->key.target_class && node 264 security/selinux/ss/avtab.c if (node->key.source_type < cur->key.source_type) node 266 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 267 security/selinux/ss/avtab.c node->key.target_type < cur->key.target_type) node 269 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 270 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && node 271 security/selinux/ss/avtab.c node->key.target_class < cur->key.target_class) node 111 security/selinux/ss/avtab.h struct avtab_node *avtab_search_node_next(struct avtab_node *node, int specified); node 88 security/selinux/ss/conditional.c int evaluate_cond_node(struct policydb *p, struct cond_node *node) node 93 security/selinux/ss/conditional.c new_state = cond_evaluate_expr(p, node->expr); node 94 security/selinux/ss/conditional.c if (new_state != node->cur_state) { node 95 security/selinux/ss/conditional.c node->cur_state = new_state; node 99 security/selinux/ss/conditional.c for (cur = node->true_list; cur; cur = cur->next) { node 101 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; node 103 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; node 106 security/selinux/ss/conditional.c for (cur = node->false_list; cur; cur = cur->next) { node 109 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; node 111 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; node 141 security/selinux/ss/conditional.c static void cond_node_destroy(struct cond_node *node) node 145 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr; cur_expr = next_expr) { node 149 security/selinux/ss/conditional.c cond_av_list_destroy(node->true_list); node 150 security/selinux/ss/conditional.c cond_av_list_destroy(node->false_list); node 151 security/selinux/ss/conditional.c kfree(node); node 304 security/selinux/ss/conditional.c if (cur->node == node_ptr) { node 335 security/selinux/ss/conditional.c list->node = node_ptr; node 395 security/selinux/ss/conditional.c static int cond_read_node(struct policydb *p, struct cond_node *node, void *fp) node 406 security/selinux/ss/conditional.c node->cur_state = le32_to_cpu(buf[0]); node 431 security/selinux/ss/conditional.c node->expr = expr; node 437 security/selinux/ss/conditional.c rc = cond_read_av_list(p, fp, &node->true_list, NULL); node 440 security/selinux/ss/conditional.c rc = cond_read_av_list(p, fp, &node->false_list, node->true_list); node 445 security/selinux/ss/conditional.c cond_node_destroy(node); node 451 security/selinux/ss/conditional.c struct cond_node *node, *last = NULL; node 468 security/selinux/ss/conditional.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 469 security/selinux/ss/conditional.c if (!node) node 472 security/selinux/ss/conditional.c rc = cond_read_node(p, node, fp); node 477 security/selinux/ss/conditional.c p->cond_list = node; node 479 security/selinux/ss/conditional.c last->next = node; node 480 security/selinux/ss/conditional.c last = node; node 542 security/selinux/ss/conditional.c rc = avtab_write_item(p, cur_list->node, fp); node 550 security/selinux/ss/conditional.c static int cond_write_node(struct policydb *p, struct cond_node *node, node 558 security/selinux/ss/conditional.c buf[0] = cpu_to_le32(node->cur_state); node 563 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) node 571 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) { node 579 security/selinux/ss/conditional.c rc = cond_write_av_list(p, node->true_list, fp); node 582 security/selinux/ss/conditional.c rc = cond_write_av_list(p, node->false_list, fp); node 616 security/selinux/ss/conditional.c struct avtab_node *node; node 621 security/selinux/ss/conditional.c for (node = avtab_search_node(ctab, key); node; node 622 security/selinux/ss/conditional.c node = avtab_search_node_next(node, key->specified)) { node 623 security/selinux/ss/conditional.c if (node->key.specified & AVTAB_ENABLED) node 624 security/selinux/ss/conditional.c services_compute_xperms_decision(xpermd, node); node 635 security/selinux/ss/conditional.c struct avtab_node *node; node 640 security/selinux/ss/conditional.c for (node = avtab_search_node(ctab, key); node; node 641 security/selinux/ss/conditional.c node = avtab_search_node_next(node, key->specified)) { node 643 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_ALLOWED|AVTAB_ENABLED))) node 644 security/selinux/ss/conditional.c avd->allowed |= node->datum.u.data; node 646 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITDENY|AVTAB_ENABLED))) node 652 security/selinux/ss/conditional.c avd->auditdeny &= node->datum.u.data; node 654 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITALLOW|AVTAB_ENABLED))) node 655 security/selinux/ss/conditional.c avd->auditallow |= node->datum.u.data; node 656 security/selinux/ss/conditional.c if (xperms && (node->key.specified & AVTAB_ENABLED) && node 657 security/selinux/ss/conditional.c (node->key.specified & AVTAB_XPERMS)) node 658 security/selinux/ss/conditional.c services_compute_xperms_drivers(xperms, node); node 42 security/selinux/ss/conditional.h struct avtab_node *node; node 78 security/selinux/ss/conditional.h int evaluate_cond_node(struct policydb *p, struct cond_node *node); node 37 security/selinux/ss/ebitmap.c n1 = e1->node; node 38 security/selinux/ss/ebitmap.c n2 = e2->node; node 57 security/selinux/ss/ebitmap.c n = src->node; node 71 security/selinux/ss/ebitmap.c dst->node = new; node 94 security/selinux/ss/ebitmap.c struct ebitmap_node *e_iter = ebmap->node; node 173 security/selinux/ss/ebitmap.c ebmap->node = e_iter; node 209 security/selinux/ss/ebitmap.c n1 = e1->node; node 210 security/selinux/ss/ebitmap.c n2 = e2->node; node 249 security/selinux/ss/ebitmap.c n = e->node; node 264 security/selinux/ss/ebitmap.c n = e->node; node 293 security/selinux/ss/ebitmap.c e->node = n->next; node 320 security/selinux/ss/ebitmap.c new->next = e->node; node 321 security/selinux/ss/ebitmap.c e->node = new; node 334 security/selinux/ss/ebitmap.c n = e->node; node 342 security/selinux/ss/ebitmap.c e->node = NULL; node 378 security/selinux/ss/ebitmap.c e->node = NULL; node 419 security/selinux/ss/ebitmap.c e->node = tmp; node 41 security/selinux/ss/ebitmap.h struct ebitmap_node *node; /* first node in the bitmap */ node 52 security/selinux/ss/ebitmap.h for (*n = e->node; *n; *n = (*n)->next) { node 83 security/selinux/ss/ebitmap.h #define EBITMAP_NODE_INDEX(node, bit) \ node 84 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) / EBITMAP_UNIT_SIZE) node 85 security/selinux/ss/ebitmap.h #define EBITMAP_NODE_OFFSET(node, bit) \ node 86 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) % EBITMAP_UNIT_SIZE) node 41 security/selinux/ss/mls.c struct ebitmap_node *node; node 55 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) { node 96 security/selinux/ss/mls.c struct ebitmap_node *node; node 115 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) { node 454 security/selinux/ss/mls.c struct ebitmap_node *node; node 470 security/selinux/ss/mls.c node, i) { node 1648 security/selinux/ss/policydb.c struct ebitmap_node *node; node 1659 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&user->roles, node, bit) { node 1684 security/selinux/ss/policydb.c struct ebitmap_node *node; node 1695 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&role->types, node, bit) { node 2140 security/selinux/ss/policydb.c c->u.node.addr = nodebuf[0]; /* network order */ node 2141 security/selinux/ss/policydb.c c->u.node.mask = nodebuf[1]; /* network order */ node 2806 security/selinux/ss/policydb.c static int write_cons_helper(struct policydb *p, struct constraint_node *node, node 2815 security/selinux/ss/policydb.c for (c = node; c; c = c->next) { node 3132 security/selinux/ss/policydb.c nodebuf[0] = c->u.node.addr; /* network order */ node 3133 security/selinux/ss/policydb.c nodebuf[1] = c->u.node.mask; /* network order */ node 181 security/selinux/ss/policydb.h } node; /* node information */ node 590 security/selinux/ss/services.c struct avtab_node *node) node 594 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) { node 597 security/selinux/ss/services.c xperms->drivers.p[i] |= node->datum.u.xperms->perms.p[i]; node 598 security/selinux/ss/services.c } else if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) { node 601 security/selinux/ss/services.c node->datum.u.xperms->driver); node 605 security/selinux/ss/services.c if (node->key.specified & AVTAB_XPERMS_ALLOWED) node 623 security/selinux/ss/services.c struct avtab_node *node; node 657 security/selinux/ss/services.c for (node = avtab_search_node(&policydb->te_avtab, node 659 security/selinux/ss/services.c node; node 660 security/selinux/ss/services.c node = avtab_search_node_next(node, avkey.specified)) { node 661 security/selinux/ss/services.c if (node->key.specified == AVTAB_ALLOWED) node 662 security/selinux/ss/services.c avd->allowed |= node->datum.u.data; node 663 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITALLOW) node 664 security/selinux/ss/services.c avd->auditallow |= node->datum.u.data; node 665 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITDENY) node 666 security/selinux/ss/services.c avd->auditdeny &= node->datum.u.data; node 667 security/selinux/ss/services.c else if (xperms && (node->key.specified & AVTAB_XPERMS)) node 668 security/selinux/ss/services.c services_compute_xperms_drivers(xperms, node); node 945 security/selinux/ss/services.c struct avtab_node *node) node 949 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) { node 950 security/selinux/ss/services.c if (xpermd->driver != node->datum.u.xperms->driver) node 952 security/selinux/ss/services.c } else if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) { node 953 security/selinux/ss/services.c if (!security_xperm_test(node->datum.u.xperms->perms.p, node 960 security/selinux/ss/services.c if (node->key.specified == AVTAB_XPERMS_ALLOWED) { node 962 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) { node 966 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) { node 969 security/selinux/ss/services.c node->datum.u.xperms->perms.p[i]; node 971 security/selinux/ss/services.c } else if (node->key.specified == AVTAB_XPERMS_AUDITALLOW) { node 973 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) { node 977 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) { node 980 security/selinux/ss/services.c node->datum.u.xperms->perms.p[i]; node 982 security/selinux/ss/services.c } else if (node->key.specified == AVTAB_XPERMS_DONTAUDIT) { node 984 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) { node 988 security/selinux/ss/services.c if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) { node 991 security/selinux/ss/services.c node->datum.u.xperms->perms.p[i]; node 1010 security/selinux/ss/services.c struct avtab_node *node; node 1063 security/selinux/ss/services.c for (node = avtab_search_node(&policydb->te_avtab, node 1065 security/selinux/ss/services.c node; node 1066 security/selinux/ss/services.c node = avtab_search_node_next(node, avkey.specified)) node 1067 security/selinux/ss/services.c services_compute_xperms_decision(xpermd, node); node 1652 security/selinux/ss/services.c struct avtab_node *node; node 1757 security/selinux/ss/services.c node = avtab_search_node(&policydb->te_cond_avtab, &avkey); node 1758 security/selinux/ss/services.c for (; node; node = avtab_search_node_next(node, specified)) { node 1759 security/selinux/ss/services.c if (node->key.specified & AVTAB_ENABLED) { node 1760 security/selinux/ss/services.c avdatum = &node->datum; node 2054 security/selinux/ss/services.c struct ebitmap_node *node; node 2064 security/selinux/ss/services.c ebitmap_for_each_positive_bit(&p->policycaps, node, i) { node 2487 security/selinux/ss/services.c if (c->u.node.addr == (addr & c->u.node.mask)) node 37 security/selinux/ss/services.h struct avtab_node *node); node 40 security/selinux/ss/services.h struct avtab_node *node); node 181 security/selinux/ss/sidtab.c struct sidtab_node_inner *node = entry.ptr_inner; node 185 security/selinux/ss/sidtab.c rc = sidtab_find_context(node->entries[i], node 193 security/selinux/ss/sidtab.c struct sidtab_node_leaf *node = entry.ptr_leaf; node 197 security/selinux/ss/sidtab.c if (context_cmp(&node->entries[i].context, context)) { node 461 security/selinux/ss/sidtab.c struct sidtab_node_inner *node = entry.ptr_inner; node 463 security/selinux/ss/sidtab.c if (!node) node 467 security/selinux/ss/sidtab.c sidtab_destroy_tree(node->entries[i], level - 1); node 468 security/selinux/ss/sidtab.c kfree(node); node 470 security/selinux/ss/sidtab.c struct sidtab_node_leaf *node = entry.ptr_leaf; node 472 security/selinux/ss/sidtab.c if (!node) node 476 security/selinux/ss/sidtab.c context_destroy(&node->entries[i].context); node 477 security/selinux/ss/sidtab.c kfree(node); node 34 security/yama/yama_lsm.c struct list_head node; node 121 security/yama/yama_lsm.c list_for_each_entry_rcu(relation, &ptracer_relations, node) { node 123 security/yama/yama_lsm.c list_del_rcu(&relation->node); node 156 security/yama/yama_lsm.c list_for_each_entry_rcu(relation, &ptracer_relations, node) { node 160 security/yama/yama_lsm.c list_replace_rcu(&relation->node, &added->node); node 166 security/yama/yama_lsm.c list_add_rcu(&added->node, &ptracer_relations); node 186 security/yama/yama_lsm.c list_for_each_entry_rcu(relation, &ptracer_relations, node) { node 328 security/yama/yama_lsm.c list_for_each_entry_rcu(relation, &ptracer_relations, node) { node 73 sound/ac97/bus.c struct device_node *node; node 80 sound/ac97/bus.c for_each_child_of_node(ac97_ctrl->parent->of_node, node) { node 81 sound/ac97/bus.c if ((idx != of_property_read_u32(node, "reg", ®)) || node 82 sound/ac97/bus.c !of_device_is_compatible(node, compat)) node 84 sound/ac97/bus.c return node; node 72 sound/aoa/aoa-gpio.h struct device_node *node; node 35 sound/aoa/aoa.h struct device_node *node; node 996 sound/aoa/codecs/onyx.c struct device_node *node = client->dev.of_node; node 1020 sound/aoa/codecs/onyx.c onyx->codec.node = of_node_get(node); node 1037 sound/aoa/codecs/onyx.c of_node_put(onyx->codec.node); node 882 sound/aoa/codecs/tas.c struct device_node *node = client->dev.of_node; node 901 sound/aoa/codecs/tas.c tas->codec.node = of_node_get(node); node 908 sound/aoa/codecs/tas.c (unsigned int)client->addr, node); node 922 sound/aoa/codecs/tas.c of_node_put(tas->codec.node); node 20 sound/aoa/core/gpio-pmf.c rc = pmf_call_function(rt->node, #name "-mute", &args); \ node 43 sound/aoa/core/gpio-pmf.c rc = pmf_call_function(rt->node, "hw-reset", &args); node 188 sound/aoa/core/gpio-pmf.c err = pmf_register_irq_client(rt->node, node 229 sound/aoa/core/gpio-pmf.c err = pmf_call_function(rt->node, name, &args); node 777 sound/aoa/fabrics/layout.c if (of_node_name_eq(codec->node, "codec")) { node 786 sound/aoa/fabrics/layout.c if (*ref != codec->node->phandle) { node 1038 sound/aoa/fabrics/layout.c ldev->gpio.node = sound->parent; node 2137 sound/pci/asihpi/hpifunc.c u32 node, index; node 2139 sound/pci/asihpi/hpifunc.c HPI_MULTIPLEXER_SOURCE, &node, node 2142 sound/pci/asihpi/hpifunc.c *source_node_type = (u16)node; node 820 sound/ppc/pmac.c ppc_md.feature_call(PMAC_FTR_SOUND_CHIP_ENABLE, chip->node, 0, enable); node 836 sound/ppc/pmac.c if (chip->node) node 862 sound/ppc/pmac.c if (chip->node) { node 872 sound/ppc/pmac.c of_node_put(chip->node); node 897 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) { node 949 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "awacs"); node 950 sound/ppc/pmac.c sound = of_node_get(chip->node); node 956 sound/ppc/pmac.c if (!chip->node) node 957 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "davbus"); node 962 sound/ppc/pmac.c if (! chip->node) { node 963 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "i2s-a"); node 964 sound/ppc/pmac.c if (chip->node && chip->node->parent && node 965 sound/ppc/pmac.c chip->node->parent->parent) { node 966 sound/ppc/pmac.c if (of_device_is_compatible(chip->node->parent->parent, node 971 sound/ppc/pmac.c if (! chip->node) node 976 sound/ppc/pmac.c if (sound->parent == chip->node) node 980 sound/ppc/pmac.c of_node_put(chip->node); node 981 sound/ppc/pmac.c chip->node = NULL; node 995 sound/ppc/pmac.c of_node_put(chip->node); node 996 sound/ppc/pmac.c chip->node = NULL; node 1046 sound/ppc/pmac.c macio = macio_find(chip->node, macio_unknown); node 1198 sound/ppc/pmac.c np = chip->node; node 1302 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) { node 82 sound/ppc/pmac.h struct device_node *node; node 1086 sound/ppc/tumbler.c struct device_node *node; node 1092 sound/ppc/tumbler.c node = find_compatible_audio_device(device); node 1094 sound/ppc/tumbler.c node = find_audio_device(device); node 1095 sound/ppc/tumbler.c if (! node) { node 1101 sound/ppc/tumbler.c base = of_get_property(node, "AAPL,address", NULL); node 1103 sound/ppc/tumbler.c base = of_get_property(node, "reg", NULL); node 1107 sound/ppc/tumbler.c of_node_put(node); node 1118 sound/ppc/tumbler.c base = of_get_property(node, "audio-gpio-active-state", NULL); node 1134 sound/ppc/tumbler.c prop = of_get_property(node, platform, NULL); node 1150 sound/ppc/tumbler.c ret = irq_of_parse_and_map(node, 0); node 1151 sound/ppc/tumbler.c of_node_put(node); node 1354 sound/ppc/tumbler.c for_each_child_of_node(chip->node, np) { node 311 sound/soc/codecs/max98504.c struct device_node *node = dev->of_node; node 319 sound/soc/codecs/max98504.c if (node) { node 320 sound/soc/codecs/max98504.c if (!of_property_read_u32(node, "maxim,brownout-threshold", node 324 sound/soc/codecs/max98504.c of_property_read_u32(node, "maxim,brownout-attenuation", node 326 sound/soc/codecs/max98504.c of_property_read_u32(node, "maxim,brownout-attack-hold-ms", node 328 sound/soc/codecs/max98504.c of_property_read_u32(node, "maxim,brownout-timed-hold-ms", node 330 sound/soc/codecs/max98504.c of_property_read_u32(node, "maxim,brownout-release-rate-ms", node 197 sound/soc/codecs/twl4030.c struct device_node *node) node 201 sound/soc/codecs/twl4030.c of_property_read_u32(node, "ti,digimic_delay", node 203 sound/soc/codecs/twl4030.c of_property_read_u32(node, "ti,ramp_delay_value", node 205 sound/soc/codecs/twl4030.c of_property_read_u32(node, "ti,offset_cncl_path", node 207 sound/soc/codecs/twl4030.c if (!of_property_read_u32(node, "ti,hs_extmute", &value)) node 210 sound/soc/codecs/twl4030.c pdata->hs_extmute_gpio = of_get_named_gpio(node, node 62 sound/soc/generic/audio-graph-card.c struct device_node *node; node 86 sound/soc/generic/audio-graph-card.c node = of_get_parent(ep); node 87 sound/soc/generic/audio-graph-card.c reg = of_get_property(node, "reg", NULL); node 88 sound/soc/generic/audio-graph-card.c of_node_put(node); node 92 sound/soc/generic/audio-graph-card.c node = of_graph_get_port_parent(ep); node 100 sound/soc/generic/audio-graph-card.c for_each_endpoint_of_node(node, endpoint) { node 106 sound/soc/generic/audio-graph-card.c of_node_put(node); node 118 sound/soc/generic/audio-graph-card.c struct device_node *node; node 125 sound/soc/generic/audio-graph-card.c node = of_graph_get_port_parent(ep); node 128 sound/soc/generic/audio-graph-card.c args.np = node; node 130 sound/soc/generic/audio-graph-card.c args.args_count = (of_graph_get_endpoint_count(node) > 1); node 155 sound/soc/generic/audio-graph-card.c dlc->of_node = node; node 158 sound/soc/generic/audio-graph-card.c *is_single_link = of_graph_get_endpoint_count(node) == 1; node 170 sound/soc/generic/audio-graph-card.c struct device_node *node = of_graph_get_port_parent(ep); node 173 sound/soc/generic/audio-graph-card.c asoc_simple_parse_convert(dev, node, PREFIX, adata); node 180 sound/soc/generic/audio-graph-card.c of_node_put(node); node 189 sound/soc/generic/audio-graph-card.c struct device_node *node = of_graph_get_port_parent(ep); node 198 sound/soc/generic/audio-graph-card.c of_node_put(node); node 214 sound/soc/generic/audio-graph-card.c struct device_node *node; node 226 sound/soc/generic/audio-graph-card.c node = of_graph_get_port_parent(ep); node 298 sound/soc/generic/audio-graph-card.c snd_soc_of_parse_node_prefix(node, cconf, codecs->of_node, node 328 sound/soc/generic/audio-graph-card.c of_node_put(node); node 420 sound/soc/generic/audio-graph-card.c struct device_node *node = dev->of_node; node 431 sound/soc/generic/audio-graph-card.c of_for_each_phandle(&it, rc, node, "dais", NULL, 0) { node 432 sound/soc/generic/audio-graph-card.c cpu_port = it.node; node 56 sound/soc/generic/simple-card-utils.c struct device_node *node, node 65 sound/soc/generic/simple-card-utils.c daifmt = snd_soc_of_parse_daifmt(node, prefix, node 161 sound/soc/generic/simple-card-utils.c struct device_node *node, node 174 sound/soc/generic/simple-card-utils.c clk = devm_get_clk_from_child(dev, node, NULL); node 179 sound/soc/generic/simple-card-utils.c } else if (!of_property_read_u32(node, "system-clock-frequency", &val)) { node 187 sound/soc/generic/simple-card-utils.c if (of_property_read_bool(node, "system-clock-direction-out")) node 402 sound/soc/generic/simple-card-utils.c struct device_node *node = card->dev->of_node; node 410 sound/soc/generic/simple-card-utils.c if (!of_property_read_bool(node, prop)) node 420 sound/soc/generic/simple-card-utils.c struct device_node *node = card->dev->of_node; node 428 sound/soc/generic/simple-card-utils.c if (of_property_read_bool(node, prop)) node 31 sound/soc/generic/simple-card.c static int asoc_simple_parse_dai(struct device_node *node, node 38 sound/soc/generic/simple-card.c if (!node) node 45 sound/soc/generic/simple-card.c ret = of_parse_phandle_with_args(node, DAI, CELL, 0, &args); node 68 sound/soc/generic/simple-card.c ret = snd_soc_of_get_dai_name(node, &dlc->dai_name); node 85 sound/soc/generic/simple-card.c struct device_node *node = of_get_parent(np); node 88 sound/soc/generic/simple-card.c asoc_simple_parse_convert(dev, node, PREFIX, adata); node 89 sound/soc/generic/simple-card.c asoc_simple_parse_convert(dev, node, NULL, adata); node 92 sound/soc/generic/simple-card.c of_node_put(node); node 101 sound/soc/generic/simple-card.c struct device_node *node = of_get_parent(cpu); node 108 sound/soc/generic/simple-card.c of_property_read_u32(node, prop, &props->mclk_fs); node 112 sound/soc/generic/simple-card.c of_node_put(node); node 128 sound/soc/generic/simple-card.c struct device_node *node = of_get_parent(np); node 214 sound/soc/generic/simple-card.c snd_soc_of_parse_node_prefix(node, cconf, codecs->of_node, node 229 sound/soc/generic/simple-card.c ret = asoc_simple_parse_daifmt(dev, node, codec, node 240 sound/soc/generic/simple-card.c of_node_put(node); node 257 sound/soc/generic/simple-card.c struct device_node *node = NULL; node 273 sound/soc/generic/simple-card.c node = of_get_parent(np); node 276 sound/soc/generic/simple-card.c dev_dbg(dev, "link_of (%pOF)\n", node); node 283 sound/soc/generic/simple-card.c plat = of_get_child_by_name(node, prop); node 290 sound/soc/generic/simple-card.c ret = asoc_simple_parse_daifmt(dev, node, codec, node 340 sound/soc/generic/simple-card.c of_node_put(node); node 358 sound/soc/generic/simple-card.c struct device_node *node; node 364 sound/soc/generic/simple-card.c node = of_get_child_by_name(top, PREFIX "dai-link"); node 365 sound/soc/generic/simple-card.c if (!node) { node 366 sound/soc/generic/simple-card.c node = of_node_get(top); node 375 sound/soc/generic/simple-card.c int num = of_get_child_count(node); node 378 sound/soc/generic/simple-card.c codec = of_get_child_by_name(node, is_top ? node 387 sound/soc/generic/simple-card.c for_each_child_of_node(node, np) node 391 sound/soc/generic/simple-card.c for_each_child_of_node(node, np) { node 413 sound/soc/generic/simple-card.c node = of_get_next_child(top, node); node 414 sound/soc/generic/simple-card.c } while (!is_top && node); node 417 sound/soc/generic/simple-card.c of_node_put(node); node 421 sound/soc/generic/simple-card.c static int simple_parse_aux_devs(struct device_node *node, node 429 sound/soc/generic/simple-card.c if (!of_find_property(node, PREFIX "aux-devs", &len)) node 442 sound/soc/generic/simple-card.c aux_node = of_parse_phandle(node, PREFIX "aux-devs", i); node 544 sound/soc/hisilicon/hi6210-i2s.c struct device_node *node = pdev->dev.of_node; node 567 sound/soc/hisilicon/hi6210-i2s.c i2s->sysctrl = syscon_regmap_lookup_by_phandle(node, node 295 sound/soc/intel/atom/sst-atom-controls.c list_for_each_entry(algo, &ids->algo_list, node) { node 589 sound/soc/intel/atom/sst-atom-controls.c list_for_each_entry(gain, &ids->gain_list, node) { node 1386 sound/soc/intel/atom/sst-atom-controls.c list_add_tail(&module->node, &ids->gain_list); node 1392 sound/soc/intel/atom/sst-atom-controls.c list_add_tail(&module->node, &ids->algo_list); node 548 sound/soc/intel/atom/sst-atom-controls.h struct list_head node; node 484 sound/soc/intel/atom/sst-mfld-dsp.h struct list_head node; node 106 sound/soc/intel/atom/sst/sst.c list_add_tail(&msg->node, &drv->rx_list); node 126 sound/soc/intel/atom/sst/sst.c list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) { node 127 sound/soc/intel/atom/sst/sst.c list_del(&msg->node); node 166 sound/soc/intel/atom/sst/sst.h struct list_head node; node 43 sound/soc/intel/atom/sst/sst_ipc.c list_add_tail(&msg->node, &ctx->block_list); node 72 sound/soc/intel/atom/sst/sst_ipc.c list_for_each_entry(block, &ctx->block_list, node) { node 99 sound/soc/intel/atom/sst/sst_ipc.c list_for_each_entry_safe(block, __block, &ctx->block_list, node) { node 103 sound/soc/intel/atom/sst/sst_ipc.c list_del(&freed->node); node 157 sound/soc/intel/atom/sst/sst_ipc.c struct ipc_post, node); node 158 sound/soc/intel/atom/sst/sst_ipc.c list_del(&msg->node); node 404 sound/soc/intel/atom/sst/sst_pvt.c list_add_tail(&msg->node, &sst->ipc_dispatch_list); node 142 sound/soc/intel/baytrail/sst-baytrail-ipc.c struct list_head node; node 203 sound/soc/intel/baytrail/sst-baytrail-ipc.c list_for_each_entry(stream, &byt->stream_list, node) { node 351 sound/soc/intel/baytrail/sst-baytrail-ipc.c list_add(&stream->node, &byt->stream_list); node 455 sound/soc/intel/baytrail/sst-baytrail-ipc.c list_del(&stream->node); node 243 sound/soc/intel/haswell/sst-haswell-ipc.c struct list_head node; node 416 sound/soc/intel/haswell/sst-haswell-ipc.c list_for_each_entry(stream, &hsw->stream_list, node) { node 976 sound/soc/intel/haswell/sst-haswell-ipc.c list_add(&stream->node, &hsw->stream_list); node 1024 sound/soc/intel/haswell/sst-haswell-ipc.c list_del(&stream->node); node 507 sound/soc/intel/skylake/skl-messages.c node_id.node.dma_type = node 511 sound/soc/intel/skylake/skl-messages.c node_id.node.vindex = params->host_dma_id + node 516 sound/soc/intel/skylake/skl-messages.c node_id.node.dma_type = node 522 sound/soc/intel/skylake/skl-messages.c node_id.node.vindex = ssp_node.val; node 526 sound/soc/intel/skylake/skl-messages.c node_id.node.dma_type = SKL_DMA_DMIC_LINK_INPUT_CLASS; node 527 sound/soc/intel/skylake/skl-messages.c node_id.node.vindex = mconfig->vbus_id + node 532 sound/soc/intel/skylake/skl-messages.c node_id.node.dma_type = node 536 sound/soc/intel/skylake/skl-messages.c node_id.node.vindex = params->link_dma_id; node 540 sound/soc/intel/skylake/skl-messages.c node_id.node.dma_type = node 544 sound/soc/intel/skylake/skl-messages.c node_id.node.vindex = params->host_dma_id; node 1388 sound/soc/intel/skylake/skl-pcm.c list_for_each_entry(p, &skl->ppl_list, node) { node 1389 sound/soc/intel/skylake/skl-pcm.c list_for_each_entry(m, &p->pipe->w_list, node) { node 1525 sound/soc/intel/skylake/skl-pcm.c list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) { node 1526 sound/soc/intel/skylake/skl-pcm.c list_del(&modules->node); node 478 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &pipe->w_list, node) { node 555 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &pipe->w_list, node) { node 674 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &s_pipe->w_list, node) { node 693 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &s_pipe->w_list, node) { node 697 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(modules, &skl->bind_list, node) { node 897 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(modules, &skl->bind_list, node) { node 910 sound/soc/intel/skylake/skl-topology.c list_add(&m_list->node, &skl->bind_list); node 1178 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &s_pipe->w_list, node) { node 1184 sound/soc/intel/skylake/skl-topology.c list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) { node 1199 sound/soc/intel/skylake/skl-topology.c list_del(&modules->node); node 1207 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &s_pipe->w_list, node) { node 1221 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &s_pipe->w_list, node) { node 1956 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(ppl, &skl->ppl_list, node) { node 1980 sound/soc/intel/skylake/skl-topology.c list_add(&ppl->node, &skl->ppl_list); node 2610 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(ppl, &skl->ppl_list, node) { node 2638 sound/soc/intel/skylake/skl-topology.c list_add(&ppl->node, &skl->ppl_list); node 3526 sound/soc/intel/skylake/skl-topology.c list_add_tail(&p_module->node, &pipe->w_list); node 3540 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(w_module, &pipe->w_list, node) { node 3595 sound/soc/intel/skylake/skl-topology.c list_for_each_entry(ppl, &skl->ppl_list, node) node 3609 sound/soc/intel/skylake/skl-topology.c list_for_each_entry_safe(ppl, tmp, &skl->ppl_list, node) node 3610 sound/soc/intel/skylake/skl-topology.c list_del(&ppl->node); node 189 sound/soc/intel/skylake/skl-topology.h } node; node 261 sound/soc/intel/skylake/skl-topology.h struct list_head node; node 419 sound/soc/intel/skylake/skl-topology.h struct list_head node; node 425 sound/soc/intel/skylake/skl-topology.h struct list_head node; node 71 sound/soc/meson/axg-card.c struct device_node *node, node 78 sound/soc/meson/axg-card.c if (!dai_name || !dai_of_node || !node) node 81 sound/soc/meson/axg-card.c ret = of_parse_phandle_with_args(node, "sound-dai", node 95 sound/soc/meson/axg-card.c struct device_node *node, node 99 sound/soc/meson/axg-card.c prefix, node->full_name); node 137 sound/soc/meson/axg-card.c struct device_node *node = card->dev->of_node; node 141 sound/soc/meson/axg-card.c num = of_count_phandle_with_args(node, "audio-aux-devs", NULL); node 163 sound/soc/meson/axg-card.c of_parse_phandle(node, "audio-aux-devs", i); node 307 sound/soc/meson/axg-card.c static unsigned int axg_card_parse_daifmt(struct device_node *node, node 314 sound/soc/meson/axg-card.c daifmt = snd_soc_of_parse_daifmt(node, PREFIX, node 335 sound/soc/meson/axg-card.c struct device_node *node, node 351 sound/soc/meson/axg-card.c snd_soc_of_get_slot_mask(node, propname, &be->tx_mask[i]); node 361 sound/soc/meson/axg-card.c snd_soc_of_get_slot_mask(node, propname, &be->rx_mask[i]); node 375 sound/soc/meson/axg-card.c of_property_read_u32(node, "dai-tdm-slot-num", &be->slots); node 391 sound/soc/meson/axg-card.c of_property_read_u32(node, "dai-tdm-slot-width", &be->slot_width); node 398 sound/soc/meson/axg-card.c struct device_node *node, node 411 sound/soc/meson/axg-card.c for_each_child_of_node(node, np) { node 424 sound/soc/meson/axg-card.c struct device_node *node, node 441 sound/soc/meson/axg-card.c link->dai_fmt = axg_card_parse_daifmt(node, link->cpus->of_node); node 443 sound/soc/meson/axg-card.c of_property_read_u32(node, "mclk-fs", &be->mclk_fs); node 445 sound/soc/meson/axg-card.c ret = axg_card_parse_cpu_tdm_slots(card, link, node, be); node 451 sound/soc/meson/axg-card.c ret = axg_card_parse_codecs_masks(card, link, node, be); node 467 sound/soc/meson/axg-card.c struct device_node *node) node 477 sound/soc/meson/axg-card.c num_codecs = of_get_child_count(node); node 480 sound/soc/meson/axg-card.c node->full_name); node 491 sound/soc/meson/axg-card.c for_each_child_of_node(node, np) { node 502 sound/soc/meson/axg-card.c ret = axg_card_set_link_name(card, link, node, "be"); node 511 sound/soc/meson/axg-card.c struct device_node *node, node 535 sound/soc/meson/axg-card.c return axg_card_set_link_name(card, link, node, "fe"); node 598 sound/soc/meson/axg-card.c struct device_node *node = card->dev->of_node; node 602 sound/soc/meson/axg-card.c num = of_get_child_count(node); node 613 sound/soc/meson/axg-card.c for_each_child_of_node(node, np) { node 118 sound/soc/qcom/apq8016_sbc.c struct device_node *np, *codec, *cpu, *node = dev->of_node; node 130 sound/soc/qcom/apq8016_sbc.c if (of_property_read_bool(node, "qcom,audio-routing")) { node 139 sound/soc/qcom/apq8016_sbc.c num_links = of_get_child_count(node); node 153 sound/soc/qcom/apq8016_sbc.c for_each_child_of_node(node, np) { node 52 sound/soc/qcom/qdsp6/q6adm.c struct list_head node; node 98 sound/soc/qcom/qdsp6/q6adm.c list_for_each_entry(c, &adm->copps_list, node) { node 120 sound/soc/qcom/qdsp6/q6adm.c list_del(&c->node); node 308 sound/soc/qcom/qdsp6/q6adm.c list_for_each_entry(c, &adm->copps_list, node) { node 411 sound/soc/qcom/qdsp6/q6adm.c list_add_tail(&copp->node, &adm->copps_list); node 465 sound/soc/qcom/qdsp6/q6adm.c struct q6adm_session_map_node_v5 *node; node 473 sound/soc/qcom/qdsp6/q6adm.c pkt_size = (APR_HDR_SIZE + sizeof(*route) + sizeof(*node) + node 482 sound/soc/qcom/qdsp6/q6adm.c node = matrix_map + APR_HDR_SIZE + sizeof(*route); node 483 sound/soc/qcom/qdsp6/q6adm.c copps_list = matrix_map + APR_HDR_SIZE + sizeof(*route) + sizeof(*node); node 505 sound/soc/qcom/qdsp6/q6adm.c node->session_id = payload_map.session_id; node 506 sound/soc/qcom/qdsp6/q6adm.c node->num_copps = payload_map.num_copps; node 1368 sound/soc/qcom/qdsp6/q6afe-dai.c struct device_node *node; node 1371 sound/soc/qcom/qdsp6/q6afe-dai.c for_each_child_of_node(dev->of_node, node) { node 1376 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "reg", &id); node 1386 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_variable_u32_array(node, node 1403 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-sync-mode", node 1409 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-sync-src", node 1415 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-data-out", node 1421 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-invert-sync", node 1427 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-data-delay", node 1433 sound/soc/qcom/qdsp6/q6afe-dai.c ret = of_property_read_u32(node, "qcom,tdm-data-align", node 486 sound/soc/qcom/qdsp6/q6afe.c struct list_head node; node 721 sound/soc/qcom/qdsp6/q6afe.c list_del(&port->node); node 734 sound/soc/qcom/qdsp6/q6afe.c list_for_each_entry(p, &afe->port_list, node) node 1442 sound/soc/qcom/qdsp6/q6afe.c list_add_tail(&port->node, &afe->port_list); node 877 sound/soc/qcom/qdsp6/q6asm-dai.c struct device_node *node; node 882 sound/soc/qcom/qdsp6/q6asm-dai.c for_each_child_of_node(dev->of_node, node) { node 883 sound/soc/qcom/qdsp6/q6asm-dai.c ret = of_property_read_u32(node, "reg", &id); node 891 sound/soc/qcom/qdsp6/q6asm-dai.c ret = of_property_read_u32(node, "direction", &dir); node 900 sound/soc/qcom/qdsp6/q6asm-dai.c if (of_property_read_bool(node, "is-compress-dai")) node 910 sound/soc/qcom/qdsp6/q6asm-dai.c struct device_node *node = dev->of_node; node 919 sound/soc/qcom/qdsp6/q6asm-dai.c rc = of_parse_phandle_with_fixed_args(node, "iommus", 1, 0, &args); node 579 sound/soc/rockchip/rockchip_i2s.c struct device_node *node = pdev->dev.of_node; node 594 sound/soc/rockchip/rockchip_i2s.c i2s->grf = syscon_regmap_lookup_by_phandle(node, "rockchip,grf"); node 658 sound/soc/rockchip/rockchip_i2s.c if (!of_property_read_u32(node, "rockchip,playback-channels", &val)) { node 663 sound/soc/rockchip/rockchip_i2s.c if (!of_property_read_u32(node, "rockchip,capture-channels", &val)) { node 1079 sound/soc/sh/rcar/core.c struct device_node *node = is_play ? node 1083 sound/soc/sh/rcar/core.c if (!node) node 1088 sound/soc/sh/rcar/core.c if (np == node) { node 1135 sound/soc/sh/rcar/core.c struct device_node *node, node 1144 sound/soc/sh/rcar/core.c if (!node) node 1148 sound/soc/sh/rcar/core.c for_each_child_of_node(node, np) { node 1157 sound/soc/sh/rcar/core.c of_node_put(node); node 298 sound/soc/sh/rcar/ctu.c struct device_node *node; node 310 sound/soc/sh/rcar/ctu.c node = rsnd_ctu_of_node(priv); node 311 sound/soc/sh/rcar/ctu.c if (!node) node 314 sound/soc/sh/rcar/ctu.c nr = of_get_child_count(node); node 331 sound/soc/sh/rcar/ctu.c for_each_child_of_node(node, np) { node 360 sound/soc/sh/rcar/ctu.c of_node_put(node); node 308 sound/soc/sh/rcar/dvc.c struct device_node *node; node 320 sound/soc/sh/rcar/dvc.c node = rsnd_dvc_of_node(priv); node 321 sound/soc/sh/rcar/dvc.c if (!node) node 324 sound/soc/sh/rcar/dvc.c nr = of_get_child_count(node); node 341 sound/soc/sh/rcar/dvc.c for_each_child_of_node(node, np) { node 365 sound/soc/sh/rcar/dvc.c of_node_put(node); node 272 sound/soc/sh/rcar/mix.c struct device_node *node; node 284 sound/soc/sh/rcar/mix.c node = rsnd_mix_of_node(priv); node 285 sound/soc/sh/rcar/mix.c if (!node) node 288 sound/soc/sh/rcar/mix.c nr = of_get_child_count(node); node 305 sound/soc/sh/rcar/mix.c for_each_child_of_node(node, np) { node 329 sound/soc/sh/rcar/mix.c of_node_put(node); node 451 sound/soc/sh/rcar/rsnd.h struct device_node *node, node 475 sound/soc/sh/rcar/rsnd.h #define rsnd_parse_of_node(priv, node) \ node 476 sound/soc/sh/rcar/rsnd.h of_get_child_by_name(rsnd_priv_to_dev(priv)->of_node, node) node 614 sound/soc/sh/rcar/src.c struct device_node *node; node 626 sound/soc/sh/rcar/src.c node = rsnd_src_of_node(priv); node 627 sound/soc/sh/rcar/src.c if (!node) node 630 sound/soc/sh/rcar/src.c nr = of_get_child_count(node); node 646 sound/soc/sh/rcar/src.c for_each_child_of_node(node, np) { node 683 sound/soc/sh/rcar/src.c of_node_put(node); node 1066 sound/soc/sh/rcar/ssi.c struct device_node *node; node 1071 sound/soc/sh/rcar/ssi.c node = rsnd_ssi_of_node(priv); node 1072 sound/soc/sh/rcar/ssi.c if (!node) node 1076 sound/soc/sh/rcar/ssi.c for_each_child_of_node(node, np) { node 1085 sound/soc/sh/rcar/ssi.c of_node_put(node); node 1106 sound/soc/sh/rcar/ssi.c struct device_node *node; node 1115 sound/soc/sh/rcar/ssi.c node = rsnd_ssi_of_node(priv); node 1116 sound/soc/sh/rcar/ssi.c if (!node) node 1119 sound/soc/sh/rcar/ssi.c nr = of_get_child_count(node); node 1135 sound/soc/sh/rcar/ssi.c for_each_child_of_node(node, np) { node 1182 sound/soc/sh/rcar/ssi.c of_node_put(node); node 361 sound/soc/sh/rcar/ssiu.c struct device_node *node = rsnd_ssiu_of_node(priv); node 369 sound/soc/sh/rcar/ssiu.c if (node) { node 371 sound/soc/sh/rcar/ssiu.c for_each_child_of_node(node, np) { node 380 sound/soc/sh/rcar/ssiu.c of_node_put(node); node 393 sound/soc/sh/rcar/ssiu.c struct device_node *node; node 406 sound/soc/sh/rcar/ssiu.c node = rsnd_ssiu_of_node(priv); node 407 sound/soc/sh/rcar/ssiu.c if (node) node 408 sound/soc/sh/rcar/ssiu.c nr = of_get_child_count(node); node 426 sound/soc/sh/rcar/ssiu.c if ((node) && node 444 sound/soc/sh/rcar/ssiu.c if (node) { node 208 sound/soc/soc-dapm.c struct snd_soc_dapm_widget *node; node 224 sound/soc/soc-dapm.c node = p->node[rdir]; node 225 sound/soc/soc-dapm.c if (node->endpoints[dir] != -1) { node 226 sound/soc/soc-dapm.c node->endpoints[dir] = -1; node 227 sound/soc/soc-dapm.c list_add_tail(&node->work_list, &list); node 1173 sound/soc/soc-dapm.c invalidate_paths_ep(path->node[dir], dir); node 1230 sound/soc/soc-dapm.c con += fn(path->node[dir], list, custom_stop_condition); node 2151 sound/soc/soc-dapm.c p->node[rdir]->name); node 2832 sound/soc/soc-dapm.c path->node[SND_SOC_DAPM_DIR_IN] = wsource; node 2833 sound/soc/soc-dapm.c path->node[SND_SOC_DAPM_DIR_OUT] = wsink; node 381 sound/soc/sti/sti_uniperif.c static int sti_uniperiph_cpu_dai_of(struct device_node *node, node 395 sound/soc/sti/sti_uniperif.c of_id = of_match_node(snd_soc_sti_match, node); node 436 sound/soc/sti/sti_uniperif.c if (!of_property_read_string(node, "st,tdm-mode", &mode)) node 473 sound/soc/sti/sti_uniperif.c struct device_node *node = pdev->dev.of_node; node 486 sound/soc/sti/sti_uniperif.c ret = sti_uniperiph_cpu_dai_of(node, priv); node 1013 sound/soc/sti/uniperif_player.c struct device_node *node = pdev->dev.of_node; node 1024 sound/soc/sti/uniperif_player.c regmap = syscon_regmap_lookup_by_phandle(node, "st,syscfg"); node 1686 sound/soc/ti/davinci-mcasp.c struct device_node *node = pdev->dev.of_node; node 1691 sound/soc/ti/davinci-mcasp.c if (!node) node 1694 sound/soc/ti/davinci-mcasp.c parent_name = of_get_property(node, "fck_parent", NULL); node 213 sound/soc/ti/omap-abe-twl6040.c struct device_node *node = pdev->dev.of_node; node 220 sound/soc/ti/omap-abe-twl6040.c if (!node) { node 248 sound/soc/ti/omap-abe-twl6040.c dai_node = of_parse_phandle(node, "ti,mcpdm", 0); node 267 sound/soc/ti/omap-abe-twl6040.c dai_node = of_parse_phandle(node, "ti,dmic", 0); node 286 sound/soc/ti/omap-abe-twl6040.c priv->jack_detection = of_property_read_bool(node, "ti,jack-detection"); node 287 sound/soc/ti/omap-abe-twl6040.c of_property_read_u32(node, "ti,mclk-freq", &priv->mclk_freq); node 1384 sound/soc/ti/omap-mcbsp.c struct device_node *node = pdev->dev.of_node; node 1395 sound/soc/ti/omap-mcbsp.c if (!of_property_read_u32(node, "ti,buffer-size", &buffer_size)) node 240 sound/soc/ti/omap-twl4030.c struct device_node *node = pdev->dev.of_node; node 251 sound/soc/ti/omap-twl4030.c if (node) { node 260 sound/soc/ti/omap-twl4030.c dai_node = of_parse_phandle(node, "ti,mcbsp", 0); node 271 sound/soc/ti/omap-twl4030.c dai_node = of_parse_phandle(node, "ti,mcbsp-voice", 0); node 282 sound/soc/ti/omap-twl4030.c priv->jack_detect = of_get_named_gpio(node, node 286 sound/soc/ti/omap-twl4030.c prop = of_find_property(node, "ti,audio-routing", NULL); node 103 sound/soc/xilinx/xlnx_i2s.c struct device_node *node = dev->of_node; node 113 sound/soc/xilinx/xlnx_i2s.c ret = of_property_read_u32(node, "xlnx,num-channels", &ch); node 120 sound/soc/xilinx/xlnx_i2s.c ret = of_property_read_u32(node, "xlnx,dwidth", &data_width); node 136 sound/soc/xilinx/xlnx_i2s.c if (of_device_is_compatible(node, "xlnx,i2s-transmitter-1.0")) { node 144 sound/soc/xilinx/xlnx_i2s.c } else if (of_device_is_compatible(node, "xlnx,i2s-receiver-1.0")) { node 245 sound/soc/xilinx/xlnx_spdif.c struct device_node *node = dev->of_node; node 268 sound/soc/xilinx/xlnx_spdif.c ret = of_property_read_u32(node, "xlnx,spdif-mode", &ctx->mode); node 295 sound/soc/xilinx/xlnx_spdif.c ret = of_property_read_u32(node, "xlnx,aud_clk_i", &ctx->aclk); node 393 sound/xen/xen_snd_front_cfg.c char node[3]; node 417 sound/xen/xen_snd_front_cfg.c snprintf(node, sizeof(node), "%d", num_streams); node 418 sound/xen/xen_snd_front_cfg.c if (!xenbus_exists(XBT_NIL, device_path, node)) node 481 sound/xen/xen_snd_front_cfg.c char node[3]; node 486 sound/xen/xen_snd_front_cfg.c snprintf(node, sizeof(node), "%d", num_devices); node 487 sound/xen/xen_snd_front_cfg.c if (!xenbus_exists(XBT_NIL, xb_dev->nodename, node)) node 57 tools/include/linux/hashtable.h #define hash_add(hashtable, node, key) \ node 58 tools/include/linux/hashtable.h hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))]) node 64 tools/include/linux/hashtable.h static inline bool hash_hashed(struct hlist_node *node) node 66 tools/include/linux/hashtable.h return !hlist_unhashed(node); node 93 tools/include/linux/hashtable.h static inline void hash_del(struct hlist_node *node) node 95 tools/include/linux/hashtable.h hlist_del_init(node); node 42 tools/include/linux/rbtree.h #define RB_EMPTY_NODE(node) \ node 43 tools/include/linux/rbtree.h ((node)->__rb_parent_color == (unsigned long)(node)) node 44 tools/include/linux/rbtree.h #define RB_CLEAR_NODE(node) \ node 45 tools/include/linux/rbtree.h ((node)->__rb_parent_color = (unsigned long)(node)) node 66 tools/include/linux/rbtree.h static inline void rb_link_node(struct rb_node *node, struct rb_node *parent, node 69 tools/include/linux/rbtree.h node->__rb_parent_color = (unsigned long)parent; node 70 tools/include/linux/rbtree.h node->rb_left = node->rb_right = NULL; node 72 tools/include/linux/rbtree.h *rb_link = node; node 129 tools/include/linux/rbtree.h static inline void rb_insert_color_cached(struct rb_node *node, node 134 tools/include/linux/rbtree.h root->rb_leftmost = node; node 135 tools/include/linux/rbtree.h rb_insert_color(node, &root->rb_root); node 138 tools/include/linux/rbtree.h static inline void rb_erase_cached(struct rb_node *node, node 141 tools/include/linux/rbtree.h if (root->rb_leftmost == node) node 142 tools/include/linux/rbtree.h root->rb_leftmost = rb_next(node); node 143 tools/include/linux/rbtree.h rb_erase(node, &root->rb_root); node 30 tools/include/linux/rbtree_augmented.h void (*propagate)(struct rb_node *node, struct rb_node *stop); node 35 tools/include/linux/rbtree_augmented.h extern void __rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 49 tools/include/linux/rbtree_augmented.h rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 52 tools/include/linux/rbtree_augmented.h __rb_insert_augmented(node, root, augment->rotate); node 56 tools/include/linux/rbtree_augmented.h rb_insert_augmented_cached(struct rb_node *node, node 61 tools/include/linux/rbtree_augmented.h root->rb_leftmost = node; node 62 tools/include/linux/rbtree_augmented.h rb_insert_augmented(node, &root->rb_root, augment); node 82 tools/include/linux/rbtree_augmented.h RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ node 83 tools/include/linux/rbtree_augmented.h if (RBCOMPUTE(node, true)) \ node 85 tools/include/linux/rbtree_augmented.h rb = rb_parent(&node->RBFIELD); \ node 124 tools/include/linux/rbtree_augmented.h static inline bool RBNAME ## _compute_max(RBSTRUCT *node, bool exit) \ node 127 tools/include/linux/rbtree_augmented.h RBTYPE max = RBCOMPUTE(node); \ node 128 tools/include/linux/rbtree_augmented.h if (node->RBFIELD.rb_left) { \ node 129 tools/include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ node 133 tools/include/linux/rbtree_augmented.h if (node->RBFIELD.rb_right) { \ node 134 tools/include/linux/rbtree_augmented.h child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \ node 138 tools/include/linux/rbtree_augmented.h if (exit && node->RBAUGMENTED == max) \ node 140 tools/include/linux/rbtree_augmented.h node->RBAUGMENTED = max; \ node 187 tools/include/linux/rbtree_augmented.h __rb_erase_augmented(struct rb_node *node, struct rb_root *root, node 190 tools/include/linux/rbtree_augmented.h struct rb_node *child = node->rb_right; node 191 tools/include/linux/rbtree_augmented.h struct rb_node *tmp = node->rb_left; node 203 tools/include/linux/rbtree_augmented.h pc = node->__rb_parent_color; node 205 tools/include/linux/rbtree_augmented.h __rb_change_child(node, child, parent, root); node 214 tools/include/linux/rbtree_augmented.h tmp->__rb_parent_color = pc = node->__rb_parent_color; node 216 tools/include/linux/rbtree_augmented.h __rb_change_child(node, tmp, parent, root); node 236 tools/include/linux/rbtree_augmented.h augment->copy(node, successor); node 262 tools/include/linux/rbtree_augmented.h augment->copy(node, successor); node 266 tools/include/linux/rbtree_augmented.h tmp = node->rb_left; node 270 tools/include/linux/rbtree_augmented.h pc = node->__rb_parent_color; node 272 tools/include/linux/rbtree_augmented.h __rb_change_child(node, successor, tmp, root); node 291 tools/include/linux/rbtree_augmented.h rb_erase_augmented(struct rb_node *node, struct rb_root *root, node 294 tools/include/linux/rbtree_augmented.h struct rb_node *rebalance = __rb_erase_augmented(node, root, augment); node 300 tools/include/linux/rbtree_augmented.h rb_erase_augmented_cached(struct rb_node *node, struct rb_root_cached *root, node 303 tools/include/linux/rbtree_augmented.h if (root->rb_leftmost == node) node 304 tools/include/linux/rbtree_augmented.h root->rb_leftmost = rb_next(node); node 305 tools/include/linux/rbtree_augmented.h rb_erase_augmented(node, &root->rb_root, augment); node 105 tools/lib/bpf/bpf.c __u32 map_flags, int node) node 115 tools/lib/bpf/bpf.c if (node >= 0) { node 116 tools/lib/bpf/bpf.c map_attr.numa_node = node; node 155 tools/lib/bpf/bpf.c __u32 map_flags, int node) node 171 tools/lib/bpf/bpf.c if (node >= 0) { node 173 tools/lib/bpf/bpf.c attr.numa_node = node; node 58 tools/lib/bpf/bpf.h int max_entries, __u32 map_flags, int node); node 67 tools/lib/bpf/bpf.h __u32 map_flags, int node); node 1318 tools/lib/bpf/btf.c #define for_each_dedup_cand(d, node, hash) \ node 1319 tools/lib/bpf/btf.c hashmap__for_each_key_entry(d->dedup_table, node, (void *)hash) node 33 tools/lib/lockdep/preload.c struct rb_node node; node 103 tools/lib/lockdep/preload.c struct rb_node **node = &locks.rb_node; node 108 tools/lib/lockdep/preload.c while (*node) { node 109 tools/lib/lockdep/preload.c l = rb_entry(*node, struct lock_lookup, node); node 111 tools/lib/lockdep/preload.c *parent = *node; node 113 tools/lib/lockdep/preload.c node = &l->node.rb_left; node 115 tools/lib/lockdep/preload.c node = &l->node.rb_right; node 117 tools/lib/lockdep/preload.c return node; node 120 tools/lib/lockdep/preload.c return node; node 177 tools/lib/lockdep/preload.c struct rb_node **node, *parent; node 181 tools/lib/lockdep/preload.c node = __get_lock_node(lock, &parent); node 183 tools/lib/lockdep/preload.c if (*node) { node 184 tools/lib/lockdep/preload.c return rb_entry(*node, struct lock_lookup, node); node 204 tools/lib/lockdep/preload.c node = __get_lock_node(lock, &parent); node 205 tools/lib/lockdep/preload.c rb_link_node(&l->node, parent, node); node 206 tools/lib/lockdep/preload.c rb_insert_color(&l->node, &locks); node 215 tools/lib/lockdep/preload.c rb_erase(&lock->node, &locks); node 85 tools/lib/rbtree.c __rb_insert(struct rb_node *node, struct rb_root *root, node 88 tools/lib/rbtree.c struct rb_node *parent = rb_red_parent(node), *gparent, *tmp; node 100 tools/lib/rbtree.c rb_set_parent_color(node, NULL, RB_BLACK); node 133 tools/lib/rbtree.c node = gparent; node 134 tools/lib/rbtree.c parent = rb_parent(node); node 135 tools/lib/rbtree.c rb_set_parent_color(node, parent, RB_RED); node 140 tools/lib/rbtree.c if (node == tmp) { node 154 tools/lib/rbtree.c tmp = node->rb_left; node 156 tools/lib/rbtree.c WRITE_ONCE(node->rb_left, parent); node 160 tools/lib/rbtree.c rb_set_parent_color(parent, node, RB_RED); node 161 tools/lib/rbtree.c augment_rotate(parent, node); node 162 tools/lib/rbtree.c parent = node; node 163 tools/lib/rbtree.c tmp = node->rb_right; node 189 tools/lib/rbtree.c node = gparent; node 190 tools/lib/rbtree.c parent = rb_parent(node); node 191 tools/lib/rbtree.c rb_set_parent_color(node, parent, RB_RED); node 196 tools/lib/rbtree.c if (node == tmp) { node 198 tools/lib/rbtree.c tmp = node->rb_right; node 200 tools/lib/rbtree.c WRITE_ONCE(node->rb_right, parent); node 204 tools/lib/rbtree.c rb_set_parent_color(parent, node, RB_RED); node 205 tools/lib/rbtree.c augment_rotate(parent, node); node 206 tools/lib/rbtree.c parent = node; node 207 tools/lib/rbtree.c tmp = node->rb_left; node 230 tools/lib/rbtree.c struct rb_node *node = NULL, *sibling, *tmp1, *tmp2; node 241 tools/lib/rbtree.c if (node != sibling) { /* node == parent->rb_left */ node 285 tools/lib/rbtree.c node = parent; node 286 tools/lib/rbtree.c parent = rb_parent(node); node 375 tools/lib/rbtree.c node = parent; node 376 tools/lib/rbtree.c parent = rb_parent(node); node 423 tools/lib/rbtree.c static inline void dummy_propagate(struct rb_node *node, struct rb_node *stop) {} node 433 tools/lib/rbtree.c void rb_insert_color(struct rb_node *node, struct rb_root *root) node 435 tools/lib/rbtree.c __rb_insert(node, root, dummy_rotate); node 438 tools/lib/rbtree.c void rb_erase(struct rb_node *node, struct rb_root *root) node 441 tools/lib/rbtree.c rebalance = __rb_erase_augmented(node, root, &dummy_callbacks); node 453 tools/lib/rbtree.c void __rb_insert_augmented(struct rb_node *node, struct rb_root *root, node 456 tools/lib/rbtree.c __rb_insert(node, root, augment_rotate); node 486 tools/lib/rbtree.c struct rb_node *rb_next(const struct rb_node *node) node 490 tools/lib/rbtree.c if (RB_EMPTY_NODE(node)) node 497 tools/lib/rbtree.c if (node->rb_right) { node 498 tools/lib/rbtree.c node = node->rb_right; node 499 tools/lib/rbtree.c while (node->rb_left) node 500 tools/lib/rbtree.c node=node->rb_left; node 501 tools/lib/rbtree.c return (struct rb_node *)node; node 511 tools/lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_right) node 512 tools/lib/rbtree.c node = parent; node 517 tools/lib/rbtree.c struct rb_node *rb_prev(const struct rb_node *node) node 521 tools/lib/rbtree.c if (RB_EMPTY_NODE(node)) node 528 tools/lib/rbtree.c if (node->rb_left) { node 529 tools/lib/rbtree.c node = node->rb_left; node 530 tools/lib/rbtree.c while (node->rb_right) node 531 tools/lib/rbtree.c node=node->rb_right; node 532 tools/lib/rbtree.c return (struct rb_node *)node; node 539 tools/lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_left) node 540 tools/lib/rbtree.c node = parent; node 561 tools/lib/rbtree.c static struct rb_node *rb_left_deepest_node(const struct rb_node *node) node 564 tools/lib/rbtree.c if (node->rb_left) node 565 tools/lib/rbtree.c node = node->rb_left; node 566 tools/lib/rbtree.c else if (node->rb_right) node 567 tools/lib/rbtree.c node = node->rb_right; node 569 tools/lib/rbtree.c return (struct rb_node *)node; node 573 tools/lib/rbtree.c struct rb_node *rb_next_postorder(const struct rb_node *node) node 576 tools/lib/rbtree.c if (!node) node 578 tools/lib/rbtree.c parent = rb_parent(node); node 581 tools/lib/rbtree.c if (parent && node == parent->rb_left && parent->rb_right) { node 240 tools/perf/bench/numa.c static int is_node_present(int node) node 242 tools/perf/bench/numa.c return numa_bitmask_isbitset(numa_nodes_ptr, node); node 248 tools/perf/bench/numa.c static bool node_has_cpus(int node) node 253 tools/perf/bench/numa.c if (cpu && !numa_node_to_cpus(node, cpu)) { node 340 tools/perf/bench/numa.c static void bind_to_memnode(int node) node 345 tools/perf/bench/numa.c if (node == NUMA_NO_NODE) node 349 tools/perf/bench/numa.c nodemask = 1L << node; node 352 tools/perf/bench/numa.c dprintf("binding to node %d, mask: %016lx => %d\n", node, nodemask, ret); node 380 tools/perf/bench/numa.c int node = numa_node_of_cpu(0); node 382 tools/perf/bench/numa.c orig_mask = bind_to_node(node); node 383 tools/perf/bench/numa.c bind_to_memnode(node); node 884 tools/perf/bench/numa.c int node; node 889 tools/perf/bench/numa.c node = numa_node_of_cpu(td->curr_cpu); node 890 tools/perf/bench/numa.c if (node < 0) /* curr_cpu was likely still -1 */ node 893 tools/perf/bench/numa.c node_present[node] = 1; node 911 tools/perf/bench/numa.c static int count_node_processes(int node) node 926 tools/perf/bench/numa.c if (n == node) { node 976 tools/perf/bench/numa.c int node; node 983 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) node 984 tools/perf/bench/numa.c nodes[node] = 0; node 999 tools/perf/bench/numa.c node = numa_node_of_cpu(cpu); node 1001 tools/perf/bench/numa.c nodes[node]++; node 1012 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) { node 1013 tools/perf/bench/numa.c if (!is_node_present(node)) node 1015 tools/perf/bench/numa.c nr = nodes[node]; node 1034 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) { node 1037 tools/perf/bench/numa.c if (!is_node_present(node)) node 1039 tools/perf/bench/numa.c processes = count_node_processes(node); node 1040 tools/perf/bench/numa.c nr = nodes[node]; node 223 tools/perf/builtin-c2c.c int node; node 230 tools/perf/builtin-c2c.c node = mem2node__node(&c2c.mem2node, sample->phys_addr); node 231 tools/perf/builtin-c2c.c if (WARN_ONCE(node < 0, "WARNING: failed to find node\n")) node 234 tools/perf/builtin-c2c.c set_bit(node, c2c_he->nodeset); node 320 tools/perf/builtin-c2c.c int node = c2c.cpu2node[cpu]; node 337 tools/perf/builtin-c2c.c c2c_add_stats(&c2c_he->node_stats[node], &stats); node 1088 tools/perf/builtin-c2c.c int node; node 1093 tools/perf/builtin-c2c.c for (node = 0; node < c2c.nodes_cnt; node++) { node 1097 tools/perf/builtin-c2c.c bitmap_and(set, c2c_he->cpuset, c2c.nodes[node], c2c.cpus_cnt); node 1114 tools/perf/builtin-c2c.c ret = scnprintf(hpp->buf, hpp->size, "%2d", node); node 1120 tools/perf/builtin-c2c.c struct c2c_stats *stats = &c2c_he->node_stats[node]; node 1122 tools/perf/builtin-c2c.c ret = scnprintf(hpp->buf, hpp->size, "%2d{%2d ", node, num); node 1161 tools/perf/builtin-c2c.c ret = scnprintf(hpp->buf, hpp->size, "%2d{", node); node 2033 tools/perf/builtin-c2c.c int node, cpu; node 2061 tools/perf/builtin-c2c.c for (node = 0; node < c2c.nodes_cnt; node++) { node 2062 tools/perf/builtin-c2c.c struct perf_cpu_map *map = n[node].map; node 2069 tools/perf/builtin-c2c.c nodes[node] = set; node 2081 tools/perf/builtin-c2c.c cpu2node[map->map[cpu]] = node; node 495 tools/perf/builtin-diff.c list_for_each_entry(pair, &he->pairs.head, pairs.node) node 51 tools/perf/builtin-inject.c struct list_head node; node 476 tools/perf/builtin-inject.c list_for_each_entry(ent, &inject->samples, node) { node 478 tools/perf/builtin-inject.c list_del_init(&ent->node); node 507 tools/perf/builtin-inject.c list_add(&ent->node, &inject->samples); node 523 tools/perf/builtin-inject.c list_for_each_entry(ent, &inject->samples, node) { node 70 tools/perf/builtin-kmem.c struct rb_node node; node 88 tools/perf/builtin-kmem.c struct rb_node **node = &root_alloc_stat.rb_node; node 92 tools/perf/builtin-kmem.c while (*node) { node 93 tools/perf/builtin-kmem.c parent = *node; node 94 tools/perf/builtin-kmem.c data = rb_entry(*node, struct alloc_stat, node); node 97 tools/perf/builtin-kmem.c node = &(*node)->rb_right; node 99 tools/perf/builtin-kmem.c node = &(*node)->rb_left; node 120 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, node); node 121 tools/perf/builtin-kmem.c rb_insert_color(&data->node, &root_alloc_stat); node 133 tools/perf/builtin-kmem.c struct rb_node **node = &root_caller_stat.rb_node; node 137 tools/perf/builtin-kmem.c while (*node) { node 138 tools/perf/builtin-kmem.c parent = *node; node 139 tools/perf/builtin-kmem.c data = rb_entry(*node, struct alloc_stat, node); node 142 tools/perf/builtin-kmem.c node = &(*node)->rb_right; node 144 tools/perf/builtin-kmem.c node = &(*node)->rb_left; node 165 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, node); node 166 tools/perf/builtin-kmem.c rb_insert_color(&data->node, &root_caller_stat); node 215 tools/perf/builtin-kmem.c struct rb_node *node = root->rb_node; node 218 tools/perf/builtin-kmem.c while (node) { node 222 tools/perf/builtin-kmem.c data = rb_entry(node, struct alloc_stat, node); node 226 tools/perf/builtin-kmem.c node = node->rb_left; node 228 tools/perf/builtin-kmem.c node = node->rb_right; node 281 tools/perf/builtin-kmem.c struct rb_node node; node 338 tools/perf/builtin-kmem.c struct rb_node *node; node 359 tools/perf/builtin-kmem.c map__for_each_symbol(kernel_map, sym, node) { node 391 tools/perf/builtin-kmem.c struct callchain_cursor_node *node; node 406 tools/perf/builtin-kmem.c node = callchain_cursor_current(&callchain_cursor); node 407 tools/perf/builtin-kmem.c if (node == NULL) node 410 tools/perf/builtin-kmem.c key.start = key.end = node->ip; node 415 tools/perf/builtin-kmem.c if (node->map) node 416 tools/perf/builtin-kmem.c addr = map__unmap_ip(node->map, node->ip); node 418 tools/perf/builtin-kmem.c addr = node->ip; node 444 tools/perf/builtin-kmem.c struct rb_node **node = &page_live_tree.rb_node; node 448 tools/perf/builtin-kmem.c while (*node) { node 451 tools/perf/builtin-kmem.c parent = *node; node 452 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); node 456 tools/perf/builtin-kmem.c node = &parent->rb_left; node 458 tools/perf/builtin-kmem.c node = &parent->rb_right; node 473 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, node); node 474 tools/perf/builtin-kmem.c rb_insert_color(&data->node, &page_live_tree); node 493 tools/perf/builtin-kmem.c struct rb_node **node = &page_alloc_tree.rb_node; node 498 tools/perf/builtin-kmem.c while (*node) { node 501 tools/perf/builtin-kmem.c parent = *node; node 502 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); node 511 tools/perf/builtin-kmem.c node = &parent->rb_left; node 513 tools/perf/builtin-kmem.c node = &parent->rb_right; node 528 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, node); node 529 tools/perf/builtin-kmem.c rb_insert_color(&data->node, &page_alloc_tree); node 548 tools/perf/builtin-kmem.c struct rb_node **node = &page_caller_tree.rb_node; node 553 tools/perf/builtin-kmem.c while (*node) { node 556 tools/perf/builtin-kmem.c parent = *node; node 557 tools/perf/builtin-kmem.c data = rb_entry(*node, struct page_stat, node); node 566 tools/perf/builtin-kmem.c node = &parent->rb_left; node 568 tools/perf/builtin-kmem.c node = &parent->rb_right; node 583 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, node); node 584 tools/perf/builtin-kmem.c rb_insert_color(&data->node, &page_caller_tree); node 895 tools/perf/builtin-kmem.c rb_erase(&pstat->node, &page_live_tree); node 921 tools/perf/builtin-kmem.c rb_erase(&pstat->node, &page_caller_tree); node 1005 tools/perf/builtin-kmem.c node); node 1077 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); node 1119 tools/perf/builtin-kmem.c data = rb_entry(next, struct page_stat, node); node 1264 tools/perf/builtin-kmem.c this = rb_entry(*new, struct alloc_stat, node); node 1279 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, new); node 1280 tools/perf/builtin-kmem.c rb_insert_color(&data->node, root); node 1286 tools/perf/builtin-kmem.c struct rb_node *node; node 1290 tools/perf/builtin-kmem.c node = rb_first(root); node 1291 tools/perf/builtin-kmem.c if (!node) node 1294 tools/perf/builtin-kmem.c rb_erase(node, root); node 1295 tools/perf/builtin-kmem.c data = rb_entry(node, struct alloc_stat, node); node 1311 tools/perf/builtin-kmem.c this = rb_entry(*new, struct page_stat, node); node 1326 tools/perf/builtin-kmem.c rb_link_node(&data->node, parent, new); node 1327 tools/perf/builtin-kmem.c rb_insert_color(&data->node, root); node 1333 tools/perf/builtin-kmem.c struct rb_node *node; node 1337 tools/perf/builtin-kmem.c node = rb_first(root); node 1338 tools/perf/builtin-kmem.c if (!node) node 1341 tools/perf/builtin-kmem.c rb_erase(node, root); node 1342 tools/perf/builtin-kmem.c data = rb_entry(node, struct page_stat, node); node 562 tools/perf/builtin-kvm.c struct rb_node *node = rb_first(result); node 564 tools/perf/builtin-kvm.c if (!node) node 567 tools/perf/builtin-kvm.c rb_erase(node, result); node 568 tools/perf/builtin-kvm.c return container_of(node, struct kvm_event, rb); node 125 tools/perf/builtin-lock.c struct rb_node *node; node 128 tools/perf/builtin-lock.c node = thread_stats.rb_node; node 129 tools/perf/builtin-lock.c while (node) { node 130 tools/perf/builtin-lock.c st = container_of(node, struct thread_stat, rb); node 134 tools/perf/builtin-lock.c node = node->rb_left; node 136 tools/perf/builtin-lock.c node = node->rb_right; node 306 tools/perf/builtin-lock.c struct rb_node *node = result.rb_node; node 308 tools/perf/builtin-lock.c if (!node) node 311 tools/perf/builtin-lock.c while (node->rb_left) node 312 tools/perf/builtin-lock.c node = node->rb_left; node 314 tools/perf/builtin-lock.c rb_erase(node, &result); node 315 tools/perf/builtin-lock.c return container_of(node, struct lock_stat, rb); node 766 tools/perf/builtin-lock.c struct rb_node *node; node 771 tools/perf/builtin-lock.c node = rb_first(&thread_stats); node 772 tools/perf/builtin-lock.c while (node) { node 773 tools/perf/builtin-lock.c st = container_of(node, struct thread_stat, rb); node 776 tools/perf/builtin-lock.c node = rb_next(node); node 128 tools/perf/builtin-sched.c struct rb_node node; node 963 tools/perf/builtin-sched.c struct rb_node *node = root->rb_root.rb_node; node 966 tools/perf/builtin-sched.c while (node) { node 970 tools/perf/builtin-sched.c atoms = container_of(node, struct work_atoms, node); node 974 tools/perf/builtin-sched.c node = node->rb_left; node 976 tools/perf/builtin-sched.c node = node->rb_right; node 996 tools/perf/builtin-sched.c this = container_of(*new, struct work_atoms, node); node 1009 tools/perf/builtin-sched.c rb_link_node(&data->node, parent, new); node 1010 tools/perf/builtin-sched.c rb_insert_color_cached(&data->node, root, leftmost); node 1459 tools/perf/builtin-sched.c struct rb_node *node; node 1464 tools/perf/builtin-sched.c node = rb_first_cached(root); node 1465 tools/perf/builtin-sched.c if (!node) node 1468 tools/perf/builtin-sched.c rb_erase_cached(node, root); node 1469 tools/perf/builtin-sched.c data = rb_entry(node, struct work_atoms, node); node 2168 tools/perf/builtin-sched.c struct callchain_cursor_node *node; node 2171 tools/perf/builtin-sched.c node = callchain_cursor_current(cursor); node 2172 tools/perf/builtin-sched.c if (node == NULL) node 2175 tools/perf/builtin-sched.c sym = node->sym; node 2745 tools/perf/builtin-sched.c static size_t callchain__fprintf_folded(FILE *fp, struct callchain_node *node) node 2753 tools/perf/builtin-sched.c if (node == NULL) node 2756 tools/perf/builtin-sched.c ret = callchain__fprintf_folded(fp, node->parent); node 2759 tools/perf/builtin-sched.c list_for_each_entry(chain, &node->val, list) { node 2938 tools/perf/builtin-sched.c list_for_each_entry(evsel, &evlist->core.entries, core.node) { node 3093 tools/perf/builtin-sched.c this = container_of(*new, struct work_atoms, node); node 3119 tools/perf/builtin-sched.c rb_link_node(&data->node, parent, new); node 3120 tools/perf/builtin-sched.c rb_insert_color_cached(&data->node, root, leftmost); node 3126 tools/perf/builtin-sched.c struct rb_node *node; node 3131 tools/perf/builtin-sched.c while ((node = rb_first_cached(&sched->atom_root))) { node 3132 tools/perf/builtin-sched.c rb_erase_cached(node, &sched->atom_root); node 3133 tools/perf/builtin-sched.c data = rb_entry(node, struct work_atoms, node); node 3159 tools/perf/builtin-sched.c work_list = rb_entry(next, struct work_atoms, node); node 2527 tools/perf/builtin-script.c struct list_head node; node 2549 tools/perf/builtin-script.c list_add_tail(&s->node, &script_specs); node 2556 tools/perf/builtin-script.c list_for_each_entry(s, &script_specs, node) node 2596 tools/perf/builtin-script.c list_for_each_entry(s, &script_specs, node) node 2827 tools/perf/builtin-script.c struct list_head node; node 2855 tools/perf/builtin-script.c list_add_tail(&s->node, &script_descs); node 2862 tools/perf/builtin-script.c list_for_each_entry(s, &script_descs, node) node 3001 tools/perf/builtin-script.c list_for_each_entry(desc, &script_descs, node) { node 2651 tools/perf/builtin-trace.c list_del_init(&evsel->core.node); node 62 tools/perf/lib/evlist.c list_add_tail(&evsel->node, &evlist->entries); node 70 tools/perf/lib/evlist.c list_del_init(&evsel->node); node 92 tools/perf/lib/evlist.c node); node 94 tools/perf/lib/evlist.c next = list_next_entry(prev, node); node 98 tools/perf/lib/evlist.c if (&next->node == &evlist->entries) node 194 tools/perf/lib/evlist.c hlist_add_head(&sid->node, &evlist->heads[hash]); node 21 tools/perf/lib/evsel.c INIT_LIST_HEAD(&evsel->node); node 37 tools/perf/lib/include/internal/evlist.h list_for_each_entry(evsel, list, node) node 53 tools/perf/lib/include/internal/evlist.h list_for_each_entry_reverse(evsel, list, node) node 65 tools/perf/lib/include/internal/evlist.h return list_entry(evlist->entries.next, struct perf_evsel, node); node 70 tools/perf/lib/include/internal/evlist.h return list_entry(evlist->entries.prev, struct perf_evsel, node); node 19 tools/perf/lib/include/internal/evsel.h struct hlist_node node; node 38 tools/perf/lib/include/internal/evsel.h struct list_head node; node 167 tools/perf/tests/hists_common.c struct rb_node *node; node 175 tools/perf/tests/hists_common.c node = rb_first_cached(root); node 176 tools/perf/tests/hists_common.c while (node) { node 179 tools/perf/tests/hists_common.c he = rb_entry(node, struct hist_entry, rb_node_in); node 189 tools/perf/tests/hists_common.c node = rb_next(node); node 197 tools/perf/tests/hists_common.c struct rb_node *node; node 202 tools/perf/tests/hists_common.c node = rb_first_cached(root); node 203 tools/perf/tests/hists_common.c while (node) { node 206 tools/perf/tests/hists_common.c he = rb_entry(node, struct hist_entry, rb_node); node 217 tools/perf/tests/hists_common.c node = rb_next(node); node 131 tools/perf/tests/hists_cumulate.c struct rb_node *node; node 141 tools/perf/tests/hists_cumulate.c node = rb_first_cached(root_out); node 143 tools/perf/tests/hists_cumulate.c he = rb_entry(node, struct hist_entry, rb_node); node 144 tools/perf/tests/hists_cumulate.c rb_erase_cached(node, root_out); node 174 tools/perf/tests/hists_cumulate.c } node[10]; node 184 tools/perf/tests/hists_cumulate.c struct rb_node *node; node 203 tools/perf/tests/hists_cumulate.c for (node = rb_first(root), i = 0; node 204 tools/perf/tests/hists_cumulate.c node && (he = rb_entry(node, struct hist_entry, rb_node)); node 205 tools/perf/tests/hists_cumulate.c node = rb_next(node), i++) { node 222 tools/perf/tests/hists_cumulate.c root = &he->callchain->node.rb_root; node 234 tools/perf/tests/hists_cumulate.c !strcmp(CDSO(clist), expected_callchain[i].node[c].dso) && node 235 tools/perf/tests/hists_cumulate.c !strcmp(CSYM(clist), expected_callchain[i].node[c].sym)); node 145 tools/perf/tests/hists_link.c struct rb_node *node; node 155 tools/perf/tests/hists_link.c node = rb_first_cached(root); node 156 tools/perf/tests/hists_link.c while (node) { node 159 tools/perf/tests/hists_link.c he = rb_entry(node, struct hist_entry, rb_node_in); node 172 tools/perf/tests/hists_link.c node = rb_next(node); node 195 tools/perf/tests/hists_link.c struct rb_node *node; node 207 tools/perf/tests/hists_link.c node = rb_first_cached(root); node 208 tools/perf/tests/hists_link.c while (node) { node 211 tools/perf/tests/hists_link.c he = rb_entry(node, struct hist_entry, rb_node_in); node 229 tools/perf/tests/hists_link.c node = rb_next(node); node 97 tools/perf/tests/hists_output.c struct rb_node *node; node 107 tools/perf/tests/hists_output.c node = rb_first_cached(root_out); node 109 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 110 tools/perf/tests/hists_output.c rb_erase_cached(node, root_out); node 131 tools/perf/tests/hists_output.c struct rb_node *node; node 166 tools/perf/tests/hists_output.c node = rb_first_cached(root); node 167 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 172 tools/perf/tests/hists_output.c node = rb_next(node); node 173 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 178 tools/perf/tests/hists_output.c node = rb_next(node); node 179 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 184 tools/perf/tests/hists_output.c node = rb_next(node); node 185 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 190 tools/perf/tests/hists_output.c node = rb_next(node); node 191 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 196 tools/perf/tests/hists_output.c node = rb_next(node); node 197 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 202 tools/perf/tests/hists_output.c node = rb_next(node); node 203 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 208 tools/perf/tests/hists_output.c node = rb_next(node); node 209 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 214 tools/perf/tests/hists_output.c node = rb_next(node); node 215 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 233 tools/perf/tests/hists_output.c struct rb_node *node; node 266 tools/perf/tests/hists_output.c node = rb_first_cached(root); node 267 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 271 tools/perf/tests/hists_output.c node = rb_next(node); node 272 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 289 tools/perf/tests/hists_output.c struct rb_node *node; node 320 tools/perf/tests/hists_output.c node = rb_first_cached(root); node 321 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 326 tools/perf/tests/hists_output.c node = rb_next(node); node 327 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 332 tools/perf/tests/hists_output.c node = rb_next(node); node 333 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 338 tools/perf/tests/hists_output.c node = rb_next(node); node 339 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 344 tools/perf/tests/hists_output.c node = rb_next(node); node 345 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 363 tools/perf/tests/hists_output.c struct rb_node *node; node 398 tools/perf/tests/hists_output.c node = rb_first_cached(root); node 399 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 404 tools/perf/tests/hists_output.c node = rb_next(node); node 405 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 410 tools/perf/tests/hists_output.c node = rb_next(node); node 411 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 416 tools/perf/tests/hists_output.c node = rb_next(node); node 417 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 422 tools/perf/tests/hists_output.c node = rb_next(node); node 423 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 428 tools/perf/tests/hists_output.c node = rb_next(node); node 429 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 434 tools/perf/tests/hists_output.c node = rb_next(node); node 435 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 440 tools/perf/tests/hists_output.c node = rb_next(node); node 441 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 446 tools/perf/tests/hists_output.c node = rb_next(node); node 447 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 465 tools/perf/tests/hists_output.c struct rb_node *node; node 501 tools/perf/tests/hists_output.c node = rb_first_cached(root); node 502 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 509 tools/perf/tests/hists_output.c node = rb_next(node); node 510 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 516 tools/perf/tests/hists_output.c node = rb_next(node); node 517 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 523 tools/perf/tests/hists_output.c node = rb_next(node); node 524 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 530 tools/perf/tests/hists_output.c node = rb_next(node); node 531 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 537 tools/perf/tests/hists_output.c node = rb_next(node); node 538 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 544 tools/perf/tests/hists_output.c node = rb_next(node); node 545 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 551 tools/perf/tests/hists_output.c node = rb_next(node); node 552 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 558 tools/perf/tests/hists_output.c node = rb_next(node); node 559 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 565 tools/perf/tests/hists_output.c node = rb_next(node); node 566 tools/perf/tests/hists_output.c he = rb_entry(node, struct hist_entry, rb_node); node 14 tools/perf/tests/mem2node.c int node; node 17 tools/perf/tests/mem2node.c { .node = 0, .map = "0" }, node 18 tools/perf/tests/mem2node.c { .node = 1, .map = "1-2" }, node 19 tools/perf/tests/mem2node.c { .node = 3, .map = "5-7,9" }, node 58 tools/perf/tests/mem2node.c nodes[i].node = test_nodes[i].node; node 215 tools/perf/tests/switch-tracking.c struct event_node *node; node 217 tools/perf/tests/switch-tracking.c node = malloc(sizeof(struct event_node)); node 218 tools/perf/tests/switch-tracking.c if (!node) { node 222 tools/perf/tests/switch-tracking.c node->event = event; node 223 tools/perf/tests/switch-tracking.c list_add(&node->list, events); node 235 tools/perf/tests/switch-tracking.c node->event_time = sample.time; node 242 tools/perf/tests/switch-tracking.c struct event_node *node; node 245 tools/perf/tests/switch-tracking.c node = list_entry(events->next, struct event_node, list); node 246 tools/perf/tests/switch-tracking.c list_del_init(&node->list); node 247 tools/perf/tests/switch-tracking.c free(node); node 266 tools/perf/tests/switch-tracking.c struct event_node *events_array, *node; node 293 tools/perf/tests/switch-tracking.c list_for_each_entry(node, &events, list) node 294 tools/perf/tests/switch-tracking.c events_array[pos++] = *node; node 49 tools/perf/ui/browsers/annotate.c struct annotation_line *al = list_entry(entry, struct annotation_line, node); node 100 tools/perf/ui/browsers/annotate.c struct annotation_line *al = list_entry(entry, struct annotation_line, node); node 130 tools/perf/ui/browsers/annotate.c struct disasm_line *pos = list_prev_entry(cursor, al.node); node 276 tools/perf/ui/browsers/annotate.c pos = list_entry(pos->node.prev, struct annotation_line, node); node 316 tools/perf/ui/browsers/annotate.c list_for_each_entry(pos, ¬es->src->source, al.node) { node 353 tools/perf/ui/browsers/annotate.c al = list_entry(browser->b.top, struct annotation_line, node); node 448 tools/perf/ui/browsers/annotate.c list_for_each_entry(pos, ¬es->src->source, al.node) { node 494 tools/perf/ui/browsers/annotate.c list_for_each_entry_continue(al, ¬es->src->source, node) { node 531 tools/perf/ui/browsers/annotate.c list_for_each_entry_continue_reverse(al, ¬es->src->source, node) { node 173 tools/perf/ui/browsers/hists.c static int callchain_node__count_rows_rb_tree(struct callchain_node *node) node 178 tools/perf/ui/browsers/hists.c for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { node 199 tools/perf/ui/browsers/hists.c static int callchain_node__count_flat_rows(struct callchain_node *node) node 205 tools/perf/ui/browsers/hists.c list_for_each_entry(chain, &node->parent_val, list) { node 215 tools/perf/ui/browsers/hists.c list_for_each_entry(chain, &node->val, list) { node 228 tools/perf/ui/browsers/hists.c static int callchain_node__count_folded_rows(struct callchain_node *node __maybe_unused) node 233 tools/perf/ui/browsers/hists.c static int callchain_node__count_rows(struct callchain_node *node) node 240 tools/perf/ui/browsers/hists.c return callchain_node__count_flat_rows(node); node 242 tools/perf/ui/browsers/hists.c return callchain_node__count_folded_rows(node); node 244 tools/perf/ui/browsers/hists.c list_for_each_entry(chain, &node->val, list) { node 251 tools/perf/ui/browsers/hists.c n += callchain_node__count_rows_rb_tree(node); node 262 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); node 263 tools/perf/ui/browsers/hists.c n += callchain_node__count_rows(node); node 273 tools/perf/ui/browsers/hists.c struct rb_node *node; node 282 tools/perf/ui/browsers/hists.c node = rb_first_cached(&he->hroot_out); node 283 tools/perf/ui/browsers/hists.c while (node) { node 286 tools/perf/ui/browsers/hists.c child = rb_entry(node, struct hist_entry, rb_node); node 296 tools/perf/ui/browsers/hists.c node = rb_next(node); node 325 tools/perf/ui/browsers/hists.c static void callchain_node__init_have_children_rb_tree(struct callchain_node *node) node 327 tools/perf/ui/browsers/hists.c struct rb_node *nd = rb_first(&node->rb_root); node 329 tools/perf/ui/browsers/hists.c for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { node 348 tools/perf/ui/browsers/hists.c static void callchain_node__init_have_children(struct callchain_node *node, node 353 tools/perf/ui/browsers/hists.c chain = list_entry(node->val.next, struct callchain_list, list); node 356 tools/perf/ui/browsers/hists.c if (!list_empty(&node->val)) { node 357 tools/perf/ui/browsers/hists.c chain = list_entry(node->val.prev, struct callchain_list, list); node 358 tools/perf/ui/browsers/hists.c chain->has_children = !RB_EMPTY_ROOT(&node->rb_root); node 361 tools/perf/ui/browsers/hists.c callchain_node__init_have_children_rb_tree(node); node 370 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); node 371 tools/perf/ui/browsers/hists.c callchain_node__init_have_children(node, has_sibling); node 374 tools/perf/ui/browsers/hists.c callchain_node__make_parent_list(node); node 461 tools/perf/ui/browsers/hists.c static int callchain_node__set_folding_rb_tree(struct callchain_node *node, bool unfold) node 466 tools/perf/ui/browsers/hists.c for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { node 484 tools/perf/ui/browsers/hists.c static int callchain_node__set_folding(struct callchain_node *node, bool unfold) node 490 tools/perf/ui/browsers/hists.c list_for_each_entry(chain, &node->val, list) { node 497 tools/perf/ui/browsers/hists.c n += callchain_node__set_folding_rb_tree(node, unfold); node 508 tools/perf/ui/browsers/hists.c struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node); node 509 tools/perf/ui/browsers/hists.c n += callchain_node__set_folding(node, unfold); node 793 tools/perf/ui/browsers/hists.c struct callchain_node *node, node 827 tools/perf/ui/browsers/hists.c callchain_node__scnprintf_value(node, buf, sizeof(buf), node 843 tools/perf/ui/browsers/hists.c static bool check_percent_display(struct rb_node *node, u64 parent_total) node 847 tools/perf/ui/browsers/hists.c if (node == NULL) node 850 tools/perf/ui/browsers/hists.c if (rb_next(node)) node 853 tools/perf/ui/browsers/hists.c child = rb_entry(node, struct callchain_node, rb_node); node 865 tools/perf/ui/browsers/hists.c struct rb_node *node; node 869 tools/perf/ui/browsers/hists.c node = rb_first(root); node 870 tools/perf/ui/browsers/hists.c need_percent = check_percent_display(node, parent_total); node 872 tools/perf/ui/browsers/hists.c while (node) { node 873 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); node 874 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); node 929 tools/perf/ui/browsers/hists.c node = next; node 969 tools/perf/ui/browsers/hists.c struct rb_node *node; node 973 tools/perf/ui/browsers/hists.c node = rb_first(root); node 974 tools/perf/ui/browsers/hists.c need_percent = check_percent_display(node, parent_total); node 976 tools/perf/ui/browsers/hists.c while (node) { node 977 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); node 978 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); node 1040 tools/perf/ui/browsers/hists.c node = next; node 1054 tools/perf/ui/browsers/hists.c struct rb_node *node; node 1062 tools/perf/ui/browsers/hists.c node = rb_first(root); node 1063 tools/perf/ui/browsers/hists.c need_percent = check_percent_display(node, parent_total); node 1065 tools/perf/ui/browsers/hists.c while (node) { node 1066 tools/perf/ui/browsers/hists.c struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node); node 1067 tools/perf/ui/browsers/hists.c struct rb_node *next = rb_next(node); node 1106 tools/perf/ui/browsers/hists.c node = next; node 3225 tools/perf/ui/browsers/hists.c struct evsel *evsel = list_entry(entry, struct evsel, core.node); node 3322 tools/perf/ui/browsers/hists.c if (pos->core.node.next == &evlist->core.entries) node 3328 tools/perf/ui/browsers/hists.c if (pos->core.node.prev == &evlist->core.entries) node 3364 tools/perf/ui/browsers/hists.c struct evsel *evsel = list_entry(entry, struct evsel, core.node); node 126 tools/perf/ui/gtk/annotate.c list_for_each_entry(pos, ¬es->src->source, al.node) { node 155 tools/perf/ui/gtk/annotate.c list_for_each_entry_safe(pos, n, ¬es->src->source, al.node) { node 156 tools/perf/ui/gtk/annotate.c list_del_init(&pos->al.node); node 104 tools/perf/ui/gtk/hists.c struct callchain_node *node; node 109 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); node 114 tools/perf/ui/gtk/hists.c callchain_node__make_parent_list(node); node 116 tools/perf/ui/gtk/hists.c list_for_each_entry(chain, &node->parent_val, list) { node 121 tools/perf/ui/gtk/hists.c callchain_node__scnprintf_value(node, buf, sizeof(buf), total); node 137 tools/perf/ui/gtk/hists.c list_for_each_entry(chain, &node->val, list) { node 142 tools/perf/ui/gtk/hists.c callchain_node__scnprintf_value(node, buf, sizeof(buf), total); node 166 tools/perf/ui/gtk/hists.c struct callchain_node *node; node 173 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); node 175 tools/perf/ui/gtk/hists.c callchain_node__make_parent_list(node); node 177 tools/perf/ui/gtk/hists.c list_for_each_entry(chain, &node->parent_val, list) { node 193 tools/perf/ui/gtk/hists.c list_for_each_entry(chain, &node->val, list) { node 211 tools/perf/ui/gtk/hists.c callchain_node__scnprintf_value(node, buf, sizeof(buf), total); node 227 tools/perf/ui/gtk/hists.c struct callchain_node *node; node 233 tools/perf/ui/gtk/hists.c node = rb_entry(nd, struct callchain_node, rb_node); node 236 tools/perf/ui/gtk/hists.c need_new_parent = !has_single_node && (node->val_nr > 1); node 238 tools/perf/ui/gtk/hists.c list_for_each_entry(chain, &node->val, list) { node 243 tools/perf/ui/gtk/hists.c callchain_node__scnprintf_value(node, buf, sizeof(buf), total); node 260 tools/perf/ui/gtk/hists.c child_total = node->children_hit; node 265 tools/perf/ui/gtk/hists.c perf_gtk__add_callchain_graph(&node->rb_root, store, &iter, col, node 413 tools/perf/ui/gtk/hists.c struct rb_node *node; node 420 tools/perf/ui/gtk/hists.c for (node = rb_first_cached(root); node; node = rb_next(node)) { node 425 tools/perf/ui/gtk/hists.c he = rb_entry(node, struct hist_entry, rb_node); node 53 tools/perf/ui/hist.c list_for_each_entry(pair, &he->pairs.head, pairs.node) { node 178 tools/perf/ui/hist.c list_for_each_entry(pair, &a->pairs.head, pairs.node) { node 183 tools/perf/ui/hist.c list_for_each_entry(pair, &b->pairs.head, pairs.node) { node 730 tools/perf/ui/hist.c struct perf_hpp_list_node *node; node 736 tools/perf/ui/hist.c list_for_each_entry(node, &hists->hpp_formats, list) { node 737 tools/perf/ui/hist.c perf_hpp_list__for_each_format(&node->hpp, fmt) node 762 tools/perf/ui/hist.c struct perf_hpp_list_node *node = NULL; node 767 tools/perf/ui/hist.c list_for_each_entry(node, &hists->hpp_formats, list) { node 768 tools/perf/ui/hist.c if (node->level == fmt->level) { node 775 tools/perf/ui/hist.c node = malloc(sizeof(*node)); node 776 tools/perf/ui/hist.c if (node == NULL) node 779 tools/perf/ui/hist.c node->skip = skip; node 780 tools/perf/ui/hist.c node->level = fmt->level; node 781 tools/perf/ui/hist.c perf_hpp_list__init(&node->hpp); node 784 tools/perf/ui/hist.c list_add_tail(&node->list, &hists->hpp_formats); node 792 tools/perf/ui/hist.c node->skip = false; node 794 tools/perf/ui/hist.c list_add_tail(&fmt_copy->list, &node->hpp.fields); node 795 tools/perf/ui/hist.c list_add_tail(&fmt_copy->sort_list, &node->hpp.sorts); node 49 tools/perf/ui/stdio/hist.c static size_t ipchain__fprintf_graph(FILE *fp, struct callchain_node *node, node 68 tools/perf/ui/stdio/hist.c ret += callchain_node__fprintf_value(node, fp, total_samples); node 112 tools/perf/ui/stdio/hist.c struct rb_node *node, *next; node 124 tools/perf/ui/stdio/hist.c node = rb_first(root); node 125 tools/perf/ui/stdio/hist.c while (node) { node 129 tools/perf/ui/stdio/hist.c child = rb_entry(node, struct callchain_node, rb_node); node 141 tools/perf/ui/stdio/hist.c next = rb_next(node); node 168 tools/perf/ui/stdio/hist.c node = next; node 205 tools/perf/ui/stdio/hist.c static bool need_percent_display(struct rb_node *node, u64 parent_samples) node 209 tools/perf/ui/stdio/hist.c if (rb_next(node)) node 212 tools/perf/ui/stdio/hist.c cnode = rb_entry(node, struct callchain_node, rb_node); node 224 tools/perf/ui/stdio/hist.c struct rb_node *node; node 229 tools/perf/ui/stdio/hist.c node = rb_first(root); node 230 tools/perf/ui/stdio/hist.c if (node && !need_percent_display(node, parent_samples)) { node 231 tools/perf/ui/stdio/hist.c cnode = rb_entry(node, struct callchain_node, rb_node); node 281 tools/perf/ui/stdio/hist.c static size_t __callchain__fprintf_flat(FILE *fp, struct callchain_node *node, node 288 tools/perf/ui/stdio/hist.c if (!node) node 291 tools/perf/ui/stdio/hist.c ret += __callchain__fprintf_flat(fp, node->parent, total_samples); node 294 tools/perf/ui/stdio/hist.c list_for_each_entry(chain, &node->val, list) { node 329 tools/perf/ui/stdio/hist.c static size_t __callchain__fprintf_folded(FILE *fp, struct callchain_node *node) node 337 tools/perf/ui/stdio/hist.c if (!node) node 340 tools/perf/ui/stdio/hist.c ret += __callchain__fprintf_folded(fp, node->parent); node 343 tools/perf/ui/stdio/hist.c list_for_each_entry(chain, &node->val, list) { node 1258 tools/perf/util/annotate.c list_add_tail(&al->node, head); node 1264 tools/perf/util/annotate.c list_for_each_entry_continue(pos, head, node) node 1402 tools/perf/util/annotate.c list_for_each_entry_from(queue, ¬es->src->source, node) { node 1583 tools/perf/util/annotate.c dl = list_entry(list->prev, struct disasm_line, al.node); node 1595 tools/perf/util/annotate.c list_del_init(&dl->al.node); node 1774 tools/perf/util/annotate.c struct btf_node *node; node 1776 tools/perf/util/annotate.c node = perf_env__find_btf(dso->bpf_prog.env, node 1778 tools/perf/util/annotate.c if (node) node 1779 tools/perf/util/annotate.c btf = btf__new((__u8 *)(node->data), node 1780 tools/perf/util/annotate.c node->data_size); node 2042 tools/perf/util/annotate.c list_for_each_entry(al, ¬es->src->source, node) { node 2187 tools/perf/util/annotate.c struct rb_node *node; node 2189 tools/perf/util/annotate.c node = rb_first(src_root); node 2190 tools/perf/util/annotate.c while (node) { node 2193 tools/perf/util/annotate.c al = rb_entry(node, struct annotation_line, rb_node); node 2194 tools/perf/util/annotate.c next = rb_next(node); node 2195 tools/perf/util/annotate.c rb_erase(node, src_root); node 2198 tools/perf/util/annotate.c node = next; node 2205 tools/perf/util/annotate.c struct rb_node *node; node 2215 tools/perf/util/annotate.c node = rb_first(root); node 2216 tools/perf/util/annotate.c while (node) { node 2222 tools/perf/util/annotate.c al = rb_entry(node, struct annotation_line, rb_node); node 2236 tools/perf/util/annotate.c node = rb_next(node); node 2258 tools/perf/util/annotate.c list_for_each_entry_reverse(line, lines, node) { node 2318 tools/perf/util/annotate.c list_for_each_entry(pos, ¬es->src->source, node) { node 2352 tools/perf/util/annotate.c queue = list_entry(queue->node.next, typeof(*queue), node); node 2420 tools/perf/util/annotate.c list_for_each_entry(al, ¬es->src->source, node) { node 2488 tools/perf/util/annotate.c list_for_each_entry_safe(al, n, &as->source, node) { node 2489 tools/perf/util/annotate.c list_del_init(&al->node); node 2516 tools/perf/util/annotate.c list_for_each_entry(pos, head, al.node) node 2571 tools/perf/util/annotate.c list_for_each_entry(al, ¬es->src->source, node) { node 2607 tools/perf/util/annotate.c list_for_each_entry(al, ¬es->src->source, node) { node 2648 tools/perf/util/annotate.c list_for_each_entry(al, ¬es->src->source, node) { node 131 tools/perf/util/annotate.h struct list_head node; node 23 tools/perf/util/block-range.c struct block_range *entry = rb_entry(rb, struct block_range, node); node 41 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); node 54 tools/perf/util/block-range.c static inline void rb_link_left_of_node(struct rb_node *left, struct rb_node *node) node 56 tools/perf/util/block-range.c struct rb_node **p = &node->rb_left; node 58 tools/perf/util/block-range.c node = *p; node 59 tools/perf/util/block-range.c p = &node->rb_right; node 61 tools/perf/util/block-range.c rb_link_node(left, node, p); node 64 tools/perf/util/block-range.c static inline void rb_link_right_of_node(struct rb_node *right, struct rb_node *node) node 66 tools/perf/util/block-range.c struct rb_node **p = &node->rb_right; node 68 tools/perf/util/block-range.c node = *p; node 69 tools/perf/util/block-range.c p = &node->rb_left; node 71 tools/perf/util/block-range.c rb_link_node(right, node, p); node 90 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); node 117 tools/perf/util/block-range.c next = rb_entry(n, struct block_range, node); node 131 tools/perf/util/block-range.c rb_link_left_of_node(&head->node, &next->node); node 132 tools/perf/util/block-range.c rb_insert_color(&head->node, &block_ranges.root); node 154 tools/perf/util/block-range.c rb_link_node(&entry->node, parent, p); node 155 tools/perf/util/block-range.c rb_insert_color(&entry->node, &block_ranges.root); node 185 tools/perf/util/block-range.c rb_link_left_of_node(&head->node, &entry->node); node 186 tools/perf/util/block-range.c rb_insert_color(&head->node, &block_ranges.root); node 225 tools/perf/util/block-range.c rb_link_right_of_node(&tail->node, &entry->node); node 226 tools/perf/util/block-range.c rb_insert_color(&tail->node, &block_ranges.root); node 263 tools/perf/util/block-range.c rb_link_right_of_node(&tail->node, &entry->node); node 264 tools/perf/util/block-range.c rb_insert_color(&tail->node, &block_ranges.root); node 286 tools/perf/util/block-range.c rb_link_left_of_node(&hole->node, &next->node); node 287 tools/perf/util/block-range.c rb_insert_color(&hole->node, &block_ranges.root); node 23 tools/perf/util/block-range.h struct rb_node node; node 40 tools/perf/util/block-range.h struct rb_node *n = rb_next(&br->node); node 43 tools/perf/util/block-range.h return rb_entry(n, struct block_range, node); node 97 tools/perf/util/bpf-event.c struct btf_node *node; node 103 tools/perf/util/bpf-event.c node = malloc(data_size + sizeof(struct btf_node)); node 104 tools/perf/util/bpf-event.c if (!node) node 107 tools/perf/util/bpf-event.c node->id = btf_id; node 108 tools/perf/util/bpf-event.c node->data_size = data_size; node 109 tools/perf/util/bpf-event.c memcpy(node->data, data, data_size); node 111 tools/perf/util/bpf-event.c perf_env__insert_btf(env, node); node 458 tools/perf/util/bpf-event.c struct btf_node *node; node 460 tools/perf/util/bpf-event.c node = perf_env__find_btf(env, info->btf_id); node 461 tools/perf/util/bpf-event.c if (node) node 462 tools/perf/util/bpf-event.c btf = btf__new((__u8 *)(node->data), node 463 tools/perf/util/bpf-event.c node->data_size); node 291 tools/perf/util/build-id.c list_for_each_entry(pos, head, node) \ node 386 tools/perf/util/build-id.c list_for_each_entry(pos, head, node) node 42 tools/perf/util/call-path.c list_for_each_entry_safe(pos, n, &cpr->blocks, node) { node 43 tools/perf/util/call-path.c list_del_init(&pos->node); node 60 tools/perf/util/call-path.c node); node 65 tools/perf/util/call-path.c list_add_tail(&cpb->node, &cpr->blocks); node 44 tools/perf/util/call-path.h struct list_head node; node 413 tools/perf/util/callchain.c __sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node, node 419 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 427 tools/perf/util/callchain.c if (node->hit && node->hit >= min_hit) node 428 tools/perf/util/callchain.c rb_insert_callchain(rb_root, node, CHAIN_FLAT); node 440 tools/perf/util/callchain.c __sort_chain_flat(rb_root, &root->node, min_hit); node 443 tools/perf/util/callchain.c static void __sort_chain_graph_abs(struct callchain_node *node, node 449 tools/perf/util/callchain.c node->rb_root = RB_ROOT; node 450 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 458 tools/perf/util/callchain.c rb_insert_callchain(&node->rb_root, child, node 467 tools/perf/util/callchain.c __sort_chain_graph_abs(&chain_root->node, min_hit); node 468 tools/perf/util/callchain.c rb_root->rb_node = chain_root->node.rb_root.rb_node; node 471 tools/perf/util/callchain.c static void __sort_chain_graph_rel(struct callchain_node *node, node 478 tools/perf/util/callchain.c node->rb_root = RB_ROOT; node 479 tools/perf/util/callchain.c min_hit = ceil(node->children_hit * min_percent); node 481 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 488 tools/perf/util/callchain.c rb_insert_callchain(&node->rb_root, child, node 497 tools/perf/util/callchain.c __sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0); node 498 tools/perf/util/callchain.c rb_root->rb_node = chain_root->node.rb_root.rb_node; node 566 tools/perf/util/callchain.c fill_node(struct callchain_node *node, struct callchain_cursor *cursor) node 570 tools/perf/util/callchain.c node->val_nr = cursor->nr - cursor->pos; node 571 tools/perf/util/callchain.c if (!node->val_nr) node 621 tools/perf/util/callchain.c list_add_tail(&call->list, &node->val); node 710 tools/perf/util/callchain.c static enum match_result match_chain(struct callchain_cursor_node *node, node 717 tools/perf/util/callchain.c match = match_chain_strings(cnode->srcline, node->srcline); node 723 tools/perf/util/callchain.c if (node->sym && cnode->ms.sym) { node 730 tools/perf/util/callchain.c if (cnode->ms.sym->inlined || node->sym->inlined) { node 732 tools/perf/util/callchain.c node->sym->name); node 737 tools/perf/util/callchain.c node->map, node->sym->start); node 745 tools/perf/util/callchain.c match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->map, node->ip); node 749 tools/perf/util/callchain.c if (match == MATCH_EQ && node->branch) { node 752 tools/perf/util/callchain.c if (node->branch_from) { node 758 tools/perf/util/callchain.c if (node->branch_flags.predicted) node 761 tools/perf/util/callchain.c if (node->branch_flags.abort) node 765 tools/perf/util/callchain.c &node->branch_flags, node 766 tools/perf/util/callchain.c node->branch_from, node 767 tools/perf/util/callchain.c node->ip); node 773 tools/perf/util/callchain.c cnode->cycles_count += node->branch_flags.cycles; node 774 tools/perf/util/callchain.c cnode->iter_count += node->nr_loop_iter; node 775 tools/perf/util/callchain.c cnode->iter_cycles += node->iter_cycles; node 825 tools/perf/util/callchain.c struct callchain_cursor_node *node; node 833 tools/perf/util/callchain.c node = callchain_cursor_current(cursor); node 847 tools/perf/util/callchain.c if (match_chain(node, cnode) == MATCH_LT) node 872 tools/perf/util/callchain.c struct callchain_cursor_node *node; node 876 tools/perf/util/callchain.c node = callchain_cursor_current(cursor); node 877 tools/perf/util/callchain.c if (!node) node 931 tools/perf/util/callchain.c struct callchain_cursor_node *node; node 933 tools/perf/util/callchain.c node = callchain_cursor_current(cursor); node 934 tools/perf/util/callchain.c if (!node) node 937 tools/perf/util/callchain.c cmp = match_chain(node, cnode); node 986 tools/perf/util/callchain.c if (append_chain_children(&root->node, cursor, period) < 0) node 1043 tools/perf/util/callchain.c return merge_chain_branch(cursor, &dst->node, &src->node); node 1052 tools/perf/util/callchain.c struct callchain_cursor_node *node = *cursor->last; node 1054 tools/perf/util/callchain.c if (!node) { node 1055 tools/perf/util/callchain.c node = calloc(1, sizeof(*node)); node 1056 tools/perf/util/callchain.c if (!node) node 1059 tools/perf/util/callchain.c *cursor->last = node; node 1062 tools/perf/util/callchain.c node->ip = ip; node 1063 tools/perf/util/callchain.c map__zput(node->map); node 1064 tools/perf/util/callchain.c node->map = map__get(map); node 1065 tools/perf/util/callchain.c node->sym = sym; node 1066 tools/perf/util/callchain.c node->branch = branch; node 1067 tools/perf/util/callchain.c node->nr_loop_iter = nr_loop_iter; node 1068 tools/perf/util/callchain.c node->iter_cycles = iter_cycles; node 1069 tools/perf/util/callchain.c node->srcline = srcline; node 1072 tools/perf/util/callchain.c memcpy(&node->branch_flags, flags, node 1075 tools/perf/util/callchain.c node->branch_from = branch_from; node 1078 tools/perf/util/callchain.c cursor->last = &node->next; node 1107 tools/perf/util/callchain.c int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node, node 1110 tools/perf/util/callchain.c al->map = node->map; node 1111 tools/perf/util/callchain.c al->sym = node->sym; node 1112 tools/perf/util/callchain.c al->srcline = node->srcline; node 1113 tools/perf/util/callchain.c al->addr = node->ip; node 1176 tools/perf/util/callchain.c char *callchain_node__scnprintf_value(struct callchain_node *node, node 1180 tools/perf/util/callchain.c u64 period = callchain_cumul_hits(node); node 1181 tools/perf/util/callchain.c unsigned count = callchain_cumul_counts(node); node 1184 tools/perf/util/callchain.c period = node->hit; node 1185 tools/perf/util/callchain.c count = node->count; node 1205 tools/perf/util/callchain.c int callchain_node__fprintf_value(struct callchain_node *node, node 1209 tools/perf/util/callchain.c u64 period = callchain_cumul_hits(node); node 1210 tools/perf/util/callchain.c unsigned count = callchain_cumul_counts(node); node 1213 tools/perf/util/callchain.c period = node->hit; node 1214 tools/perf/util/callchain.c count = node->count; node 1231 tools/perf/util/callchain.c static void callchain_counts_value(struct callchain_node *node, node 1237 tools/perf/util/callchain.c list_for_each_entry(clist, &node->val, list) { node 1252 tools/perf/util/callchain.c static int callchain_node_branch_counts_cumul(struct callchain_node *node, node 1261 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 1295 tools/perf/util/callchain.c return callchain_node_branch_counts_cumul(&root->node, node 1454 tools/perf/util/callchain.c static void free_callchain_node(struct callchain_node *node) node 1460 tools/perf/util/callchain.c list_for_each_entry_safe(list, tmp, &node->parent_val, list) { node 1466 tools/perf/util/callchain.c list_for_each_entry_safe(list, tmp, &node->val, list) { node 1472 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 1476 tools/perf/util/callchain.c rb_erase(&child->rb_node_in, &node->rb_root_in); node 1488 tools/perf/util/callchain.c free_callchain_node(&root->node); node 1491 tools/perf/util/callchain.c static u64 decay_callchain_node(struct callchain_node *node) node 1497 tools/perf/util/callchain.c n = rb_first(&node->rb_root_in); node 1505 tools/perf/util/callchain.c node->hit = (node->hit * 7) / 8; node 1506 tools/perf/util/callchain.c node->children_hit = child_hits; node 1508 tools/perf/util/callchain.c return node->hit; node 1516 tools/perf/util/callchain.c decay_callchain_node(&root->node); node 1519 tools/perf/util/callchain.c int callchain_node__make_parent_list(struct callchain_node *node) node 1521 tools/perf/util/callchain.c struct callchain_node *parent = node->parent; node 1539 tools/perf/util/callchain.c list_move_tail(&chain->list, &node->parent_val); node 1541 tools/perf/util/callchain.c if (!list_empty(&node->parent_val)) { node 1542 tools/perf/util/callchain.c chain = list_first_entry(&node->parent_val, struct callchain_list, list); node 1543 tools/perf/util/callchain.c chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node); node 1545 tools/perf/util/callchain.c chain = list_first_entry(&node->val, struct callchain_list, list); node 1568 tools/perf/util/callchain.c struct callchain_cursor_node *node; node 1570 tools/perf/util/callchain.c node = callchain_cursor_current(src); node 1571 tools/perf/util/callchain.c if (node == NULL) node 1574 tools/perf/util/callchain.c rc = callchain_cursor_append(dst, node->ip, node->map, node->sym, node 1575 tools/perf/util/callchain.c node->branch, &node->branch_flags, node 1576 tools/perf/util/callchain.c node->nr_loop_iter, node 1577 tools/perf/util/callchain.c node->iter_cycles, node 1578 tools/perf/util/callchain.c node->branch_from, node->srcline); node 1594 tools/perf/util/callchain.c struct callchain_cursor_node *node; node 1599 tools/perf/util/callchain.c for (node = cursor->first; node != NULL; node = node->next) node 1600 tools/perf/util/callchain.c map__zput(node->map); node 76 tools/perf/util/callchain.h struct callchain_node node; node 167 tools/perf/util/callchain.h INIT_LIST_HEAD(&root->node.val); node 168 tools/perf/util/callchain.h INIT_LIST_HEAD(&root->node.parent_val); node 170 tools/perf/util/callchain.h root->node.parent = NULL; node 171 tools/perf/util/callchain.h root->node.hit = 0; node 172 tools/perf/util/callchain.h root->node.children_hit = 0; node 173 tools/perf/util/callchain.h root->node.rb_root_in = RB_ROOT; node 177 tools/perf/util/callchain.h static inline u64 callchain_cumul_hits(struct callchain_node *node) node 179 tools/perf/util/callchain.h return node->hit + node->children_hit; node 182 tools/perf/util/callchain.h static inline unsigned callchain_cumul_counts(struct callchain_node *node) node 184 tools/perf/util/callchain.h return node->count + node->children_count; node 246 tools/perf/util/callchain.h int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node, node 277 tools/perf/util/callchain.h char *callchain_node__scnprintf_value(struct callchain_node *node, node 279 tools/perf/util/callchain.h int callchain_node__fprintf_value(struct callchain_node *node, node 287 tools/perf/util/callchain.h int callchain_node__make_parent_list(struct callchain_node *node); node 515 tools/perf/util/config.c list_for_each_entry(section, sections, node) node 527 tools/perf/util/config.c list_for_each_entry(item, §ion->items, node) node 550 tools/perf/util/config.c list_add_tail(§ion->node, sections); node 569 tools/perf/util/config.c list_add_tail(&item->node, §ion->items); node 780 tools/perf/util/config.c list_for_each_entry_safe(item, tmp, §ion->items, node) { node 781 tools/perf/util/config.c list_del_init(&item->node); node 797 tools/perf/util/config.c list_for_each_entry_safe(section, tmp, &set->sections, node) { node 798 tools/perf/util/config.c list_del_init(§ion->node); node 12 tools/perf/util/config.h struct list_head node; node 19 tools/perf/util/config.h struct list_head node; node 50 tools/perf/util/config.h list_for_each_entry(section, list, node) node 58 tools/perf/util/config.h list_for_each_entry(item, list, node) node 233 tools/perf/util/cputopo.c static int load_numa_node(struct numa_topology_node *node, int nr) node 243 tools/perf/util/cputopo.c node->node = (u32) nr; node 258 tools/perf/util/cputopo.c node->mem_total = mem; node 260 tools/perf/util/cputopo.c node->mem_free = mem; node 261 tools/perf/util/cputopo.c if (node->mem_total && node->mem_free) node 282 tools/perf/util/cputopo.c node->cpus = buf; node 18 tools/perf/util/cputopo.h u32 node; node 2566 tools/perf/util/cs-etm.c INIT_LIST_HEAD(&etm->unknown_thread->node); node 238 tools/perf/util/db-export.c struct callchain_cursor_node *node; node 244 tools/perf/util/db-export.c node = callchain_cursor_current(&callchain_cursor); node 245 tools/perf/util/db-export.c if (!node) node 252 tools/perf/util/db-export.c al.sym = node->sym; node 253 tools/perf/util/db-export.c al.map = node->map; node 255 tools/perf/util/db-export.c al.addr = node->ip; node 264 tools/perf/util/db-export.c al.sym, node->ip, node 717 tools/perf/util/dso.c struct bpf_prog_info_node *node; node 722 tools/perf/util/dso.c node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id); node 723 tools/perf/util/dso.c if (!node || !node->info_linear) { node 728 tools/perf/util/dso.c len = node->info_linear->info.jited_prog_len; node 729 tools/perf/util/dso.c buf = (u8 *)(uintptr_t)node->info_linear->info.jited_prog_insns; node 741 tools/perf/util/dso.c struct bpf_prog_info_node *node; node 743 tools/perf/util/dso.c node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id); node 744 tools/perf/util/dso.c if (!node || !node->info_linear) { node 749 tools/perf/util/dso.c dso->data.file_size = node->info_linear->info.jited_prog_len; node 1193 tools/perf/util/dso.c INIT_LIST_HEAD(&dso->node); node 136 tools/perf/util/dso.h struct list_head node; node 18 tools/perf/util/dsos.c list_for_each_entry(pos, head, node) { node 92 tools/perf/util/dsos.c list_add_tail(&dso->node, &dsos->head); node 129 tools/perf/util/dsos.c list_for_each_entry(pos, &dsos->head, node) node 213 tools/perf/util/dsos.c list_for_each_entry(pos, head, node) { node 227 tools/perf/util/dsos.c list_for_each_entry(pos, head, node) { node 20 tools/perf/util/env.c struct bpf_prog_info_node *node; node 29 tools/perf/util/env.c node = rb_entry(parent, struct bpf_prog_info_node, rb_node); node 30 tools/perf/util/env.c if (prog_id < node->info_linear->info.id) { node 32 tools/perf/util/env.c } else if (prog_id > node->info_linear->info.id) { node 50 tools/perf/util/env.c struct bpf_prog_info_node *node = NULL; node 57 tools/perf/util/env.c node = rb_entry(n, struct bpf_prog_info_node, rb_node); node 58 tools/perf/util/env.c if (prog_id < node->info_linear->info.id) node 60 tools/perf/util/env.c else if (prog_id > node->info_linear->info.id) node 65 tools/perf/util/env.c node = NULL; node 69 tools/perf/util/env.c return node; node 76 tools/perf/util/env.c struct btf_node *node; node 84 tools/perf/util/env.c node = rb_entry(parent, struct btf_node, rb_node); node 85 tools/perf/util/env.c if (btf_id < node->id) { node 87 tools/perf/util/env.c } else if (btf_id > node->id) { node 104 tools/perf/util/env.c struct btf_node *node = NULL; node 111 tools/perf/util/env.c node = rb_entry(n, struct btf_node, rb_node); node 112 tools/perf/util/env.c if (btf_id < node->id) node 114 tools/perf/util/env.c else if (btf_id > node->id) node 119 tools/perf/util/env.c node = NULL; node 123 tools/perf/util/env.c return node; node 138 tools/perf/util/env.c struct bpf_prog_info_node *node; node 140 tools/perf/util/env.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); node 141 tools/perf/util/env.c next = rb_next(&node->rb_node); node 142 tools/perf/util/env.c rb_erase(&node->rb_node, root); node 143 tools/perf/util/env.c free(node); node 152 tools/perf/util/env.c struct btf_node *node; node 154 tools/perf/util/env.c node = rb_entry(next, struct btf_node, rb_node); node 155 tools/perf/util/env.c next = rb_next(&node->rb_node); node 156 tools/perf/util/env.c rb_erase(&node->rb_node, root); node 157 tools/perf/util/env.c free(node); node 28 tools/perf/util/env.h u32 node; node 35 tools/perf/util/env.h u64 node; node 129 tools/perf/util/evlist.c list_del_init(&pos->core.node); node 184 tools/perf/util/evlist.c list_del_init(&evsel->core.node); node 193 tools/perf/util/evlist.c leader = list_entry(list->next, struct evsel, core.node); node 194 tools/perf/util/evlist.c evsel = list_entry(list->prev, struct evsel, core.node); node 249 tools/perf/util/evlist.c list_add_tail(&evsel->core.node, &head); node 451 tools/perf/util/evlist.c hlist_for_each_entry(sid, head, node) node 537 tools/perf/util/evlist.c hlist_for_each_entry(sid, head, node) { node 1520 tools/perf/util/evlist.c list_move_tail(&evsel->core.node, &move); node 264 tools/perf/util/evlist.h list_for_each_entry(evsel, list, core.node) node 280 tools/perf/util/evlist.h list_for_each_entry_continue(evsel, list, core.node) node 296 tools/perf/util/evlist.h list_for_each_entry_reverse(evsel, list, core.node) node 313 tools/perf/util/evlist.h list_for_each_entry_safe(evsel, tmp, list, core.node) node 1245 tools/perf/util/evsel.c assert(list_empty(&evsel->core.node)); node 309 tools/perf/util/evsel.h return list_entry(evsel->core.node.next, struct evsel, core.node); node 314 tools/perf/util/evsel.h return list_entry(evsel->core.node.prev, struct evsel, core.node); node 370 tools/perf/util/evsel.h for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \ node 372 tools/perf/util/evsel.h (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node)) node 378 tools/perf/util/evsel.h (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node)) node 109 tools/perf/util/evsel_fprintf.c struct callchain_cursor_node *node; node 130 tools/perf/util/evsel_fprintf.c node = callchain_cursor_current(cursor); node 131 tools/perf/util/evsel_fprintf.c if (!node) node 134 tools/perf/util/evsel_fprintf.c if (node->sym && node->sym->ignore && print_skip_ignored) node 143 tools/perf/util/evsel_fprintf.c printed += fprintf(fp, "%c%16" PRIx64, s, node->ip); node 145 tools/perf/util/evsel_fprintf.c if (node->map) node 146 tools/perf/util/evsel_fprintf.c addr = node->map->map_ip(node->map, node->ip); node 151 tools/perf/util/evsel_fprintf.c node_al.map = node->map; node 154 tools/perf/util/evsel_fprintf.c printed += __symbol__fprintf_symname_offs(node->sym, &node_al, node 158 tools/perf/util/evsel_fprintf.c printed += __symbol__fprintf_symname(node->sym, &node_al, node 163 tools/perf/util/evsel_fprintf.c if (print_dso && (!node->sym || !node->sym->inlined)) { node 165 tools/perf/util/evsel_fprintf.c printed += map__fprintf_dsoname(node->map, fp); node 170 tools/perf/util/evsel_fprintf.c printed += map__fprintf_srcline(node->map, addr, "\n ", fp); node 172 tools/perf/util/evsel_fprintf.c if (node->sym && node->sym->inlined) node 179 tools/perf/util/evsel_fprintf.c if (bt_stop_list && node->sym && node 180 tools/perf/util/evsel_fprintf.c strlist__has_entry(bt_stop_list, node->sym->name)) { node 689 tools/perf/util/header.c ret = do_write(ff, &n->node, sizeof(u32)); node 933 tools/perf/util/header.c struct bpf_prog_info_node *node; node 936 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); node 937 tools/perf/util/header.c next = rb_next(&node->rb_node); node 939 tools/perf/util/header.c node->info_linear->data_len; node 942 tools/perf/util/header.c bpf_program__bpil_addr_to_offs(node->info_linear); node 943 tools/perf/util/header.c ret = do_write(ff, node->info_linear, len); node 948 tools/perf/util/header.c bpf_program__bpil_offs_to_addr(node->info_linear); node 983 tools/perf/util/header.c struct btf_node *node; node 985 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); node 986 tools/perf/util/header.c next = rb_next(&node->rb_node); node 987 tools/perf/util/header.c ret = do_write(ff, &node->id, node 988 tools/perf/util/header.c sizeof(u32) * 2 + node->data_size); node 1238 tools/perf/util/header.c n->node = idx; node 1256 tools/perf/util/header.c return na->node - nb->node; node 1360 tools/perf/util/header.c _W(node) node 1544 tools/perf/util/header.c struct bpf_prog_info_node *node; node 1546 tools/perf/util/header.c node = rb_entry(next, struct bpf_prog_info_node, rb_node); node 1547 tools/perf/util/header.c next = rb_next(&node->rb_node); node 1549 tools/perf/util/header.c bpf_event__print_bpf_prog_info(&node->info_linear->info, node 1568 tools/perf/util/header.c struct btf_node *node; node 1570 tools/perf/util/header.c node = rb_entry(next, struct btf_node, rb_node); node 1571 tools/perf/util/header.c next = rb_next(&node->rb_node); node 1572 tools/perf/util/header.c fprintf(fp, "# btf info of id %u\n", node->id); node 1730 tools/perf/util/header.c n->node, n->mem_total, n->mem_free); node 1732 tools/perf/util/header.c fprintf(fp, "# node%u cpu list : ", n->node); node 1867 tools/perf/util/header.c fprintf(fp, "# %3" PRIu64 " [%s]: %s\n", n->node, buf_size, buf_map); node 2347 tools/perf/util/header.c if (do_read_u32(ff, &n->node)) node 2623 tools/perf/util/header.c _R(node) node 2748 tools/perf/util/header.c struct btf_node *node = NULL; node 2770 tools/perf/util/header.c node = malloc(sizeof(struct btf_node) + data_size); node 2771 tools/perf/util/header.c if (!node) node 2774 tools/perf/util/header.c node->id = id; node 2775 tools/perf/util/header.c node->data_size = data_size; node 2777 tools/perf/util/header.c if (__do_read(ff, node->data, data_size)) node 2780 tools/perf/util/header.c perf_env__insert_btf(env, node); node 2781 tools/perf/util/header.c node = NULL; node 2787 tools/perf/util/header.c free(node); node 321 tools/perf/util/hist.c struct rb_node *node = rb_first_cached(&he->hroot_out); node 322 tools/perf/util/hist.c while (node) { node 323 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); node 324 tools/perf/util/hist.c node = rb_next(node); node 474 tools/perf/util/hist.c INIT_LIST_HEAD(&he->pairs.node); node 1041 tools/perf/util/hist.c struct callchain_cursor_node *node; node 1043 tools/perf/util/hist.c node = callchain_cursor_current(&callchain_cursor); node 1044 tools/perf/util/hist.c if (node == NULL) node 1047 tools/perf/util/hist.c return fill_callchain_info(al, node, iter->hide_unresolved); node 1470 tools/perf/util/hist.c struct perf_hpp_list_node *node; node 1476 tools/perf/util/hist.c list_for_each_entry(node, &hists->hpp_formats, list) { node 1478 tools/perf/util/hist.c if (node->level == 0 || node->skip) node 1482 tools/perf/util/hist.c new_he = hierarchy_insert_entry(hists, root, he, parent, &node->hpp); node 1677 tools/perf/util/hist.c struct rb_node *node; node 1680 tools/perf/util/hist.c node = rb_first_cached(&hists->entries); node 1690 tools/perf/util/hist.c while (node) { node 1691 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node); node 1692 tools/perf/util/hist.c node = rb_next(node); node 1738 tools/perf/util/hist.c struct rb_node *node; node 1742 tools/perf/util/hist.c node = rb_first_cached(root_in); node 1744 tools/perf/util/hist.c while (node) { node 1745 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node_in); node 1746 tools/perf/util/hist.c node = rb_next(node); node 1928 tools/perf/util/hist.c struct rb_node *rb_hierarchy_last(struct rb_node *node) node 1930 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); node 1933 tools/perf/util/hist.c node = rb_last(&he->hroot_out.rb_root); node 1934 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node); node 1936 tools/perf/util/hist.c return node; node 1939 tools/perf/util/hist.c struct rb_node *__rb_hierarchy_next(struct rb_node *node, enum hierarchy_move_dir hmd) node 1941 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); node 1944 tools/perf/util/hist.c node = rb_first_cached(&he->hroot_out); node 1946 tools/perf/util/hist.c node = rb_next(node); node 1948 tools/perf/util/hist.c while (node == NULL) { node 1953 tools/perf/util/hist.c node = rb_next(&he->rb_node); node 1955 tools/perf/util/hist.c return node; node 1958 tools/perf/util/hist.c struct rb_node *rb_hierarchy_prev(struct rb_node *node) node 1960 tools/perf/util/hist.c struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node); node 1962 tools/perf/util/hist.c node = rb_prev(node); node 1963 tools/perf/util/hist.c if (node) node 1964 tools/perf/util/hist.c return rb_hierarchy_last(node); node 1975 tools/perf/util/hist.c struct rb_node *node; node 1982 tools/perf/util/hist.c node = rb_first_cached(&he->hroot_out); node 1983 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); node 1985 tools/perf/util/hist.c while (node && child->filtered) { node 1986 tools/perf/util/hist.c node = rb_next(node); node 1987 tools/perf/util/hist.c child = rb_entry(node, struct hist_entry, rb_node); node 1990 tools/perf/util/hist.c if (node) node 1995 tools/perf/util/hist.c return node && percent >= limit; node 2485 tools/perf/util/hist.c list_for_each_entry(leader, &pos->pairs.head, pairs.node) { node 2565 tools/perf/util/hist.c list_del_init(&pos->pairs.node); node 2743 tools/perf/util/hist.c struct rb_node *node; node 2747 tools/perf/util/hist.c node = rb_first_cached(root); node 2748 tools/perf/util/hist.c rb_erase_cached(node, root); node 2750 tools/perf/util/hist.c he = rb_entry(node, struct hist_entry, rb_node_in); node 2767 tools/perf/util/hist.c struct perf_hpp_list_node *node, *tmp; node 2771 tools/perf/util/hist.c list_for_each_entry_safe(node, tmp, &hists->hpp_formats, list) { node 2772 tools/perf/util/hist.c perf_hpp_list__for_each_format_safe(&node->hpp, fmt, pos) { node 2776 tools/perf/util/hist.c list_del_init(&node->list); node 2777 tools/perf/util/hist.c free(node); node 544 tools/perf/util/hist.h struct rb_node *rb_hierarchy_last(struct rb_node *node); node 545 tools/perf/util/hist.h struct rb_node *__rb_hierarchy_next(struct rb_node *node, node 547 tools/perf/util/hist.h struct rb_node *rb_hierarchy_prev(struct rb_node *node); node 549 tools/perf/util/hist.h static inline struct rb_node *rb_hierarchy_next(struct rb_node *node) node 551 tools/perf/util/hist.h return __rb_hierarchy_next(node, HMD_NORMAL); node 3196 tools/perf/util/intel-pt.c INIT_LIST_HEAD(&pt->unknown_thread->node); node 18 tools/perf/util/intlist.c struct int_node *node = malloc(sizeof(*node)); node 20 tools/perf/util/intlist.c if (node != NULL) { node 21 tools/perf/util/intlist.c node->i = i; node 22 tools/perf/util/intlist.c node->priv = NULL; node 23 tools/perf/util/intlist.c rc = &node->rb_node; node 37 tools/perf/util/intlist.c struct int_node *node = container_of(rb_node, struct int_node, rb_node); node 39 tools/perf/util/intlist.c int_node__delete(node); node 45 tools/perf/util/intlist.c struct int_node *node = container_of(rb_node, struct int_node, rb_node); node 47 tools/perf/util/intlist.c return node->i - i; node 55 tools/perf/util/intlist.c void intlist__remove(struct intlist *ilist, struct int_node *node) node 57 tools/perf/util/intlist.c rblist__remove_node(&ilist->rblist, &node->rb_node); node 63 tools/perf/util/intlist.c struct int_node *node = NULL; node 75 tools/perf/util/intlist.c node = container_of(rb_node, struct int_node, rb_node); node 77 tools/perf/util/intlist.c return node; node 137 tools/perf/util/intlist.c struct int_node *node = NULL; node 142 tools/perf/util/intlist.c node = container_of(rb_node, struct int_node, rb_node); node 144 tools/perf/util/intlist.c return node; node 172 tools/perf/util/machine.c list_for_each_entry_safe(pos, n, &dsos->head, node) { node 175 tools/perf/util/machine.c list_del_init(&pos->node); node 233 tools/perf/util/machine.c list_for_each_entry_safe(thread, n, &threads->dead, node) node 234 tools/perf/util/machine.c list_del_init(&thread->node); node 379 tools/perf/util/machine.c struct rb_node *node; node 384 tools/perf/util/machine.c for (node = rb_first_cached(&machines->guests); node; node 385 tools/perf/util/machine.c node = rb_next(node)) { node 386 tools/perf/util/machine.c machine = rb_entry(node, struct machine, rb_node); node 1503 tools/perf/util/machine.c list_for_each_entry(dso, &machine->dsos.head, node) { node 1576 tools/perf/util/machine.c list_for_each_entry(dso, &machine->dsos.head, node) { node 1774 tools/perf/util/machine.c list_add_tail(&th->node, &threads->dead); node 2558 tools/perf/util/machine.c list_for_each_entry(thread, &threads->dead, node) { node 24 tools/perf/util/map.h struct list_head node; node 14 tools/perf/util/mem2node.c u64 node; node 38 tools/perf/util/mem2node.c phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node) node 42 tools/perf/util/mem2node.c entry->node = node; node 86 tools/perf/util/mem2node.c (prev->node == n->node)) { node 92 tools/perf/util/mem2node.c phys_entry__init(&entries[j++], start, bsize, n->node); node 103 tools/perf/util/mem2node.c entries[i].node, entries[i].start, entries[i].end); node 136 tools/perf/util/mem2node.c return entry ? (int) entry->node : -1; node 306 tools/perf/util/metricgroup.c struct rb_node *node, *next; node 387 tools/perf/util/metricgroup.c for (node = rb_first_cached(&groups.entries); node; node = next) { node 388 tools/perf/util/metricgroup.c struct mep *me = container_of(node, struct mep, nd); node 394 tools/perf/util/metricgroup.c next = rb_next(node); node 395 tools/perf/util/metricgroup.c rblist__remove_node(&groups, node); node 330 tools/perf/util/mmap.c static void build_node_mask(int node, cpu_set_t *mask) node 342 tools/perf/util/mmap.c if (cpu__get_node(cpu) == node) node 348 tools/perf/util/parse-events.c list_add_tail(&evsel->core.node, list); node 531 tools/perf/util/parse-events.c list_add_tail(&evsel->core.node, list); node 665 tools/perf/util/parse-events.c list_for_each_entry_safe(evsel, tmp, &new_evsels, core.node) { node 666 tools/perf/util/parse-events.c list_del_init(&evsel->core.node); node 673 tools/perf/util/parse-events.c list_for_each_entry(pos, &new_evsels, core.node) { node 1478 tools/perf/util/parse-events.c leader = list_first_entry(list, struct evsel, core.node); node 1479 tools/perf/util/parse-events.c evsel = list_last_entry(list, struct evsel, core.node); node 1575 tools/perf/util/parse-events.c leader = list_entry(list->next, struct evsel, core.node); node 2075 tools/perf/util/parse-events.c if (last->core.node.prev == &evlist->core.entries) node 2077 tools/perf/util/parse-events.c last = list_entry(last->core.node.prev, struct evsel, core.node); node 2392 tools/perf/util/parse-events.c list_for_each_entry(ent, &pcache->entries, node) { node 1047 tools/perf/util/probe-event.c struct str_node *node; node 1089 tools/perf/util/probe-event.c strlist__for_each_entry(node, vl->vars) { node 1090 tools/perf/util/probe-event.c var = strchr(node->s, '\t') + 1; node 1092 tools/perf/util/probe-event.c fprintf(stdout, "\t\t%s\n", node->s); node 2345 tools/perf/util/probe-event.c struct kprobe_blacklist_node *node; node 2348 tools/perf/util/probe-event.c node = list_first_entry(blacklist, node 2350 tools/perf/util/probe-event.c list_del_init(&node->list); node 2351 tools/perf/util/probe-event.c zfree(&node->symbol); node 2352 tools/perf/util/probe-event.c free(node); node 2358 tools/perf/util/probe-event.c struct kprobe_blacklist_node *node; node 2377 tools/perf/util/probe-event.c node = zalloc(sizeof(*node)); node 2378 tools/perf/util/probe-event.c if (!node) { node 2382 tools/perf/util/probe-event.c INIT_LIST_HEAD(&node->list); node 2383 tools/perf/util/probe-event.c list_add_tail(&node->list, blacklist); node 2384 tools/perf/util/probe-event.c if (sscanf(buf, "0x%lx-0x%lx", &node->start, &node->end) != 2) { node 2395 tools/perf/util/probe-event.c node->symbol = strdup(p); node 2396 tools/perf/util/probe-event.c if (!node->symbol) { node 2401 tools/perf/util/probe-event.c node->start, node->end, node->symbol); node 2415 tools/perf/util/probe-event.c struct kprobe_blacklist_node *node; node 2417 tools/perf/util/probe-event.c list_for_each_entry(node, blacklist, list) { node 2418 tools/perf/util/probe-event.c if (node->start <= address && address < node->end) node 2419 tools/perf/util/probe-event.c return node; node 3262 tools/perf/util/probe-event.c struct str_node *node; node 3298 tools/perf/util/probe-event.c strlist__for_each_entry(node, entry->tevlist) { node 3300 tools/perf/util/probe-event.c ret = parse_probe_trace_command(node->s, tev); node 352 tools/perf/util/probe-file.c BUG_ON(!list_empty(&entry->node)); node 367 tools/perf/util/probe-file.c INIT_LIST_HEAD(&entry->node); node 388 tools/perf/util/probe-file.c struct str_node *node; node 400 tools/perf/util/probe-file.c strlist__for_each_entry(node, entry->tevlist) { node 402 tools/perf/util/probe-file.c ret = parse_probe_trace_command(node->s, tev); node 511 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); node 540 tools/perf/util/probe-file.c list_for_each_entry_safe(entry, n, &pcache->entries, node) { node 541 tools/perf/util/probe-file.c list_del_init(&entry->node); node 662 tools/perf/util/probe-file.c list_del_init(&entry->node); node 681 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); node 851 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); node 864 tools/perf/util/probe-file.c list_del_init(&entry->node); node 953 tools/perf/util/probe-file.c list_for_each_entry_safe(entry, tmp, &pcache->entries, node) { node 956 tools/perf/util/probe-file.c list_del_init(&entry->node); node 12 tools/perf/util/probe-file.h struct list_head node; node 36 tools/perf/util/probe-file.h list_for_each_entry(entry, &pcache->entries, node) node 131 tools/perf/util/rblist.c struct rb_node *node; node 133 tools/perf/util/rblist.c for (node = rb_first_cached(&rblist->entries); node; node 134 tools/perf/util/rblist.c node = rb_next(node)) { node 136 tools/perf/util/rblist.c return node; node 283 tools/perf/util/scripting-engines/trace-event-perl.c struct callchain_cursor_node *node; node 284 tools/perf/util/scripting-engines/trace-event-perl.c node = callchain_cursor_current(&callchain_cursor); node 285 tools/perf/util/scripting-engines/trace-event-perl.c if (!node) node 292 tools/perf/util/scripting-engines/trace-event-perl.c if (!hv_stores(elem, "ip", newSVuv(node->ip))) { node 297 tools/perf/util/scripting-engines/trace-event-perl.c if (node->sym) { node 303 tools/perf/util/scripting-engines/trace-event-perl.c if (!hv_stores(sym, "start", newSVuv(node->sym->start)) || node 304 tools/perf/util/scripting-engines/trace-event-perl.c !hv_stores(sym, "end", newSVuv(node->sym->end)) || node 305 tools/perf/util/scripting-engines/trace-event-perl.c !hv_stores(sym, "binding", newSVuv(node->sym->binding)) || node 306 tools/perf/util/scripting-engines/trace-event-perl.c !hv_stores(sym, "name", newSVpvn(node->sym->name, node 307 tools/perf/util/scripting-engines/trace-event-perl.c node->sym->namelen)) || node 315 tools/perf/util/scripting-engines/trace-event-perl.c if (node->map) { node 316 tools/perf/util/scripting-engines/trace-event-perl.c struct map *map = node->map; node 418 tools/perf/util/scripting-engines/trace-event-python.c struct callchain_cursor_node *node; node 419 tools/perf/util/scripting-engines/trace-event-python.c node = callchain_cursor_current(&callchain_cursor); node 420 tools/perf/util/scripting-engines/trace-event-python.c if (!node) node 429 tools/perf/util/scripting-engines/trace-event-python.c PyLong_FromUnsignedLongLong(node->ip)); node 431 tools/perf/util/scripting-engines/trace-event-python.c if (node->sym) { node 436 tools/perf/util/scripting-engines/trace-event-python.c PyLong_FromUnsignedLongLong(node->sym->start)); node 438 tools/perf/util/scripting-engines/trace-event-python.c PyLong_FromUnsignedLongLong(node->sym->end)); node 440 tools/perf/util/scripting-engines/trace-event-python.c _PyLong_FromLong(node->sym->binding)); node 442 tools/perf/util/scripting-engines/trace-event-python.c _PyUnicode_FromStringAndSize(node->sym->name, node 443 tools/perf/util/scripting-engines/trace-event-python.c node->sym->namelen)); node 447 tools/perf/util/scripting-engines/trace-event-python.c if (node->map) { node 448 tools/perf/util/scripting-engines/trace-event-python.c const char *dsoname = get_dsoname(node->map); node 91 tools/perf/util/sort.h struct list_head node; node 165 tools/perf/util/sort.h return !list_empty(&he->pairs.node); node 171 tools/perf/util/sort.h return list_entry(he->pairs.node.next, struct hist_entry, pairs.node); node 178 tools/perf/util/sort.h list_add_tail(&pair->pairs.node, &he->pairs.head); node 40 tools/perf/util/srcline.c struct inline_node *node) node 52 tools/perf/util/srcline.c list_add_tail(&ilist->list, &node->val); node 54 tools/perf/util/srcline.c list_add(&ilist->list, &node->val); node 279 tools/perf/util/srcline.c struct inline_node *node, node 289 tools/perf/util/srcline.c return inline_list__append(inline_sym, srcline, node); node 294 tools/perf/util/srcline.c bool unwind_inlines, struct inline_node *node, node 322 tools/perf/util/srcline.c if (node && inline_list__append_dso_a2l(dso, node, sym)) node 332 tools/perf/util/srcline.c if (node != NULL) { node 333 tools/perf/util/srcline.c if (inline_list__append_dso_a2l(dso, node, sym)) node 367 tools/perf/util/srcline.c struct inline_node *node; node 369 tools/perf/util/srcline.c node = zalloc(sizeof(*node)); node 370 tools/perf/util/srcline.c if (node == NULL) { node 375 tools/perf/util/srcline.c INIT_LIST_HEAD(&node->val); node 376 tools/perf/util/srcline.c node->addr = addr; node 378 tools/perf/util/srcline.c addr2line(dso_name, addr, NULL, NULL, dso, true, node, sym); node 379 tools/perf/util/srcline.c return node; node 409 tools/perf/util/srcline.c struct inline_node *node __maybe_unused, node 455 tools/perf/util/srcline.c struct inline_node *node; node 470 tools/perf/util/srcline.c node = zalloc(sizeof(*node)); node 471 tools/perf/util/srcline.c if (node == NULL) { node 476 tools/perf/util/srcline.c INIT_LIST_HEAD(&node->val); node 477 tools/perf/util/srcline.c node->addr = addr; node 495 tools/perf/util/srcline.c if (inline_list__append(inline_sym, srcline, node) != 0) { node 508 tools/perf/util/srcline.c return node; node 618 tools/perf/util/srcline.c struct srcline_node *i, *node; node 621 tools/perf/util/srcline.c node = zalloc(sizeof(struct srcline_node)); node 622 tools/perf/util/srcline.c if (!node) { node 627 tools/perf/util/srcline.c node->addr = addr; node 628 tools/perf/util/srcline.c node->srcline = srcline; node 640 tools/perf/util/srcline.c rb_link_node(&node->rb_node, parent, p); node 641 tools/perf/util/srcline.c rb_insert_color_cached(&node->rb_node, tree, leftmost); node 689 tools/perf/util/srcline.c void inline_node__delete(struct inline_node *node) node 693 tools/perf/util/srcline.c list_for_each_entry_safe(ilist, tmp, &node->val, list) { node 702 tools/perf/util/srcline.c free(node); node 46 tools/perf/util/srcline.h void inline_node__delete(struct inline_node *node); node 552 tools/perf/util/stat-display.c alias = list_prepare_entry(counter, &(evlist->core.entries), core.node); node 553 tools/perf/util/stat-display.c list_for_each_entry_continue (alias, &evlist->core.entries, core.node) { node 19 tools/perf/util/strfilter.c static void strfilter_node__delete(struct strfilter_node *node) node 21 tools/perf/util/strfilter.c if (node) { node 22 tools/perf/util/strfilter.c if (node->p && !is_operator(*node->p)) node 23 tools/perf/util/strfilter.c zfree((char **)&node->p); node 24 tools/perf/util/strfilter.c strfilter_node__delete(node->l); node 25 tools/perf/util/strfilter.c strfilter_node__delete(node->r); node 26 tools/perf/util/strfilter.c free(node); node 70 tools/perf/util/strfilter.c struct strfilter_node *node = zalloc(sizeof(*node)); node 72 tools/perf/util/strfilter.c if (node) { node 73 tools/perf/util/strfilter.c node->p = op; node 74 tools/perf/util/strfilter.c node->l = l; node 75 tools/perf/util/strfilter.c node->r = r; node 78 tools/perf/util/strfilter.c return node; node 218 tools/perf/util/strfilter.c static bool strfilter_node__compare(struct strfilter_node *node, node 221 tools/perf/util/strfilter.c if (!node || !node->p) node 224 tools/perf/util/strfilter.c switch (*node->p) { node 226 tools/perf/util/strfilter.c return strfilter_node__compare(node->l, str) || node 227 tools/perf/util/strfilter.c strfilter_node__compare(node->r, str); node 229 tools/perf/util/strfilter.c return strfilter_node__compare(node->l, str) && node 230 tools/perf/util/strfilter.c strfilter_node__compare(node->r, str); node 232 tools/perf/util/strfilter.c return !strfilter_node__compare(node->r, str); node 234 tools/perf/util/strfilter.c return strglobmatch(str, node->p); node 246 tools/perf/util/strfilter.c static int strfilter_node__sprint(struct strfilter_node *node, char *buf); node 249 tools/perf/util/strfilter.c static int strfilter_node__sprint_pt(struct strfilter_node *node, char *buf) node 252 tools/perf/util/strfilter.c int pt = node->r ? 2 : 0; /* don't need to check node->l */ node 256 tools/perf/util/strfilter.c len = strfilter_node__sprint(node, buf); node 264 tools/perf/util/strfilter.c static int strfilter_node__sprint(struct strfilter_node *node, char *buf) node 268 tools/perf/util/strfilter.c if (!node || !node->p) node 271 tools/perf/util/strfilter.c switch (*node->p) { node 274 tools/perf/util/strfilter.c len = strfilter_node__sprint_pt(node->l, buf); node 280 tools/perf/util/strfilter.c *(buf + len++) = *node->p; node 284 tools/perf/util/strfilter.c rlen = strfilter_node__sprint_pt(node->r, buf); node 290 tools/perf/util/strfilter.c len = strlen(node->p); node 292 tools/perf/util/strfilter.c strcpy(buf, node->p); node 1437 tools/perf/util/symbol-elf.c struct list_head node; node 1443 tools/perf/util/symbol-elf.c struct list_head node; node 1459 tools/perf/util/symbol-elf.c list_for_each_entry((p), &(k)->phdrs, node) node 1481 tools/perf/util/symbol-elf.c list_add_tail(&p->node, &kci->phdrs); node 1490 tools/perf/util/symbol-elf.c list_for_each_entry_safe(p, tmp, &kci->phdrs, node) { node 1491 tools/perf/util/symbol-elf.c list_del_init(&p->node); node 1503 tools/perf/util/symbol-elf.c list_add_tail(&s->node, &kci->syms); node 1513 tools/perf/util/symbol-elf.c list_for_each_entry_safe(s, tmp, &kci->syms, node) { node 1514 tools/perf/util/symbol-elf.c list_del_init(&s->node); node 1628 tools/perf/util/symbol-elf.c list_for_each_entry(sdat, &kci->syms, node) { node 1178 tools/perf/util/symbol.c list_add(&map->node, &md->maps); node 1222 tools/perf/util/symbol.c list_add_tail(&m->node, &merged); node 1249 tools/perf/util/symbol.c old_map = list_entry(merged.next, struct map, node); node 1250 tools/perf/util/symbol.c list_del_init(&old_map->node); node 1331 tools/perf/util/symbol.c list_for_each_entry(new_map, &md.maps, node) { node 1340 tools/perf/util/symbol.c replacement_map = list_entry(md.maps.next, struct map, node); node 1344 tools/perf/util/symbol.c new_map = list_entry(md.maps.next, struct map, node); node 1345 tools/perf/util/symbol.c list_del_init(&new_map->node); node 1403 tools/perf/util/symbol.c map = list_entry(md.maps.next, struct map, node); node 1404 tools/perf/util/symbol.c list_del_init(&map->node); node 1791 tools/perf/util/symbol.c struct rb_node *node; node 1795 tools/perf/util/symbol.c for (node = maps->names.rb_node; node; ) { node 1798 tools/perf/util/symbol.c map = rb_entry(node, struct map, rb_node_name); node 1802 tools/perf/util/symbol.c node = node->rb_left; node 1804 tools/perf/util/symbol.c node = node->rb_right; node 146 tools/perf/util/thread.c if (!list_empty(&thread->node)) node 147 tools/perf/util/thread.c list_del_init(&thread->node); node 26 tools/perf/util/thread.h struct list_head node; node 410 tools/perf/util/trace-event-info.c list_for_each_entry(pos, pattrs, core.node) { node 445 tools/perf/util/trace-event-info.c list_for_each_entry(pos, pattrs, core.node) node 349 tools/perf/util/util.c struct str_node *node; node 364 tools/perf/util/util.c node = strlist__entry(tips, random() % strlist__nr_entries(tips)); node 365 tools/perf/util/util.c if (asprintf(&tip, "Tip: %s", node->s) < 0) node 2573 tools/power/x86/turbostat/turbostat.c int pkg, node, lnode, cpu, cpux; node 2590 tools/power/x86/turbostat/turbostat.c node = cpus[cpu].physical_node_id; node 2598 tools/power/x86/turbostat/turbostat.c (cpus[cpux].physical_node_id == node)) { node 122 tools/testing/nvdimm/test/nfit.c #define NFIT_DIMM_HANDLE(node, socket, imc, chan, dimm) \ node 123 tools/testing/nvdimm/test/nfit.c (((node & 0xfff) << 16) | ((socket & 0xf) << 12) \ node 30 tools/testing/radix-tree/linux.c struct radix_tree_node *node; node 38 tools/testing/radix-tree/linux.c node = cachep->objs; node 39 tools/testing/radix-tree/linux.c cachep->objs = node->parent; node 41 tools/testing/radix-tree/linux.c node->parent = NULL; node 44 tools/testing/radix-tree/linux.c node = malloc(cachep->size); node 46 tools/testing/radix-tree/linux.c cachep->ctor(node); node 51 tools/testing/radix-tree/linux.c printf("Allocating %p from slab\n", node); node 52 tools/testing/radix-tree/linux.c return node; node 66 tools/testing/radix-tree/linux.c struct radix_tree_node *node = objp; node 68 tools/testing/radix-tree/linux.c node->parent = cachep->objs; node 69 tools/testing/radix-tree/linux.c cachep->objs = node; node 249 tools/testing/radix-tree/test.c struct radix_tree_node *node = root->xa_head; node 250 tools/testing/radix-tree/test.c if (!radix_tree_is_internal_node(node)) node 252 tools/testing/radix-tree/test.c verify_node(node, tag, !!root_tag_get(root, tag)); node 273 tools/testing/radix-tree/test.c struct radix_tree_node *node = root->xa_head; node 274 tools/testing/radix-tree/test.c if (!radix_tree_is_internal_node(node)) { node 279 tools/testing/radix-tree/test.c node = entry_to_node(node); node 280 tools/testing/radix-tree/test.c assert(maxindex <= node_maxindex(node)); node 282 tools/testing/radix-tree/test.c shift = node->shift; node 45 tools/testing/selftests/bpf/test_lpm_map.c struct tlpm_node *node; node 51 tools/testing/selftests/bpf/test_lpm_map.c node = tlpm_match(list, key, n_bits); node 52 tools/testing/selftests/bpf/test_lpm_map.c if (node && node->n_bits == n_bits) { node 53 tools/testing/selftests/bpf/test_lpm_map.c memcpy(node->key, key, n); node 59 tools/testing/selftests/bpf/test_lpm_map.c node = malloc(sizeof(*node) + n); node 60 tools/testing/selftests/bpf/test_lpm_map.c assert(node); node 62 tools/testing/selftests/bpf/test_lpm_map.c node->next = list; node 63 tools/testing/selftests/bpf/test_lpm_map.c node->n_bits = n_bits; node 64 tools/testing/selftests/bpf/test_lpm_map.c memcpy(node->key, key, n); node 66 tools/testing/selftests/bpf/test_lpm_map.c return node; node 71 tools/testing/selftests/bpf/test_lpm_map.c struct tlpm_node *node; node 75 tools/testing/selftests/bpf/test_lpm_map.c while ((node = list)) { node 77 tools/testing/selftests/bpf/test_lpm_map.c free(node); node 115 tools/testing/selftests/bpf/test_lpm_map.c struct tlpm_node *node; node 121 tools/testing/selftests/bpf/test_lpm_map.c node = best->next; node 123 tools/testing/selftests/bpf/test_lpm_map.c return node; node 126 tools/testing/selftests/bpf/test_lpm_map.c for (node = list; node; node = node->next) { node 127 tools/testing/selftests/bpf/test_lpm_map.c if (node->next == best) { node 128 tools/testing/selftests/bpf/test_lpm_map.c node->next = best->next; node 169 tools/testing/selftests/kvm/lib/sparsebit.c struct node *parent; node 170 tools/testing/selftests/kvm/lib/sparsebit.c struct node *left; node 171 tools/testing/selftests/kvm/lib/sparsebit.c struct node *right; node 183 tools/testing/selftests/kvm/lib/sparsebit.c struct node *root; node 197 tools/testing/selftests/kvm/lib/sparsebit.c static sparsebit_num_t node_num_set(struct node *nodep) node 205 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_first(struct sparsebit *s) node 207 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 219 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_next(struct sparsebit *s, struct node *np) node 221 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep = np; node 247 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_prev(struct sparsebit *s, struct node *np) node 249 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep = np; node 258 tools/testing/selftests/kvm/lib/sparsebit.c return (struct node *) nodep; node 268 tools/testing/selftests/kvm/lib/sparsebit.c return (struct node *) nodep->parent; node 276 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_copy_subtree(struct node *subtree) node 278 tools/testing/selftests/kvm/lib/sparsebit.c struct node *root; node 310 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_find(struct sparsebit *s, sparsebit_idx_t idx) node 312 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 333 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_add(struct sparsebit *s, sparsebit_idx_t idx) node 335 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *parentp, *prev; node 409 tools/testing/selftests/kvm/lib/sparsebit.c static void node_rm(struct sparsebit *s, struct node *nodep) node 411 tools/testing/selftests/kvm/lib/sparsebit.c struct node *tmp; node 498 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_split(struct sparsebit *s, sparsebit_idx_t idx) node 500 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep1, *nodep2; node 599 tools/testing/selftests/kvm/lib/sparsebit.c static void node_reduce(struct sparsebit *s, struct node *nodep) node 605 tools/testing/selftests/kvm/lib/sparsebit.c struct node *prev, *next, *tmp; node 781 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 807 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 834 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 872 tools/testing/selftests/kvm/lib/sparsebit.c static void dump_nodes(FILE *stream, struct node *nodep, node 901 tools/testing/selftests/kvm/lib/sparsebit.c static inline sparsebit_idx_t node_first_set(struct node *nodep, int start) node 909 tools/testing/selftests/kvm/lib/sparsebit.c static inline sparsebit_idx_t node_first_clear(struct node *nodep, int start) node 1089 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 1103 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep1, *nodep2; node 1160 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 1170 tools/testing/selftests/kvm/lib/sparsebit.c struct node *candidate = NULL; node 1253 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep1, *nodep2; node 1374 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *next; node 1456 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *next; node 1592 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep; node 1688 tools/testing/selftests/kvm/lib/sparsebit.c struct node *nodep, *prev = NULL; node 197 tools/testing/selftests/net/reuseport_bpf_numa.c int epfd, node; node 205 tools/testing/selftests/net/reuseport_bpf_numa.c for (node = 0; node < len; ++node) { node 207 tools/testing/selftests/net/reuseport_bpf_numa.c ev.data.fd = rcv_fd[node]; node 208 tools/testing/selftests/net/reuseport_bpf_numa.c if (epoll_ctl(epfd, EPOLL_CTL_ADD, rcv_fd[node], &ev)) node 213 tools/testing/selftests/net/reuseport_bpf_numa.c for (node = 0; node < len; ++node) { node 214 tools/testing/selftests/net/reuseport_bpf_numa.c send_from_node(node, family, proto); node 215 tools/testing/selftests/net/reuseport_bpf_numa.c receive_on_node(rcv_fd, len, epfd, node, proto); node 219 tools/testing/selftests/net/reuseport_bpf_numa.c for (node = len - 1; node >= 0; --node) { node 220 tools/testing/selftests/net/reuseport_bpf_numa.c send_from_node(node, family, proto); node 221 tools/testing/selftests/net/reuseport_bpf_numa.c receive_on_node(rcv_fd, len, epfd, node, proto); node 225 tools/testing/selftests/net/reuseport_bpf_numa.c for (node = 0; node < len; ++node) node 226 tools/testing/selftests/net/reuseport_bpf_numa.c close(rcv_fd[node]); node 136 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *node, node 148 tools/testing/selftests/rseq/basic_percpu_ops_test.c newval = (intptr_t)node; node 150 tools/testing/selftests/rseq/basic_percpu_ops_test.c node->next = (struct percpu_list_node *)expect; node 198 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *node; node 200 tools/testing/selftests/rseq/basic_percpu_ops_test.c node = list->c[cpu].head; node 201 tools/testing/selftests/rseq/basic_percpu_ops_test.c if (!node) node 203 tools/testing/selftests/rseq/basic_percpu_ops_test.c list->c[cpu].head = node->next; node 204 tools/testing/selftests/rseq/basic_percpu_ops_test.c return node; node 219 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *node; node 221 tools/testing/selftests/rseq/basic_percpu_ops_test.c node = this_cpu_list_pop(list, NULL); node 223 tools/testing/selftests/rseq/basic_percpu_ops_test.c if (node) node 224 tools/testing/selftests/rseq/basic_percpu_ops_test.c this_cpu_list_push(list, node, NULL); node 253 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *node; node 257 tools/testing/selftests/rseq/basic_percpu_ops_test.c node = malloc(sizeof(*node)); node 258 tools/testing/selftests/rseq/basic_percpu_ops_test.c assert(node); node 259 tools/testing/selftests/rseq/basic_percpu_ops_test.c node->data = j; node 260 tools/testing/selftests/rseq/basic_percpu_ops_test.c node->next = list.c[i].head; node 261 tools/testing/selftests/rseq/basic_percpu_ops_test.c list.c[i].head = node; node 273 tools/testing/selftests/rseq/basic_percpu_ops_test.c struct percpu_list_node *node; node 278 tools/testing/selftests/rseq/basic_percpu_ops_test.c while ((node = __percpu_list_pop(&list, i))) { node 279 tools/testing/selftests/rseq/basic_percpu_ops_test.c sum += node->data; node 280 tools/testing/selftests/rseq/basic_percpu_ops_test.c free(node); node 514 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node, node 526 tools/testing/selftests/rseq/param_test.c newval = (intptr_t)node; node 528 tools/testing/selftests/rseq/param_test.c node->next = (struct percpu_list_node *)expect; node 546 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node = NULL; node 563 tools/testing/selftests/rseq/param_test.c node = head; node 572 tools/testing/selftests/rseq/param_test.c return node; node 581 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node; node 583 tools/testing/selftests/rseq/param_test.c node = list->c[cpu].head; node 584 tools/testing/selftests/rseq/param_test.c if (!node) node 586 tools/testing/selftests/rseq/param_test.c list->c[cpu].head = node->next; node 587 tools/testing/selftests/rseq/param_test.c return node; node 600 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node; node 602 tools/testing/selftests/rseq/param_test.c node = this_cpu_list_pop(list, NULL); node 605 tools/testing/selftests/rseq/param_test.c if (node) node 606 tools/testing/selftests/rseq/param_test.c this_cpu_list_push(list, node, NULL); node 635 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node; node 639 tools/testing/selftests/rseq/param_test.c node = malloc(sizeof(*node)); node 640 tools/testing/selftests/rseq/param_test.c assert(node); node 641 tools/testing/selftests/rseq/param_test.c node->data = j; node 642 tools/testing/selftests/rseq/param_test.c node->next = list.c[i].head; node 643 tools/testing/selftests/rseq/param_test.c list.c[i].head = node; node 667 tools/testing/selftests/rseq/param_test.c struct percpu_list_node *node; node 672 tools/testing/selftests/rseq/param_test.c while ((node = __percpu_list_pop(&list, i))) { node 673 tools/testing/selftests/rseq/param_test.c sum += node->data; node 674 tools/testing/selftests/rseq/param_test.c free(node); node 687 tools/testing/selftests/rseq/param_test.c struct percpu_buffer_node *node, node 703 tools/testing/selftests/rseq/param_test.c newval_spec = (intptr_t)node; node 787 tools/testing/selftests/rseq/param_test.c struct percpu_buffer_node *node; node 789 tools/testing/selftests/rseq/param_test.c node = this_cpu_buffer_pop(buffer, NULL); node 792 tools/testing/selftests/rseq/param_test.c if (node) { node 793 tools/testing/selftests/rseq/param_test.c if (!this_cpu_buffer_push(buffer, node, NULL)) { node 832 tools/testing/selftests/rseq/param_test.c struct percpu_buffer_node *node; node 843 tools/testing/selftests/rseq/param_test.c node = malloc(sizeof(*node)); node 844 tools/testing/selftests/rseq/param_test.c assert(node); node 845 tools/testing/selftests/rseq/param_test.c node->data = j; node 846 tools/testing/selftests/rseq/param_test.c buffer.c[i].array[j - 1] = node; node 871 tools/testing/selftests/rseq/param_test.c struct percpu_buffer_node *node; node 876 tools/testing/selftests/rseq/param_test.c while ((node = __percpu_buffer_pop(&buffer, i))) { node 877 tools/testing/selftests/rseq/param_test.c sum += node->data; node 878 tools/testing/selftests/rseq/param_test.c free(node); node 118 tools/testing/selftests/x86/test_vdso.c static long sys_getcpu(unsigned * cpu, unsigned * node, node 121 tools/testing/selftests/x86/test_vdso.c return syscall(__NR_getcpu, cpu, node, cache); node 148 tools/testing/selftests/x86/test_vdso.c unsigned node; node 157 tools/testing/selftests/x86/test_vdso.c node = node_sys; node 159 tools/testing/selftests/x86/test_vdso.c node = node_vdso; node 161 tools/testing/selftests/x86/test_vdso.c node = node_vsys; node 164 tools/testing/selftests/x86/test_vdso.c if (!ret_sys && (cpu_sys != cpu || node_sys != node)) node 166 tools/testing/selftests/x86/test_vdso.c if (!ret_vdso && (cpu_vdso != cpu || node_vdso != node)) node 168 tools/testing/selftests/x86/test_vdso.c if (!ret_vsys && (cpu_vsys != cpu || node_vsys != node)) node 174 tools/testing/selftests/x86/test_vsyscall.c static inline long sys_getcpu(unsigned * cpu, unsigned * node, node 177 tools/testing/selftests/x86/test_vsyscall.c return syscall(SYS_getcpu, cpu, node, cache); node 327 tools/testing/selftests/x86/test_vsyscall.c unsigned node = 0; node 342 tools/testing/selftests/x86/test_vsyscall.c node = node_sys; node 352 tools/testing/selftests/x86/test_vsyscall.c node = node_vdso; node 362 tools/testing/selftests/x86/test_vsyscall.c if (node_vdso != node) { node 363 tools/testing/selftests/x86/test_vsyscall.c printf("[FAIL]\tvDSO reported node %hu but should be %hu\n", node_vdso, node); node 378 tools/testing/selftests/x86/test_vsyscall.c node = node_vsys; node 388 tools/testing/selftests/x86/test_vsyscall.c if (node_vsys != node) { node 389 tools/testing/selftests/x86/test_vsyscall.c printf("[FAIL]\tvsyscall reported node %hu but should be %hu\n", node_vsys, node); node 151 tools/testing/vsock/vsock_diag_test.c struct list_head *node; node 154 tools/testing/vsock/vsock_diag_test.c list_for_each(node, head) node 147 tools/usb/usbip/libsrc/usbip_host_common.c list_add(&edev->node, &hdriver->edev_list); node 161 tools/usb/usbip/libsrc/usbip_host_common.c edev = list_entry(i, struct usbip_exported_device, node); node 275 tools/usb/usbip/libsrc/usbip_host_common.c edev = list_entry(i, struct usbip_exported_device, node); node 50 tools/usb/usbip/libsrc/usbip_host_common.h struct list_head node; node 111 tools/usb/usbip/src/usbipd.c edev = list_entry(i, struct usbip_exported_device, node); node 178 tools/usb/usbip/src/usbipd.c edev = list_entry(j, struct usbip_exported_device, node); node 198 tools/usb/usbip/src/usbipd.c edev = list_entry(j, struct usbip_exported_device, node); node 276 tools/vm/slabinfo.c int node; node 286 tools/vm/slabinfo.c node = strtoul(t, &t, 10); node 290 tools/vm/slabinfo.c numa[node] = nr; node 291 tools/vm/slabinfo.c if (node > highest_node) node 292 tools/vm/slabinfo.c highest_node = node; node 365 tools/vm/slabinfo.c int node; node 380 tools/vm/slabinfo.c for(node = 0; node <= highest_node; node++) node 381 tools/vm/slabinfo.c printf(" %4d", node); node 383 tools/vm/slabinfo.c for(node = 0; node <= highest_node; node++) node 388 tools/vm/slabinfo.c for(node = 0; node <= highest_node; node++) { node 391 tools/vm/slabinfo.c store_size(b, s->numa[node]); node 397 tools/vm/slabinfo.c for(node = 0; node <= highest_node; node++) { node 400 tools/vm/slabinfo.c store_size(b, s->numa_partial[node]); node 25 virt/kvm/vfio.c struct list_head node; node 165 virt/kvm/vfio.c list_for_each_entry(kvg, &kv->group_list, node) { node 211 virt/kvm/vfio.c list_for_each_entry(kvg, &kv->group_list, node) { node 226 virt/kvm/vfio.c list_add_tail(&kvg->node, &kv->group_list); node 251 virt/kvm/vfio.c list_for_each_entry(kvg, &kv->group_list, node) { node 256 virt/kvm/vfio.c list_del(&kvg->node); node 310 virt/kvm/vfio.c list_for_each_entry(kvg, &kv->group_list, node) { node 368 virt/kvm/vfio.c list_for_each_entry_safe(kvg, tmp, &kv->group_list, node) { node 374 virt/kvm/vfio.c list_del(&kvg->node); node 99 virt/lib/irqbypass.c list_for_each_entry(tmp, &producers, node) { node 107 virt/lib/irqbypass.c list_for_each_entry(consumer, &consumers, node) { node 119 virt/lib/irqbypass.c list_add(&producer->node, &producers); node 149 virt/lib/irqbypass.c list_for_each_entry(tmp, &producers, node) { node 153 virt/lib/irqbypass.c list_for_each_entry(consumer, &consumers, node) { node 160 virt/lib/irqbypass.c list_del(&producer->node); node 194 virt/lib/irqbypass.c list_for_each_entry(tmp, &consumers, node) { node 202 virt/lib/irqbypass.c list_for_each_entry(producer, &producers, node) { node 214 virt/lib/irqbypass.c list_add(&consumer->node, &consumers); node 244 virt/lib/irqbypass.c list_for_each_entry(tmp, &consumers, node) { node 248 virt/lib/irqbypass.c list_for_each_entry(producer, &producers, node) { node 255 virt/lib/irqbypass.c list_del(&consumer->node);