Home
last modified time | relevance | path

Searched refs:NUMA_NO_NODE (Results 1 – 105 of 105) sorted by relevance

/linux-4.4.14/drivers/acpi/
Dnuma.c41 = { [0 ... MAX_PXM_DOMAINS - 1] = NUMA_NO_NODE };
50 return NUMA_NO_NODE; in pxm_to_node()
63 if (pxm_to_node_map[pxm] == NUMA_NO_NODE || node < pxm_to_node_map[pxm]) in __acpi_map_pxm_to_node()
74 return NUMA_NO_NODE; in acpi_map_pxm_to_node()
78 if (node == NUMA_NO_NODE) { in acpi_map_pxm_to_node()
80 return NUMA_NO_NODE; in acpi_map_pxm_to_node()
110 if (node == NUMA_NO_NODE) in acpi_map_pxm_to_online_node()
Dacpi_memhotplug.c281 if (nid == NUMA_NO_NODE) in acpi_memory_remove_memory()
Dpci_root.c857 if (node != NUMA_NO_NODE) in acpi_pci_root_create()
Dnfit.c1559 ndr_desc->numa_node = NUMA_NO_NODE; in acpi_nfit_register_region()
/linux-4.4.14/arch/x86/mm/
Dnuma_emulation.c62 if (emu_nid_to_phys[nid] == NUMA_NO_NODE) in emu_setup_memblk()
327 emu_nid_to_phys[i] = NUMA_NO_NODE; in numa_emulation()
380 dfl_phys_nid = NUMA_NO_NODE; in numa_emulation()
382 if (emu_nid_to_phys[i] != NUMA_NO_NODE) { in numa_emulation()
384 if (dfl_phys_nid == NUMA_NO_NODE) in numa_emulation()
388 if (dfl_phys_nid == NUMA_NO_NODE) { in numa_emulation()
402 if (__apicid_to_node[i] == NUMA_NO_NODE) in numa_emulation()
412 if (emu_nid_to_phys[i] == NUMA_NO_NODE) in numa_emulation()
452 BUG_ON(nid == NUMA_NO_NODE || !node_online(nid)); in numa_add_cpu()
478 if (nid == NUMA_NO_NODE) { in numa_set_cpumask()
Dnuma.c60 [0 ... MAX_LOCAL_APIC-1] = NUMA_NO_NODE
69 return NUMA_NO_NODE; in numa_cpu_node()
78 DEFINE_EARLY_PER_CPU(int, x86_cpu_to_node_map, NUMA_NO_NODE);
105 numa_set_node(cpu, NUMA_NO_NODE); in numa_clear_node()
312 mi->blk[i].nid = NUMA_NO_NODE; in numa_cleanup_meminfo()
328 mi->blk[i].nid != NUMA_NO_NODE) in numa_nodemask_from_meminfo()
596 if (early_cpu_to_node(i) != NUMA_NO_NODE) in numa_init_array()
611 set_apicid_to_node(i, NUMA_NO_NODE); in numa_init()
652 if (nid == NUMA_NO_NODE) in numa_init()
749 if (node == NUMA_NO_NODE) in init_cpu_to_node()
[all …]
Dkasan_init_64.c28 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); in map_range()
118 NUMA_NO_NODE); in kasan_init()
Dsrat.c57 if (from_node == NUMA_NO_NODE) in acpi_numa_slit_init()
63 if (to_node == NUMA_NO_NODE) in acpi_numa_slit_init()
/linux-4.4.14/mm/kasan/
Dkasan_init.c78 early_alloc(PAGE_SIZE, NUMA_NO_NODE)); in zero_pmd_populate()
103 early_alloc(PAGE_SIZE, NUMA_NO_NODE)); in zero_pud_populate()
148 early_alloc(PAGE_SIZE, NUMA_NO_NODE)); in kasan_populate_zero_shadow()
Dkasan.c443 PAGE_KERNEL, VM_NO_GUARD, NUMA_NO_NODE, in kasan_module_alloc()
/linux-4.4.14/include/linux/
Dbootmem.h167 NUMA_NO_NODE); in memblock_virt_alloc()
176 NUMA_NO_NODE); in memblock_virt_alloc_nopanic()
189 NUMA_NO_NODE); in memblock_virt_alloc_low()
197 NUMA_NO_NODE); in memblock_virt_alloc_low_nopanic()
205 NUMA_NO_NODE); in memblock_virt_alloc_from_nopanic()
Dnuma.h13 #define NUMA_NO_NODE (-1) macro
Dkthread.h14 kthread_create_on_node(threadfn, data, NUMA_NO_NODE, namefmt, ##arg)
Dgfp.h439 if (nid == NUMA_NO_NODE) in alloc_pages_node()
Ddevice.h610 __devres_alloc_node(release, size, gfp, NUMA_NO_NODE, #release)
618 return devres_alloc_node(release, size, gfp, NUMA_NO_NODE); in devres_alloc()
Dskbuff.h815 return __alloc_skb(size, priority, 0, NUMA_NO_NODE); in alloc_skb()
856 return __alloc_skb(size, priority, SKB_ALLOC_FCLONE, NUMA_NO_NODE); in alloc_skb_fclone()
2368 return alloc_pages_node(NUMA_NO_NODE, gfp_mask, order); in __dev_alloc_pages()
Dof.h684 return NUMA_NO_NODE; in of_node_to_nid()
Dnetdevice.h600 return NUMA_NO_NODE; in netdev_queue_numa_node_read()
/linux-4.4.14/arch/ia64/mm/
Dnuma.c31 { [0 ... NR_CPUS-1] = { .phys_id = 0, .nid = NUMA_NO_NODE } };
86 unmap_cpu_from_node(cpu, NUMA_NO_NODE); in numa_clear_node()
/linux-4.4.14/arch/x86/pci/
Dacpi.c242 if (node == NUMA_NO_NODE) { in pci_acpi_root_get_node()
244 if (node != 0 && node != NUMA_NO_NODE) in pci_acpi_root_get_node()
248 if (node != NUMA_NO_NODE && !node_online(node)) in pci_acpi_root_get_node()
249 node = NUMA_NO_NODE; in pci_acpi_root_get_node()
Dbus_numa.c25 return NUMA_NO_NODE; in x86_pci_root_bus_node()
/linux-4.4.14/mm/
Dnobootmem.c136 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, in free_low_memory_core_early()
246 ptr = __alloc_memory_core_early(NUMA_NO_NODE, size, align, goal, limit); in ___alloc_bootmem_nopanic()
330 ptr = __alloc_memory_core_early(NUMA_NO_NODE, size, align, in ___alloc_bootmem_node_nopanic()
Dslob.c195 if (node != NUMA_NO_NODE) in slob_new_pages()
291 if (node != NUMA_NO_NODE && page_to_nid(sp) != node) in slob_alloc()
467 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, _RET_IP_); in __kmalloc()
473 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, caller); in __kmalloc_track_caller()
564 return slob_alloc_node(cachep, flags, NUMA_NO_NODE); in kmem_cache_alloc()
Dmemblock.c273 NUMA_NO_NODE, flags); in memblock_find_in_range()
895 nid = NUMA_NO_NODE; in __next_mem_range()
905 if (nid != NUMA_NO_NODE && nid != m_nid) in __next_mem_range()
999 nid = NUMA_NO_NODE; in __next_mem_range_rev()
1014 if (nid != NUMA_NO_NODE && nid != m_nid) in __next_mem_range_rev()
1164 return memblock_alloc_range_nid(size, align, start, end, NUMA_NO_NODE, in memblock_alloc_range()
1193 return memblock_alloc_base_nid(size, align, max_addr, NUMA_NO_NODE, in __memblock_alloc_base()
1261 nid = NUMA_NO_NODE; in memblock_virt_alloc_internal()
1283 if (nid != NUMA_NO_NODE) { in memblock_virt_alloc_internal()
1285 max_addr, NUMA_NO_NODE, in memblock_virt_alloc_internal()
Dvmalloc.c1367 return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE, in __get_vm_area()
1376 return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE, in __get_vm_area_caller()
1392 NUMA_NO_NODE, GFP_KERNEL, in get_vm_area()
1400 NUMA_NO_NODE, GFP_KERNEL, caller); in get_vm_area_caller()
1610 if (node == NUMA_NO_NODE) in __vmalloc_area_node()
1721 return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE, in __vmalloc()
1744 return __vmalloc_node_flags(size, NUMA_NO_NODE, in vmalloc()
1761 return __vmalloc_node_flags(size, NUMA_NO_NODE, in vzalloc()
1780 PAGE_KERNEL, NUMA_NO_NODE, in vmalloc_user()
1846 NUMA_NO_NODE, __builtin_return_address(0)); in vmalloc_exec()
[all …]
Dmemtest.c70 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &this_start, in do_one_pass()
Dmempolicy.c135 if (node != NUMA_NO_NODE) { in get_task_policy()
259 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE); in mpol_new()
1035 int source = NUMA_NO_NODE; in do_migrate_pages()
1070 if (source == NUMA_NO_NODE) in do_migrate_pages()
1180 nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE); in do_mbind()
1740 int nid = NUMA_NO_NODE; in offset_il_node()
1781 int w, bit = NUMA_NO_NODE; in node_random()
2479 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE); in mpol_set_shared_policy()
Dhugetlb.c1477 if (nid != NUMA_NO_NODE) in __hugetlb_alloc_buddy_huge_page()
1536 VM_WARN_ON_ONCE(nid != NUMA_NO_NODE); in __alloc_buddy_huge_page()
1615 return __alloc_buddy_huge_page(h, vma, addr, NUMA_NO_NODE); in __alloc_buddy_huge_page_with_mpol()
1663 page = __alloc_buddy_huge_page_no_mpol(h, NUMA_NO_NODE); in gather_surplus_pages()
2238 *nidp = NUMA_NO_NODE; in kobj_to_hstate()
2253 if (nid == NUMA_NO_NODE) in nr_hugepages_show_common()
2273 if (nid == NUMA_NO_NODE) { in __nr_hugepages_store_common()
2391 if (nid == NUMA_NO_NODE) in free_hugepages_show()
2416 if (nid == NUMA_NO_NODE) in surplus_hugepages_show()
2798 NUMA_NO_NODE, tmp, *length); in hugetlb_sysctl_handler_common()
Dmempool.c183 GFP_KERNEL, NUMA_NO_NODE); in mempool_create()
Dslub.c1382 if (node == NUMA_NO_NODE) in alloc_slab_page()
1778 if (node == NUMA_NO_NODE) in get_partial()
1784 if (object || node != NUMA_NO_NODE) in get_partial()
2185 if (!page || (node != NUMA_NO_NODE && page_to_nid(page) != node)) in node_match()
2365 if (node != NUMA_NO_NODE && !node_present_pages(node)) in ___slab_alloc()
2572 return slab_alloc_node(s, gfpflags, NUMA_NO_NODE, addr); in slab_alloc()
2938 p[i] = ___slab_alloc(s, flags, NUMA_NO_NODE, in kmem_cache_alloc_bulk()
Dhuge_memory.c2372 static int last_khugepaged_target_node = NUMA_NO_NODE; in khugepaged_find_target_node()
2650 int node = NUMA_NO_NODE; in khugepaged_scan_pmd()
Dpage_alloc.c3345 page = alloc_pages_node(NUMA_NO_NODE, gfp_mask, in __page_frag_refill()
3350 page = alloc_pages_node(NUMA_NO_NODE, gfp, 0); in __page_frag_refill()
4028 int best_node = NUMA_NO_NODE; in find_next_best_node()
Dslab.c3163 if (nodeid == NUMA_NO_NODE) in slab_alloc_node()
Dmemcontrol.c5647 node_online(node) ? node : NUMA_NO_NODE); in mem_cgroup_init()
/linux-4.4.14/arch/x86/include/asm/
Dnuma.h50 return NUMA_NO_NODE; in numa_cpu_node()
/linux-4.4.14/drivers/misc/
Dsram.c95 NUMA_NO_NODE, block->label); in sram_add_pool()
100 block->size, NUMA_NO_NODE); in sram_add_pool()
368 NUMA_NO_NODE, NULL); in sram_probe()
/linux-4.4.14/arch/x86/kernel/cpu/
Damd.c281 if (node != NUMA_NO_NODE && node_online(node)) in nearby_node()
286 if (node != NUMA_NO_NODE && node_online(node)) in nearby_node()
404 if (node == NUMA_NO_NODE) in srat_detect_node()
438 __apicid_to_node[ht_nodeid] != NUMA_NO_NODE) in srat_detect_node()
Dintel.c311 if (node == NUMA_NO_NODE || !node_online(node)) { in srat_detect_node()
Dperf_event_intel_uncore.c614 fake_box = uncore_alloc_box(pmu->type, NUMA_NO_NODE); in uncore_validate_group()
878 box = uncore_alloc_box(type, NUMA_NO_NODE); in uncore_pci_probe()
Dcommon.c1352 early_cpu_to_node(cpu) != NUMA_NO_NODE) in cpu_init()
/linux-4.4.14/drivers/nvdimm/
De820.c37 return NUMA_NO_NODE; in e820_range_to_nid()
/linux-4.4.14/arch/ia64/include/asm/
Dacpi.h135 if (node_cpuid[cpu].nid == NUMA_NO_NODE) { in per_cpu_scan_finalize()
/linux-4.4.14/block/
Dblk-mq-cpumap.c119 return NUMA_NO_NODE; in blk_mq_hw_queue_to_node()
Dioprio.c51 ioc = get_task_io_context(task, GFP_ATOMIC, NUMA_NO_NODE); in set_task_ioprio()
Dblk-core.c645 return blk_alloc_queue_node(gfp_mask, NUMA_NO_NODE); in blk_alloc_queue()
807 return blk_init_queue_node(rfn, lock, NUMA_NO_NODE); in blk_init_queue()
837 q->fq = blk_alloc_flush_queue(q, NUMA_NO_NODE, 0); in blk_init_allocated_queue()
Dblk-mq.c1684 if (node == NUMA_NO_NODE) in blk_mq_init_hctx()
1796 if (nr_hw_queues > 1 && hctx->numa_node == NUMA_NO_NODE) in blk_mq_init_cpu_queues()
Dgenhd.c1257 return alloc_disk_node(minors, NUMA_NO_NODE); in alloc_disk()
/linux-4.4.14/arch/unicore32/kernel/
Dmodule.c28 GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.4.14/lib/
Dcpumask.c94 return alloc_cpumask_var_node(mask, flags, NUMA_NO_NODE); in alloc_cpumask_var()
/linux-4.4.14/fs/nfs/
Dcallback.c98 return svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in nfs4_callback_up()
153 rqstp = svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in nfs41_callback_up()
/linux-4.4.14/arch/x86/kernel/
Dcheck.c94 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, in setup_bios_corruption_check()
Ddevicetree.c212 ioapic_set_alloc_attr(&tmp, NUMA_NO_NODE, it->trigger, it->polarity); in dt_irqdomain_alloc()
Dmodule.c88 PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
De820.c1146 for_each_free_mem_range(u, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, in memblock_find_dma_reserve()
/linux-4.4.14/arch/arm/kernel/
Dmodule.c44 GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
49 GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.4.14/arch/s390/include/asm/
Dpci.h205 return NUMA_NO_NODE; in __pcibus_to_node()
/linux-4.4.14/arch/sparc/kernel/
Dmodule.c32 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_map()
/linux-4.4.14/Documentation/block/
Dnull_blk.txt32 home_node=[0--nr_nodes]: Default: NUMA_NO_NODE
/linux-4.4.14/arch/mips/kernel/
Dmodule.c50 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.4.14/arch/s390/kernel/
Dcrash_dump.c508 for_each_mem_range(idx, &memblock.physmem, &oldmem_type, NUMA_NO_NODE, in get_mem_chunk_cnt()
522 for_each_mem_range(idx, &memblock.physmem, &oldmem_type, NUMA_NO_NODE, in loads_init()
Dmodule.c48 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.4.14/net/openvswitch/
Dflow_table.c92 flow->stats_last_writer = NUMA_NO_NODE; in ovs_flow_alloc()
97 node_online(0) ? 0 : NUMA_NO_NODE); in ovs_flow_alloc()
Dflow.c98 if (likely(flow->stats_last_writer != NUMA_NO_NODE) in ovs_flow_stats_update()
/linux-4.4.14/arch/x86/kernel/apic/
Dmsi.c355 return irq_domain_alloc_irqs(domain, 1, NUMA_NO_NODE, &info); in hpet_assign_irq()
Dio_apic.c898 dst->ioapic_node = NUMA_NO_NODE; in ioapic_copy_alloc_attr()
915 return (info && info->ioapic_valid) ? info->ioapic_node : NUMA_NO_NODE; in ioapic_alloc_attr_node()
2029 ioapic_set_alloc_attr(&info, NUMA_NO_NODE, 0, 0); in mp_alloc_timer_irq()
Dvector.c246 if (node != NUMA_NO_NODE && in assign_irq_vector_policy()
/linux-4.4.14/arch/arm64/kernel/
Dmodule.c42 NUMA_NO_NODE, __builtin_return_address(0)); in module_alloc()
/linux-4.4.14/kernel/
Dworkqueue.c578 if (unlikely(node == NUMA_NO_NODE)) in unbound_pwq_by_node()
3109 pool->node = NUMA_NO_NODE; in init_worker_pool()
3241 int target_node = NUMA_NO_NODE; in get_unbound_pool()
3885 rescuer = alloc_worker(NUMA_NO_NODE); in __alloc_workqueue_key()
4210 if (pool->node != NUMA_NO_NODE) in pr_cont_pool_info()
5246 node_online(node) ? node : NUMA_NO_NODE)); in wq_numa_init()
5250 if (WARN_ON(node == NUMA_NO_NODE)) { in wq_numa_init()
Dkthread.c222 return NUMA_NO_NODE; in tsk_fork_get_node()
Dfork.c1056 new_ioc = get_task_io_context(tsk, GFP_KERNEL, NUMA_NO_NODE); in copy_io()
1383 p->cpuset_mem_spread_rotor = NUMA_NO_NODE; in copy_process()
1384 p->cpuset_slab_spread_rotor = NUMA_NO_NODE; in copy_process()
Dcpuset.c2601 if (current->cpuset_mem_spread_rotor == NUMA_NO_NODE) in cpuset_mem_spread_node()
2610 if (current->cpuset_slab_spread_rotor == NUMA_NO_NODE) in cpuset_slab_spread_node()
/linux-4.4.14/drivers/base/
Dcpu.c52 if (from_nid == NUMA_NO_NODE) in cpu_subsys_online()
Dnode.c536 if (nid != NUMA_NO_NODE) in node_memory_callback()
Dcore.c1069 if (parent && (dev_to_node(dev) == NUMA_NO_NODE)) in device_add()
/linux-4.4.14/arch/x86/kernel/acpi/
Dboot.c627 node = dev ? dev_to_node(dev) : NUMA_NO_NODE; in acpi_register_gsi_ioapic()
735 set_apicid_to_node(per_cpu(x86_cpu_to_apicid, cpu), NUMA_NO_NODE); in acpi_unmap_cpu()
/linux-4.4.14/net/core/
Dskbuff.c420 skb = __alloc_skb(len, gfp_mask, SKB_ALLOC_RX, NUMA_NO_NODE); in __netdev_alloc_skb()
487 skb = __alloc_skb(len, gfp_mask, SKB_ALLOC_RX, NUMA_NO_NODE); in __napi_alloc_skb()
1021 skb_alloc_rx_flag(skb), NUMA_NO_NODE); in skb_copy()
1061 struct sk_buff *n = __alloc_skb(size, gfp_mask, flags, NUMA_NO_NODE); in __pskb_copy_fclone()
1137 gfp_mask, NUMA_NO_NODE, NULL); in pskb_expand_head()
1244 NUMA_NO_NODE); in skb_copy_expand()
3094 NUMA_NO_NODE); in skb_segment()
Ddev.c1989 NUMA_NO_NODE); in netif_reset_xps_queues_gt()
2114 NUMA_NO_NODE); in netif_set_xps_queue()
6555 netdev_queue_numa_node_write(queue, NUMA_NO_NODE); in netdev_init_one_queue()
/linux-4.4.14/arch/x86/platform/intel-mid/
Dsfi.c470 ioapic_set_alloc_attr(&info, NUMA_NO_NODE, 1, polarity); in sfi_parse_devs()
/linux-4.4.14/fs/lockd/
Dsvc.c292 nlmsvc_rqst = svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in lockd_start_svc()
/linux-4.4.14/arch/ia64/hp/common/
Dsba_iommu.c1906 if (ioc->node != NUMA_NO_NODE) in ioc_show()
2014 if (node != NUMA_NO_NODE && !node_online(node)) in sba_map_ioc_to_node()
2015 node = NUMA_NO_NODE; in sba_map_ioc_to_node()
/linux-4.4.14/drivers/gpio/
Dgpio-sta2x11.c395 err = irq_alloc_descs(-1, 384, GSTA_NR_GPIO, NUMA_NO_NODE); in gsta_probe()
Dgpio-pch.c403 irq_base = irq_alloc_descs(-1, 0, gpio_pins[chip->ioh], NUMA_NO_NODE); in pch_gpio_probe()
Dgpio-ml-ioh.c463 NUMA_NO_NODE); in ioh_gpio_probe()
/linux-4.4.14/drivers/mtd/ubi/
Dblock.c410 dev->tag_set.numa_node = NUMA_NO_NODE; in ubiblock_create()
/linux-4.4.14/drivers/md/
Ddm-stats.c283 s = dm_kvzalloc(shared_alloc_size, NUMA_NO_NODE); in dm_stats_create()
322 hi = dm_kvzalloc(s->histogram_alloc_size, NUMA_NO_NODE); in dm_stats_create()
Ddm.c2725 md->tag_set.numa_node = NUMA_NO_NODE; in dm_init_request_based_blk_mq_queue()
Draid5.c66 #define ANY_GROUP NUMA_NO_NODE
/linux-4.4.14/arch/ia64/kernel/
Dacpi.c947 if (node == NUMA_NO_NODE || !node_online(node) || in acpi_map_iosapic()
/linux-4.4.14/drivers/pci/
Dpci-sysfs.c219 if ((node < 0 && node != NUMA_NO_NODE) || node >= MAX_NUMNODES) in numa_node_store()
222 if (node != NUMA_NO_NODE && !node_online(node)) in numa_node_store()
/linux-4.4.14/drivers/block/
Dnull_blk.c69 static int home_node = NUMA_NO_NODE;
Dvirtio_blk.c632 vblk->tag_set.numa_node = NUMA_NO_NODE; in virtblk_probe()
Dloop.c1746 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
Dxen-blkfront.c793 info->tag_set.numa_node = NUMA_NO_NODE; in xlvbd_init_blk_queue()
Drbd.c3748 rbd_dev->tag_set.numa_node = NUMA_NO_NODE; in rbd_init_disk()
/linux-4.4.14/arch/parisc/kernel/
Dmodule.c222 PAGE_KERNEL_RWX, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/
Dixgbe_lib.c807 int node = NUMA_NO_NODE; in ixgbe_alloc_q_vector()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c2463 &iq->phys_addr, NULL, 0, NUMA_NO_NODE); in t4_sge_alloc_rxq()
2503 &fl->sdesc, s->stat_len, NUMA_NO_NODE); in t4_sge_alloc_rxq()
2762 NUMA_NO_NODE); in t4_sge_alloc_ofld_txq()
/linux-4.4.14/kernel/irq/
Dirqdomain.c599 virq = irq_domain_alloc_irqs(domain, 1, NUMA_NO_NODE, fwspec); in irq_create_fwspec_mapping()
Dmanage.c355 if (node != NUMA_NO_NODE) { in setup_affinity()
/linux-4.4.14/arch/sparc/mm/
Dinit_64.c2026 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &pa_start, in available_memory()
2053 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &pa_start, in reduce_memory()
/linux-4.4.14/net/sunrpc/
Dsvc.c295 return NUMA_NO_NODE; in svc_pool_map_get_node()
/linux-4.4.14/drivers/of/
Dbase.c92 return NUMA_NO_NODE; in of_node_to_nid()
/linux-4.4.14/drivers/block/mtip32xx/
Dmtip32xx.c4389 int my_node = NUMA_NO_NODE; in mtip_pci_probe()
4394 if (my_node != NUMA_NO_NODE) { in mtip_pci_probe()
/linux-4.4.14/drivers/scsi/
Dscsi_lib.c2203 shost->tag_set.numa_node = NUMA_NO_NODE; in scsi_mq_setup_tags()
/linux-4.4.14/drivers/nvme/host/
Dpci.c3301 if (node == NUMA_NO_NODE) in nvme_probe()