Home
last modified time | relevance | path

Searched refs:NUMA_NO_NODE (Results 1 – 90 of 90) sorted by relevance

/linux-4.1.27/drivers/acpi/
Dnuma.c43 = { [0 ... MAX_PXM_DOMAINS - 1] = NUMA_NO_NODE };
52 return NUMA_NO_NODE; in pxm_to_node()
65 if (pxm_to_node_map[pxm] == NUMA_NO_NODE || node < pxm_to_node_map[pxm]) in __acpi_map_pxm_to_node()
75 if (node == NUMA_NO_NODE) { in acpi_map_pxm_to_node()
77 return NUMA_NO_NODE; in acpi_map_pxm_to_node()
332 return NUMA_NO_NODE; in acpi_get_node()
Dacpi_memhotplug.c286 if (nid == NUMA_NO_NODE) in acpi_memory_remove_memory()
/linux-4.1.27/arch/x86/mm/
Dnuma_emulation.c62 if (emu_nid_to_phys[nid] == NUMA_NO_NODE) in emu_setup_memblk()
327 emu_nid_to_phys[i] = NUMA_NO_NODE; in numa_emulation()
380 dfl_phys_nid = NUMA_NO_NODE; in numa_emulation()
382 if (emu_nid_to_phys[i] != NUMA_NO_NODE) { in numa_emulation()
384 if (dfl_phys_nid == NUMA_NO_NODE) in numa_emulation()
388 if (dfl_phys_nid == NUMA_NO_NODE) { in numa_emulation()
402 if (__apicid_to_node[i] == NUMA_NO_NODE) in numa_emulation()
412 if (emu_nid_to_phys[i] == NUMA_NO_NODE) in numa_emulation()
452 BUG_ON(nid == NUMA_NO_NODE || !node_online(nid)); in numa_add_cpu()
478 if (nid == NUMA_NO_NODE) { in numa_set_cpumask()
Dnuma.c60 [0 ... MAX_LOCAL_APIC-1] = NUMA_NO_NODE
69 return NUMA_NO_NODE; in numa_cpu_node()
78 DEFINE_EARLY_PER_CPU(int, x86_cpu_to_node_map, NUMA_NO_NODE);
105 numa_set_node(cpu, NUMA_NO_NODE); in numa_clear_node()
310 mi->blk[i].nid = NUMA_NO_NODE; in numa_cleanup_meminfo()
326 mi->blk[i].nid != NUMA_NO_NODE) in numa_nodemask_from_meminfo()
594 if (early_cpu_to_node(i) != NUMA_NO_NODE) in numa_init_array()
609 set_apicid_to_node(i, NUMA_NO_NODE); in numa_init()
650 if (nid == NUMA_NO_NODE) in numa_init()
747 if (node == NUMA_NO_NODE) in init_cpu_to_node()
[all …]
Dkasan_init_64.c41 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); in map_range()
94 void *p = vmemmap_alloc_block(PAGE_SIZE, NUMA_NO_NODE); in zero_pmd_populate()
121 void *p = vmemmap_alloc_block(PAGE_SIZE, NUMA_NO_NODE); in zero_pud_populate()
146 void *p = vmemmap_alloc_block(PAGE_SIZE, NUMA_NO_NODE); in zero_pgd_populate()
230 NUMA_NO_NODE); in kasan_init()
Dsrat.c57 if (from_node == NUMA_NO_NODE) in acpi_numa_slit_init()
63 if (to_node == NUMA_NO_NODE) in acpi_numa_slit_init()
/linux-4.1.27/include/linux/
Dbootmem.h167 NUMA_NO_NODE); in memblock_virt_alloc()
176 NUMA_NO_NODE); in memblock_virt_alloc_nopanic()
189 NUMA_NO_NODE); in memblock_virt_alloc_low()
197 NUMA_NO_NODE); in memblock_virt_alloc_low_nopanic()
205 NUMA_NO_NODE); in memblock_virt_alloc_from_nopanic()
Dnuma.h13 #define NUMA_NO_NODE (-1) macro
Dskbuff.h783 return __alloc_skb(size, priority, 0, NUMA_NO_NODE); in alloc_skb()
824 return __alloc_skb(size, priority, SKB_ALLOC_FCLONE, NUMA_NO_NODE); in alloc_skb_fclone()
2216 return alloc_pages_node(NUMA_NO_NODE, gfp_mask, order); in __dev_alloc_pages()
Dof.h678 return NUMA_NO_NODE; in of_node_to_nid()
Dnetdevice.h598 return NUMA_NO_NODE; in netdev_queue_numa_node_read()
/linux-4.1.27/arch/ia64/mm/
Dnuma.c31 { [0 ... NR_CPUS-1] = { .phys_id = 0, .nid = NUMA_NO_NODE } };
91 unmap_cpu_from_node(cpu, NUMA_NO_NODE); in numa_clear_node()
/linux-4.1.27/arch/x86/pci/
Dacpi.c414 if (node == NUMA_NO_NODE) { in pci_acpi_scan_root()
416 if (node != 0 && node != NUMA_NO_NODE) in pci_acpi_scan_root()
421 if (node != NUMA_NO_NODE && !node_online(node)) in pci_acpi_scan_root()
422 node = NUMA_NO_NODE; in pci_acpi_scan_root()
490 if (bus && node != NUMA_NO_NODE) in pci_acpi_scan_root()
Dbus_numa.c25 return NUMA_NO_NODE; in x86_pci_root_bus_node()
/linux-4.1.27/mm/
Dnobootmem.c124 for_each_free_mem_range(i, NUMA_NO_NODE, &start, &end, NULL) in free_low_memory_core_early()
233 ptr = __alloc_memory_core_early(NUMA_NO_NODE, size, align, goal, limit); in ___alloc_bootmem_nopanic()
317 ptr = __alloc_memory_core_early(NUMA_NO_NODE, size, align, in ___alloc_bootmem_node_nopanic()
Dslob.c195 if (node != NUMA_NO_NODE) in slob_new_pages()
291 if (node != NUMA_NO_NODE && page_to_nid(sp) != node) in slob_alloc()
467 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, _RET_IP_); in __kmalloc()
473 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, caller); in __kmalloc_track_caller()
564 return slob_alloc_node(cachep, flags, NUMA_NO_NODE); in kmem_cache_alloc()
Dmemblock.c256 NUMA_NO_NODE); in memblock_find_in_range()
817 nid = NUMA_NO_NODE; in __next_mem_range()
827 if (nid != NUMA_NO_NODE && nid != m_nid) in __next_mem_range()
916 nid = NUMA_NO_NODE; in __next_mem_range_rev()
931 if (nid != NUMA_NO_NODE && nid != m_nid) in __next_mem_range_rev()
1075 return memblock_alloc_range_nid(size, align, start, end, NUMA_NO_NODE); in memblock_alloc_range()
1092 return memblock_alloc_base_nid(size, align, max_addr, NUMA_NO_NODE); in __memblock_alloc_base()
1158 nid = NUMA_NO_NODE; in memblock_virt_alloc_internal()
1180 if (nid != NUMA_NO_NODE) { in memblock_virt_alloc_internal()
1182 max_addr, NUMA_NO_NODE); in memblock_virt_alloc_internal()
Dvmalloc.c1365 return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE, in __get_vm_area()
1374 return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE, in __get_vm_area_caller()
1390 NUMA_NO_NODE, GFP_KERNEL, in get_vm_area()
1398 NUMA_NO_NODE, GFP_KERNEL, caller); in get_vm_area_caller()
1609 if (node == NUMA_NO_NODE) in __vmalloc_area_node()
1720 return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE, in __vmalloc()
1743 return __vmalloc_node_flags(size, NUMA_NO_NODE, in vmalloc()
1760 return __vmalloc_node_flags(size, NUMA_NO_NODE, in vzalloc()
1779 PAGE_KERNEL, NUMA_NO_NODE, in vmalloc_user()
1845 NUMA_NO_NODE, __builtin_return_address(0)); in vmalloc_exec()
[all …]
Dmemtest.c77 for_each_free_mem_range(i, NUMA_NO_NODE, &this_start, &this_end, NULL) { in do_one_pass()
Dmempolicy.c135 if (node != NUMA_NO_NODE) { in get_task_policy()
259 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE); in mpol_new()
1038 int source = NUMA_NO_NODE; in do_migrate_pages()
1073 if (source == NUMA_NO_NODE) in do_migrate_pages()
1183 nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE); in do_mbind()
1743 int nid = NUMA_NO_NODE; in offset_il_node()
1784 int w, bit = NUMA_NO_NODE; in node_random()
2482 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE); in mpol_set_shared_policy()
Dhugetlb.c1247 if (nid == NUMA_NO_NODE) in alloc_buddy_huge_page()
1328 page = alloc_buddy_huge_page(h, NUMA_NO_NODE); in gather_surplus_pages()
1501 page = alloc_buddy_huge_page(h, NUMA_NO_NODE); in alloc_huge_page()
1824 *nidp = NUMA_NO_NODE; in kobj_to_hstate()
1839 if (nid == NUMA_NO_NODE) in nr_hugepages_show_common()
1859 if (nid == NUMA_NO_NODE) { in __nr_hugepages_store_common()
1977 if (nid == NUMA_NO_NODE) in free_hugepages_show()
2002 if (nid == NUMA_NO_NODE) in surplus_hugepages_show()
2384 NUMA_NO_NODE, tmp, *length); in hugetlb_sysctl_handler_common()
Dmempool.c180 GFP_KERNEL, NUMA_NO_NODE); in mempool_create()
Dslub.c1323 if (node == NUMA_NO_NODE) in alloc_slab_page()
1735 if (node == NUMA_NO_NODE) in get_partial()
1741 if (object || node != NUMA_NO_NODE) in get_partial()
2142 if (!page || (node != NUMA_NO_NODE && page_to_nid(page) != node)) in node_match()
2330 if (node != NUMA_NO_NODE && !node_present_pages(node)) in __slab_alloc()
2515 return slab_alloc_node(s, gfpflags, NUMA_NO_NODE, addr); in slab_alloc()
Dhuge_memory.c2293 static int last_khugepaged_target_node = NUMA_NO_NODE; in khugepaged_find_target_node()
2573 int node = NUMA_NO_NODE; in khugepaged_scan_pmd()
Dslab.c3172 if (nodeid == NUMA_NO_NODE) in slab_alloc_node()
Dpage_alloc.c3604 int best_node = NUMA_NO_NODE; in find_next_best_node()
Dmemcontrol.c5796 node_online(node) ? node : NUMA_NO_NODE); in mem_cgroup_init()
/linux-4.1.27/arch/x86/include/asm/
Dnuma.h50 return NUMA_NO_NODE; in numa_cpu_node()
/linux-4.1.27/arch/x86/kernel/cpu/
Damd.c273 if (node != NUMA_NO_NODE && node_online(node)) in nearby_node()
278 if (node != NUMA_NO_NODE && node_online(node)) in nearby_node()
377 if (node == NUMA_NO_NODE) in srat_detect_node()
411 __apicid_to_node[ht_nodeid] != NUMA_NO_NODE) in srat_detect_node()
Dintel.c310 if (node == NUMA_NO_NODE || !node_online(node)) { in srat_detect_node()
Dperf_event_intel_uncore.c561 fake_box = uncore_alloc_box(pmu->type, NUMA_NO_NODE); in uncore_validate_group()
825 box = uncore_alloc_box(type, NUMA_NO_NODE); in uncore_pci_probe()
Dcommon.c1374 early_cpu_to_node(cpu) != NUMA_NO_NODE) in cpu_init()
/linux-4.1.27/block/
Dblk-mq-cpumap.c118 return NUMA_NO_NODE; in blk_mq_hw_queue_to_node()
Dioprio.c51 ioc = get_task_io_context(task, GFP_ATOMIC, NUMA_NO_NODE); in set_task_ioprio()
Dblk-core.c604 return blk_alloc_queue_node(gfp_mask, NUMA_NO_NODE); in blk_alloc_queue()
716 return blk_init_queue_node(rfn, lock, NUMA_NO_NODE); in blk_init_queue()
746 q->fq = blk_alloc_flush_queue(q, NUMA_NO_NODE, 0); in blk_init_allocated_queue()
Dblk-mq.c1656 if (node == NUMA_NO_NODE) in blk_mq_init_hctx()
1768 if (nr_hw_queues > 1 && hctx->numa_node == NUMA_NO_NODE) in blk_mq_init_cpu_queues()
Dgenhd.c1255 return alloc_disk_node(minors, NUMA_NO_NODE); in alloc_disk()
/linux-4.1.27/arch/ia64/include/asm/
Dacpi.h135 if (node_cpuid[cpu].nid == NUMA_NO_NODE) { in per_cpu_scan_finalize()
/linux-4.1.27/arch/unicore32/kernel/
Dmodule.c28 GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.1.27/lib/
Dcpumask.c94 return alloc_cpumask_var_node(mask, flags, NUMA_NO_NODE); in alloc_cpumask_var()
/linux-4.1.27/fs/nfs/
Dcallback.c98 return svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in nfs4_callback_up()
164 rqstp = svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in nfs41_callback_up()
/linux-4.1.27/arch/x86/kernel/
Dcheck.c94 for_each_free_mem_range(i, NUMA_NO_NODE, &start, &end, NULL) { in setup_bios_corruption_check()
Dmodule.c88 PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
De820.c1126 for_each_free_mem_range(u, NUMA_NO_NODE, &start, &end, NULL) { in memblock_find_dma_reserve()
/linux-4.1.27/arch/sparc/kernel/
Dmodule.c32 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_map()
/linux-4.1.27/Documentation/block/
Dnull_blk.txt32 home_node=[0--nr_nodes]: Default: NUMA_NO_NODE
/linux-4.1.27/arch/mips/kernel/
Dmodule.c50 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.1.27/arch/s390/kernel/
Dcrash_dump.c542 for_each_dump_mem_range(idx, NUMA_NO_NODE, NULL, NULL, NULL) in get_mem_chunk_cnt()
555 for_each_dump_mem_range(idx, NUMA_NO_NODE, &start, &end, NULL) { in loads_init()
Dmodule.c48 GFP_KERNEL, PAGE_KERNEL, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.1.27/net/openvswitch/
Dflow_table.c91 flow->stats_last_writer = NUMA_NO_NODE; in ovs_flow_alloc()
96 node_online(0) ? 0 : NUMA_NO_NODE); in ovs_flow_alloc()
Dflow.c96 if (likely(flow->stats_last_writer != NUMA_NO_NODE) in ovs_flow_stats_update()
/linux-4.1.27/arch/ia64/hp/common/
Dsba_iommu.c1146 if (node == NUMA_NO_NODE) in sba_alloc_coherent()
1910 if (ioc->node != NUMA_NO_NODE) in ioc_show()
2018 if (node != NUMA_NO_NODE && !node_online(node)) in sba_map_ioc_to_node()
2019 node = NUMA_NO_NODE; in sba_map_ioc_to_node()
/linux-4.1.27/arch/arm64/kernel/
Dmodule.c39 NUMA_NO_NODE, __builtin_return_address(0)); in module_alloc()
/linux-4.1.27/arch/arm/kernel/
Dmodule.c44 GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.1.27/net/core/
Dskbuff.c372 page = alloc_pages_node(NUMA_NO_NODE, gfp_mask, order); in __page_frag_refill()
377 page = alloc_pages_node(NUMA_NO_NODE, gfp, 0); in __page_frag_refill()
508 SKB_ALLOC_RX, NUMA_NO_NODE); in __alloc_rx_skb()
1069 skb_alloc_rx_flag(skb), NUMA_NO_NODE); in skb_copy()
1109 struct sk_buff *n = __alloc_skb(size, gfp_mask, flags, NUMA_NO_NODE); in __pskb_copy_fclone()
1185 gfp_mask, NUMA_NO_NODE, NULL); in pskb_expand_head()
1292 NUMA_NO_NODE); in skb_copy_expand()
3094 NUMA_NO_NODE); in skb_segment()
Ddev.c1957 NUMA_NO_NODE); in netif_reset_xps_queues_gt()
2082 NUMA_NO_NODE); in netif_set_xps_queue()
6287 netdev_queue_numa_node_write(queue, NUMA_NO_NODE); in netdev_init_one_queue()
/linux-4.1.27/drivers/base/
Dcpu.c51 if (from_nid == NUMA_NO_NODE) in cpu_subsys_online()
Dnode.c532 if (nid != NUMA_NO_NODE) in node_memory_callback()
/linux-4.1.27/mm/kasan/
Dkasan.c429 PAGE_KERNEL, VM_NO_GUARD, NUMA_NO_NODE, in kasan_module_alloc()
/linux-4.1.27/arch/x86/kernel/acpi/
Dboot.c413 node = dev ? dev_to_node(dev) : NUMA_NO_NODE; in mp_register_gsi()
770 set_apicid_to_node(per_cpu(x86_cpu_to_apicid, cpu), NUMA_NO_NODE); in acpi_unmap_cpu()
/linux-4.1.27/kernel/
Dworkqueue.c583 if (unlikely(node == NUMA_NO_NODE)) in unbound_pwq_by_node()
3146 pool->node = NUMA_NO_NODE; in init_worker_pool()
3905 rescuer = alloc_worker(NUMA_NO_NODE); in __alloc_workqueue_key()
4230 if (pool->node != NUMA_NO_NODE) in pr_cont_pool_info()
5138 node_online(node) ? node : NUMA_NO_NODE)); in wq_numa_init()
5142 if (WARN_ON(node == NUMA_NO_NODE)) { in wq_numa_init()
Dkthread.c220 return NUMA_NO_NODE; in tsk_fork_get_node()
Dfork.c1047 new_ioc = get_task_io_context(tsk, GFP_KERNEL, NUMA_NO_NODE); in copy_io()
1380 p->cpuset_mem_spread_rotor = NUMA_NO_NODE; in copy_process()
1381 p->cpuset_slab_spread_rotor = NUMA_NO_NODE; in copy_process()
Dcpuset.c2565 if (current->cpuset_mem_spread_rotor == NUMA_NO_NODE) in cpuset_mem_spread_node()
2574 if (current->cpuset_slab_spread_rotor == NUMA_NO_NODE) in cpuset_slab_spread_node()
/linux-4.1.27/arch/x86/platform/intel-mid/
Dsfi.c472 ret = mp_set_gsi_attr(irq, 1, polarity, NUMA_NO_NODE); in sfi_parse_devs()
/linux-4.1.27/fs/lockd/
Dsvc.c292 nlmsvc_rqst = svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); in lockd_start_svc()
/linux-4.1.27/drivers/gpio/
Dgpio-sta2x11.c395 err = irq_alloc_descs(-1, 384, GSTA_NR_GPIO, NUMA_NO_NODE); in gsta_probe()
Dgpio-pch.c403 irq_base = irq_alloc_descs(-1, 0, gpio_pins[chip->ioh], NUMA_NO_NODE); in pch_gpio_probe()
Dgpio-ml-ioh.c463 NUMA_NO_NODE); in ioh_gpio_probe()
/linux-4.1.27/drivers/mtd/ubi/
Dblock.c401 dev->tag_set.numa_node = NUMA_NO_NODE; in ubiblock_create()
/linux-4.1.27/drivers/block/
Dnull_blk.c76 static int home_node = NUMA_NO_NODE;
Dvirtio_blk.c633 vblk->tag_set.numa_node = NUMA_NO_NODE; in virtblk_probe()
Dloop.c1559 lo->tag_set.numa_node = NUMA_NO_NODE; in loop_add()
Dnvme-core.c2944 if (node == NUMA_NO_NODE) in nvme_probe()
Drbd.c3769 rbd_dev->tag_set.numa_node = NUMA_NO_NODE; in rbd_init_disk()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c2453 &iq->phys_addr, NULL, 0, NUMA_NO_NODE); in t4_sge_alloc_rxq()
2479 &fl->sdesc, s->stat_len, NUMA_NO_NODE); in t4_sge_alloc_rxq()
2640 NULL, 0, NUMA_NO_NODE); in t4_sge_alloc_ctrl_txq()
2693 NUMA_NO_NODE); in t4_sge_alloc_ofld_txq()
/linux-4.1.27/arch/ia64/kernel/
Dacpi.c947 if (node == NUMA_NO_NODE || !node_online(node) || in acpi_map_iosapic()
/linux-4.1.27/arch/parisc/kernel/
Dmodule.c222 PAGE_KERNEL_RWX, 0, NUMA_NO_NODE, in module_alloc()
/linux-4.1.27/kernel/irq/
Dirqdomain.c499 virq = irq_domain_alloc_irqs(domain, 1, NUMA_NO_NODE, irq_data); in irq_create_of_mapping()
Dmanage.c354 if (node != NUMA_NO_NODE) { in setup_affinity()
/linux-4.1.27/drivers/md/
Ddm-stats.c267 s = dm_kvzalloc(shared_alloc_size, NUMA_NO_NODE); in dm_stats_create()
Ddm.c2782 md->tag_set.numa_node = NUMA_NO_NODE; in dm_init_request_based_blk_mq_queue()
Draid5.c66 #define ANY_GROUP NUMA_NO_NODE
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/
Dixgbe_lib.c807 int node = NUMA_NO_NODE; in ixgbe_alloc_q_vector()
/linux-4.1.27/arch/sparc/mm/
Dinit_64.c1969 for_each_free_mem_range(i, NUMA_NO_NODE, &pa_start, &pa_end, NULL) in available_memory()
1995 for_each_free_mem_range(i, NUMA_NO_NODE, &pa_start, &pa_end, NULL) { in reduce_memory()
/linux-4.1.27/net/sunrpc/
Dsvc.c312 return NUMA_NO_NODE; in svc_pool_map_get_node()
/linux-4.1.27/drivers/of/
Dbase.c92 return NUMA_NO_NODE; in of_node_to_nid()
/linux-4.1.27/drivers/block/mtip32xx/
Dmtip32xx.c4272 int my_node = NUMA_NO_NODE; in mtip_pci_probe()
4277 if (my_node != NUMA_NO_NODE) { in mtip_pci_probe()
/linux-4.1.27/drivers/scsi/
Dscsi_lib.c2203 shost->tag_set.numa_node = NUMA_NO_NODE; in scsi_mq_setup_tags()
/linux-4.1.27/arch/x86/kernel/apic/
Dio_apic.c3039 info->node = NUMA_NO_NODE; in mp_irqdomain_map()