Home
last modified time | relevance | path

Searched refs:NR_CPUS (Results 1 – 200 of 322) sorted by relevance

12

/linux-4.1.27/include/linux/
Dcpumask.h15 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t;
34 #if NR_CPUS == 1
45 #define nr_cpumask_bits NR_CPUS
93 #if NR_CPUS > 1
122 #if NR_CPUS == 1
257 [0 ... BITS_TO_LONGS(NR_CPUS)-1] = 0UL \
708 extern const DECLARE_BITMAP(cpu_all_bits, NR_CPUS);
754 cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)];
765 #if NR_CPUS <= BITS_PER_LONG
768 [BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \
[all …]
Dblockgroup_lock.h18 #if NR_CPUS >= 32
20 #elif NR_CPUS >= 16
22 #elif NR_CPUS >= 8
24 #elif NR_CPUS >= 4
26 #elif NR_CPUS >= 2
Dthreads.h20 #define NR_CPUS CONFIG_NR_CPUS macro
Drelay.h66 struct rchan_buf *buf[NR_CPUS]; /* per-cpu channel buffers */
Dmm_types.h26 #define USE_SPLIT_PTE_PTLOCKS (NR_CPUS >= CONFIG_SPLIT_PTLOCK_CPUS)
/linux-4.1.27/kernel/rcu/
Dtree.h44 #if NR_CPUS <= RCU_FANOUT_1
47 # define NUM_RCU_LVL_1 (NR_CPUS)
51 #elif NR_CPUS <= RCU_FANOUT_2
54 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
55 # define NUM_RCU_LVL_2 (NR_CPUS)
58 #elif NR_CPUS <= RCU_FANOUT_3
61 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2)
62 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
63 # define NUM_RCU_LVL_3 (NR_CPUS)
65 #elif NR_CPUS <= RCU_FANOUT_4
[all …]
/linux-4.1.27/arch/powerpc/include/asm/
Dirq.h51 extern struct thread_info *critirq_ctx[NR_CPUS];
52 extern struct thread_info *dbgirq_ctx[NR_CPUS];
53 extern struct thread_info *mcheckirq_ctx[NR_CPUS];
62 extern struct thread_info *hardirq_ctx[NR_CPUS];
63 extern struct thread_info *softirq_ctx[NR_CPUS];
Dfixmap.h52 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
Dcputhreads.h48 for (i = 0; i < NR_CPUS; i += threads_per_core) { in cpu_thread_mask_to_cores()
Dkvm_host.h39 #define KVM_MAX_VCPUS NR_CPUS
40 #define KVM_MAX_VCORES NR_CPUS
/linux-4.1.27/arch/blackfin/kernel/cplb-nompu/
Dcplbmgr.c29 int nr_dcplb_miss[NR_CPUS], nr_icplb_miss[NR_CPUS];
30 int nr_dcplb_supv_miss[NR_CPUS], nr_icplb_supv_miss[NR_CPUS];
31 int nr_cplb_flush[NR_CPUS], nr_dcplb_prot[NR_CPUS];
68 static int icplb_rr_index[NR_CPUS] PDT_ATTR;
69 static int dcplb_rr_index[NR_CPUS] PDT_ATTR;
Dcplbinit.c17 struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR;
18 struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR;
/linux-4.1.27/arch/mips/paravirt/
Dparavirt-smp.c23 unsigned long paravirt_smp_sp[NR_CPUS];
24 unsigned long paravirt_smp_gp[NR_CPUS];
33 if (newval < 1 || newval >= NR_CPUS) in set_numcpus()
49 if (WARN_ON(cpunum >= NR_CPUS)) in paravirt_smp_setup()
53 for (id = 0; id < NR_CPUS; id++) { in paravirt_smp_setup()
/linux-4.1.27/arch/ia64/kernel/
Derr_inject.c43 static u64 call_start[NR_CPUS];
44 static u64 phys_addr[NR_CPUS];
45 static u64 err_type_info[NR_CPUS];
46 static u64 err_struct_info[NR_CPUS];
51 } __attribute__((__aligned__(16))) err_data_buffer[NR_CPUS];
52 static s64 status[NR_CPUS];
53 static u64 capabilities[NR_CPUS];
54 static u64 resources[NR_CPUS];
Dnuma.c27 u16 cpu_to_node_map[NR_CPUS] __cacheline_aligned;
78 for (i = 0; i < NR_CPUS; ++i) in build_cpu_to_node_map()
Dsmp.c58 } __attribute__((__aligned__(32))) local_tlb_flush_counts[NR_CPUS];
60 static DEFINE_PER_CPU_SHARED_ALIGNED(unsigned short [NR_CPUS],
Dsmpboot.c80 struct sal_to_os_boot sal_boot_rendez_state[NR_CPUS];
119 cpumask_t cpu_core_map[NR_CPUS] __cacheline_aligned;
127 volatile int ia64_cpu_to_sapicid[NR_CPUS];
513 for (cpu = 0; cpu < NR_CPUS; cpu++) { in smp_build_cpu_map()
Dtopology.c88 sysfs_cpus = kzalloc(sizeof(struct ia64_cpu) * NR_CPUS, GFP_KERNEL); in topology_init()
138 static struct cpu_cache_info all_cpu_cache_info[NR_CPUS];
Dcrash.c23 int kdump_status[NR_CPUS];
/linux-4.1.27/arch/blackfin/include/asm/
Dsmp.h26 extern unsigned long blackfin_iflush_l1_entry[NR_CPUS];
35 extern unsigned long icache_invld_count[NR_CPUS];
38 extern unsigned long dcache_invld_count[NR_CPUS];
Dcplbinit.h35 extern struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS];
36 extern struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS];
50 extern unsigned long *current_rwx_mask[NR_CPUS];
/linux-4.1.27/arch/arm/mach-shmobile/
Dheadsmp.S68 cmp r1, #NR_CPUS
85 1: .space NR_CPUS * 4
88 2: .space NR_CPUS * 4
91 3: .space NR_CPUS * 4
Dplatsmp-apmu.c31 } apmu_cpus[NR_CPUS];
/linux-4.1.27/arch/blackfin/kernel/cplb-mpu/
Dcplbmgr.c28 unsigned long *current_rwx_mask[NR_CPUS];
30 int nr_dcplb_miss[NR_CPUS], nr_icplb_miss[NR_CPUS];
31 int nr_icplb_supv_miss[NR_CPUS], nr_dcplb_prot[NR_CPUS];
32 int nr_cplb_flush[NR_CPUS];
63 static int icplb_rr_index[NR_CPUS], dcplb_rr_index[NR_CPUS];
Dcplbinit.c16 struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS];
17 struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS];
/linux-4.1.27/arch/m32r/kernel/
Dsmpboot.c83 struct cpuinfo_m32r cpu_data[NR_CPUS] __cacheline_aligned;
94 static volatile int physid_2_cpu[NR_CPUS];
98 volatile int cpu_2_physid[NR_CPUS];
169 if (nr_cpu > NR_CPUS) { in smp_prepare_cpus()
171 nr_cpu, NR_CPUS); in smp_prepare_cpus()
202 for (phys_id = 0 ; phys_id < NR_CPUS ; phys_id++) { in smp_prepare_cpus()
602 for (i = 0 ; i < NR_CPUS ; i++) { in init_cpu_to_physid()
Dsetup.c274 static struct cpu cpu_devices[NR_CPUS];
362 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
/linux-4.1.27/arch/ia64/include/asm/
Dsmp.h55 int cpu_phys_id[NR_CPUS];
60 extern cpumask_t cpu_core_map[NR_CPUS];
80 for (i = 0; i < NR_CPUS; ++i) in cpu_logical_id()
Dnuma.h25 extern u16 cpu_to_node_map[NR_CPUS] __cacheline_aligned;
51 extern struct node_cpuid_s node_cpuid[NR_CPUS];
Dmca.h62 u8 imi_rendez_checkin[NR_CPUS];
141 extern unsigned long __per_cpu_mca[NR_CPUS];
Dacpi.h131 high_cpu = min(high_cpu + reserve_cpus, NR_CPUS); in per_cpu_scan_finalize()
Dtlb.h78 extern struct ia64_tr_entry *ia64_idtrs[NR_CPUS];
/linux-4.1.27/arch/ia64/include/asm/native/
Dirq.h27 #if (NR_VECTORS + 32 * NR_CPUS) < 1024
28 #define IA64_NATIVE_NR_IRQS (NR_VECTORS + 32 * NR_CPUS)
/linux-4.1.27/drivers/hv/
Dhyperv_vmbus.h523 void *synic_message_page[NR_CPUS];
524 void *synic_event_page[NR_CPUS];
534 u32 vp_index[NR_CPUS];
540 struct tasklet_struct *event_dpc[NR_CPUS];
545 struct list_head percpu_list[NR_CPUS];
549 void *post_msg_page[NR_CPUS];
553 struct clock_event_device *clk_evt[NR_CPUS];
Dhv.c144 memset(hv_context.synic_event_page, 0, sizeof(void *) * NR_CPUS); in hv_init()
146 sizeof(void *) * NR_CPUS); in hv_init()
148 sizeof(void *) * NR_CPUS); in hv_init()
150 sizeof(int) * NR_CPUS); in hv_init()
152 sizeof(void *) * NR_CPUS); in hv_init()
154 sizeof(void *) * NR_CPUS); in hv_init()
/linux-4.1.27/arch/sparc/kernel/
Dleon_smp.c55 extern volatile unsigned long cpu_callin_map[NR_CPUS];
168 (unsigned int)nrcpu, (unsigned int)NR_CPUS, in leon_boot_cpus()
238 for (i = 0; i < NR_CPUS; i++) { in leon_smp_done()
369 unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */
370 unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */
381 register int high = NR_CPUS - 1; in leon_cross_call()
Dsmp_32.c43 volatile unsigned long cpu_callin_map[NR_CPUS] = {0,};
191 if (cpuid >= NR_CPUS) in smp_prepare_cpus()
235 if (mid < NR_CPUS) { in smp_setup_cpu_possible_map()
247 if (cpuid >= NR_CPUS) { in smp_prepare_boot_cpu()
Dsun4d_smp.c276 unsigned char processors_in[NR_CPUS]; /* Set when ipi entered. */
277 unsigned char processors_out[NR_CPUS]; /* Set when ipi exited. */
368 static int cpu_tick[NR_CPUS]; in smp4d_percpu_timer_interrupt()
408 for (i = 0; i < NR_CPUS; i++) { in sun4d_init_smp()
Dmdesc.c561 if (max_cpu > NR_CPUS) in report_platform_properties()
562 max_cpu = NR_CPUS; in report_platform_properties()
564 max_cpu = NR_CPUS; in report_platform_properties()
773 if (*id < NR_CPUS) in mark_proc_ids()
864 if (cpuid >= NR_CPUS) { in mdesc_iterate_over_cpus()
867 cpuid, NR_CPUS); in mdesc_iterate_over_cpus()
Dprom_64.c451 if (cpuid >= NR_CPUS) { in of_iterate_over_cpus()
454 cpuid, NR_CPUS); in of_iterate_over_cpus()
Dsetup_64.c357 if (cpu >= NR_CPUS) { in start_early_boot()
359 cpu, NR_CPUS); in start_early_boot()
Dsmp_64.c60 cpumask_t cpu_core_map[NR_CPUS] __read_mostly =
61 { [0 ... NR_CPUS-1] = CPU_MASK_NONE };
63 cpumask_t cpu_core_sib_map[NR_CPUS] __read_mostly = {
64 [0 ... NR_CPUS-1] = CPU_MASK_NONE };
Dirq_64.c806 void *hardirq_stack[NR_CPUS];
807 void *softirq_stack[NR_CPUS];
1033 BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > (PAGE_SIZE - 64)); in init_cpu_send_mondo_info()
/linux-4.1.27/arch/sh/kernel/
Dirq.c64 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly;
65 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly;
67 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
68 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
Dtopology.c21 cpumask_t cpu_core_map[NR_CPUS];
Dsmp.c32 int __cpu_number_map[NR_CPUS]; /* Map physical to logical */
33 int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
Dsetup.c50 struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
/linux-4.1.27/arch/metag/kernel/
Dirq.c24 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly;
25 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly;
119 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
121 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
Dtopology.c21 cpumask_t cpu_core_map[NR_CPUS];
Dsetup.c120 u8 cpu_2_hwthread_id[NR_CPUS] __read_mostly = {
121 [0 ... NR_CPUS-1] = BAD_HWTHREAD_ID
/linux-4.1.27/arch/tile/include/asm/
Dfixmap.h56 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
60 FIX_HOMECACHE_END = FIX_HOMECACHE_BEGIN+(NR_CPUS)-1,
Dsmp.h126 __cpulist_parse_crop((buf), (dst), NR_CPUS)
/linux-4.1.27/arch/sh/include/asm/
Dfixmap.h57 FIX_CMAP_END = FIX_CMAP_BEGIN + (FIX_N_COLOURS * NR_CPUS) - 1,
61 FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_TYPE_NR * NR_CPUS) - 1,
Dsmp.h18 extern int __cpu_number_map[NR_CPUS];
22 extern int __cpu_logical_map[NR_CPUS];
Dtopology.h22 extern cpumask_t cpu_core_map[NR_CPUS];
Dmmu.h41 typedef unsigned long mm_context_id_t[NR_CPUS];
/linux-4.1.27/fs/pstore/
Dinternal.h8 #if NR_CPUS <= 2 && defined(CONFIG_ARM_THUMB)
10 #elif NR_CPUS <= 4 && defined(CONFIG_ARM)
/linux-4.1.27/arch/sparc/include/asm/
Dtopology_64.h47 extern cpumask_t cpu_core_map[NR_CPUS];
48 extern cpumask_t cpu_core_sib_map[NR_CPUS];
Dirq_64.h92 extern void *hardirq_stack[NR_CPUS];
93 extern void *softirq_stack[NR_CPUS];
Dsmp_64.h34 extern cpumask_t cpu_core_map[NR_CPUS];
Dvaddrs.h40 FIX_KMAP_END = (KM_TYPE_NR * NR_CPUS),
Dswitch_to_32.h6 extern struct thread_info *current_set[NR_CPUS];
Dsmp_32.h31 extern volatile unsigned long cpu_callin_map[NR_CPUS];
Dptrace.h59 extern union global_cpu_snapshot global_cpu_snapshot[NR_CPUS];
Dtrap_block.h53 extern struct trap_per_cpu trap_block[NR_CPUS];
/linux-4.1.27/arch/mn10300/mm/
Dmmu-context.c20 unsigned long mmu_context_cache[NR_CPUS] = {
21 [0 ... NR_CPUS - 1] =
Dtlb-smp.c121 #if NR_CPUS <= BITS_PER_LONG in flush_tlb_others()
/linux-4.1.27/arch/mips/kernel/
Dsmp-cps.c29 static DECLARE_BITMAP(core_power, NR_CPUS);
61 for (v = 0; v < min_t(int, core_vpes, NR_CPUS - nvpes); v++) { in cps_smp_setup()
73 for (v = 0; v < min_t(unsigned, nvpes, NR_CPUS); v++) { in cps_smp_setup()
253 for (remote = 0; remote < NR_CPUS; remote++) { in cps_boot_secondary()
259 BUG_ON(remote >= NR_CPUS); in cps_boot_secondary()
Dsmp.c48 int __cpu_number_map[NR_CPUS]; /* Map physical to logical */
51 int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
59 cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly;
63 cpumask_t cpu_core_map[NR_CPUS] __read_mostly;
497 for (cpu = 0; cpu < NR_CPUS; cpu++) { in tick_broadcast_init()
Dsmp-cmp.c116 for (i = 1; i < NR_CPUS; i++) { in cmp_smp_setup()
Dsetup.c41 struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
753 for (; i < NR_CPUS; i++) in prefill_possible_map()
791 unsigned long kernelsp[NR_CPUS];
Dproc.c157 return i < NR_CPUS ? (void *) (i + 1) : NULL; in c_start()
/linux-4.1.27/arch/m32r/mm/
Dfault-nommu.c39 unsigned int tlb_entry_i_dat[NR_CPUS];
40 unsigned int tlb_entry_d_dat[NR_CPUS];
Dinit.c41 unsigned long mmu_context_cache_dat[NR_CPUS];
Dfault.c42 unsigned int tlb_entry_i_dat[NR_CPUS];
43 unsigned int tlb_entry_d_dat[NR_CPUS];
/linux-4.1.27/arch/powerpc/lib/
Dlocks.c34 BUG_ON(holder_cpu >= NR_CPUS); in __spin_yield()
60 BUG_ON(holder_cpu >= NR_CPUS); in __rw_yield()
/linux-4.1.27/arch/mips/include/asm/
Dsmp.h32 extern int __cpu_number_map[NR_CPUS];
36 extern int __cpu_logical_map[NR_CPUS];
Dmmu.h7 unsigned long asid[NR_CPUS];
Dfixmap.h55 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
/linux-4.1.27/arch/arm64/kernel/
Dsmp.c363 for (i = 1; (i < cpu) && (i < NR_CPUS); i++) { in of_smp_init_cpus()
395 if (cpu >= NR_CPUS) in of_smp_init_cpus()
411 if (cpu > NR_CPUS) in of_smp_init_cpus()
413 cpu, NR_CPUS); in of_smp_init_cpus()
424 for (i = 0; i < NR_CPUS; i++) in of_smp_init_cpus()
Dacpi.c118 if (enabled_cpus >= NR_CPUS) { in acpi_map_gic_cpu_interface()
120 NR_CPUS, total_cpus, mpidr); in acpi_map_gic_cpu_interface()
Dcpu_ops.c28 const struct cpu_operations *cpu_ops[NR_CPUS];
Dsmp_spin_table.c34 static phys_addr_t cpu_release_addr[NR_CPUS];
Dsetup.c371 u64 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID };
Dfpsimd.c217 t->thread.fpsimd_state.cpu = NR_CPUS; in fpsimd_flush_task_state()
/linux-4.1.27/arch/alpha/kernel/
Dsmp.c55 struct cpuinfo_alpha cpu_data[NR_CPUS];
61 } ipi_data[NR_CPUS] __cacheline_aligned;
248 for (i = 0; i < NR_CPUS; i++) { in recv_secondary_console_msg()
491 for(cpu = 0; cpu < NR_CPUS; cpu++) in smp_cpus_done()
658 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_mm()
707 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_page()
763 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_icache_user_range()
Dirq.c59 cpu = (cpu < (NR_CPUS-1) ? cpu + 1 : 0); in irq_select_affinity()
/linux-4.1.27/arch/powerpc/platforms/86xx/
Dmpc86xx_smp.c43 if (nr < 0 || nr >= NR_CPUS) in smp_86xx_release_core()
67 if (nr < 0 || nr >= NR_CPUS) in smp_86xx_kick_cpu()
/linux-4.1.27/arch/ia64/mm/
Dnuma.c30 struct node_cpuid_s node_cpuid[NR_CPUS] =
31 { [0 ... NR_CPUS-1] = { .phys_id = 0, .nid = NUMA_NO_NODE } };
/linux-4.1.27/arch/blackfin/mm/
Dinit.c29 static unsigned long exception_stack[NR_CPUS][1024];
32 struct blackfin_pda cpu_pda[NR_CPUS];
/linux-4.1.27/arch/mn10300/kernel/
Dsmp.c44 static unsigned long sleep_mode[NR_CPUS];
86 unsigned long start_stack[NR_CPUS - 1];
91 struct mn10300_cpuinfo cpu_data[NR_CPUS] __cacheline_aligned;
318 for (i = 0; i < NR_CPUS; i++) { in send_IPI_mask()
567 for (i = 0; i < NR_CPUS; i++) { in smp_init_cpus()
705 for (phy_id = 0; phy_id < NR_CPUS; phy_id++) { in smp_prepare_cpus()
1172 #if NR_CPUS == 2 in run_wakeup_cpu()
1179 #error not support NR_CPUS > 2, when CONFIG_HOTPLUG_CPU=y. in run_wakeup_cpu()
Dmn10300-watchdog.c32 unsigned int watchdog_alert_counter[NR_CPUS];
47 static unsigned int last_irq_sums[NR_CPUS];
Dirq.c19 unsigned long __mn10300_irq_enabled_epsw[NR_CPUS] __cacheline_aligned_in_smp = {
20 [0 ... NR_CPUS - 1] = EPSW_IE | EPSW_IM_7
298 for (j = 0; j < NR_CPUS; j++) in arch_show_interrupts()
Dsetup.c184 static struct cpu cpu_devices[NR_CPUS];
265 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
Dmn10300-watchdog-low.S62 cmp NR_CPUS, d1
Dgdb-io-serial.c123 for (cpu = 0; cpu < NR_CPUS; cpu++) in gdbstub_io_rx_char()
Dgdb-io-ttysm.c205 for (cpu = 0; cpu < NR_CPUS; cpu++) in gdbstub_io_rx_char()
/linux-4.1.27/arch/x86/kernel/
Dsetup_percpu.c36 unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = {
37 [0 ... NR_CPUS-1] = BOOT_PERCPU_OFFSET,
174 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
Dcpuid.c188 if (__register_chrdev(CPUID_MAJOR, 0, NR_CPUS, in cpuid_init()
222 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_init()
235 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_exit()
Dmsr.c252 if (__register_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr", &msr_fops)) { in msr_init()
283 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr"); in msr_init()
296 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr"); in msr_exit()
Dkvmclock.c223 size = PAGE_ALIGN(sizeof(struct pvclock_vsyscall_time_info)*NR_CPUS); in kvmclock_init()
282 size = PAGE_ALIGN(sizeof(struct pvclock_vsyscall_time_info)*NR_CPUS); in kvm_setup_vsyscall_timeinfo()
Dnmi_selftest.c28 static DECLARE_BITMAP(nmi_ipi_mask, NR_CPUS) __initdata;
Dvsmp_64.c186 if (setup_max_cpus != NR_CPUS) in vsmp_cap_cpus()
/linux-4.1.27/arch/powerpc/kernel/
Dirq.c558 struct thread_info *critirq_ctx[NR_CPUS] __read_mostly;
559 struct thread_info *dbgirq_ctx[NR_CPUS] __read_mostly;
560 struct thread_info *mcheckirq_ctx[NR_CPUS] __read_mostly;
598 struct thread_info *softirq_ctx[NR_CPUS] __read_mostly;
599 struct thread_info *hardirq_ctx[NR_CPUS] __read_mostly;
Didle_6xx.S191 .space 4*NR_CPUS
194 .space 4*NR_CPUS
Dsmp.c81 volatile unsigned int cpu_callin_map[NR_CPUS];
112 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_generic_kick_cpu()
348 struct thread_info *current_set[NR_CPUS];
Dtau_6xx.c36 } tau[NR_CPUS];
Dsetup_32.c49 int smp_hw_index[NR_CPUS];
/linux-4.1.27/drivers/acpi/
Dacpi_pad.c91 static unsigned long cpu_weight[NR_CPUS];
92 static int tsk_in_cpu[NR_CPUS] = {[0 ... NR_CPUS-1] = -1};
93 static DECLARE_BITMAP(pad_busy_cpus_bits, NR_CPUS);
215 static struct task_struct *ps_tsks[NR_CPUS];
/linux-4.1.27/arch/mn10300/include/asm/
Dmmu_context.h56 extern unsigned long mmu_context_cache[NR_CPUS];
106 int num_cpus = NR_CPUS, i; in init_new_context()
Dmmu.h15 unsigned long tlbpid[NR_CPUS]; /* TLB PID for this process on
Dprocessor.h59 # error Sorry, NR_CPUS should be 2 to 8
/linux-4.1.27/Documentation/
Dcputopology.txt70 [NR_CPUS-1]
75 above). [~cpu_online_mask + cpus >= NR_CPUS]
89 the kernel max which is limited to 0..31 by the NR_CPUS config option
99 In this example, the NR_CPUS config option is 128, but the kernel was
Dpreempt-locking.txt24 struct this_needs_locking tux[NR_CPUS];
107 int buf[NR_CPUS];
/linux-4.1.27/arch/arm/kernel/
Ddevtree.c82 u32 tmp_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID }; in arm_dt_init_cpu_maps()
Dcpuidle.c22 static struct cpuidle_ops cpuidle_ops[NR_CPUS];
Dsetup.c141 static struct stack stacks[NR_CPUS];
454 if (cpu >= NR_CPUS) { in cpu_init()
509 u32 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID };
Dtopology.c180 struct cputopo_arm cpu_topology[NR_CPUS];
/linux-4.1.27/arch/mips/cavium-octeon/
Dsmp.c111 for (id = 0; id < NR_CPUS; id++) { in octeon_smp_setup()
121 for (id = 0; id < NR_CPUS; id++) { in octeon_smp_setup()
138 id < num_cores && id < NR_CPUS; id++) { in octeon_smp_setup()
/linux-4.1.27/mm/
Dpercpu.c1396 ai->groups[0].cpu_map[unit] = NR_CPUS; in pcpu_alloc_alloc_info()
1462 if (gi->cpu_map[unit] != NR_CPUS) in pcpu_dump_alloc_info()
1578 pcpu_low_unit_cpu = NR_CPUS; in pcpu_setup_first_chunk()
1579 pcpu_high_unit_cpu = NR_CPUS; in pcpu_setup_first_chunk()
1589 if (cpu == NR_CPUS) in pcpu_setup_first_chunk()
1600 if (pcpu_low_unit_cpu == NR_CPUS || in pcpu_setup_first_chunk()
1603 if (pcpu_high_unit_cpu == NR_CPUS || in pcpu_setup_first_chunk()
1780 static int group_map[NR_CPUS] __initdata; in pcpu_build_alloc_info()
1781 static int group_cnt[NR_CPUS] __initdata; in pcpu_build_alloc_info()
1972 unsigned int cpu = NR_CPUS; in pcpu_embed_first_chunk()
[all …]
/linux-4.1.27/arch/alpha/include/asm/
Dmmu.h5 typedef unsigned long mm_context_t[NR_CPUS];
Dsmp.h40 extern struct cpuinfo_alpha cpu_data[NR_CPUS];
/linux-4.1.27/arch/blackfin/mach-common/
Dsmp.c46 unsigned long blackfin_iflush_l1_entry[NR_CPUS];
379 unsigned long icache_invld_count[NR_CPUS];
391 unsigned long dcache_invld_count[NR_CPUS];
/linux-4.1.27/arch/xtensa/include/asm/
Dmmu.h17 unsigned long asid[NR_CPUS];
Dfixmap.h42 (KM_TYPE_NR * NR_CPUS * DCACHE_N_COLORS) - 1,
/linux-4.1.27/arch/m32r/include/asm/
Dmmu.h16 typedef unsigned long mm_context_t[NR_CPUS];
Dsmp.h60 extern volatile int cpu_2_physid[NR_CPUS];
/linux-4.1.27/arch/arm/include/asm/
Dfixmap.h12 FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_TYPE_NR * NR_CPUS) - 1,
Dtopology.h16 extern struct cputopo_arm cpu_topology[NR_CPUS];
/linux-4.1.27/arch/arm64/include/asm/
Dsmp_plat.h40 extern u64 __cpu_logical_map[NR_CPUS];
Dcpu_ops.h66 extern const struct cpu_operations *cpu_ops[NR_CPUS];
Dtopology.h16 extern struct cpu_topology cpu_topology[NR_CPUS];
/linux-4.1.27/arch/mips/netlogic/common/
Dsmp.c178 for (i = 0; i < NR_CPUS; i++) { in nlm_smp_setup()
237 for (i = 0; i < NR_CPUS; i += NLM_THREADS_PER_CORE) { in nlm_parse_cpumask()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/
Dtracefile.h184 extern union cfs_trace_data_union (*cfs_trace_data[TCD_MAX_TYPES])[NR_CPUS];
273 extern char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
/linux-4.1.27/arch/tile/kernel/
Dsmp.c34 static unsigned long __iomem *ipi_mappings[NR_CPUS];
75 HV_Recipient recip[NR_CPUS]; in send_IPI_many()
Dsetup.c489 NR_CPUS * (PFN_UP(per_cpu_size) >> PAGE_SHIFT); in setup_memory()
851 char cpu_2_node[NR_CPUS] __write_once __attribute__((aligned(L2_CACHE_BYTES)));
885 int distance[MAX_NUMNODES][NR_CPUS]; in setup_numa_mapping()
986 static struct cpu cpu_devices[NR_CPUS];
1357 for (i = smp_height * smp_width; i < NR_CPUS; ++i) in setup_cpu_maps()
1511 unsigned long __per_cpu_offset[NR_CPUS] __write_once;
1515 static unsigned long __initdata percpu_pfn[NR_CPUS] = { 0 };
Dtlb.c36 HV_Remote_ASID asids[NR_CPUS]; in flush_tlb_mm()
Dsmpboot.c85 for (cpu = 0; cpu < NR_CPUS; ++cpu) { in smp_prepare_cpus()
/linux-4.1.27/arch/metag/include/asm/
Dtopology.h22 extern cpumask_t cpu_core_map[NR_CPUS];
Dfixmap.h45 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
/linux-4.1.27/arch/ia64/include/asm/sn/
Dnodepda.h55 struct phys_cpuid phys_cpuid[NR_CPUS];
/linux-4.1.27/arch/x86/include/asm/
Dspinlock.h148 BUILD_BUG_ON(((__ticket_t)NR_CPUS) != NR_CPUS); in arch_spin_unlock()
Dirq_vectors.h163 # define CPU_VECTOR_LIMIT (64 * NR_CPUS)
Dpercpu.h568 __typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \
569 { [0 ... NR_CPUS-1] = _initvalue }; \
574 __typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \
575 { [0 ... NR_CPUS-1] = _initvalue }; \
Dpvclock.h101 #define PVCLOCK_VSYSCALL_NR_PAGES (((NR_CPUS-1)/(PAGE_SIZE/PVTI_SIZE))+1)
Dfixmap.h95 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
/linux-4.1.27/arch/arc/include/asm/
Dmmu.h51 unsigned long asid[NR_CPUS]; /* 8 bit MMU PID + Generation cycle */
/linux-4.1.27/arch/mips/include/asm/mach-ip27/
Dtopology.h23 extern struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
/linux-4.1.27/arch/mips/netlogic/xlr/
Dwakeup.c64 for (i = 0; i < NR_CPUS; i++) { in xlr_wakeup_secondary_cpus()
/linux-4.1.27/arch/arm/vfp/
Dvfpmodule.c59 union vfp_state *vfp_current_hw_state[NR_CPUS];
86 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_force_reload()
117 vfp->hard.cpu = NR_CPUS; in vfp_thread_flush()
139 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_thread_copy()
/linux-4.1.27/arch/mips/loongson/common/
Denv.c140 if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0) in prom_init_env()
141 loongson_sysconf.nr_cpus = NR_CPUS; in prom_init_env()
/linux-4.1.27/arch/um/include/shared/
Dcommon-offsets.h14 DEFINE(UM_NR_CPUS, NR_CPUS);
/linux-4.1.27/arch/cris/arch-v32/kernel/
Dirq.c58 static unsigned long irq_regs[NR_CPUS] =
69 unsigned long cpu_irq_counters[NR_CPUS];
283 for (cpu = 0; cpu < NR_CPUS; cpu++) in crisv32_mask_irq()
Dtime.c43 unsigned long timer_regs[NR_CPUS] =
/linux-4.1.27/arch/mips/include/asm/mach-paravirt/
Dkernel-entry-init.h30 slti t1, t0, NR_CPUS
/linux-4.1.27/arch/powerpc/platforms/powermac/
Dsmp.c329 if (ncpus > NR_CPUS) in smp_psurge_probe()
330 ncpus = NR_CPUS; in smp_psurge_probe()
347 extern volatile unsigned int cpu_callin_map[NR_CPUS]; in smp_psurge_kick_cpu()
1022 for (cpu = 1; cpu < 4 && cpu < NR_CPUS; ++cpu) in pmac_setup_smp()
/linux-4.1.27/kernel/
Dsmp.c499 unsigned int setup_max_cpus = NR_CPUS;
552 int nr_cpu_ids __read_mostly = NR_CPUS;
558 nr_cpu_ids = find_last_bit(cpumask_bits(cpu_possible_mask),NR_CPUS) + 1; in setup_nr_cpu_ids()
DKconfig.hz15 environment leading to NR_CPUS * HZ number of timer interrupts
Dcpu.c743 const unsigned long cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)] = {
754 const DECLARE_BITMAP(cpu_all_bits, NR_CPUS) = CPU_BITS_ALL;
/linux-4.1.27/arch/powerpc/platforms/cell/
Dsmp.c118 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_cell_kick_cpu()
Dspu_base.c112 int nr = (NR_CPUS > 1) ? NR_CPUS : NR_CPUS + 1; in mm_needs_global_tlbie()
/linux-4.1.27/arch/parisc/kernel/
Dprocessor.c157 if (cpu_info.cpu_num >= NR_CPUS) { in processor_probe()
161 dev->hpa.start, cpu_info.cpu_num, NR_CPUS); in processor_probe()
/linux-4.1.27/arch/arm/mach-imx/
Dplatsmp.c66 for (i = ncores; i < NR_CPUS; i++) in imx_smp_init_cpus()
/linux-4.1.27/arch/powerpc/mm/
Dinit_32.c84 extern struct task_struct *current_set[NR_CPUS];
Dmmu_context_nohash.c60 static unsigned long *stale_map[NR_CPUS];
422 init_mm.context.active = NR_CPUS; in mmu_context_init()
/linux-4.1.27/arch/um/os-Linux/skas/
Dprocess.c245 #undef NR_CPUS
246 #define NR_CPUS 1 macro
247 int userspace_pid[NR_CPUS];
/linux-4.1.27/arch/mips/sibyte/sb1250/
Dsmp.c135 for (i = 1, num = 0; i < NR_CPUS; i++) { in sb1250_smp_setup()
Dirq.c157 for (i=0; i<NR_CPUS; i++) { in ack_sb1250_irq()
/linux-4.1.27/arch/um/kernel/
Dprocess.c36 struct cpu_task cpu_tasks[NR_CPUS] = { [0 ... NR_CPUS - 1] = { -1, NULL } };
/linux-4.1.27/arch/arm64/mm/
Dcontext.c162 cpu_last_asid += NR_CPUS - 1; in __new_context()
/linux-4.1.27/kernel/trace/
Dtrace_kdb.c119 if (*cp || cpu_file >= NR_CPUS || cpu_file < 0 || in kdb_ftdump()
/linux-4.1.27/arch/cris/kernel/
Dsetup.c43 static struct cpu cpu_devices[NR_CPUS];
/linux-4.1.27/drivers/edac/
Docteon_edac-pc.c27 extern unsigned long long cache_err_dcache[NR_CPUS];
/linux-4.1.27/arch/mips/sibyte/bcm1480/
Dsmp.c147 for (i = 1, num = 0; i < NR_CPUS; i++) { in bcm1480_smp_setup()
/linux-4.1.27/arch/powerpc/platforms/85xx/
Dsmp.c200 WARN_ON(nr < 0 || nr >= NR_CPUS); in smp_85xx_kick_cpu()
201 WARN_ON(hw_cpu < 0 || hw_cpu >= NR_CPUS); in smp_85xx_kick_cpu()
/linux-4.1.27/arch/arm/common/
DbL_switcher.c140 static int bL_switcher_cpu_pairing[NR_CPUS];
270 static struct bL_thread bL_threads[NR_CPUS];
384 static unsigned int bL_switcher_cpu_original_cluster[NR_CPUS];
/linux-4.1.27/arch/tile/mm/
Dhomecache.c168 BUG_ON(home < 0 || home >= NR_CPUS); in homecache_finv_page_va()
321 BUG_ON(home < 0 || home >= NR_CPUS || in pte_set_home()
/linux-4.1.27/arch/microblaze/kernel/cpu/
Dmb.c140 return i < NR_CPUS ? (void *) (i + 1) : NULL; in c_start()
/linux-4.1.27/arch/x86/kernel/apic/
Dhw_nmi.c32 static DECLARE_BITMAP(backtrace_mask, NR_CPUS) __read_mostly;
/linux-4.1.27/arch/sh/kernel/cpu/sh4a/
Dsmp-shx3.c63 for (i = 1, num = 0; i < NR_CPUS; i++) { in shx3_smp_setup()
/linux-4.1.27/arch/openrisc/kernel/
Dprocess.c52 struct thread_info *current_thread_info_set[NR_CPUS] = { &init_thread_info, };
/linux-4.1.27/arch/hexagon/kernel/
Dsmp.c264 for (i = 0; i < NR_CPUS; i++) in smp_start_cpus()
/linux-4.1.27/arch/powerpc/platforms/powernv/
Dsmp.c65 BUG_ON(nr < 0 || nr >= NR_CPUS); in pnv_smp_kick_cpu()
/linux-4.1.27/arch/mips/sgi-ip27/
Dip27-init.c50 struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
Dip27-smp.c78 (tot_cpus_found != NR_CPUS)) { in do_cpumask()
/linux-4.1.27/arch/powerpc/platforms/pseries/
Dsmp.c162 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_pSeries_kick_cpu()
/linux-4.1.27/arch/sh/kernel/cpu/
Dproc.c134 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
/linux-4.1.27/drivers/net/hyperv/
Dhyperv_net.h659 struct vmbus_channel *chn_table[NR_CPUS];
663 atomic_t queue_sends[NR_CPUS];
675 struct multi_send_data msd[NR_CPUS];
/linux-4.1.27/arch/s390/kernel/
Dmachine_kexec.c219 VMCOREINFO_LENGTH(lowcore_ptr, NR_CPUS); in arch_crash_save_vmcoreinfo()
/linux-4.1.27/arch/blackfin/kernel/
Dnmi.c37 static atomic_t nmi_touched[NR_CPUS];
/linux-4.1.27/arch/arc/kernel/
Dsmp.c52 for (i = 0; i < NR_CPUS; i++) in smp_init_cpus()
Dsetup.c39 struct task_struct *_current_task[NR_CPUS]; /* For stack switching */
41 struct cpuinfo_arc cpuinfo_arc700[NR_CPUS];
/linux-4.1.27/drivers/irqchip/
Dirq-mips-gic.c34 static struct gic_pcpu_mask pcpu_masks[NR_CPUS];
405 for (i = 0; i < NR_CPUS; i++) in gic_set_affinity()
576 for (i = 0; i < NR_CPUS; i++) in gic_ipi_init_one()
/linux-4.1.27/drivers/cpufreq/
Dsparc-us3-cpufreq.c189 us3_freq_table = kzalloc((NR_CPUS * sizeof(*us3_freq_table)), in us3_freq_init()
Dp4-clockmod.c52 static int has_N44_O17_errata[NR_CPUS];
/linux-4.1.27/arch/m68k/kernel/
Dsetup_no.c297 return *pos < NR_CPUS ? ((void *) 0x12345678) : NULL; in c_start()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/
Dlinux-tracefile.c50 char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
/linux-4.1.27/arch/xtensa/kernel/
Dprocess.c50 struct task_struct *current_set[NR_CPUS] = {&init_task, };
/linux-4.1.27/arch/powerpc/sysdev/xics/
Dicp-native.c50 static struct icp_ipl __iomem *icp_native_regs[NR_CPUS];
/linux-4.1.27/drivers/lguest/
Dlg.h87 struct lg_cpu cpus[NR_CPUS];

12