/linux-4.1.27/include/linux/ |
D | cpumask.h | 15 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; 34 #if NR_CPUS == 1 45 #define nr_cpumask_bits NR_CPUS 93 #if NR_CPUS > 1 122 #if NR_CPUS == 1 257 [0 ... BITS_TO_LONGS(NR_CPUS)-1] = 0UL \ 708 extern const DECLARE_BITMAP(cpu_all_bits, NR_CPUS); 754 cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)]; 765 #if NR_CPUS <= BITS_PER_LONG 768 [BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \ [all …]
|
D | blockgroup_lock.h | 18 #if NR_CPUS >= 32 20 #elif NR_CPUS >= 16 22 #elif NR_CPUS >= 8 24 #elif NR_CPUS >= 4 26 #elif NR_CPUS >= 2
|
D | threads.h | 20 #define NR_CPUS CONFIG_NR_CPUS macro
|
D | relay.h | 66 struct rchan_buf *buf[NR_CPUS]; /* per-cpu channel buffers */
|
D | mm_types.h | 26 #define USE_SPLIT_PTE_PTLOCKS (NR_CPUS >= CONFIG_SPLIT_PTLOCK_CPUS)
|
/linux-4.1.27/kernel/rcu/ |
D | tree.h | 44 #if NR_CPUS <= RCU_FANOUT_1 47 # define NUM_RCU_LVL_1 (NR_CPUS) 51 #elif NR_CPUS <= RCU_FANOUT_2 54 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 55 # define NUM_RCU_LVL_2 (NR_CPUS) 58 #elif NR_CPUS <= RCU_FANOUT_3 61 # define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2) 62 # define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1) 63 # define NUM_RCU_LVL_3 (NR_CPUS) 65 #elif NR_CPUS <= RCU_FANOUT_4 [all …]
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | irq.h | 51 extern struct thread_info *critirq_ctx[NR_CPUS]; 52 extern struct thread_info *dbgirq_ctx[NR_CPUS]; 53 extern struct thread_info *mcheckirq_ctx[NR_CPUS]; 62 extern struct thread_info *hardirq_ctx[NR_CPUS]; 63 extern struct thread_info *softirq_ctx[NR_CPUS];
|
D | fixmap.h | 52 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
|
D | cputhreads.h | 48 for (i = 0; i < NR_CPUS; i += threads_per_core) { in cpu_thread_mask_to_cores()
|
D | kvm_host.h | 39 #define KVM_MAX_VCPUS NR_CPUS 40 #define KVM_MAX_VCORES NR_CPUS
|
/linux-4.1.27/arch/blackfin/kernel/cplb-nompu/ |
D | cplbmgr.c | 29 int nr_dcplb_miss[NR_CPUS], nr_icplb_miss[NR_CPUS]; 30 int nr_dcplb_supv_miss[NR_CPUS], nr_icplb_supv_miss[NR_CPUS]; 31 int nr_cplb_flush[NR_CPUS], nr_dcplb_prot[NR_CPUS]; 68 static int icplb_rr_index[NR_CPUS] PDT_ATTR; 69 static int dcplb_rr_index[NR_CPUS] PDT_ATTR;
|
D | cplbinit.c | 17 struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR; 18 struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR;
|
/linux-4.1.27/arch/mips/paravirt/ |
D | paravirt-smp.c | 23 unsigned long paravirt_smp_sp[NR_CPUS]; 24 unsigned long paravirt_smp_gp[NR_CPUS]; 33 if (newval < 1 || newval >= NR_CPUS) in set_numcpus() 49 if (WARN_ON(cpunum >= NR_CPUS)) in paravirt_smp_setup() 53 for (id = 0; id < NR_CPUS; id++) { in paravirt_smp_setup()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | err_inject.c | 43 static u64 call_start[NR_CPUS]; 44 static u64 phys_addr[NR_CPUS]; 45 static u64 err_type_info[NR_CPUS]; 46 static u64 err_struct_info[NR_CPUS]; 51 } __attribute__((__aligned__(16))) err_data_buffer[NR_CPUS]; 52 static s64 status[NR_CPUS]; 53 static u64 capabilities[NR_CPUS]; 54 static u64 resources[NR_CPUS];
|
D | numa.c | 27 u16 cpu_to_node_map[NR_CPUS] __cacheline_aligned; 78 for (i = 0; i < NR_CPUS; ++i) in build_cpu_to_node_map()
|
D | smp.c | 58 } __attribute__((__aligned__(32))) local_tlb_flush_counts[NR_CPUS]; 60 static DEFINE_PER_CPU_SHARED_ALIGNED(unsigned short [NR_CPUS],
|
D | smpboot.c | 80 struct sal_to_os_boot sal_boot_rendez_state[NR_CPUS]; 119 cpumask_t cpu_core_map[NR_CPUS] __cacheline_aligned; 127 volatile int ia64_cpu_to_sapicid[NR_CPUS]; 513 for (cpu = 0; cpu < NR_CPUS; cpu++) { in smp_build_cpu_map()
|
D | topology.c | 88 sysfs_cpus = kzalloc(sizeof(struct ia64_cpu) * NR_CPUS, GFP_KERNEL); in topology_init() 138 static struct cpu_cache_info all_cpu_cache_info[NR_CPUS];
|
D | crash.c | 23 int kdump_status[NR_CPUS];
|
/linux-4.1.27/arch/blackfin/include/asm/ |
D | smp.h | 26 extern unsigned long blackfin_iflush_l1_entry[NR_CPUS]; 35 extern unsigned long icache_invld_count[NR_CPUS]; 38 extern unsigned long dcache_invld_count[NR_CPUS];
|
D | cplbinit.h | 35 extern struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS]; 36 extern struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS]; 50 extern unsigned long *current_rwx_mask[NR_CPUS];
|
/linux-4.1.27/arch/arm/mach-shmobile/ |
D | headsmp.S | 68 cmp r1, #NR_CPUS 85 1: .space NR_CPUS * 4 88 2: .space NR_CPUS * 4 91 3: .space NR_CPUS * 4
|
D | platsmp-apmu.c | 31 } apmu_cpus[NR_CPUS];
|
/linux-4.1.27/arch/blackfin/kernel/cplb-mpu/ |
D | cplbmgr.c | 28 unsigned long *current_rwx_mask[NR_CPUS]; 30 int nr_dcplb_miss[NR_CPUS], nr_icplb_miss[NR_CPUS]; 31 int nr_icplb_supv_miss[NR_CPUS], nr_dcplb_prot[NR_CPUS]; 32 int nr_cplb_flush[NR_CPUS]; 63 static int icplb_rr_index[NR_CPUS], dcplb_rr_index[NR_CPUS];
|
D | cplbinit.c | 16 struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS]; 17 struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS];
|
/linux-4.1.27/arch/m32r/kernel/ |
D | smpboot.c | 83 struct cpuinfo_m32r cpu_data[NR_CPUS] __cacheline_aligned; 94 static volatile int physid_2_cpu[NR_CPUS]; 98 volatile int cpu_2_physid[NR_CPUS]; 169 if (nr_cpu > NR_CPUS) { in smp_prepare_cpus() 171 nr_cpu, NR_CPUS); in smp_prepare_cpus() 202 for (phys_id = 0 ; phys_id < NR_CPUS ; phys_id++) { in smp_prepare_cpus() 602 for (i = 0 ; i < NR_CPUS ; i++) { in init_cpu_to_physid()
|
D | setup.c | 274 static struct cpu cpu_devices[NR_CPUS]; 362 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | smp.h | 55 int cpu_phys_id[NR_CPUS]; 60 extern cpumask_t cpu_core_map[NR_CPUS]; 80 for (i = 0; i < NR_CPUS; ++i) in cpu_logical_id()
|
D | numa.h | 25 extern u16 cpu_to_node_map[NR_CPUS] __cacheline_aligned; 51 extern struct node_cpuid_s node_cpuid[NR_CPUS];
|
D | mca.h | 62 u8 imi_rendez_checkin[NR_CPUS]; 141 extern unsigned long __per_cpu_mca[NR_CPUS];
|
D | acpi.h | 131 high_cpu = min(high_cpu + reserve_cpus, NR_CPUS); in per_cpu_scan_finalize()
|
D | tlb.h | 78 extern struct ia64_tr_entry *ia64_idtrs[NR_CPUS];
|
/linux-4.1.27/arch/ia64/include/asm/native/ |
D | irq.h | 27 #if (NR_VECTORS + 32 * NR_CPUS) < 1024 28 #define IA64_NATIVE_NR_IRQS (NR_VECTORS + 32 * NR_CPUS)
|
/linux-4.1.27/drivers/hv/ |
D | hyperv_vmbus.h | 523 void *synic_message_page[NR_CPUS]; 524 void *synic_event_page[NR_CPUS]; 534 u32 vp_index[NR_CPUS]; 540 struct tasklet_struct *event_dpc[NR_CPUS]; 545 struct list_head percpu_list[NR_CPUS]; 549 void *post_msg_page[NR_CPUS]; 553 struct clock_event_device *clk_evt[NR_CPUS];
|
D | hv.c | 144 memset(hv_context.synic_event_page, 0, sizeof(void *) * NR_CPUS); in hv_init() 146 sizeof(void *) * NR_CPUS); in hv_init() 148 sizeof(void *) * NR_CPUS); in hv_init() 150 sizeof(int) * NR_CPUS); in hv_init() 152 sizeof(void *) * NR_CPUS); in hv_init() 154 sizeof(void *) * NR_CPUS); in hv_init()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | leon_smp.c | 55 extern volatile unsigned long cpu_callin_map[NR_CPUS]; 168 (unsigned int)nrcpu, (unsigned int)NR_CPUS, in leon_boot_cpus() 238 for (i = 0; i < NR_CPUS; i++) { in leon_smp_done() 369 unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */ 370 unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */ 381 register int high = NR_CPUS - 1; in leon_cross_call()
|
D | smp_32.c | 43 volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; 191 if (cpuid >= NR_CPUS) in smp_prepare_cpus() 235 if (mid < NR_CPUS) { in smp_setup_cpu_possible_map() 247 if (cpuid >= NR_CPUS) { in smp_prepare_boot_cpu()
|
D | sun4d_smp.c | 276 unsigned char processors_in[NR_CPUS]; /* Set when ipi entered. */ 277 unsigned char processors_out[NR_CPUS]; /* Set when ipi exited. */ 368 static int cpu_tick[NR_CPUS]; in smp4d_percpu_timer_interrupt() 408 for (i = 0; i < NR_CPUS; i++) { in sun4d_init_smp()
|
D | mdesc.c | 561 if (max_cpu > NR_CPUS) in report_platform_properties() 562 max_cpu = NR_CPUS; in report_platform_properties() 564 max_cpu = NR_CPUS; in report_platform_properties() 773 if (*id < NR_CPUS) in mark_proc_ids() 864 if (cpuid >= NR_CPUS) { in mdesc_iterate_over_cpus() 867 cpuid, NR_CPUS); in mdesc_iterate_over_cpus()
|
D | prom_64.c | 451 if (cpuid >= NR_CPUS) { in of_iterate_over_cpus() 454 cpuid, NR_CPUS); in of_iterate_over_cpus()
|
D | setup_64.c | 357 if (cpu >= NR_CPUS) { in start_early_boot() 359 cpu, NR_CPUS); in start_early_boot()
|
D | smp_64.c | 60 cpumask_t cpu_core_map[NR_CPUS] __read_mostly = 61 { [0 ... NR_CPUS-1] = CPU_MASK_NONE }; 63 cpumask_t cpu_core_sib_map[NR_CPUS] __read_mostly = { 64 [0 ... NR_CPUS-1] = CPU_MASK_NONE };
|
D | irq_64.c | 806 void *hardirq_stack[NR_CPUS]; 807 void *softirq_stack[NR_CPUS]; 1033 BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > (PAGE_SIZE - 64)); in init_cpu_send_mondo_info()
|
/linux-4.1.27/arch/sh/kernel/ |
D | irq.c | 64 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 65 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 67 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 68 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
|
D | topology.c | 21 cpumask_t cpu_core_map[NR_CPUS];
|
D | smp.c | 32 int __cpu_number_map[NR_CPUS]; /* Map physical to logical */ 33 int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
|
D | setup.c | 50 struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
|
/linux-4.1.27/arch/metag/kernel/ |
D | irq.c | 24 static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 25 static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 119 static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 121 static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
|
D | topology.c | 21 cpumask_t cpu_core_map[NR_CPUS];
|
D | setup.c | 120 u8 cpu_2_hwthread_id[NR_CPUS] __read_mostly = { 121 [0 ... NR_CPUS-1] = BAD_HWTHREAD_ID
|
/linux-4.1.27/arch/tile/include/asm/ |
D | fixmap.h | 56 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1, 60 FIX_HOMECACHE_END = FIX_HOMECACHE_BEGIN+(NR_CPUS)-1,
|
D | smp.h | 126 __cpulist_parse_crop((buf), (dst), NR_CPUS)
|
/linux-4.1.27/arch/sh/include/asm/ |
D | fixmap.h | 57 FIX_CMAP_END = FIX_CMAP_BEGIN + (FIX_N_COLOURS * NR_CPUS) - 1, 61 FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_TYPE_NR * NR_CPUS) - 1,
|
D | smp.h | 18 extern int __cpu_number_map[NR_CPUS]; 22 extern int __cpu_logical_map[NR_CPUS];
|
D | topology.h | 22 extern cpumask_t cpu_core_map[NR_CPUS];
|
D | mmu.h | 41 typedef unsigned long mm_context_id_t[NR_CPUS];
|
/linux-4.1.27/fs/pstore/ |
D | internal.h | 8 #if NR_CPUS <= 2 && defined(CONFIG_ARM_THUMB) 10 #elif NR_CPUS <= 4 && defined(CONFIG_ARM)
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | topology_64.h | 47 extern cpumask_t cpu_core_map[NR_CPUS]; 48 extern cpumask_t cpu_core_sib_map[NR_CPUS];
|
D | irq_64.h | 92 extern void *hardirq_stack[NR_CPUS]; 93 extern void *softirq_stack[NR_CPUS];
|
D | smp_64.h | 34 extern cpumask_t cpu_core_map[NR_CPUS];
|
D | vaddrs.h | 40 FIX_KMAP_END = (KM_TYPE_NR * NR_CPUS),
|
D | switch_to_32.h | 6 extern struct thread_info *current_set[NR_CPUS];
|
D | smp_32.h | 31 extern volatile unsigned long cpu_callin_map[NR_CPUS];
|
D | ptrace.h | 59 extern union global_cpu_snapshot global_cpu_snapshot[NR_CPUS];
|
D | trap_block.h | 53 extern struct trap_per_cpu trap_block[NR_CPUS];
|
/linux-4.1.27/arch/mn10300/mm/ |
D | mmu-context.c | 20 unsigned long mmu_context_cache[NR_CPUS] = { 21 [0 ... NR_CPUS - 1] =
|
D | tlb-smp.c | 121 #if NR_CPUS <= BITS_PER_LONG in flush_tlb_others()
|
/linux-4.1.27/arch/mips/kernel/ |
D | smp-cps.c | 29 static DECLARE_BITMAP(core_power, NR_CPUS); 61 for (v = 0; v < min_t(int, core_vpes, NR_CPUS - nvpes); v++) { in cps_smp_setup() 73 for (v = 0; v < min_t(unsigned, nvpes, NR_CPUS); v++) { in cps_smp_setup() 253 for (remote = 0; remote < NR_CPUS; remote++) { in cps_boot_secondary() 259 BUG_ON(remote >= NR_CPUS); in cps_boot_secondary()
|
D | smp.c | 48 int __cpu_number_map[NR_CPUS]; /* Map physical to logical */ 51 int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */ 59 cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly; 63 cpumask_t cpu_core_map[NR_CPUS] __read_mostly; 497 for (cpu = 0; cpu < NR_CPUS; cpu++) { in tick_broadcast_init()
|
D | smp-cmp.c | 116 for (i = 1; i < NR_CPUS; i++) { in cmp_smp_setup()
|
D | setup.c | 41 struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly; 753 for (; i < NR_CPUS; i++) in prefill_possible_map() 791 unsigned long kernelsp[NR_CPUS];
|
D | proc.c | 157 return i < NR_CPUS ? (void *) (i + 1) : NULL; in c_start()
|
/linux-4.1.27/arch/m32r/mm/ |
D | fault-nommu.c | 39 unsigned int tlb_entry_i_dat[NR_CPUS]; 40 unsigned int tlb_entry_d_dat[NR_CPUS];
|
D | init.c | 41 unsigned long mmu_context_cache_dat[NR_CPUS];
|
D | fault.c | 42 unsigned int tlb_entry_i_dat[NR_CPUS]; 43 unsigned int tlb_entry_d_dat[NR_CPUS];
|
/linux-4.1.27/arch/powerpc/lib/ |
D | locks.c | 34 BUG_ON(holder_cpu >= NR_CPUS); in __spin_yield() 60 BUG_ON(holder_cpu >= NR_CPUS); in __rw_yield()
|
/linux-4.1.27/arch/mips/include/asm/ |
D | smp.h | 32 extern int __cpu_number_map[NR_CPUS]; 36 extern int __cpu_logical_map[NR_CPUS];
|
D | mmu.h | 7 unsigned long asid[NR_CPUS];
|
D | fixmap.h | 55 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
|
/linux-4.1.27/arch/arm64/kernel/ |
D | smp.c | 363 for (i = 1; (i < cpu) && (i < NR_CPUS); i++) { in of_smp_init_cpus() 395 if (cpu >= NR_CPUS) in of_smp_init_cpus() 411 if (cpu > NR_CPUS) in of_smp_init_cpus() 413 cpu, NR_CPUS); in of_smp_init_cpus() 424 for (i = 0; i < NR_CPUS; i++) in of_smp_init_cpus()
|
D | acpi.c | 118 if (enabled_cpus >= NR_CPUS) { in acpi_map_gic_cpu_interface() 120 NR_CPUS, total_cpus, mpidr); in acpi_map_gic_cpu_interface()
|
D | cpu_ops.c | 28 const struct cpu_operations *cpu_ops[NR_CPUS];
|
D | smp_spin_table.c | 34 static phys_addr_t cpu_release_addr[NR_CPUS];
|
D | setup.c | 371 u64 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID };
|
D | fpsimd.c | 217 t->thread.fpsimd_state.cpu = NR_CPUS; in fpsimd_flush_task_state()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | smp.c | 55 struct cpuinfo_alpha cpu_data[NR_CPUS]; 61 } ipi_data[NR_CPUS] __cacheline_aligned; 248 for (i = 0; i < NR_CPUS; i++) { in recv_secondary_console_msg() 491 for(cpu = 0; cpu < NR_CPUS; cpu++) in smp_cpus_done() 658 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_mm() 707 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_tlb_page() 763 for (cpu = 0; cpu < NR_CPUS; cpu++) { in flush_icache_user_range()
|
D | irq.c | 59 cpu = (cpu < (NR_CPUS-1) ? cpu + 1 : 0); in irq_select_affinity()
|
/linux-4.1.27/arch/powerpc/platforms/86xx/ |
D | mpc86xx_smp.c | 43 if (nr < 0 || nr >= NR_CPUS) in smp_86xx_release_core() 67 if (nr < 0 || nr >= NR_CPUS) in smp_86xx_kick_cpu()
|
/linux-4.1.27/arch/ia64/mm/ |
D | numa.c | 30 struct node_cpuid_s node_cpuid[NR_CPUS] = 31 { [0 ... NR_CPUS-1] = { .phys_id = 0, .nid = NUMA_NO_NODE } };
|
/linux-4.1.27/arch/blackfin/mm/ |
D | init.c | 29 static unsigned long exception_stack[NR_CPUS][1024]; 32 struct blackfin_pda cpu_pda[NR_CPUS];
|
/linux-4.1.27/arch/mn10300/kernel/ |
D | smp.c | 44 static unsigned long sleep_mode[NR_CPUS]; 86 unsigned long start_stack[NR_CPUS - 1]; 91 struct mn10300_cpuinfo cpu_data[NR_CPUS] __cacheline_aligned; 318 for (i = 0; i < NR_CPUS; i++) { in send_IPI_mask() 567 for (i = 0; i < NR_CPUS; i++) { in smp_init_cpus() 705 for (phy_id = 0; phy_id < NR_CPUS; phy_id++) { in smp_prepare_cpus() 1172 #if NR_CPUS == 2 in run_wakeup_cpu() 1179 #error not support NR_CPUS > 2, when CONFIG_HOTPLUG_CPU=y. in run_wakeup_cpu()
|
D | mn10300-watchdog.c | 32 unsigned int watchdog_alert_counter[NR_CPUS]; 47 static unsigned int last_irq_sums[NR_CPUS];
|
D | irq.c | 19 unsigned long __mn10300_irq_enabled_epsw[NR_CPUS] __cacheline_aligned_in_smp = { 20 [0 ... NR_CPUS - 1] = EPSW_IE | EPSW_IM_7 298 for (j = 0; j < NR_CPUS; j++) in arch_show_interrupts()
|
D | setup.c | 184 static struct cpu cpu_devices[NR_CPUS]; 265 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
D | mn10300-watchdog-low.S | 62 cmp NR_CPUS, d1
|
D | gdb-io-serial.c | 123 for (cpu = 0; cpu < NR_CPUS; cpu++) in gdbstub_io_rx_char()
|
D | gdb-io-ttysm.c | 205 for (cpu = 0; cpu < NR_CPUS; cpu++) in gdbstub_io_rx_char()
|
/linux-4.1.27/arch/x86/kernel/ |
D | setup_percpu.c | 36 unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = { 37 [0 ... NR_CPUS-1] = BOOT_PERCPU_OFFSET, 174 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
|
D | cpuid.c | 188 if (__register_chrdev(CPUID_MAJOR, 0, NR_CPUS, in cpuid_init() 222 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_init() 235 __unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid"); in cpuid_exit()
|
D | msr.c | 252 if (__register_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr", &msr_fops)) { in msr_init() 283 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr"); in msr_init() 296 __unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr"); in msr_exit()
|
D | kvmclock.c | 223 size = PAGE_ALIGN(sizeof(struct pvclock_vsyscall_time_info)*NR_CPUS); in kvmclock_init() 282 size = PAGE_ALIGN(sizeof(struct pvclock_vsyscall_time_info)*NR_CPUS); in kvm_setup_vsyscall_timeinfo()
|
D | nmi_selftest.c | 28 static DECLARE_BITMAP(nmi_ipi_mask, NR_CPUS) __initdata;
|
D | vsmp_64.c | 186 if (setup_max_cpus != NR_CPUS) in vsmp_cap_cpus()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | irq.c | 558 struct thread_info *critirq_ctx[NR_CPUS] __read_mostly; 559 struct thread_info *dbgirq_ctx[NR_CPUS] __read_mostly; 560 struct thread_info *mcheckirq_ctx[NR_CPUS] __read_mostly; 598 struct thread_info *softirq_ctx[NR_CPUS] __read_mostly; 599 struct thread_info *hardirq_ctx[NR_CPUS] __read_mostly;
|
D | idle_6xx.S | 191 .space 4*NR_CPUS 194 .space 4*NR_CPUS
|
D | smp.c | 81 volatile unsigned int cpu_callin_map[NR_CPUS]; 112 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_generic_kick_cpu() 348 struct thread_info *current_set[NR_CPUS];
|
D | tau_6xx.c | 36 } tau[NR_CPUS];
|
D | setup_32.c | 49 int smp_hw_index[NR_CPUS];
|
/linux-4.1.27/drivers/acpi/ |
D | acpi_pad.c | 91 static unsigned long cpu_weight[NR_CPUS]; 92 static int tsk_in_cpu[NR_CPUS] = {[0 ... NR_CPUS-1] = -1}; 93 static DECLARE_BITMAP(pad_busy_cpus_bits, NR_CPUS); 215 static struct task_struct *ps_tsks[NR_CPUS];
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | mmu_context.h | 56 extern unsigned long mmu_context_cache[NR_CPUS]; 106 int num_cpus = NR_CPUS, i; in init_new_context()
|
D | mmu.h | 15 unsigned long tlbpid[NR_CPUS]; /* TLB PID for this process on
|
D | processor.h | 59 # error Sorry, NR_CPUS should be 2 to 8
|
/linux-4.1.27/Documentation/ |
D | cputopology.txt | 70 [NR_CPUS-1] 75 above). [~cpu_online_mask + cpus >= NR_CPUS] 89 the kernel max which is limited to 0..31 by the NR_CPUS config option 99 In this example, the NR_CPUS config option is 128, but the kernel was
|
D | preempt-locking.txt | 24 struct this_needs_locking tux[NR_CPUS]; 107 int buf[NR_CPUS];
|
/linux-4.1.27/arch/arm/kernel/ |
D | devtree.c | 82 u32 tmp_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID }; in arm_dt_init_cpu_maps()
|
D | cpuidle.c | 22 static struct cpuidle_ops cpuidle_ops[NR_CPUS];
|
D | setup.c | 141 static struct stack stacks[NR_CPUS]; 454 if (cpu >= NR_CPUS) { in cpu_init() 509 u32 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID };
|
D | topology.c | 180 struct cputopo_arm cpu_topology[NR_CPUS];
|
/linux-4.1.27/arch/mips/cavium-octeon/ |
D | smp.c | 111 for (id = 0; id < NR_CPUS; id++) { in octeon_smp_setup() 121 for (id = 0; id < NR_CPUS; id++) { in octeon_smp_setup() 138 id < num_cores && id < NR_CPUS; id++) { in octeon_smp_setup()
|
/linux-4.1.27/mm/ |
D | percpu.c | 1396 ai->groups[0].cpu_map[unit] = NR_CPUS; in pcpu_alloc_alloc_info() 1462 if (gi->cpu_map[unit] != NR_CPUS) in pcpu_dump_alloc_info() 1578 pcpu_low_unit_cpu = NR_CPUS; in pcpu_setup_first_chunk() 1579 pcpu_high_unit_cpu = NR_CPUS; in pcpu_setup_first_chunk() 1589 if (cpu == NR_CPUS) in pcpu_setup_first_chunk() 1600 if (pcpu_low_unit_cpu == NR_CPUS || in pcpu_setup_first_chunk() 1603 if (pcpu_high_unit_cpu == NR_CPUS || in pcpu_setup_first_chunk() 1780 static int group_map[NR_CPUS] __initdata; in pcpu_build_alloc_info() 1781 static int group_cnt[NR_CPUS] __initdata; in pcpu_build_alloc_info() 1972 unsigned int cpu = NR_CPUS; in pcpu_embed_first_chunk() [all …]
|
/linux-4.1.27/arch/alpha/include/asm/ |
D | mmu.h | 5 typedef unsigned long mm_context_t[NR_CPUS];
|
D | smp.h | 40 extern struct cpuinfo_alpha cpu_data[NR_CPUS];
|
/linux-4.1.27/arch/blackfin/mach-common/ |
D | smp.c | 46 unsigned long blackfin_iflush_l1_entry[NR_CPUS]; 379 unsigned long icache_invld_count[NR_CPUS]; 391 unsigned long dcache_invld_count[NR_CPUS];
|
/linux-4.1.27/arch/xtensa/include/asm/ |
D | mmu.h | 17 unsigned long asid[NR_CPUS];
|
D | fixmap.h | 42 (KM_TYPE_NR * NR_CPUS * DCACHE_N_COLORS) - 1,
|
/linux-4.1.27/arch/m32r/include/asm/ |
D | mmu.h | 16 typedef unsigned long mm_context_t[NR_CPUS];
|
D | smp.h | 60 extern volatile int cpu_2_physid[NR_CPUS];
|
/linux-4.1.27/arch/arm/include/asm/ |
D | fixmap.h | 12 FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_TYPE_NR * NR_CPUS) - 1,
|
D | topology.h | 16 extern struct cputopo_arm cpu_topology[NR_CPUS];
|
/linux-4.1.27/arch/arm64/include/asm/ |
D | smp_plat.h | 40 extern u64 __cpu_logical_map[NR_CPUS];
|
D | cpu_ops.h | 66 extern const struct cpu_operations *cpu_ops[NR_CPUS];
|
D | topology.h | 16 extern struct cpu_topology cpu_topology[NR_CPUS];
|
/linux-4.1.27/arch/mips/netlogic/common/ |
D | smp.c | 178 for (i = 0; i < NR_CPUS; i++) { in nlm_smp_setup() 237 for (i = 0; i < NR_CPUS; i += NLM_THREADS_PER_CORE) { in nlm_parse_cpumask()
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/ |
D | tracefile.h | 184 extern union cfs_trace_data_union (*cfs_trace_data[TCD_MAX_TYPES])[NR_CPUS]; 273 extern char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
|
/linux-4.1.27/arch/tile/kernel/ |
D | smp.c | 34 static unsigned long __iomem *ipi_mappings[NR_CPUS]; 75 HV_Recipient recip[NR_CPUS]; in send_IPI_many()
|
D | setup.c | 489 NR_CPUS * (PFN_UP(per_cpu_size) >> PAGE_SHIFT); in setup_memory() 851 char cpu_2_node[NR_CPUS] __write_once __attribute__((aligned(L2_CACHE_BYTES))); 885 int distance[MAX_NUMNODES][NR_CPUS]; in setup_numa_mapping() 986 static struct cpu cpu_devices[NR_CPUS]; 1357 for (i = smp_height * smp_width; i < NR_CPUS; ++i) in setup_cpu_maps() 1511 unsigned long __per_cpu_offset[NR_CPUS] __write_once; 1515 static unsigned long __initdata percpu_pfn[NR_CPUS] = { 0 };
|
D | tlb.c | 36 HV_Remote_ASID asids[NR_CPUS]; in flush_tlb_mm()
|
D | smpboot.c | 85 for (cpu = 0; cpu < NR_CPUS; ++cpu) { in smp_prepare_cpus()
|
/linux-4.1.27/arch/metag/include/asm/ |
D | topology.h | 22 extern cpumask_t cpu_core_map[NR_CPUS];
|
D | fixmap.h | 45 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
|
/linux-4.1.27/arch/ia64/include/asm/sn/ |
D | nodepda.h | 55 struct phys_cpuid phys_cpuid[NR_CPUS];
|
/linux-4.1.27/arch/x86/include/asm/ |
D | spinlock.h | 148 BUILD_BUG_ON(((__ticket_t)NR_CPUS) != NR_CPUS); in arch_spin_unlock()
|
D | irq_vectors.h | 163 # define CPU_VECTOR_LIMIT (64 * NR_CPUS)
|
D | percpu.h | 568 __typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \ 569 { [0 ... NR_CPUS-1] = _initvalue }; \ 574 __typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \ 575 { [0 ... NR_CPUS-1] = _initvalue }; \
|
D | pvclock.h | 101 #define PVCLOCK_VSYSCALL_NR_PAGES (((NR_CPUS-1)/(PAGE_SIZE/PVTI_SIZE))+1)
|
D | fixmap.h | 95 FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
|
/linux-4.1.27/arch/arc/include/asm/ |
D | mmu.h | 51 unsigned long asid[NR_CPUS]; /* 8 bit MMU PID + Generation cycle */
|
/linux-4.1.27/arch/mips/include/asm/mach-ip27/ |
D | topology.h | 23 extern struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
|
/linux-4.1.27/arch/mips/netlogic/xlr/ |
D | wakeup.c | 64 for (i = 0; i < NR_CPUS; i++) { in xlr_wakeup_secondary_cpus()
|
/linux-4.1.27/arch/arm/vfp/ |
D | vfpmodule.c | 59 union vfp_state *vfp_current_hw_state[NR_CPUS]; 86 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_force_reload() 117 vfp->hard.cpu = NR_CPUS; in vfp_thread_flush() 139 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_thread_copy()
|
/linux-4.1.27/arch/mips/loongson/common/ |
D | env.c | 140 if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0) in prom_init_env() 141 loongson_sysconf.nr_cpus = NR_CPUS; in prom_init_env()
|
/linux-4.1.27/arch/um/include/shared/ |
D | common-offsets.h | 14 DEFINE(UM_NR_CPUS, NR_CPUS);
|
/linux-4.1.27/arch/cris/arch-v32/kernel/ |
D | irq.c | 58 static unsigned long irq_regs[NR_CPUS] = 69 unsigned long cpu_irq_counters[NR_CPUS]; 283 for (cpu = 0; cpu < NR_CPUS; cpu++) in crisv32_mask_irq()
|
D | time.c | 43 unsigned long timer_regs[NR_CPUS] =
|
/linux-4.1.27/arch/mips/include/asm/mach-paravirt/ |
D | kernel-entry-init.h | 30 slti t1, t0, NR_CPUS
|
/linux-4.1.27/arch/powerpc/platforms/powermac/ |
D | smp.c | 329 if (ncpus > NR_CPUS) in smp_psurge_probe() 330 ncpus = NR_CPUS; in smp_psurge_probe() 347 extern volatile unsigned int cpu_callin_map[NR_CPUS]; in smp_psurge_kick_cpu() 1022 for (cpu = 1; cpu < 4 && cpu < NR_CPUS; ++cpu) in pmac_setup_smp()
|
/linux-4.1.27/kernel/ |
D | smp.c | 499 unsigned int setup_max_cpus = NR_CPUS; 552 int nr_cpu_ids __read_mostly = NR_CPUS; 558 nr_cpu_ids = find_last_bit(cpumask_bits(cpu_possible_mask),NR_CPUS) + 1; in setup_nr_cpu_ids()
|
D | Kconfig.hz | 15 environment leading to NR_CPUS * HZ number of timer interrupts
|
D | cpu.c | 743 const unsigned long cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)] = { 754 const DECLARE_BITMAP(cpu_all_bits, NR_CPUS) = CPU_BITS_ALL;
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
D | smp.c | 118 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_cell_kick_cpu()
|
D | spu_base.c | 112 int nr = (NR_CPUS > 1) ? NR_CPUS : NR_CPUS + 1; in mm_needs_global_tlbie()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | processor.c | 157 if (cpu_info.cpu_num >= NR_CPUS) { in processor_probe() 161 dev->hpa.start, cpu_info.cpu_num, NR_CPUS); in processor_probe()
|
/linux-4.1.27/arch/arm/mach-imx/ |
D | platsmp.c | 66 for (i = ncores; i < NR_CPUS; i++) in imx_smp_init_cpus()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | init_32.c | 84 extern struct task_struct *current_set[NR_CPUS];
|
D | mmu_context_nohash.c | 60 static unsigned long *stale_map[NR_CPUS]; 422 init_mm.context.active = NR_CPUS; in mmu_context_init()
|
/linux-4.1.27/arch/um/os-Linux/skas/ |
D | process.c | 245 #undef NR_CPUS 246 #define NR_CPUS 1 macro 247 int userspace_pid[NR_CPUS];
|
/linux-4.1.27/arch/mips/sibyte/sb1250/ |
D | smp.c | 135 for (i = 1, num = 0; i < NR_CPUS; i++) { in sb1250_smp_setup()
|
D | irq.c | 157 for (i=0; i<NR_CPUS; i++) { in ack_sb1250_irq()
|
/linux-4.1.27/arch/um/kernel/ |
D | process.c | 36 struct cpu_task cpu_tasks[NR_CPUS] = { [0 ... NR_CPUS - 1] = { -1, NULL } };
|
/linux-4.1.27/arch/arm64/mm/ |
D | context.c | 162 cpu_last_asid += NR_CPUS - 1; in __new_context()
|
/linux-4.1.27/kernel/trace/ |
D | trace_kdb.c | 119 if (*cp || cpu_file >= NR_CPUS || cpu_file < 0 || in kdb_ftdump()
|
/linux-4.1.27/arch/cris/kernel/ |
D | setup.c | 43 static struct cpu cpu_devices[NR_CPUS];
|
/linux-4.1.27/drivers/edac/ |
D | octeon_edac-pc.c | 27 extern unsigned long long cache_err_dcache[NR_CPUS];
|
/linux-4.1.27/arch/mips/sibyte/bcm1480/ |
D | smp.c | 147 for (i = 1, num = 0; i < NR_CPUS; i++) { in bcm1480_smp_setup()
|
/linux-4.1.27/arch/powerpc/platforms/85xx/ |
D | smp.c | 200 WARN_ON(nr < 0 || nr >= NR_CPUS); in smp_85xx_kick_cpu() 201 WARN_ON(hw_cpu < 0 || hw_cpu >= NR_CPUS); in smp_85xx_kick_cpu()
|
/linux-4.1.27/arch/arm/common/ |
D | bL_switcher.c | 140 static int bL_switcher_cpu_pairing[NR_CPUS]; 270 static struct bL_thread bL_threads[NR_CPUS]; 384 static unsigned int bL_switcher_cpu_original_cluster[NR_CPUS];
|
/linux-4.1.27/arch/tile/mm/ |
D | homecache.c | 168 BUG_ON(home < 0 || home >= NR_CPUS); in homecache_finv_page_va() 321 BUG_ON(home < 0 || home >= NR_CPUS || in pte_set_home()
|
/linux-4.1.27/arch/microblaze/kernel/cpu/ |
D | mb.c | 140 return i < NR_CPUS ? (void *) (i + 1) : NULL; in c_start()
|
/linux-4.1.27/arch/x86/kernel/apic/ |
D | hw_nmi.c | 32 static DECLARE_BITMAP(backtrace_mask, NR_CPUS) __read_mostly;
|
/linux-4.1.27/arch/sh/kernel/cpu/sh4a/ |
D | smp-shx3.c | 63 for (i = 1, num = 0; i < NR_CPUS; i++) { in shx3_smp_setup()
|
/linux-4.1.27/arch/openrisc/kernel/ |
D | process.c | 52 struct thread_info *current_thread_info_set[NR_CPUS] = { &init_thread_info, };
|
/linux-4.1.27/arch/hexagon/kernel/ |
D | smp.c | 264 for (i = 0; i < NR_CPUS; i++) in smp_start_cpus()
|
/linux-4.1.27/arch/powerpc/platforms/powernv/ |
D | smp.c | 65 BUG_ON(nr < 0 || nr >= NR_CPUS); in pnv_smp_kick_cpu()
|
/linux-4.1.27/arch/mips/sgi-ip27/ |
D | ip27-init.c | 50 struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
|
D | ip27-smp.c | 78 (tot_cpus_found != NR_CPUS)) { in do_cpumask()
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
D | smp.c | 162 BUG_ON(nr < 0 || nr >= NR_CPUS); in smp_pSeries_kick_cpu()
|
/linux-4.1.27/arch/sh/kernel/cpu/ |
D | proc.c | 134 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
/linux-4.1.27/drivers/net/hyperv/ |
D | hyperv_net.h | 659 struct vmbus_channel *chn_table[NR_CPUS]; 663 atomic_t queue_sends[NR_CPUS]; 675 struct multi_send_data msd[NR_CPUS];
|
/linux-4.1.27/arch/s390/kernel/ |
D | machine_kexec.c | 219 VMCOREINFO_LENGTH(lowcore_ptr, NR_CPUS); in arch_crash_save_vmcoreinfo()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | nmi.c | 37 static atomic_t nmi_touched[NR_CPUS];
|
/linux-4.1.27/arch/arc/kernel/ |
D | smp.c | 52 for (i = 0; i < NR_CPUS; i++) in smp_init_cpus()
|
D | setup.c | 39 struct task_struct *_current_task[NR_CPUS]; /* For stack switching */ 41 struct cpuinfo_arc cpuinfo_arc700[NR_CPUS];
|
/linux-4.1.27/drivers/irqchip/ |
D | irq-mips-gic.c | 34 static struct gic_pcpu_mask pcpu_masks[NR_CPUS]; 405 for (i = 0; i < NR_CPUS; i++) in gic_set_affinity() 576 for (i = 0; i < NR_CPUS; i++) in gic_ipi_init_one()
|
/linux-4.1.27/drivers/cpufreq/ |
D | sparc-us3-cpufreq.c | 189 us3_freq_table = kzalloc((NR_CPUS * sizeof(*us3_freq_table)), in us3_freq_init()
|
D | p4-clockmod.c | 52 static int has_N44_O17_errata[NR_CPUS];
|
/linux-4.1.27/arch/m68k/kernel/ |
D | setup_no.c | 297 return *pos < NR_CPUS ? ((void *) 0x12345678) : NULL; in c_start()
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/ |
D | linux-tracefile.c | 50 char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
|
/linux-4.1.27/arch/xtensa/kernel/ |
D | process.c | 50 struct task_struct *current_set[NR_CPUS] = {&init_task, };
|
/linux-4.1.27/arch/powerpc/sysdev/xics/ |
D | icp-native.c | 50 static struct icp_ipl __iomem *icp_native_regs[NR_CPUS];
|
/linux-4.1.27/drivers/lguest/ |
D | lg.h | 87 struct lg_cpu cpus[NR_CPUS];
|