/linux-4.1.27/arch/mips/include/asm/ |
D | cpu-features.h | 21 #define cpu_has_tlb (cpu_data[0].options & MIPS_CPU_TLB) 24 #define cpu_has_tlbinv (cpu_data[0].options & MIPS_CPU_TLBINV) 27 #define cpu_has_segments (cpu_data[0].options & MIPS_CPU_SEGMENTS) 30 #define cpu_has_eva (cpu_data[0].options & MIPS_CPU_EVA) 33 #define cpu_has_htw (cpu_data[0].options & MIPS_CPU_HTW) 36 #define cpu_has_rixiex (cpu_data[0].options & MIPS_CPU_RIXIEX) 39 #define cpu_has_maar (cpu_data[0].options & MIPS_CPU_MAAR) 42 #define cpu_has_rw_llb (cpu_data[0].options & MIPS_CPU_RW_LLB) 55 #define cpu_has_4kex (cpu_data[0].options & MIPS_CPU_4KEX) 58 #define cpu_has_3k_cache (cpu_data[0].options & MIPS_CPU_3K_CACHE) [all …]
|
D | cpu-info.h | 96 extern struct cpuinfo_mips cpu_data[]; 97 #define current_cpu_data cpu_data[smp_processor_id()] 98 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()] 99 #define boot_cpu_data cpu_data[0]
|
D | topology.h | 15 #define topology_physical_package_id(cpu) (cpu_data[cpu].package) 16 #define topology_core_id(cpu) (cpu_data[cpu].core)
|
D | bugs.h | 37 cpu_data[cpu].udelay_val = loops_per_jiffy; in check_bugs()
|
D | timex.h | 64 asm volatile("" : "=m" (cpu_data[0].options)); in can_use_mips_counter()
|
D | cpu-type.h | 219 const int cpu_type = cpu_data[0].cputype; in boot_cpu_type()
|
D | mmu_context.h | 87 #define asid_cache(cpu) (cpu_data[cpu].asid_cache)
|
D | pgtable.h | 420 prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine; in pgprot_writecombine()
|
/linux-4.1.27/arch/sh/mm/ |
D | tlb-sh5.c | 23 cpu_data->dtlb.entries = 64; in sh64_tlb_init() 24 cpu_data->dtlb.step = 0x10; in sh64_tlb_init() 26 cpu_data->dtlb.first = DTLB_FIXED | cpu_data->dtlb.step; in sh64_tlb_init() 27 cpu_data->dtlb.next = cpu_data->dtlb.first; in sh64_tlb_init() 29 cpu_data->dtlb.last = DTLB_FIXED | in sh64_tlb_init() 30 ((cpu_data->dtlb.entries - 1) * in sh64_tlb_init() 31 cpu_data->dtlb.step); in sh64_tlb_init() 34 cpu_data->itlb.entries = 64; in sh64_tlb_init() 35 cpu_data->itlb.step = 0x10; in sh64_tlb_init() 37 cpu_data->itlb.first = ITLB_FIXED | cpu_data->itlb.step; in sh64_tlb_init() [all …]
|
D | cache-sh5.c | 249 cpu_data->dcache.entry_mask) >> in sh64_dcache_purge_sets() 250 cpu_data->dcache.entry_shift; in sh64_dcache_purge_sets() 254 set_offset &= (cpu_data->dcache.sets - 1); in sh64_dcache_purge_sets() 256 (set_offset << cpu_data->dcache.entry_shift); in sh64_dcache_purge_sets() 265 eaddr1 = eaddr0 + cpu_data->dcache.way_size * in sh64_dcache_purge_sets() 266 cpu_data->dcache.ways; in sh64_dcache_purge_sets() 269 eaddr += cpu_data->dcache.way_size) { in sh64_dcache_purge_sets() 274 eaddr1 = eaddr0 + cpu_data->dcache.way_size * in sh64_dcache_purge_sets() 275 cpu_data->dcache.ways; in sh64_dcache_purge_sets() 278 eaddr += cpu_data->dcache.way_size) { in sh64_dcache_purge_sets() [all …]
|
D | cache-sh4.c | 74 cpu_data->icache.entry_mask); in sh4_flush_icache_range() 78 for (i = 0; i < cpu_data->icache.ways; i++) { in sh4_flush_icache_range() 81 icacheaddr += cpu_data->icache.way_incr; in sh4_flush_icache_range()
|
D | tlb-sh4.c | 45 if (cpu_data->flags & CPU_HAS_PTEA) { in __update_tlb()
|
/linux-4.1.27/arch/mips/kernel/ |
D | proc.c | 39 unsigned int version = cpu_data[n].processor_id; in show_cpuinfo() 40 unsigned int fp_vers = cpu_data[n].fpu_id; in show_cpuinfo() 61 cpu_data[n].options & MIPS_CPU_FPU ? " FPU V%d.%d" : ""); in show_cpuinfo() 66 cpu_data[n].udelay_val / (500000/HZ), in show_cpuinfo() 67 (cpu_data[n].udelay_val / (5000/HZ)) % 100); in show_cpuinfo() 71 seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize); in show_cpuinfo() 78 cpu_data[n].watch_reg_count); in show_cpuinfo() 79 for (i = 0; i < cpu_data[n].watch_reg_count; i++) in show_cpuinfo() 81 cpu_data[n].watch_reg_masks[i]); in show_cpuinfo() 131 cpu_data[n].srsets); in show_cpuinfo() [all …]
|
D | smp-cps.c | 62 cpu_data[nvpes + v].core = c; in cps_smp_setup() 64 cpu_data[nvpes + v].vpe_id = v; in cps_smp_setup() 128 if (cpu_data[c].core) in cps_prepare_cpus() 230 unsigned core = cpu_data[cpu].core; in cps_boot_secondary() 231 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); in cps_boot_secondary() 241 atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask); in cps_boot_secondary() 254 if (cpu_data[remote].core != core) in cps_boot_secondary() 338 core = cpu_data[cpu].core; in play_dead() 342 if (cpu_data[cpu_death_sibling].core != core) in play_dead() 373 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); in wait_for_sibling_halt() [all …]
|
D | smp.c | 89 if (cpu_data[cpu].package == cpu_data[i].package && in set_cpu_sibling_map() 90 cpu_data[cpu].core == cpu_data[i].core) { in set_cpu_sibling_map() 106 if (cpu_data[cpu].package == cpu_data[i].package) { in set_cpu_core_map() 127 if (cpu_data[i].package == cpu_data[k].package && in calculate_cpu_foreign_map() 128 cpu_data[i].core == cpu_data[k].core) in calculate_cpu_foreign_map() 170 cpu_data[cpu].udelay_val = loops_per_jiffy; in start_secondary()
|
D | pm-cps.c | 460 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache, in cps_gen_entry_code() 464 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache, in cps_gen_entry_code() 476 uasm_i_addiu(&p, t0, zero, 1 << cpu_data[cpu].core); in cps_gen_entry_code() 489 err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu], in cps_gen_entry_code() 626 unsigned core = cpu_data[cpu].core; in cps_gen_core_entries() 627 unsigned dlinesz = cpu_data[cpu].dcache.linesz; in cps_gen_core_entries()
|
D | smp-gic.c | 25 unsigned int core = cpu_data[cpu].core; in gic_send_ipi_single()
|
D | cpu-probe.c | 166 cpu_data[0].ases &= ~(MIPS_ASE_DSP | MIPS_ASE_DSP2P); in dsp_disable() 179 cpu_data[0].options &= ~MIPS_CPU_HTW; in htw_disable() 205 set_ftlb_enable(&cpu_data[0], 0); in ftlb_disable() 234 cpu_data[0].tlbsize -= cpu_data[0].tlbsizeftlbways * in ftlb_disable() 235 cpu_data[0].tlbsizeftlbsets; in ftlb_disable() 236 cpu_data[0].tlbsizeftlbsets = 0; in ftlb_disable() 237 cpu_data[0].tlbsizeftlbways = 0; in ftlb_disable()
|
D | smp-mt.c | 303 seq_printf(m, "VPE\t\t\t: %d\n", cpu_data[n].vpe_id); in proc_cpuinfo_chain_call()
|
D | setup.c | 41 struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly; variable 43 EXPORT_SYMBOL(cpu_data);
|
D | traps.c | 2136 if (!cpu_data[cpu].asid_cache) in per_cpu_trap_init() 2137 cpu_data[cpu].asid_cache = ASID_FIRST_VERSION; in per_cpu_trap_init()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | prom_64.c | 412 int proc_id = cpu_data(cpu).proc_id; in arch_find_n_match_cpu_physical_id() 526 cpu_data(cpuid).clock_tick = in fill_in_one_cpu() 530 cpu_data(cpuid).dcache_size = in fill_in_one_cpu() 533 cpu_data(cpuid).dcache_line_size = in fill_in_one_cpu() 536 cpu_data(cpuid).icache_size = in fill_in_one_cpu() 539 cpu_data(cpuid).icache_line_size = in fill_in_one_cpu() 542 cpu_data(cpuid).ecache_size = in fill_in_one_cpu() 544 cpu_data(cpuid).ecache_line_size = in fill_in_one_cpu() 546 if (!cpu_data(cpuid).ecache_size || in fill_in_one_cpu() 547 !cpu_data(cpuid).ecache_line_size) { in fill_in_one_cpu() [all …]
|
D | smp_32.c | 62 cpu_data(id).udelay_val = loops_per_jiffy; in smp_store_cpu_info() 65 cpu_data(id).clock_tick = prom_getintdefault(cpu_node, in smp_store_cpu_info() 67 cpu_data(id).prom_node = cpu_node; in smp_store_cpu_info() 74 cpu_data(id).mid = mid; in smp_store_cpu_info() 84 bogosum += cpu_data(cpu).udelay_val; in smp_cpus_done() 386 cpu_data(i).udelay_val/(500000/HZ), in smp_bogo() 387 (cpu_data(i).udelay_val/(5000/HZ))%100); in smp_bogo()
|
D | cpu.c | 347 , cpu_data(0).udelay_val/(500000/HZ), in show_cpuinfo() 348 (cpu_data(0).udelay_val/(5000/HZ)) % 100, in show_cpuinfo() 349 cpu_data(0).clock_tick in show_cpuinfo() 397 , cpu_data(0).clock_tick in show_cpuinfo()
|
D | irq_32.c | 203 seq_printf(p, "%10u ", cpu_data(j).irq_resched_count); in arch_show_interrupts() 207 seq_printf(p, "%10u ", cpu_data(j).irq_call_count); in arch_show_interrupts() 212 seq_printf(p, "%10u ", cpu_data(j).counter); in arch_show_interrupts()
|
D | smp_64.c | 88 i, cpu_data(i).clock_tick); in smp_bogo() 1238 if (cpu_data(i).core_id == 0) { in smp_fill_in_sib_core_maps() 1244 if (cpu_data(i).core_id == in smp_fill_in_sib_core_maps() 1245 cpu_data(j).core_id) in smp_fill_in_sib_core_maps() 1254 if (cpu_data(i).sock_id == cpu_data(j).sock_id) in smp_fill_in_sib_core_maps() 1263 if (cpu_data(i).proc_id == -1) { in smp_fill_in_sib_core_maps() 1269 if (cpu_data(i).proc_id == in smp_fill_in_sib_core_maps() 1270 cpu_data(j).proc_id) in smp_fill_in_sib_core_maps() 1347 c = &cpu_data(cpu); in __cpu_disable()
|
D | devices.c | 129 cpu_data(0).clock_tick = prom_getintdefault(cpu_node, in device_scan()
|
D | time_64.c | 641 return cpu_data(cpu).clock_tick; in sparc64_get_clock_tick() 656 ft->clock_tick_ref = cpu_data(cpu).clock_tick; in sparc64_cpufreq_notifier() 660 cpu_data(cpu).clock_tick = in sparc64_cpufreq_notifier()
|
D | sysfs.c | 192 cpuinfo_sparc *c = &cpu_data(dev->id); \ 200 cpuinfo_sparc *c = &cpu_data(dev->id); \
|
D | cpumap.c | 107 id = cpu_data(cpu).core_id; in cpuinfo_id() 110 id = cpu_data(cpu).proc_id; in cpuinfo_id()
|
D | mdesc.c | 645 cpu_data(*id).core_id = core_id; in __mark_core_id() 654 cpu_data(*id).sock_id = sock_id; in __mark_sock_id() 738 cpu_data(*id).sock_id = idx; in set_sock_ids_by_socket() 774 cpu_data(*id).proc_id = proc_id; in mark_proc_ids() 944 c = &cpu_data(cpuid); in fill_in_one_cpu()
|
D | sun4m_smp.c | 130 prev = &cpu_data(i).next; in smp4m_smp_done()
|
D | nmi.c | 128 return cpu_data(cpu).__nmi_count; in get_nmi_count()
|
D | sun4d_smp.c | 172 prev = &cpu_data(i).next; in smp4d_smp_done()
|
D | leon_smp.c | 241 prev = &cpu_data(i).next; in leon_smp_done()
|
D | irq_64.c | 308 seq_printf(p, "%10u ", cpu_data(j).__nmi_count); in arch_show_interrupts()
|
D | traps_64.c | 838 val = cpu_data(i).ecache_size; in cheetah_ecache_flush_init() 845 val = cpu_data(i).ecache_line_size; in cheetah_ecache_flush_init()
|
/linux-4.1.27/arch/sh/include/asm/ |
D | tlb_64.h | 29 for (tlb = cpu_data->dtlb.first; \ 30 tlb <= cpu_data->dtlb.last; \ 31 tlb += cpu_data->dtlb.step) 39 for (tlb = cpu_data->itlb.first; \ 40 tlb <= cpu_data->itlb.last; \ 41 tlb += cpu_data->itlb.step)
|
D | processor.h | 93 extern struct sh_cpuinfo cpu_data[]; 94 #define boot_cpu_data cpu_data[0] 95 #define current_cpu_data cpu_data[smp_processor_id()] 96 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
|
D | mmu_context.h | 35 #define asid_cache(cpu) (cpu_data[cpu].asid_cache)
|
/linux-4.1.27/kernel/trace/ |
D | trace_functions_graph.c | 56 struct fgraph_cpu_data __percpu *cpu_data; member 544 last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid); in verif_pid() 777 struct fgraph_cpu_data *cpu_data; in print_graph_entry_leaf() local 780 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_leaf() 787 cpu_data->depth = call->depth - 1; in print_graph_entry_leaf() 791 cpu_data->enter_funcs[call->depth] = 0; in print_graph_entry_leaf() 816 struct fgraph_cpu_data *cpu_data; in print_graph_entry_nested() local 819 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_nested() 820 cpu_data->depth = call->depth; in print_graph_entry_nested() 824 cpu_data->enter_funcs[call->depth] = call->func; in print_graph_entry_nested() [all …]
|
/linux-4.1.27/arch/ia64/kernel/ |
D | smpboot.c | 422 last_cpuinfo = cpu_data(cpuid - 1); in smp_callin() 594 if (cpu_data(cpu)->threads_per_core == 1 && in remove_siblinginfo() 595 cpu_data(cpu)->cores_per_socket == 1) { in remove_siblinginfo() 709 bogosum += cpu_data(cpu)->loops_per_jiffy; in smp_cpus_done() 721 if ((cpu_data(cpu)->socket_id == cpu_data(i)->socket_id)) { in set_cpu_sibling_map() 724 if (cpu_data(cpu)->core_id == cpu_data(i)->core_id) { in set_cpu_sibling_map() 757 if (cpu_data(cpu)->threads_per_core == 1 && in __cpu_up() 758 cpu_data(cpu)->cores_per_socket == 1) { in __cpu_up() 853 if ((cpu_data(j)->socket_id == cpu_data(i)->socket_id)) { in is_multithreading_enabled() 854 if (cpu_data(j)->core_id == cpu_data(i)->core_id) in is_multithreading_enabled()
|
D | topology.c | 35 if (cpu_data(num)->socket_id == -1) in arch_fix_phys_package_id() 36 cpu_data(num)->socket_id = slot; in arch_fix_phys_package_id() 149 if (cpu_data(cpu)->threads_per_core <= 1 && in cache_shared_cpu_map_setup() 150 cpu_data(cpu)->cores_per_socket <= 1) { in cache_shared_cpu_map_setup() 164 if (cpu_data(cpu)->socket_id == cpu_data(j)->socket_id in cache_shared_cpu_map_setup() 165 && cpu_data(j)->core_id == csi.log1_cid in cache_shared_cpu_map_setup() 166 && cpu_data(j)->thread_id == csi.log1_tid) in cache_shared_cpu_map_setup()
|
D | setup.c | 727 return *pos < nr_cpu_ids ? cpu_data(*pos) : NULL; in c_start() 928 void *cpu_data; in cpu_init() local 930 cpu_data = per_cpu_init(); in cpu_init() 948 ia64_tpa(cpu_data) - (long) __per_cpu_start); in cpu_init() 960 cpu_info = cpu_data + ((char *) &__ia64_per_cpu_var(ia64_cpu_info) - __per_cpu_start); in cpu_init() 1006 ia64_mmu_init(ia64_imva(cpu_data)); in cpu_init() 1007 ia64_mca_cpu_init(ia64_imva(cpu_data)); in cpu_init()
|
D | mca.c | 1842 ia64_mca_cpu_init(void *cpu_data) in ia64_mca_cpu_init() argument 1878 pte_val(mk_pte_phys(__pa(cpu_data), PAGE_KERNEL))); in ia64_mca_cpu_init()
|
/linux-4.1.27/arch/alpha/include/asm/ |
D | mmu_context.h | 89 #define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn) 141 cpu_data[cpu].asn_lock = 1; in ev5_switch_mm() 152 cpu_data[cpu].need_new_asn = 1; in ev5_switch_mm() 187 cpu_data[cpu].asn_lock = 0; \ 189 if (cpu_data[cpu].need_new_asn) { \ 191 cpu_data[cpu].need_new_asn = 0; \
|
D | smp.h | 40 extern struct cpuinfo_alpha cpu_data[NR_CPUS];
|
/linux-4.1.27/arch/ia64/mm/ |
D | contig.c | 85 static void *cpu_data; variable 109 memcpy(cpu_data, src, __per_cpu_end - __per_cpu_start); in per_cpu_init() 110 __per_cpu_offset[cpu] = (char *)cpu_data - __per_cpu_start; in per_cpu_init() 123 ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) - in per_cpu_init() 126 cpu_data += PERCPU_PAGE_SIZE; in per_cpu_init() 135 cpu_data = __alloc_bootmem(PERCPU_PAGE_SIZE * num_possible_cpus(), in alloc_per_cpu_data()
|
D | discontig.c | 140 static void *per_cpu_node_setup(void *cpu_data, int node) in per_cpu_node_setup() argument 151 memcpy(__va(cpu_data), src, __per_cpu_end - __per_cpu_start); in per_cpu_node_setup() 152 __per_cpu_offset[cpu] = (char *)__va(cpu_data) - in per_cpu_node_setup() 166 (unsigned long)cpu_data - in per_cpu_node_setup() 169 cpu_data += PERCPU_PAGE_SIZE; in per_cpu_node_setup() 172 return cpu_data; in per_cpu_node_setup() 269 void *cpu_data; in fill_pernode() local 277 cpu_data = (void *)pernode; in fill_pernode() 290 cpu_data = per_cpu_node_setup(cpu_data, node); in fill_pernode()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | smp.c | 55 struct cpuinfo_alpha cpu_data[NR_CPUS]; variable 56 EXPORT_SYMBOL(cpu_data); 83 cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy; in smp_store_cpu_info() 84 cpu_data[cpuid].last_asn = ASN_FIRST_VERSION; in smp_store_cpu_info() 85 cpu_data[cpuid].need_new_asn = 0; in smp_store_cpu_info() 86 cpu_data[cpuid].asn_lock = 0; in smp_store_cpu_info() 95 cpu_data[cpuid].prof_counter = 1; in smp_setup_percpu_timer() 96 cpu_data[cpuid].prof_multiplier = 1; in smp_setup_percpu_timer() 493 bogosum += cpu_data[cpu].loops_per_jiffy; in smp_cpus_done() 566 cpu_data[this_cpu].ipi_count++; in handle_ipi() [all …]
|
D | proto.h | 205 #define mcheck_expected(cpu) (cpu_data[cpu].mcheck_expected) 206 #define mcheck_taken(cpu) (cpu_data[cpu].mcheck_taken) 207 #define mcheck_extra(cpu) (cpu_data[cpu].mcheck_extra)
|
D | irq.c | 75 seq_printf(p, "%10lu ", cpu_data[j].ipi_count); in arch_show_interrupts()
|
/linux-4.1.27/arch/mips/loongson/loongson-3/ |
D | smp.c | 299 cpu_data[cpu].core = in loongson3_init_secondary() 301 cpu_data[cpu].package = in loongson3_init_secondary() 362 cpu_data[0].core = cpu_logical_map(0) % loongson_sysconf.cores_per_package; in loongson3_smp_setup() 363 cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package; in loongson3_smp_setup() 462 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3a_play_dead() 524 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3b_play_dead() 585 uint64_t core_id = cpu_data[cpu].core; in loongson3_disable_clock() 586 uint64_t package_id = cpu_data[cpu].package; in loongson3_disable_clock() 598 uint64_t core_id = cpu_data[cpu].core; in loongson3_enable_clock() 599 uint64_t package_id = cpu_data[cpu].package; in loongson3_enable_clock()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | processor.c | 48 DEFINE_PER_CPU(struct cpuinfo_parisc, cpu_data); 81 p = &per_cpu(cpu_data, cpunum); in init_percpu_prof() 172 p = &per_cpu(cpu_data, cpuid); in processor_probe() 316 per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision; in init_per_cpu() 317 per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model; in init_per_cpu() 356 const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); in show_cpuinfo()
|
D | smp.c | 124 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt() 191 struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu); in ipi_send() 319 const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid); in smp_boot_one_cpu() 377 int bootstrap_processor = per_cpu(cpu_data, 0).cpuid; in smp_prepare_boot_cpu()
|
D | time.c | 65 struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); in timer_interrupt() 231 per_cpu(cpu_data, cpu).it_value = next_tick; in start_cpu_itimer()
|
D | setup.c | 392 per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision; in start_parisc() 393 per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model; in start_parisc()
|
D | irq.c | 345 return per_cpu(cpu_data, cpu).txn_addr; in txn_affinity_addr() 357 (!per_cpu(cpu_data, next_cpu).txn_addr || in txn_alloc_addr() 538 per_cpu(cpu_data, cpu).hpa); in do_cpu_irq_mask()
|
D | perf.c | 540 cpu_device = per_cpu(cpu_data, 0).dev; in perf_init() 542 per_cpu(cpu_data, 0).dev->name); in perf_init()
|
/linux-4.1.27/arch/m32r/include/asm/ |
D | processor.h | 47 extern struct cpuinfo_m32r cpu_data[]; 48 #define current_cpu_data cpu_data[smp_processor_id()] 50 #define cpu_data (&boot_cpu_data) macro
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | processor.h | 61 extern struct mn10300_cpuinfo cpu_data[]; 62 #define current_cpu_data cpu_data[smp_processor_id()] 64 #define cpu_data &boot_cpu_data macro
|
/linux-4.1.27/arch/x86/kernel/ |
D | smpboot.c | 192 cpu_data(cpuid).loops_per_jiffy = loops_per_jiffy; in smp_callin() 275 struct cpuinfo_x86 *c = &cpu_data(id); in smp_store_boot_cpu_info() 287 struct cpuinfo_x86 *c = &cpu_data(id); in smp_store_cpu_info() 394 struct cpuinfo_x86 *c = &cpu_data(cpu); in set_cpu_sibling_map() 409 o = &cpu_data(i); in set_cpu_sibling_map() 424 o = &cpu_data(i); in set_cpu_sibling_map() 444 cpu_data(i).booted_cores++; in set_cpu_sibling_map() 446 c->booted_cores = cpu_data(i).booted_cores; in set_cpu_sibling_map() 469 bogosum += cpu_data(cpu).loops_per_jiffy; in impress_friends() 1114 c = &cpu_data(i); in smp_cpu_index_default() [all …]
|
D | tsc.c | 817 cpu_data(0).loops_per_jiffy = in recalibrate_cpu_khz() 818 cpufreq_scale(cpu_data(0).loops_per_jiffy, in recalibrate_cpu_khz() 902 if (cpu_has(&cpu_data(freq->cpu), X86_FEATURE_CONSTANT_TSC)) in time_cpufreq_notifier() 908 lpj = &cpu_data(freq->cpu).loops_per_jiffy; in time_cpufreq_notifier() 1235 if (!tsc_disabled && !cpu_has(&cpu_data(cpu), X86_FEATURE_CONSTANT_TSC)) in calibrate_delay_is_known() 1239 if (cpu_data(i).phys_proc_id == cpu_data(cpu).phys_proc_id) in calibrate_delay_is_known() 1240 return cpu_data(i).loops_per_jiffy; in calibrate_delay_is_known()
|
D | amd_nb.c | 182 cuid = cpu_data(cpu).compute_unit_id; in amd_get_subcaches() 209 cuid = cpu_data(cpu).compute_unit_id; in amd_set_subcaches()
|
D | topology.c | 109 struct cpuinfo_x86 *c = &cpu_data(num); in arch_register_cpu()
|
D | cpuid.c | 123 c = &cpu_data(cpu); in cpuid_open()
|
D | msr.c | 185 c = &cpu_data(cpu); in msr_open()
|
D | alternative.c | 535 clear_cpu_cap(&cpu_data(0), X86_FEATURE_UP); in alternatives_enable_smp()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | topology.h | 53 #define topology_physical_package_id(cpu) (cpu_data(cpu)->socket_id) 54 #define topology_core_id(cpu) (cpu_data(cpu)->core_id)
|
D | processor.h | 243 #define cpu_data(cpu) (&per_cpu(ia64_cpu_info, cpu)) macro
|
/linux-4.1.27/arch/mips/sgi-ip27/ |
D | ip27-irq-pci.c | 67 struct slice_data *si = cpu_data[cpu].data; in alloc_level() 85 struct slice_data *si = cpu_data[cpu].data; in find_level() 101 struct slice_data *si = cpu_data[cpu].data; in intr_connect_level() 119 struct slice_data *si = cpu_data[cpu].data; in intr_disconnect_level()
|
D | ip27-irq.c | 118 struct slice_data *si = cpu_data[cpu].data; in ip27_do_irq_mask0() 133 struct slice_data *si = cpu_data[cpu].data; in ip27_do_irq_mask1() 187 struct slice_data *si = cpu_data[cpu].data; in install_ipi()
|
D | ip27-init.c | 136 cpu_data[cpu].data = si; in per_cpu_init()
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | topology_64.h | 41 #define topology_physical_package_id(cpu) (cpu_data(cpu).proc_id) 42 #define topology_core_id(cpu) (cpu_data(cpu).core_id)
|
D | bugs.h | 15 cpu_data(0).udelay_val = loops_per_jiffy; in check_bugs()
|
D | cpudata_32.h | 28 #define cpu_data(__cpu) per_cpu(__cpu_data, (__cpu)) macro
|
D | cpudata_64.h | 33 #define cpu_data(__cpu) per_cpu(__cpu_data, (__cpu)) macro
|
D | delay_32.h | 27 #define __udelay_val cpu_data(smp_processor_id()).udelay_val
|
/linux-4.1.27/arch/metag/kernel/ |
D | topology.c | 19 DEFINE_PER_CPU(struct cpuinfo_metag, cpu_data); 54 struct cpuinfo_metag *cpuinfo = &per_cpu(cpu_data, i); in topology_init()
|
D | smp.c | 328 struct cpuinfo_metag *cpu_info = &per_cpu(cpu_data, cpuid); in smp_store_cpu_info() 406 bogosum += per_cpu(cpu_data, cpu).loops_per_jiffy; in smp_cpus_done()
|
D | setup.c | 481 lpj = per_cpu(cpu_data, i).loops_per_jiffy; in show_cpuinfo()
|
/linux-4.1.27/arch/x86/kernel/cpu/microcode/ |
D | intel.c | 92 struct cpuinfo_x86 *c = &cpu_data(cpu_num); in collect_cpu_info() 136 struct cpuinfo_x86 *c = &cpu_data(cpu_num); in apply_microcode_intel() 279 struct cpuinfo_x86 *c = &cpu_data(cpu); in request_microcode_fw() 328 struct cpuinfo_x86 *c = &cpu_data(0); in init_intel_microcode()
|
D | amd.c | 127 struct cpuinfo_x86 *c = &cpu_data(cpu); in collect_cpu_info_amd() 196 struct cpuinfo_x86 *c = &cpu_data(cpu); in apply_microcode_amd() 392 if (cpu_data(cpu).cpu_index == boot_cpu_data.cpu_index) { in load_microcode_amd() 424 struct cpuinfo_x86 *c = &cpu_data(cpu); in request_microcode_amd() 478 struct cpuinfo_x86 *c = &cpu_data(0); in init_amd_microcode()
|
D | core.c | 551 struct cpuinfo_x86 *c = &cpu_data(0); in microcode_init() 626 struct cpuinfo_x86 *c = &cpu_data(0); in microcode_exit()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | topology.h | 122 #define topology_physical_package_id(cpu) (cpu_data(cpu).phys_proc_id) 123 #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id)
|
D | acpi.h | 101 struct cpuinfo_x86 *c = &cpu_data(0); in arch_has_acpi_pdc() 108 struct cpuinfo_x86 *c = &cpu_data(0); in arch_acpi_set_pdc_bits()
|
D | processor.h | 158 #define cpu_data(cpu) per_cpu(cpu_info, cpu) macro 161 #define cpu_data(cpu) boot_cpu_data macro
|
/linux-4.1.27/drivers/cpufreq/ |
D | qoriq-cpufreq.c | 33 struct cpu_data { struct 204 struct cpu_data *data; in qoriq_cpufreq_cpu_init() 294 struct cpu_data *data = policy->driver_data; in qoriq_cpufreq_cpu_exit() 308 struct cpu_data *data = policy->driver_data; in qoriq_cpufreq_target()
|
D | amd_freq_sensitivity.c | 38 static DEFINE_PER_CPU(struct cpu_data_t, cpu_data); 47 struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu); in amd_powersave_bias_target()
|
D | sc520_freq.c | 77 struct cpuinfo_x86 *c = &cpu_data(0); in sc520_freq_cpu_init()
|
D | elanfreq.c | 149 struct cpuinfo_x86 *c = &cpu_data(0); in elanfreq_cpu_init()
|
D | speedstep-centrino.c | 234 struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); in centrino_cpu_init_table() 345 struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); in centrino_cpu_init()
|
D | p4-clockmod.c | 170 struct cpuinfo_x86 *c = &cpu_data(policy->cpu); in cpufreq_p4_cpu_init()
|
D | acpi-cpufreq.c | 188 struct cpuinfo_x86 *cpu = &cpu_data(cpuid); in check_est_cpu() 195 struct cpuinfo_x86 *cpu = &cpu_data(cpuid); in check_amd_hwpstate_cpu() 650 struct cpuinfo_x86 *c = &cpu_data(policy->cpu); in acpi_cpufreq_cpu_init()
|
D | longrun.c | 171 struct cpuinfo_x86 *c = &cpu_data(0); in longrun_determine_freqs()
|
D | speedstep-lib.c | 255 struct cpuinfo_x86 *c = &cpu_data(0); in speedstep_detect_processor()
|
D | longhaul.c | 773 struct cpuinfo_x86 *c = &cpu_data(0); in longhaul_cpu_init() 931 struct cpuinfo_x86 *c = &cpu_data(0); in longhaul_init()
|
D | e_powersaver.c | 184 struct cpuinfo_x86 *c = &cpu_data(0); in eps_cpu_init()
|
D | powernow-k7.c | 120 struct cpuinfo_x86 *c = &cpu_data(0); in check_powernow()
|
/linux-4.1.27/arch/m32r/kernel/ |
D | smpboot.c | 83 struct cpuinfo_m32r cpu_data[NR_CPUS] __cacheline_aligned; variable 388 bogosum += cpu_data[cpu_id].loops_per_jiffy; in smp_cpus_done() 548 struct cpuinfo_m32r *ci = cpu_data + cpu_id; in smp_store_cpu_info() 556 struct cpuinfo_m32r *ci = &cpu_data[cpu_id]; in show_cpu_info()
|
D | m32r_ksyms.c | 38 EXPORT_SYMBOL(cpu_data);
|
D | setup.c | 295 unsigned long cpu = c - cpu_data; in show_cpuinfo() 362 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
/linux-4.1.27/drivers/cpuidle/ |
D | cpuidle-cps.c | 40 if (!cpu_data[dev->cpu].core && (index > STATE_NC_WAIT)) in cps_nc_enter() 163 core = cpu_data[cpu].core; in cps_cpuidle_init()
|
/linux-4.1.27/arch/x86/kernel/acpi/ |
D | cstate.c | 32 struct cpuinfo_x86 *c = &cpu_data(cpu); in acpi_processor_power_init_bm_check() 121 struct cpuinfo_x86 *c = &cpu_data(cpu); in acpi_processor_ffh_cstate_probe()
|
/linux-4.1.27/arch/arm/kernel/ |
D | smp.c | 322 struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid); in smp_store_cpu_info() 399 bogosum += per_cpu(cpu_data, cpu).loops_per_jiffy; in smp_cpus_done() 692 per_cpu(cpu_data, cpu).loops_per_jiffy; in cpufreq_callback() 705 per_cpu(cpu_data, cpu).loops_per_jiffy = in cpufreq_callback()
|
D | setup.c | 155 DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data); 994 struct cpuinfo_arm *cpuinfo = &per_cpu(cpu_data, cpu); in topology_init() 1063 cpuid = is_smp() ? per_cpu(cpu_data, i).cpuid : read_cpuid_id(); in c_show() 1069 per_cpu(cpu_data, i).loops_per_jiffy / (500000UL/HZ), in c_show() 1070 (per_cpu(cpu_data, i).loops_per_jiffy / (5000UL/HZ)) % 100); in c_show()
|
/linux-4.1.27/arch/arm64/kernel/ |
D | cpuinfo.c | 36 DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data); 245 struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data); in cpuinfo_store_cpu() 252 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0); in cpuinfo_store_boot_cpu()
|
D | setup.c | 452 struct cpu *cpu = &per_cpu(cpu_data.cpu, i); in topology_init() 514 struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, i); in c_show()
|
/linux-4.1.27/arch/sh/kernel/cpu/ |
D | proc.c | 81 unsigned int cpu = c - cpu_data; in show_cpuinfo() 134 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
/linux-4.1.27/arch/metag/include/asm/ |
D | cpu.h | 13 DECLARE_PER_CPU(struct cpuinfo_metag, cpu_data);
|
/linux-4.1.27/arch/arm/include/asm/ |
D | cpu.h | 24 DECLARE_PER_CPU(struct cpuinfo_arm, cpu_data);
|
D | smp_plat.h | 37 struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpu); in smp_cpuid_part()
|
/linux-4.1.27/arch/blackfin/include/asm/ |
D | cpu.h | 22 DECLARE_PER_CPU(struct blackfin_cpudata, cpu_data);
|
/linux-4.1.27/arch/sh/lib64/ |
D | udelay.c | 38 __delay(xloops * (HZ * cpu_data[raw_smp_processor_id()].loops_per_jiffy)); in __const_udelay()
|
/linux-4.1.27/arch/sh/lib/ |
D | delay.c | 39 "r" (cpu_data[raw_smp_processor_id()].loops_per_jiffy * (HZ/4)) in __const_udelay()
|
/linux-4.1.27/arch/alpha/lib/ |
D | udelay.c | 35 #define LPJ cpu_data[smp_processor_id()].loops_per_jiffy
|
/linux-4.1.27/arch/c6x/kernel/ |
D | setup.c | 85 static DEFINE_PER_CPU(struct cpuinfo_c6x, cpu_data); 101 p = &per_cpu(cpu_data, smp_processor_id()); in get_cpuinfo() 447 struct cpuinfo_c6x *p = &per_cpu(cpu_data, n); in show_cpuinfo()
|
/linux-4.1.27/drivers/hwmon/ |
D | coretemp.c | 61 #define TO_PHYS_ID(cpu) (cpu_data(cpu).phys_proc_id) 62 #define TO_CORE_ID(cpu) (cpu_data(cpu).cpu_core_id) 421 struct cpuinfo_x86 *c = &cpu_data(cpu); in chk_ucode_version() 475 struct cpuinfo_x86 *c = &cpu_data(cpu); in create_core_data() 680 struct cpuinfo_x86 *c = &cpu_data(cpu); in get_core_online()
|
D | via-cputemp.c | 127 struct cpuinfo_x86 *c = &cpu_data(pdev->id); in via_cputemp_probe() 323 struct cpuinfo_x86 *c = &cpu_data(i); in via_cputemp_init()
|
D | hwmon-vid.c | 290 struct cpuinfo_x86 *c = &cpu_data(0); in vid_which_vrm()
|
/linux-4.1.27/arch/mips/include/asm/mach-cavium-octeon/ |
D | cpu-feature-overrides.h | 61 #define cpu_has_rixi (cpu_data[0].cputype != CPU_CAVIUM_OCTEON)
|
/linux-4.1.27/arch/sh/kernel/ |
D | setup.c | 50 struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = { variable 58 EXPORT_SYMBOL(cpu_data);
|
D | smp.c | 50 struct sh_cpuinfo *c = cpu_data + cpu; in smp_store_cpu_info() 257 bogosum += cpu_data[cpu].loops_per_jiffy; in smp_cpus_done()
|
/linux-4.1.27/arch/mn10300/kernel/ |
D | setup.c | 205 unsigned long cpu_id = c - cpu_data; in show_cpuinfo() 265 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
D | smp.c | 91 struct mn10300_cpuinfo cpu_data[NR_CPUS] __cacheline_aligned; variable 727 struct mn10300_cpuinfo *ci = &cpu_data[cpu]; in smp_store_cpu_info() 823 struct mn10300_cpuinfo *ci = &cpu_data[cpu]; in smp_show_cpu_info()
|
/linux-4.1.27/arch/arm64/include/asm/ |
D | cpu.h | 61 DECLARE_PER_CPU(struct cpuinfo_arm64, cpu_data);
|
/linux-4.1.27/arch/um/include/asm/ |
D | processor-generic.h | 101 #define cpu_data (&boot_cpu_data) macro
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | intel_cacheinfo.c | 817 apicid = cpu_data(cpu).apicid; in __cache_amd_cpumap_setup() 826 apicid = cpu_data(i).apicid; in __cache_amd_cpumap_setup() 833 apicid = cpu_data(sibling).apicid; in __cache_amd_cpumap_setup() 866 struct cpuinfo_x86 *c = &cpu_data(cpu); in __cache_cpumap_setup() 883 if (cpu_data(i).apicid >> index_msb == c->apicid >> index_msb) { in __cache_cpumap_setup()
|
D | proc.c | 143 return &cpu_data(*pos); in c_start()
|
D | perf_event_intel_cqm.c | 1255 struct cpuinfo_x86 *c = &cpu_data(cpu); in intel_cqm_cpu_prepare() 1334 struct cpuinfo_x86 *c = &cpu_data(cpu); in intel_cqm_init()
|
D | perf_event_amd_uncore.c | 372 unsigned int apicid = cpu_data(cpu).apicid; in amd_uncore_cpu_starting()
|
D | perf_event_intel.c | 2805 switch (cpu_data(cpu).x86_model) { in intel_snb_pebs_broken() 2811 switch (cpu_data(cpu).x86_mask) { in intel_snb_pebs_broken() 2817 return (cpu_data(cpu).microcode < rev); in intel_snb_pebs_broken()
|
/linux-4.1.27/arch/mips/oprofile/ |
D | op_model_mipsxx.c | 55 M_PERFCTL_VPEID(cpu_data[smp_processor_id()].vpe_id)) 57 0 : cpu_data[smp_processor_id()].vpe_id)
|
/linux-4.1.27/drivers/char/hw_random/ |
D | via-rng.c | 135 struct cpuinfo_x86 *c = &cpu_data(0); in via_rng_init()
|
/linux-4.1.27/arch/xtensa/kernel/ |
D | setup.c | 536 static DEFINE_PER_CPU(struct cpu, cpu_data); 543 struct cpu *cpu = &per_cpu(cpu_data, i); in topology_init()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | setup.c | 89 DEFINE_PER_CPU(struct blackfin_cpudata, cpu_data); 173 struct blackfin_cpudata *cpudata = &per_cpu(cpu_data, cpu); in bfin_setup_cpudata() 1117 register_cpu(&per_cpu(cpu_data, cpu).cpu, cpu); 1277 struct blackfin_cpudata *cpudata = &per_cpu(cpu_data, cpu_num);
|
/linux-4.1.27/drivers/misc/mic/card/ |
D | mic_x100.c | 264 struct cpuinfo_x86 *c = &cpu_data(0); in mic_init()
|
/linux-4.1.27/arch/x86/vdso/ |
D | vma.c | 251 if (cpu_has(&cpu_data(cpu), X86_FEATURE_RDTSCP)) in vgetcpu_cpu_init()
|
/linux-4.1.27/arch/parisc/include/asm/ |
D | processor.h | 116 DECLARE_PER_CPU(struct cpuinfo_parisc, cpu_data);
|
/linux-4.1.27/arch/um/kernel/ |
D | um_arch.c | 84 return *pos < NR_CPUS ? cpu_data + *pos : NULL; in c_start()
|
/linux-4.1.27/drivers/crypto/ |
D | padlock-sha.c | 539 struct cpuinfo_x86 *c = &cpu_data(0); in padlock_init() 578 struct cpuinfo_x86 *c = &cpu_data(0); in padlock_fini()
|
D | padlock-aes.c | 513 struct cpuinfo_x86 *c = &cpu_data(0); in padlock_init()
|
/linux-4.1.27/arch/x86/xen/ |
D | smp.c | 84 cpu_data(cpu).x86_max_cores = 1; in cpu_bringup() 329 cpu_data(0).x86_max_cores = 1; in xen_smp_prepare_cpus()
|
/linux-4.1.27/drivers/video/fbdev/geode/ |
D | video_gx.c | 130 if (cpu_data(0).x86_mask == 1) { in gx_set_dclk_frequency()
|
/linux-4.1.27/arch/ia64/sn/kernel/sn2/ |
D | sn_hwperf.c | 476 c = cpu_data(i); in sn_topology_show() 762 cdata = cpu_data(j); in sn_hwperf_ioctl()
|
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/ |
D | therm_throt.c | 246 struct cpuinfo_x86 *c = &cpu_data(cpu); in thermal_throttle_add_dev()
|
D | mce.c | 130 m->socketid = cpu_data(m->extcpu).phys_proc_id; in mce_setup() 131 m->apicid = cpu_data(m->extcpu).initial_apicid; in mce_setup() 277 cpu_data(m->extcpu).microcode); in print_mce()
|
/linux-4.1.27/drivers/thermal/ |
D | x86_pkg_temp_thermal.c | 519 struct cpuinfo_x86 *c = &cpu_data(cpu); in get_core_online()
|
/linux-4.1.27/drivers/xen/ |
D | xen-acpi-processor.c | 428 struct cpuinfo_x86 *c = &cpu_data(0); in check_prereq()
|
/linux-4.1.27/drivers/edac/ |
D | mce_amd.c | 753 struct cpuinfo_x86 *c = &cpu_data(m->extcpu); in amd_decode_mce()
|
D | e752x_edac.c | 1215 char *cpu_id = cpu_data(0).x86_model_id; in e752x_init_sysbus_parity_mask()
|
/linux-4.1.27/drivers/acpi/ |
D | processor_idle.c | 140 if (cpu_has(&cpu_data(pr->id), X86_FEATURE_ARAT)) in lapic_timer_check_state()
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
D | qat_algs.c | 135 return cpu_data(current_thread_info()->cpu).phys_proc_id; in get_current_node()
|
/linux-4.1.27/arch/mips/mm/ |
D | tlbex.c | 331 unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask; in allocate_kscratch()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
D | lpfc_init.c | 8695 cpuinfo = &cpu_data(cpu); in lpfc_sli4_set_affinity()
|