cpu_data 93 arch/alpha/include/asm/mmu_context.h #define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn) cpu_data 145 arch/alpha/include/asm/mmu_context.h cpu_data[cpu].asn_lock = 1; cpu_data 156 arch/alpha/include/asm/mmu_context.h cpu_data[cpu].need_new_asn = 1; cpu_data 191 arch/alpha/include/asm/mmu_context.h cpu_data[cpu].asn_lock = 0; \ cpu_data 193 arch/alpha/include/asm/mmu_context.h if (cpu_data[cpu].need_new_asn) { \ cpu_data 195 arch/alpha/include/asm/mmu_context.h cpu_data[cpu].need_new_asn = 0; \ cpu_data 41 arch/alpha/include/asm/smp.h extern struct cpuinfo_alpha cpu_data[NR_CPUS]; cpu_data 76 arch/alpha/kernel/irq.c seq_printf(p, "%10lu ", cpu_data[j].ipi_count); cpu_data 206 arch/alpha/kernel/proto.h #define mcheck_expected(cpu) (cpu_data[cpu].mcheck_expected) cpu_data 207 arch/alpha/kernel/proto.h #define mcheck_taken(cpu) (cpu_data[cpu].mcheck_taken) cpu_data 208 arch/alpha/kernel/proto.h #define mcheck_extra(cpu) (cpu_data[cpu].mcheck_extra) cpu_data 56 arch/alpha/kernel/smp.c struct cpuinfo_alpha cpu_data[NR_CPUS]; cpu_data 57 arch/alpha/kernel/smp.c EXPORT_SYMBOL(cpu_data); cpu_data 84 arch/alpha/kernel/smp.c cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy; cpu_data 85 arch/alpha/kernel/smp.c cpu_data[cpuid].last_asn = ASN_FIRST_VERSION; cpu_data 86 arch/alpha/kernel/smp.c cpu_data[cpuid].need_new_asn = 0; cpu_data 87 arch/alpha/kernel/smp.c cpu_data[cpuid].asn_lock = 0; cpu_data 96 arch/alpha/kernel/smp.c cpu_data[cpuid].prof_counter = 1; cpu_data 97 arch/alpha/kernel/smp.c cpu_data[cpuid].prof_multiplier = 1; cpu_data 494 arch/alpha/kernel/smp.c bogosum += cpu_data[cpu].loops_per_jiffy; cpu_data 567 arch/alpha/kernel/smp.c cpu_data[this_cpu].ipi_count++; cpu_data 635 arch/alpha/kernel/smp.c #define asn_locked() (cpu_data[smp_processor_id()].asn_lock) cpu_data 37 arch/alpha/lib/udelay.c #define LPJ cpu_data[smp_processor_id()].loops_per_jiffy cpu_data 21 arch/arm/include/asm/cpu.h DECLARE_PER_CPU(struct cpuinfo_arm, cpu_data); cpu_data 38 arch/arm/include/asm/smp_plat.h struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpu); cpu_data 163 arch/arm/kernel/setup.c DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data); cpu_data 1182 arch/arm/kernel/setup.c struct cpuinfo_arm *cpuinfo = &per_cpu(cpu_data, cpu); cpu_data 1251 arch/arm/kernel/setup.c cpuid = is_smp() ? per_cpu(cpu_data, i).cpuid : read_cpuid_id(); cpu_data 1257 arch/arm/kernel/setup.c per_cpu(cpu_data, i).loops_per_jiffy / (500000UL/HZ), cpu_data 1258 arch/arm/kernel/setup.c (per_cpu(cpu_data, i).loops_per_jiffy / (5000UL/HZ)) % 100); cpu_data 371 arch/arm/kernel/smp.c struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid); cpu_data 458 arch/arm/kernel/smp.c bogosum += per_cpu(cpu_data, cpu).loops_per_jiffy; cpu_data 771 arch/arm/kernel/smp.c per_cpu(cpu_data, cpu).loops_per_jiffy; cpu_data 790 arch/arm/kernel/smp.c per_cpu(cpu_data, cpu).loops_per_jiffy = lpj; cpu_data 57 arch/arm64/include/asm/cpu.h DECLARE_PER_CPU(struct cpuinfo_arm64, cpu_data); cpu_data 33 arch/arm64/kernel/cpuinfo.c DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data); cpu_data 133 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, i); cpu_data 255 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); cpu_data 275 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); cpu_data 293 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); cpu_data 384 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data); cpu_data 391 arch/arm64/kernel/cpuinfo.c struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0); cpu_data 384 arch/arm64/kernel/setup.c struct cpu *cpu = &per_cpu(cpu_data.cpu, i); cpu_data 81 arch/c6x/kernel/setup.c static DEFINE_PER_CPU(struct cpuinfo_c6x, cpu_data); cpu_data 97 arch/c6x/kernel/setup.c p = &per_cpu(cpu_data, smp_processor_id()); cpu_data 412 arch/c6x/kernel/setup.c struct cpuinfo_c6x *p = &per_cpu(cpu_data, n); cpu_data 23 arch/csky/include/asm/processor.h extern struct cpuinfo_csky cpu_data[]; cpu_data 17 arch/csky/kernel/process.c struct cpuinfo_csky cpu_data[NR_CPUS]; cpu_data 42 arch/ia64/include/asm/topology.h #define topology_physical_package_id(cpu) (cpu_data(cpu)->socket_id) cpu_data 43 arch/ia64/include/asm/topology.h #define topology_core_id(cpu) (cpu_data(cpu)->core_id) cpu_data 1835 arch/ia64/kernel/mca.c ia64_mca_cpu_init(void *cpu_data) cpu_data 1871 arch/ia64/kernel/mca.c pte_val(mk_pte_phys(__pa(cpu_data), PAGE_KERNEL))); cpu_data 747 arch/ia64/kernel/setup.c return *pos < nr_cpu_ids ? cpu_data(*pos) : NULL; cpu_data 948 arch/ia64/kernel/setup.c void *cpu_data; cpu_data 950 arch/ia64/kernel/setup.c cpu_data = per_cpu_init(); cpu_data 968 arch/ia64/kernel/setup.c ia64_tpa(cpu_data) - (long) __per_cpu_start); cpu_data 980 arch/ia64/kernel/setup.c cpu_info = cpu_data + ((char *) &__ia64_per_cpu_var(ia64_cpu_info) - __per_cpu_start); cpu_data 1026 arch/ia64/kernel/setup.c ia64_mmu_init(ia64_imva(cpu_data)); cpu_data 1027 arch/ia64/kernel/setup.c ia64_mca_cpu_init(ia64_imva(cpu_data)); cpu_data 420 arch/ia64/kernel/smpboot.c last_cpuinfo = cpu_data(cpuid - 1); cpu_data 591 arch/ia64/kernel/smpboot.c if (cpu_data(cpu)->threads_per_core == 1 && cpu_data 592 arch/ia64/kernel/smpboot.c cpu_data(cpu)->cores_per_socket == 1) { cpu_data 701 arch/ia64/kernel/smpboot.c bogosum += cpu_data(cpu)->loops_per_jiffy; cpu_data 713 arch/ia64/kernel/smpboot.c if ((cpu_data(cpu)->socket_id == cpu_data(i)->socket_id)) { cpu_data 716 arch/ia64/kernel/smpboot.c if (cpu_data(cpu)->core_id == cpu_data(i)->core_id) { cpu_data 749 arch/ia64/kernel/smpboot.c if (cpu_data(cpu)->threads_per_core == 1 && cpu_data 750 arch/ia64/kernel/smpboot.c cpu_data(cpu)->cores_per_socket == 1) { cpu_data 845 arch/ia64/kernel/smpboot.c if ((cpu_data(j)->socket_id == cpu_data(i)->socket_id)) { cpu_data 846 arch/ia64/kernel/smpboot.c if (cpu_data(j)->core_id == cpu_data(i)->core_id) cpu_data 35 arch/ia64/kernel/topology.c if (cpu_data(num)->socket_id == -1) cpu_data 36 arch/ia64/kernel/topology.c cpu_data(num)->socket_id = slot; cpu_data 145 arch/ia64/kernel/topology.c if (cpu_data(cpu)->threads_per_core <= 1 && cpu_data 146 arch/ia64/kernel/topology.c cpu_data(cpu)->cores_per_socket <= 1) { cpu_data 160 arch/ia64/kernel/topology.c if (cpu_data(cpu)->socket_id == cpu_data(j)->socket_id cpu_data 161 arch/ia64/kernel/topology.c && cpu_data(j)->core_id == csi.log1_cid cpu_data 162 arch/ia64/kernel/topology.c && cpu_data(j)->thread_id == csi.log1_tid) cpu_data 37 arch/ia64/mm/contig.c static void *cpu_data; cpu_data 61 arch/ia64/mm/contig.c memcpy(cpu_data, src, __per_cpu_end - __per_cpu_start); cpu_data 62 arch/ia64/mm/contig.c __per_cpu_offset[cpu] = (char *)cpu_data - __per_cpu_start; cpu_data 75 arch/ia64/mm/contig.c ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) - cpu_data 78 arch/ia64/mm/contig.c cpu_data += PERCPU_PAGE_SIZE; cpu_data 89 arch/ia64/mm/contig.c cpu_data = memblock_alloc_from(size, PERCPU_PAGE_SIZE, cpu_data 91 arch/ia64/mm/contig.c if (!cpu_data) cpu_data 136 arch/ia64/mm/discontig.c static void *per_cpu_node_setup(void *cpu_data, int node) cpu_data 147 arch/ia64/mm/discontig.c memcpy(__va(cpu_data), src, __per_cpu_end - __per_cpu_start); cpu_data 148 arch/ia64/mm/discontig.c __per_cpu_offset[cpu] = (char *)__va(cpu_data) - cpu_data 162 arch/ia64/mm/discontig.c (unsigned long)cpu_data - cpu_data 165 arch/ia64/mm/discontig.c cpu_data += PERCPU_PAGE_SIZE; cpu_data 168 arch/ia64/mm/discontig.c return cpu_data; cpu_data 262 arch/ia64/mm/discontig.c void *cpu_data; cpu_data 269 arch/ia64/mm/discontig.c cpu_data = (void *)pernode; cpu_data 280 arch/ia64/mm/discontig.c cpu_data = per_cpu_node_setup(cpu_data, node); cpu_data 38 arch/mips/include/asm/bugs.h cpu_data[cpu].udelay_val = loops_per_jiffy; cpu_data 17 arch/mips/include/asm/cpu-features.h #define __ase(ase) (cpu_data[0].ases & (ase)) cpu_data 18 arch/mips/include/asm/cpu-features.h #define __isa(isa) (cpu_data[0].isa_level & (isa)) cpu_data 19 arch/mips/include/asm/cpu-features.h #define __opt(opt) (cpu_data[0].options & (opt)) cpu_data 235 arch/mips/include/asm/cpu-features.h #define cpu_has_vtag_icache (cpu_data[0].icache.flags & MIPS_CACHE_VTAG) cpu_data 238 arch/mips/include/asm/cpu-features.h #define cpu_has_dc_aliases (cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES) cpu_data 241 arch/mips/include/asm/cpu-features.h #define cpu_has_ic_fills_f_dc (cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC) cpu_data 244 arch/mips/include/asm/cpu-features.h #define cpu_has_pindexed_dcache (cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX) cpu_data 260 arch/mips/include/asm/cpu-features.h #define cpu_icache_snoops_remote_store (cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE) cpu_data 430 arch/mips/include/asm/cpu-features.h # define cpu_has_64bits (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT) cpu_data 433 arch/mips/include/asm/cpu-features.h # define cpu_has_64bit_zero_reg (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT) cpu_data 463 arch/mips/include/asm/cpu-features.h # define cpu_vmbits cpu_data[0].vmbits cpu_data 485 arch/mips/include/asm/cpu-features.h #define cpu_dcache_line_size() cpu_data[0].dcache.linesz cpu_data 488 arch/mips/include/asm/cpu-features.h #define cpu_icache_line_size() cpu_data[0].icache.linesz cpu_data 491 arch/mips/include/asm/cpu-features.h #define cpu_scache_line_size() cpu_data[0].scache.linesz cpu_data 494 arch/mips/include/asm/cpu-features.h #define cpu_tcache_line_size() cpu_data[0].tcache.linesz cpu_data 623 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1)) cpu_data 626 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2)) cpu_data 629 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf3 (cpu_data[0].guest.conf & (1 << 3)) cpu_data 632 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf4 (cpu_data[0].guest.conf & (1 << 4)) cpu_data 635 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf5 (cpu_data[0].guest.conf & (1 << 5)) cpu_data 638 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf6 (cpu_data[0].guest.conf & (1 << 6)) cpu_data 641 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf7 (cpu_data[0].guest.conf & (1 << 7)) cpu_data 644 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_fpu (cpu_data[0].guest.options & MIPS_CPU_FPU) cpu_data 647 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_watch (cpu_data[0].guest.options & MIPS_CPU_WATCH) cpu_data 650 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC) cpu_data 653 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_segments (cpu_data[0].guest.options & MIPS_CPU_SEGMENTS) cpu_data 656 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_badinstr (cpu_data[0].guest.options & MIPS_CPU_BADINSTR) cpu_data 659 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_badinstrp (cpu_data[0].guest.options & MIPS_CPU_BADINSTRP) cpu_data 662 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_htw (cpu_data[0].guest.options & MIPS_CPU_HTW) cpu_data 665 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_mvh (cpu_data[0].guest.options & MIPS_CPU_MVH) cpu_data 668 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_msa (cpu_data[0].guest.ases & MIPS_ASE_MSA) cpu_data 671 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_kscr(n) (cpu_data[0].guest.kscratch_mask & (1u << (n))) cpu_data 674 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_rw_llb (cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB)) cpu_data 677 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_perf (cpu_data[0].guest.options & MIPS_CPU_PERF) cpu_data 680 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_maar (cpu_data[0].guest.options & MIPS_CPU_MAAR) cpu_data 683 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_userlocal (cpu_data[0].guest.options & MIPS_CPU_ULRI) cpu_data 690 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_fpu (cpu_data[0].guest.options_dyn & MIPS_CPU_FPU) cpu_data 693 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_watch (cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH) cpu_data 696 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC) cpu_data 699 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_perf (cpu_data[0].guest.options_dyn & MIPS_CPU_PERF) cpu_data 702 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_msa (cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA) cpu_data 705 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_maar (cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR) cpu_data 110 arch/mips/include/asm/cpu-info.h extern struct cpuinfo_mips cpu_data[]; cpu_data 111 arch/mips/include/asm/cpu-info.h #define current_cpu_data cpu_data[smp_processor_id()] cpu_data 112 arch/mips/include/asm/cpu-info.h #define raw_current_cpu_data cpu_data[raw_smp_processor_id()] cpu_data 113 arch/mips/include/asm/cpu-info.h #define boot_cpu_data cpu_data[0] cpu_data 174 arch/mips/include/asm/cpu-info.h struct cpuinfo_mips *infoa = &cpu_data[cpua]; cpu_data 175 arch/mips/include/asm/cpu-info.h struct cpuinfo_mips *infob = &cpu_data[cpub]; cpu_data 212 arch/mips/include/asm/cpu-type.h const int cpu_type = cpu_data[0].cputype; cpu_data 60 arch/mips/include/asm/mach-cavium-octeon/cpu-feature-overrides.h #define cpu_has_rixi (cpu_data[0].cputype != CPU_CAVIUM_OCTEON) cpu_data 398 arch/mips/include/asm/mips-cm.h unsigned int core = cpu_core(&cpu_data[cpu]); cpu_data 399 arch/mips/include/asm/mips-cm.h unsigned int vp = cpu_vpe_id(&cpu_data[cpu]); cpu_data 454 arch/mips/include/asm/mips-cm.h struct cpuinfo_mips *d = &cpu_data[cpu]; cpu_data 96 arch/mips/include/asm/mmu_context.h unsigned long asid_mask = cpu_asid_mask(&cpu_data[cpu]); cpu_data 123 arch/mips/include/asm/mmu_context.h #define asid_cache(cpu) (cpu_data[cpu].asid_cache) cpu_data 125 arch/mips/include/asm/mmu_context.h (cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu])) cpu_data 216 arch/mips/include/asm/mmu_context.h write_c0_memorymapid(ctx & cpu_asid_mask(&cpu_data[cpu])); cpu_data 436 arch/mips/include/asm/pgtable.h prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine; cpu_data 60 arch/mips/include/asm/processor.h #define TASK_SIZE64 (0x1UL << ((cpu_data[0].vmbits>48)?48:cpu_data[0].vmbits)) cpu_data 64 arch/mips/include/asm/timex.h asm volatile("" : "=m" (cpu_data[0].options)); cpu_data 15 arch/mips/include/asm/topology.h #define topology_physical_package_id(cpu) (cpu_data[cpu].package) cpu_data 16 arch/mips/include/asm/topology.h #define topology_core_id(cpu) (cpu_core(&cpu_data[cpu])) cpu_data 65 arch/mips/kernel/cacheinfo.c int cluster = cpu_cluster(&cpu_data[cpu]); cpu_data 68 arch/mips/kernel/cacheinfo.c if (cpu_cluster(&cpu_data[cpu1]) == cluster) cpu_data 357 arch/mips/kernel/cpu-probe.c cpu_data[0].ases &= ~(MIPS_ASE_DSP | MIPS_ASE_DSP2P); cpu_data 370 arch/mips/kernel/cpu-probe.c cpu_data[0].options &= ~MIPS_CPU_HTW; cpu_data 401 arch/mips/kernel/cpu-probe.c if (set_ftlb_enable(&cpu_data[0], 0)) { cpu_data 431 arch/mips/kernel/cpu-probe.c cpu_data[0].tlbsize -= cpu_data[0].tlbsizeftlbways * cpu_data 432 arch/mips/kernel/cpu-probe.c cpu_data[0].tlbsizeftlbsets; cpu_data 433 arch/mips/kernel/cpu-probe.c cpu_data[0].tlbsizeftlbsets = 0; cpu_data 434 arch/mips/kernel/cpu-probe.c cpu_data[0].tlbsizeftlbways = 0; cpu_data 346 arch/mips/kernel/perf_event_mipsxx.c ctrl = M_PERFCTL_VPEID(cpu_vpe_id(&cpu_data[cpu])); cpu_data 447 arch/mips/kernel/pm-cps.c vpe_id = cpu_vpe_id(&cpu_data[cpu]); cpu_data 468 arch/mips/kernel/pm-cps.c cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache, cpu_data 472 arch/mips/kernel/pm-cps.c cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache, cpu_data 485 arch/mips/kernel/pm-cps.c uasm_i_addiu(&p, t0, zero, 1 << cpu_core(&cpu_data[cpu])); cpu_data 499 arch/mips/kernel/pm-cps.c err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu], cpu_data 639 arch/mips/kernel/pm-cps.c unsigned core = cpu_core(&cpu_data[cpu]); cpu_data 40 arch/mips/kernel/proc.c unsigned int version = cpu_data[n].processor_id; cpu_data 41 arch/mips/kernel/proc.c unsigned int fp_vers = cpu_data[n].fpu_id; cpu_data 62 arch/mips/kernel/proc.c cpu_data[n].options & MIPS_CPU_FPU ? " FPU V%d.%d" : ""); cpu_data 67 arch/mips/kernel/proc.c cpu_data[n].udelay_val / (500000/HZ), cpu_data 68 arch/mips/kernel/proc.c (cpu_data[n].udelay_val / (5000/HZ)) % 100); cpu_data 72 arch/mips/kernel/proc.c seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize); cpu_data 79 arch/mips/kernel/proc.c cpu_data[n].watch_reg_count); cpu_data 80 arch/mips/kernel/proc.c for (i = 0; i < cpu_data[n].watch_reg_count; i++) cpu_data 82 arch/mips/kernel/proc.c cpu_data[n].watch_reg_masks[i]); cpu_data 138 arch/mips/kernel/proc.c cpu_data[n].srsets); cpu_data 140 arch/mips/kernel/proc.c hweight8(cpu_data[n].kscratch_mask)); cpu_data 141 arch/mips/kernel/proc.c seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package); cpu_data 142 arch/mips/kernel/proc.c seq_printf(m, "core\t\t\t: %d\n", cpu_core(&cpu_data[n])); cpu_data 146 arch/mips/kernel/proc.c seq_printf(m, "VPE\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n])); cpu_data 148 arch/mips/kernel/proc.c seq_printf(m, "VP\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n])); cpu_data 49 arch/mips/kernel/setup.c struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly; cpu_data 51 arch/mips/kernel/setup.c EXPORT_SYMBOL(cpu_data); cpu_data 74 arch/mips/kernel/smp-cps.c cpu_set_cluster(&cpu_data[nvpes + v], cl); cpu_data 75 arch/mips/kernel/smp-cps.c cpu_set_core(&cpu_data[nvpes + v], c); cpu_data 76 arch/mips/kernel/smp-cps.c cpu_set_vpe_id(&cpu_data[nvpes + v], v); cpu_data 88 arch/mips/kernel/smp-cps.c set_cpu_possible(v, cpu_cluster(&cpu_data[v]) == 0); cpu_data 89 arch/mips/kernel/smp-cps.c set_cpu_present(v, cpu_cluster(&cpu_data[v]) == 0); cpu_data 296 arch/mips/kernel/smp-cps.c unsigned core = cpu_core(&cpu_data[cpu]); cpu_data 297 arch/mips/kernel/smp-cps.c unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); cpu_data 305 arch/mips/kernel/smp-cps.c if (cpu_cluster(&cpu_data[cpu]) != cpu_cluster(&raw_current_cpu_data)) cpu_data 312 arch/mips/kernel/smp-cps.c atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask); cpu_data 409 arch/mips/kernel/smp-cps.c core = cpu_core(&cpu_data[cpu]); cpu_data 412 arch/mips/kernel/smp-cps.c vpe_id = cpu_vpe_id(&cpu_data[cpu]); cpu_data 509 arch/mips/kernel/smp-cps.c unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); cpu_data 523 arch/mips/kernel/smp-cps.c unsigned core = cpu_core(&cpu_data[cpu]); cpu_data 524 arch/mips/kernel/smp-cps.c unsigned int vpe_id = cpu_vpe_id(&cpu_data[cpu]); cpu_data 75 arch/mips/kernel/smp-mt.c cpu_set_vpe_id(&cpu_data[ncpu], tc); cpu_data 105 arch/mips/kernel/smp.c if (cpu_data[cpu].package == cpu_data[i].package) { cpu_data 180 arch/mips/kernel/smp.c core = cpu_core(&cpu_data[cpu]); cpu_data 362 arch/mips/kernel/smp.c cpu_data[cpu].udelay_val = loops_per_jiffy; cpu_data 2213 arch/mips/kernel/traps.c cpu_data[cpu].asid_cache = 0; cpu_data 2214 arch/mips/kernel/traps.c else if (!cpu_data[cpu].asid_cache) cpu_data 2215 arch/mips/kernel/traps.c cpu_data[cpu].asid_cache = asid_first_version(cpu); cpu_data 129 arch/mips/kvm/entry.c unsigned int kscratch_mask = cpu_data[0].kscratch_mask; cpu_data 398 arch/mips/kvm/entry.c UASM_i_LA_mostly(&p, AT, (long)&cpu_data[0].asid_mask); cpu_data 400 arch/mips/kvm/entry.c UASM_i_LW(&p, T2, uasm_rel_lo((long)&cpu_data[0].asid_mask), AT); cpu_data 1713 arch/mips/kvm/vz.c ret += __arch_hweight8(cpu_data[0].guest.kscratch_mask); cpu_data 2312 arch/mips/kvm/vz.c #define guestid_cache(cpu) (cpu_data[cpu].guestid_cache) cpu_data 28 arch/mips/loongson64/loongson-3/irq.c if (cpu_data[cpu].package > 0) cpu_data 313 arch/mips/loongson64/loongson-3/smp.c cpu_set_core(&cpu_data[cpu], cpu_data 315 arch/mips/loongson64/loongson-3/smp.c cpu_data[cpu].package = cpu_data 328 arch/mips/loongson64/loongson-3/smp.c if (cpu_data[cpu].package) cpu_data 380 arch/mips/loongson64/loongson-3/smp.c cpu_set_core(&cpu_data[0], cpu_data 382 arch/mips/loongson64/loongson-3/smp.c cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package; cpu_data 481 arch/mips/loongson64/loongson-3/smp.c [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); cpu_data 563 arch/mips/loongson64/loongson-3/smp.c [sets] "r" (cpu_data[smp_processor_id()].dcache.sets), cpu_data 564 arch/mips/loongson64/loongson-3/smp.c [vsets] "r" (cpu_data[smp_processor_id()].vcache.sets)); cpu_data 626 arch/mips/loongson64/loongson-3/smp.c [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); cpu_data 695 arch/mips/loongson64/loongson-3/smp.c uint64_t core_id = cpu_core(&cpu_data[cpu]); cpu_data 696 arch/mips/loongson64/loongson-3/smp.c uint64_t package_id = cpu_data[cpu].package; cpu_data 709 arch/mips/loongson64/loongson-3/smp.c uint64_t core_id = cpu_core(&cpu_data[cpu]); cpu_data 710 arch/mips/loongson64/loongson-3/smp.c uint64_t package_id = cpu_data[cpu].package; cpu_data 36 arch/mips/mm/context.c if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) { cpu_data 76 arch/mips/mm/context.c mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0); cpu_data 88 arch/mips/mm/context.c __set_bit(mmid & cpu_asid_mask(&cpu_data[cpu]), mmid_map); cpu_data 210 arch/mips/mm/context.c old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache); cpu_data 213 arch/mips/mm/context.c !cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) { cpu_data 220 arch/mips/mm/context.c WRITE_ONCE(cpu_data[cpu].asid_cache, ctx); cpu_data 336 arch/mips/mm/tlbex.c unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask; cpu_data 104 arch/parisc/include/asm/processor.h DECLARE_PER_CPU(struct cpuinfo_parisc, cpu_data); cpu_data 341 arch/parisc/kernel/irq.c return per_cpu(cpu_data, cpu).txn_addr; cpu_data 353 arch/parisc/kernel/irq.c (!per_cpu(cpu_data, next_cpu).txn_addr || cpu_data 541 arch/parisc/kernel/irq.c per_cpu(cpu_data, cpu).hpa); cpu_data 524 arch/parisc/kernel/perf.c cpu_device = per_cpu(cpu_data, 0).dev; cpu_data 526 arch/parisc/kernel/perf.c per_cpu(cpu_data, 0).dev->name); cpu_data 39 arch/parisc/kernel/processor.c DEFINE_PER_CPU(struct cpuinfo_parisc, cpu_data); cpu_data 159 arch/parisc/kernel/processor.c p = &per_cpu(cpu_data, cpuid); cpu_data 337 arch/parisc/kernel/processor.c per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision; cpu_data 338 arch/parisc/kernel/processor.c per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model; cpu_data 378 arch/parisc/kernel/processor.c const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); cpu_data 405 arch/parisc/kernel/setup.c per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision; cpu_data 406 arch/parisc/kernel/setup.c per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model; cpu_data 122 arch/parisc/kernel/smp.c struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); cpu_data 190 arch/parisc/kernel/smp.c struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu); cpu_data 322 arch/parisc/kernel/smp.c const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid); cpu_data 380 arch/parisc/kernel/smp.c int bootstrap_processor = per_cpu(cpu_data, 0).cpuid; cpu_data 68 arch/parisc/kernel/time.c struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); cpu_data 163 arch/parisc/kernel/time.c per_cpu(cpu_data, cpu).it_value = next_tick; cpu_data 255 arch/parisc/kernel/time.c cpu0_loc = per_cpu(cpu_data, 0).cpu_loc; cpu_data 261 arch/parisc/kernel/time.c (cpu0_loc == per_cpu(cpu_data, cpu).cpu_loc)) cpu_data 78 arch/parisc/kernel/topology.c p = &per_cpu(cpu_data, cpuid); cpu_data 80 arch/parisc/kernel/topology.c const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); cpu_data 161 arch/s390/pci/pci_irq.c struct cpu_irq_data *cpu_data; cpu_data 177 arch/s390/pci/pci_irq.c cpu_data = &per_cpu(irq_data, cpu); cpu_data 178 arch/s390/pci/pci_irq.c if (atomic_inc_return(&cpu_data->scheduled) > 1) cpu_data 181 arch/s390/pci/pci_irq.c cpu_data->csd.func = zpci_handle_remote_irq; cpu_data 182 arch/s390/pci/pci_irq.c cpu_data->csd.info = &cpu_data->scheduled; cpu_data 183 arch/s390/pci/pci_irq.c cpu_data->csd.flags = 0; cpu_data 184 arch/s390/pci/pci_irq.c smp_call_function_single_async(cpu, &cpu_data->csd); cpu_data 38 arch/sh/include/asm/mmu_context.h #define asid_cache(cpu) (cpu_data[cpu].asid_cache) cpu_data 94 arch/sh/include/asm/processor.h extern struct sh_cpuinfo cpu_data[]; cpu_data 95 arch/sh/include/asm/processor.h #define boot_cpu_data cpu_data[0] cpu_data 96 arch/sh/include/asm/processor.h #define current_cpu_data cpu_data[smp_processor_id()] cpu_data 97 arch/sh/include/asm/processor.h #define raw_current_cpu_data cpu_data[raw_smp_processor_id()] cpu_data 26 arch/sh/include/asm/tlb_64.h for (tlb = cpu_data->dtlb.first; \ cpu_data 27 arch/sh/include/asm/tlb_64.h tlb <= cpu_data->dtlb.last; \ cpu_data 28 arch/sh/include/asm/tlb_64.h tlb += cpu_data->dtlb.step) cpu_data 36 arch/sh/include/asm/tlb_64.h for (tlb = cpu_data->itlb.first; \ cpu_data 37 arch/sh/include/asm/tlb_64.h tlb <= cpu_data->itlb.last; \ cpu_data 38 arch/sh/include/asm/tlb_64.h tlb += cpu_data->itlb.step) cpu_data 83 arch/sh/kernel/cpu/proc.c unsigned int cpu = c - cpu_data; cpu_data 136 arch/sh/kernel/cpu/proc.c return *pos < NR_CPUS ? cpu_data + *pos : NULL; cpu_data 53 arch/sh/kernel/setup.c struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = { cpu_data 61 arch/sh/kernel/setup.c EXPORT_SYMBOL(cpu_data); cpu_data 49 arch/sh/kernel/smp.c struct sh_cpuinfo *c = cpu_data + cpu; cpu_data 253 arch/sh/kernel/smp.c bogosum += cpu_data[cpu].loops_per_jiffy; cpu_data 40 arch/sh/lib/delay.c "r" (cpu_data[raw_smp_processor_id()].loops_per_jiffy * (HZ/4)) cpu_data 38 arch/sh/lib64/udelay.c __delay(xloops * (HZ * cpu_data[raw_smp_processor_id()].loops_per_jiffy)); cpu_data 74 arch/sh/mm/cache-sh4.c cpu_data->icache.entry_mask); cpu_data 78 arch/sh/mm/cache-sh4.c for (i = 0; i < cpu_data->icache.ways; i++) { cpu_data 81 arch/sh/mm/cache-sh4.c icacheaddr += cpu_data->icache.way_incr; cpu_data 249 arch/sh/mm/cache-sh5.c cpu_data->dcache.entry_mask) >> cpu_data 250 arch/sh/mm/cache-sh5.c cpu_data->dcache.entry_shift; cpu_data 254 arch/sh/mm/cache-sh5.c set_offset &= (cpu_data->dcache.sets - 1); cpu_data 256 arch/sh/mm/cache-sh5.c (set_offset << cpu_data->dcache.entry_shift); cpu_data 265 arch/sh/mm/cache-sh5.c eaddr1 = eaddr0 + cpu_data->dcache.way_size * cpu_data 266 arch/sh/mm/cache-sh5.c cpu_data->dcache.ways; cpu_data 269 arch/sh/mm/cache-sh5.c eaddr += cpu_data->dcache.way_size) { cpu_data 274 arch/sh/mm/cache-sh5.c eaddr1 = eaddr0 + cpu_data->dcache.way_size * cpu_data 275 arch/sh/mm/cache-sh5.c cpu_data->dcache.ways; cpu_data 278 arch/sh/mm/cache-sh5.c eaddr += cpu_data->dcache.way_size) { cpu_data 283 arch/sh/mm/cache-sh5.c if (test_bit(SH_CACHE_MODE_WT, &(cpu_data->dcache.flags))) cpu_data 306 arch/sh/mm/cache-sh5.c sh64_dcache_purge_sets(0, cpu_data->dcache.sets); cpu_data 44 arch/sh/mm/tlb-sh4.c if (cpu_data->flags & CPU_HAS_PTEA) { cpu_data 23 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.entries = 64; cpu_data 24 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.step = 0x10; cpu_data 26 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.first = DTLB_FIXED | cpu_data->dtlb.step; cpu_data 27 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next = cpu_data->dtlb.first; cpu_data 29 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.last = DTLB_FIXED | cpu_data 30 arch/sh/mm/tlb-sh5.c ((cpu_data->dtlb.entries - 1) * cpu_data 31 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.step); cpu_data 34 arch/sh/mm/tlb-sh5.c cpu_data->itlb.entries = 64; cpu_data 35 arch/sh/mm/tlb-sh5.c cpu_data->itlb.step = 0x10; cpu_data 37 arch/sh/mm/tlb-sh5.c cpu_data->itlb.first = ITLB_FIXED | cpu_data->itlb.step; cpu_data 38 arch/sh/mm/tlb-sh5.c cpu_data->itlb.next = cpu_data->itlb.first; cpu_data 39 arch/sh/mm/tlb-sh5.c cpu_data->itlb.last = ITLB_FIXED | cpu_data 40 arch/sh/mm/tlb-sh5.c ((cpu_data->itlb.entries - 1) * cpu_data 41 arch/sh/mm/tlb-sh5.c cpu_data->itlb.step); cpu_data 51 arch/sh/mm/tlb-sh5.c return cpu_data->dtlb.next; cpu_data 61 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.first += cpu_data->dtlb.step; cpu_data 62 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next += cpu_data->dtlb.step; cpu_data 99 arch/sh/mm/tlb-sh5.c if (entry < (cpu_data->dtlb.first - cpu_data->dtlb.step)) cpu_data 103 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.first -= cpu_data->dtlb.step; cpu_data 104 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next = entry; cpu_data 210 arch/sh/mm/tlb-sh5.c tlbp = &cpu_data->itlb; cpu_data 212 arch/sh/mm/tlb-sh5.c tlbp = &cpu_data->dtlb; cpu_data 16 arch/sparc/include/asm/bugs.h cpu_data(0).udelay_val = loops_per_jiffy; cpu_data 28 arch/sparc/include/asm/delay_32.h #define __udelay_val cpu_data(smp_processor_id()).udelay_val cpu_data 46 arch/sparc/include/asm/topology_64.h #define topology_physical_package_id(cpu) (cpu_data(cpu).proc_id) cpu_data 47 arch/sparc/include/asm/topology_64.h #define topology_core_id(cpu) (cpu_data(cpu).core_id) cpu_data 348 arch/sparc/kernel/cpu.c , cpu_data(0).udelay_val/(500000/HZ), cpu_data 349 arch/sparc/kernel/cpu.c (cpu_data(0).udelay_val/(5000/HZ)) % 100, cpu_data 350 arch/sparc/kernel/cpu.c cpu_data(0).clock_tick cpu_data 398 arch/sparc/kernel/cpu.c , cpu_data(0).clock_tick cpu_data 108 arch/sparc/kernel/cpumap.c id = cpu_data(cpu).core_id; cpu_data 111 arch/sparc/kernel/cpumap.c id = cpu_data(cpu).proc_id; cpu_data 130 arch/sparc/kernel/devices.c cpu_data(0).clock_tick = prom_getintdefault(cpu_node, cpu_data 204 arch/sparc/kernel/irq_32.c seq_printf(p, "%10u ", cpu_data(j).irq_resched_count); cpu_data 208 arch/sparc/kernel/irq_32.c seq_printf(p, "%10u ", cpu_data(j).irq_call_count); cpu_data 213 arch/sparc/kernel/irq_32.c seq_printf(p, "%10u ", cpu_data(j).counter); cpu_data 308 arch/sparc/kernel/irq_64.c seq_printf(p, "%10u ", cpu_data(j).__nmi_count); cpu_data 242 arch/sparc/kernel/leon_smp.c prev = &cpu_data(i).next; cpu_data 888 arch/sparc/kernel/mdesc.c cpu_data(*id).core_id = core_id; cpu_data 897 arch/sparc/kernel/mdesc.c cpu_data(*id).max_cache_id = max_cache_id; cpu_data 903 arch/sparc/kernel/mdesc.c cpu_data(*id).sock_id = max_cache_id; cpu_data 989 arch/sparc/kernel/mdesc.c cpu_data(*id).sock_id = idx; cpu_data 1027 arch/sparc/kernel/mdesc.c cpu_data(*id).proc_id = proc_id; cpu_data 1197 arch/sparc/kernel/mdesc.c c = &cpu_data(cpuid); cpu_data 127 arch/sparc/kernel/nmi.c return cpu_data(cpu).__nmi_count; cpu_data 416 arch/sparc/kernel/prom_64.c int proc_id = cpu_data(cpu).proc_id; cpu_data 530 arch/sparc/kernel/prom_64.c cpu_data(cpuid).clock_tick = cpu_data 534 arch/sparc/kernel/prom_64.c cpu_data(cpuid).dcache_size = cpu_data 537 arch/sparc/kernel/prom_64.c cpu_data(cpuid).dcache_line_size = cpu_data 540 arch/sparc/kernel/prom_64.c cpu_data(cpuid).icache_size = cpu_data 543 arch/sparc/kernel/prom_64.c cpu_data(cpuid).icache_line_size = cpu_data 546 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_size = cpu_data 548 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_line_size = cpu_data 550 arch/sparc/kernel/prom_64.c if (!cpu_data(cpuid).ecache_size || cpu_data 551 arch/sparc/kernel/prom_64.c !cpu_data(cpuid).ecache_line_size) { cpu_data 552 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_size = cpu_data 556 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_line_size = cpu_data 561 arch/sparc/kernel/prom_64.c cpu_data(cpuid).core_id = portid + 1; cpu_data 562 arch/sparc/kernel/prom_64.c cpu_data(cpuid).proc_id = portid; cpu_data 564 arch/sparc/kernel/prom_64.c cpu_data(cpuid).dcache_size = cpu_data 566 arch/sparc/kernel/prom_64.c cpu_data(cpuid).dcache_line_size = cpu_data 569 arch/sparc/kernel/prom_64.c cpu_data(cpuid).icache_size = cpu_data 571 arch/sparc/kernel/prom_64.c cpu_data(cpuid).icache_line_size = cpu_data 574 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_size = cpu_data 577 arch/sparc/kernel/prom_64.c cpu_data(cpuid).ecache_line_size = cpu_data 580 arch/sparc/kernel/prom_64.c cpu_data(cpuid).core_id = 0; cpu_data 581 arch/sparc/kernel/prom_64.c cpu_data(cpuid).proc_id = -1; cpu_data 63 arch/sparc/kernel/smp_32.c cpu_data(id).udelay_val = loops_per_jiffy; cpu_data 66 arch/sparc/kernel/smp_32.c cpu_data(id).clock_tick = prom_getintdefault(cpu_node, cpu_data 68 arch/sparc/kernel/smp_32.c cpu_data(id).prom_node = cpu_node; cpu_data 75 arch/sparc/kernel/smp_32.c cpu_data(id).mid = mid; cpu_data 85 arch/sparc/kernel/smp_32.c bogosum += cpu_data(cpu).udelay_val; cpu_data 385 arch/sparc/kernel/smp_32.c cpu_data(i).udelay_val/(500000/HZ), cpu_data 386 arch/sparc/kernel/smp_32.c (cpu_data(i).udelay_val/(5000/HZ))%100); cpu_data 97 arch/sparc/kernel/smp_64.c i, cpu_data(i).clock_tick); cpu_data 1275 arch/sparc/kernel/smp_64.c if (cpu_data(i).core_id == 0) { cpu_data 1281 arch/sparc/kernel/smp_64.c if (cpu_data(i).core_id == cpu_data 1282 arch/sparc/kernel/smp_64.c cpu_data(j).core_id) cpu_data 1291 arch/sparc/kernel/smp_64.c if (cpu_data(i).max_cache_id == cpu_data 1292 arch/sparc/kernel/smp_64.c cpu_data(j).max_cache_id) cpu_data 1295 arch/sparc/kernel/smp_64.c if (cpu_data(i).sock_id == cpu_data(j).sock_id) cpu_data 1304 arch/sparc/kernel/smp_64.c if (cpu_data(i).proc_id == -1) { cpu_data 1310 arch/sparc/kernel/smp_64.c if (cpu_data(i).proc_id == cpu_data 1311 arch/sparc/kernel/smp_64.c cpu_data(j).proc_id) cpu_data 1388 arch/sparc/kernel/smp_64.c c = &cpu_data(cpu); cpu_data 173 arch/sparc/kernel/sun4d_smp.c prev = &cpu_data(i).next; cpu_data 131 arch/sparc/kernel/sun4m_smp.c prev = &cpu_data(i).next; cpu_data 176 arch/sparc/kernel/sysfs.c cpuinfo_sparc *c = &cpu_data(dev->id); \ cpu_data 184 arch/sparc/kernel/sysfs.c cpuinfo_sparc *c = &cpu_data(dev->id); \ cpu_data 646 arch/sparc/kernel/time_64.c return cpu_data(cpu).clock_tick; cpu_data 664 arch/sparc/kernel/time_64.c ft->clock_tick_ref = cpu_data(cpu).clock_tick; cpu_data 669 arch/sparc/kernel/time_64.c cpu_data(cpu).clock_tick = cpu_data 869 arch/sparc/kernel/traps_64.c val = cpu_data(i).ecache_size; cpu_data 876 arch/sparc/kernel/traps_64.c val = cpu_data(i).ecache_line_size; cpu_data 80 arch/um/kernel/um_arch.c return *pos < NR_CPUS ? cpu_data + *pos : NULL; cpu_data 210 arch/x86/events/amd/uncore.c int thread = 2 * (cpu_data(event->cpu).cpu_core_id % 4); cpu_data 213 arch/x86/events/amd/uncore.c thread += cpu_data(event->cpu).apicid & 1; cpu_data 91 arch/x86/include/asm/acpi.h struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 98 arch/x86/include/asm/acpi.h struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 107 arch/x86/include/asm/topology.h #define topology_logical_package_id(cpu) (cpu_data(cpu).logical_proc_id) cpu_data 108 arch/x86/include/asm/topology.h #define topology_physical_package_id(cpu) (cpu_data(cpu).phys_proc_id) cpu_data 109 arch/x86/include/asm/topology.h #define topology_logical_die_id(cpu) (cpu_data(cpu).logical_die_id) cpu_data 110 arch/x86/include/asm/topology.h #define topology_die_id(cpu) (cpu_data(cpu).cpu_die_id) cpu_data 111 arch/x86/include/asm/topology.h #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id) cpu_data 32 arch/x86/kernel/acpi/cstate.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 149 arch/x86/kernel/acpi/cstate.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 555 arch/x86/kernel/alternative.c clear_cpu_cap(&cpu_data(0), X86_FEATURE_UP); cpu_data 389 arch/x86/kernel/amd_nb.c return (mask >> (4 * cpu_data(cpu).cpu_core_id)) & 0xf; cpu_data 415 arch/x86/kernel/amd_nb.c cuid = cpu_data(cpu).cpu_core_id; cpu_data 905 arch/x86/kernel/cpu/cacheinfo.c apicid = cpu_data(cpu).apicid; cpu_data 914 arch/x86/kernel/cpu/cacheinfo.c apicid = cpu_data(i).apicid; cpu_data 921 arch/x86/kernel/cpu/cacheinfo.c apicid = cpu_data(sibling).apicid; cpu_data 941 arch/x86/kernel/cpu/cacheinfo.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 959 arch/x86/kernel/cpu/cacheinfo.c if (cpu_data(i).apicid >> index_msb == c->apicid >> index_msb) { cpu_data 1008 arch/x86/kernel/cpu/cacheinfo.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 141 arch/x86/kernel/cpu/mce/core.c m->socketid = cpu_data(m->extcpu).phys_proc_id; cpu_data 142 arch/x86/kernel/cpu/mce/core.c m->apicid = cpu_data(m->extcpu).initial_apicid; cpu_data 241 arch/x86/kernel/cpu/mce/therm_throt.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 649 arch/x86/kernel/cpu/microcode/amd.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 671 arch/x86/kernel/cpu/microcode/amd.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 892 arch/x86/kernel/cpu/microcode/amd.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 765 arch/x86/kernel/cpu/microcode/intel.c struct cpuinfo_x86 *c = &cpu_data(cpu_num); cpu_data 793 arch/x86/kernel/cpu/microcode/intel.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 947 arch/x86/kernel/cpu/microcode/intel.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 971 arch/x86/kernel/cpu/microcode/intel.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 148 arch/x86/kernel/cpu/proc.c return &cpu_data(*pos); cpu_data 111 arch/x86/kernel/cpuid.c c = &cpu_data(cpu); cpu_data 167 arch/x86/kernel/msr.c c = &cpu_data(cpu); cpu_data 193 arch/x86/kernel/smpboot.c cpu_data(cpuid).loops_per_jiffy = loops_per_jiffy; cpu_data 302 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 318 arch/x86/kernel/smpboot.c int proc_id = cpu_data(cur_cpu).phys_proc_id; cpu_data 321 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 351 arch/x86/kernel/smpboot.c cpu_data(cpu).logical_proc_id = new; cpu_data 374 arch/x86/kernel/smpboot.c cpu_data(cpu).logical_die_id = new; cpu_data 381 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(id); cpu_data 396 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(id); cpu_data 577 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 593 arch/x86/kernel/smpboot.c o = &cpu_data(i); cpu_data 608 arch/x86/kernel/smpboot.c o = &cpu_data(i); cpu_data 630 arch/x86/kernel/smpboot.c cpu_data(i).booted_cores++; cpu_data 632 arch/x86/kernel/smpboot.c c->booted_cores = cpu_data(i).booted_cores; cpu_data 662 arch/x86/kernel/smpboot.c bogosum += cpu_data(cpu).loops_per_jiffy; cpu_data 1299 arch/x86/kernel/smpboot.c c = &cpu_data(i); cpu_data 1372 arch/x86/kernel/smpboot.c print_cpu_info(&cpu_data(0)); cpu_data 1414 arch/x86/kernel/smpboot.c ncpus = cpu_data(0).booted_cores * topology_max_smt_threads(); cpu_data 1543 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 1551 arch/x86/kernel/smpboot.c cpu_data(sibling).booted_cores--; cpu_data 109 arch/x86/kernel/topology.c struct cpuinfo_x86 *c = &cpu_data(num); cpu_data 903 arch/x86/kernel/tsc.c cpu_data(0).loops_per_jiffy = cpufreq_scale(cpu_data(0).loops_per_jiffy, cpu_data 1525 arch/x86/kernel/tsc.c int constant_tsc = cpu_has(&cpu_data(cpu), X86_FEATURE_CONSTANT_TSC); cpu_data 1533 arch/x86/kernel/tsc.c return cpu_data(sibling).loops_per_jiffy; cpu_data 149 arch/x86/xen/apic.c return cpu_data(cpu).apicid; cpu_data 73 arch/x86/xen/smp_pv.c cpu_data(cpu).x86_max_cores = 1; cpu_data 251 arch/x86/xen/smp_pv.c cpu_data(0).x86_max_cores = 1; cpu_data 407 arch/xtensa/kernel/setup.c static DEFINE_PER_CPU(struct cpu, cpu_data); cpu_data 414 arch/xtensa/kernel/setup.c struct cpu *cpu = &per_cpu(cpu_data, i); cpu_data 130 drivers/acpi/processor_idle.c if (cpu_has(&cpu_data(pr->id), X86_FEATURE_ARAT)) cpu_data 131 drivers/char/hw_random/via-rng.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 179 drivers/cpufreq/acpi-cpufreq.c struct cpuinfo_x86 *cpu = &cpu_data(cpuid); cpu_data 186 drivers/cpufreq/acpi-cpufreq.c struct cpuinfo_x86 *cpu = &cpu_data(cpuid); cpu_data 635 drivers/cpufreq/acpi-cpufreq.c struct cpuinfo_x86 *c = &cpu_data(policy->cpu); cpu_data 36 drivers/cpufreq/amd_freq_sensitivity.c static DEFINE_PER_CPU(struct cpu_data_t, cpu_data); cpu_data 45 drivers/cpufreq/amd_freq_sensitivity.c struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu); cpu_data 183 drivers/cpufreq/e_powersaver.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 146 drivers/cpufreq/elanfreq.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 522 drivers/cpufreq/intel_pstate.c static s16 intel_pstate_get_epb(struct cpudata *cpu_data) cpu_data 530 drivers/cpufreq/intel_pstate.c ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); cpu_data 537 drivers/cpufreq/intel_pstate.c static s16 intel_pstate_get_epp(struct cpudata *cpu_data, u64 hwp_req_data) cpu_data 547 drivers/cpufreq/intel_pstate.c epp = rdmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, cpu_data 555 drivers/cpufreq/intel_pstate.c epp = intel_pstate_get_epb(cpu_data); cpu_data 605 drivers/cpufreq/intel_pstate.c static int intel_pstate_get_energy_pref_index(struct cpudata *cpu_data) cpu_data 610 drivers/cpufreq/intel_pstate.c epp = intel_pstate_get_epp(cpu_data, 0); cpu_data 640 drivers/cpufreq/intel_pstate.c static int intel_pstate_set_energy_pref_index(struct cpudata *cpu_data, cpu_data 647 drivers/cpufreq/intel_pstate.c epp = cpu_data->epp_default; cpu_data 654 drivers/cpufreq/intel_pstate.c ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, &value); cpu_data 664 drivers/cpufreq/intel_pstate.c ret = wrmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, value); cpu_data 668 drivers/cpufreq/intel_pstate.c ret = intel_pstate_set_epb(cpu_data->cpu, epp); cpu_data 695 drivers/cpufreq/intel_pstate.c struct cpudata *cpu_data = all_cpu_data[policy->cpu]; cpu_data 707 drivers/cpufreq/intel_pstate.c intel_pstate_set_energy_pref_index(cpu_data, ret); cpu_data 714 drivers/cpufreq/intel_pstate.c struct cpudata *cpu_data = all_cpu_data[policy->cpu]; cpu_data 717 drivers/cpufreq/intel_pstate.c preference = intel_pstate_get_energy_pref_index(cpu_data); cpu_data 769 drivers/cpufreq/intel_pstate.c struct cpudata *cpu_data = all_cpu_data[cpu]; cpu_data 774 drivers/cpufreq/intel_pstate.c max = cpu_data->max_perf_ratio; cpu_data 775 drivers/cpufreq/intel_pstate.c min = cpu_data->min_perf_ratio; cpu_data 777 drivers/cpufreq/intel_pstate.c if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) cpu_data 788 drivers/cpufreq/intel_pstate.c if (cpu_data->epp_policy == cpu_data->policy) cpu_data 791 drivers/cpufreq/intel_pstate.c cpu_data->epp_policy = cpu_data->policy; cpu_data 793 drivers/cpufreq/intel_pstate.c if (cpu_data->epp_saved >= 0) { cpu_data 794 drivers/cpufreq/intel_pstate.c epp = cpu_data->epp_saved; cpu_data 795 drivers/cpufreq/intel_pstate.c cpu_data->epp_saved = -EINVAL; cpu_data 799 drivers/cpufreq/intel_pstate.c if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) { cpu_data 800 drivers/cpufreq/intel_pstate.c epp = intel_pstate_get_epp(cpu_data, value); cpu_data 801 drivers/cpufreq/intel_pstate.c cpu_data->epp_powersave = epp; cpu_data 809 drivers/cpufreq/intel_pstate.c if (cpu_data->epp_powersave < 0) cpu_data 819 drivers/cpufreq/intel_pstate.c epp = intel_pstate_get_epp(cpu_data, value); cpu_data 823 drivers/cpufreq/intel_pstate.c epp = cpu_data->epp_powersave; cpu_data 833 drivers/cpufreq/intel_pstate.c WRITE_ONCE(cpu_data->hwp_req_cached, value); cpu_data 859 drivers/cpufreq/intel_pstate.c struct cpudata *cpu_data = all_cpu_data[policy->cpu]; cpu_data 864 drivers/cpufreq/intel_pstate.c cpu_data->epp_saved = intel_pstate_get_epp(cpu_data, 0); cpu_data 2023 drivers/cpufreq/intel_pstate.c struct cpudata *cpu_data = all_cpu_data[cpu]; cpu_data 2025 drivers/cpufreq/intel_pstate.c if (!cpu_data->update_util_set) cpu_data 2029 drivers/cpufreq/intel_pstate.c cpu_data->update_util_set = false; cpu_data 762 drivers/cpufreq/longhaul.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 920 drivers/cpufreq/longhaul.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 166 drivers/cpufreq/longrun.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 157 drivers/cpufreq/p4-clockmod.c struct cpuinfo_x86 *c = &cpu_data(policy->cpu); cpu_data 119 drivers/cpufreq/powernow-k7.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 168 drivers/cpufreq/qoriq-cpufreq.c struct cpu_data *data; cpu_data 235 drivers/cpufreq/qoriq-cpufreq.c struct cpu_data *data = policy->driver_data; cpu_data 249 drivers/cpufreq/qoriq-cpufreq.c struct cpu_data *data = policy->driver_data; cpu_data 73 drivers/cpufreq/sc520_freq.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 236 drivers/cpufreq/speedstep-centrino.c struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); cpu_data 347 drivers/cpufreq/speedstep-centrino.c struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); cpu_data 256 drivers/cpufreq/speedstep-lib.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 119 drivers/cpuidle/governors/teo.c struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); cpu_data 120 drivers/cpuidle/governors/teo.c unsigned int sleep_length_us = ktime_to_us(cpu_data->sleep_length_ns); cpu_data 124 drivers/cpuidle/governors/teo.c if (cpu_data->time_span_ns >= cpu_data->sleep_length_ns) { cpu_data 136 drivers/cpuidle/governors/teo.c measured_us = ktime_to_us(cpu_data->time_span_ns); cpu_data 154 drivers/cpuidle/governors/teo.c unsigned int early_hits = cpu_data->states[i].early_hits; cpu_data 156 drivers/cpuidle/governors/teo.c cpu_data->states[i].early_hits -= early_hits >> DECAY_SHIFT; cpu_data 174 drivers/cpuidle/governors/teo.c unsigned int hits = cpu_data->states[idx_timer].hits; cpu_data 175 drivers/cpuidle/governors/teo.c unsigned int misses = cpu_data->states[idx_timer].misses; cpu_data 183 drivers/cpuidle/governors/teo.c cpu_data->states[idx_hit].early_hits += PULSE; cpu_data 188 drivers/cpuidle/governors/teo.c cpu_data->states[idx_timer].misses = misses; cpu_data 189 drivers/cpuidle/governors/teo.c cpu_data->states[idx_timer].hits = hits; cpu_data 196 drivers/cpuidle/governors/teo.c cpu_data->intervals[cpu_data->interval_idx++] = measured_us; cpu_data 197 drivers/cpuidle/governors/teo.c if (cpu_data->interval_idx >= INTERVALS) cpu_data 198 drivers/cpuidle/governors/teo.c cpu_data->interval_idx = 0; cpu_data 234 drivers/cpuidle/governors/teo.c struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); cpu_data 245 drivers/cpuidle/governors/teo.c cpu_data->time_span_ns = local_clock(); cpu_data 247 drivers/cpuidle/governors/teo.c cpu_data->sleep_length_ns = tick_nohz_get_sleep_length(&delta_tick); cpu_data 248 drivers/cpuidle/governors/teo.c duration_us = ktime_to_us(cpu_data->sleep_length_ns); cpu_data 278 drivers/cpuidle/governors/teo.c hits = cpu_data->states[i].hits; cpu_data 279 drivers/cpuidle/governors/teo.c misses = cpu_data->states[i].misses; cpu_data 281 drivers/cpuidle/governors/teo.c if (early_hits >= cpu_data->states[i].early_hits || cpu_data 293 drivers/cpuidle/governors/teo.c early_hits = cpu_data->states[i].early_hits; cpu_data 308 drivers/cpuidle/governors/teo.c early_hits = cpu_data->states[i].early_hits; cpu_data 317 drivers/cpuidle/governors/teo.c hits = cpu_data->states[i].hits; cpu_data 318 drivers/cpuidle/governors/teo.c misses = cpu_data->states[i].misses; cpu_data 328 drivers/cpuidle/governors/teo.c hits = cpu_data->states[i].hits; cpu_data 329 drivers/cpuidle/governors/teo.c misses = cpu_data->states[i].misses; cpu_data 331 drivers/cpuidle/governors/teo.c if (early_hits < cpu_data->states[i].early_hits && cpu_data 335 drivers/cpuidle/governors/teo.c early_hits = cpu_data->states[i].early_hits; cpu_data 381 drivers/cpuidle/governors/teo.c unsigned int val = cpu_data->intervals[i]; cpu_data 440 drivers/cpuidle/governors/teo.c struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); cpu_data 449 drivers/cpuidle/governors/teo.c (tick_nohz_idle_got_tick() && cpu_data->sleep_length_ns > TICK_NSEC)) { cpu_data 451 drivers/cpuidle/governors/teo.c cpu_data->time_span_ns = cpu_data->sleep_length_ns; cpu_data 453 drivers/cpuidle/governors/teo.c cpu_data->time_span_ns = local_clock() - cpu_data->time_span_ns; cpu_data 465 drivers/cpuidle/governors/teo.c struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); cpu_data 468 drivers/cpuidle/governors/teo.c memset(cpu_data, 0, sizeof(*cpu_data)); cpu_data 471 drivers/cpuidle/governors/teo.c cpu_data->intervals[i] = UINT_MAX; cpu_data 499 drivers/crypto/padlock-aes.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 507 drivers/crypto/padlock-sha.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 546 drivers/crypto/padlock-sha.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 1214 drivers/edac/e752x_edac.c char *cpu_id = cpu_data(0).x86_model_id; cpu_data 884 drivers/gpu/drm/amd/amdkfd/kfd_crat.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 1443 drivers/gpu/drm/amd/amdkfd/kfd_topology.c return cpu_data(first_cpu_of_numa_node).apicid; cpu_data 49 drivers/hwmon/coretemp.c #define TO_CORE_ID(cpu) (cpu_data(cpu).cpu_core_id) cpu_data 409 drivers/hwmon/coretemp.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 455 drivers/hwmon/coretemp.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 592 drivers/hwmon/coretemp.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 149 drivers/hwmon/fam15h_power.c cu = cpu_data(cpu).cpu_core_id; cpu_data 277 drivers/hwmon/hwmon-vid.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 113 drivers/hwmon/via-cputemp.c struct cpuinfo_x86 *c = &cpu_data(pdev->id); cpu_data 103 drivers/irqchip/irq-mips-cpu.c settc(cpu_vpe_id(&cpu_data[cpu])); cpu_data 302 drivers/misc/mic/card/mic_x100.c struct cpuinfo_x86 *c = &cpu_data(0); cpu_data 1293 drivers/powercap/intel_rapl_common.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 10675 drivers/scsi/lpfc/lpfc_init.c cpuinfo = &cpu_data(cpu); cpu_data 466 drivers/thermal/intel/x86_pkg_temp_thermal.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu_data 126 drivers/video/fbdev/geode/video_gx.c if (cpu_data(0).x86_stepping == 1) { cpu_data 31 kernel/trace/trace_functions_graph.c struct fgraph_cpu_data __percpu *cpu_data; cpu_data 393 kernel/trace/trace_functions_graph.c last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid); cpu_data 644 kernel/trace/trace_functions_graph.c struct fgraph_cpu_data *cpu_data; cpu_data 646 kernel/trace/trace_functions_graph.c cpu_data = per_cpu_ptr(data->cpu_data, cpu); cpu_data 653 kernel/trace/trace_functions_graph.c cpu_data->depth = call->depth - 1; cpu_data 658 kernel/trace/trace_functions_graph.c cpu_data->enter_funcs[call->depth] = 0; cpu_data 687 kernel/trace/trace_functions_graph.c struct fgraph_cpu_data *cpu_data; cpu_data 690 kernel/trace/trace_functions_graph.c cpu_data = per_cpu_ptr(data->cpu_data, cpu); cpu_data 691 kernel/trace/trace_functions_graph.c cpu_data->depth = call->depth; cpu_data 696 kernel/trace/trace_functions_graph.c cpu_data->enter_funcs[call->depth] = call->func; cpu_data 790 kernel/trace/trace_functions_graph.c depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq); cpu_data 836 kernel/trace/trace_functions_graph.c depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq); cpu_data 917 kernel/trace/trace_functions_graph.c struct fgraph_cpu_data *cpu_data; cpu_data 920 kernel/trace/trace_functions_graph.c cpu_data = per_cpu_ptr(data->cpu_data, cpu); cpu_data 927 kernel/trace/trace_functions_graph.c cpu_data->depth = trace->depth - 1; cpu_data 931 kernel/trace/trace_functions_graph.c if (cpu_data->enter_funcs[trace->depth] != trace->func) cpu_data 933 kernel/trace/trace_functions_graph.c cpu_data->enter_funcs[trace->depth] = 0; cpu_data 982 kernel/trace/trace_functions_graph.c depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth; cpu_data 1048 kernel/trace/trace_functions_graph.c if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) { cpu_data 1049 kernel/trace/trace_functions_graph.c per_cpu_ptr(data->cpu_data, cpu)->ignore = 0; cpu_data 1062 kernel/trace/trace_functions_graph.c per_cpu_ptr(data->cpu_data, iter->cpu)->ignore = 1; cpu_data 1216 kernel/trace/trace_functions_graph.c data->cpu_data = alloc_percpu_gfp(struct fgraph_cpu_data, gfpflags); cpu_data 1217 kernel/trace/trace_functions_graph.c if (!data->cpu_data) cpu_data 1221 kernel/trace/trace_functions_graph.c pid_t *pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid); cpu_data 1222 kernel/trace/trace_functions_graph.c int *depth = &(per_cpu_ptr(data->cpu_data, cpu)->depth); cpu_data 1223 kernel/trace/trace_functions_graph.c int *ignore = &(per_cpu_ptr(data->cpu_data, cpu)->ignore); cpu_data 1224 kernel/trace/trace_functions_graph.c int *depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq); cpu_data 1247 kernel/trace/trace_functions_graph.c free_percpu(data->cpu_data); cpu_data 11 tools/testing/selftests/bpf/prog_tests/perf_buffer.c int cpu_data = *(int *)data, duration = 0; cpu_data 14 tools/testing/selftests/bpf/prog_tests/perf_buffer.c if (cpu_data != cpu) cpu_data 15 tools/testing/selftests/bpf/prog_tests/perf_buffer.c CHECK(cpu_data != cpu, "check_cpu_data", cpu_data 16 tools/testing/selftests/bpf/prog_tests/perf_buffer.c "cpu_data %d != cpu %d\n", cpu_data, cpu); cpu_data 370 virt/kvm/arm/arm.c kvm_host_data_t *cpu_data; cpu_data 373 virt/kvm/arm/arm.c cpu_data = this_cpu_ptr(&kvm_host_data); cpu_data 385 virt/kvm/arm/arm.c vcpu->arch.host_cpu_context = &cpu_data->host_ctxt; cpu_data 1601 virt/kvm/arm/arm.c kvm_host_data_t *cpu_data; cpu_data 1603 virt/kvm/arm/arm.c cpu_data = per_cpu_ptr(&kvm_host_data, cpu); cpu_data 1604 virt/kvm/arm/arm.c err = create_hyp_mappings(cpu_data, cpu_data + 1, PAGE_HYP);