cpu_data           93 arch/alpha/include/asm/mmu_context.h #define cpu_last_asn(cpuid)	(cpu_data[cpuid].last_asn)
cpu_data          145 arch/alpha/include/asm/mmu_context.h 	cpu_data[cpu].asn_lock = 1;
cpu_data          156 arch/alpha/include/asm/mmu_context.h 		cpu_data[cpu].need_new_asn = 1;
cpu_data          191 arch/alpha/include/asm/mmu_context.h 	cpu_data[cpu].asn_lock = 0;				\
cpu_data          193 arch/alpha/include/asm/mmu_context.h 	if (cpu_data[cpu].need_new_asn) {			\
cpu_data          195 arch/alpha/include/asm/mmu_context.h 		cpu_data[cpu].need_new_asn = 0;			\
cpu_data           41 arch/alpha/include/asm/smp.h extern struct cpuinfo_alpha cpu_data[NR_CPUS];
cpu_data           76 arch/alpha/kernel/irq.c 		seq_printf(p, "%10lu ", cpu_data[j].ipi_count);
cpu_data          206 arch/alpha/kernel/proto.h #define mcheck_expected(cpu)	(cpu_data[cpu].mcheck_expected)
cpu_data          207 arch/alpha/kernel/proto.h #define mcheck_taken(cpu)	(cpu_data[cpu].mcheck_taken)
cpu_data          208 arch/alpha/kernel/proto.h #define mcheck_extra(cpu)	(cpu_data[cpu].mcheck_extra)
cpu_data           56 arch/alpha/kernel/smp.c struct cpuinfo_alpha cpu_data[NR_CPUS];
cpu_data           57 arch/alpha/kernel/smp.c EXPORT_SYMBOL(cpu_data);
cpu_data           84 arch/alpha/kernel/smp.c 	cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy;
cpu_data           85 arch/alpha/kernel/smp.c 	cpu_data[cpuid].last_asn = ASN_FIRST_VERSION;
cpu_data           86 arch/alpha/kernel/smp.c 	cpu_data[cpuid].need_new_asn = 0;
cpu_data           87 arch/alpha/kernel/smp.c 	cpu_data[cpuid].asn_lock = 0;
cpu_data           96 arch/alpha/kernel/smp.c 	cpu_data[cpuid].prof_counter = 1;
cpu_data           97 arch/alpha/kernel/smp.c 	cpu_data[cpuid].prof_multiplier = 1;
cpu_data          494 arch/alpha/kernel/smp.c 			bogosum += cpu_data[cpu].loops_per_jiffy;
cpu_data          567 arch/alpha/kernel/smp.c 	cpu_data[this_cpu].ipi_count++;
cpu_data          635 arch/alpha/kernel/smp.c #define asn_locked() (cpu_data[smp_processor_id()].asn_lock)
cpu_data           37 arch/alpha/lib/udelay.c #define LPJ	 cpu_data[smp_processor_id()].loops_per_jiffy
cpu_data           21 arch/arm/include/asm/cpu.h DECLARE_PER_CPU(struct cpuinfo_arm, cpu_data);
cpu_data           38 arch/arm/include/asm/smp_plat.h 	struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpu);
cpu_data          163 arch/arm/kernel/setup.c DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data);
cpu_data         1182 arch/arm/kernel/setup.c 		struct cpuinfo_arm *cpuinfo = &per_cpu(cpu_data, cpu);
cpu_data         1251 arch/arm/kernel/setup.c 		cpuid = is_smp() ? per_cpu(cpu_data, i).cpuid : read_cpuid_id();
cpu_data         1257 arch/arm/kernel/setup.c 			   per_cpu(cpu_data, i).loops_per_jiffy / (500000UL/HZ),
cpu_data         1258 arch/arm/kernel/setup.c 			   (per_cpu(cpu_data, i).loops_per_jiffy / (5000UL/HZ)) % 100);
cpu_data          371 arch/arm/kernel/smp.c 	struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid);
cpu_data          458 arch/arm/kernel/smp.c 		bogosum += per_cpu(cpu_data, cpu).loops_per_jiffy;
cpu_data          771 arch/arm/kernel/smp.c 				per_cpu(cpu_data, cpu).loops_per_jiffy;
cpu_data          790 arch/arm/kernel/smp.c 			per_cpu(cpu_data, cpu).loops_per_jiffy = lpj;
cpu_data           57 arch/arm64/include/asm/cpu.h DECLARE_PER_CPU(struct cpuinfo_arm64, cpu_data);
cpu_data           33 arch/arm64/kernel/cpuinfo.c DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data);
cpu_data          133 arch/arm64/kernel/cpuinfo.c 		struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, i);
cpu_data          255 arch/arm64/kernel/cpuinfo.c 	struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
cpu_data          275 arch/arm64/kernel/cpuinfo.c 	struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
cpu_data          293 arch/arm64/kernel/cpuinfo.c 		struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
cpu_data          384 arch/arm64/kernel/cpuinfo.c 	struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data);
cpu_data          391 arch/arm64/kernel/cpuinfo.c 	struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0);
cpu_data          384 arch/arm64/kernel/setup.c 		struct cpu *cpu = &per_cpu(cpu_data.cpu, i);
cpu_data           81 arch/c6x/kernel/setup.c static DEFINE_PER_CPU(struct cpuinfo_c6x, cpu_data);
cpu_data           97 arch/c6x/kernel/setup.c 	p = &per_cpu(cpu_data, smp_processor_id());
cpu_data          412 arch/c6x/kernel/setup.c 	struct cpuinfo_c6x *p = &per_cpu(cpu_data, n);
cpu_data           23 arch/csky/include/asm/processor.h extern struct cpuinfo_csky cpu_data[];
cpu_data           17 arch/csky/kernel/process.c struct cpuinfo_csky cpu_data[NR_CPUS];
cpu_data           42 arch/ia64/include/asm/topology.h #define topology_physical_package_id(cpu)	(cpu_data(cpu)->socket_id)
cpu_data           43 arch/ia64/include/asm/topology.h #define topology_core_id(cpu)			(cpu_data(cpu)->core_id)
cpu_data         1835 arch/ia64/kernel/mca.c ia64_mca_cpu_init(void *cpu_data)
cpu_data         1871 arch/ia64/kernel/mca.c 		pte_val(mk_pte_phys(__pa(cpu_data), PAGE_KERNEL)));
cpu_data          747 arch/ia64/kernel/setup.c 	return *pos < nr_cpu_ids ? cpu_data(*pos) : NULL;
cpu_data          948 arch/ia64/kernel/setup.c 	void *cpu_data;
cpu_data          950 arch/ia64/kernel/setup.c 	cpu_data = per_cpu_init();
cpu_data          968 arch/ia64/kernel/setup.c 			    ia64_tpa(cpu_data) - (long) __per_cpu_start);
cpu_data          980 arch/ia64/kernel/setup.c 	cpu_info = cpu_data + ((char *) &__ia64_per_cpu_var(ia64_cpu_info) - __per_cpu_start);
cpu_data         1026 arch/ia64/kernel/setup.c 	ia64_mmu_init(ia64_imva(cpu_data));
cpu_data         1027 arch/ia64/kernel/setup.c 	ia64_mca_cpu_init(ia64_imva(cpu_data));
cpu_data          420 arch/ia64/kernel/smpboot.c 	last_cpuinfo = cpu_data(cpuid - 1);
cpu_data          591 arch/ia64/kernel/smpboot.c 	if (cpu_data(cpu)->threads_per_core == 1 &&
cpu_data          592 arch/ia64/kernel/smpboot.c 	    cpu_data(cpu)->cores_per_socket == 1) {
cpu_data          701 arch/ia64/kernel/smpboot.c 		bogosum += cpu_data(cpu)->loops_per_jiffy;
cpu_data          713 arch/ia64/kernel/smpboot.c 		if ((cpu_data(cpu)->socket_id == cpu_data(i)->socket_id)) {
cpu_data          716 arch/ia64/kernel/smpboot.c 			if (cpu_data(cpu)->core_id == cpu_data(i)->core_id) {
cpu_data          749 arch/ia64/kernel/smpboot.c 	if (cpu_data(cpu)->threads_per_core == 1 &&
cpu_data          750 arch/ia64/kernel/smpboot.c 	    cpu_data(cpu)->cores_per_socket == 1) {
cpu_data          845 arch/ia64/kernel/smpboot.c 			if ((cpu_data(j)->socket_id == cpu_data(i)->socket_id)) {
cpu_data          846 arch/ia64/kernel/smpboot.c 				if (cpu_data(j)->core_id == cpu_data(i)->core_id)
cpu_data           35 arch/ia64/kernel/topology.c 	if (cpu_data(num)->socket_id == -1)
cpu_data           36 arch/ia64/kernel/topology.c 		cpu_data(num)->socket_id = slot;
cpu_data          145 arch/ia64/kernel/topology.c 	if (cpu_data(cpu)->threads_per_core <= 1 &&
cpu_data          146 arch/ia64/kernel/topology.c 		cpu_data(cpu)->cores_per_socket <= 1) {
cpu_data          160 arch/ia64/kernel/topology.c 			if (cpu_data(cpu)->socket_id == cpu_data(j)->socket_id
cpu_data          161 arch/ia64/kernel/topology.c 				&& cpu_data(j)->core_id == csi.log1_cid
cpu_data          162 arch/ia64/kernel/topology.c 				&& cpu_data(j)->thread_id == csi.log1_tid)
cpu_data           37 arch/ia64/mm/contig.c static void *cpu_data;
cpu_data           61 arch/ia64/mm/contig.c 		memcpy(cpu_data, src, __per_cpu_end - __per_cpu_start);
cpu_data           62 arch/ia64/mm/contig.c 		__per_cpu_offset[cpu] = (char *)cpu_data - __per_cpu_start;
cpu_data           75 arch/ia64/mm/contig.c 			ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) -
cpu_data           78 arch/ia64/mm/contig.c 		cpu_data += PERCPU_PAGE_SIZE;
cpu_data           89 arch/ia64/mm/contig.c 	cpu_data = memblock_alloc_from(size, PERCPU_PAGE_SIZE,
cpu_data           91 arch/ia64/mm/contig.c 	if (!cpu_data)
cpu_data          136 arch/ia64/mm/discontig.c static void *per_cpu_node_setup(void *cpu_data, int node)
cpu_data          147 arch/ia64/mm/discontig.c 		memcpy(__va(cpu_data), src, __per_cpu_end - __per_cpu_start);
cpu_data          148 arch/ia64/mm/discontig.c 		__per_cpu_offset[cpu] = (char *)__va(cpu_data) -
cpu_data          162 arch/ia64/mm/discontig.c 				    (unsigned long)cpu_data -
cpu_data          165 arch/ia64/mm/discontig.c 		cpu_data += PERCPU_PAGE_SIZE;
cpu_data          168 arch/ia64/mm/discontig.c 	return cpu_data;
cpu_data          262 arch/ia64/mm/discontig.c 	void *cpu_data;
cpu_data          269 arch/ia64/mm/discontig.c 	cpu_data = (void *)pernode;
cpu_data          280 arch/ia64/mm/discontig.c 	cpu_data = per_cpu_node_setup(cpu_data, node);
cpu_data           38 arch/mips/include/asm/bugs.h 	cpu_data[cpu].udelay_val = loops_per_jiffy;
cpu_data           17 arch/mips/include/asm/cpu-features.h #define __ase(ase)			(cpu_data[0].ases & (ase))
cpu_data           18 arch/mips/include/asm/cpu-features.h #define __isa(isa)			(cpu_data[0].isa_level & (isa))
cpu_data           19 arch/mips/include/asm/cpu-features.h #define __opt(opt)			(cpu_data[0].options & (opt))
cpu_data          235 arch/mips/include/asm/cpu-features.h #define cpu_has_vtag_icache	(cpu_data[0].icache.flags & MIPS_CACHE_VTAG)
cpu_data          238 arch/mips/include/asm/cpu-features.h #define cpu_has_dc_aliases	(cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES)
cpu_data          241 arch/mips/include/asm/cpu-features.h #define cpu_has_ic_fills_f_dc	(cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC)
cpu_data          244 arch/mips/include/asm/cpu-features.h #define cpu_has_pindexed_dcache	(cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX)
cpu_data          260 arch/mips/include/asm/cpu-features.h #define cpu_icache_snoops_remote_store	(cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE)
cpu_data          430 arch/mips/include/asm/cpu-features.h # define cpu_has_64bits		(cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
cpu_data          433 arch/mips/include/asm/cpu-features.h # define cpu_has_64bit_zero_reg	(cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
cpu_data          463 arch/mips/include/asm/cpu-features.h # define cpu_vmbits cpu_data[0].vmbits
cpu_data          485 arch/mips/include/asm/cpu-features.h #define cpu_dcache_line_size()	cpu_data[0].dcache.linesz
cpu_data          488 arch/mips/include/asm/cpu-features.h #define cpu_icache_line_size()	cpu_data[0].icache.linesz
cpu_data          491 arch/mips/include/asm/cpu-features.h #define cpu_scache_line_size()	cpu_data[0].scache.linesz
cpu_data          494 arch/mips/include/asm/cpu-features.h #define cpu_tcache_line_size()	cpu_data[0].tcache.linesz
cpu_data          623 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf1	(cpu_data[0].guest.conf & (1 << 1))
cpu_data          626 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf2	(cpu_data[0].guest.conf & (1 << 2))
cpu_data          629 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf3	(cpu_data[0].guest.conf & (1 << 3))
cpu_data          632 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf4	(cpu_data[0].guest.conf & (1 << 4))
cpu_data          635 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf5	(cpu_data[0].guest.conf & (1 << 5))
cpu_data          638 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf6	(cpu_data[0].guest.conf & (1 << 6))
cpu_data          641 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_conf7	(cpu_data[0].guest.conf & (1 << 7))
cpu_data          644 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_fpu	(cpu_data[0].guest.options & MIPS_CPU_FPU)
cpu_data          647 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_watch	(cpu_data[0].guest.options & MIPS_CPU_WATCH)
cpu_data          650 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC)
cpu_data          653 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_segments	(cpu_data[0].guest.options & MIPS_CPU_SEGMENTS)
cpu_data          656 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_badinstr	(cpu_data[0].guest.options & MIPS_CPU_BADINSTR)
cpu_data          659 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_badinstrp	(cpu_data[0].guest.options & MIPS_CPU_BADINSTRP)
cpu_data          662 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_htw	(cpu_data[0].guest.options & MIPS_CPU_HTW)
cpu_data          665 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_mvh	(cpu_data[0].guest.options & MIPS_CPU_MVH)
cpu_data          668 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_msa	(cpu_data[0].guest.ases & MIPS_ASE_MSA)
cpu_data          671 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_kscr(n)	(cpu_data[0].guest.kscratch_mask & (1u << (n)))
cpu_data          674 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_rw_llb	(cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB))
cpu_data          677 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_perf	(cpu_data[0].guest.options & MIPS_CPU_PERF)
cpu_data          680 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_maar	(cpu_data[0].guest.options & MIPS_CPU_MAAR)
cpu_data          683 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_userlocal	(cpu_data[0].guest.options & MIPS_CPU_ULRI)
cpu_data          690 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_fpu	(cpu_data[0].guest.options_dyn & MIPS_CPU_FPU)
cpu_data          693 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_watch	(cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH)
cpu_data          696 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC)
cpu_data          699 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_perf	(cpu_data[0].guest.options_dyn & MIPS_CPU_PERF)
cpu_data          702 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_msa	(cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA)
cpu_data          705 arch/mips/include/asm/cpu-features.h #define cpu_guest_has_dyn_maar	(cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR)
cpu_data          110 arch/mips/include/asm/cpu-info.h extern struct cpuinfo_mips cpu_data[];
cpu_data          111 arch/mips/include/asm/cpu-info.h #define current_cpu_data cpu_data[smp_processor_id()]
cpu_data          112 arch/mips/include/asm/cpu-info.h #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
cpu_data          113 arch/mips/include/asm/cpu-info.h #define boot_cpu_data cpu_data[0]
cpu_data          174 arch/mips/include/asm/cpu-info.h 	struct cpuinfo_mips *infoa = &cpu_data[cpua];
cpu_data          175 arch/mips/include/asm/cpu-info.h 	struct cpuinfo_mips *infob = &cpu_data[cpub];
cpu_data          212 arch/mips/include/asm/cpu-type.h 	const int cpu_type = cpu_data[0].cputype;
cpu_data           60 arch/mips/include/asm/mach-cavium-octeon/cpu-feature-overrides.h #define cpu_has_rixi		(cpu_data[0].cputype != CPU_CAVIUM_OCTEON)
cpu_data          398 arch/mips/include/asm/mips-cm.h 	unsigned int core = cpu_core(&cpu_data[cpu]);
cpu_data          399 arch/mips/include/asm/mips-cm.h 	unsigned int vp = cpu_vpe_id(&cpu_data[cpu]);
cpu_data          454 arch/mips/include/asm/mips-cm.h 	struct cpuinfo_mips *d = &cpu_data[cpu];
cpu_data           96 arch/mips/include/asm/mmu_context.h 	unsigned long asid_mask = cpu_asid_mask(&cpu_data[cpu]);
cpu_data          123 arch/mips/include/asm/mmu_context.h #define asid_cache(cpu)		(cpu_data[cpu].asid_cache)
cpu_data          125 arch/mips/include/asm/mmu_context.h 	(cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu]))
cpu_data          216 arch/mips/include/asm/mmu_context.h 		write_c0_memorymapid(ctx & cpu_asid_mask(&cpu_data[cpu]));
cpu_data          436 arch/mips/include/asm/pgtable.h 	prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine;
cpu_data           60 arch/mips/include/asm/processor.h #define TASK_SIZE64     (0x1UL << ((cpu_data[0].vmbits>48)?48:cpu_data[0].vmbits))
cpu_data           64 arch/mips/include/asm/timex.h 		asm volatile("" : "=m" (cpu_data[0].options));
cpu_data           15 arch/mips/include/asm/topology.h #define topology_physical_package_id(cpu)	(cpu_data[cpu].package)
cpu_data           16 arch/mips/include/asm/topology.h #define topology_core_id(cpu)			(cpu_core(&cpu_data[cpu]))
cpu_data           65 arch/mips/kernel/cacheinfo.c 	int cluster = cpu_cluster(&cpu_data[cpu]);
cpu_data           68 arch/mips/kernel/cacheinfo.c 		if (cpu_cluster(&cpu_data[cpu1]) == cluster)
cpu_data          357 arch/mips/kernel/cpu-probe.c 	cpu_data[0].ases &= ~(MIPS_ASE_DSP | MIPS_ASE_DSP2P);
cpu_data          370 arch/mips/kernel/cpu-probe.c 	cpu_data[0].options &= ~MIPS_CPU_HTW;
cpu_data          401 arch/mips/kernel/cpu-probe.c 	if (set_ftlb_enable(&cpu_data[0], 0)) {
cpu_data          431 arch/mips/kernel/cpu-probe.c 	cpu_data[0].tlbsize -= cpu_data[0].tlbsizeftlbways *
cpu_data          432 arch/mips/kernel/cpu-probe.c 			       cpu_data[0].tlbsizeftlbsets;
cpu_data          433 arch/mips/kernel/cpu-probe.c 	cpu_data[0].tlbsizeftlbsets = 0;
cpu_data          434 arch/mips/kernel/cpu-probe.c 	cpu_data[0].tlbsizeftlbways = 0;
cpu_data          346 arch/mips/kernel/perf_event_mipsxx.c 		ctrl = M_PERFCTL_VPEID(cpu_vpe_id(&cpu_data[cpu]));
cpu_data          447 arch/mips/kernel/pm-cps.c 				vpe_id = cpu_vpe_id(&cpu_data[cpu]);
cpu_data          468 arch/mips/kernel/pm-cps.c 	cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache,
cpu_data          472 arch/mips/kernel/pm-cps.c 	cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache,
cpu_data          485 arch/mips/kernel/pm-cps.c 		uasm_i_addiu(&p, t0, zero, 1 << cpu_core(&cpu_data[cpu]));
cpu_data          499 arch/mips/kernel/pm-cps.c 		err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu],
cpu_data          639 arch/mips/kernel/pm-cps.c 	unsigned core = cpu_core(&cpu_data[cpu]);
cpu_data           40 arch/mips/kernel/proc.c 	unsigned int version = cpu_data[n].processor_id;
cpu_data           41 arch/mips/kernel/proc.c 	unsigned int fp_vers = cpu_data[n].fpu_id;
cpu_data           62 arch/mips/kernel/proc.c 		      cpu_data[n].options & MIPS_CPU_FPU ? "  FPU V%d.%d" : "");
cpu_data           67 arch/mips/kernel/proc.c 		      cpu_data[n].udelay_val / (500000/HZ),
cpu_data           68 arch/mips/kernel/proc.c 		      (cpu_data[n].udelay_val / (5000/HZ)) % 100);
cpu_data           72 arch/mips/kernel/proc.c 	seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize);
cpu_data           79 arch/mips/kernel/proc.c 		      cpu_data[n].watch_reg_count);
cpu_data           80 arch/mips/kernel/proc.c 		for (i = 0; i < cpu_data[n].watch_reg_count; i++)
cpu_data           82 arch/mips/kernel/proc.c 				cpu_data[n].watch_reg_masks[i]);
cpu_data          138 arch/mips/kernel/proc.c 		      cpu_data[n].srsets);
cpu_data          140 arch/mips/kernel/proc.c 		      hweight8(cpu_data[n].kscratch_mask));
cpu_data          141 arch/mips/kernel/proc.c 	seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package);
cpu_data          142 arch/mips/kernel/proc.c 	seq_printf(m, "core\t\t\t: %d\n", cpu_core(&cpu_data[n]));
cpu_data          146 arch/mips/kernel/proc.c 		seq_printf(m, "VPE\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n]));
cpu_data          148 arch/mips/kernel/proc.c 		seq_printf(m, "VP\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n]));
cpu_data           49 arch/mips/kernel/setup.c struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
cpu_data           51 arch/mips/kernel/setup.c EXPORT_SYMBOL(cpu_data);
cpu_data           74 arch/mips/kernel/smp-cps.c 				cpu_set_cluster(&cpu_data[nvpes + v], cl);
cpu_data           75 arch/mips/kernel/smp-cps.c 				cpu_set_core(&cpu_data[nvpes + v], c);
cpu_data           76 arch/mips/kernel/smp-cps.c 				cpu_set_vpe_id(&cpu_data[nvpes + v], v);
cpu_data           88 arch/mips/kernel/smp-cps.c 		set_cpu_possible(v, cpu_cluster(&cpu_data[v]) == 0);
cpu_data           89 arch/mips/kernel/smp-cps.c 		set_cpu_present(v, cpu_cluster(&cpu_data[v]) == 0);
cpu_data          296 arch/mips/kernel/smp-cps.c 	unsigned core = cpu_core(&cpu_data[cpu]);
cpu_data          297 arch/mips/kernel/smp-cps.c 	unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
cpu_data          305 arch/mips/kernel/smp-cps.c 	if (cpu_cluster(&cpu_data[cpu]) != cpu_cluster(&raw_current_cpu_data))
cpu_data          312 arch/mips/kernel/smp-cps.c 	atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask);
cpu_data          409 arch/mips/kernel/smp-cps.c 	core = cpu_core(&cpu_data[cpu]);
cpu_data          412 arch/mips/kernel/smp-cps.c 		vpe_id = cpu_vpe_id(&cpu_data[cpu]);
cpu_data          509 arch/mips/kernel/smp-cps.c 	unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
cpu_data          523 arch/mips/kernel/smp-cps.c 	unsigned core = cpu_core(&cpu_data[cpu]);
cpu_data          524 arch/mips/kernel/smp-cps.c 	unsigned int vpe_id = cpu_vpe_id(&cpu_data[cpu]);
cpu_data           75 arch/mips/kernel/smp-mt.c 	cpu_set_vpe_id(&cpu_data[ncpu], tc);
cpu_data          105 arch/mips/kernel/smp.c 		if (cpu_data[cpu].package == cpu_data[i].package) {
cpu_data          180 arch/mips/kernel/smp.c 			core = cpu_core(&cpu_data[cpu]);
cpu_data          362 arch/mips/kernel/smp.c 	cpu_data[cpu].udelay_val = loops_per_jiffy;
cpu_data         2213 arch/mips/kernel/traps.c 		cpu_data[cpu].asid_cache = 0;
cpu_data         2214 arch/mips/kernel/traps.c 	else if (!cpu_data[cpu].asid_cache)
cpu_data         2215 arch/mips/kernel/traps.c 		cpu_data[cpu].asid_cache = asid_first_version(cpu);
cpu_data          129 arch/mips/kvm/entry.c 	unsigned int kscratch_mask = cpu_data[0].kscratch_mask;
cpu_data          398 arch/mips/kvm/entry.c 	UASM_i_LA_mostly(&p, AT, (long)&cpu_data[0].asid_mask);
cpu_data          400 arch/mips/kvm/entry.c 	UASM_i_LW(&p, T2, uasm_rel_lo((long)&cpu_data[0].asid_mask), AT);
cpu_data         1713 arch/mips/kvm/vz.c 	ret += __arch_hweight8(cpu_data[0].guest.kscratch_mask);
cpu_data         2312 arch/mips/kvm/vz.c #define guestid_cache(cpu)	(cpu_data[cpu].guestid_cache)
cpu_data           28 arch/mips/loongson64/loongson-3/irq.c 		if (cpu_data[cpu].package > 0)
cpu_data          313 arch/mips/loongson64/loongson-3/smp.c 	cpu_set_core(&cpu_data[cpu],
cpu_data          315 arch/mips/loongson64/loongson-3/smp.c 	cpu_data[cpu].package =
cpu_data          328 arch/mips/loongson64/loongson-3/smp.c 	if (cpu_data[cpu].package)
cpu_data          380 arch/mips/loongson64/loongson-3/smp.c 	cpu_set_core(&cpu_data[0],
cpu_data          382 arch/mips/loongson64/loongson-3/smp.c 	cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package;
cpu_data          481 arch/mips/loongson64/loongson-3/smp.c 		  [sets] "r" (cpu_data[smp_processor_id()].dcache.sets));
cpu_data          563 arch/mips/loongson64/loongson-3/smp.c 		  [sets] "r" (cpu_data[smp_processor_id()].dcache.sets),
cpu_data          564 arch/mips/loongson64/loongson-3/smp.c 		  [vsets] "r" (cpu_data[smp_processor_id()].vcache.sets));
cpu_data          626 arch/mips/loongson64/loongson-3/smp.c 		  [sets] "r" (cpu_data[smp_processor_id()].dcache.sets));
cpu_data          695 arch/mips/loongson64/loongson-3/smp.c 	uint64_t core_id = cpu_core(&cpu_data[cpu]);
cpu_data          696 arch/mips/loongson64/loongson-3/smp.c 	uint64_t package_id = cpu_data[cpu].package;
cpu_data          709 arch/mips/loongson64/loongson-3/smp.c 	uint64_t core_id = cpu_core(&cpu_data[cpu]);
cpu_data          710 arch/mips/loongson64/loongson-3/smp.c 	uint64_t package_id = cpu_data[cpu].package;
cpu_data           36 arch/mips/mm/context.c 	if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) {
cpu_data           76 arch/mips/mm/context.c 		mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0);
cpu_data           88 arch/mips/mm/context.c 		__set_bit(mmid & cpu_asid_mask(&cpu_data[cpu]), mmid_map);
cpu_data          210 arch/mips/mm/context.c 	old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache);
cpu_data          213 arch/mips/mm/context.c 	    !cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) {
cpu_data          220 arch/mips/mm/context.c 		WRITE_ONCE(cpu_data[cpu].asid_cache, ctx);
cpu_data          336 arch/mips/mm/tlbex.c 	unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask;
cpu_data          104 arch/parisc/include/asm/processor.h DECLARE_PER_CPU(struct cpuinfo_parisc, cpu_data);
cpu_data          341 arch/parisc/kernel/irq.c 	return per_cpu(cpu_data, cpu).txn_addr;
cpu_data          353 arch/parisc/kernel/irq.c 		(!per_cpu(cpu_data, next_cpu).txn_addr ||
cpu_data          541 arch/parisc/kernel/irq.c 			   per_cpu(cpu_data, cpu).hpa);
cpu_data          524 arch/parisc/kernel/perf.c 	cpu_device = per_cpu(cpu_data, 0).dev;
cpu_data          526 arch/parisc/kernel/perf.c 		per_cpu(cpu_data, 0).dev->name);
cpu_data           39 arch/parisc/kernel/processor.c DEFINE_PER_CPU(struct cpuinfo_parisc, cpu_data);
cpu_data          159 arch/parisc/kernel/processor.c 	p = &per_cpu(cpu_data, cpuid);
cpu_data          337 arch/parisc/kernel/processor.c 		per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision;
cpu_data          338 arch/parisc/kernel/processor.c 		per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model;
cpu_data          378 arch/parisc/kernel/processor.c 		const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
cpu_data          405 arch/parisc/kernel/setup.c 		per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision;
cpu_data          406 arch/parisc/kernel/setup.c 		per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model;
cpu_data          122 arch/parisc/kernel/smp.c 	struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu);
cpu_data          190 arch/parisc/kernel/smp.c 	struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu);
cpu_data          322 arch/parisc/kernel/smp.c 	const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid);
cpu_data          380 arch/parisc/kernel/smp.c 	int bootstrap_processor = per_cpu(cpu_data, 0).cpuid;
cpu_data           68 arch/parisc/kernel/time.c 	struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
cpu_data          163 arch/parisc/kernel/time.c 	per_cpu(cpu_data, cpu).it_value = next_tick;
cpu_data          255 arch/parisc/kernel/time.c 		cpu0_loc = per_cpu(cpu_data, 0).cpu_loc;
cpu_data          261 arch/parisc/kernel/time.c 			    (cpu0_loc == per_cpu(cpu_data, cpu).cpu_loc))
cpu_data           78 arch/parisc/kernel/topology.c 	p = &per_cpu(cpu_data, cpuid);
cpu_data           80 arch/parisc/kernel/topology.c 		const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
cpu_data          161 arch/s390/pci/pci_irq.c 	struct cpu_irq_data *cpu_data;
cpu_data          177 arch/s390/pci/pci_irq.c 		cpu_data = &per_cpu(irq_data, cpu);
cpu_data          178 arch/s390/pci/pci_irq.c 		if (atomic_inc_return(&cpu_data->scheduled) > 1)
cpu_data          181 arch/s390/pci/pci_irq.c 		cpu_data->csd.func = zpci_handle_remote_irq;
cpu_data          182 arch/s390/pci/pci_irq.c 		cpu_data->csd.info = &cpu_data->scheduled;
cpu_data          183 arch/s390/pci/pci_irq.c 		cpu_data->csd.flags = 0;
cpu_data          184 arch/s390/pci/pci_irq.c 		smp_call_function_single_async(cpu, &cpu_data->csd);
cpu_data           38 arch/sh/include/asm/mmu_context.h #define asid_cache(cpu)		(cpu_data[cpu].asid_cache)
cpu_data           94 arch/sh/include/asm/processor.h extern struct sh_cpuinfo cpu_data[];
cpu_data           95 arch/sh/include/asm/processor.h #define boot_cpu_data cpu_data[0]
cpu_data           96 arch/sh/include/asm/processor.h #define current_cpu_data cpu_data[smp_processor_id()]
cpu_data           97 arch/sh/include/asm/processor.h #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
cpu_data           26 arch/sh/include/asm/tlb_64.h 	for (tlb  = cpu_data->dtlb.first;	\
cpu_data           27 arch/sh/include/asm/tlb_64.h 	     tlb <= cpu_data->dtlb.last;	\
cpu_data           28 arch/sh/include/asm/tlb_64.h 	     tlb += cpu_data->dtlb.step)
cpu_data           36 arch/sh/include/asm/tlb_64.h 	for (tlb  = cpu_data->itlb.first;	\
cpu_data           37 arch/sh/include/asm/tlb_64.h 	     tlb <= cpu_data->itlb.last;	\
cpu_data           38 arch/sh/include/asm/tlb_64.h 	     tlb += cpu_data->itlb.step)
cpu_data           83 arch/sh/kernel/cpu/proc.c 	unsigned int cpu = c - cpu_data;
cpu_data          136 arch/sh/kernel/cpu/proc.c 	return *pos < NR_CPUS ? cpu_data + *pos : NULL;
cpu_data           53 arch/sh/kernel/setup.c struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
cpu_data           61 arch/sh/kernel/setup.c EXPORT_SYMBOL(cpu_data);
cpu_data           49 arch/sh/kernel/smp.c 	struct sh_cpuinfo *c = cpu_data + cpu;
cpu_data          253 arch/sh/kernel/smp.c 		bogosum += cpu_data[cpu].loops_per_jiffy;
cpu_data           40 arch/sh/lib/delay.c 		  "r" (cpu_data[raw_smp_processor_id()].loops_per_jiffy * (HZ/4))
cpu_data           38 arch/sh/lib64/udelay.c 	__delay(xloops * (HZ * cpu_data[raw_smp_processor_id()].loops_per_jiffy));
cpu_data           74 arch/sh/mm/cache-sh4.c 				cpu_data->icache.entry_mask);
cpu_data           78 arch/sh/mm/cache-sh4.c 		for (i = 0; i < cpu_data->icache.ways; i++) {
cpu_data           81 arch/sh/mm/cache-sh4.c 			icacheaddr += cpu_data->icache.way_incr;
cpu_data          249 arch/sh/mm/cache-sh5.c 				 cpu_data->dcache.entry_mask) >>
cpu_data          250 arch/sh/mm/cache-sh5.c 				 cpu_data->dcache.entry_shift;
cpu_data          254 arch/sh/mm/cache-sh5.c 		set_offset &= (cpu_data->dcache.sets - 1);
cpu_data          256 arch/sh/mm/cache-sh5.c 			(set_offset << cpu_data->dcache.entry_shift);
cpu_data          265 arch/sh/mm/cache-sh5.c 		eaddr1 = eaddr0 + cpu_data->dcache.way_size *
cpu_data          266 arch/sh/mm/cache-sh5.c 				  cpu_data->dcache.ways;
cpu_data          269 arch/sh/mm/cache-sh5.c 		     eaddr += cpu_data->dcache.way_size) {
cpu_data          274 arch/sh/mm/cache-sh5.c 		eaddr1 = eaddr0 + cpu_data->dcache.way_size *
cpu_data          275 arch/sh/mm/cache-sh5.c 				  cpu_data->dcache.ways;
cpu_data          278 arch/sh/mm/cache-sh5.c 		     eaddr += cpu_data->dcache.way_size) {
cpu_data          283 arch/sh/mm/cache-sh5.c 			if (test_bit(SH_CACHE_MODE_WT, &(cpu_data->dcache.flags)))
cpu_data          306 arch/sh/mm/cache-sh5.c 	sh64_dcache_purge_sets(0, cpu_data->dcache.sets);
cpu_data           44 arch/sh/mm/tlb-sh4.c 	if (cpu_data->flags & CPU_HAS_PTEA) {
cpu_data           23 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.entries	= 64;
cpu_data           24 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.step	= 0x10;
cpu_data           26 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.first	= DTLB_FIXED | cpu_data->dtlb.step;
cpu_data           27 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next	= cpu_data->dtlb.first;
cpu_data           29 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.last	= DTLB_FIXED |
cpu_data           30 arch/sh/mm/tlb-sh5.c 				  ((cpu_data->dtlb.entries - 1) *
cpu_data           31 arch/sh/mm/tlb-sh5.c 				   cpu_data->dtlb.step);
cpu_data           34 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.entries	= 64;
cpu_data           35 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.step	= 0x10;
cpu_data           37 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.first	= ITLB_FIXED | cpu_data->itlb.step;
cpu_data           38 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.next	= cpu_data->itlb.first;
cpu_data           39 arch/sh/mm/tlb-sh5.c 	cpu_data->itlb.last	= ITLB_FIXED |
cpu_data           40 arch/sh/mm/tlb-sh5.c 				  ((cpu_data->itlb.entries - 1) *
cpu_data           41 arch/sh/mm/tlb-sh5.c 				   cpu_data->itlb.step);
cpu_data           51 arch/sh/mm/tlb-sh5.c 	return cpu_data->dtlb.next;
cpu_data           61 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.first += cpu_data->dtlb.step;
cpu_data           62 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next  += cpu_data->dtlb.step;
cpu_data           99 arch/sh/mm/tlb-sh5.c 	if (entry < (cpu_data->dtlb.first - cpu_data->dtlb.step))
cpu_data          103 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.first	-= cpu_data->dtlb.step;
cpu_data          104 arch/sh/mm/tlb-sh5.c 	cpu_data->dtlb.next	= entry;
cpu_data          210 arch/sh/mm/tlb-sh5.c 		tlbp = &cpu_data->itlb;
cpu_data          212 arch/sh/mm/tlb-sh5.c 		tlbp = &cpu_data->dtlb;
cpu_data           16 arch/sparc/include/asm/bugs.h 	cpu_data(0).udelay_val = loops_per_jiffy;
cpu_data           28 arch/sparc/include/asm/delay_32.h #define __udelay_val	cpu_data(smp_processor_id()).udelay_val
cpu_data           46 arch/sparc/include/asm/topology_64.h #define topology_physical_package_id(cpu)	(cpu_data(cpu).proc_id)
cpu_data           47 arch/sparc/include/asm/topology_64.h #define topology_core_id(cpu)			(cpu_data(cpu).core_id)
cpu_data          348 arch/sparc/kernel/cpu.c 		   , cpu_data(0).udelay_val/(500000/HZ),
cpu_data          349 arch/sparc/kernel/cpu.c 		   (cpu_data(0).udelay_val/(5000/HZ)) % 100,
cpu_data          350 arch/sparc/kernel/cpu.c 		   cpu_data(0).clock_tick
cpu_data          398 arch/sparc/kernel/cpu.c 		   , cpu_data(0).clock_tick
cpu_data          108 arch/sparc/kernel/cpumap.c 		id = cpu_data(cpu).core_id;
cpu_data          111 arch/sparc/kernel/cpumap.c 		id = cpu_data(cpu).proc_id;
cpu_data          130 arch/sparc/kernel/devices.c 		cpu_data(0).clock_tick = prom_getintdefault(cpu_node,
cpu_data          204 arch/sparc/kernel/irq_32.c 		seq_printf(p, "%10u ", cpu_data(j).irq_resched_count);
cpu_data          208 arch/sparc/kernel/irq_32.c 		seq_printf(p, "%10u ", cpu_data(j).irq_call_count);
cpu_data          213 arch/sparc/kernel/irq_32.c 		seq_printf(p, "%10u ", cpu_data(j).counter);
cpu_data          308 arch/sparc/kernel/irq_64.c 		seq_printf(p, "%10u ", cpu_data(j).__nmi_count);
cpu_data          242 arch/sparc/kernel/leon_smp.c 			prev = &cpu_data(i).next;
cpu_data          888 arch/sparc/kernel/mdesc.c 		cpu_data(*id).core_id = core_id;
cpu_data          897 arch/sparc/kernel/mdesc.c 		cpu_data(*id).max_cache_id = max_cache_id;
cpu_data          903 arch/sparc/kernel/mdesc.c 		cpu_data(*id).sock_id = max_cache_id;
cpu_data          989 arch/sparc/kernel/mdesc.c 				cpu_data(*id).sock_id = idx;
cpu_data         1027 arch/sparc/kernel/mdesc.c 			cpu_data(*id).proc_id = proc_id;
cpu_data         1197 arch/sparc/kernel/mdesc.c 	c = &cpu_data(cpuid);
cpu_data          127 arch/sparc/kernel/nmi.c 	return cpu_data(cpu).__nmi_count;
cpu_data          416 arch/sparc/kernel/prom_64.c 			int proc_id = cpu_data(cpu).proc_id;
cpu_data          530 arch/sparc/kernel/prom_64.c 	cpu_data(cpuid).clock_tick =
cpu_data          534 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).dcache_size =
cpu_data          537 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).dcache_line_size =
cpu_data          540 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).icache_size =
cpu_data          543 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).icache_line_size =
cpu_data          546 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).ecache_size =
cpu_data          548 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).ecache_line_size =
cpu_data          550 arch/sparc/kernel/prom_64.c 		if (!cpu_data(cpuid).ecache_size ||
cpu_data          551 arch/sparc/kernel/prom_64.c 		    !cpu_data(cpuid).ecache_line_size) {
cpu_data          552 arch/sparc/kernel/prom_64.c 			cpu_data(cpuid).ecache_size =
cpu_data          556 arch/sparc/kernel/prom_64.c 			cpu_data(cpuid).ecache_line_size =
cpu_data          561 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).core_id = portid + 1;
cpu_data          562 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).proc_id = portid;
cpu_data          564 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).dcache_size =
cpu_data          566 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).dcache_line_size =
cpu_data          569 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).icache_size =
cpu_data          571 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).icache_line_size =
cpu_data          574 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).ecache_size =
cpu_data          577 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).ecache_line_size =
cpu_data          580 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).core_id = 0;
cpu_data          581 arch/sparc/kernel/prom_64.c 		cpu_data(cpuid).proc_id = -1;
cpu_data           63 arch/sparc/kernel/smp_32.c 	cpu_data(id).udelay_val = loops_per_jiffy;
cpu_data           66 arch/sparc/kernel/smp_32.c 	cpu_data(id).clock_tick = prom_getintdefault(cpu_node,
cpu_data           68 arch/sparc/kernel/smp_32.c 	cpu_data(id).prom_node = cpu_node;
cpu_data           75 arch/sparc/kernel/smp_32.c 	cpu_data(id).mid = mid;
cpu_data           85 arch/sparc/kernel/smp_32.c 		bogosum += cpu_data(cpu).udelay_val;
cpu_data          385 arch/sparc/kernel/smp_32.c 			   cpu_data(i).udelay_val/(500000/HZ),
cpu_data          386 arch/sparc/kernel/smp_32.c 			   (cpu_data(i).udelay_val/(5000/HZ))%100);
cpu_data           97 arch/sparc/kernel/smp_64.c 			   i, cpu_data(i).clock_tick);
cpu_data         1275 arch/sparc/kernel/smp_64.c 		if (cpu_data(i).core_id == 0) {
cpu_data         1281 arch/sparc/kernel/smp_64.c 			if (cpu_data(i).core_id ==
cpu_data         1282 arch/sparc/kernel/smp_64.c 			    cpu_data(j).core_id)
cpu_data         1291 arch/sparc/kernel/smp_64.c 			if (cpu_data(i).max_cache_id ==
cpu_data         1292 arch/sparc/kernel/smp_64.c 			    cpu_data(j).max_cache_id)
cpu_data         1295 arch/sparc/kernel/smp_64.c 			if (cpu_data(i).sock_id == cpu_data(j).sock_id)
cpu_data         1304 arch/sparc/kernel/smp_64.c 		if (cpu_data(i).proc_id == -1) {
cpu_data         1310 arch/sparc/kernel/smp_64.c 			if (cpu_data(i).proc_id ==
cpu_data         1311 arch/sparc/kernel/smp_64.c 			    cpu_data(j).proc_id)
cpu_data         1388 arch/sparc/kernel/smp_64.c 	c = &cpu_data(cpu);
cpu_data          173 arch/sparc/kernel/sun4d_smp.c 		prev = &cpu_data(i).next;
cpu_data          131 arch/sparc/kernel/sun4m_smp.c 		prev = &cpu_data(i).next;
cpu_data          176 arch/sparc/kernel/sysfs.c 	cpuinfo_sparc *c = &cpu_data(dev->id); \
cpu_data          184 arch/sparc/kernel/sysfs.c 	cpuinfo_sparc *c = &cpu_data(dev->id); \
cpu_data          646 arch/sparc/kernel/time_64.c 	return cpu_data(cpu).clock_tick;
cpu_data          664 arch/sparc/kernel/time_64.c 			ft->clock_tick_ref = cpu_data(cpu).clock_tick;
cpu_data          669 arch/sparc/kernel/time_64.c 			cpu_data(cpu).clock_tick =
cpu_data          869 arch/sparc/kernel/traps_64.c 		val = cpu_data(i).ecache_size;
cpu_data          876 arch/sparc/kernel/traps_64.c 		val = cpu_data(i).ecache_line_size;
cpu_data           80 arch/um/kernel/um_arch.c 	return *pos < NR_CPUS ? cpu_data + *pos : NULL;
cpu_data          210 arch/x86/events/amd/uncore.c 		int thread = 2 * (cpu_data(event->cpu).cpu_core_id % 4);
cpu_data          213 arch/x86/events/amd/uncore.c 			thread += cpu_data(event->cpu).apicid & 1;
cpu_data           91 arch/x86/include/asm/acpi.h 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data           98 arch/x86/include/asm/acpi.h 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          107 arch/x86/include/asm/topology.h #define topology_logical_package_id(cpu)	(cpu_data(cpu).logical_proc_id)
cpu_data          108 arch/x86/include/asm/topology.h #define topology_physical_package_id(cpu)	(cpu_data(cpu).phys_proc_id)
cpu_data          109 arch/x86/include/asm/topology.h #define topology_logical_die_id(cpu)		(cpu_data(cpu).logical_die_id)
cpu_data          110 arch/x86/include/asm/topology.h #define topology_die_id(cpu)			(cpu_data(cpu).cpu_die_id)
cpu_data          111 arch/x86/include/asm/topology.h #define topology_core_id(cpu)			(cpu_data(cpu).cpu_core_id)
cpu_data           32 arch/x86/kernel/acpi/cstate.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          149 arch/x86/kernel/acpi/cstate.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          555 arch/x86/kernel/alternative.c 		clear_cpu_cap(&cpu_data(0), X86_FEATURE_UP);
cpu_data          389 arch/x86/kernel/amd_nb.c 	return (mask >> (4 * cpu_data(cpu).cpu_core_id)) & 0xf;
cpu_data          415 arch/x86/kernel/amd_nb.c 	cuid = cpu_data(cpu).cpu_core_id;
cpu_data          905 arch/x86/kernel/cpu/cacheinfo.c 		apicid = cpu_data(cpu).apicid;
cpu_data          914 arch/x86/kernel/cpu/cacheinfo.c 			apicid = cpu_data(i).apicid;
cpu_data          921 arch/x86/kernel/cpu/cacheinfo.c 				apicid = cpu_data(sibling).apicid;
cpu_data          941 arch/x86/kernel/cpu/cacheinfo.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          959 arch/x86/kernel/cpu/cacheinfo.c 		if (cpu_data(i).apicid >> index_msb == c->apicid >> index_msb) {
cpu_data         1008 arch/x86/kernel/cpu/cacheinfo.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          141 arch/x86/kernel/cpu/mce/core.c 	m->socketid = cpu_data(m->extcpu).phys_proc_id;
cpu_data          142 arch/x86/kernel/cpu/mce/core.c 	m->apicid = cpu_data(m->extcpu).initial_apicid;
cpu_data          241 arch/x86/kernel/cpu/mce/therm_throt.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          649 arch/x86/kernel/cpu/microcode/amd.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          671 arch/x86/kernel/cpu/microcode/amd.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          892 arch/x86/kernel/cpu/microcode/amd.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          765 arch/x86/kernel/cpu/microcode/intel.c 	struct cpuinfo_x86 *c = &cpu_data(cpu_num);
cpu_data          793 arch/x86/kernel/cpu/microcode/intel.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          947 arch/x86/kernel/cpu/microcode/intel.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          971 arch/x86/kernel/cpu/microcode/intel.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          148 arch/x86/kernel/cpu/proc.c 		return &cpu_data(*pos);
cpu_data          111 arch/x86/kernel/cpuid.c 	c = &cpu_data(cpu);
cpu_data          167 arch/x86/kernel/msr.c 	c = &cpu_data(cpu);
cpu_data          193 arch/x86/kernel/smpboot.c 	cpu_data(cpuid).loops_per_jiffy = loops_per_jiffy;
cpu_data          302 arch/x86/kernel/smpboot.c 		struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          318 arch/x86/kernel/smpboot.c 	int proc_id = cpu_data(cur_cpu).phys_proc_id;
cpu_data          321 arch/x86/kernel/smpboot.c 		struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          351 arch/x86/kernel/smpboot.c 	cpu_data(cpu).logical_proc_id = new;
cpu_data          374 arch/x86/kernel/smpboot.c 	cpu_data(cpu).logical_die_id = new;
cpu_data          381 arch/x86/kernel/smpboot.c 	struct cpuinfo_x86 *c = &cpu_data(id);
cpu_data          396 arch/x86/kernel/smpboot.c 	struct cpuinfo_x86 *c = &cpu_data(id);
cpu_data          577 arch/x86/kernel/smpboot.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          593 arch/x86/kernel/smpboot.c 		o = &cpu_data(i);
cpu_data          608 arch/x86/kernel/smpboot.c 		o = &cpu_data(i);
cpu_data          630 arch/x86/kernel/smpboot.c 					cpu_data(i).booted_cores++;
cpu_data          632 arch/x86/kernel/smpboot.c 				c->booted_cores = cpu_data(i).booted_cores;
cpu_data          662 arch/x86/kernel/smpboot.c 			bogosum += cpu_data(cpu).loops_per_jiffy;
cpu_data         1299 arch/x86/kernel/smpboot.c 		c = &cpu_data(i);
cpu_data         1372 arch/x86/kernel/smpboot.c 	print_cpu_info(&cpu_data(0));
cpu_data         1414 arch/x86/kernel/smpboot.c 	ncpus = cpu_data(0).booted_cores * topology_max_smt_threads();
cpu_data         1543 arch/x86/kernel/smpboot.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data         1551 arch/x86/kernel/smpboot.c 			cpu_data(sibling).booted_cores--;
cpu_data          109 arch/x86/kernel/topology.c 	struct cpuinfo_x86 *c = &cpu_data(num);
cpu_data          903 arch/x86/kernel/tsc.c 	cpu_data(0).loops_per_jiffy = cpufreq_scale(cpu_data(0).loops_per_jiffy,
cpu_data         1525 arch/x86/kernel/tsc.c 	int constant_tsc = cpu_has(&cpu_data(cpu), X86_FEATURE_CONSTANT_TSC);
cpu_data         1533 arch/x86/kernel/tsc.c 		return cpu_data(sibling).loops_per_jiffy;
cpu_data          149 arch/x86/xen/apic.c 		return cpu_data(cpu).apicid;
cpu_data           73 arch/x86/xen/smp_pv.c 	cpu_data(cpu).x86_max_cores = 1;
cpu_data          251 arch/x86/xen/smp_pv.c 	cpu_data(0).x86_max_cores = 1;
cpu_data          407 arch/xtensa/kernel/setup.c static DEFINE_PER_CPU(struct cpu, cpu_data);
cpu_data          414 arch/xtensa/kernel/setup.c 		struct cpu *cpu = &per_cpu(cpu_data, i);
cpu_data          130 drivers/acpi/processor_idle.c 	if (cpu_has(&cpu_data(pr->id), X86_FEATURE_ARAT))
cpu_data          131 drivers/char/hw_random/via-rng.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          179 drivers/cpufreq/acpi-cpufreq.c 	struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
cpu_data          186 drivers/cpufreq/acpi-cpufreq.c 	struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
cpu_data          635 drivers/cpufreq/acpi-cpufreq.c 	struct cpuinfo_x86 *c = &cpu_data(policy->cpu);
cpu_data           36 drivers/cpufreq/amd_freq_sensitivity.c static DEFINE_PER_CPU(struct cpu_data_t, cpu_data);
cpu_data           45 drivers/cpufreq/amd_freq_sensitivity.c 	struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu);
cpu_data          183 drivers/cpufreq/e_powersaver.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          146 drivers/cpufreq/elanfreq.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          522 drivers/cpufreq/intel_pstate.c static s16 intel_pstate_get_epb(struct cpudata *cpu_data)
cpu_data          530 drivers/cpufreq/intel_pstate.c 	ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb);
cpu_data          537 drivers/cpufreq/intel_pstate.c static s16 intel_pstate_get_epp(struct cpudata *cpu_data, u64 hwp_req_data)
cpu_data          547 drivers/cpufreq/intel_pstate.c 			epp = rdmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST,
cpu_data          555 drivers/cpufreq/intel_pstate.c 		epp = intel_pstate_get_epb(cpu_data);
cpu_data          605 drivers/cpufreq/intel_pstate.c static int intel_pstate_get_energy_pref_index(struct cpudata *cpu_data)
cpu_data          610 drivers/cpufreq/intel_pstate.c 	epp = intel_pstate_get_epp(cpu_data, 0);
cpu_data          640 drivers/cpufreq/intel_pstate.c static int intel_pstate_set_energy_pref_index(struct cpudata *cpu_data,
cpu_data          647 drivers/cpufreq/intel_pstate.c 		epp = cpu_data->epp_default;
cpu_data          654 drivers/cpufreq/intel_pstate.c 		ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, &value);
cpu_data          664 drivers/cpufreq/intel_pstate.c 		ret = wrmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, value);
cpu_data          668 drivers/cpufreq/intel_pstate.c 		ret = intel_pstate_set_epb(cpu_data->cpu, epp);
cpu_data          695 drivers/cpufreq/intel_pstate.c 	struct cpudata *cpu_data = all_cpu_data[policy->cpu];
cpu_data          707 drivers/cpufreq/intel_pstate.c 	intel_pstate_set_energy_pref_index(cpu_data, ret);
cpu_data          714 drivers/cpufreq/intel_pstate.c 	struct cpudata *cpu_data = all_cpu_data[policy->cpu];
cpu_data          717 drivers/cpufreq/intel_pstate.c 	preference = intel_pstate_get_energy_pref_index(cpu_data);
cpu_data          769 drivers/cpufreq/intel_pstate.c 	struct cpudata *cpu_data = all_cpu_data[cpu];
cpu_data          774 drivers/cpufreq/intel_pstate.c 	max = cpu_data->max_perf_ratio;
cpu_data          775 drivers/cpufreq/intel_pstate.c 	min = cpu_data->min_perf_ratio;
cpu_data          777 drivers/cpufreq/intel_pstate.c 	if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE)
cpu_data          788 drivers/cpufreq/intel_pstate.c 	if (cpu_data->epp_policy == cpu_data->policy)
cpu_data          791 drivers/cpufreq/intel_pstate.c 	cpu_data->epp_policy = cpu_data->policy;
cpu_data          793 drivers/cpufreq/intel_pstate.c 	if (cpu_data->epp_saved >= 0) {
cpu_data          794 drivers/cpufreq/intel_pstate.c 		epp = cpu_data->epp_saved;
cpu_data          795 drivers/cpufreq/intel_pstate.c 		cpu_data->epp_saved = -EINVAL;
cpu_data          799 drivers/cpufreq/intel_pstate.c 	if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) {
cpu_data          800 drivers/cpufreq/intel_pstate.c 		epp = intel_pstate_get_epp(cpu_data, value);
cpu_data          801 drivers/cpufreq/intel_pstate.c 		cpu_data->epp_powersave = epp;
cpu_data          809 drivers/cpufreq/intel_pstate.c 		if (cpu_data->epp_powersave < 0)
cpu_data          819 drivers/cpufreq/intel_pstate.c 		epp = intel_pstate_get_epp(cpu_data, value);
cpu_data          823 drivers/cpufreq/intel_pstate.c 		epp = cpu_data->epp_powersave;
cpu_data          833 drivers/cpufreq/intel_pstate.c 	WRITE_ONCE(cpu_data->hwp_req_cached, value);
cpu_data          859 drivers/cpufreq/intel_pstate.c 	struct cpudata *cpu_data = all_cpu_data[policy->cpu];
cpu_data          864 drivers/cpufreq/intel_pstate.c 	cpu_data->epp_saved = intel_pstate_get_epp(cpu_data, 0);
cpu_data         2023 drivers/cpufreq/intel_pstate.c 	struct cpudata *cpu_data = all_cpu_data[cpu];
cpu_data         2025 drivers/cpufreq/intel_pstate.c 	if (!cpu_data->update_util_set)
cpu_data         2029 drivers/cpufreq/intel_pstate.c 	cpu_data->update_util_set = false;
cpu_data          762 drivers/cpufreq/longhaul.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          920 drivers/cpufreq/longhaul.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          166 drivers/cpufreq/longrun.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          157 drivers/cpufreq/p4-clockmod.c 	struct cpuinfo_x86 *c = &cpu_data(policy->cpu);
cpu_data          119 drivers/cpufreq/powernow-k7.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          168 drivers/cpufreq/qoriq-cpufreq.c 	struct cpu_data *data;
cpu_data          235 drivers/cpufreq/qoriq-cpufreq.c 	struct cpu_data *data = policy->driver_data;
cpu_data          249 drivers/cpufreq/qoriq-cpufreq.c 	struct cpu_data *data = policy->driver_data;
cpu_data           73 drivers/cpufreq/sc520_freq.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          236 drivers/cpufreq/speedstep-centrino.c 	struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
cpu_data          347 drivers/cpufreq/speedstep-centrino.c 	struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
cpu_data          256 drivers/cpufreq/speedstep-lib.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          119 drivers/cpuidle/governors/teo.c 	struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
cpu_data          120 drivers/cpuidle/governors/teo.c 	unsigned int sleep_length_us = ktime_to_us(cpu_data->sleep_length_ns);
cpu_data          124 drivers/cpuidle/governors/teo.c 	if (cpu_data->time_span_ns >= cpu_data->sleep_length_ns) {
cpu_data          136 drivers/cpuidle/governors/teo.c 		measured_us = ktime_to_us(cpu_data->time_span_ns);
cpu_data          154 drivers/cpuidle/governors/teo.c 		unsigned int early_hits = cpu_data->states[i].early_hits;
cpu_data          156 drivers/cpuidle/governors/teo.c 		cpu_data->states[i].early_hits -= early_hits >> DECAY_SHIFT;
cpu_data          174 drivers/cpuidle/governors/teo.c 		unsigned int hits = cpu_data->states[idx_timer].hits;
cpu_data          175 drivers/cpuidle/governors/teo.c 		unsigned int misses = cpu_data->states[idx_timer].misses;
cpu_data          183 drivers/cpuidle/governors/teo.c 				cpu_data->states[idx_hit].early_hits += PULSE;
cpu_data          188 drivers/cpuidle/governors/teo.c 		cpu_data->states[idx_timer].misses = misses;
cpu_data          189 drivers/cpuidle/governors/teo.c 		cpu_data->states[idx_timer].hits = hits;
cpu_data          196 drivers/cpuidle/governors/teo.c 	cpu_data->intervals[cpu_data->interval_idx++] = measured_us;
cpu_data          197 drivers/cpuidle/governors/teo.c 	if (cpu_data->interval_idx >= INTERVALS)
cpu_data          198 drivers/cpuidle/governors/teo.c 		cpu_data->interval_idx = 0;
cpu_data          234 drivers/cpuidle/governors/teo.c 	struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
cpu_data          245 drivers/cpuidle/governors/teo.c 	cpu_data->time_span_ns = local_clock();
cpu_data          247 drivers/cpuidle/governors/teo.c 	cpu_data->sleep_length_ns = tick_nohz_get_sleep_length(&delta_tick);
cpu_data          248 drivers/cpuidle/governors/teo.c 	duration_us = ktime_to_us(cpu_data->sleep_length_ns);
cpu_data          278 drivers/cpuidle/governors/teo.c 			hits = cpu_data->states[i].hits;
cpu_data          279 drivers/cpuidle/governors/teo.c 			misses = cpu_data->states[i].misses;
cpu_data          281 drivers/cpuidle/governors/teo.c 			if (early_hits >= cpu_data->states[i].early_hits ||
cpu_data          293 drivers/cpuidle/governors/teo.c 				early_hits = cpu_data->states[i].early_hits;
cpu_data          308 drivers/cpuidle/governors/teo.c 				early_hits = cpu_data->states[i].early_hits;
cpu_data          317 drivers/cpuidle/governors/teo.c 			hits = cpu_data->states[i].hits;
cpu_data          318 drivers/cpuidle/governors/teo.c 			misses = cpu_data->states[i].misses;
cpu_data          328 drivers/cpuidle/governors/teo.c 		hits = cpu_data->states[i].hits;
cpu_data          329 drivers/cpuidle/governors/teo.c 		misses = cpu_data->states[i].misses;
cpu_data          331 drivers/cpuidle/governors/teo.c 		if (early_hits < cpu_data->states[i].early_hits &&
cpu_data          335 drivers/cpuidle/governors/teo.c 			early_hits = cpu_data->states[i].early_hits;
cpu_data          381 drivers/cpuidle/governors/teo.c 			unsigned int val = cpu_data->intervals[i];
cpu_data          440 drivers/cpuidle/governors/teo.c 	struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
cpu_data          449 drivers/cpuidle/governors/teo.c 	    (tick_nohz_idle_got_tick() && cpu_data->sleep_length_ns > TICK_NSEC)) {
cpu_data          451 drivers/cpuidle/governors/teo.c 		cpu_data->time_span_ns = cpu_data->sleep_length_ns;
cpu_data          453 drivers/cpuidle/governors/teo.c 		cpu_data->time_span_ns = local_clock() - cpu_data->time_span_ns;
cpu_data          465 drivers/cpuidle/governors/teo.c 	struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
cpu_data          468 drivers/cpuidle/governors/teo.c 	memset(cpu_data, 0, sizeof(*cpu_data));
cpu_data          471 drivers/cpuidle/governors/teo.c 		cpu_data->intervals[i] = UINT_MAX;
cpu_data          499 drivers/crypto/padlock-aes.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          507 drivers/crypto/padlock-sha.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          546 drivers/crypto/padlock-sha.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data         1214 drivers/edac/e752x_edac.c 	char *cpu_id = cpu_data(0).x86_model_id;
cpu_data          884 drivers/gpu/drm/amd/amdkfd/kfd_crat.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data         1443 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	return cpu_data(first_cpu_of_numa_node).apicid;
cpu_data           49 drivers/hwmon/coretemp.c #define TO_CORE_ID(cpu)		(cpu_data(cpu).cpu_core_id)
cpu_data          409 drivers/hwmon/coretemp.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          455 drivers/hwmon/coretemp.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          592 drivers/hwmon/coretemp.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          149 drivers/hwmon/fam15h_power.c 	cu = cpu_data(cpu).cpu_core_id;
cpu_data          277 drivers/hwmon/hwmon-vid.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data          113 drivers/hwmon/via-cputemp.c 	struct cpuinfo_x86 *c = &cpu_data(pdev->id);
cpu_data          103 drivers/irqchip/irq-mips-cpu.c 	settc(cpu_vpe_id(&cpu_data[cpu]));
cpu_data          302 drivers/misc/mic/card/mic_x100.c 	struct cpuinfo_x86 *c = &cpu_data(0);
cpu_data         1293 drivers/powercap/intel_rapl_common.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data         10675 drivers/scsi/lpfc/lpfc_init.c 		cpuinfo = &cpu_data(cpu);
cpu_data          466 drivers/thermal/intel/x86_pkg_temp_thermal.c 	struct cpuinfo_x86 *c = &cpu_data(cpu);
cpu_data          126 drivers/video/fbdev/geode/video_gx.c 	if (cpu_data(0).x86_stepping == 1) {
cpu_data           31 kernel/trace/trace_functions_graph.c 	struct fgraph_cpu_data __percpu *cpu_data;
cpu_data          393 kernel/trace/trace_functions_graph.c 	last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
cpu_data          644 kernel/trace/trace_functions_graph.c 		struct fgraph_cpu_data *cpu_data;
cpu_data          646 kernel/trace/trace_functions_graph.c 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);
cpu_data          653 kernel/trace/trace_functions_graph.c 		cpu_data->depth = call->depth - 1;
cpu_data          658 kernel/trace/trace_functions_graph.c 			cpu_data->enter_funcs[call->depth] = 0;
cpu_data          687 kernel/trace/trace_functions_graph.c 		struct fgraph_cpu_data *cpu_data;
cpu_data          690 kernel/trace/trace_functions_graph.c 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);
cpu_data          691 kernel/trace/trace_functions_graph.c 		cpu_data->depth = call->depth;
cpu_data          696 kernel/trace/trace_functions_graph.c 			cpu_data->enter_funcs[call->depth] = call->func;
cpu_data          790 kernel/trace/trace_functions_graph.c 	depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
cpu_data          836 kernel/trace/trace_functions_graph.c 	depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
cpu_data          917 kernel/trace/trace_functions_graph.c 		struct fgraph_cpu_data *cpu_data;
cpu_data          920 kernel/trace/trace_functions_graph.c 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);
cpu_data          927 kernel/trace/trace_functions_graph.c 		cpu_data->depth = trace->depth - 1;
cpu_data          931 kernel/trace/trace_functions_graph.c 			if (cpu_data->enter_funcs[trace->depth] != trace->func)
cpu_data          933 kernel/trace/trace_functions_graph.c 			cpu_data->enter_funcs[trace->depth] = 0;
cpu_data          982 kernel/trace/trace_functions_graph.c 		depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth;
cpu_data         1048 kernel/trace/trace_functions_graph.c 	if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) {
cpu_data         1049 kernel/trace/trace_functions_graph.c 		per_cpu_ptr(data->cpu_data, cpu)->ignore = 0;
cpu_data         1062 kernel/trace/trace_functions_graph.c 			per_cpu_ptr(data->cpu_data, iter->cpu)->ignore = 1;
cpu_data         1216 kernel/trace/trace_functions_graph.c 	data->cpu_data = alloc_percpu_gfp(struct fgraph_cpu_data, gfpflags);
cpu_data         1217 kernel/trace/trace_functions_graph.c 	if (!data->cpu_data)
cpu_data         1221 kernel/trace/trace_functions_graph.c 		pid_t *pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
cpu_data         1222 kernel/trace/trace_functions_graph.c 		int *depth = &(per_cpu_ptr(data->cpu_data, cpu)->depth);
cpu_data         1223 kernel/trace/trace_functions_graph.c 		int *ignore = &(per_cpu_ptr(data->cpu_data, cpu)->ignore);
cpu_data         1224 kernel/trace/trace_functions_graph.c 		int *depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
cpu_data         1247 kernel/trace/trace_functions_graph.c 		free_percpu(data->cpu_data);
cpu_data           11 tools/testing/selftests/bpf/prog_tests/perf_buffer.c 	int cpu_data = *(int *)data, duration = 0;
cpu_data           14 tools/testing/selftests/bpf/prog_tests/perf_buffer.c 	if (cpu_data != cpu)
cpu_data           15 tools/testing/selftests/bpf/prog_tests/perf_buffer.c 		CHECK(cpu_data != cpu, "check_cpu_data",
cpu_data           16 tools/testing/selftests/bpf/prog_tests/perf_buffer.c 		      "cpu_data %d != cpu %d\n", cpu_data, cpu);
cpu_data          370 virt/kvm/arm/arm.c 	kvm_host_data_t *cpu_data;
cpu_data          373 virt/kvm/arm/arm.c 	cpu_data = this_cpu_ptr(&kvm_host_data);
cpu_data          385 virt/kvm/arm/arm.c 	vcpu->arch.host_cpu_context = &cpu_data->host_ctxt;
cpu_data         1601 virt/kvm/arm/arm.c 		kvm_host_data_t *cpu_data;
cpu_data         1603 virt/kvm/arm/arm.c 		cpu_data = per_cpu_ptr(&kvm_host_data, cpu);
cpu_data         1604 virt/kvm/arm/arm.c 		err = create_hyp_mappings(cpu_data, cpu_data + 1, PAGE_HYP);