lcpu               29 arch/arm/mach-shmobile/smp-sh73a0.c 	unsigned int lcpu = cpu_logical_map(cpu);
lcpu               31 arch/arm/mach-shmobile/smp-sh73a0.c 	if (((__raw_readl(PSTR) >> (4 * lcpu)) & 3) == 3)
lcpu               32 arch/arm/mach-shmobile/smp-sh73a0.c 		__raw_writel(1 << lcpu, WUPCR);	/* wake up */
lcpu               34 arch/arm/mach-shmobile/smp-sh73a0.c 		__raw_writel(1 << lcpu, SRESCR);	/* reset */
lcpu              468 arch/powerpc/mm/numa.c static int numa_setup_cpu(unsigned long lcpu)
lcpu              478 arch/powerpc/mm/numa.c 	if ((nid = numa_cpu_lookup_table[lcpu]) >= 0) {
lcpu              479 arch/powerpc/mm/numa.c 		map_cpu_to_node(lcpu, nid);
lcpu              483 arch/powerpc/mm/numa.c 	cpu = of_get_cpu_node(lcpu, NULL);
lcpu              487 arch/powerpc/mm/numa.c 		if (cpu_present(lcpu))
lcpu              499 arch/powerpc/mm/numa.c 	map_cpu_to_node(lcpu, nid);
lcpu               67 arch/powerpc/platforms/cell/smp.c static inline int smp_startup_cpu(unsigned int lcpu)
lcpu               75 arch/powerpc/platforms/cell/smp.c 	if (cpumask_test_cpu(lcpu, &of_spin_map))
lcpu               79 arch/powerpc/platforms/cell/smp.c 	pcpu = get_hard_smp_processor_id(lcpu);
lcpu               82 arch/powerpc/platforms/cell/smp.c 	task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count	= 0;
lcpu               92 arch/powerpc/platforms/cell/smp.c 	status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, lcpu);
lcpu               89 arch/powerpc/platforms/pseries/smp.c static inline int smp_startup_cpu(unsigned int lcpu)
lcpu               97 arch/powerpc/platforms/pseries/smp.c 	if (cpumask_test_cpu(lcpu, of_spin_mask))
lcpu              101 arch/powerpc/platforms/pseries/smp.c 	pcpu = get_hard_smp_processor_id(lcpu);
lcpu              105 arch/powerpc/platforms/pseries/smp.c 		cpumask_set_cpu(lcpu, of_spin_mask);
lcpu              110 arch/powerpc/platforms/pseries/smp.c 	task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count	= 0;
lcpu              112 arch/powerpc/platforms/pseries/smp.c 	if (get_cpu_current_state(lcpu) == CPU_STATE_INACTIVE)
lcpu              124 arch/s390/kernel/topology.c 		int lcpu, i;
lcpu              127 arch/s390/kernel/topology.c 		lcpu = smp_find_processor_id(rcore << smp_cpu_mt_shift);
lcpu              128 arch/s390/kernel/topology.c 		if (lcpu < 0)
lcpu              131 arch/s390/kernel/topology.c 			topo = &cpu_topology[lcpu + i];
lcpu              136 arch/s390/kernel/topology.c 			topo->thread_id = lcpu + i;
lcpu              138 arch/s390/kernel/topology.c 			cpumask_set_cpu(lcpu + i, &drawer->mask);
lcpu              139 arch/s390/kernel/topology.c 			cpumask_set_cpu(lcpu + i, &book->mask);
lcpu              140 arch/s390/kernel/topology.c 			cpumask_set_cpu(lcpu + i, &socket->mask);
lcpu              141 arch/s390/kernel/topology.c 			cpumask_set_cpu(lcpu + i, &cpus_with_topology);
lcpu              142 arch/s390/kernel/topology.c 			smp_cpu_set_polarization(lcpu + i, tl_core->pp);
lcpu             1487 arch/x86/kernel/cpu/mce/amd.c 	unsigned lcpu = 0;
lcpu             1490 arch/x86/kernel/cpu/mce/amd.c 	for_each_online_cpu(lcpu) {
lcpu             1491 arch/x86/kernel/cpu/mce/amd.c 		int err = mce_threshold_create_device(lcpu);
lcpu               53 arch/x86/platform/uv/uv_time.c 		int	lcpu;		/* systemwide logical cpu number */
lcpu              173 arch/x86/platform/uv/uv_time.c 		head->cpu[bcpu].lcpu = cpu;
lcpu              196 arch/x86/platform/uv/uv_time.c 		c = head->cpu[bcpu].lcpu;
lcpu              257 drivers/misc/sgi-gru/grukservices.c 	int lcpu;
lcpu              262 drivers/misc/sgi-gru/grukservices.c 	lcpu = uv_blade_processor_id();
lcpu              263 drivers/misc/sgi-gru/grukservices.c 	*cb = bs->kernel_cb + lcpu * GRU_HANDLE_STRIDE;
lcpu              264 drivers/misc/sgi-gru/grukservices.c 	*dsr = bs->kernel_dsr + lcpu * GRU_NUM_KERNEL_DSR_BYTES;