lcpu 29 arch/arm/mach-shmobile/smp-sh73a0.c unsigned int lcpu = cpu_logical_map(cpu); lcpu 31 arch/arm/mach-shmobile/smp-sh73a0.c if (((__raw_readl(PSTR) >> (4 * lcpu)) & 3) == 3) lcpu 32 arch/arm/mach-shmobile/smp-sh73a0.c __raw_writel(1 << lcpu, WUPCR); /* wake up */ lcpu 34 arch/arm/mach-shmobile/smp-sh73a0.c __raw_writel(1 << lcpu, SRESCR); /* reset */ lcpu 468 arch/powerpc/mm/numa.c static int numa_setup_cpu(unsigned long lcpu) lcpu 478 arch/powerpc/mm/numa.c if ((nid = numa_cpu_lookup_table[lcpu]) >= 0) { lcpu 479 arch/powerpc/mm/numa.c map_cpu_to_node(lcpu, nid); lcpu 483 arch/powerpc/mm/numa.c cpu = of_get_cpu_node(lcpu, NULL); lcpu 487 arch/powerpc/mm/numa.c if (cpu_present(lcpu)) lcpu 499 arch/powerpc/mm/numa.c map_cpu_to_node(lcpu, nid); lcpu 67 arch/powerpc/platforms/cell/smp.c static inline int smp_startup_cpu(unsigned int lcpu) lcpu 75 arch/powerpc/platforms/cell/smp.c if (cpumask_test_cpu(lcpu, &of_spin_map)) lcpu 79 arch/powerpc/platforms/cell/smp.c pcpu = get_hard_smp_processor_id(lcpu); lcpu 82 arch/powerpc/platforms/cell/smp.c task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count = 0; lcpu 92 arch/powerpc/platforms/cell/smp.c status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, lcpu); lcpu 89 arch/powerpc/platforms/pseries/smp.c static inline int smp_startup_cpu(unsigned int lcpu) lcpu 97 arch/powerpc/platforms/pseries/smp.c if (cpumask_test_cpu(lcpu, of_spin_mask)) lcpu 101 arch/powerpc/platforms/pseries/smp.c pcpu = get_hard_smp_processor_id(lcpu); lcpu 105 arch/powerpc/platforms/pseries/smp.c cpumask_set_cpu(lcpu, of_spin_mask); lcpu 110 arch/powerpc/platforms/pseries/smp.c task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count = 0; lcpu 112 arch/powerpc/platforms/pseries/smp.c if (get_cpu_current_state(lcpu) == CPU_STATE_INACTIVE) lcpu 124 arch/s390/kernel/topology.c int lcpu, i; lcpu 127 arch/s390/kernel/topology.c lcpu = smp_find_processor_id(rcore << smp_cpu_mt_shift); lcpu 128 arch/s390/kernel/topology.c if (lcpu < 0) lcpu 131 arch/s390/kernel/topology.c topo = &cpu_topology[lcpu + i]; lcpu 136 arch/s390/kernel/topology.c topo->thread_id = lcpu + i; lcpu 138 arch/s390/kernel/topology.c cpumask_set_cpu(lcpu + i, &drawer->mask); lcpu 139 arch/s390/kernel/topology.c cpumask_set_cpu(lcpu + i, &book->mask); lcpu 140 arch/s390/kernel/topology.c cpumask_set_cpu(lcpu + i, &socket->mask); lcpu 141 arch/s390/kernel/topology.c cpumask_set_cpu(lcpu + i, &cpus_with_topology); lcpu 142 arch/s390/kernel/topology.c smp_cpu_set_polarization(lcpu + i, tl_core->pp); lcpu 1487 arch/x86/kernel/cpu/mce/amd.c unsigned lcpu = 0; lcpu 1490 arch/x86/kernel/cpu/mce/amd.c for_each_online_cpu(lcpu) { lcpu 1491 arch/x86/kernel/cpu/mce/amd.c int err = mce_threshold_create_device(lcpu); lcpu 53 arch/x86/platform/uv/uv_time.c int lcpu; /* systemwide logical cpu number */ lcpu 173 arch/x86/platform/uv/uv_time.c head->cpu[bcpu].lcpu = cpu; lcpu 196 arch/x86/platform/uv/uv_time.c c = head->cpu[bcpu].lcpu; lcpu 257 drivers/misc/sgi-gru/grukservices.c int lcpu; lcpu 262 drivers/misc/sgi-gru/grukservices.c lcpu = uv_blade_processor_id(); lcpu 263 drivers/misc/sgi-gru/grukservices.c *cb = bs->kernel_cb + lcpu * GRU_HANDLE_STRIDE; lcpu 264 drivers/misc/sgi-gru/grukservices.c *dsr = bs->kernel_dsr + lcpu * GRU_NUM_KERNEL_DSR_BYTES;