pcpu              242 arch/alpha/kernel/err_common.c cdl_process_console_data_log(int cpu, struct percpu_struct *pcpu)
pcpu              245 arch/alpha/kernel/err_common.c 		(IDENT_ADDR | pcpu->console_data_log_pa);
pcpu              256 arch/alpha/kernel/err_common.c 	pcpu->console_data_log_pa = 0;
pcpu              266 arch/alpha/kernel/err_common.c 	struct percpu_struct *pcpu;
pcpu              270 arch/alpha/kernel/err_common.c 		pcpu = (struct percpu_struct *)
pcpu              273 arch/alpha/kernel/err_common.c 		if (pcpu->console_data_log_pa)
pcpu              274 arch/alpha/kernel/err_common.c 			cdl_process_console_data_log(cpu, pcpu);
pcpu               20 arch/arm/common/mcpm_platsmp.c 			unsigned int *pcpu, unsigned int *pcluster)
pcpu               25 arch/arm/common/mcpm_platsmp.c 	*pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0);
pcpu               31 arch/arm/common/mcpm_platsmp.c 	unsigned int pcpu, pcluster, ret;
pcpu               34 arch/arm/common/mcpm_platsmp.c 	cpu_to_pcpu(cpu, &pcpu, &pcluster);
pcpu               37 arch/arm/common/mcpm_platsmp.c 		 __func__, cpu, pcpu, pcluster);
pcpu               39 arch/arm/common/mcpm_platsmp.c 	mcpm_set_entry_vector(pcpu, pcluster, NULL);
pcpu               40 arch/arm/common/mcpm_platsmp.c 	ret = mcpm_cpu_power_up(pcpu, pcluster);
pcpu               43 arch/arm/common/mcpm_platsmp.c 	mcpm_set_entry_vector(pcpu, pcluster, secondary_startup);
pcpu               58 arch/arm/common/mcpm_platsmp.c 	unsigned int pcpu, pcluster;
pcpu               60 arch/arm/common/mcpm_platsmp.c 	cpu_to_pcpu(cpu, &pcpu, &pcluster);
pcpu               62 arch/arm/common/mcpm_platsmp.c 	return !mcpm_wait_for_cpu_powerdown(pcpu, pcluster);
pcpu               73 arch/arm/common/mcpm_platsmp.c 	unsigned int mpidr, pcpu, pcluster;
pcpu               75 arch/arm/common/mcpm_platsmp.c 	pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0);
pcpu               77 arch/arm/common/mcpm_platsmp.c 	mcpm_set_entry_vector(pcpu, pcluster, NULL);
pcpu              783 arch/ia64/kernel/acpi.c static int _acpi_map_lsapic(acpi_handle handle, int physid, int *pcpu)
pcpu              800 arch/ia64/kernel/acpi.c 	*pcpu = cpu;
pcpu              806 arch/ia64/kernel/acpi.c 		       int *pcpu)
pcpu              808 arch/ia64/kernel/acpi.c 	return _acpi_map_lsapic(handle, physid, pcpu);
pcpu               74 arch/powerpc/include/asm/kvm_book3s.h 	u16 pcpu;
pcpu              560 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu,
pcpu              722 arch/powerpc/kvm/book3s_hv.c 	dt->processor_id = cpu_to_be16(vc->pcpu + vcpu->arch.ptid);
pcpu             1163 arch/powerpc/kvm/book3s_hv.c 	int thr, cpu, pcpu, nthreads;
pcpu             1179 arch/powerpc/kvm/book3s_hv.c 		pcpu = READ_ONCE(v->cpu);
pcpu             1180 arch/powerpc/kvm/book3s_hv.c 		if (pcpu >= 0)
pcpu             1181 arch/powerpc/kvm/book3s_hv.c 			smp_call_function_single(pcpu, do_nothing, NULL, 1);
pcpu             2558 arch/powerpc/kvm/book3s_hv.c static void kvmppc_prepare_radix_vcpu(struct kvm_vcpu *vcpu, int pcpu)
pcpu             2584 arch/powerpc/kvm/book3s_hv.c 	if (prev_cpu != pcpu) {
pcpu             2587 arch/powerpc/kvm/book3s_hv.c 		    cpu_first_thread_sibling(pcpu))
pcpu             2590 arch/powerpc/kvm/book3s_hv.c 			nested->prev_cpu[vcpu->arch.nested_vcpu_id] = pcpu;
pcpu             2592 arch/powerpc/kvm/book3s_hv.c 			vcpu->arch.prev_cpu = pcpu;
pcpu             2602 arch/powerpc/kvm/book3s_hv.c 	cpu = vc->pcpu;
pcpu             2609 arch/powerpc/kvm/book3s_hv.c 		vcpu->cpu = vc->pcpu;
pcpu             2615 arch/powerpc/kvm/book3s_hv.c 	tpaca->kvm_hstate.ptid = cpu - vc->pcpu;
pcpu             2713 arch/powerpc/kvm/book3s_hv.c 	vc->pcpu = smp_processor_id();
pcpu             2730 arch/powerpc/kvm/book3s_hv.c 		lp = &per_cpu(preempted_vcores, vc->pcpu);
pcpu             3054 arch/powerpc/kvm/book3s_hv.c 	int pcpu, thr;
pcpu             3110 arch/powerpc/kvm/book3s_hv.c 	pcpu = smp_processor_id();
pcpu             3122 arch/powerpc/kvm/book3s_hv.c 	pcpu = smp_processor_id();
pcpu             3126 arch/powerpc/kvm/book3s_hv.c 				kvmppc_prepare_radix_vcpu(vcpu, pcpu);
pcpu             3150 arch/powerpc/kvm/book3s_hv.c 			kvmppc_release_hwthread(pcpu + i);
pcpu             3154 arch/powerpc/kvm/book3s_hv.c 	kvmppc_clear_host_core(pcpu);
pcpu             3200 arch/powerpc/kvm/book3s_hv.c 		struct paca_struct *paca = paca_ptrs[pcpu + thr];
pcpu             3249 arch/powerpc/kvm/book3s_hv.c 		pvc->pcpu = pcpu + thr;
pcpu             3283 arch/powerpc/kvm/book3s_hv.c 				kvmppc_ipi_thread(pcpu + thr);
pcpu             3343 arch/powerpc/kvm/book3s_hv.c 			struct paca_struct *paca = paca_ptrs[pcpu + thr];
pcpu             3354 arch/powerpc/kvm/book3s_hv.c 	kvmppc_set_host_core(pcpu);
pcpu             3361 arch/powerpc/kvm/book3s_hv.c 		kvmppc_release_hwthread(pcpu + i);
pcpu             3363 arch/powerpc/kvm/book3s_hv.c 			kvmppc_ipi_thread(pcpu + i);
pcpu             3364 arch/powerpc/kvm/book3s_hv.c 		cpumask_clear_cpu(pcpu + i, &vc->kvm->arch.cpu_in_guest);
pcpu             4040 arch/powerpc/kvm/book3s_hv.c 	int trap, r, pcpu;
pcpu             4077 arch/powerpc/kvm/book3s_hv.c 	pcpu = smp_processor_id();
pcpu             4078 arch/powerpc/kvm/book3s_hv.c 	vc->pcpu = pcpu;
pcpu             4079 arch/powerpc/kvm/book3s_hv.c 	kvmppc_prepare_radix_vcpu(vcpu, pcpu);
pcpu             4105 arch/powerpc/kvm/book3s_hv.c 	kvmppc_clear_host_core(pcpu);
pcpu             4121 arch/powerpc/kvm/book3s_hv.c 		kvmppc_check_need_tlb_flush(kvm, pcpu, nested);
pcpu             4149 arch/powerpc/kvm/book3s_hv.c 	kvmppc_set_host_core(pcpu);
pcpu             4154 arch/powerpc/kvm/book3s_hv.c 	cpumask_clear_cpu(pcpu, &kvm->arch.cpu_in_guest);
pcpu              274 arch/powerpc/kvm/book3s_hv_builtin.c 	int cpu = vc->pcpu;
pcpu              340 arch/powerpc/kvm/book3s_hv_builtin.c 			cpu0 = vc->pcpu + ptid - local_paca->kvm_hstate.tid;
pcpu              838 arch/powerpc/kvm/book3s_hv_builtin.c void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu,
pcpu              850 arch/powerpc/kvm/book3s_hv_builtin.c 		pcpu = cpu_first_thread_sibling(pcpu);
pcpu              857 arch/powerpc/kvm/book3s_hv_builtin.c 	if (cpumask_test_cpu(pcpu, need_tlb_flush)) {
pcpu              861 arch/powerpc/kvm/book3s_hv_builtin.c 		cpumask_clear_cpu(pcpu, need_tlb_flush);
pcpu               64 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		cpu = local_paca->kvm_hstate.kvm_vcore->pcpu;
pcpu              719 arch/powerpc/kvm/book3s_hv_rm_xics.c 			int pcpu = raw_smp_processor_id();
pcpu              721 arch/powerpc/kvm/book3s_hv_rm_xics.c 			pcpu = cpu_first_thread_sibling(pcpu);
pcpu              723 arch/powerpc/kvm/book3s_hv_rm_xics.c 			if (state->intr_cpu != pcpu) {
pcpu              725 arch/powerpc/kvm/book3s_hv_rm_xics.c 				xics_opal_set_server(state->host_irq, pcpu);
pcpu               72 arch/powerpc/platforms/cell/smp.c 	unsigned int pcpu;
pcpu               79 arch/powerpc/platforms/cell/smp.c 	pcpu = get_hard_smp_processor_id(lcpu);
pcpu               92 arch/powerpc/platforms/cell/smp.c 	status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, lcpu);
pcpu               66 arch/powerpc/platforms/powernv/smp.c 	unsigned int pcpu;
pcpu               75 arch/powerpc/platforms/powernv/smp.c 	pcpu = get_hard_smp_processor_id(nr);
pcpu               89 arch/powerpc/platforms/powernv/smp.c 	rc = opal_query_cpu_status(pcpu, &status);
pcpu              106 arch/powerpc/platforms/powernv/smp.c 		pr_devel("OPAL: Starting CPU %d (HW 0x%x)...\n", nr, pcpu);
pcpu              107 arch/powerpc/platforms/powernv/smp.c 		rc = opal_start_cpu(pcpu, start_here);
pcpu              120 arch/powerpc/platforms/powernv/smp.c 			 " (status %d)...\n", nr, pcpu, status);
pcpu              201 arch/powerpc/platforms/pseries/hotplug-cpu.c 	unsigned int pcpu = get_hard_smp_processor_id(cpu);
pcpu              215 arch/powerpc/platforms/pseries/hotplug-cpu.c 			cpu_status = smp_query_cpu_stopped(pcpu);
pcpu              225 arch/powerpc/platforms/pseries/hotplug-cpu.c 		       cpu, pcpu, cpu_status);
pcpu               29 arch/powerpc/platforms/pseries/pseries.h int smp_query_cpu_stopped(unsigned int pcpu);
pcpu               57 arch/powerpc/platforms/pseries/smp.c int smp_query_cpu_stopped(unsigned int pcpu)
pcpu               68 arch/powerpc/platforms/pseries/smp.c 	status = rtas_call(qcss_tok, 1, 2, &cpu_status, pcpu);
pcpu               94 arch/powerpc/platforms/pseries/smp.c 	unsigned int pcpu;
pcpu              101 arch/powerpc/platforms/pseries/smp.c 	pcpu = get_hard_smp_processor_id(lcpu);
pcpu              104 arch/powerpc/platforms/pseries/smp.c 	if (smp_query_cpu_stopped(pcpu) == QCSS_NOT_STOPPED){
pcpu              123 arch/powerpc/platforms/pseries/smp.c 	status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, pcpu);
pcpu               42 arch/s390/include/asm/smp.h 	u16 pcpu = stap();
pcpu               45 arch/s390/include/asm/smp.h 		__pcpu_sigp(pcpu, SIGP_STOP, 0, NULL);
pcpu               83 arch/s390/kernel/smp.c static struct pcpu pcpu_devices[NR_CPUS];
pcpu              132 arch/s390/kernel/smp.c static int pcpu_sigp_retry(struct pcpu *pcpu, u8 order, u32 parm)
pcpu              137 arch/s390/kernel/smp.c 		cc = __pcpu_sigp(pcpu->address, order, parm, NULL);
pcpu              146 arch/s390/kernel/smp.c static inline int pcpu_stopped(struct pcpu *pcpu)
pcpu              150 arch/s390/kernel/smp.c 	if (__pcpu_sigp(pcpu->address, SIGP_SENSE,
pcpu              156 arch/s390/kernel/smp.c static inline int pcpu_running(struct pcpu *pcpu)
pcpu              158 arch/s390/kernel/smp.c 	if (__pcpu_sigp(pcpu->address, SIGP_SENSE_RUNNING,
pcpu              168 arch/s390/kernel/smp.c static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address)
pcpu              178 arch/s390/kernel/smp.c static void pcpu_ec_call(struct pcpu *pcpu, int ec_bit)
pcpu              182 arch/s390/kernel/smp.c 	if (test_and_set_bit(ec_bit, &pcpu->ec_mask))
pcpu              184 arch/s390/kernel/smp.c 	order = pcpu_running(pcpu) ? SIGP_EXTERNAL_CALL : SIGP_EMERGENCY_SIGNAL;
pcpu              185 arch/s390/kernel/smp.c 	pcpu->ec_clk = get_tod_clock_fast();
pcpu              186 arch/s390/kernel/smp.c 	pcpu_sigp_retry(pcpu, order, 0);
pcpu              189 arch/s390/kernel/smp.c static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
pcpu              194 arch/s390/kernel/smp.c 	if (pcpu != &pcpu_devices[0]) {
pcpu              195 arch/s390/kernel/smp.c 		pcpu->lowcore =	(struct lowcore *)
pcpu              198 arch/s390/kernel/smp.c 		if (!pcpu->lowcore || !nodat_stack)
pcpu              201 arch/s390/kernel/smp.c 		nodat_stack = pcpu->lowcore->nodat_stack - STACK_INIT_OFFSET;
pcpu              206 arch/s390/kernel/smp.c 	lc = pcpu->lowcore;
pcpu              222 arch/s390/kernel/smp.c 	pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, (u32)(unsigned long) lc);
pcpu              230 arch/s390/kernel/smp.c 	if (pcpu != &pcpu_devices[0]) {
pcpu              232 arch/s390/kernel/smp.c 		free_pages((unsigned long) pcpu->lowcore, LC_ORDER);
pcpu              237 arch/s390/kernel/smp.c static void pcpu_free_lowcore(struct pcpu *pcpu)
pcpu              241 arch/s390/kernel/smp.c 	nodat_stack = pcpu->lowcore->nodat_stack - STACK_INIT_OFFSET;
pcpu              242 arch/s390/kernel/smp.c 	async_stack = pcpu->lowcore->async_stack - STACK_INIT_OFFSET;
pcpu              243 arch/s390/kernel/smp.c 	lowcore = (unsigned long) pcpu->lowcore;
pcpu              245 arch/s390/kernel/smp.c 	pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, 0);
pcpu              246 arch/s390/kernel/smp.c 	lowcore_ptr[pcpu - pcpu_devices] = NULL;
pcpu              247 arch/s390/kernel/smp.c 	vdso_free_per_cpu(pcpu->lowcore);
pcpu              248 arch/s390/kernel/smp.c 	nmi_free_per_cpu(pcpu->lowcore);
pcpu              250 arch/s390/kernel/smp.c 	if (pcpu == &pcpu_devices[0])
pcpu              256 arch/s390/kernel/smp.c static void pcpu_prepare_secondary(struct pcpu *pcpu, int cpu)
pcpu              258 arch/s390/kernel/smp.c 	struct lowcore *lc = pcpu->lowcore;
pcpu              282 arch/s390/kernel/smp.c static void pcpu_attach_task(struct pcpu *pcpu, struct task_struct *tsk)
pcpu              284 arch/s390/kernel/smp.c 	struct lowcore *lc = pcpu->lowcore;
pcpu              299 arch/s390/kernel/smp.c static void pcpu_start_fn(struct pcpu *pcpu, void (*func)(void *), void *data)
pcpu              301 arch/s390/kernel/smp.c 	struct lowcore *lc = pcpu->lowcore;
pcpu              307 arch/s390/kernel/smp.c 	pcpu_sigp_retry(pcpu, SIGP_RESTART, 0);
pcpu              318 arch/s390/kernel/smp.c static void __no_sanitize_address pcpu_delegate(struct pcpu *pcpu,
pcpu              322 arch/s390/kernel/smp.c 	struct lowcore *lc = lowcore_ptr[pcpu - pcpu_devices];
pcpu              326 arch/s390/kernel/smp.c 	if (pcpu->address == source_cpu)
pcpu              329 arch/s390/kernel/smp.c 	pcpu_sigp_retry(pcpu, SIGP_STOP, 0);
pcpu              341 arch/s390/kernel/smp.c 		: : "d" (pcpu->address), "d" (source_cpu),
pcpu              372 arch/s390/kernel/smp.c 	struct pcpu *pcpu;
pcpu              375 arch/s390/kernel/smp.c 	pcpu = pcpu_find_address(cpu_online_mask, stap());
pcpu              376 arch/s390/kernel/smp.c 	if (!pcpu)
pcpu              378 arch/s390/kernel/smp.c 		pcpu = pcpu_devices + cpumask_first(cpu_online_mask);
pcpu              379 arch/s390/kernel/smp.c 	pcpu_delegate(pcpu, func, data, (unsigned long) restart_stack);
pcpu              443 arch/s390/kernel/smp.c 		struct pcpu *pcpu = pcpu_devices + cpu;
pcpu              444 arch/s390/kernel/smp.c 		set_bit(ec_stop_cpu, &pcpu->ec_mask);
pcpu              445 arch/s390/kernel/smp.c 		while (__pcpu_sigp(pcpu->address, SIGP_EMERGENCY_SIGNAL,
pcpu              583 arch/s390/kernel/smp.c 	struct pcpu *pcpu = pcpu_devices + cpu;
pcpu              586 arch/s390/kernel/smp.c 	pa = __pa(&pcpu->lowcore->floating_pt_save_area);
pcpu              587 arch/s390/kernel/smp.c 	if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_STATUS_AT_ADDRESS,
pcpu              592 arch/s390/kernel/smp.c 	pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK);
pcpu              594 arch/s390/kernel/smp.c 		pa |= pcpu->lowcore->mcesad & MCESA_LC_MASK;
pcpu              595 arch/s390/kernel/smp.c 	if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_ADDITIONAL_STATUS,
pcpu              735 arch/s390/kernel/smp.c 	struct pcpu *pcpu;
pcpu              747 arch/s390/kernel/smp.c 		pcpu = pcpu_devices + cpu;
pcpu              748 arch/s390/kernel/smp.c 		pcpu->address = address + i;
pcpu              750 arch/s390/kernel/smp.c 			pcpu->state = CPU_STATE_CONFIGURED;
pcpu              752 arch/s390/kernel/smp.c 			pcpu->state = CPU_STATE_STANDBY;
pcpu              887 arch/s390/kernel/smp.c 	struct pcpu *pcpu;
pcpu              890 arch/s390/kernel/smp.c 	pcpu = pcpu_devices + cpu;
pcpu              891 arch/s390/kernel/smp.c 	if (pcpu->state != CPU_STATE_CONFIGURED)
pcpu              908 arch/s390/kernel/smp.c 	rc = pcpu_alloc_lowcore(pcpu, cpu);
pcpu              911 arch/s390/kernel/smp.c 	pcpu_prepare_secondary(pcpu, cpu);
pcpu              912 arch/s390/kernel/smp.c 	pcpu_attach_task(pcpu, tidle);
pcpu              913 arch/s390/kernel/smp.c 	pcpu_start_fn(pcpu, smp_start_secondary, NULL);
pcpu              950 arch/s390/kernel/smp.c 	struct pcpu *pcpu;
pcpu              953 arch/s390/kernel/smp.c 	pcpu = pcpu_devices + cpu;
pcpu              954 arch/s390/kernel/smp.c 	while (!pcpu_stopped(pcpu))
pcpu              956 arch/s390/kernel/smp.c 	pcpu_free_lowcore(pcpu);
pcpu              994 arch/s390/kernel/smp.c 	struct pcpu *pcpu = pcpu_devices;
pcpu              997 arch/s390/kernel/smp.c 	pcpu->state = CPU_STATE_CONFIGURED;
pcpu              998 arch/s390/kernel/smp.c 	pcpu->lowcore = (struct lowcore *)(unsigned long) store_prefix();
pcpu             1041 arch/s390/kernel/smp.c 	struct pcpu *pcpu;
pcpu             1060 arch/s390/kernel/smp.c 	pcpu = pcpu_devices + cpu;
pcpu             1064 arch/s390/kernel/smp.c 		if (pcpu->state != CPU_STATE_CONFIGURED)
pcpu             1066 arch/s390/kernel/smp.c 		rc = sclp_core_deconfigure(pcpu->address >> smp_cpu_mt_shift);
pcpu             1072 arch/s390/kernel/smp.c 			pcpu[i].state = CPU_STATE_STANDBY;
pcpu             1079 arch/s390/kernel/smp.c 		if (pcpu->state != CPU_STATE_STANDBY)
pcpu             1081 arch/s390/kernel/smp.c 		rc = sclp_core_configure(pcpu->address >> smp_cpu_mt_shift);
pcpu             1087 arch/s390/kernel/smp.c 			pcpu[i].state = CPU_STATE_CONFIGURED;
pcpu               92 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs __percpu	*pcpu;
pcpu              396 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
pcpu              410 arch/x86/events/amd/ibs.c 	set_bit(IBS_STARTED,    pcpu->state);
pcpu              411 arch/x86/events/amd/ibs.c 	clear_bit(IBS_STOPPING, pcpu->state);
pcpu              421 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
pcpu              425 arch/x86/events/amd/ibs.c 	if (test_and_set_bit(IBS_STOPPING, pcpu->state))
pcpu              428 arch/x86/events/amd/ibs.c 	stopping = test_bit(IBS_STARTED, pcpu->state);
pcpu              442 arch/x86/events/amd/ibs.c 		set_bit(IBS_STOPPED, pcpu->state);
pcpu              453 arch/x86/events/amd/ibs.c 		clear_bit(IBS_STARTED, pcpu->state);
pcpu              474 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
pcpu              476 arch/x86/events/amd/ibs.c 	if (test_and_set_bit(IBS_ENABLED, pcpu->state))
pcpu              481 arch/x86/events/amd/ibs.c 	pcpu->event = event;
pcpu              492 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
pcpu              494 arch/x86/events/amd/ibs.c 	if (!test_and_clear_bit(IBS_ENABLED, pcpu->state))
pcpu              499 arch/x86/events/amd/ibs.c 	pcpu->event = NULL;
pcpu              571 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
pcpu              572 arch/x86/events/amd/ibs.c 	struct perf_event *event = pcpu->event;
pcpu              582 arch/x86/events/amd/ibs.c 	if (!test_bit(IBS_STARTED, pcpu->state)) {
pcpu              590 arch/x86/events/amd/ibs.c 		if (test_and_clear_bit(IBS_STOPPED, pcpu->state))
pcpu              703 arch/x86/events/amd/ibs.c 	struct cpu_perf_ibs __percpu *pcpu;
pcpu              706 arch/x86/events/amd/ibs.c 	pcpu = alloc_percpu(struct cpu_perf_ibs);
pcpu              707 arch/x86/events/amd/ibs.c 	if (!pcpu)
pcpu              710 arch/x86/events/amd/ibs.c 	perf_ibs->pcpu = pcpu;
pcpu              725 arch/x86/events/amd/ibs.c 		perf_ibs->pcpu = NULL;
pcpu              726 arch/x86/events/amd/ibs.c 		free_percpu(pcpu);
pcpu              750 arch/x86/kernel/acpi/boot.c 		 int *pcpu)
pcpu              763 arch/x86/kernel/acpi/boot.c 	*pcpu = cpu;
pcpu              327 arch/x86/xen/enlighten.c 	pin_override.pcpu = cpu;
pcpu              166 drivers/acpi/acpi_processor.c 		phys_cpuid_t physid, u32 acpi_id, int *pcpu)
pcpu             1085 drivers/base/devres.c 	void __percpu *pcpu;
pcpu             1087 drivers/base/devres.c 	pcpu = __alloc_percpu(size, align);
pcpu             1088 drivers/base/devres.c 	if (!pcpu)
pcpu             1093 drivers/base/devres.c 		free_percpu(pcpu);
pcpu             1097 drivers/base/devres.c 	*(void __percpu **)p = pcpu;
pcpu             1101 drivers/base/devres.c 	return pcpu;
pcpu              432 drivers/crypto/caam/qi.c 		int *pcpu = &get_cpu_var(last_cpu);
pcpu              434 drivers/crypto/caam/qi.c 		*pcpu = cpumask_next(*pcpu, cpus);
pcpu              435 drivers/crypto/caam/qi.c 		if (*pcpu >= nr_cpu_ids)
pcpu              436 drivers/crypto/caam/qi.c 			*pcpu = cpumask_first(cpus);
pcpu              437 drivers/crypto/caam/qi.c 		*cpu = *pcpu;
pcpu              915 drivers/net/ethernet/marvell/mvpp2/mvpp2.h 	struct mvpp2_port_pcpu __percpu *pcpu;
pcpu             1120 drivers/net/ethernet/marvell/mvpp2/mvpp2.h 	struct mvpp2_txq_pcpu __percpu *pcpu;
pcpu             1986 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread);
pcpu             2300 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             2515 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             2552 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             2622 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             3021 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             3039 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             3217 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             3303 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		struct mvpp2_port_pcpu *port_pcpu = per_cpu_ptr(port->pcpu, thread);
pcpu             3743 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			port_pcpu = per_cpu_ptr(port->pcpu, thread);
pcpu             4580 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->pcpu = alloc_percpu(struct mvpp2_txq_pcpu);
pcpu             4581 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (!txq->pcpu) {
pcpu             4590 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
pcpu             4664 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		free_percpu(port->txqs[queue]->pcpu);
pcpu             5353 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	port->pcpu = alloc_percpu(struct mvpp2_port_pcpu);
pcpu             5354 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (!port->pcpu) {
pcpu             5361 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			port_pcpu = per_cpu_ptr(port->pcpu, thread);
pcpu             5426 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	free_percpu(port->pcpu);
pcpu             5429 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		free_percpu(port->txqs[i]->pcpu);
pcpu             5450 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	free_percpu(port->pcpu);
pcpu             5453 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		free_percpu(port->txqs[i]->pcpu);
pcpu               99 drivers/xen/pcpu.c 	struct pcpu *cpu = container_of(dev, struct pcpu, dev);
pcpu              108 drivers/xen/pcpu.c 	struct pcpu *pcpu = container_of(dev, struct pcpu, dev);
pcpu              120 drivers/xen/pcpu.c 		ret = xen_pcpu_down(pcpu->cpu_id);
pcpu              123 drivers/xen/pcpu.c 		ret = xen_pcpu_up(pcpu->cpu_id);
pcpu              168 drivers/xen/pcpu.c 			       struct pcpu *pcpu)
pcpu              171 drivers/xen/pcpu.c 	   !xen_pcpu_online(pcpu->flags)) {
pcpu              173 drivers/xen/pcpu.c 		pcpu->flags |= XEN_PCPU_FLAGS_ONLINE;
pcpu              174 drivers/xen/pcpu.c 		kobject_uevent(&pcpu->dev.kobj, KOBJ_ONLINE);
pcpu              176 drivers/xen/pcpu.c 		    xen_pcpu_online(pcpu->flags)) {
pcpu              178 drivers/xen/pcpu.c 		pcpu->flags &= ~XEN_PCPU_FLAGS_ONLINE;
pcpu              179 drivers/xen/pcpu.c 		kobject_uevent(&pcpu->dev.kobj, KOBJ_OFFLINE);
pcpu              183 drivers/xen/pcpu.c static struct pcpu *get_pcpu(uint32_t cpu_id)
pcpu              185 drivers/xen/pcpu.c 	struct pcpu *pcpu;
pcpu              187 drivers/xen/pcpu.c 	list_for_each_entry(pcpu, &xen_pcpus, list) {
pcpu              188 drivers/xen/pcpu.c 		if (pcpu->cpu_id == cpu_id)
pcpu              189 drivers/xen/pcpu.c 			return pcpu;
pcpu              197 drivers/xen/pcpu.c 	struct pcpu *pcpu = container_of(dev, struct pcpu, dev);
pcpu              199 drivers/xen/pcpu.c 	list_del(&pcpu->list);
pcpu              200 drivers/xen/pcpu.c 	kfree(pcpu);
pcpu              203 drivers/xen/pcpu.c static void unregister_and_remove_pcpu(struct pcpu *pcpu)
pcpu              207 drivers/xen/pcpu.c 	if (!pcpu)
pcpu              210 drivers/xen/pcpu.c 	dev = &pcpu->dev;
pcpu              215 drivers/xen/pcpu.c static int register_pcpu(struct pcpu *pcpu)
pcpu              220 drivers/xen/pcpu.c 	if (!pcpu)
pcpu              223 drivers/xen/pcpu.c 	dev = &pcpu->dev;
pcpu              225 drivers/xen/pcpu.c 	dev->id = pcpu->cpu_id;
pcpu              238 drivers/xen/pcpu.c static struct pcpu *create_and_register_pcpu(struct xenpf_pcpuinfo *info)
pcpu              240 drivers/xen/pcpu.c 	struct pcpu *pcpu;
pcpu              246 drivers/xen/pcpu.c 	pcpu = kzalloc(sizeof(struct pcpu), GFP_KERNEL);
pcpu              247 drivers/xen/pcpu.c 	if (!pcpu)
pcpu              250 drivers/xen/pcpu.c 	INIT_LIST_HEAD(&pcpu->list);
pcpu              251 drivers/xen/pcpu.c 	pcpu->cpu_id = info->xen_cpuid;
pcpu              252 drivers/xen/pcpu.c 	pcpu->flags = info->flags;
pcpu              255 drivers/xen/pcpu.c 	list_add_tail(&pcpu->list, &xen_pcpus);
pcpu              257 drivers/xen/pcpu.c 	err = register_pcpu(pcpu);
pcpu              263 drivers/xen/pcpu.c 	return pcpu;
pcpu              272 drivers/xen/pcpu.c 	struct pcpu *pcpu = NULL;
pcpu              288 drivers/xen/pcpu.c 	pcpu = get_pcpu(cpu);
pcpu              294 drivers/xen/pcpu.c 		unregister_and_remove_pcpu(pcpu);
pcpu              298 drivers/xen/pcpu.c 	if (!pcpu) {
pcpu              299 drivers/xen/pcpu.c 		pcpu = create_and_register_pcpu(info);
pcpu              300 drivers/xen/pcpu.c 		if (IS_ERR_OR_NULL(pcpu))
pcpu              303 drivers/xen/pcpu.c 		pcpu_online_status(info, pcpu);
pcpu              318 drivers/xen/pcpu.c 	struct pcpu *pcpu, *tmp;
pcpu              328 drivers/xen/pcpu.c 		list_for_each_entry_safe(pcpu, tmp, &xen_pcpus, list)
pcpu              329 drivers/xen/pcpu.c 			unregister_and_remove_pcpu(pcpu);
pcpu              286 include/linux/acpi.h 		 int *pcpu);
pcpu              106 include/linux/notifier.h #define SRCU_NOTIFIER_INIT(name, pcpu)				\
pcpu              110 include/linux/notifier.h 		.srcu = __SRCU_STRUCT_INIT(name.srcu, pcpu),	\
pcpu               94 include/linux/padata.h 	cpumask_var_t	pcpu;
pcpu              133 include/linux/psi_types.h 	struct psi_group_cpu __percpu *pcpu;
pcpu              158 include/xen/interface/sched.h     int32_t pcpu;
pcpu               27 kernel/module-internal.h 		unsigned int sym, str, mod, vers, info, pcpu;
pcpu              647 kernel/module.c 	Elf_Shdr *pcpusec = &info->sechdrs[info->index.pcpu];
pcpu              744 kernel/module.c 	if (info->sechdrs[info->index.pcpu].sh_size != 0)
pcpu             2351 kernel/module.c 			if (sym[i].st_shndx == info->index.pcpu)
pcpu             2630 kernel/module.c 	if (sym->st_shndx == SHN_ABS || sym->st_shndx == info->index.pcpu)
pcpu             2711 kernel/module.c 				   info->index.pcpu)) {
pcpu             2775 kernel/module.c 				   info->index.pcpu)) {
pcpu             3103 kernel/module.c 	info->index.pcpu = find_pcpusec(info);
pcpu             3425 kernel/module.c 	info->sechdrs[info->index.pcpu].sh_flags &= ~(unsigned long)SHF_ALLOC;
pcpu             3484 kernel/module.c 	percpu_modcopy(mod, (void *)info->sechdrs[info->index.pcpu].sh_addr,
pcpu             3485 kernel/module.c 		       info->sechdrs[info->index.pcpu].sh_size);
pcpu               44 kernel/padata.c 	target_cpu = cpumask_first(pd->cpumask.pcpu);
pcpu               46 kernel/padata.c 		target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu);
pcpu               57 kernel/padata.c 	int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu);
pcpu              207 kernel/padata.c 		pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false);
pcpu              363 kernel/padata.c 	cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu);
pcpu              376 kernel/padata.c 	if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL))
pcpu              381 kernel/padata.c 	cpumask_copy(pd->cpumask.pcpu, pcpumask);
pcpu              387 kernel/padata.c 	free_cpumask_var(pd->cpumask.pcpu);
pcpu              418 kernel/padata.c 	for_each_cpu(cpu, pd->cpumask.pcpu) {
pcpu              437 kernel/padata.c 	pcpumask = pinst->rcpumask.pcpu;
pcpu              461 kernel/padata.c 	pd->cpu = cpumask_first(pd->cpumask.pcpu);
pcpu              478 kernel/padata.c 	free_cpumask_var(pd->cpumask.pcpu);
pcpu              523 kernel/padata.c 	cpumask_copy(pinst->omask, pinst->rcpumask.pcpu);
pcpu              524 kernel/padata.c 	cpumask_and(pinst->rcpumask.pcpu, pinst->cpumask.pcpu,
pcpu              526 kernel/padata.c 	if (!cpumask_equal(pinst->omask, pinst->rcpumask.pcpu))
pcpu              620 kernel/padata.c 	cpumask_copy(pinst->cpumask.pcpu, pcpumask);
pcpu              655 kernel/padata.c 		parallel_mask = pinst->cpumask.pcpu;
pcpu              716 kernel/padata.c 		if (padata_validate_cpumask(pinst, pinst->cpumask.pcpu) &&
pcpu              729 kernel/padata.c 		if (!padata_validate_cpumask(pinst, pinst->cpumask.pcpu) ||
pcpu              763 kernel/padata.c 		cpumask_clear_cpu(cpu, pinst->cpumask.pcpu);
pcpu              776 kernel/padata.c 	return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) ||
pcpu              826 kernel/padata.c 	free_cpumask_var(pinst->rcpumask.pcpu);
pcpu              827 kernel/padata.c 	free_cpumask_var(pinst->cpumask.pcpu);
pcpu              862 kernel/padata.c 		cpumask = pinst->cpumask.pcpu;
pcpu              990 kernel/padata.c 	if (!alloc_cpumask_var(&pinst->cpumask.pcpu, GFP_KERNEL))
pcpu              993 kernel/padata.c 		free_cpumask_var(pinst->cpumask.pcpu);
pcpu             1000 kernel/padata.c 	if (!alloc_cpumask_var(&pinst->rcpumask.pcpu, GFP_KERNEL))
pcpu             1009 kernel/padata.c 	cpumask_copy(pinst->cpumask.pcpu, pcpumask);
pcpu             1011 kernel/padata.c 	cpumask_and(pinst->rcpumask.pcpu, pcpumask, cpu_online_mask);
pcpu             1039 kernel/padata.c 	free_cpumask_var(pinst->rcpumask.pcpu);
pcpu             1041 kernel/padata.c 	free_cpumask_var(pinst->cpumask.pcpu);
pcpu              177 kernel/sched/psi.c 	.pcpu = &system_group_pcpu,
pcpu              187 kernel/sched/psi.c 		seqcount_init(&per_cpu_ptr(group->pcpu, cpu)->seq);
pcpu              241 kernel/sched/psi.c 	struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
pcpu              680 kernel/sched/psi.c 	groupc = per_cpu_ptr(group->pcpu, cpu);
pcpu              799 kernel/sched/psi.c 		groupc = per_cpu_ptr(group->pcpu, cpu);
pcpu              872 kernel/sched/psi.c 	cgroup->psi.pcpu = alloc_percpu(struct psi_group_cpu);
pcpu              873 kernel/sched/psi.c 	if (!cgroup->psi.pcpu)
pcpu              885 kernel/sched/psi.c 	free_percpu(cgroup->psi.pcpu);
pcpu              298 lib/test_vmalloc.c 	void __percpu **pcpu;
pcpu              302 lib/test_vmalloc.c 	pcpu = vmalloc(sizeof(void __percpu *) * 35000);
pcpu              303 lib/test_vmalloc.c 	if (!pcpu)
pcpu              318 lib/test_vmalloc.c 		pcpu[i] = __alloc_percpu(size, align);
pcpu              319 lib/test_vmalloc.c 		if (!pcpu[i])
pcpu              324 lib/test_vmalloc.c 		free_percpu(pcpu[i]);
pcpu              326 lib/test_vmalloc.c 	vfree(pcpu);
pcpu              245 mm/percpu.c    static void pcpu_set_page_chunk(struct page *page, struct pcpu_chunk *pcpu)
pcpu              247 mm/percpu.c    	page->index = (unsigned long)pcpu;
pcpu             2416 net/ipv4/fib_trie.c 		const struct trie_use_stats *pcpu = per_cpu_ptr(stats, cpu);
pcpu             2418 net/ipv4/fib_trie.c 		s.gets += pcpu->gets;
pcpu             2419 net/ipv4/fib_trie.c 		s.backtrack += pcpu->backtrack;
pcpu             2420 net/ipv4/fib_trie.c 		s.semantic_match_passed += pcpu->semantic_match_passed;
pcpu             2421 net/ipv4/fib_trie.c 		s.semantic_match_miss += pcpu->semantic_match_miss;
pcpu             2422 net/ipv4/fib_trie.c 		s.null_node_hit += pcpu->null_node_hit;
pcpu             2423 net/ipv4/fib_trie.c 		s.resize_node_skipped += pcpu->resize_node_skipped;
pcpu              497 net/netfilter/nf_conntrack_core.c 	struct ct_pcpu *pcpu;
pcpu              501 net/netfilter/nf_conntrack_core.c 	pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu);
pcpu              503 net/netfilter/nf_conntrack_core.c 	spin_lock(&pcpu->lock);
pcpu              505 net/netfilter/nf_conntrack_core.c 			     &pcpu->dying);
pcpu              506 net/netfilter/nf_conntrack_core.c 	spin_unlock(&pcpu->lock);
pcpu              512 net/netfilter/nf_conntrack_core.c 	struct ct_pcpu *pcpu;
pcpu              516 net/netfilter/nf_conntrack_core.c 	pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu);
pcpu              518 net/netfilter/nf_conntrack_core.c 	spin_lock(&pcpu->lock);
pcpu              520 net/netfilter/nf_conntrack_core.c 			     &pcpu->unconfirmed);
pcpu              521 net/netfilter/nf_conntrack_core.c 	spin_unlock(&pcpu->lock);
pcpu              527 net/netfilter/nf_conntrack_core.c 	struct ct_pcpu *pcpu;
pcpu              530 net/netfilter/nf_conntrack_core.c 	pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu);
pcpu              532 net/netfilter/nf_conntrack_core.c 	spin_lock(&pcpu->lock);
pcpu              535 net/netfilter/nf_conntrack_core.c 	spin_unlock(&pcpu->lock);
pcpu             2140 net/netfilter/nf_conntrack_core.c 		struct ct_pcpu *pcpu;
pcpu             2142 net/netfilter/nf_conntrack_core.c 		pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu);
pcpu             2144 net/netfilter/nf_conntrack_core.c 		spin_lock_bh(&pcpu->lock);
pcpu             2145 net/netfilter/nf_conntrack_core.c 		hlist_nulls_for_each_entry(h, n, &pcpu->unconfirmed, hnnode) {
pcpu             2155 net/netfilter/nf_conntrack_core.c 		spin_unlock_bh(&pcpu->lock);
pcpu             2585 net/netfilter/nf_conntrack_core.c 		struct ct_pcpu *pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu);
pcpu             2587 net/netfilter/nf_conntrack_core.c 		spin_lock_init(&pcpu->lock);
pcpu             2588 net/netfilter/nf_conntrack_core.c 		INIT_HLIST_NULLS_HEAD(&pcpu->unconfirmed, UNCONFIRMED_NULLS_VAL);
pcpu             2589 net/netfilter/nf_conntrack_core.c 		INIT_HLIST_NULLS_HEAD(&pcpu->dying, DYING_NULLS_VAL);
pcpu               40 net/netfilter/nf_conntrack_ecache.c static enum retry_state ecache_work_evict_list(struct ct_pcpu *pcpu)
pcpu               48 net/netfilter/nf_conntrack_ecache.c 	spin_lock(&pcpu->lock);
pcpu               50 net/netfilter/nf_conntrack_ecache.c 	hlist_nulls_for_each_entry(h, n, &pcpu->dying, hnnode) {
pcpu               75 net/netfilter/nf_conntrack_ecache.c 	spin_unlock(&pcpu->lock);
pcpu               89 net/netfilter/nf_conntrack_ecache.c 	struct ct_pcpu *pcpu;
pcpu               96 net/netfilter/nf_conntrack_ecache.c 		pcpu = per_cpu_ptr(ctnet->pcpu_lists, cpu);
pcpu               98 net/netfilter/nf_conntrack_ecache.c 		ret = ecache_work_evict_list(pcpu);
pcpu             1414 net/netfilter/nf_conntrack_netlink.c 		struct ct_pcpu *pcpu;
pcpu             1419 net/netfilter/nf_conntrack_netlink.c 		pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu);
pcpu             1420 net/netfilter/nf_conntrack_netlink.c 		spin_lock_bh(&pcpu->lock);
pcpu             1421 net/netfilter/nf_conntrack_netlink.c 		list = dying ? &pcpu->dying : &pcpu->unconfirmed;
pcpu             1443 net/netfilter/nf_conntrack_netlink.c 				spin_unlock_bh(&pcpu->lock);
pcpu             1451 net/netfilter/nf_conntrack_netlink.c 		spin_unlock_bh(&pcpu->lock);