pcpu 242 arch/alpha/kernel/err_common.c cdl_process_console_data_log(int cpu, struct percpu_struct *pcpu) pcpu 245 arch/alpha/kernel/err_common.c (IDENT_ADDR | pcpu->console_data_log_pa); pcpu 256 arch/alpha/kernel/err_common.c pcpu->console_data_log_pa = 0; pcpu 266 arch/alpha/kernel/err_common.c struct percpu_struct *pcpu; pcpu 270 arch/alpha/kernel/err_common.c pcpu = (struct percpu_struct *) pcpu 273 arch/alpha/kernel/err_common.c if (pcpu->console_data_log_pa) pcpu 274 arch/alpha/kernel/err_common.c cdl_process_console_data_log(cpu, pcpu); pcpu 20 arch/arm/common/mcpm_platsmp.c unsigned int *pcpu, unsigned int *pcluster) pcpu 25 arch/arm/common/mcpm_platsmp.c *pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0); pcpu 31 arch/arm/common/mcpm_platsmp.c unsigned int pcpu, pcluster, ret; pcpu 34 arch/arm/common/mcpm_platsmp.c cpu_to_pcpu(cpu, &pcpu, &pcluster); pcpu 37 arch/arm/common/mcpm_platsmp.c __func__, cpu, pcpu, pcluster); pcpu 39 arch/arm/common/mcpm_platsmp.c mcpm_set_entry_vector(pcpu, pcluster, NULL); pcpu 40 arch/arm/common/mcpm_platsmp.c ret = mcpm_cpu_power_up(pcpu, pcluster); pcpu 43 arch/arm/common/mcpm_platsmp.c mcpm_set_entry_vector(pcpu, pcluster, secondary_startup); pcpu 58 arch/arm/common/mcpm_platsmp.c unsigned int pcpu, pcluster; pcpu 60 arch/arm/common/mcpm_platsmp.c cpu_to_pcpu(cpu, &pcpu, &pcluster); pcpu 62 arch/arm/common/mcpm_platsmp.c return !mcpm_wait_for_cpu_powerdown(pcpu, pcluster); pcpu 73 arch/arm/common/mcpm_platsmp.c unsigned int mpidr, pcpu, pcluster; pcpu 75 arch/arm/common/mcpm_platsmp.c pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0); pcpu 77 arch/arm/common/mcpm_platsmp.c mcpm_set_entry_vector(pcpu, pcluster, NULL); pcpu 783 arch/ia64/kernel/acpi.c static int _acpi_map_lsapic(acpi_handle handle, int physid, int *pcpu) pcpu 800 arch/ia64/kernel/acpi.c *pcpu = cpu; pcpu 806 arch/ia64/kernel/acpi.c int *pcpu) pcpu 808 arch/ia64/kernel/acpi.c return _acpi_map_lsapic(handle, physid, pcpu); pcpu 74 arch/powerpc/include/asm/kvm_book3s.h u16 pcpu; pcpu 560 arch/powerpc/include/asm/kvm_ppc.h extern void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu, pcpu 722 arch/powerpc/kvm/book3s_hv.c dt->processor_id = cpu_to_be16(vc->pcpu + vcpu->arch.ptid); pcpu 1163 arch/powerpc/kvm/book3s_hv.c int thr, cpu, pcpu, nthreads; pcpu 1179 arch/powerpc/kvm/book3s_hv.c pcpu = READ_ONCE(v->cpu); pcpu 1180 arch/powerpc/kvm/book3s_hv.c if (pcpu >= 0) pcpu 1181 arch/powerpc/kvm/book3s_hv.c smp_call_function_single(pcpu, do_nothing, NULL, 1); pcpu 2558 arch/powerpc/kvm/book3s_hv.c static void kvmppc_prepare_radix_vcpu(struct kvm_vcpu *vcpu, int pcpu) pcpu 2584 arch/powerpc/kvm/book3s_hv.c if (prev_cpu != pcpu) { pcpu 2587 arch/powerpc/kvm/book3s_hv.c cpu_first_thread_sibling(pcpu)) pcpu 2590 arch/powerpc/kvm/book3s_hv.c nested->prev_cpu[vcpu->arch.nested_vcpu_id] = pcpu; pcpu 2592 arch/powerpc/kvm/book3s_hv.c vcpu->arch.prev_cpu = pcpu; pcpu 2602 arch/powerpc/kvm/book3s_hv.c cpu = vc->pcpu; pcpu 2609 arch/powerpc/kvm/book3s_hv.c vcpu->cpu = vc->pcpu; pcpu 2615 arch/powerpc/kvm/book3s_hv.c tpaca->kvm_hstate.ptid = cpu - vc->pcpu; pcpu 2713 arch/powerpc/kvm/book3s_hv.c vc->pcpu = smp_processor_id(); pcpu 2730 arch/powerpc/kvm/book3s_hv.c lp = &per_cpu(preempted_vcores, vc->pcpu); pcpu 3054 arch/powerpc/kvm/book3s_hv.c int pcpu, thr; pcpu 3110 arch/powerpc/kvm/book3s_hv.c pcpu = smp_processor_id(); pcpu 3122 arch/powerpc/kvm/book3s_hv.c pcpu = smp_processor_id(); pcpu 3126 arch/powerpc/kvm/book3s_hv.c kvmppc_prepare_radix_vcpu(vcpu, pcpu); pcpu 3150 arch/powerpc/kvm/book3s_hv.c kvmppc_release_hwthread(pcpu + i); pcpu 3154 arch/powerpc/kvm/book3s_hv.c kvmppc_clear_host_core(pcpu); pcpu 3200 arch/powerpc/kvm/book3s_hv.c struct paca_struct *paca = paca_ptrs[pcpu + thr]; pcpu 3249 arch/powerpc/kvm/book3s_hv.c pvc->pcpu = pcpu + thr; pcpu 3283 arch/powerpc/kvm/book3s_hv.c kvmppc_ipi_thread(pcpu + thr); pcpu 3343 arch/powerpc/kvm/book3s_hv.c struct paca_struct *paca = paca_ptrs[pcpu + thr]; pcpu 3354 arch/powerpc/kvm/book3s_hv.c kvmppc_set_host_core(pcpu); pcpu 3361 arch/powerpc/kvm/book3s_hv.c kvmppc_release_hwthread(pcpu + i); pcpu 3363 arch/powerpc/kvm/book3s_hv.c kvmppc_ipi_thread(pcpu + i); pcpu 3364 arch/powerpc/kvm/book3s_hv.c cpumask_clear_cpu(pcpu + i, &vc->kvm->arch.cpu_in_guest); pcpu 4040 arch/powerpc/kvm/book3s_hv.c int trap, r, pcpu; pcpu 4077 arch/powerpc/kvm/book3s_hv.c pcpu = smp_processor_id(); pcpu 4078 arch/powerpc/kvm/book3s_hv.c vc->pcpu = pcpu; pcpu 4079 arch/powerpc/kvm/book3s_hv.c kvmppc_prepare_radix_vcpu(vcpu, pcpu); pcpu 4105 arch/powerpc/kvm/book3s_hv.c kvmppc_clear_host_core(pcpu); pcpu 4121 arch/powerpc/kvm/book3s_hv.c kvmppc_check_need_tlb_flush(kvm, pcpu, nested); pcpu 4149 arch/powerpc/kvm/book3s_hv.c kvmppc_set_host_core(pcpu); pcpu 4154 arch/powerpc/kvm/book3s_hv.c cpumask_clear_cpu(pcpu, &kvm->arch.cpu_in_guest); pcpu 274 arch/powerpc/kvm/book3s_hv_builtin.c int cpu = vc->pcpu; pcpu 340 arch/powerpc/kvm/book3s_hv_builtin.c cpu0 = vc->pcpu + ptid - local_paca->kvm_hstate.tid; pcpu 838 arch/powerpc/kvm/book3s_hv_builtin.c void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu, pcpu 850 arch/powerpc/kvm/book3s_hv_builtin.c pcpu = cpu_first_thread_sibling(pcpu); pcpu 857 arch/powerpc/kvm/book3s_hv_builtin.c if (cpumask_test_cpu(pcpu, need_tlb_flush)) { pcpu 861 arch/powerpc/kvm/book3s_hv_builtin.c cpumask_clear_cpu(pcpu, need_tlb_flush); pcpu 64 arch/powerpc/kvm/book3s_hv_rm_mmu.c cpu = local_paca->kvm_hstate.kvm_vcore->pcpu; pcpu 719 arch/powerpc/kvm/book3s_hv_rm_xics.c int pcpu = raw_smp_processor_id(); pcpu 721 arch/powerpc/kvm/book3s_hv_rm_xics.c pcpu = cpu_first_thread_sibling(pcpu); pcpu 723 arch/powerpc/kvm/book3s_hv_rm_xics.c if (state->intr_cpu != pcpu) { pcpu 725 arch/powerpc/kvm/book3s_hv_rm_xics.c xics_opal_set_server(state->host_irq, pcpu); pcpu 72 arch/powerpc/platforms/cell/smp.c unsigned int pcpu; pcpu 79 arch/powerpc/platforms/cell/smp.c pcpu = get_hard_smp_processor_id(lcpu); pcpu 92 arch/powerpc/platforms/cell/smp.c status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, lcpu); pcpu 66 arch/powerpc/platforms/powernv/smp.c unsigned int pcpu; pcpu 75 arch/powerpc/platforms/powernv/smp.c pcpu = get_hard_smp_processor_id(nr); pcpu 89 arch/powerpc/platforms/powernv/smp.c rc = opal_query_cpu_status(pcpu, &status); pcpu 106 arch/powerpc/platforms/powernv/smp.c pr_devel("OPAL: Starting CPU %d (HW 0x%x)...\n", nr, pcpu); pcpu 107 arch/powerpc/platforms/powernv/smp.c rc = opal_start_cpu(pcpu, start_here); pcpu 120 arch/powerpc/platforms/powernv/smp.c " (status %d)...\n", nr, pcpu, status); pcpu 201 arch/powerpc/platforms/pseries/hotplug-cpu.c unsigned int pcpu = get_hard_smp_processor_id(cpu); pcpu 215 arch/powerpc/platforms/pseries/hotplug-cpu.c cpu_status = smp_query_cpu_stopped(pcpu); pcpu 225 arch/powerpc/platforms/pseries/hotplug-cpu.c cpu, pcpu, cpu_status); pcpu 29 arch/powerpc/platforms/pseries/pseries.h int smp_query_cpu_stopped(unsigned int pcpu); pcpu 57 arch/powerpc/platforms/pseries/smp.c int smp_query_cpu_stopped(unsigned int pcpu) pcpu 68 arch/powerpc/platforms/pseries/smp.c status = rtas_call(qcss_tok, 1, 2, &cpu_status, pcpu); pcpu 94 arch/powerpc/platforms/pseries/smp.c unsigned int pcpu; pcpu 101 arch/powerpc/platforms/pseries/smp.c pcpu = get_hard_smp_processor_id(lcpu); pcpu 104 arch/powerpc/platforms/pseries/smp.c if (smp_query_cpu_stopped(pcpu) == QCSS_NOT_STOPPED){ pcpu 123 arch/powerpc/platforms/pseries/smp.c status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, pcpu); pcpu 42 arch/s390/include/asm/smp.h u16 pcpu = stap(); pcpu 45 arch/s390/include/asm/smp.h __pcpu_sigp(pcpu, SIGP_STOP, 0, NULL); pcpu 83 arch/s390/kernel/smp.c static struct pcpu pcpu_devices[NR_CPUS]; pcpu 132 arch/s390/kernel/smp.c static int pcpu_sigp_retry(struct pcpu *pcpu, u8 order, u32 parm) pcpu 137 arch/s390/kernel/smp.c cc = __pcpu_sigp(pcpu->address, order, parm, NULL); pcpu 146 arch/s390/kernel/smp.c static inline int pcpu_stopped(struct pcpu *pcpu) pcpu 150 arch/s390/kernel/smp.c if (__pcpu_sigp(pcpu->address, SIGP_SENSE, pcpu 156 arch/s390/kernel/smp.c static inline int pcpu_running(struct pcpu *pcpu) pcpu 158 arch/s390/kernel/smp.c if (__pcpu_sigp(pcpu->address, SIGP_SENSE_RUNNING, pcpu 168 arch/s390/kernel/smp.c static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address) pcpu 178 arch/s390/kernel/smp.c static void pcpu_ec_call(struct pcpu *pcpu, int ec_bit) pcpu 182 arch/s390/kernel/smp.c if (test_and_set_bit(ec_bit, &pcpu->ec_mask)) pcpu 184 arch/s390/kernel/smp.c order = pcpu_running(pcpu) ? SIGP_EXTERNAL_CALL : SIGP_EMERGENCY_SIGNAL; pcpu 185 arch/s390/kernel/smp.c pcpu->ec_clk = get_tod_clock_fast(); pcpu 186 arch/s390/kernel/smp.c pcpu_sigp_retry(pcpu, order, 0); pcpu 189 arch/s390/kernel/smp.c static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) pcpu 194 arch/s390/kernel/smp.c if (pcpu != &pcpu_devices[0]) { pcpu 195 arch/s390/kernel/smp.c pcpu->lowcore = (struct lowcore *) pcpu 198 arch/s390/kernel/smp.c if (!pcpu->lowcore || !nodat_stack) pcpu 201 arch/s390/kernel/smp.c nodat_stack = pcpu->lowcore->nodat_stack - STACK_INIT_OFFSET; pcpu 206 arch/s390/kernel/smp.c lc = pcpu->lowcore; pcpu 222 arch/s390/kernel/smp.c pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, (u32)(unsigned long) lc); pcpu 230 arch/s390/kernel/smp.c if (pcpu != &pcpu_devices[0]) { pcpu 232 arch/s390/kernel/smp.c free_pages((unsigned long) pcpu->lowcore, LC_ORDER); pcpu 237 arch/s390/kernel/smp.c static void pcpu_free_lowcore(struct pcpu *pcpu) pcpu 241 arch/s390/kernel/smp.c nodat_stack = pcpu->lowcore->nodat_stack - STACK_INIT_OFFSET; pcpu 242 arch/s390/kernel/smp.c async_stack = pcpu->lowcore->async_stack - STACK_INIT_OFFSET; pcpu 243 arch/s390/kernel/smp.c lowcore = (unsigned long) pcpu->lowcore; pcpu 245 arch/s390/kernel/smp.c pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, 0); pcpu 246 arch/s390/kernel/smp.c lowcore_ptr[pcpu - pcpu_devices] = NULL; pcpu 247 arch/s390/kernel/smp.c vdso_free_per_cpu(pcpu->lowcore); pcpu 248 arch/s390/kernel/smp.c nmi_free_per_cpu(pcpu->lowcore); pcpu 250 arch/s390/kernel/smp.c if (pcpu == &pcpu_devices[0]) pcpu 256 arch/s390/kernel/smp.c static void pcpu_prepare_secondary(struct pcpu *pcpu, int cpu) pcpu 258 arch/s390/kernel/smp.c struct lowcore *lc = pcpu->lowcore; pcpu 282 arch/s390/kernel/smp.c static void pcpu_attach_task(struct pcpu *pcpu, struct task_struct *tsk) pcpu 284 arch/s390/kernel/smp.c struct lowcore *lc = pcpu->lowcore; pcpu 299 arch/s390/kernel/smp.c static void pcpu_start_fn(struct pcpu *pcpu, void (*func)(void *), void *data) pcpu 301 arch/s390/kernel/smp.c struct lowcore *lc = pcpu->lowcore; pcpu 307 arch/s390/kernel/smp.c pcpu_sigp_retry(pcpu, SIGP_RESTART, 0); pcpu 318 arch/s390/kernel/smp.c static void __no_sanitize_address pcpu_delegate(struct pcpu *pcpu, pcpu 322 arch/s390/kernel/smp.c struct lowcore *lc = lowcore_ptr[pcpu - pcpu_devices]; pcpu 326 arch/s390/kernel/smp.c if (pcpu->address == source_cpu) pcpu 329 arch/s390/kernel/smp.c pcpu_sigp_retry(pcpu, SIGP_STOP, 0); pcpu 341 arch/s390/kernel/smp.c : : "d" (pcpu->address), "d" (source_cpu), pcpu 372 arch/s390/kernel/smp.c struct pcpu *pcpu; pcpu 375 arch/s390/kernel/smp.c pcpu = pcpu_find_address(cpu_online_mask, stap()); pcpu 376 arch/s390/kernel/smp.c if (!pcpu) pcpu 378 arch/s390/kernel/smp.c pcpu = pcpu_devices + cpumask_first(cpu_online_mask); pcpu 379 arch/s390/kernel/smp.c pcpu_delegate(pcpu, func, data, (unsigned long) restart_stack); pcpu 443 arch/s390/kernel/smp.c struct pcpu *pcpu = pcpu_devices + cpu; pcpu 444 arch/s390/kernel/smp.c set_bit(ec_stop_cpu, &pcpu->ec_mask); pcpu 445 arch/s390/kernel/smp.c while (__pcpu_sigp(pcpu->address, SIGP_EMERGENCY_SIGNAL, pcpu 583 arch/s390/kernel/smp.c struct pcpu *pcpu = pcpu_devices + cpu; pcpu 586 arch/s390/kernel/smp.c pa = __pa(&pcpu->lowcore->floating_pt_save_area); pcpu 587 arch/s390/kernel/smp.c if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_STATUS_AT_ADDRESS, pcpu 592 arch/s390/kernel/smp.c pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK); pcpu 594 arch/s390/kernel/smp.c pa |= pcpu->lowcore->mcesad & MCESA_LC_MASK; pcpu 595 arch/s390/kernel/smp.c if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_ADDITIONAL_STATUS, pcpu 735 arch/s390/kernel/smp.c struct pcpu *pcpu; pcpu 747 arch/s390/kernel/smp.c pcpu = pcpu_devices + cpu; pcpu 748 arch/s390/kernel/smp.c pcpu->address = address + i; pcpu 750 arch/s390/kernel/smp.c pcpu->state = CPU_STATE_CONFIGURED; pcpu 752 arch/s390/kernel/smp.c pcpu->state = CPU_STATE_STANDBY; pcpu 887 arch/s390/kernel/smp.c struct pcpu *pcpu; pcpu 890 arch/s390/kernel/smp.c pcpu = pcpu_devices + cpu; pcpu 891 arch/s390/kernel/smp.c if (pcpu->state != CPU_STATE_CONFIGURED) pcpu 908 arch/s390/kernel/smp.c rc = pcpu_alloc_lowcore(pcpu, cpu); pcpu 911 arch/s390/kernel/smp.c pcpu_prepare_secondary(pcpu, cpu); pcpu 912 arch/s390/kernel/smp.c pcpu_attach_task(pcpu, tidle); pcpu 913 arch/s390/kernel/smp.c pcpu_start_fn(pcpu, smp_start_secondary, NULL); pcpu 950 arch/s390/kernel/smp.c struct pcpu *pcpu; pcpu 953 arch/s390/kernel/smp.c pcpu = pcpu_devices + cpu; pcpu 954 arch/s390/kernel/smp.c while (!pcpu_stopped(pcpu)) pcpu 956 arch/s390/kernel/smp.c pcpu_free_lowcore(pcpu); pcpu 994 arch/s390/kernel/smp.c struct pcpu *pcpu = pcpu_devices; pcpu 997 arch/s390/kernel/smp.c pcpu->state = CPU_STATE_CONFIGURED; pcpu 998 arch/s390/kernel/smp.c pcpu->lowcore = (struct lowcore *)(unsigned long) store_prefix(); pcpu 1041 arch/s390/kernel/smp.c struct pcpu *pcpu; pcpu 1060 arch/s390/kernel/smp.c pcpu = pcpu_devices + cpu; pcpu 1064 arch/s390/kernel/smp.c if (pcpu->state != CPU_STATE_CONFIGURED) pcpu 1066 arch/s390/kernel/smp.c rc = sclp_core_deconfigure(pcpu->address >> smp_cpu_mt_shift); pcpu 1072 arch/s390/kernel/smp.c pcpu[i].state = CPU_STATE_STANDBY; pcpu 1079 arch/s390/kernel/smp.c if (pcpu->state != CPU_STATE_STANDBY) pcpu 1081 arch/s390/kernel/smp.c rc = sclp_core_configure(pcpu->address >> smp_cpu_mt_shift); pcpu 1087 arch/s390/kernel/smp.c pcpu[i].state = CPU_STATE_CONFIGURED; pcpu 92 arch/x86/events/amd/ibs.c struct cpu_perf_ibs __percpu *pcpu; pcpu 396 arch/x86/events/amd/ibs.c struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); pcpu 410 arch/x86/events/amd/ibs.c set_bit(IBS_STARTED, pcpu->state); pcpu 411 arch/x86/events/amd/ibs.c clear_bit(IBS_STOPPING, pcpu->state); pcpu 421 arch/x86/events/amd/ibs.c struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); pcpu 425 arch/x86/events/amd/ibs.c if (test_and_set_bit(IBS_STOPPING, pcpu->state)) pcpu 428 arch/x86/events/amd/ibs.c stopping = test_bit(IBS_STARTED, pcpu->state); pcpu 442 arch/x86/events/amd/ibs.c set_bit(IBS_STOPPED, pcpu->state); pcpu 453 arch/x86/events/amd/ibs.c clear_bit(IBS_STARTED, pcpu->state); pcpu 474 arch/x86/events/amd/ibs.c struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); pcpu 476 arch/x86/events/amd/ibs.c if (test_and_set_bit(IBS_ENABLED, pcpu->state)) pcpu 481 arch/x86/events/amd/ibs.c pcpu->event = event; pcpu 492 arch/x86/events/amd/ibs.c struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); pcpu 494 arch/x86/events/amd/ibs.c if (!test_and_clear_bit(IBS_ENABLED, pcpu->state)) pcpu 499 arch/x86/events/amd/ibs.c pcpu->event = NULL; pcpu 571 arch/x86/events/amd/ibs.c struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); pcpu 572 arch/x86/events/amd/ibs.c struct perf_event *event = pcpu->event; pcpu 582 arch/x86/events/amd/ibs.c if (!test_bit(IBS_STARTED, pcpu->state)) { pcpu 590 arch/x86/events/amd/ibs.c if (test_and_clear_bit(IBS_STOPPED, pcpu->state)) pcpu 703 arch/x86/events/amd/ibs.c struct cpu_perf_ibs __percpu *pcpu; pcpu 706 arch/x86/events/amd/ibs.c pcpu = alloc_percpu(struct cpu_perf_ibs); pcpu 707 arch/x86/events/amd/ibs.c if (!pcpu) pcpu 710 arch/x86/events/amd/ibs.c perf_ibs->pcpu = pcpu; pcpu 725 arch/x86/events/amd/ibs.c perf_ibs->pcpu = NULL; pcpu 726 arch/x86/events/amd/ibs.c free_percpu(pcpu); pcpu 750 arch/x86/kernel/acpi/boot.c int *pcpu) pcpu 763 arch/x86/kernel/acpi/boot.c *pcpu = cpu; pcpu 327 arch/x86/xen/enlighten.c pin_override.pcpu = cpu; pcpu 166 drivers/acpi/acpi_processor.c phys_cpuid_t physid, u32 acpi_id, int *pcpu) pcpu 1085 drivers/base/devres.c void __percpu *pcpu; pcpu 1087 drivers/base/devres.c pcpu = __alloc_percpu(size, align); pcpu 1088 drivers/base/devres.c if (!pcpu) pcpu 1093 drivers/base/devres.c free_percpu(pcpu); pcpu 1097 drivers/base/devres.c *(void __percpu **)p = pcpu; pcpu 1101 drivers/base/devres.c return pcpu; pcpu 432 drivers/crypto/caam/qi.c int *pcpu = &get_cpu_var(last_cpu); pcpu 434 drivers/crypto/caam/qi.c *pcpu = cpumask_next(*pcpu, cpus); pcpu 435 drivers/crypto/caam/qi.c if (*pcpu >= nr_cpu_ids) pcpu 436 drivers/crypto/caam/qi.c *pcpu = cpumask_first(cpus); pcpu 437 drivers/crypto/caam/qi.c *cpu = *pcpu; pcpu 915 drivers/net/ethernet/marvell/mvpp2/mvpp2.h struct mvpp2_port_pcpu __percpu *pcpu; pcpu 1120 drivers/net/ethernet/marvell/mvpp2/mvpp2.h struct mvpp2_txq_pcpu __percpu *pcpu; pcpu 1986 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread); pcpu 2300 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 2515 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 2552 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 2622 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 3021 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 3039 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 3217 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 3303 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_port_pcpu *port_pcpu = per_cpu_ptr(port->pcpu, thread); pcpu 3743 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port_pcpu = per_cpu_ptr(port->pcpu, thread); pcpu 4580 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->pcpu = alloc_percpu(struct mvpp2_txq_pcpu); pcpu 4581 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!txq->pcpu) { pcpu 4590 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); pcpu 4664 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[queue]->pcpu); pcpu 5353 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->pcpu = alloc_percpu(struct mvpp2_port_pcpu); pcpu 5354 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!port->pcpu) { pcpu 5361 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port_pcpu = per_cpu_ptr(port->pcpu, thread); pcpu 5426 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->pcpu); pcpu 5429 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[i]->pcpu); pcpu 5450 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->pcpu); pcpu 5453 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[i]->pcpu); pcpu 99 drivers/xen/pcpu.c struct pcpu *cpu = container_of(dev, struct pcpu, dev); pcpu 108 drivers/xen/pcpu.c struct pcpu *pcpu = container_of(dev, struct pcpu, dev); pcpu 120 drivers/xen/pcpu.c ret = xen_pcpu_down(pcpu->cpu_id); pcpu 123 drivers/xen/pcpu.c ret = xen_pcpu_up(pcpu->cpu_id); pcpu 168 drivers/xen/pcpu.c struct pcpu *pcpu) pcpu 171 drivers/xen/pcpu.c !xen_pcpu_online(pcpu->flags)) { pcpu 173 drivers/xen/pcpu.c pcpu->flags |= XEN_PCPU_FLAGS_ONLINE; pcpu 174 drivers/xen/pcpu.c kobject_uevent(&pcpu->dev.kobj, KOBJ_ONLINE); pcpu 176 drivers/xen/pcpu.c xen_pcpu_online(pcpu->flags)) { pcpu 178 drivers/xen/pcpu.c pcpu->flags &= ~XEN_PCPU_FLAGS_ONLINE; pcpu 179 drivers/xen/pcpu.c kobject_uevent(&pcpu->dev.kobj, KOBJ_OFFLINE); pcpu 183 drivers/xen/pcpu.c static struct pcpu *get_pcpu(uint32_t cpu_id) pcpu 185 drivers/xen/pcpu.c struct pcpu *pcpu; pcpu 187 drivers/xen/pcpu.c list_for_each_entry(pcpu, &xen_pcpus, list) { pcpu 188 drivers/xen/pcpu.c if (pcpu->cpu_id == cpu_id) pcpu 189 drivers/xen/pcpu.c return pcpu; pcpu 197 drivers/xen/pcpu.c struct pcpu *pcpu = container_of(dev, struct pcpu, dev); pcpu 199 drivers/xen/pcpu.c list_del(&pcpu->list); pcpu 200 drivers/xen/pcpu.c kfree(pcpu); pcpu 203 drivers/xen/pcpu.c static void unregister_and_remove_pcpu(struct pcpu *pcpu) pcpu 207 drivers/xen/pcpu.c if (!pcpu) pcpu 210 drivers/xen/pcpu.c dev = &pcpu->dev; pcpu 215 drivers/xen/pcpu.c static int register_pcpu(struct pcpu *pcpu) pcpu 220 drivers/xen/pcpu.c if (!pcpu) pcpu 223 drivers/xen/pcpu.c dev = &pcpu->dev; pcpu 225 drivers/xen/pcpu.c dev->id = pcpu->cpu_id; pcpu 238 drivers/xen/pcpu.c static struct pcpu *create_and_register_pcpu(struct xenpf_pcpuinfo *info) pcpu 240 drivers/xen/pcpu.c struct pcpu *pcpu; pcpu 246 drivers/xen/pcpu.c pcpu = kzalloc(sizeof(struct pcpu), GFP_KERNEL); pcpu 247 drivers/xen/pcpu.c if (!pcpu) pcpu 250 drivers/xen/pcpu.c INIT_LIST_HEAD(&pcpu->list); pcpu 251 drivers/xen/pcpu.c pcpu->cpu_id = info->xen_cpuid; pcpu 252 drivers/xen/pcpu.c pcpu->flags = info->flags; pcpu 255 drivers/xen/pcpu.c list_add_tail(&pcpu->list, &xen_pcpus); pcpu 257 drivers/xen/pcpu.c err = register_pcpu(pcpu); pcpu 263 drivers/xen/pcpu.c return pcpu; pcpu 272 drivers/xen/pcpu.c struct pcpu *pcpu = NULL; pcpu 288 drivers/xen/pcpu.c pcpu = get_pcpu(cpu); pcpu 294 drivers/xen/pcpu.c unregister_and_remove_pcpu(pcpu); pcpu 298 drivers/xen/pcpu.c if (!pcpu) { pcpu 299 drivers/xen/pcpu.c pcpu = create_and_register_pcpu(info); pcpu 300 drivers/xen/pcpu.c if (IS_ERR_OR_NULL(pcpu)) pcpu 303 drivers/xen/pcpu.c pcpu_online_status(info, pcpu); pcpu 318 drivers/xen/pcpu.c struct pcpu *pcpu, *tmp; pcpu 328 drivers/xen/pcpu.c list_for_each_entry_safe(pcpu, tmp, &xen_pcpus, list) pcpu 329 drivers/xen/pcpu.c unregister_and_remove_pcpu(pcpu); pcpu 286 include/linux/acpi.h int *pcpu); pcpu 106 include/linux/notifier.h #define SRCU_NOTIFIER_INIT(name, pcpu) \ pcpu 110 include/linux/notifier.h .srcu = __SRCU_STRUCT_INIT(name.srcu, pcpu), \ pcpu 94 include/linux/padata.h cpumask_var_t pcpu; pcpu 133 include/linux/psi_types.h struct psi_group_cpu __percpu *pcpu; pcpu 158 include/xen/interface/sched.h int32_t pcpu; pcpu 27 kernel/module-internal.h unsigned int sym, str, mod, vers, info, pcpu; pcpu 647 kernel/module.c Elf_Shdr *pcpusec = &info->sechdrs[info->index.pcpu]; pcpu 744 kernel/module.c if (info->sechdrs[info->index.pcpu].sh_size != 0) pcpu 2351 kernel/module.c if (sym[i].st_shndx == info->index.pcpu) pcpu 2630 kernel/module.c if (sym->st_shndx == SHN_ABS || sym->st_shndx == info->index.pcpu) pcpu 2711 kernel/module.c info->index.pcpu)) { pcpu 2775 kernel/module.c info->index.pcpu)) { pcpu 3103 kernel/module.c info->index.pcpu = find_pcpusec(info); pcpu 3425 kernel/module.c info->sechdrs[info->index.pcpu].sh_flags &= ~(unsigned long)SHF_ALLOC; pcpu 3484 kernel/module.c percpu_modcopy(mod, (void *)info->sechdrs[info->index.pcpu].sh_addr, pcpu 3485 kernel/module.c info->sechdrs[info->index.pcpu].sh_size); pcpu 44 kernel/padata.c target_cpu = cpumask_first(pd->cpumask.pcpu); pcpu 46 kernel/padata.c target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); pcpu 57 kernel/padata.c int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); pcpu 207 kernel/padata.c pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false); pcpu 363 kernel/padata.c cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu); pcpu 376 kernel/padata.c if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) pcpu 381 kernel/padata.c cpumask_copy(pd->cpumask.pcpu, pcpumask); pcpu 387 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); pcpu 418 kernel/padata.c for_each_cpu(cpu, pd->cpumask.pcpu) { pcpu 437 kernel/padata.c pcpumask = pinst->rcpumask.pcpu; pcpu 461 kernel/padata.c pd->cpu = cpumask_first(pd->cpumask.pcpu); pcpu 478 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); pcpu 523 kernel/padata.c cpumask_copy(pinst->omask, pinst->rcpumask.pcpu); pcpu 524 kernel/padata.c cpumask_and(pinst->rcpumask.pcpu, pinst->cpumask.pcpu, pcpu 526 kernel/padata.c if (!cpumask_equal(pinst->omask, pinst->rcpumask.pcpu)) pcpu 620 kernel/padata.c cpumask_copy(pinst->cpumask.pcpu, pcpumask); pcpu 655 kernel/padata.c parallel_mask = pinst->cpumask.pcpu; pcpu 716 kernel/padata.c if (padata_validate_cpumask(pinst, pinst->cpumask.pcpu) && pcpu 729 kernel/padata.c if (!padata_validate_cpumask(pinst, pinst->cpumask.pcpu) || pcpu 763 kernel/padata.c cpumask_clear_cpu(cpu, pinst->cpumask.pcpu); pcpu 776 kernel/padata.c return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) || pcpu 826 kernel/padata.c free_cpumask_var(pinst->rcpumask.pcpu); pcpu 827 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); pcpu 862 kernel/padata.c cpumask = pinst->cpumask.pcpu; pcpu 990 kernel/padata.c if (!alloc_cpumask_var(&pinst->cpumask.pcpu, GFP_KERNEL)) pcpu 993 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); pcpu 1000 kernel/padata.c if (!alloc_cpumask_var(&pinst->rcpumask.pcpu, GFP_KERNEL)) pcpu 1009 kernel/padata.c cpumask_copy(pinst->cpumask.pcpu, pcpumask); pcpu 1011 kernel/padata.c cpumask_and(pinst->rcpumask.pcpu, pcpumask, cpu_online_mask); pcpu 1039 kernel/padata.c free_cpumask_var(pinst->rcpumask.pcpu); pcpu 1041 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); pcpu 177 kernel/sched/psi.c .pcpu = &system_group_pcpu, pcpu 187 kernel/sched/psi.c seqcount_init(&per_cpu_ptr(group->pcpu, cpu)->seq); pcpu 241 kernel/sched/psi.c struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu); pcpu 680 kernel/sched/psi.c groupc = per_cpu_ptr(group->pcpu, cpu); pcpu 799 kernel/sched/psi.c groupc = per_cpu_ptr(group->pcpu, cpu); pcpu 872 kernel/sched/psi.c cgroup->psi.pcpu = alloc_percpu(struct psi_group_cpu); pcpu 873 kernel/sched/psi.c if (!cgroup->psi.pcpu) pcpu 885 kernel/sched/psi.c free_percpu(cgroup->psi.pcpu); pcpu 298 lib/test_vmalloc.c void __percpu **pcpu; pcpu 302 lib/test_vmalloc.c pcpu = vmalloc(sizeof(void __percpu *) * 35000); pcpu 303 lib/test_vmalloc.c if (!pcpu) pcpu 318 lib/test_vmalloc.c pcpu[i] = __alloc_percpu(size, align); pcpu 319 lib/test_vmalloc.c if (!pcpu[i]) pcpu 324 lib/test_vmalloc.c free_percpu(pcpu[i]); pcpu 326 lib/test_vmalloc.c vfree(pcpu); pcpu 245 mm/percpu.c static void pcpu_set_page_chunk(struct page *page, struct pcpu_chunk *pcpu) pcpu 247 mm/percpu.c page->index = (unsigned long)pcpu; pcpu 2416 net/ipv4/fib_trie.c const struct trie_use_stats *pcpu = per_cpu_ptr(stats, cpu); pcpu 2418 net/ipv4/fib_trie.c s.gets += pcpu->gets; pcpu 2419 net/ipv4/fib_trie.c s.backtrack += pcpu->backtrack; pcpu 2420 net/ipv4/fib_trie.c s.semantic_match_passed += pcpu->semantic_match_passed; pcpu 2421 net/ipv4/fib_trie.c s.semantic_match_miss += pcpu->semantic_match_miss; pcpu 2422 net/ipv4/fib_trie.c s.null_node_hit += pcpu->null_node_hit; pcpu 2423 net/ipv4/fib_trie.c s.resize_node_skipped += pcpu->resize_node_skipped; pcpu 497 net/netfilter/nf_conntrack_core.c struct ct_pcpu *pcpu; pcpu 501 net/netfilter/nf_conntrack_core.c pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu); pcpu 503 net/netfilter/nf_conntrack_core.c spin_lock(&pcpu->lock); pcpu 505 net/netfilter/nf_conntrack_core.c &pcpu->dying); pcpu 506 net/netfilter/nf_conntrack_core.c spin_unlock(&pcpu->lock); pcpu 512 net/netfilter/nf_conntrack_core.c struct ct_pcpu *pcpu; pcpu 516 net/netfilter/nf_conntrack_core.c pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu); pcpu 518 net/netfilter/nf_conntrack_core.c spin_lock(&pcpu->lock); pcpu 520 net/netfilter/nf_conntrack_core.c &pcpu->unconfirmed); pcpu 521 net/netfilter/nf_conntrack_core.c spin_unlock(&pcpu->lock); pcpu 527 net/netfilter/nf_conntrack_core.c struct ct_pcpu *pcpu; pcpu 530 net/netfilter/nf_conntrack_core.c pcpu = per_cpu_ptr(nf_ct_net(ct)->ct.pcpu_lists, ct->cpu); pcpu 532 net/netfilter/nf_conntrack_core.c spin_lock(&pcpu->lock); pcpu 535 net/netfilter/nf_conntrack_core.c spin_unlock(&pcpu->lock); pcpu 2140 net/netfilter/nf_conntrack_core.c struct ct_pcpu *pcpu; pcpu 2142 net/netfilter/nf_conntrack_core.c pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu); pcpu 2144 net/netfilter/nf_conntrack_core.c spin_lock_bh(&pcpu->lock); pcpu 2145 net/netfilter/nf_conntrack_core.c hlist_nulls_for_each_entry(h, n, &pcpu->unconfirmed, hnnode) { pcpu 2155 net/netfilter/nf_conntrack_core.c spin_unlock_bh(&pcpu->lock); pcpu 2585 net/netfilter/nf_conntrack_core.c struct ct_pcpu *pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu); pcpu 2587 net/netfilter/nf_conntrack_core.c spin_lock_init(&pcpu->lock); pcpu 2588 net/netfilter/nf_conntrack_core.c INIT_HLIST_NULLS_HEAD(&pcpu->unconfirmed, UNCONFIRMED_NULLS_VAL); pcpu 2589 net/netfilter/nf_conntrack_core.c INIT_HLIST_NULLS_HEAD(&pcpu->dying, DYING_NULLS_VAL); pcpu 40 net/netfilter/nf_conntrack_ecache.c static enum retry_state ecache_work_evict_list(struct ct_pcpu *pcpu) pcpu 48 net/netfilter/nf_conntrack_ecache.c spin_lock(&pcpu->lock); pcpu 50 net/netfilter/nf_conntrack_ecache.c hlist_nulls_for_each_entry(h, n, &pcpu->dying, hnnode) { pcpu 75 net/netfilter/nf_conntrack_ecache.c spin_unlock(&pcpu->lock); pcpu 89 net/netfilter/nf_conntrack_ecache.c struct ct_pcpu *pcpu; pcpu 96 net/netfilter/nf_conntrack_ecache.c pcpu = per_cpu_ptr(ctnet->pcpu_lists, cpu); pcpu 98 net/netfilter/nf_conntrack_ecache.c ret = ecache_work_evict_list(pcpu); pcpu 1414 net/netfilter/nf_conntrack_netlink.c struct ct_pcpu *pcpu; pcpu 1419 net/netfilter/nf_conntrack_netlink.c pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu); pcpu 1420 net/netfilter/nf_conntrack_netlink.c spin_lock_bh(&pcpu->lock); pcpu 1421 net/netfilter/nf_conntrack_netlink.c list = dying ? &pcpu->dying : &pcpu->unconfirmed; pcpu 1443 net/netfilter/nf_conntrack_netlink.c spin_unlock_bh(&pcpu->lock); pcpu 1451 net/netfilter/nf_conntrack_netlink.c spin_unlock_bh(&pcpu->lock);