percpu 72 arch/alpha/boot/bootp.c struct percpu_struct * percpu; percpu 102 arch/alpha/boot/bootp.c percpu = (struct percpu_struct *) percpu 104 arch/alpha/boot/bootp.c rev = percpu->pal_revision = percpu->palcode_avail[2]; percpu 120 arch/alpha/boot/bootpz.c struct percpu_struct * percpu; percpu 150 arch/alpha/boot/bootpz.c percpu = (struct percpu_struct *) percpu 152 arch/alpha/boot/bootpz.c rev = percpu->pal_revision = percpu->palcode_avail[2]; percpu 66 arch/alpha/boot/main.c struct percpu_struct * percpu; percpu 96 arch/alpha/boot/main.c percpu = (struct percpu_struct *) percpu 98 arch/alpha/boot/main.c rev = percpu->pal_revision = percpu->palcode_avail[2]; percpu 526 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep, pte_t pte, int percpu) percpu 535 arch/powerpc/include/asm/book3s/32/pgtable.h if (percpu) percpu 550 arch/powerpc/include/asm/book3s/32/pgtable.h if (percpu) { percpu 225 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, pte_t pte, int percpu) percpu 816 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, pte_t pte, int percpu) percpu 819 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__set_pte_at(mm, addr, ptep, pte, percpu); percpu 820 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__set_pte_at(mm, addr, ptep, pte, percpu); percpu 191 arch/powerpc/include/asm/book3s/64/radix.h pte_t *ptep, pte_t pte, int percpu) percpu 193 arch/powerpc/include/asm/nohash/pgtable.h pte_t *ptep, pte_t pte, int percpu) percpu 200 arch/powerpc/include/asm/nohash/pgtable.h if (IS_ENABLED(CONFIG_PPC32) && IS_ENABLED(CONFIG_PTE_64BIT) && !percpu) { percpu 108 arch/sparc/kernel/sun4m_irq.c bool percpu; percpu 201 arch/sparc/kernel/sun4m_irq.c if (handler_data->percpu) { percpu 220 arch/sparc/kernel/sun4m_irq.c if (handler_data->percpu) { percpu 279 arch/sparc/kernel/sun4m_irq.c handler_data->percpu = real_irq < OBP_INT_LEVEL_ONBOARD; percpu 28 drivers/clocksource/timer-of.c if (of_irq->percpu) percpu 72 drivers/clocksource/timer-of.c ret = of_irq->percpu ? percpu 14 drivers/clocksource/timer-of.h int percpu; percpu 153 drivers/clocksource/timer-qcom.c bool percpu) percpu 159 drivers/clocksource/timer-qcom.c msm_timer_has_ppi = percpu; percpu 167 drivers/clocksource/timer-qcom.c if (percpu) percpu 38 drivers/md/raid5-log.h ops_run_partial_parity(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 158 drivers/md/raid5-ppl.c ops_run_partial_parity(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 162 drivers/md/raid5-ppl.c struct page **srcs = percpu->scribble; percpu 1392 drivers/md/raid5.c static struct page **to_addr_page(struct raid5_percpu *percpu, int i) percpu 1394 drivers/md/raid5.c return percpu->scribble + i * percpu->scribble_obj_size; percpu 1399 drivers/md/raid5.c struct raid5_percpu *percpu, int i) percpu 1401 drivers/md/raid5.c return (void *) (to_addr_page(percpu, i) + sh->disks + 2); percpu 1405 drivers/md/raid5.c ops_run_compute5(struct stripe_head *sh, struct raid5_percpu *percpu) percpu 1408 drivers/md/raid5.c struct page **xor_srcs = to_addr_page(percpu, 0); percpu 1430 drivers/md/raid5.c ops_complete_compute, sh, to_addr_conv(sh, percpu, 0)); percpu 1487 drivers/md/raid5.c ops_run_compute6_1(struct stripe_head *sh, struct raid5_percpu *percpu) percpu 1490 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); percpu 1524 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1537 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1545 drivers/md/raid5.c ops_run_compute6_2(struct stripe_head *sh, struct raid5_percpu *percpu) percpu 1556 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); percpu 1599 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1623 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1630 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1637 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 1668 drivers/md/raid5.c ops_run_prexor5(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 1672 drivers/md/raid5.c struct page **xor_srcs = to_addr_page(percpu, 0); percpu 1693 drivers/md/raid5.c ops_complete_prexor, sh, to_addr_conv(sh, percpu, 0)); percpu 1700 drivers/md/raid5.c ops_run_prexor6(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 1703 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); percpu 1713 drivers/md/raid5.c ops_complete_prexor, sh, to_addr_conv(sh, percpu, 0)); percpu 1838 drivers/md/raid5.c ops_run_reconstruct5(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 1869 drivers/md/raid5.c xor_srcs = to_addr_page(percpu, j); percpu 1905 drivers/md/raid5.c to_addr_conv(sh, percpu, j)); percpu 1909 drivers/md/raid5.c to_addr_conv(sh, percpu, j)); percpu 1925 drivers/md/raid5.c ops_run_reconstruct6(struct stripe_head *sh, struct raid5_percpu *percpu, percpu 1953 drivers/md/raid5.c blocks = to_addr_page(percpu, j); percpu 1971 drivers/md/raid5.c head_sh, to_addr_conv(sh, percpu, j)); percpu 1974 drivers/md/raid5.c to_addr_conv(sh, percpu, j)); percpu 1996 drivers/md/raid5.c static void ops_run_check_p(struct stripe_head *sh, struct raid5_percpu *percpu) percpu 2002 drivers/md/raid5.c struct page **xor_srcs = to_addr_page(percpu, 0); percpu 2022 drivers/md/raid5.c to_addr_conv(sh, percpu, 0)); percpu 2031 drivers/md/raid5.c static void ops_run_check_pq(struct stripe_head *sh, struct raid5_percpu *percpu, int checkp) percpu 2033 drivers/md/raid5.c struct page **srcs = to_addr_page(percpu, 0); percpu 2047 drivers/md/raid5.c sh, to_addr_conv(sh, percpu, 0)); percpu 2049 drivers/md/raid5.c &sh->ops.zero_sum_result, percpu->spare_page, &submit); percpu 2058 drivers/md/raid5.c struct raid5_percpu *percpu; percpu 2062 drivers/md/raid5.c percpu = per_cpu_ptr(conf->percpu, cpu); percpu 2070 drivers/md/raid5.c tx = ops_run_compute5(sh, percpu); percpu 2073 drivers/md/raid5.c tx = ops_run_compute6_1(sh, percpu); percpu 2075 drivers/md/raid5.c tx = ops_run_compute6_2(sh, percpu); percpu 2084 drivers/md/raid5.c tx = ops_run_prexor5(sh, percpu, tx); percpu 2086 drivers/md/raid5.c tx = ops_run_prexor6(sh, percpu, tx); percpu 2090 drivers/md/raid5.c tx = ops_run_partial_parity(sh, percpu, tx); percpu 2099 drivers/md/raid5.c ops_run_reconstruct5(sh, percpu, tx); percpu 2101 drivers/md/raid5.c ops_run_reconstruct6(sh, percpu, tx); percpu 2106 drivers/md/raid5.c ops_run_check_p(sh, percpu); percpu 2108 drivers/md/raid5.c ops_run_check_pq(sh, percpu, 0); percpu 2110 drivers/md/raid5.c ops_run_check_pq(sh, percpu, 1); percpu 2230 drivers/md/raid5.c static int scribble_alloc(struct raid5_percpu *percpu, percpu 2242 drivers/md/raid5.c kvfree(percpu->scribble); percpu 2244 drivers/md/raid5.c percpu->scribble = scribble; percpu 2245 drivers/md/raid5.c percpu->scribble_obj_size = obj_size; percpu 2266 drivers/md/raid5.c struct raid5_percpu *percpu; percpu 2268 drivers/md/raid5.c percpu = per_cpu_ptr(conf->percpu, cpu); percpu 2269 drivers/md/raid5.c err = scribble_alloc(percpu, new_disks, percpu 6747 drivers/md/raid5.c static void free_scratch_buffer(struct r5conf *conf, struct raid5_percpu *percpu) percpu 6749 drivers/md/raid5.c safe_put_page(percpu->spare_page); percpu 6750 drivers/md/raid5.c percpu->spare_page = NULL; percpu 6751 drivers/md/raid5.c kvfree(percpu->scribble); percpu 6752 drivers/md/raid5.c percpu->scribble = NULL; percpu 6755 drivers/md/raid5.c static int alloc_scratch_buffer(struct r5conf *conf, struct raid5_percpu *percpu) percpu 6757 drivers/md/raid5.c if (conf->level == 6 && !percpu->spare_page) { percpu 6758 drivers/md/raid5.c percpu->spare_page = alloc_page(GFP_KERNEL); percpu 6759 drivers/md/raid5.c if (!percpu->spare_page) percpu 6763 drivers/md/raid5.c if (scribble_alloc(percpu, percpu 6770 drivers/md/raid5.c free_scratch_buffer(conf, percpu); percpu 6781 drivers/md/raid5.c free_scratch_buffer(conf, per_cpu_ptr(conf->percpu, cpu)); percpu 6787 drivers/md/raid5.c if (!conf->percpu) percpu 6791 drivers/md/raid5.c free_percpu(conf->percpu); percpu 6817 drivers/md/raid5.c struct raid5_percpu *percpu = per_cpu_ptr(conf->percpu, cpu); percpu 6819 drivers/md/raid5.c if (alloc_scratch_buffer(conf, percpu)) { percpu 6831 drivers/md/raid5.c conf->percpu = alloc_percpu(struct raid5_percpu); percpu 6832 drivers/md/raid5.c if (!conf->percpu) percpu 643 drivers/md/raid5.h } __percpu *percpu; percpu 3840 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static int mvpp2_bm_switch_buffers(struct mvpp2 *priv, bool percpu) percpu 3861 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c priv->percpu_pools = percpu; percpu 962 drivers/xen/events/events_base.c int bind_virq_to_irq(unsigned int virq, unsigned int cpu, bool percpu) percpu 976 drivers/xen/events/events_base.c if (percpu) percpu 30 fs/squashfs/decompressor_multi_percpu.c struct squashfs_stream __percpu *percpu; percpu 33 fs/squashfs/decompressor_multi_percpu.c percpu = alloc_percpu(struct squashfs_stream); percpu 34 fs/squashfs/decompressor_multi_percpu.c if (percpu == NULL) percpu 38 fs/squashfs/decompressor_multi_percpu.c stream = per_cpu_ptr(percpu, cpu); percpu 47 fs/squashfs/decompressor_multi_percpu.c return (__force void *) percpu; percpu 51 fs/squashfs/decompressor_multi_percpu.c stream = per_cpu_ptr(percpu, cpu); percpu 55 fs/squashfs/decompressor_multi_percpu.c free_percpu(percpu); percpu 61 fs/squashfs/decompressor_multi_percpu.c struct squashfs_stream __percpu *percpu = percpu 68 fs/squashfs/decompressor_multi_percpu.c stream = per_cpu_ptr(percpu, cpu); percpu 71 fs/squashfs/decompressor_multi_percpu.c free_percpu(percpu); percpu 78 fs/squashfs/decompressor_multi_percpu.c struct squashfs_stream __percpu *percpu = percpu 80 fs/squashfs/decompressor_multi_percpu.c struct squashfs_stream *stream = get_cpu_ptr(percpu); percpu 863 include/asm-generic/vmlinux.lds.h *(.data..percpu..decrypted) \ percpu 901 include/asm-generic/vmlinux.lds.h *(.data..percpu..first) \ percpu 903 include/asm-generic/vmlinux.lds.h *(.data..percpu..page_aligned) \ percpu 905 include/asm-generic/vmlinux.lds.h *(.data..percpu..read_mostly) \ percpu 907 include/asm-generic/vmlinux.lds.h *(.data..percpu) \ percpu 908 include/asm-generic/vmlinux.lds.h *(.data..percpu..shared_aligned) \ percpu 938 include/asm-generic/vmlinux.lds.h .data..percpu vaddr : AT(__per_cpu_load - LOAD_OFFSET) { \ percpu 941 include/asm-generic/vmlinux.lds.h . = __per_cpu_load + SIZEOF(.data..percpu); percpu 957 include/asm-generic/vmlinux.lds.h .data..percpu : AT(ADDR(.data..percpu) - LOAD_OFFSET) { \ percpu 445 include/linux/module.h void __percpu *percpu; percpu 118 include/linux/stm.h unsigned int percpu; percpu 3 include/trace/events/percpu.h #define TRACE_SYSTEM percpu percpu 22 include/xen/events.h int bind_virq_to_irq(unsigned int virq, unsigned int cpu, bool percpu); percpu 51 kernel/bpf/arraymap.c bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; percpu 59 kernel/bpf/arraymap.c (percpu && numa_node != NUMA_NO_NODE)) percpu 73 kernel/bpf/arraymap.c bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; percpu 105 kernel/bpf/arraymap.c if (percpu) percpu 112 kernel/bpf/arraymap.c if (percpu) percpu 133 kernel/bpf/arraymap.c if (percpu && bpf_array_alloc_percpu(array)) { percpu 496 kernel/bpf/bpf_lru_list.c if (lru->percpu) percpu 552 kernel/bpf/bpf_lru_list.c if (lru->percpu) percpu 610 kernel/bpf/bpf_lru_list.c if (lru->percpu) percpu 645 kernel/bpf/bpf_lru_list.c int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, percpu 650 kernel/bpf/bpf_lru_list.c if (percpu) { percpu 680 kernel/bpf/bpf_lru_list.c lru->percpu = percpu; percpu 690 kernel/bpf/bpf_lru_list.c if (lru->percpu) percpu 61 kernel/bpf/bpf_lru_list.h bool percpu; percpu 73 kernel/bpf/bpf_lru_list.h int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, percpu 228 kernel/bpf/hashtab.c bool percpu = (attr->map_type == BPF_MAP_TYPE_PERCPU_HASH || percpu 267 kernel/bpf/hashtab.c if (numa_node != NUMA_NO_NODE && (percpu || percpu_lru)) percpu 297 kernel/bpf/hashtab.c bool percpu = (attr->map_type == BPF_MAP_TYPE_PERCPU_HASH || percpu 335 kernel/bpf/hashtab.c if (percpu) percpu 349 kernel/bpf/hashtab.c if (percpu) percpu 382 kernel/bpf/hashtab.c if (!percpu && !lru) { percpu 723 kernel/bpf/hashtab.c bool percpu, bool onallcpus, percpu 769 kernel/bpf/hashtab.c if (percpu) { percpu 642 kernel/module.c return mod->percpu; percpu 659 kernel/module.c mod->percpu = __alloc_reserved_percpu(pcpusec->sh_size, align); percpu 660 kernel/module.c if (!mod->percpu) { percpu 671 kernel/module.c free_percpu(mod->percpu); percpu 685 kernel/module.c memcpy(per_cpu_ptr(mod->percpu, cpu), from, size); percpu 701 kernel/module.c void *start = per_cpu_ptr(mod->percpu, cpu); percpu 708 kernel/module.c per_cpu_ptr(mod->percpu, percpu 207 kernel/sched/cpuacct.c u64 percpu; percpu 211 kernel/sched/cpuacct.c percpu = cpuacct_cpuusage_read(ca, i, index); percpu 212 kernel/sched/cpuacct.c seq_printf(m, "%llu ", (unsigned long long) percpu); percpu 54 net/rds/ib.h struct rds_ib_cache_head __percpu *percpu; percpu 106 net/rds/ib_recv.c cache->percpu = alloc_percpu_gfp(struct rds_ib_cache_head, gfp); percpu 107 net/rds/ib_recv.c if (!cache->percpu) percpu 111 net/rds/ib_recv.c head = per_cpu_ptr(cache->percpu, cpu); percpu 129 net/rds/ib_recv.c free_percpu(ic->i_cache_incs.percpu); percpu 142 net/rds/ib_recv.c head = per_cpu_ptr(cache->percpu, cpu); percpu 165 net/rds/ib_recv.c free_percpu(ic->i_cache_incs.percpu); percpu 176 net/rds/ib_recv.c free_percpu(ic->i_cache_frags.percpu); percpu 485 net/rds/ib_recv.c chpfirst = __this_cpu_read(cache->percpu->first); percpu 491 net/rds/ib_recv.c __this_cpu_write(cache->percpu->first, new_item); percpu 492 net/rds/ib_recv.c __this_cpu_inc(cache->percpu->count); percpu 494 net/rds/ib_recv.c if (__this_cpu_read(cache->percpu->count) < RDS_IB_RECYCLE_BATCH_COUNT) percpu 511 net/rds/ib_recv.c __this_cpu_write(cache->percpu->first, NULL); percpu 512 net/rds/ib_recv.c __this_cpu_write(cache->percpu->count, 0); percpu 498 tools/testing/selftests/bpf/test_verifier.c static int create_cgroup_storage(bool percpu) percpu 500 tools/testing/selftests/bpf/test_verifier.c enum bpf_map_type type = percpu ? BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE :