cpumask 49 arch/alpha/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 27 arch/alpha/include/asm/topology.h extern struct cpumask node_to_cpumask_map[]; cpumask 29 arch/alpha/include/asm/topology.h static const struct cpumask *cpumask_of_node(int node) cpumask 84 arch/alpha/kernel/setup.c struct cpumask node_to_cpumask_map[MAX_NUMNODES] __read_mostly; cpumask 177 arch/alpha/kernel/smp.c wait_for_txrdy (unsigned long cpumask) cpumask 181 arch/alpha/kernel/smp.c if (!(hwrpb->txrdy & cpumask)) cpumask 186 arch/alpha/kernel/smp.c if (!(hwrpb->txrdy & cpumask)) cpumask 204 arch/alpha/kernel/smp.c unsigned long cpumask; cpumask 212 arch/alpha/kernel/smp.c cpumask = (1UL << cpuid); cpumask 213 arch/alpha/kernel/smp.c if (wait_for_txrdy(cpumask)) cpumask 226 arch/alpha/kernel/smp.c if (wait_for_txrdy(cpumask)) cpumask 510 arch/alpha/kernel/smp.c send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation) cpumask 597 arch/alpha/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 152 arch/alpha/kernel/sys_dp264.c dp264_set_affinity(struct irq_data *d, const struct cpumask *affinity, cpumask 164 arch/alpha/kernel/sys_dp264.c clipper_set_affinity(struct irq_data *d, const struct cpumask *affinity, cpumask 150 arch/alpha/kernel/sys_titan.c titan_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, cpumask 125 arch/alpha/kernel/time.c .cpumask = cpumask_of(cpu), cpumask 192 arch/alpha/kernel/time.c .cpumask = cpumask_of(cpu), cpumask 18 arch/arc/include/asm/smp.h struct cpumask; cpumask 24 arch/arc/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 266 arch/arc/kernel/mcip.c idu_irq_set_affinity(struct irq_data *data, const struct cpumask *cpumask, cpumask 275 arch/arc/kernel/mcip.c if (!cpumask_and(&online, cpumask, cpu_online_mask)) cpumask 48 arch/arc/kernel/smp.c static int __init arc_get_cpu_map(const char *name, struct cpumask *cpumask) cpumask 57 arch/arc/kernel/smp.c if (cpulist_parse(buf, cpumask)) cpumask 69 arch/arc/kernel/smp.c struct cpumask cpumask; cpumask 71 arch/arc/kernel/smp.c if (arc_get_cpu_map("possible-cpus", &cpumask)) { cpumask 75 arch/arc/kernel/smp.c cpumask_setall(&cpumask); cpumask 78 arch/arc/kernel/smp.c if (!cpumask_test_cpu(0, &cpumask)) cpumask 81 arch/arc/kernel/smp.c init_cpu_possible(&cpumask); cpumask 298 arch/arc/kernel/smp.c static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) cpumask 313 arch/arc/kernel/smp.c struct cpumask targets; cpumask 324 arch/arc/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 21 arch/arc/plat-eznps/smp.c static int __init eznps_get_map(const char *name, struct cpumask *cpumask) cpumask 30 arch/arc/plat-eznps/smp.c cpulist_parse(buf, cpumask); cpumask 38 arch/arc/plat-eznps/smp.c struct cpumask cpumask; cpumask 40 arch/arc/plat-eznps/smp.c if (eznps_get_map("present-cpus", &cpumask)) { cpumask 44 arch/arc/plat-eznps/smp.c init_cpu_present(&cpumask); cpumask 46 arch/arc/plat-eznps/smp.c if (eznps_get_map("possible-cpus", &cpumask)) { cpumask 50 arch/arc/plat-eznps/smp.c init_cpu_possible(&cpumask); cpumask 46 arch/arm/include/asm/smp.h extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); cpumask 75 arch/arm/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 76 arch/arm/include/asm/smp.h extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask); cpumask 505 arch/arm/kernel/smp.c static void (*__smp_cross_call)(const struct cpumask *, unsigned int); cpumask 507 arch/arm/kernel/smp.c void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) cpumask 524 arch/arm/kernel/smp.c static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) cpumask 556 arch/arm/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 561 arch/arm/kernel/smp.c void arch_send_wakeup_ipi_mask(const struct cpumask *mask) cpumask 580 arch/arm/kernel/smp.c void tick_broadcast(const struct cpumask *mask) cpumask 711 arch/arm/kernel/smp.c struct cpumask mask; cpumask 761 arch/arm/kernel/smp.c struct cpumask *cpus = freq->policy->cpus; cpumask 253 arch/arm/kernel/smp_twd.c clk->cpumask = cpumask_of(cpu); cpumask 184 arch/arm/kernel/topology.c const struct cpumask *cpu_corepower_mask(int cpu) cpumask 163 arch/arm/mach-cns3xxx/core.c .cpumask = cpu_all_mask, cpumask 394 arch/arm/mach-davinci/time.c clockevent_davinci.cpumask = cpumask_of(0); cpumask 123 arch/arm/mach-footbridge/dc21285-timer.c ce->cpumask = cpumask_of(smp_processor_id()); cpumask 131 arch/arm/mach-imx/mmdc.c __ATTR(cpumask, S_IRUGO, mmdc_pmu_cpumask_show, NULL); cpumask 172 arch/arm/mach-iop32x/time.c iop_clockevent.cpumask = cpumask_of(0); cpumask 190 arch/arm/mach-mmp/time.c ckevt.cpumask = cpumask_of(0); cpumask 169 arch/arm/mach-omap1/time.c clockevent_mpu_timer1.cpumask = cpumask_of(0); cpumask 161 arch/arm/mach-omap1/timer32k.c clockevent_32k_timer.cpumask = cpumask_of(0); cpumask 360 arch/arm/mach-omap2/timer.c void tick_broadcast(const struct cpumask *mask) cpumask 390 arch/arm/mach-omap2/timer.c clockevent_gpt.cpumask = cpu_possible_mask; cpumask 200 arch/arm/mach-spear/time.c clkevt.cpumask = cpumask_of(0); cpumask 40 arch/arm/mach-vexpress/dcscb.c unsigned int rst_hold, cpumask = (1 << cpu); cpumask 43 arch/arm/mach-vexpress/dcscb.c if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) cpumask 47 arch/arm/mach-vexpress/dcscb.c rst_hold &= ~(cpumask | (cpumask << 4)); cpumask 397 arch/arm/mm/cache-l2x0-pmu.c __ATTR(cpumask, S_IRUGO, l2x0_pmu_cpumask_show, NULL); cpumask 240 arch/arm/plat-orion/time.c orion_clkevt.cpumask = cpumask_of(0); cpumask 23 arch/arm64/include/asm/numa.h const struct cpumask *cpumask_of_node(int node); cpumask 26 arch/arm64/include/asm/numa.h static inline const struct cpumask *cpumask_of_node(int node) cpumask 72 arch/arm64/include/asm/smp.h extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); cpumask 74 arch/arm64/include/asm/smp.h extern void (*__smp_cross_call)(const struct cpumask *, unsigned int); cpumask 98 arch/arm64/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 101 arch/arm64/include/asm/smp.h extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask); cpumask 103 arch/arm64/include/asm/smp.h static inline void arch_send_wakeup_ipi_mask(const struct cpumask *mask) cpumask 756 arch/arm64/kernel/smp.c void (*__smp_cross_call)(const struct cpumask *, unsigned int); cpumask 758 arch/arm64/kernel/smp.c void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) cpumask 774 arch/arm64/kernel/smp.c static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) cpumask 805 arch/arm64/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 816 arch/arm64/kernel/smp.c void arch_send_wakeup_ipi_mask(const struct cpumask *mask) cpumask 952 arch/arm64/kernel/smp.c void tick_broadcast(const struct cpumask *mask) cpumask 47 arch/arm64/mm/numa.c const struct cpumask *cpumask_of_node(int node) cpumask 238 arch/c6x/platforms/timer64.c cd->cpumask = cpumask_of(smp_processor_id()); cpumask 16 arch/csky/include/asm/smp.h void arch_send_call_function_ipi_mask(struct cpumask *mask); cpumask 20 arch/csky/include/asm/smp.h void __init set_send_ipi(void (*func)(const struct cpumask *mask), int irq); cpumask 62 arch/csky/kernel/smp.c static void (*send_arch_ipi)(const struct cpumask *mask); cpumask 65 arch/csky/kernel/smp.c void __init set_send_ipi(void (*func)(const struct cpumask *mask), int irq) cpumask 75 arch/csky/kernel/smp.c send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation) cpumask 86 arch/csky/kernel/smp.c void arch_send_call_function_ipi_mask(struct cpumask *mask) cpumask 23 arch/hexagon/include/asm/smp.h extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg); cpumask 26 arch/hexagon/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 93 arch/hexagon/kernel/smp.c void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) cpumask 101 arch/hexagon/kernel/smp.c for_each_cpu(cpu, cpumask) { cpumask 227 arch/hexagon/kernel/smp.c struct cpumask targets; cpumask 238 arch/hexagon/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 89 arch/hexagon/kernel/time.c static void broadcast(const struct cpumask *mask) cpumask 121 arch/hexagon/kernel/time.c dummy_clock_dev->cpumask = cpumask_of(cpu); cpumask 167 arch/hexagon/kernel/time.c ce_dev->cpumask = cpu_all_mask; cpumask 130 arch/ia64/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 324 arch/ia64/kernel/iosapic.c iosapic_set_affinity(struct irq_data *data, const struct cpumask *mask, cpumask 651 arch/ia64/kernel/iosapic.c const struct cpumask *cpu_mask; cpumask 117 arch/ia64/kernel/msi_ia64.c const struct cpumask *mask, bool force) cpumask 169 arch/ia64/kernel/smp.c send_IPI_mask(const struct cpumask *mask, int op) cpumask 259 arch/ia64/kernel/smp.c cpumask_t cpumask = xcpumask; cpumask 265 arch/ia64/kernel/smp.c for_each_cpu(cpu, &cpumask) cpumask 269 arch/ia64/kernel/smp.c for_each_cpu(cpu, &cpumask) { cpumask 279 arch/ia64/kernel/smp.c for_each_cpu(cpu, &cpumask) cpumask 324 arch/ia64/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 610 arch/ia64/kernel/smpboot.c const struct cpumask *mask; cpumask 149 arch/m68k/coldfire/pit.c cf_pit_clockevent.cpumask = cpumask_of(smp_processor_id()); cpumask 182 arch/microblaze/kernel/timer.c clockevent_xilinx_timer.cpumask = cpumask_of(0); cpumask 72 arch/mips/alchemy/common/time.c .cpumask = cpu_possible_mask, cpumask 34 arch/mips/bcm63xx/irq.c static void (*internal_irq_unmask)(struct irq_data *d, const struct cpumask *m); cpumask 54 arch/mips/bcm63xx/irq.c const struct cpumask *m) cpumask 134 arch/mips/bcm63xx/irq.c const struct cpumask *m) \ cpumask 376 arch/mips/bcm63xx/irq.c const struct cpumask *dest, cpumask 266 arch/mips/cavium-octeon/octeon-irq.c struct cpumask *mask = irq_data_get_affinity_mask(data); cpumask 761 arch/mips/cavium-octeon/octeon-irq.c struct cpumask *mask = irq_data_get_affinity_mask(data); cpumask 782 arch/mips/cavium-octeon/octeon-irq.c const struct cpumask *dest, bool force) cpumask 843 arch/mips/cavium-octeon/octeon-irq.c const struct cpumask *dest, cpumask 888 arch/mips/cavium-octeon/octeon-irq.c const struct cpumask *dest, cpumask 1781 arch/mips/cavium-octeon/octeon-irq.c const struct cpumask *dest, bool force) cpumask 2490 arch/mips/cavium-octeon/octeon-irq.c const struct cpumask *dest, bool force) cpumask 109 arch/mips/cavium-octeon/smp.c static inline void octeon_send_ipi_mask(const struct cpumask *mask, cpumask 487 arch/mips/cavium-octeon/smp.c static void octeon_78xx_send_ipi_mask(const struct cpumask *mask, cpumask 81 arch/mips/include/asm/irq.h void arch_trigger_cpumask_backtrace(const struct cpumask *mask, cpumask 39 arch/mips/include/asm/mach-loongson64/mmzone.h cpumask_t cpumask; cpumask 8 arch/mips/include/asm/mach-loongson64/topology.h #define cpumask_of_node(node) (&__node_data[(node)]->cpumask) cpumask 59 arch/mips/include/asm/mach-netlogic/multi-node.h cpumask_t cpumask; /* logical cpu mask for node */ cpumask 26 arch/mips/include/asm/smp-ops.h void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action); cpumask 52 arch/mips/include/asm/smp-ops.h extern void mips_smp_send_ipi_mask(const struct cpumask *mask, cpumask 115 arch/mips/include/asm/smp.h int mips_smp_ipi_allocate(const struct cpumask *mask); cpumask 122 arch/mips/include/asm/smp.h int mips_smp_ipi_free(const struct cpumask *mask); cpumask 131 arch/mips/include/asm/smp.h static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 142 arch/mips/jazz/irq.c cd->cpumask = cpumask_of(cpu); cpumask 118 arch/mips/kernel/cevt-bcm1480.c cd->cpumask = cpumask_of(cpu); cpumask 121 arch/mips/kernel/cevt-ds1287.c cd->cpumask = cpumask_of(0); cpumask 145 arch/mips/kernel/cevt-gt641xx.c cd->cpumask = cpumask_of(0); cpumask 283 arch/mips/kernel/cevt-r4k.c cd->cpumask = cpumask_of(cpu); cpumask 118 arch/mips/kernel/cevt-sb1250.c cd->cpumask = cpumask_of(cpu); cpumask 203 arch/mips/kernel/cevt-txx9.c cd->cpumask = cpumask_of(0), cpumask 685 arch/mips/kernel/process.c static struct cpumask backtrace_csd_busy; cpumask 750 arch/mips/kernel/process.c struct cpumask process_cpus; cpumask 298 arch/mips/kernel/smp-bmips.c static void bmips5000_send_ipi_mask(const struct cpumask *mask, cpumask 350 arch/mips/kernel/smp-bmips.c static void bmips43xx_send_ipi_mask(const struct cpumask *mask, cpumask 21 arch/mips/kernel/smp-up.c static inline void up_send_ipi_mask(const struct cpumask *mask, cpumask 154 arch/mips/kernel/smp.c void mips_smp_send_ipi_mask(const struct cpumask *mask, unsigned int action) cpumask 234 arch/mips/kernel/smp.c int mips_smp_ipi_allocate(const struct cpumask *mask) cpumask 292 arch/mips/kernel/smp.c int mips_smp_ipi_free(const struct cpumask *mask) cpumask 702 arch/mips/kernel/smp.c void tick_broadcast(const struct cpumask *mask) cpumask 243 arch/mips/lantiq/irq.c const struct cpumask *cpumask, bool force) cpumask 245 arch/mips/lantiq/irq.c struct cpumask tmask; cpumask 247 arch/mips/lantiq/irq.c if (!cpumask_and(&tmask, cpumask, cpu_online_mask)) cpumask 201 arch/mips/loongson32/common/time.c cd->cpumask = cpumask_of(smp_processor_id()); cpumask 119 arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c cd->cpumask = cpumask_of(cpu); cpumask 242 arch/mips/loongson64/loongson-3/hpet.c cd->cpumask = cpumask_of(cpu); cpumask 19 arch/mips/loongson64/loongson-3/irq.c int plat_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, cpumask 23 arch/mips/loongson64/loongson-3/irq.c struct cpumask new_affinity; cpumask 43 arch/mips/loongson64/loongson-3/irq.c struct cpumask affinity; cpumask 217 arch/mips/loongson64/loongson-3/numa.c cpumask_clear(&__node_data[(node)]->cpumask); cpumask 231 arch/mips/loongson64/loongson-3/numa.c cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask); cpumask 240 arch/mips/loongson64/loongson-3/smp.c loongson3_send_ipi_mask(const struct cpumask *mask, unsigned int action) cpumask 76 arch/mips/netlogic/common/smp.c void nlm_send_ipi_mask(const struct cpumask *mask, unsigned int action) cpumask 194 arch/mips/netlogic/common/smp.c cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask); cpumask 75 arch/mips/paravirt/paravirt-smp.c static void paravirt_send_ipi_mask(const struct cpumask *mask, unsigned int action) cpumask 275 arch/mips/pci/pci-xtalk-bridge.c static int bridge_set_affinity(struct irq_data *d, const struct cpumask *mask, cpumask 126 arch/mips/pmcs-msp71xx/msp_irq_cic.c const struct cpumask *cpumask, bool force) cpumask 140 arch/mips/pmcs-msp71xx/msp_irq_cic.c if (cpumask_test_cpu(cpu, cpumask)) cpumask 85 arch/mips/pmcs-msp71xx/msp_irq_per.c const struct cpumask *affinity, bool force) cpumask 70 arch/mips/sgi-ip27/ip27-irq.c static void setup_hub_mask(struct hub_irq_data *hd, const struct cpumask *mask) cpumask 90 arch/mips/sgi-ip27/ip27-irq.c static int set_affinity_hub_irq(struct irq_data *d, const struct cpumask *mask, cpumask 172 arch/mips/sgi-ip27/ip27-smp.c static void ip27_send_ipi_mask(const struct cpumask *mask, unsigned int action) cpumask 106 arch/mips/sgi-ip27/ip27-timer.c cd->cpumask = cpumask_of(cpu); cpumask 77 arch/mips/sibyte/bcm1480/irq.c static int bcm1480_set_affinity(struct irq_data *d, const struct cpumask *mask, cpumask 72 arch/mips/sibyte/bcm1480/smp.c static void bcm1480_send_ipi_mask(const struct cpumask *mask, cpumask 70 arch/mips/sibyte/sb1250/irq.c static int sb1250_set_affinity(struct irq_data *d, const struct cpumask *mask, cpumask 61 arch/mips/sibyte/sb1250/smp.c static inline void sb1250_send_ipi_mask(const struct cpumask *mask, cpumask 74 arch/mips/sni/time.c cd->cpumask = cpumask_of(cpu); cpumask 262 arch/nios2/kernel/time.c nios2_ce.ced.cpumask = cpumask_of(0); cpumask 21 arch/openrisc/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 23 arch/openrisc/include/asm/smp.h extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); cpumask 24 arch/openrisc/kernel/smp.c static void (*smp_cross_call)(const struct cpumask *, unsigned int); cpumask 200 arch/openrisc/kernel/smp.c void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) cpumask 210 arch/openrisc/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 82 arch/openrisc/kernel/time.c evt->cpumask = cpumask_of(cpu); cpumask 48 arch/parisc/include/asm/irq.h extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest); cpumask 33 arch/parisc/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 25 arch/parisc/include/asm/topology.h const struct cpumask *cpu_coregroup_mask(int cpu); cpumask 98 arch/parisc/kernel/irq.c int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest) cpumask 115 arch/parisc/kernel/irq.c static int cpu_set_affinity_irq(struct irq_data *d, const struct cpumask *dest, cpumask 201 arch/parisc/kernel/smp.c send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op) cpumask 241 arch/parisc/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 25 arch/parisc/kernel/topology.c const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 47 arch/powerpc/include/asm/cputhreads.h static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads) cpumask 58 arch/powerpc/include/asm/fadump-internal.h struct cpumask online_mask; cpumask 65 arch/powerpc/include/asm/irq.h int irq_choose_cpu(const struct cpumask *mask); cpumask 116 arch/powerpc/include/asm/smp.h static inline struct cpumask *cpu_sibling_mask(int cpu) cpumask 121 arch/powerpc/include/asm/smp.h static inline struct cpumask *cpu_core_mask(int cpu) cpumask 126 arch/powerpc/include/asm/smp.h static inline struct cpumask *cpu_l2_cache_mask(int cpu) cpumask 131 arch/powerpc/include/asm/smp.h static inline struct cpumask *cpu_smallcore_mask(int cpu) cpumask 181 arch/powerpc/include/asm/smp.h static inline const struct cpumask *cpu_sibling_mask(int cpu) cpumask 186 arch/powerpc/include/asm/smp.h static inline const struct cpumask *cpu_smallcore_mask(int cpu) cpumask 239 arch/powerpc/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 167 arch/powerpc/include/asm/xics.h extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, cpumask 170 arch/powerpc/include/asm/xics.h #define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server) cpumask 118 arch/powerpc/kernel/cacheinfo.c struct cpumask shared_cpu_map; /* online CPUs using this cache */ cpumask 642 arch/powerpc/kernel/cacheinfo.c static const struct cpumask *get_big_core_shared_cpu_map(int cpu, struct cache *cache) cpumask 654 arch/powerpc/kernel/cacheinfo.c const struct cpumask *mask; cpumask 695 arch/powerpc/kernel/irq.c int irq_choose_cpu(const struct cpumask *mask) cpumask 724 arch/powerpc/kernel/irq.c int irq_choose_cpu(const struct cpumask *mask) cpumask 342 arch/powerpc/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 373 arch/powerpc/kernel/smp.c static struct cpumask nmi_ipi_pending_mask; cpumask 539 arch/powerpc/kernel/smp.c void tick_broadcast(const struct cpumask *mask) cpumask 646 arch/powerpc/kernel/smp.c static void set_cpus_related(int i, int j, struct cpumask *(*get_cpumask)(int)) cpumask 654 arch/powerpc/kernel/smp.c struct cpumask *(*get_cpumask)(int)) cpumask 1128 arch/powerpc/kernel/smp.c static bool update_mask_by_l2(int cpu, struct cpumask *(*mask_fn)(int)) cpumask 1174 arch/powerpc/kernel/smp.c struct cpumask *this_l1_cache_map = per_cpu(cpu_l1_cache_map, cpu); cpumask 1234 arch/powerpc/kernel/smp.c struct cpumask *(*sibling_mask)(int) = cpu_sibling_mask; cpumask 1330 arch/powerpc/kernel/smp.c static const struct cpumask *shared_cache_mask(int cpu) cpumask 1336 arch/powerpc/kernel/smp.c static const struct cpumask *smallcore_smt_mask(int cpu) cpumask 998 arch/powerpc/kernel/time.c dec->cpumask = cpumask_of(cpu); cpumask 131 arch/powerpc/kernel/watchdog.c static void set_cpumask_stuck(const struct cpumask *cpumask, u64 tb) cpumask 133 arch/powerpc/kernel/watchdog.c cpumask_or(&wd_smp_cpus_stuck, &wd_smp_cpus_stuck, cpumask); cpumask 134 arch/powerpc/kernel/watchdog.c cpumask_andnot(&wd_smp_cpus_pending, &wd_smp_cpus_pending, cpumask); cpumask 355 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpumask 494 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpumask 545 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpumask 293 arch/powerpc/mm/nohash/tlb.c struct cpumask *cpu_mask; cpumask 111 arch/powerpc/perf/imc-pmu.c static DEVICE_ATTR(cpumask, S_IRUGO, imc_pmu_cpumask_get_attr, NULL); cpumask 336 arch/powerpc/perf/imc-pmu.c const struct cpumask *l_cpumask; cpumask 399 arch/powerpc/perf/imc-pmu.c const struct cpumask *l_cpumask; cpumask 400 arch/powerpc/perf/imc-pmu.c static struct cpumask tmp_mask; cpumask 627 arch/powerpc/perf/imc-pmu.c const struct cpumask *l_cpumask; cpumask 628 arch/powerpc/perf/imc-pmu.c static struct cpumask tmp_mask; cpumask 69 arch/powerpc/platforms/cell/spu_priv1_mmio.c const struct cpumask *spumask = cpumask_of_node(spu->node), cpumask 70 arch/powerpc/platforms/cell/spu_priv1_mmio.c *cpumask = cpumask_of_node(cpu_to_node(cpu)); cpumask 72 arch/powerpc/platforms/cell/spu_priv1_mmio.c if (!cpumask_intersects(spumask, cpumask)) cpumask 158 arch/powerpc/platforms/cell/spufs/sched.c const struct cpumask *mask = cpumask_of_node(node); cpumask 197 arch/powerpc/platforms/powernv/opal-imc.c const struct cpumask *l_cpumask; cpumask 70 arch/powerpc/sysdev/ehv_pic.c int ehv_pic_set_affinity(struct irq_data *d, const struct cpumask *dest, cpumask 624 arch/powerpc/sysdev/mpic.c static inline u32 mpic_physmask(u32 cpumask) cpumask 629 arch/powerpc/sysdev/mpic.c for (i = 0; i < min(32, NR_CPUS) && cpu_possible(i); ++i, cpumask >>= 1) cpumask 630 arch/powerpc/sysdev/mpic.c mask |= (cpumask & 1) << get_hard_smp_processor_id(i); cpumask 824 arch/powerpc/sysdev/mpic.c int mpic_set_affinity(struct irq_data *d, const struct cpumask *cpumask, cpumask 831 arch/powerpc/sysdev/mpic.c int cpuid = irq_choose_cpu(cpumask); cpumask 835 arch/powerpc/sysdev/mpic.c u32 mask = cpumask_bits(cpumask)[0]; cpumask 35 arch/powerpc/sysdev/mpic.h const struct cpumask *cpumask, bool force); cpumask 107 arch/powerpc/sysdev/xics/ics-opal.c const struct cpumask *cpumask, cpumask 128 arch/powerpc/sysdev/xics/ics-opal.c wanted_server = xics_get_irq_server(d->irq, cpumask, 1); cpumask 131 arch/powerpc/sysdev/xics/ics-opal.c __func__, cpumask_pr_args(cpumask), d->irq); cpumask 123 arch/powerpc/sysdev/xics/ics-rtas.c const struct cpumask *cpumask, cpumask 142 arch/powerpc/sysdev/xics/ics-rtas.c irq_server = xics_get_irq_server(d->irq, cpumask, 1); cpumask 145 arch/powerpc/sysdev/xics/ics-rtas.c __func__, cpumask_pr_args(cpumask), d->irq); cpumask 278 arch/powerpc/sysdev/xics/xics-common.c int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, cpumask 285 arch/powerpc/sysdev/xics/xics-common.c if (!cpumask_subset(cpu_possible_mask, cpumask)) { cpumask 286 arch/powerpc/sysdev/xics/xics-common.c int server = cpumask_first_and(cpu_online_mask, cpumask); cpumask 509 arch/powerpc/sysdev/xive/common.c static int xive_find_target_in_mask(const struct cpumask *mask, cpumask 556 arch/powerpc/sysdev/xive/common.c const struct cpumask *affinity) cpumask 718 arch/powerpc/sysdev/xive/common.c const struct cpumask *cpumask, cpumask 729 arch/powerpc/sysdev/xive/common.c if (cpumask_any_and(cpumask, cpu_online_mask) >= nr_cpu_ids) cpumask 742 arch/powerpc/sysdev/xive/common.c cpumask_test_cpu(xd->target, cpumask)) cpumask 746 arch/powerpc/sysdev/xive/common.c target = xive_pick_irq_target(d, cpumask); cpumask 32 arch/riscv/include/asm/smp.h void arch_send_call_function_ipi_mask(struct cpumask *mask); cpumask 38 arch/riscv/include/asm/smp.h void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out); cpumask 57 arch/riscv/kernel/smp.c void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out) cpumask 84 arch/riscv/kernel/smp.c static void send_ipi_mask(const struct cpumask *mask, enum ipi_message_type op) cpumask 86 arch/riscv/kernel/smp.c struct cpumask hartid_mask; cpumask 172 arch/riscv/kernel/smp.c void arch_send_call_function_ipi_mask(struct cpumask *mask) cpumask 12 arch/riscv/mm/tlbflush.c static void __sbi_tlb_flush_range(struct cpumask *cmask, unsigned long start, cpumask 15 arch/riscv/mm/tlbflush.c struct cpumask hmask; cpumask 261 arch/s390/include/asm/pci.h static inline const struct cpumask * cpumask 24 arch/s390/include/asm/smp.h extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 49 arch/s390/include/asm/topology.h const struct cpumask *cpu_coregroup_mask(int cpu); cpumask 79 arch/s390/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node) cpumask 63 arch/s390/kernel/processor.c void notrace stop_machine_yield(const struct cpumask *cpumask) cpumask 70 arch/s390/kernel/processor.c cpu = cpumask_next_wrap(this_cpu, cpumask, this_cpu, false); cpumask 168 arch/s390/kernel/smp.c static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address) cpumask 434 arch/s390/kernel/smp.c cpumask_t cpumask; cpumask 438 arch/s390/kernel/smp.c cpumask_copy(&cpumask, cpu_online_mask); cpumask 439 arch/s390/kernel/smp.c cpumask_clear_cpu(smp_processor_id(), &cpumask); cpumask 442 arch/s390/kernel/smp.c for_each_cpu(cpu, &cpumask) { cpumask 451 arch/s390/kernel/smp.c for_each_cpu(cpu, &cpumask) cpumask 453 arch/s390/kernel/smp.c cpumask_clear_cpu(cpu, &cpumask); cpumask 454 arch/s390/kernel/smp.c if (cpumask_empty(&cpumask)) cpumask 512 arch/s390/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 178 arch/s390/kernel/time.c cd->cpumask = cpumask_of(cpu); cpumask 483 arch/s390/kernel/topology.c static const struct cpumask *cpu_thread_mask(int cpu) cpumask 489 arch/s390/kernel/topology.c const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 494 arch/s390/kernel/topology.c static const struct cpumask *cpu_book_mask(int cpu) cpumask 499 arch/s390/kernel/topology.c static const struct cpumask *cpu_drawer_mask(int cpu) cpumask 101 arch/s390/pci/pci_irq.c static int zpci_set_irq_affinity(struct irq_data *data, const struct cpumask *dest, cpumask 40 arch/sh/include/asm/smp.h void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 20 arch/sh/include/asm/topology.h const struct cpumask *cpu_coregroup_mask(int cpu); cpumask 232 arch/sh/kernel/irq.c struct cpumask *mask = irq_data_get_affinity_mask(data); cpumask 271 arch/sh/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 285 arch/sh/kernel/smp.c void tick_broadcast(const struct cpumask *mask) cpumask 30 arch/sh/kernel/topology.c const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 90 arch/sparc/include/asm/irq_64.h void arch_trigger_cpumask_backtrace(const struct cpumask *mask, cpumask 98 arch/sparc/include/asm/smp_32.h void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 41 arch/sparc/include/asm/smp_64.h void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 60 arch/sparc/include/asm/topology_64.h static inline const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 349 arch/sparc/kernel/irq_64.c static int irq_choose_cpu(unsigned int irq, const struct cpumask *affinity) cpumask 396 arch/sparc/kernel/irq_64.c const struct cpumask *mask, bool force) cpumask 473 arch/sparc/kernel/irq_64.c const struct cpumask *mask, bool force) cpumask 538 arch/sparc/kernel/irq_64.c const struct cpumask *mask, bool force) cpumask 109 arch/sparc/kernel/leon_kernel.c static int irq_choose_cpu(const struct cpumask *affinity) cpumask 123 arch/sparc/kernel/leon_kernel.c static int leon_set_affinity(struct irq_data *data, const struct cpumask *dest, cpumask 145 arch/sparc/kernel/smp_32.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 861 arch/sparc/kernel/smp_64.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 131 arch/sparc/kernel/time_32.c ce->cpumask = cpu_possible_mask; cpumask 188 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); cpumask 196 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); cpumask 205 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); cpumask 227 arch/sparc/kernel/time_32.c ce->cpumask = cpumask_of(cpu); cpumask 764 arch/sparc/kernel/time_64.c sevt->cpumask = cpumask_of(smp_processor_id()); cpumask 109 arch/um/kernel/time.c .cpumask = cpu_possible_mask, cpumask 95 arch/unicore32/kernel/time.c ckevt_puv3_osmr0.cpumask = cpumask_of(0); cpumask 141 arch/x86/events/amd/iommu.c static DEVICE_ATTR(cpumask, S_IRUGO, _iommu_cpumask_show, NULL); cpumask 156 arch/x86/events/amd/power.c static DEVICE_ATTR(cpumask, S_IRUGO, get_attr_cpumask, NULL); cpumask 248 arch/x86/events/amd/uncore.c static DEVICE_ATTR(cpumask, S_IRUGO, amd_uncore_attr_show_cpumask, NULL); cpumask 204 arch/x86/events/intel/cstate.c static DEVICE_ATTR(cpumask, S_IRUGO, cstate_get_attr_cpumask, NULL); cpumask 387 arch/x86/events/intel/rapl.c static DEVICE_ATTR(cpumask, S_IRUGO, rapl_get_attr_cpumask, NULL); cpumask 817 arch/x86/events/intel/uncore.c static DEVICE_ATTR(cpumask, S_IRUGO, uncore_get_attr_cpumask, NULL); cpumask 100 arch/x86/hyperv/hv_apic.c static bool __send_ipi_mask_ex(const struct cpumask *mask, int vector) cpumask 139 arch/x86/hyperv/hv_apic.c static bool __send_ipi_mask(const struct cpumask *mask, int vector) cpumask 197 arch/x86/hyperv/hv_apic.c struct cpumask mask = CPU_MASK_NONE; cpumask 209 arch/x86/hyperv/hv_apic.c static void hv_send_ipi_mask(const struct cpumask *mask, int vector) cpumask 215 arch/x86/hyperv/hv_apic.c static void hv_send_ipi_mask_allbutself(const struct cpumask *mask, int vector) cpumask 218 arch/x86/hyperv/hv_apic.c struct cpumask new_mask; cpumask 219 arch/x86/hyperv/hv_apic.c const struct cpumask *local_mask; cpumask 20 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus, cpumask 55 arch/x86/hyperv/mmu.c static void hyperv_flush_tlb_others(const struct cpumask *cpus, cpumask 164 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus, cpumask 302 arch/x86/include/asm/apic.h void (*send_IPI_mask)(const struct cpumask *mask, int vector); cpumask 303 arch/x86/include/asm/apic.h void (*send_IPI_mask_allbutself)(const struct cpumask *msk, int vec); cpumask 70 arch/x86/include/asm/hw_irq.h const struct cpumask *mask; /* CPU mask for vector allocation */ cpumask 46 arch/x86/include/asm/irq.h void arch_trigger_cpumask_backtrace(const struct cpumask *mask, cpumask 18 arch/x86/include/asm/irqdomain.h const struct cpumask *mask); cpumask 343 arch/x86/include/asm/msr.h void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); cpumask 344 arch/x86/include/asm/msr.h void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); cpumask 372 arch/x86/include/asm/msr.h static inline void rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, cpumask 377 arch/x86/include/asm/msr.h static inline void wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, cpumask 65 arch/x86/include/asm/paravirt.h static inline void flush_tlb_others(const struct cpumask *cpumask, cpumask 68 arch/x86/include/asm/paravirt.h PVOP_VCALL2(mmu.flush_tlb_others, cpumask, info); cpumask 55 arch/x86/include/asm/paravirt_types.h struct cpumask; cpumask 209 arch/x86/include/asm/paravirt_types.h void (*flush_tlb_others)(const struct cpumask *cpus, cpumask 139 arch/x86/include/asm/pci.h static inline const struct cpumask * cpumask 32 arch/x86/include/asm/smp.h static inline struct cpumask *cpu_llc_shared_mask(int cpu) cpumask 60 arch/x86/include/asm/smp.h void (*send_call_func_ipi)(const struct cpumask *mask); cpumask 125 arch/x86/include/asm/smp.h static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 147 arch/x86/include/asm/smp.h void native_send_call_func_ipi(const struct cpumask *mask); cpumask 12 arch/x86/include/asm/tlbbatch.h struct cpumask cpumask; cpumask 588 arch/x86/include/asm/tlbflush.h void native_flush_tlb_others(const struct cpumask *cpumask, cpumask 606 arch/x86/include/asm/tlbflush.h cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm)); cpumask 67 arch/x86/include/asm/topology.h extern const struct cpumask *cpumask_of_node(int node); cpumask 70 arch/x86/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node) cpumask 105 arch/x86/include/asm/topology.h extern const struct cpumask *cpu_coregroup_mask(int cpu); cpumask 12 arch/x86/include/asm/trace/hyperv.h TP_PROTO(const struct cpumask *cpus, cpumask 60 arch/x86/include/asm/trace/hyperv.h TP_PROTO(const struct cpumask *cpus, cpumask 9 arch/x86/include/asm/uv/uv.h struct cpumask; cpumask 27 arch/x86/include/asm/uv/uv.h extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, cpumask 38 arch/x86/include/asm/uv/uv.h static inline const struct cpumask * cpumask 39 arch/x86/include/asm/uv/uv.h uv_flush_tlb_others(const struct cpumask *cpumask, cpumask 41 arch/x86/include/asm/uv/uv.h { return cpumask; } cpumask 618 arch/x86/include/asm/uv/uv_bau.h cpumask_t *cpumask; cpumask 519 arch/x86/kernel/apic/apic.c static void lapic_timer_broadcast(const struct cpumask *mask) cpumask 657 arch/x86/kernel/apic/apic.c levt->cpumask = cpumask_of(smp_processor_id()); cpumask 60 arch/x86/kernel/apic/apic_flat_64.c static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) cpumask 62 arch/x86/kernel/apic/apic_flat_64.c unsigned long mask = cpumask_bits(cpumask)[0]; cpumask 68 arch/x86/kernel/apic/apic_flat_64.c flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) cpumask 70 arch/x86/kernel/apic/apic_flat_64.c unsigned long mask = cpumask_bits(cpumask)[0]; cpumask 18 arch/x86/kernel/apic/apic_noop.c static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } cpumask 19 arch/x86/kernel/apic/apic_noop.c static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } cpumask 119 arch/x86/kernel/apic/apic_numachip.c static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) cpumask 127 arch/x86/kernel/apic/apic_numachip.c static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, cpumask 1885 arch/x86/kernel/apic/io_apic.c const struct cpumask *mask, bool force) cpumask 78 arch/x86/kernel/apic/ipi.c void native_send_call_func_ipi(const struct cpumask *mask) cpumask 184 arch/x86/kernel/apic/ipi.c void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) cpumask 202 arch/x86/kernel/apic/ipi.c void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, cpumask 246 arch/x86/kernel/apic/ipi.c void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, cpumask 266 arch/x86/kernel/apic/ipi.c void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, cpumask 289 arch/x86/kernel/apic/ipi.c void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) cpumask 291 arch/x86/kernel/apic/ipi.c unsigned long mask = cpumask_bits(cpumask)[0]; cpumask 58 arch/x86/kernel/apic/local.h void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector); cpumask 59 arch/x86/kernel/apic/local.h void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, int vector); cpumask 65 arch/x86/kernel/apic/local.h void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector); cpumask 66 arch/x86/kernel/apic/local.h void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, int vector); cpumask 67 arch/x86/kernel/apic/local.h void default_send_IPI_mask_logical(const struct cpumask *mask, int vector); cpumask 62 arch/x86/kernel/apic/msi.c msi_set_affinity(struct irq_data *irqd, const struct cpumask *mask, bool force) cpumask 64 arch/x86/kernel/apic/vector.c const struct cpumask *mask) cpumask 185 arch/x86/kernel/apic/vector.c const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd); cpumask 221 arch/x86/kernel/apic/vector.c assign_vector_locked(struct irq_data *irqd, const struct cpumask *dest) cpumask 257 arch/x86/kernel/apic/vector.c static int assign_irq_vector(struct irq_data *irqd, const struct cpumask *dest) cpumask 272 arch/x86/kernel/apic/vector.c const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd); cpumask 308 arch/x86/kernel/apic/vector.c assign_managed_vector(struct irq_data *irqd, const struct cpumask *dest) cpumask 310 arch/x86/kernel/apic/vector.c const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd); cpumask 417 arch/x86/kernel/apic/vector.c const struct cpumask *dest = irq_data_get_affinity_mask(irqd); cpumask 466 arch/x86/kernel/apic/vector.c const struct cpumask *dest = irq_data_get_affinity_mask(irqd); cpumask 770 arch/x86/kernel/apic/vector.c const struct cpumask *dest, bool force) cpumask 15 arch/x86/kernel/apic/x2apic_cluster.c struct cpumask mask; cpumask 37 arch/x86/kernel/apic/x2apic_cluster.c __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) cpumask 40 arch/x86/kernel/apic/x2apic_cluster.c struct cpumask *tmpmsk; cpumask 72 arch/x86/kernel/apic/x2apic_cluster.c static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) cpumask 78 arch/x86/kernel/apic/x2apic_cluster.c x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) cpumask 45 arch/x86/kernel/apic/x2apic_phys.c __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) cpumask 65 arch/x86/kernel/apic/x2apic_phys.c static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) cpumask 71 arch/x86/kernel/apic/x2apic_phys.c x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) cpumask 552 arch/x86/kernel/apic/x2apic_uv_x.c static void uv_send_IPI_mask(const struct cpumask *mask, int vector) cpumask 560 arch/x86/kernel/apic/x2apic_uv_x.c static void uv_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) cpumask 205 arch/x86/kernel/cpu/resctrl/internal.h struct cpumask cpu_mask; cpumask 318 arch/x86/kernel/cpu/resctrl/internal.h struct cpumask cpu_mask; cpumask 265 arch/x86/kernel/cpu/resctrl/rdtgroup.c struct cpumask *mask; cpumask 323 arch/x86/kernel/cpu/resctrl/rdtgroup.c update_closid_rmid(const struct cpumask *cpu_mask, struct rdtgroup *r) cpumask 376 arch/x86/kernel/cpu/resctrl/rdtgroup.c static void cpumask_rdtgrp_clear(struct rdtgroup *r, struct cpumask *m) cpumask 2183 arch/x86/kernel/cpu/resctrl/rdtgroup.c struct cpumask *mask) cpumask 403 arch/x86/kernel/hpet.c evt->cpumask = cpumask_of(hc->cpu); cpumask 434 arch/x86/kernel/kvm.c static void __send_ipi_mask(const struct cpumask *mask, int vector) cpumask 488 arch/x86/kernel/kvm.c static void kvm_send_ipi_mask(const struct cpumask *mask, int vector) cpumask 493 arch/x86/kernel/kvm.c static void kvm_send_ipi_mask_allbutself(const struct cpumask *mask, int vector) cpumask 496 arch/x86/kernel/kvm.c struct cpumask new_mask; cpumask 497 arch/x86/kernel/kvm.c const struct cpumask *local_mask; cpumask 515 arch/x86/kernel/kvm.c static void kvm_smp_send_call_func_ipi(const struct cpumask *mask) cpumask 583 arch/x86/kernel/kvm.c static void kvm_flush_tlb_others(const struct cpumask *cpumask, cpumask 589 arch/x86/kernel/kvm.c struct cpumask *flushmask = this_cpu_cpumask_var_ptr(__pv_tlb_mask); cpumask 591 arch/x86/kernel/kvm.c cpumask_copy(flushmask, cpumask); cpumask 65 arch/x86/kernel/nmi_selftest.c static void __init test_nmi_ipi(struct cpumask *mask) cpumask 647 arch/x86/kernel/smpboot.c const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 1526 arch/x86/kernel/tsc.c const struct cpumask *mask = topology_core_cpumask(cpu); cpumask 142 arch/x86/kernel/tsc_sync.c struct cpumask *mask; cpumask 99 arch/x86/lib/msr-smp.c static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, cpumask 127 arch/x86/lib/msr-smp.c void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) cpumask 141 arch/x86/lib/msr-smp.c void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) cpumask 818 arch/x86/mm/numa.c struct cpumask *mask; cpumask 862 arch/x86/mm/numa.c const struct cpumask *cpumask_of_node(int node) cpumask 663 arch/x86/mm/tlb.c void native_flush_tlb_others(const struct cpumask *cpumask, cpumask 689 arch/x86/mm/tlb.c cpumask = uv_flush_tlb_others(cpumask, info); cpumask 690 arch/x86/mm/tlb.c if (cpumask) cpumask 691 arch/x86/mm/tlb.c smp_call_function_many(cpumask, flush_tlb_func_remote, cpumask 707 arch/x86/mm/tlb.c smp_call_function_many(cpumask, flush_tlb_func_remote, cpumask 711 arch/x86/mm/tlb.c (void *)info, 1, GFP_ATOMIC, cpumask); cpumask 862 arch/x86/mm/tlb.c if (cpumask_test_cpu(cpu, &batch->cpumask)) { cpumask 869 arch/x86/mm/tlb.c if (cpumask_any_but(&batch->cpumask, cpu) < nr_cpu_ids) cpumask 870 arch/x86/mm/tlb.c flush_tlb_others(&batch->cpumask, &full_flush_tlb_info); cpumask 872 arch/x86/mm/tlb.c cpumask_clear(&batch->cpumask); cpumask 425 arch/x86/platform/uv/tlb_uv.c cpumask_t *mask = bcp->uvhub_master->cpumask; cpumask 896 arch/x86/platform/uv/tlb_uv.c static int uv_flush_send_and_wait(struct cpumask *flush_mask, cpumask 1048 arch/x86/platform/uv/tlb_uv.c static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp, cpumask 1102 arch/x86/platform/uv/tlb_uv.c const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, cpumask 1108 arch/x86/platform/uv/tlb_uv.c struct cpumask *flush_mask; cpumask 1116 arch/x86/platform/uv/tlb_uv.c return cpumask; cpumask 1127 arch/x86/platform/uv/tlb_uv.c return cpumask; cpumask 1135 arch/x86/platform/uv/tlb_uv.c return cpumask; cpumask 1144 arch/x86/platform/uv/tlb_uv.c flush_mask = (struct cpumask *)per_cpu(uv_flush_tlb_mask, cpu); cpumask 1146 arch/x86/platform/uv/tlb_uv.c cpumask_andnot(flush_mask, cpumask, cpumask_of(cpu)); cpumask 1148 arch/x86/platform/uv/tlb_uv.c if (cpumask_test_cpu(cpu, cpumask)) cpumask 1185 arch/x86/platform/uv/tlb_uv.c return cpumask; cpumask 2014 arch/x86/platform/uv/tlb_uv.c hmaster->cpumask = kzalloc_node(sz, GFP_KERNEL, hmaster->osnode); cpumask 51 arch/x86/platform/uv/uv_irq.c uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask, cpumask 304 arch/x86/platform/uv/uv_time.c int ced_cpu = cpumask_first(ced->cpumask); cpumask 314 arch/x86/platform/uv/uv_time.c int ced_cpu = cpumask_first(evt->cpumask); cpumask 347 arch/x86/platform/uv/uv_time.c ced->cpumask = cpumask_of(smp_processor_id()); cpumask 1348 arch/x86/xen/mmu_pv.c static void xen_flush_tlb_others(const struct cpumask *cpus, cpumask 156 arch/x86/xen/smp.c static void __xen_send_IPI_mask(const struct cpumask *mask, cpumask 165 arch/x86/xen/smp.c void xen_smp_send_call_function_ipi(const struct cpumask *mask) cpumask 218 arch/x86/xen/smp.c void xen_send_IPI_mask(const struct cpumask *mask, cpumask 243 arch/x86/xen/smp.c void xen_send_IPI_mask_allbutself(const struct cpumask *mask, cpumask 5 arch/x86/xen/smp.h extern void xen_send_IPI_mask(const struct cpumask *mask, cpumask 7 arch/x86/xen/smp.h extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask, cpumask 21 arch/x86/xen/smp.h void xen_smp_send_call_function_ipi(const struct cpumask *mask); cpumask 350 arch/x86/xen/time.c evt->cpumask = cpumask_of(cpu); cpumask 22 arch/xtensa/include/asm/smp.h struct cpumask; cpumask 23 arch/xtensa/include/asm/smp.h void arch_send_call_function_ipi_mask(const struct cpumask *mask); cpumask 175 arch/xtensa/kernel/irq.c struct cpumask *mask; cpumask 368 arch/xtensa/kernel/smp.c static void send_ipi_message(const struct cpumask *callmask, cpumask 380 arch/xtensa/kernel/smp.c void arch_send_call_function_ipi_mask(const struct cpumask *mask) cpumask 397 arch/xtensa/kernel/smp.c struct cpumask targets; cpumask 145 arch/xtensa/kernel/time.c clockevent->cpumask = cpumask_of(cpu); cpumask 29 block/blk-mq-pci.c const struct cpumask *mask; cpumask 27 block/blk-mq-rdma.c const struct cpumask *mask; cpumask 45 block/blk-mq-sysfs.c free_cpumask_var(hctx->cpumask); cpumask 173 block/blk-mq-sysfs.c for_each_cpu(i, hctx->cpumask) { cpumask 27 block/blk-mq-virtio.c const struct cpumask *mask; cpumask 479 block/blk-mq.c cpu = cpumask_first_and(alloc_data.hctx->cpumask, cpu_online_mask); cpumask 1388 block/blk-mq.c if (!cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask) && cpumask 1392 block/blk-mq.c cpumask_empty(hctx->cpumask) ? "inactive": "active"); cpumask 1411 block/blk-mq.c int cpu = cpumask_first_and(hctx->cpumask, cpu_online_mask); cpumask 1414 block/blk-mq.c cpu = cpumask_first(hctx->cpumask); cpumask 1434 block/blk-mq.c next_cpu = cpumask_next_and(next_cpu, hctx->cpumask, cpumask 1472 block/blk-mq.c if (cpumask_test_cpu(cpu, hctx->cpumask)) { cpumask 2361 block/blk-mq.c if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node)) cpumask 2411 block/blk-mq.c free_cpumask_var(hctx->cpumask); cpumask 2485 block/blk-mq.c cpumask_clear(hctx->cpumask); cpumask 2524 block/blk-mq.c if (cpumask_test_cpu(i, hctx->cpumask)) cpumask 2527 block/blk-mq.c cpumask_set_cpu(i, hctx->cpumask); cpumask 87 drivers/acpi/acpi_pad.c struct cpumask *pad_busy_cpus = to_cpumask(pad_busy_cpus_bits); cpumask 130 drivers/acpi/acpi_pad.c struct cpumask *pad_busy_cpus = to_cpumask(pad_busy_cpus_bits); cpumask 26 drivers/base/arch_topology.c void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpumask 434 drivers/base/arch_topology.c const struct cpumask *cpu_coregroup_mask(int cpu) cpumask 386 drivers/base/cacheinfo.c const struct cpumask *mask = &this_leaf->shared_cpu_map; cpumask 485 drivers/base/cacheinfo.c const struct cpumask *mask = &this_leaf->shared_cpu_map; cpumask 206 drivers/base/cpu.c const struct cpumask *const map; cpumask 3894 drivers/block/mtip32xx/mtip32xx.c const struct cpumask *node_mask; cpumask 4002 drivers/block/mtip32xx/mtip32xx.c const struct cpumask *node_mask; cpumask 309 drivers/clocksource/arc_timer.c evt->cpumask = cpumask_of(smp_processor_id()); cpumask 752 drivers/clocksource/arm_arch_timer.c clk->cpumask = cpumask_of(smp_processor_id()); cpumask 776 drivers/clocksource/arm_arch_timer.c clk->cpumask = cpu_possible_mask; cpumask 177 drivers/clocksource/arm_global_timer.c clk->cpumask = cpumask_of(cpu); cpumask 236 drivers/clocksource/asm9260_timer.c event_dev.cpumask = cpumask_of(0); cpumask 115 drivers/clocksource/bcm2835_timer.c timer->evt.cpumask = cpumask_of(0); cpumask 149 drivers/clocksource/bcm_kona_timer.c kona_clockevent_timer.cpumask = cpumask_of(0); cpumask 69 drivers/clocksource/clps711x-timer.c clkevt->cpumask = cpumask_of(0); cpumask 25 drivers/clocksource/dummy_timer.c evt->cpumask = cpumask_of(cpu); cpumask 127 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); cpumask 141 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); cpumask 174 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); cpumask 198 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); cpumask 260 drivers/clocksource/dw_apb_timer.c dw_ced->ced.cpumask = cpumask_of(cpu); cpumask 269 drivers/clocksource/em_sti.c ced->cpumask = cpu_possible_mask; cpumask 341 drivers/clocksource/exynos_mct.c mct_comp_device.cpumask = cpumask_of(0); cpumask 456 drivers/clocksource/exynos_mct.c evt->cpumask = cpumask_of(cpu); cpumask 120 drivers/clocksource/hyperv_timer.c ce->cpumask = cpumask_of(cpu); cpumask 191 drivers/clocksource/i8253.c i8253_clockevent.cpumask = cpumask_of(smp_processor_id()); cpumask 149 drivers/clocksource/ingenic-timer.c tcu->cevt.cpumask = cpumask_of(smp_processor_id()); cpumask 231 drivers/clocksource/jcore-pit.c pit->ced.cpumask = cpumask_of(cpu); cpumask 45 drivers/clocksource/mips-gic-timer.c int cpu = cpumask_first(evt->cpumask); cpumask 86 drivers/clocksource/mips-gic-timer.c cd->cpumask = cpumask_of(cpu); cpumask 151 drivers/clocksource/mps2-timer.c ce->clkevt.cpumask = cpu_possible_mask; cpumask 176 drivers/clocksource/mxs_timer.c mxs_clockevent_device.cpumask = cpumask_of(0); cpumask 236 drivers/clocksource/nomadik-mtu.c nmdk_clkevt.cpumask = cpumask_of(0); cpumask 68 drivers/clocksource/numachip.c ced->cpumask = cpumask_of(smp_processor_id()); cpumask 180 drivers/clocksource/renesas-ostm.c ced->cpumask = cpumask_of(0); cpumask 280 drivers/clocksource/samsung_pwm_timer.c time_event_device.cpumask = cpumask_of(0); cpumask 795 drivers/clocksource/sh_cmt.c ced->cpumask = cpu_possible_mask; cpumask 312 drivers/clocksource/sh_mtu2.c ced->cpumask = cpu_possible_mask; cpumask 411 drivers/clocksource/sh_tmu.c ced->cpumask = cpu_possible_mask; cpumask 195 drivers/clocksource/timer-armada-370-xp.c evt->cpumask = cpumask_of(cpu); cpumask 188 drivers/clocksource/timer-atcpit100.c .cpumask = cpu_possible_mask, cpumask 197 drivers/clocksource/timer-atlas7.c ce->cpumask = cpumask_of(cpu); cpumask 249 drivers/clocksource/timer-atmel-pit.c data->clkevt.cpumask = cpumask_of(0); cpumask 243 drivers/clocksource/timer-atmel-st.c clkevt.cpumask = cpumask_of(0); cpumask 283 drivers/clocksource/timer-atmel-tcb.c clkevt.clkevt.cpumask = cpumask_of(0); cpumask 442 drivers/clocksource/timer-cadence-ttc.c ttcce->ce.cpumask = cpu_possible_mask; cpumask 281 drivers/clocksource/timer-davinci.c clockevent->dev.cpumask = cpumask_of(0); cpumask 199 drivers/clocksource/timer-digicolor.c dc_timer_dev.ce.cpumask = cpu_possible_mask; cpumask 201 drivers/clocksource/timer-fsl-ftm.c ftm_clockevent.cpumask = cpumask_of(0); cpumask 377 drivers/clocksource/timer-fttmr010.c fttmr010->clkevt.cpumask = cpumask_of(0); cpumask 85 drivers/clocksource/timer-gx6605s.c .cpumask = cpu_possible_mask, cpumask 285 drivers/clocksource/timer-imx-gpt.c ced->cpumask = cpumask_of(0); cpumask 124 drivers/clocksource/timer-imx-sysctr.c to_sysctr.clkevt.cpumask = cpumask_of(0); cpumask 136 drivers/clocksource/timer-imx-tpm.c .cpumask = cpu_possible_mask, cpumask 219 drivers/clocksource/timer-ixp4xx.c tmr->clkevt.cpumask = cpumask_of(0); cpumask 210 drivers/clocksource/timer-keystone.c event_dev->cpumask = cpu_possible_mask; cpumask 250 drivers/clocksource/timer-mediatek.c .cpumask = cpu_possible_mask, cpumask 203 drivers/clocksource/timer-meson6.c meson6_clockevent.cpumask = cpu_possible_mask; cpumask 149 drivers/clocksource/timer-milbeaut.c .cpumask = cpu_possible_mask, cpumask 80 drivers/clocksource/timer-mp-csky.c to->clkevt.cpumask = cpumask_of(cpu); cpumask 163 drivers/clocksource/timer-npcm7xx.c npcm7xx_to.clkevt.cpumask = cpumask_of(0); cpumask 229 drivers/clocksource/timer-nps.c evt->cpumask = cpumask_of(smp_processor_id()); cpumask 182 drivers/clocksource/timer-orion.c orion_clkevt.cpumask = cpumask_of(0); cpumask 159 drivers/clocksource/timer-owl.c owl_clockevent.cpumask = cpumask_of(0); cpumask 159 drivers/clocksource/timer-oxnas-rps.c rps->clkevent.cpumask = cpu_possible_mask; cpumask 184 drivers/clocksource/timer-prima2.c sirfsoc_clockevent.cpumask = cpumask_of(0); cpumask 162 drivers/clocksource/timer-pxa.c ckevt_pxa_osmr0.cpumask = cpumask_of(0); cpumask 112 drivers/clocksource/timer-qcom.c evt->cpumask = cpumask_of(cpu); cpumask 61 drivers/clocksource/timer-riscv.c ce->cpumask = cpumask_of(cpu); cpumask 228 drivers/clocksource/timer-rockchip.c ce->cpumask = cpu_possible_mask; cpumask 199 drivers/clocksource/timer-sp804.c evt->cpumask = cpu_possible_mask; cpumask 135 drivers/clocksource/timer-sprd.c .cpumask = cpu_possible_mask, cpumask 155 drivers/clocksource/timer-sun4i.c .cpumask = cpu_possible_mask, cpumask 310 drivers/clocksource/timer-sun5i.c ce->clkevt.cpumask = cpu_possible_mask; cpumask 317 drivers/clocksource/timer-tegra.c cpu_to->clkevt.cpumask = cpumask_of(cpu); cpumask 140 drivers/clocksource/timer-vf-pit.c clockevent_pit.cpumask = cpumask_of(0); cpumask 140 drivers/clocksource/timer-vt8500.c clockevent.cpumask = cpumask_of(0); cpumask 161 drivers/clocksource/timer-zevio.c timer->clkevt.cpumask = cpu_possible_mask; cpumask 302 drivers/cpufreq/acpi-cpufreq.c static u32 drv_read(struct acpi_cpufreq_data *data, const struct cpumask *mask) cpumask 325 drivers/cpufreq/acpi-cpufreq.c const struct cpumask *mask, u32 val) cpumask 343 drivers/cpufreq/acpi-cpufreq.c static u32 get_cur_val(const struct cpumask *mask, struct acpi_cpufreq_data *data) cpumask 390 drivers/cpufreq/acpi-cpufreq.c const struct cpumask *mask, unsigned int freq) cpumask 410 drivers/cpufreq/acpi-cpufreq.c const struct cpumask *mask; cpumask 310 drivers/cpufreq/arm_big_little.c const struct cpumask *cpumask) cpumask 320 drivers/cpufreq/arm_big_little.c arm_bL_ops->free_opp_table(cpumask); cpumask 325 drivers/cpufreq/arm_big_little.c const struct cpumask *cpumask) cpumask 334 drivers/cpufreq/arm_big_little.c return _put_cluster_clk_and_freq_table(cpu_dev, cpumask); cpumask 343 drivers/cpufreq/arm_big_little.c _put_cluster_clk_and_freq_table(cdev, cpumask); cpumask 351 drivers/cpufreq/arm_big_little.c const struct cpumask *cpumask) cpumask 359 drivers/cpufreq/arm_big_little.c ret = arm_bL_ops->init_opp_table(cpumask); cpumask 388 drivers/cpufreq/arm_big_little.c arm_bL_ops->free_opp_table(cpumask); cpumask 396 drivers/cpufreq/arm_big_little.c const struct cpumask *cpumask) cpumask 405 drivers/cpufreq/arm_big_little.c ret = _get_cluster_clk_and_freq_table(cpu_dev, cpumask); cpumask 422 drivers/cpufreq/arm_big_little.c ret = _get_cluster_clk_and_freq_table(cdev, cpumask); cpumask 448 drivers/cpufreq/arm_big_little.c _put_cluster_clk_and_freq_table(cdev, cpumask); cpumask 33 drivers/cpufreq/arm_big_little.h int (*init_opp_table)(const struct cpumask *cpumask); cpumask 37 drivers/cpufreq/arm_big_little.h void (*free_opp_table)(const struct cpumask *cpumask); cpumask 39 drivers/cpufreq/armada-8k-cpufreq.c struct cpumask *cpumask) cpumask 58 drivers/cpufreq/armada-8k-cpufreq.c cpumask_set_cpu(cpu, cpumask); cpumask 128 drivers/cpufreq/armada-8k-cpufreq.c struct cpumask cpus; cpumask 149 drivers/cpufreq/armada-8k-cpufreq.c struct cpumask shared_cpus; cpumask 152 drivers/cpufreq/cpufreq.c __weak void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpumask 824 drivers/cpufreq/cpufreq.c ssize_t cpufreq_show_cpus(const struct cpumask *mask, char *buf) cpumask 36 drivers/cpufreq/mediatek-cpufreq.c struct cpumask cpus; cpumask 38 drivers/cpufreq/powernow-k8.h struct cpumask *available_cores; cpumask 152 drivers/cpufreq/qcom-cpufreq-hw.c static void qcom_get_related_cpus(int index, struct cpumask *m) cpumask 93 drivers/cpufreq/qoriq-cpufreq.c struct cpumask *dstp = policy->cpus; cpumask 80 drivers/cpufreq/scmi-cpufreq.c scmi_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) cpumask 99 drivers/cpufreq/scmi-cpufreq.c cpumask_set_cpu(cpu, cpumask); cpumask 72 drivers/cpufreq/scpi-cpufreq.c scpi_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) cpumask 91 drivers/cpufreq/scpi-cpufreq.c cpumask_set_cpu(cpu, cpumask); cpumask 30 drivers/cpufreq/vexpress-spc-cpufreq.c static int ve_spc_init_opp_table(const struct cpumask *cpumask) cpumask 32 drivers/cpufreq/vexpress-spc-cpufreq.c struct device *cpu_dev = get_cpu_device(cpumask_first(cpumask)); cpumask 86 drivers/cpuidle/cpuidle-arm.c drv->cpumask = (struct cpumask *)cpumask_of(cpu); cpumask 140 drivers/cpuidle/cpuidle-big_little.c struct cpumask *cpumask; cpumask 143 drivers/cpuidle/cpuidle-big_little.c cpumask = kzalloc(cpumask_size(), GFP_KERNEL); cpumask 144 drivers/cpuidle/cpuidle-big_little.c if (!cpumask) cpumask 149 drivers/cpuidle/cpuidle-big_little.c cpumask_set_cpu(cpu, cpumask); cpumask 151 drivers/cpuidle/cpuidle-big_little.c drv->cpumask = cpumask; cpumask 224 drivers/cpuidle/cpuidle-big_little.c kfree(bl_idle_big_driver.cpumask); cpumask 226 drivers/cpuidle/cpuidle-big_little.c kfree(bl_idle_little_driver.cpumask); cpumask 225 drivers/cpuidle/cpuidle-powernv.c drv->cpumask = (struct cpumask *)cpu_present_mask; cpumask 166 drivers/cpuidle/cpuidle-psci.c drv->cpumask = (struct cpumask *)cpumask_of(cpu); cpumask 671 drivers/cpuidle/cpuidle.c for_each_cpu(cpu, drv->cpumask) { cpumask 692 drivers/cpuidle/cpuidle.c const struct cpumask *const coupled_cpus) cpumask 703 drivers/cpuidle/cpuidle.c for_each_cpu(cpu, drv->cpumask) { cpumask 52 drivers/cpuidle/driver.c for_each_cpu(cpu, drv->cpumask) { cpumask 72 drivers/cpuidle/driver.c for_each_cpu(cpu, drv->cpumask) { cpumask 80 drivers/cpuidle/driver.c for_each_cpu(cpu, drv->cpumask) cpumask 165 drivers/cpuidle/driver.c if (!drv->cpumask) cpumask 166 drivers/cpuidle/driver.c drv->cpumask = (struct cpumask *)cpu_possible_mask; cpumask 215 drivers/cpuidle/driver.c on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, cpumask 237 drivers/cpuidle/driver.c on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, cpumask 98 drivers/cpuidle/dt_idle_states.c const cpumask_t *cpumask) cpumask 111 drivers/cpuidle/dt_idle_states.c for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); cpumask 112 drivers/cpuidle/dt_idle_states.c cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { cpumask 158 drivers/cpuidle/dt_idle_states.c const cpumask_t *cpumask; cpumask 169 drivers/cpuidle/dt_idle_states.c cpumask = drv->cpumask ? : cpu_possible_mask; cpumask 170 drivers/cpuidle/dt_idle_states.c cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); cpumask 188 drivers/cpuidle/dt_idle_states.c if (!idle_state_valid(state_node, i, cpumask)) { cpumask 3024 drivers/edac/amd64_edac.c static void get_cpus_on_this_dct_cpumask(struct cpumask *mask, u16 nid) cpumask 618 drivers/edac/cpc925_edac.c u32 cpumask; cpumask 622 drivers/edac/cpc925_edac.c cpumask = cpc925_cpu_mask_disabled(); cpumask 623 drivers/edac/cpc925_edac.c if (apimask & cpumask) { cpumask 626 drivers/edac/cpc925_edac.c apimask &= ~cpumask; cpumask 77 drivers/firmware/psci/psci_checker.c static unsigned int down_and_up_cpus(const struct cpumask *cpus, cpumask 78 drivers/firmware/psci/psci_checker.c struct cpumask *offlined_cpus) cpumask 166 drivers/firmware/psci/psci_checker.c const struct cpumask *cpu_group = cpumask 806 drivers/gpu/drm/amd/amdkfd/kfd_crat.c const struct cpumask *cpumask; cpumask 819 drivers/gpu/drm/amd/amdkfd/kfd_crat.c cpumask = cpumask_of_node(numa_node_id); cpumask 828 drivers/gpu/drm/amd/amdkfd/kfd_crat.c sub_type_hdr->num_cpu_cores = cpumask_weight(cpumask); cpumask 1433 drivers/gpu/drm/amd/amdkfd/kfd_topology.c static int kfd_cpumask_to_apic_id(const struct cpumask *cpumask) cpumask 1437 drivers/gpu/drm/amd/amdkfd/kfd_topology.c if (!cpumask || cpumask == cpu_none_mask) cpumask 1439 drivers/gpu/drm/amd/amdkfd/kfd_topology.c first_cpu_of_numa_node = cpumask_first(cpumask); cpumask 781 drivers/gpu/drm/i915/i915_pmu.c static DEVICE_ATTR(cpumask, 0444, i915_pmu_get_attr_cpumask, NULL); cpumask 659 drivers/hv/channel_mgmt.c struct cpumask *alloced_mask; cpumask 89 drivers/hv/hv.c hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), cpumask 155 drivers/hv/hyperv_vmbus.h struct cpumask *hv_numa_map; cpumask 94 drivers/hwmon/coretemp.c struct cpumask cpumask; cpumask 638 drivers/hwmon/coretemp.c if (!cpumask_intersects(&pdata->cpumask, topology_sibling_cpumask(cpu))) cpumask 641 drivers/hwmon/coretemp.c cpumask_set_cpu(cpu, &pdata->cpumask); cpumask 671 drivers/hwmon/coretemp.c cpumask_clear_cpu(cpu, &pd->cpumask); cpumask 678 drivers/hwmon/coretemp.c target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu)); cpumask 692 drivers/hwmon/coretemp.c if (cpumask_empty(&pd->cpumask)) { cpumask 704 drivers/hwmon/coretemp.c target = cpumask_first(&pd->cpumask); cpumask 540 drivers/infiniband/hw/hfi1/affinity.c struct cpumask *dev_comp_vect_mask = &dd->comp_vect->mask; cpumask 636 drivers/infiniband/hw/hfi1/affinity.c const struct cpumask *local_mask; cpumask 1042 drivers/infiniband/hw/hfi1/affinity.c const struct cpumask *node_mask, cpumask 68 drivers/infiniband/hw/hfi1/affinity.h struct cpumask mask; cpumask 69 drivers/infiniband/hw/hfi1/affinity.h struct cpumask used; cpumask 104 drivers/infiniband/hw/hfi1/affinity.h struct cpumask general_intr_mask; cpumask 105 drivers/infiniband/hw/hfi1/affinity.h struct cpumask comp_vect_mask; cpumask 111 drivers/infiniband/hw/hfi1/affinity.h struct cpumask real_cpu_mask; cpumask 407 drivers/infiniband/hw/hfi1/sdma.h struct cpumask cpu_mask; cpumask 1146 drivers/infiniband/hw/qib/qib_file_ops.c const struct cpumask *local_mask = cpumask_of_pcibus(dd->pcidev->bus); cpumask 3383 drivers/infiniband/hw/qib/qib_iba7322.c const struct cpumask *local_mask; cpumask 130 drivers/infiniband/sw/siw/siw_main.c struct cpumask **tx_valid_cpus; cpumask 143 drivers/infiniband/sw/siw/siw_main.c kcalloc(num_nodes, sizeof(struct cpumask *), GFP_KERNEL); cpumask 150 drivers/infiniband/sw/siw/siw_main.c kzalloc(sizeof(struct cpumask), GFP_KERNEL); cpumask 187 drivers/infiniband/sw/siw/siw_main.c const struct cpumask *tx_cpumask; cpumask 4528 drivers/iommu/amd_iommu.c const struct cpumask *mask, bool force) cpumask 39 drivers/iommu/hyperv-iommu.c const struct cpumask *mask, bool force) cpumask 1185 drivers/iommu/intel_irq_remapping.c intel_ir_set_affinity(struct irq_data *data, const struct cpumask *mask, cpumask 96 drivers/irqchip/exynos-combiner.c const struct cpumask *mask_val, bool force) cpumask 218 drivers/irqchip/irq-armada-370-xp.c const struct cpumask *mask, bool force) cpumask 317 drivers/irqchip/irq-armada-370-xp.c const struct cpumask *mask_val, bool force) cpumask 407 drivers/irqchip/irq-armada-370-xp.c static void armada_mpic_send_doorbell(const struct cpumask *mask, cpumask 148 drivers/irqchip/irq-bcm2836.c static void bcm2836_arm_irqchip_send_ipi(const struct cpumask *mask, cpumask 80 drivers/irqchip/irq-bcm6345-l1.c struct cpumask cpumask; cpumask 113 drivers/irqchip/irq-bcm6345-l1.c return cpumask_first_and(&intc->cpumask, irq_data_get_affinity_mask(d)); cpumask 197 drivers/irqchip/irq-bcm6345-l1.c const struct cpumask *dest, cpumask 205 drivers/irqchip/irq-bcm6345-l1.c struct cpumask valid; cpumask 209 drivers/irqchip/irq-bcm6345-l1.c if (!cpumask_and(&valid, &intc->cpumask, dest)) cpumask 319 drivers/irqchip/irq-bcm6345-l1.c cpumask_set_cpu(idx, &intc->cpumask); cpumask 322 drivers/irqchip/irq-bcm6345-l1.c if (!cpumask_weight(&intc->cpumask)) { cpumask 339 drivers/irqchip/irq-bcm6345-l1.c for_each_cpu(idx, &intc->cpumask) { cpumask 191 drivers/irqchip/irq-bcm7038-l1.c const struct cpumask *dest, cpumask 220 drivers/irqchip/irq-bcm7038-l1.c struct cpumask *mask = irq_data_get_affinity_mask(d); cpumask 128 drivers/irqchip/irq-csky-mpintc.c const struct cpumask *mask_val, cpumask 213 drivers/irqchip/irq-csky-mpintc.c static void csky_mpintc_send_ipi(const struct cpumask *mask) cpumask 1140 drivers/irqchip/irq-gic-v3-its.c static int its_set_affinity(struct irq_data *d, const struct cpumask *mask_val, cpumask 1144 drivers/irqchip/irq-gic-v3-its.c const struct cpumask *cpu_mask = cpu_online_mask; cpumask 2617 drivers/irqchip/irq-gic-v3-its.c const struct cpumask *cpu_mask = cpu_online_mask; cpumask 2793 drivers/irqchip/irq-gic-v3-its.c const struct cpumask *mask_val, cpumask 1049 drivers/irqchip/irq-gic-v3.c static u16 gic_compute_target_list(int *base_cpu, const struct cpumask *mask, cpumask 1095 drivers/irqchip/irq-gic-v3.c static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq) cpumask 1128 drivers/irqchip/irq-gic-v3.c static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, cpumask 329 drivers/irqchip/irq-gic.c static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, cpumask 494 drivers/irqchip/irq-gic.c u32 cpumask; cpumask 503 drivers/irqchip/irq-gic.c cpumask = gic_get_cpumask(gic); cpumask 504 drivers/irqchip/irq-gic.c cpumask |= cpumask << 8; cpumask 505 drivers/irqchip/irq-gic.c cpumask |= cpumask << 16; cpumask 507 drivers/irqchip/irq-gic.c writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4); cpumask 806 drivers/irqchip/irq-gic.c static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq) cpumask 147 drivers/irqchip/irq-hip04.c const struct cpumask *mask_val, cpumask 235 drivers/irqchip/irq-hip04.c u32 cpumask; cpumask 244 drivers/irqchip/irq-hip04.c cpumask = hip04_get_cpumask(intc); cpumask 245 drivers/irqchip/irq-hip04.c cpumask |= cpumask << 16; cpumask 247 drivers/irqchip/irq-hip04.c writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3)); cpumask 283 drivers/irqchip/irq-hip04.c static void hip04_raise_softirq(const struct cpumask *mask, unsigned int irq) cpumask 94 drivers/irqchip/irq-ls-scfg-msi.c const struct cpumask *mask; cpumask 104 drivers/irqchip/irq-ls-scfg-msi.c const struct cpumask *mask, bool force) cpumask 256 drivers/irqchip/irq-mips-gic.c static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask, cpumask 263 drivers/irqchip/irq-mips-gic.c cpu = cpumask_first_and(cpumask, cpu_online_mask); cpumask 571 drivers/irqchip/irq-mips-gic.c struct cpumask *ipimask = arg; cpumask 94 drivers/irqchip/irq-mvebu-sei.c const struct cpumask *mask_val, cpumask 97 drivers/irqchip/irq-ompic.c static void ompic_raise_softirq(const struct cpumask *mask, cpumask 86 drivers/irqchip/irq-sifive-plic.c static inline void plic_irq_toggle(const struct cpumask *mask, cpumask 116 drivers/irqchip/irq-sifive-plic.c const struct cpumask *mask_val, bool force) cpumask 553 drivers/irqchip/irq-stm32-exti.c const struct cpumask *dest, bool force) cpumask 406 drivers/irqchip/irq-ti-sci-inta.c const struct cpumask *mask_val, bool force) cpumask 131 drivers/irqchip/irq-xtensa-mx.c const struct cpumask *dest, bool force) cpumask 413 drivers/net/ethernet/cavium/liquidio/octeon_device.h struct cpumask affinity_mask; cpumask 1948 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct cpumask xps_mask; cpumask 388 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h struct cpumask dpio_cpumask; cpumask 97 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h struct cpumask affinity_mask; cpumask 850 drivers/net/ethernet/marvell/mvpp2/mvpp2.h struct cpumask *mask; cpumask 958 drivers/net/ethernet/mellanox/mlx4/en_rx.c const struct cpumask *aff; cpumask 42 drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c const struct cpumask *aff; cpumask 864 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct cpumask * cpumask 82 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h struct cpumask *mlx5_eq_comp_cpumask(struct mlx5_core_dev *dev, int ix); cpumask 170 drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h struct cpumask * cpumask 238 drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c struct cpumask * cpumask 108 drivers/opp/cpu.c void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, cpumask 114 drivers/opp/cpu.c WARN_ON(cpumask_empty(cpumask)); cpumask 116 drivers/opp/cpu.c for_each_cpu(cpu, cpumask) { cpumask 139 drivers/opp/cpu.c void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) cpumask 141 drivers/opp/cpu.c _dev_pm_opp_cpumask_remove_table(cpumask, -1); cpumask 156 drivers/opp/cpu.c const struct cpumask *cpumask) cpumask 167 drivers/opp/cpu.c for_each_cpu(cpu, cpumask) { cpumask 205 drivers/opp/cpu.c int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) cpumask 220 drivers/opp/cpu.c cpumask_clear(cpumask); cpumask 225 drivers/opp/cpu.c cpumask_set_cpu(opp_dev->dev->id, cpumask); cpumask 228 drivers/opp/cpu.c cpumask_set_cpu(cpu_dev->id, cpumask); cpumask 856 drivers/opp/of.c void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask) cpumask 858 drivers/opp/of.c _dev_pm_opp_cpumask_remove_table(cpumask, -1); cpumask 868 drivers/opp/of.c int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask) cpumask 873 drivers/opp/of.c if (WARN_ON(cpumask_empty(cpumask))) cpumask 876 drivers/opp/of.c for_each_cpu(cpu, cpumask) { cpumask 902 drivers/opp/of.c _dev_pm_opp_cpumask_remove_table(cpumask, cpu); cpumask 926 drivers/opp/of.c struct cpumask *cpumask) cpumask 938 drivers/opp/of.c cpumask_set_cpu(cpu_dev->id, cpumask); cpumask 967 drivers/opp/of.c cpumask_set_cpu(cpu, cpumask); cpumask 1103 drivers/opp/of.c void dev_pm_opp_of_register_em(struct cpumask *cpus) cpumask 218 drivers/opp/opp.h void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, int last_cpu); cpumask 669 drivers/parisc/iosapic.c const struct cpumask *dest, bool force) cpumask 190 drivers/pci/controller/dwc/pci-keystone.c const struct cpumask *mask, bool force) cpumask 142 drivers/pci/controller/dwc/pcie-designware-host.c const struct cpumask *mask, bool force) cpumask 680 drivers/pci/controller/pci-aardvark.c const struct cpumask *mask, bool force) cpumask 1162 drivers/pci/controller/pci-hyperv.c static int hv_set_affinity(struct irq_data *data, const struct cpumask *dest, cpumask 1191 drivers/pci/controller/pci-hyperv.c struct cpumask *dest; cpumask 1298 drivers/pci/controller/pci-hyperv.c struct pci_create_interrupt *int_pkt, struct cpumask *affinity, cpumask 1317 drivers/pci/controller/pci-hyperv.c struct pci_create_interrupt2 *int_pkt, struct cpumask *affinity, cpumask 1358 drivers/pci/controller/pci-hyperv.c struct cpumask *dest; cpumask 173 drivers/pci/controller/pci-xgene-msi.c const struct cpumask *mask, bool force) cpumask 105 drivers/pci/controller/pcie-altera-msi.c const struct cpumask *mask, bool force) cpumask 207 drivers/pci/controller/pcie-iproc-msi.c const struct cpumask *mask, bool force) cpumask 415 drivers/pci/controller/pcie-mediatek.c const struct cpumask *mask, bool force) cpumask 747 drivers/pci/controller/pcie-mobiveil.c const struct cpumask *mask, bool force) cpumask 82 drivers/pci/controller/pcie-tango.c static int tango_set_affinity(struct irq_data *d, const struct cpumask *mask, cpumask 465 drivers/pci/controller/pcie-xilinx-nwl.c const struct cpumask *mask, bool force) cpumask 174 drivers/pci/controller/vmd.c const struct cpumask *dest, bool force) cpumask 1292 drivers/pci/msi.c const struct cpumask *pci_irq_get_affinity(struct pci_dev *dev, int nr) cpumask 1326 drivers/pci/msi.c const struct cpumask *mask; cpumask 81 drivers/pci/pci-sysfs.c const struct cpumask *mask; cpumask 112 drivers/pci/pci-sysfs.c const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); cpumask 114 drivers/pci/pci-sysfs.c return cpumap_print_to_pagebuf(false, buf, cpumask); cpumask 121 drivers/pci/pci-sysfs.c const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); cpumask 123 drivers/pci/pci-sysfs.c return cpumap_print_to_pagebuf(true, buf, cpumask); cpumask 1372 drivers/perf/arm-cci.c __ATTR(cpumask, S_IRUGO, pmu_cpumask_attr_show, NULL); cpumask 559 drivers/perf/arm-ccn.c __ATTR(cpumask, S_IRUGO, arm_ccn_pmu_cpumask_show, NULL); cpumask 160 drivers/perf/arm_dsu_pmu.c const cpumask_t *cpumask; cpumask 164 drivers/perf/arm_dsu_pmu.c cpumask = &dsu_pmu->active_cpu; cpumask 167 drivers/perf/arm_dsu_pmu.c cpumask = &dsu_pmu->associated_cpus; cpumask 172 drivers/perf/arm_dsu_pmu.c return cpumap_print_to_pagebuf(true, buf, cpumask); cpumask 217 drivers/perf/arm_dsu_pmu.c DSU_CPUMASK_ATTR(cpumask, DSU_ACTIVE_CPU_MASK), cpumask 235 drivers/perf/arm_dsu_pmu.c struct cpumask online_supported; cpumask 487 drivers/perf/arm_smmuv3_pmu.c __ATTR(cpumask, 0444, smmu_pmu_cpumask_show, NULL); cpumask 243 drivers/perf/arm_spe_pmu.c static DEVICE_ATTR(cpumask, S_IRUGO, arm_spe_pmu_get_attr_cpumask, NULL); cpumask 91 drivers/perf/fsl_imx8_ddr_perf.c __ATTR(cpumask, 0444, ddr_perf_cpumask_show, NULL); cpumask 303 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL); cpumask 314 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL); cpumask 304 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL); cpumask 156 drivers/perf/qcom_l2_pmu.c cpumask_t cpumask; cpumask 683 drivers/perf/qcom_l2_pmu.c return cpumap_print_to_pagebuf(true, buf, &l2cache_pmu->cpumask); cpumask 687 drivers/perf/qcom_l2_pmu.c __ATTR(cpumask, S_IRUGO, l2_cache_pmu_cpumask_show, NULL); cpumask 838 drivers/perf/qcom_l2_pmu.c cpumask_set_cpu(cpu, &l2cache_pmu->cpumask); cpumask 864 drivers/perf/qcom_l2_pmu.c cpumask_clear_cpu(cpu, &l2cache_pmu->cpumask); cpumask 878 drivers/perf/qcom_l2_pmu.c cpumask_set_cpu(target, &l2cache_pmu->cpumask); cpumask 979 drivers/perf/qcom_l2_pmu.c cpumask_clear(&l2cache_pmu->cpumask); cpumask 160 drivers/perf/qcom_l3_pmu.c cpumask_t cpumask; cpumask 519 drivers/perf/qcom_l3_pmu.c event->cpu = cpumask_first(&l3pmu->cpumask); cpumask 678 drivers/perf/qcom_l3_pmu.c return cpumap_print_to_pagebuf(true, buf, &l3pmu->cpumask); cpumask 681 drivers/perf/qcom_l3_pmu.c static DEVICE_ATTR(cpumask, 0444, qcom_l3_cache_pmu_cpumask_show, NULL); cpumask 711 drivers/perf/qcom_l3_pmu.c if (cpumask_empty(&l3pmu->cpumask)) cpumask 712 drivers/perf/qcom_l3_pmu.c cpumask_set_cpu(cpu, &l3pmu->cpumask); cpumask 722 drivers/perf/qcom_l3_pmu.c if (!cpumask_test_and_clear_cpu(cpu, &l3pmu->cpumask)) cpumask 728 drivers/perf/qcom_l3_pmu.c cpumask_set_cpu(target, &l3pmu->cpumask); cpumask 188 drivers/perf/thunderx2_pmu.c static DEVICE_ATTR_RO(cpumask); cpumask 739 drivers/perf/thunderx2_pmu.c struct cpumask cpu_online_mask_temp; cpumask 615 drivers/perf/xgene_pmu.c static DEVICE_ATTR(cpumask, S_IRUGO, xgene_pmu_cpumask_show, NULL); cpumask 70 drivers/powercap/idle_inject.c unsigned long int cpumask[0]; cpumask 88 drivers/powercap/idle_inject.c for_each_cpu_and(cpu, to_cpumask(ii_dev->cpumask), cpu_online_mask) { cpumask 191 drivers/powercap/idle_inject.c cpumask_pr_args(to_cpumask(ii_dev->cpumask))); cpumask 220 drivers/powercap/idle_inject.c cpumask_pr_args(to_cpumask(ii_dev->cpumask))); cpumask 239 drivers/powercap/idle_inject.c for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { cpumask 288 drivers/powercap/idle_inject.c struct idle_inject_device *idle_inject_register(struct cpumask *cpumask) cpumask 297 drivers/powercap/idle_inject.c cpumask_copy(to_cpumask(ii_dev->cpumask), cpumask); cpumask 301 drivers/powercap/idle_inject.c for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { cpumask 314 drivers/powercap/idle_inject.c for_each_cpu(cpu_rb, to_cpumask(ii_dev->cpumask)) { cpumask 339 drivers/powercap/idle_inject.c for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) cpumask 65 drivers/powercap/intel_rapl_msr.c cpumask_set_cpu(cpu, &rp->cpumask); cpumask 78 drivers/powercap/intel_rapl_msr.c cpumask_clear_cpu(cpu, &rp->cpumask); cpumask 79 drivers/powercap/intel_rapl_msr.c lead_cpu = cpumask_first(&rp->cpumask); cpumask 180 drivers/scsi/hisi_sas/hisi_sas.h const struct cpumask *pci_irq_mask; cpumask 2130 drivers/scsi/hisi_sas/hisi_sas_main.c const struct cpumask *mask = cq->pci_irq_mask; cpumask 2349 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c const struct cpumask *mask; cpumask 7440 drivers/scsi/hpsa.c const struct cpumask *mask; cpumask 10650 drivers/scsi/lpfc/lpfc_init.c const struct cpumask *maskp; cpumask 10987 drivers/scsi/lpfc/lpfc_init.c const struct cpumask *maskp; cpumask 5697 drivers/scsi/megaraid/megaraid_sas_base.c const struct cpumask *mask; cpumask 464 drivers/scsi/storvsc_drv.c struct cpumask alloced_cpus; cpumask 1248 drivers/scsi/storvsc_drv.c const struct cpumask *node_mask; cpumask 1299 drivers/scsi/storvsc_drv.c const struct cpumask *node_mask; cpumask 69 drivers/sh/intc/chip.c const struct cpumask *cpumask, cpumask 72 drivers/sh/intc/chip.c if (!cpumask_intersects(cpumask, cpu_online_mask)) cpumask 75 drivers/sh/intc/chip.c cpumask_copy(irq_data_get_affinity_mask(data), cpumask); cpumask 34 drivers/soc/fsl/qbman/bman_portal.c static struct cpumask portal_cpus; cpumask 40 drivers/soc/fsl/qbman/qman_portal.c static struct cpumask portal_cpus; cpumask 325 drivers/soc/ti/knav_qmss.h struct cpumask *cpu_mask; cpumask 200 drivers/soc/ti/knav_qmss_acc.c struct cpumask *cpu_mask; cpumask 479 drivers/thermal/intel/int340x_thermal/processor_thermal_device.c cpumask_set_cpu(cpu, &rp->cpumask); cpumask 492 drivers/thermal/intel/int340x_thermal/processor_thermal_device.c cpumask_clear_cpu(cpu, &rp->cpumask); cpumask 493 drivers/thermal/intel/int340x_thermal/processor_thermal_device.c lead_cpu = cpumask_first(&rp->cpumask); cpumask 54 drivers/thermal/intel/x86_pkg_temp_thermal.c struct cpumask cpumask; cpumask 370 drivers/thermal/intel/x86_pkg_temp_thermal.c cpumask_set_cpu(cpu, &zonedev->cpumask); cpumask 386 drivers/thermal/intel/x86_pkg_temp_thermal.c target = cpumask_any_but(&zonedev->cpumask, cpu); cpumask 387 drivers/thermal/intel/x86_pkg_temp_thermal.c cpumask_clear_cpu(cpu, &zonedev->cpumask); cpumask 474 drivers/thermal/intel/x86_pkg_temp_thermal.c cpumask_set_cpu(cpu, &zonedev->cpumask); cpumask 423 drivers/virtio/virtio_pci_common.c int vp_set_vq_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask) cpumask 428 drivers/virtio/virtio_pci_common.c struct cpumask *mask; cpumask 447 drivers/virtio/virtio_pci_common.c const struct cpumask *vp_get_vq_affinity(struct virtio_device *vdev, int index) cpumask 141 drivers/virtio/virtio_pci_common.h int vp_set_vq_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask); cpumask 143 drivers/virtio/virtio_pci_common.h const struct cpumask *vp_get_vq_affinity(struct virtio_device *vdev, int index); cpumask 1332 drivers/xen/events/events_base.c static int set_affinity_irq(struct irq_data *data, const struct cpumask *dest, cpumask 132 include/asm-generic/mshyperv.h const struct cpumask *cpus) cpumask 56 include/linux/arch_topology.h const struct cpumask *cpu_coregroup_mask(int cpu); cpumask 23 include/linux/blk-mq.h cpumask_var_t cpumask; cpumask 119 include/linux/clockchips.h void (*broadcast)(const struct cpumask *mask); cpumask 129 include/linux/clockchips.h const struct cpumask *cpumask; cpumask 203 include/linux/clockchips.h extern void tick_broadcast(const struct cpumask *mask); cpumask 40 include/linux/cpu_rmap.h const struct cpumask *affinity); cpumask 711 include/linux/cpufreq.h ssize_t cpufreq_show_cpus(const struct cpumask *mask, char *buf); cpumask 991 include/linux/cpufreq.h extern void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpumask 123 include/linux/cpuidle.h struct cpumask *cpumask; cpumask 151 include/linux/cpuidle.h const struct cpumask *const coupled_cpus); cpumask 189 include/linux/cpuidle.h const struct cpumask *const coupled_cpus) cpumask 90 include/linux/cpumask.h extern struct cpumask __cpu_possible_mask; cpumask 91 include/linux/cpumask.h extern struct cpumask __cpu_online_mask; cpumask 92 include/linux/cpumask.h extern struct cpumask __cpu_present_mask; cpumask 93 include/linux/cpumask.h extern struct cpumask __cpu_active_mask; cpumask 94 include/linux/cpumask.h #define cpu_possible_mask ((const struct cpumask *)&__cpu_possible_mask) cpumask 95 include/linux/cpumask.h #define cpu_online_mask ((const struct cpumask *)&__cpu_online_mask) cpumask 96 include/linux/cpumask.h #define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask) cpumask 97 include/linux/cpumask.h #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) cpumask 150 include/linux/cpumask.h static inline unsigned int cpumask_first(const struct cpumask *srcp) cpumask 155 include/linux/cpumask.h static inline unsigned int cpumask_last(const struct cpumask *srcp) cpumask 161 include/linux/cpumask.h static inline unsigned int cpumask_next(int n, const struct cpumask *srcp) cpumask 166 include/linux/cpumask.h static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp) cpumask 172 include/linux/cpumask.h const struct cpumask *srcp, cpumask 173 include/linux/cpumask.h const struct cpumask *andp) cpumask 178 include/linux/cpumask.h static inline unsigned int cpumask_next_wrap(int n, const struct cpumask *mask, cpumask 186 include/linux/cpumask.h static inline unsigned int cpumask_any_but(const struct cpumask *mask, cpumask 212 include/linux/cpumask.h static inline unsigned int cpumask_first(const struct cpumask *srcp) cpumask 223 include/linux/cpumask.h static inline unsigned int cpumask_last(const struct cpumask *srcp) cpumask 228 include/linux/cpumask.h unsigned int cpumask_next(int n, const struct cpumask *srcp); cpumask 237 include/linux/cpumask.h static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp) cpumask 245 include/linux/cpumask.h int cpumask_next_and(int n, const struct cpumask *, const struct cpumask *); cpumask 246 include/linux/cpumask.h int cpumask_any_but(const struct cpumask *mask, unsigned int cpu); cpumask 273 include/linux/cpumask.h extern int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool wrap); cpumask 325 include/linux/cpumask.h static inline void cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp) cpumask 330 include/linux/cpumask.h static inline void __cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp) cpumask 341 include/linux/cpumask.h static inline void cpumask_clear_cpu(int cpu, struct cpumask *dstp) cpumask 346 include/linux/cpumask.h static inline void __cpumask_clear_cpu(int cpu, struct cpumask *dstp) cpumask 358 include/linux/cpumask.h static inline int cpumask_test_cpu(int cpu, const struct cpumask *cpumask) cpumask 360 include/linux/cpumask.h return test_bit(cpumask_check(cpu), cpumask_bits((cpumask))); cpumask 372 include/linux/cpumask.h static inline int cpumask_test_and_set_cpu(int cpu, struct cpumask *cpumask) cpumask 374 include/linux/cpumask.h return test_and_set_bit(cpumask_check(cpu), cpumask_bits(cpumask)); cpumask 386 include/linux/cpumask.h static inline int cpumask_test_and_clear_cpu(int cpu, struct cpumask *cpumask) cpumask 388 include/linux/cpumask.h return test_and_clear_bit(cpumask_check(cpu), cpumask_bits(cpumask)); cpumask 395 include/linux/cpumask.h static inline void cpumask_setall(struct cpumask *dstp) cpumask 404 include/linux/cpumask.h static inline void cpumask_clear(struct cpumask *dstp) cpumask 417 include/linux/cpumask.h static inline int cpumask_and(struct cpumask *dstp, cpumask 418 include/linux/cpumask.h const struct cpumask *src1p, cpumask 419 include/linux/cpumask.h const struct cpumask *src2p) cpumask 431 include/linux/cpumask.h static inline void cpumask_or(struct cpumask *dstp, const struct cpumask *src1p, cpumask 432 include/linux/cpumask.h const struct cpumask *src2p) cpumask 444 include/linux/cpumask.h static inline void cpumask_xor(struct cpumask *dstp, cpumask 445 include/linux/cpumask.h const struct cpumask *src1p, cpumask 446 include/linux/cpumask.h const struct cpumask *src2p) cpumask 460 include/linux/cpumask.h static inline int cpumask_andnot(struct cpumask *dstp, cpumask 461 include/linux/cpumask.h const struct cpumask *src1p, cpumask 462 include/linux/cpumask.h const struct cpumask *src2p) cpumask 473 include/linux/cpumask.h static inline void cpumask_complement(struct cpumask *dstp, cpumask 474 include/linux/cpumask.h const struct cpumask *srcp) cpumask 485 include/linux/cpumask.h static inline bool cpumask_equal(const struct cpumask *src1p, cpumask 486 include/linux/cpumask.h const struct cpumask *src2p) cpumask 498 include/linux/cpumask.h static inline bool cpumask_or_equal(const struct cpumask *src1p, cpumask 499 include/linux/cpumask.h const struct cpumask *src2p, cpumask 500 include/linux/cpumask.h const struct cpumask *src3p) cpumask 511 include/linux/cpumask.h static inline bool cpumask_intersects(const struct cpumask *src1p, cpumask 512 include/linux/cpumask.h const struct cpumask *src2p) cpumask 525 include/linux/cpumask.h static inline int cpumask_subset(const struct cpumask *src1p, cpumask 526 include/linux/cpumask.h const struct cpumask *src2p) cpumask 536 include/linux/cpumask.h static inline bool cpumask_empty(const struct cpumask *srcp) cpumask 545 include/linux/cpumask.h static inline bool cpumask_full(const struct cpumask *srcp) cpumask 554 include/linux/cpumask.h static inline unsigned int cpumask_weight(const struct cpumask *srcp) cpumask 565 include/linux/cpumask.h static inline void cpumask_shift_right(struct cpumask *dstp, cpumask 566 include/linux/cpumask.h const struct cpumask *srcp, int n) cpumask 578 include/linux/cpumask.h static inline void cpumask_shift_left(struct cpumask *dstp, cpumask 579 include/linux/cpumask.h const struct cpumask *srcp, int n) cpumask 590 include/linux/cpumask.h static inline void cpumask_copy(struct cpumask *dstp, cpumask 591 include/linux/cpumask.h const struct cpumask *srcp) cpumask 637 include/linux/cpumask.h struct cpumask *dstp) cpumask 651 include/linux/cpumask.h struct cpumask *dstp) cpumask 664 include/linux/cpumask.h static inline int cpumask_parse(const char *buf, struct cpumask *dstp) cpumask 678 include/linux/cpumask.h static inline int cpulist_parse(const char *buf, struct cpumask *dstp) cpumask 732 include/linux/cpumask.h typedef struct cpumask *cpumask_var_t; cpumask 751 include/linux/cpumask.h typedef struct cpumask cpumask_var_t[1]; cpumask 811 include/linux/cpumask.h void init_cpu_present(const struct cpumask *src); cpumask 812 include/linux/cpumask.h void init_cpu_possible(const struct cpumask *src); cpumask 813 include/linux/cpumask.h void init_cpu_online(const struct cpumask *src); cpumask 861 include/linux/cpumask.h ((struct cpumask *)(1 ? (bitmap) \ cpumask 879 include/linux/cpumask.h static inline const struct cpumask *get_cpu_mask(unsigned int cpu) cpumask 914 include/linux/cpumask.h cpumap_print_to_pagebuf(bool list, char *buf, const struct cpumask *mask) cpumask 60 include/linux/cpuset.h extern void cpuset_cpus_allowed(struct task_struct *p, struct cpumask *mask); cpumask 185 include/linux/cpuset.h struct cpumask *mask) cpumask 804 include/linux/hyperv.h struct cpumask alloced_cpus_in_node; cpumask 14 include/linux/idle_inject.h struct idle_inject_device *idle_inject_register(struct cpumask *cpumask); cpumask 143 include/linux/intel_rapl.h struct cpumask cpumask; cpumask 294 include/linux/interrupt.h struct cpumask mask; cpumask 303 include/linux/interrupt.h extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask, cpumask 314 include/linux/interrupt.h irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) cpumask 316 include/linux/interrupt.h return __irq_set_affinity(irq, cpumask, false); cpumask 331 include/linux/interrupt.h irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) cpumask 333 include/linux/interrupt.h return __irq_set_affinity(irq, cpumask, true); cpumask 339 include/linux/interrupt.h extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m); cpumask 352 include/linux/interrupt.h static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) cpumask 357 include/linux/interrupt.h static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) cpumask 370 include/linux/interrupt.h const struct cpumask *m) cpumask 481 include/linux/irq.h int (*irq_set_affinity)(struct irq_data *data, const struct cpumask *dest, bool force); cpumask 511 include/linux/irq.h void (*ipi_send_mask)(struct irq_data *data, const struct cpumask *dest); cpumask 571 include/linux/irq.h const struct cpumask *cpumask, bool force); cpumask 640 include/linux/irq.h const struct cpumask *dest, cpumask 676 include/linux/irq.h const struct cpumask *affinity); cpumask 678 include/linux/irq.h struct cpumask *affinity); cpumask 831 include/linux/irq.h static inline struct cpumask *irq_get_affinity_mask(int irq) cpumask 838 include/linux/irq.h static inline struct cpumask *irq_data_get_affinity_mask(struct irq_data *d) cpumask 845 include/linux/irq.h struct cpumask *irq_data_get_effective_affinity_mask(struct irq_data *d) cpumask 850 include/linux/irq.h const struct cpumask *m) cpumask 856 include/linux/irq.h const struct cpumask *m) cpumask 860 include/linux/irq.h struct cpumask *irq_data_get_effective_affinity_mask(struct irq_data *d) cpumask 1185 include/linux/irq.h int irq_matrix_reserve_managed(struct irq_matrix *m, const struct cpumask *msk); cpumask 1186 include/linux/irq.h void irq_matrix_remove_managed(struct irq_matrix *m, const struct cpumask *msk); cpumask 1187 include/linux/irq.h int irq_matrix_alloc_managed(struct irq_matrix *m, const struct cpumask *msk, cpumask 1191 include/linux/irq.h int irq_matrix_alloc(struct irq_matrix *m, const struct cpumask *msk, cpumask 1205 include/linux/irq.h int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest); cpumask 1207 include/linux/irq.h int ipi_send_mask(unsigned int virq, const struct cpumask *dest); cpumask 76 include/linux/irqdesc.h struct cpumask *percpu_enabled; cpumask 77 include/linux/irqdesc.h const struct cpumask *percpu_affinity; cpumask 79 include/linux/irqdesc.h const struct cpumask *affinity_hint; cpumask 44 include/linux/irqdomain.h struct cpumask; cpumask 437 include/linux/irqdomain.h int irq_reserve_ipi(struct irq_domain *domain, const struct cpumask *dest); cpumask 438 include/linux/irqdomain.h int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest); cpumask 54 include/linux/kthread.h void kthread_bind_mask(struct task_struct *k, const struct cpumask *mask); cpumask 1017 include/linux/mlx5/driver.h struct cpumask * cpumask 544 include/linux/mm_types.h cpumask_clear((struct cpumask *)cpu_bitmap); cpumask 550 include/linux/mm_types.h return (struct cpumask *)&mm->cpu_bitmap; cpumask 324 include/linux/msi.h int msi_domain_set_affinity(struct irq_data *data, const struct cpumask *mask, cpumask 3432 include/linux/netdevice.h int netif_set_xps_queue(struct net_device *dev, const struct cpumask *mask, cpumask 3522 include/linux/netdevice.h const struct cpumask *mask, cpumask 26 include/linux/nmi.h extern struct cpumask watchdog_cpumask; cpumask 156 include/linux/nmi.h static inline bool trigger_cpumask_backtrace(struct cpumask *mask) cpumask 183 include/linux/nmi.h static inline bool trigger_cpumask_backtrace(struct cpumask *mask) cpumask 123 include/linux/padata.h struct padata_cpumask cpumask; cpumask 169 include/linux/padata.h struct padata_cpumask cpumask; cpumask 189 include/linux/padata.h cpumask_var_t cpumask); cpumask 1456 include/linux/pci.h const struct cpumask *pci_irq_get_affinity(struct pci_dev *pdev, int vec); cpumask 1495 include/linux/pci.h static inline const struct cpumask *pci_irq_get_affinity(struct pci_dev *pdev, cpumask 137 include/linux/pm_opp.h int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask); cpumask 138 include/linux/pm_opp.h int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); cpumask 140 include/linux/pm_opp.h void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask); cpumask 320 include/linux/pm_opp.h static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask) cpumask 325 include/linux/pm_opp.h static inline int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) cpumask 334 include/linux/pm_opp.h static inline void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) cpumask 344 include/linux/pm_opp.h int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask); cpumask 345 include/linux/pm_opp.h void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask); cpumask 346 include/linux/pm_opp.h int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); cpumask 350 include/linux/pm_opp.h void dev_pm_opp_of_register_em(struct cpumask *cpus); cpumask 366 include/linux/pm_opp.h static inline int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask) cpumask 371 include/linux/pm_opp.h static inline void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask) cpumask 375 include/linux/pm_opp.h static inline int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) cpumask 390 include/linux/pm_opp.h static inline void dev_pm_opp_of_register_em(struct cpumask *cpus) cpumask 1571 include/linux/sched.h extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); cpumask 1572 include/linux/sched.h extern int task_can_attach(struct task_struct *p, const struct cpumask *cs_cpus_allowed); cpumask 1574 include/linux/sched.h extern void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask); cpumask 1575 include/linux/sched.h extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask); cpumask 1577 include/linux/sched.h static inline void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) cpumask 1580 include/linux/sched.h static inline int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) cpumask 1853 include/linux/sched.h extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); cpumask 1854 include/linux/sched.h extern long sched_getaffinity(pid_t pid, struct cpumask *mask); cpumask 1998 include/linux/sched.h const struct cpumask *sched_trace_rd_span(struct root_domain *rd); cpumask 21 include/linux/sched/isolation.h extern const struct cpumask *housekeeping_cpumask(enum hk_flags flags); cpumask 34 include/linux/sched/isolation.h static inline const struct cpumask *housekeeping_cpumask(enum hk_flags flags) cpumask 148 include/linux/sched/topology.h static inline struct cpumask *sched_domain_span(struct sched_domain *sd) cpumask 166 include/linux/sched/topology.h typedef const struct cpumask *(*sched_domain_mask_f)(int cpu); cpumask 44 include/linux/smp.h void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func, cpumask 58 include/linux/smp.h gfp_t gfp_flags, const struct cpumask *mask); cpumask 105 include/linux/smp.h void smp_call_function_many(const struct cpumask *mask, cpumask 108 include/linux/smp.h int smp_call_function_any(const struct cpumask *mask, cpumask 160 include/linux/smp.h smp_call_function_any(const struct cpumask *mask, smp_call_func_t func, cpumask 35 include/linux/stop_machine.h int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); cpumask 36 include/linux/stop_machine.h int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); cpumask 39 include/linux/stop_machine.h void stop_machine_yield(const struct cpumask *cpumask); cpumask 85 include/linux/stop_machine.h static inline int stop_cpus(const struct cpumask *cpumask, cpumask 88 include/linux/stop_machine.h if (cpumask_test_cpu(raw_smp_processor_id(), cpumask)) cpumask 93 include/linux/stop_machine.h static inline int try_stop_cpus(const struct cpumask *cpumask, cpumask 96 include/linux/stop_machine.h return stop_cpus(cpumask, fn, arg); cpumask 125 include/linux/stop_machine.h int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); cpumask 136 include/linux/stop_machine.h int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); cpumask 139 include/linux/stop_machine.h const struct cpumask *cpus); cpumask 143 include/linux/stop_machine.h const struct cpumask *cpus) cpumask 154 include/linux/stop_machine.h const struct cpumask *cpus) cpumask 160 include/linux/stop_machine.h const struct cpumask *cpus) cpumask 191 include/linux/tick.h static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) cpumask 265 include/linux/tick.h extern void __init tick_nohz_full_setup(cpumask_var_t cpumask); cpumask 269 include/linux/tick.h static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) { } cpumask 286 include/linux/tick.h static inline void tick_nohz_full_setup(cpumask_var_t cpumask) { } cpumask 219 include/linux/topology.h static inline const struct cpumask *cpu_smt_mask(int cpu) cpumask 225 include/linux/topology.h static inline const struct cpumask *cpu_cpu_mask(int cpu) cpumask 88 include/linux/virtio_config.h const struct cpumask *cpu_mask); cpumask 89 include/linux/virtio_config.h const struct cpumask *(*get_vq_affinity)(struct virtio_device *vdev, cpumask 245 include/linux/virtio_config.h int virtqueue_set_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask) cpumask 146 include/linux/workqueue.h cpumask_var_t cpumask; cpumask 442 include/linux/workqueue.h int workqueue_set_unbound_cpumask(cpumask_var_t cpumask); cpumask 2289 include/rdma/ib_verbs.h const struct cpumask *(*get_vector_affinity)(struct ib_device *ibdev, cpumask 4579 include/rdma/ib_verbs.h static inline const struct cpumask * cpumask 21 include/trace/events/ipi.h TP_PROTO(const struct cpumask *mask, const char *reason), cpumask 96 include/trace/events/thermal.h TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load, cpumask 102 include/trace/events/thermal.h __bitmask(cpumask, num_possible_cpus()) cpumask 110 include/trace/events/thermal.h __assign_bitmask(cpumask, cpumask_bits(cpus), cpumask 120 include/trace/events/thermal.h __get_bitmask(cpumask), __entry->freq, cpumask 126 include/trace/events/thermal.h TP_PROTO(const struct cpumask *cpus, unsigned int freq, cpumask 132 include/trace/events/thermal.h __bitmask(cpumask, num_possible_cpus()) cpumask 139 include/trace/events/thermal.h __assign_bitmask(cpumask, cpumask_bits(cpus), cpumask 147 include/trace/events/thermal.h __get_bitmask(cpumask), __entry->freq, __entry->cdev_state, cpumask 370 include/trace/events/xen.h TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm, cpumask 380 kernel/cgroup/cpuset.c static void guarantee_online_cpus(struct cpuset *cs, struct cpumask *pmask) cpumask 855 kernel/cgroup/cpuset.c struct cpumask *dp; cpumask 1055 kernel/cgroup/cpuset.c static void compute_effective_cpumask(struct cpumask *new_cpus, cpumask 1120 kernel/cgroup/cpuset.c struct cpumask *newmask, cpumask 2943 kernel/cgroup/cpuset.c struct cpumask *new_cpus, nodemask_t *new_mems, cpumask 2982 kernel/cgroup/cpuset.c struct cpumask *new_cpus, nodemask_t *new_mems, cpumask 3275 kernel/cgroup/cpuset.c void cpuset_cpus_allowed(struct task_struct *tsk, struct cpumask *pmask) cpumask 199 kernel/compat.c unsigned len, struct cpumask *new_mask) cpumask 2298 kernel/cpu.c struct cpumask __cpu_possible_mask __read_mostly cpumask 2301 kernel/cpu.c struct cpumask __cpu_possible_mask __read_mostly; cpumask 2305 kernel/cpu.c struct cpumask __cpu_online_mask __read_mostly; cpumask 2308 kernel/cpu.c struct cpumask __cpu_present_mask __read_mostly; cpumask 2311 kernel/cpu.c struct cpumask __cpu_active_mask __read_mostly; cpumask 2317 kernel/cpu.c void init_cpu_present(const struct cpumask *src) cpumask 2322 kernel/cpu.c void init_cpu_possible(const struct cpumask *src) cpumask 2327 kernel/cpu.c void init_cpu_online(const struct cpumask *src) cpumask 121 kernel/events/hw_breakpoint.c static const struct cpumask *cpumask_of_bp(struct perf_event *bp) cpumask 136 kernel/events/hw_breakpoint.c const struct cpumask *cpumask = cpumask_of_bp(bp); cpumask 139 kernel/events/hw_breakpoint.c for_each_cpu(cpu, cpumask) { cpumask 194 kernel/events/hw_breakpoint.c const struct cpumask *cpumask = cpumask_of_bp(bp); cpumask 207 kernel/events/hw_breakpoint.c for_each_cpu(cpu, cpumask) cpumask 12 kernel/irq/affinity.c static void irq_spread_init_one(struct cpumask *irqmsk, struct cpumask *nmsk, cpumask 15 kernel/irq/affinity.c const struct cpumask *siblmsk; cpumask 84 kernel/irq/affinity.c const struct cpumask *mask, nodemask_t *nodemsk) cpumask 130 kernel/irq/affinity.c const struct cpumask *cpu_mask, cpumask 132 kernel/irq/affinity.c struct cpumask *nmsk, cpumask 251 kernel/irq/affinity.c const struct cpumask *cpu_mask, cpumask 252 kernel/irq/affinity.c struct cpumask *nmsk, cpumask 194 kernel/irq/chip.c __irq_startup_managed(struct irq_desc *desc, struct cpumask *aff, bool force) cpumask 230 kernel/irq/chip.c __irq_startup_managed(struct irq_desc *desc, struct cpumask *aff, bool force) cpumask 258 kernel/irq/chip.c struct cpumask *aff = irq_data_get_affinity_mask(d); cpumask 1394 kernel/irq/chip.c const struct cpumask *dest, bool force) cpumask 21 kernel/irq/cpuhotplug.c const struct cpumask *m = irq_data_get_effective_affinity_mask(d); cpumask 57 kernel/irq/cpuhotplug.c const struct cpumask *affinity; cpumask 177 kernel/irq/cpuhotplug.c const struct cpumask *affinity = irq_data_get_affinity_mask(data); cpumask 33 kernel/irq/debugfs.c struct cpumask *msk; cpumask 134 kernel/irq/internals.h const struct cpumask *dest, bool force); cpumask 415 kernel/irq/internals.h irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) cpumask 420 kernel/irq/internals.h irq_get_pending(struct cpumask *mask, struct irq_desc *desc) cpumask 424 kernel/irq/internals.h static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc) cpumask 439 kernel/irq/internals.h irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) cpumask 443 kernel/irq/internals.h irq_get_pending(struct cpumask *mask, struct irq_desc *desc) cpumask 446 kernel/irq/internals.h static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc) cpumask 24 kernel/irq/ipi.c const struct cpumask *dest) cpumask 115 kernel/irq/ipi.c int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest) cpumask 118 kernel/irq/ipi.c struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; cpumask 165 kernel/irq/ipi.c struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; cpumask 187 kernel/irq/ipi.c const struct cpumask *dest, unsigned int cpu) cpumask 189 kernel/irq/ipi.c struct cpumask *ipimask = irq_data_get_affinity_mask(data); cpumask 263 kernel/irq/ipi.c int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest) cpumask 328 kernel/irq/ipi.c int ipi_send_mask(unsigned int virq, const struct cpumask *dest) cpumask 82 kernel/irq/irqdesc.c const struct cpumask *affinity) cpumask 100 kernel/irq/irqdesc.c desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { } cpumask 104 kernel/irq/irqdesc.c const struct cpumask *affinity, struct module *owner) cpumask 388 kernel/irq/irqdesc.c const struct cpumask *affinity, cpumask 482 kernel/irq/irqdesc.c const struct cpumask *mask = NULL; cpumask 901 kernel/irq/irqdesc.c const struct cpumask *affinity) cpumask 930 kernel/irq/irqdesc.c int irq_get_percpu_devid_partition(unsigned int irq, struct cpumask *affinity) cpumask 200 kernel/irq/manage.c const struct cpumask *m = irq_data_get_effective_affinity_mask(data); cpumask 210 kernel/irq/manage.c int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, cpumask 237 kernel/irq/manage.c const struct cpumask *dest) cpumask 247 kernel/irq/manage.c const struct cpumask *dest) cpumask 254 kernel/irq/manage.c const struct cpumask *dest, bool force) cpumask 268 kernel/irq/manage.c int irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask, cpumask 298 kernel/irq/manage.c int __irq_set_affinity(unsigned int irq, const struct cpumask *mask, bool force) cpumask 313 kernel/irq/manage.c int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m) cpumask 334 kernel/irq/manage.c cpumask_var_t cpumask; cpumask 337 kernel/irq/manage.c if (!desc || !alloc_cpumask_var(&cpumask, GFP_KERNEL)) cpumask 342 kernel/irq/manage.c irq_get_pending(cpumask, desc); cpumask 344 kernel/irq/manage.c cpumask_copy(cpumask, desc->irq_common_data.affinity); cpumask 347 kernel/irq/manage.c notify->notify(notify, cpumask); cpumask 349 kernel/irq/manage.c free_cpumask_var(cpumask); cpumask 408 kernel/irq/manage.c struct cpumask *set = irq_default_affinity; cpumask 411 kernel/irq/manage.c static struct cpumask mask; cpumask 436 kernel/irq/manage.c const struct cpumask *nodemask = cpumask_of_node(node); cpumask 962 kernel/irq/manage.c const struct cpumask *m; cpumask 130 kernel/irq/matrix.c const struct cpumask *msk) cpumask 151 kernel/irq/matrix.c const struct cpumask *msk) cpumask 210 kernel/irq/matrix.c int irq_matrix_reserve_managed(struct irq_matrix *m, const struct cpumask *msk) cpumask 251 kernel/irq/matrix.c void irq_matrix_remove_managed(struct irq_matrix *m, const struct cpumask *msk) cpumask 285 kernel/irq/matrix.c int irq_matrix_alloc_managed(struct irq_matrix *m, const struct cpumask *msk, cpumask 377 kernel/irq/matrix.c int irq_matrix_alloc(struct irq_matrix *m, const struct cpumask *msk, cpumask 102 kernel/irq/msi.c const struct cpumask *mask, bool force) cpumask 49 kernel/irq/proc.c const struct cpumask *mask; cpumask 395 kernel/kthread.c static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, long state) cpumask 416 kernel/kthread.c void kthread_bind_mask(struct task_struct *p, const struct cpumask *mask) cpumask 44 kernel/padata.c target_cpu = cpumask_first(pd->cpumask.pcpu); cpumask 46 kernel/padata.c target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); cpumask 57 kernel/padata.c int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); cpumask 119 kernel/padata.c if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) { cpumask 120 kernel/padata.c if (!cpumask_weight(pd->cpumask.cbcpu)) cpumask 124 kernel/padata.c cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu); cpumask 126 kernel/padata.c cpu = cpumask_first(pd->cpumask.cbcpu); cpumask 128 kernel/padata.c cpu = cpumask_next(cpu, pd->cpumask.cbcpu); cpumask 207 kernel/padata.c pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false); cpumask 363 kernel/padata.c cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu); cpumask 371 kernel/padata.c const struct cpumask *pcpumask, cpumask 372 kernel/padata.c const struct cpumask *cbcpumask) cpumask 376 kernel/padata.c if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) cpumask 378 kernel/padata.c if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL)) cpumask 381 kernel/padata.c cpumask_copy(pd->cpumask.pcpu, pcpumask); cpumask 382 kernel/padata.c cpumask_copy(pd->cpumask.cbcpu, cbcpumask); cpumask 387 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); cpumask 404 kernel/padata.c for_each_cpu(cpu, pd->cpumask.cbcpu) { cpumask 418 kernel/padata.c for_each_cpu(cpu, pd->cpumask.pcpu) { cpumask 432 kernel/padata.c const struct cpumask *cbcpumask; cpumask 433 kernel/padata.c const struct cpumask *pcpumask; cpumask 461 kernel/padata.c pd->cpu = cpumask_first(pd->cpumask.pcpu); cpumask 478 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); cpumask 479 kernel/padata.c free_cpumask_var(pd->cpumask.cbcpu); cpumask 524 kernel/padata.c cpumask_and(pinst->rcpumask.pcpu, pinst->cpumask.pcpu, cpumask 530 kernel/padata.c cpumask_and(pinst->rcpumask.cbcpu, pinst->cpumask.cbcpu, cpumask 550 kernel/padata.c &pinst->cpumask); cpumask 591 kernel/padata.c const struct cpumask *cpumask) cpumask 593 kernel/padata.c if (!cpumask_intersects(cpumask, cpu_online_mask)) { cpumask 620 kernel/padata.c cpumask_copy(pinst->cpumask.pcpu, pcpumask); cpumask 621 kernel/padata.c cpumask_copy(pinst->cpumask.cbcpu, cbcpumask); cpumask 641 kernel/padata.c cpumask_var_t cpumask) cpumask 643 kernel/padata.c struct cpumask *serial_mask, *parallel_mask; cpumask 651 kernel/padata.c serial_mask = pinst->cpumask.cbcpu; cpumask 652 kernel/padata.c parallel_mask = cpumask; cpumask 655 kernel/padata.c parallel_mask = pinst->cpumask.pcpu; cpumask 656 kernel/padata.c serial_mask = cpumask; cpumask 716 kernel/padata.c if (padata_validate_cpumask(pinst, pinst->cpumask.pcpu) && cpumask 717 kernel/padata.c padata_validate_cpumask(pinst, pinst->cpumask.cbcpu)) cpumask 729 kernel/padata.c if (!padata_validate_cpumask(pinst, pinst->cpumask.pcpu) || cpumask 730 kernel/padata.c !padata_validate_cpumask(pinst, pinst->cpumask.cbcpu)) cpumask 761 kernel/padata.c cpumask_clear_cpu(cpu, pinst->cpumask.cbcpu); cpumask 763 kernel/padata.c cpumask_clear_cpu(cpu, pinst->cpumask.pcpu); cpumask 776 kernel/padata.c return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) || cpumask 777 kernel/padata.c cpumask_test_cpu(cpu, pinst->cpumask.cbcpu); cpumask 827 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); cpumask 828 kernel/padata.c free_cpumask_var(pinst->cpumask.cbcpu); cpumask 855 kernel/padata.c struct cpumask *cpumask; cpumask 860 kernel/padata.c cpumask = pinst->cpumask.cbcpu; cpumask 862 kernel/padata.c cpumask = pinst->cpumask.pcpu; cpumask 865 kernel/padata.c nr_cpu_ids, cpumask_bits(cpumask)); cpumask 969 kernel/padata.c const struct cpumask *pcpumask, cpumask 970 kernel/padata.c const struct cpumask *cbcpumask) cpumask 990 kernel/padata.c if (!alloc_cpumask_var(&pinst->cpumask.pcpu, GFP_KERNEL)) cpumask 992 kernel/padata.c if (!alloc_cpumask_var(&pinst->cpumask.cbcpu, GFP_KERNEL)) { cpumask 993 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); cpumask 1009 kernel/padata.c cpumask_copy(pinst->cpumask.pcpu, pcpumask); cpumask 1010 kernel/padata.c cpumask_copy(pinst->cpumask.cbcpu, cbcpumask); cpumask 1041 kernel/padata.c free_cpumask_var(pinst->cpumask.pcpu); cpumask 1042 kernel/padata.c free_cpumask_var(pinst->cpumask.cbcpu); cpumask 485 kernel/rcu/rcu.h long rcutorture_sched_setaffinity(pid_t pid, const struct cpumask *in_mask); cpumask 453 kernel/rcu/update.c long rcutorture_sched_setaffinity(pid_t pid, const struct cpumask *in_mask) cpumask 1580 kernel/sched/core.c void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask) cpumask 1586 kernel/sched/core.c void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) cpumask 1625 kernel/sched/core.c const struct cpumask *new_mask, bool check) cpumask 1627 kernel/sched/core.c const struct cpumask *cpu_valid_mask = cpu_active_mask; cpumask 1696 kernel/sched/core.c int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) cpumask 2029 kernel/sched/core.c const struct cpumask *nodemask = NULL; cpumask 2163 kernel/sched/core.c const struct cpumask *new_mask, bool check) cpumask 5384 kernel/sched/core.c long sched_setaffinity(pid_t pid, const struct cpumask *in_mask) cpumask 5474 kernel/sched/core.c struct cpumask *new_mask) cpumask 5508 kernel/sched/core.c long sched_getaffinity(pid_t pid, struct cpumask *mask) cpumask 6080 kernel/sched/core.c int cpuset_cpumask_can_shrink(const struct cpumask *cur, cpumask 6081 kernel/sched/core.c const struct cpumask *trial) cpumask 6094 kernel/sched/core.c const struct cpumask *cs_cpus_allowed) cpumask 118 kernel/sched/cpudeadline.c struct cpumask *later_mask) cpumask 19 kernel/sched/cpudeadline.h int cpudl_find(struct cpudl *cp, struct task_struct *p, struct cpumask *later_mask); cpumask 60 kernel/sched/cpupri.c struct cpumask *lowest_mask) cpumask 21 kernel/sched/cpupri.h int cpupri_find(struct cpupri *cp, struct task_struct *p, struct cpumask *lowest_mask); cpumask 1879 kernel/sched/deadline.c struct cpumask *later_mask = this_cpu_cpumask_var_ptr(local_cpu_mask_dl); cpumask 2232 kernel/sched/deadline.c const struct cpumask *new_mask) cpumask 2716 kernel/sched/deadline.c int dl_task_can_attach(struct task_struct *p, const struct cpumask *cs_cpus_allowed) cpumask 2749 kernel/sched/deadline.c int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, cpumask 2750 kernel/sched/deadline.c const struct cpumask *trial) cpumask 5885 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask); cpumask 5960 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask); cpumask 6279 kernel/sched/fair.c struct cpumask *pd_mask = perf_domain_span(pd); cpumask 7131 kernel/sched/fair.c struct cpumask *dst_grpmask; cpumask 7136 kernel/sched/fair.c struct cpumask *cpus; cpumask 8838 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask); cpumask 10630 kernel/sched/fair.c const struct cpumask *sched_trace_rd_span(struct root_domain *rd) cpumask 40 kernel/sched/isolation.c const struct cpumask *housekeeping_cpumask(enum hk_flags flags) cpumask 541 kernel/sched/rt.c static inline const struct cpumask *sched_rt_period_mask(void) cpumask 546 kernel/sched/rt.c static inline const struct cpumask *sched_rt_period_mask(void) cpumask 609 kernel/sched/rt.c static inline const struct cpumask *sched_rt_period_mask(void) cpumask 815 kernel/sched/rt.c const struct cpumask *span; cpumask 1638 kernel/sched/rt.c struct cpumask *lowest_mask = this_cpu_cpumask_var_ptr(local_cpu_mask); cpumask 323 kernel/sched/sched.h extern int dl_task_can_attach(struct task_struct *p, const struct cpumask *cs_cpus_allowed); cpumask 324 kernel/sched/sched.h extern int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); cpumask 794 kernel/sched/sched.h extern int sched_init_domains(const struct cpumask *cpu_map); cpumask 1281 kernel/sched/sched.h extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu); cpumask 1286 kernel/sched/sched.h static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu) cpumask 1407 kernel/sched/sched.h unsigned long cpumask[0]; /* Balance mask */ cpumask 1425 kernel/sched/sched.h unsigned long cpumask[0]; cpumask 1428 kernel/sched/sched.h static inline struct cpumask *sched_group_span(struct sched_group *sg) cpumask 1430 kernel/sched/sched.h return to_cpumask(sg->cpumask); cpumask 1436 kernel/sched/sched.h static inline struct cpumask *group_balance_mask(struct sched_group *sg) cpumask 1438 kernel/sched/sched.h return to_cpumask(sg->sgc->cpumask); cpumask 1747 kernel/sched/sched.h const struct cpumask *newmask); cpumask 1836 kernel/sched/sched.h extern void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask); cpumask 29 kernel/sched/topology.c struct cpumask *groupmask) cpumask 276 kernel/sched/topology.c static void perf_domain_debug(const struct cpumask *cpu_map, cpumask 342 kernel/sched/topology.c static bool build_perf_domains(const struct cpumask *cpu_map) cpumask 840 kernel/sched/topology.c build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) cpumask 842 kernel/sched/topology.c const struct cpumask *sg_span = sched_group_span(sg); cpumask 880 kernel/sched/topology.c struct cpumask *sg_span; cpumask 901 kernel/sched/topology.c struct cpumask *mask = sched_domains_tmpmask2; cpumask 903 kernel/sched/topology.c struct cpumask *sg_span; cpumask 930 kernel/sched/topology.c const struct cpumask *span = sched_domain_span(sd); cpumask 931 kernel/sched/topology.c struct cpumask *covered = sched_domains_tmpmask; cpumask 939 kernel/sched/topology.c struct cpumask *sg_span; cpumask 1106 kernel/sched/topology.c const struct cpumask *span = sched_domain_span(sd); cpumask 1107 kernel/sched/topology.c struct cpumask *covered; cpumask 1217 kernel/sched/topology.c static void __sdt_free(const struct cpumask *cpu_map); cpumask 1218 kernel/sched/topology.c static int __sdt_alloc(const struct cpumask *cpu_map); cpumask 1221 kernel/sched/topology.c const struct cpumask *cpu_map) cpumask 1240 kernel/sched/topology.c __visit_domain_allocation_hell(struct s_data *d, const struct cpumask *cpu_map) cpumask 1286 kernel/sched/topology.c static struct cpumask ***sched_domains_numa_masks; cpumask 1316 kernel/sched/topology.c const struct cpumask *cpu_map, cpumask 1462 kernel/sched/topology.c static const struct cpumask *sd_numa_mask(int cpu) cpumask 1646 kernel/sched/topology.c struct cpumask *mask = kzalloc(cpumask_size(), GFP_KERNEL); cpumask 1736 kernel/sched/topology.c int sched_numa_find_closest(const struct cpumask *cpus, int cpu) cpumask 1750 kernel/sched/topology.c static int __sdt_alloc(const struct cpumask *cpu_map) cpumask 1819 kernel/sched/topology.c static void __sdt_free(const struct cpumask *cpu_map) cpumask 1856 kernel/sched/topology.c const struct cpumask *cpu_map, struct sched_domain_attr *attr, cpumask 1890 kernel/sched/topology.c const struct cpumask *cpu_map, int cpu) cpumask 1926 kernel/sched/topology.c *asym_cpu_capacity_level(const struct cpumask *cpu_map) cpumask 1985 kernel/sched/topology.c build_sched_domains(const struct cpumask *cpu_map, struct sched_domain_attr *attr) cpumask 2140 kernel/sched/topology.c int sched_init_domains(const struct cpumask *cpu_map) cpumask 2164 kernel/sched/topology.c static void detach_destroy_domains(const struct cpumask *cpu_map) cpumask 33 kernel/smp.c cpumask_var_t cpumask; cpumask 47 kernel/smp.c if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, cpumask 52 kernel/smp.c free_cpumask_var(cfd->cpumask); cpumask 57 kernel/smp.c free_cpumask_var(cfd->cpumask); cpumask 69 kernel/smp.c free_cpumask_var(cfd->cpumask); cpumask 369 kernel/smp.c int smp_call_function_any(const struct cpumask *mask, cpumask 373 kernel/smp.c const struct cpumask *nodemask; cpumask 412 kernel/smp.c void smp_call_function_many(const struct cpumask *mask, cpumask 457 kernel/smp.c cpumask_and(cfd->cpumask, mask, cpu_online_mask); cpumask 458 kernel/smp.c __cpumask_clear_cpu(this_cpu, cfd->cpumask); cpumask 461 kernel/smp.c if (unlikely(!cpumask_weight(cfd->cpumask))) cpumask 465 kernel/smp.c for_each_cpu(cpu, cfd->cpumask) { cpumask 481 kernel/smp.c for_each_cpu(cpu, cfd->cpumask) { cpumask 640 kernel/smp.c void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func, cpumask 685 kernel/smp.c gfp_t gfp_flags, const struct cpumask *mask) cpumask 159 kernel/stop_machine.c const struct cpumask *active_cpus; cpumask 181 kernel/stop_machine.c void __weak stop_machine_yield(const struct cpumask *cpumask) cpumask 192 kernel/stop_machine.c const struct cpumask *cpumask; cpumask 203 kernel/stop_machine.c cpumask = cpu_online_mask; cpumask 204 kernel/stop_machine.c is_active = cpu == cpumask_first(cpumask); cpumask 206 kernel/stop_machine.c cpumask = msdata->active_cpus; cpumask 207 kernel/stop_machine.c is_active = cpumask_test_cpu(cpu, cpumask); cpumask 213 kernel/stop_machine.c stop_machine_yield(cpumask); cpumask 373 kernel/stop_machine.c static bool queue_stop_cpus_work(const struct cpumask *cpumask, cpumask 389 kernel/stop_machine.c for_each_cpu(cpu, cpumask) { cpumask 404 kernel/stop_machine.c static int __stop_cpus(const struct cpumask *cpumask, cpumask 409 kernel/stop_machine.c cpu_stop_init_done(&done, cpumask_weight(cpumask)); cpumask 410 kernel/stop_machine.c if (!queue_stop_cpus_work(cpumask, fn, arg, &done)) cpumask 444 kernel/stop_machine.c int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) cpumask 450 kernel/stop_machine.c ret = __stop_cpus(cpumask, fn, arg); cpumask 473 kernel/stop_machine.c int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) cpumask 480 kernel/stop_machine.c ret = __stop_cpus(cpumask, fn, arg); cpumask 596 kernel/stop_machine.c const struct cpumask *cpus) cpumask 631 kernel/stop_machine.c int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus) cpumask 666 kernel/stop_machine.c const struct cpumask *cpus) cpumask 278 kernel/taskstats.c static int add_del_listener(pid_t pid, const struct cpumask *mask, int isadd) cpumask 337 kernel/taskstats.c static int parse(struct nlattr *na, struct cpumask *mask) cpumask 451 kernel/time/clockevents.c if (!dev->cpumask) { cpumask 453 kernel/time/clockevents.c dev->cpumask = cpumask_of(smp_processor_id()); cpumask 456 kernel/time/clockevents.c if (dev->cpumask == cpu_all_mask) { cpumask 459 kernel/time/clockevents.c dev->cpumask = cpu_possible_mask; cpumask 651 kernel/time/clockevents.c if (cpumask_test_cpu(cpu, dev->cpumask) && cpumask 652 kernel/time/clockevents.c cpumask_weight(dev->cpumask) == 1 && cpumask 96 kernel/time/tick-broadcast-hrtimer.c .cpumask = cpu_possible_mask, cpumask 59 kernel/time/tick-broadcast.c struct cpumask *tick_get_broadcast_mask(void) cpumask 143 kernel/time/tick-broadcast.c static void err_broadcast(const struct cpumask *mask) cpumask 264 kernel/time/tick-broadcast.c static bool tick_do_broadcast(struct cpumask *mask) cpumask 532 kernel/time/tick-broadcast.c struct cpumask *tick_get_broadcast_oneshot_mask(void) cpumask 553 kernel/time/tick-broadcast.c const struct cpumask *cpumask) cpumask 558 kernel/time/tick-broadcast.c if (cpumask_equal(bc->cpumask, cpumask)) cpumask 561 kernel/time/tick-broadcast.c bc->cpumask = cpumask; cpumask 562 kernel/time/tick-broadcast.c irq_set_affinity(bc->irq, bc->cpumask); cpumask 867 kernel/time/tick-broadcast.c static void tick_broadcast_init_next_event(struct cpumask *mask, cpumask 203 kernel/time/tick-common.c const struct cpumask *cpumask) cpumask 255 kernel/time/tick-common.c if (!cpumask_equal(newdev->cpumask, cpumask)) cpumask 256 kernel/time/tick-common.c irq_set_affinity(newdev->irq, cpumask); cpumask 288 kernel/time/tick-common.c if (!cpumask_test_cpu(cpu, newdev->cpumask)) cpumask 290 kernel/time/tick-common.c if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) cpumask 296 kernel/time/tick-common.c if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) cpumask 318 kernel/time/tick-common.c !cpumask_equal(curdev->cpumask, newdev->cpumask); cpumask 74 kernel/time/tick-internal.h extern struct cpumask *tick_get_broadcast_mask(void); cpumask 132 kernel/time/tick-internal.h extern struct cpumask *tick_get_broadcast_oneshot_mask(void); cpumask 398 kernel/time/tick-sched.c void __init tick_nohz_full_setup(cpumask_var_t cpumask) cpumask 401 kernel/time/tick-sched.c cpumask_copy(tick_nohz_full_mask, cpumask); cpumask 271 kernel/trace/ring_buffer.c for_each_cpu(cpu, buffer->cpumask) cpumask 490 kernel/trace/ring_buffer.c cpumask_var_t cpumask; cpumask 593 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 696 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 1392 kernel/trace/ring_buffer.c if (!zalloc_cpumask_var(&buffer->cpumask, GFP_KERNEL)) cpumask 1416 kernel/trace/ring_buffer.c cpumask_set_cpu(cpu, buffer->cpumask); cpumask 1437 kernel/trace/ring_buffer.c free_cpumask_var(buffer->cpumask); cpumask 1460 kernel/trace/ring_buffer.c free_cpumask_var(buffer->cpumask); cpumask 1730 kernel/trace/ring_buffer.c !cpumask_test_cpu(cpu_id, buffer->cpumask)) cpumask 1812 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu_id, buffer->cpumask)) cpumask 2978 kernel/trace/ring_buffer.c if (unlikely(!cpumask_test_cpu(cpu, buffer->cpumask))) cpumask 3133 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3302 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3322 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3355 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3386 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3405 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3425 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3448 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3470 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 3490 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4088 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4156 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4209 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4343 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4406 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4488 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4520 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer_a->cpumask) || cpumask 4521 kernel/trace/ring_buffer.c !cpumask_test_cpu(cpu, buffer_b->cpumask)) cpumask 4600 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4714 kernel/trace/ring_buffer.c if (!cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4877 kernel/trace/ring_buffer.c if (cpumask_test_cpu(cpu, buffer->cpumask)) cpumask 4903 kernel/trace/ring_buffer.c cpumask_set_cpu(cpu, buffer->cpumask); cpumask 267 kernel/trace/trace_hwlat.c static struct cpumask save_cpumask; cpumask 272 kernel/trace/trace_hwlat.c struct cpumask *current_mask = &save_cpumask; cpumask 353 kernel/trace/trace_hwlat.c struct cpumask *current_mask = &save_cpumask; cpumask 54 kernel/up.c void on_each_cpu_mask(const struct cpumask *mask, cpumask 73 kernel/up.c gfp_t gfp_flags, const struct cpumask *mask) cpumask 47 kernel/watchdog.c static struct cpumask watchdog_allowed_mask __read_mostly; cpumask 49 kernel/watchdog.c struct cpumask watchdog_cpumask __read_mostly; cpumask 27 kernel/watchdog_hld.c static struct cpumask dead_events_mask; cpumask 1852 kernel/workqueue.c set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); cpumask 1938 kernel/workqueue.c kthread_bind_mask(worker->task, pool->attrs->cpumask); cpumask 3344 kernel/workqueue.c free_cpumask_var(attrs->cpumask); cpumask 3364 kernel/workqueue.c if (!alloc_cpumask_var(&attrs->cpumask, GFP_KERNEL)) cpumask 3367 kernel/workqueue.c cpumask_copy(attrs->cpumask, cpu_possible_mask); cpumask 3378 kernel/workqueue.c cpumask_copy(to->cpumask, from->cpumask); cpumask 3393 kernel/workqueue.c hash = jhash(cpumask_bits(attrs->cpumask), cpumask 3404 kernel/workqueue.c if (!cpumask_equal(a->cpumask, b->cpumask)) cpumask 3607 kernel/workqueue.c if (cpumask_subset(attrs->cpumask, cpumask 3822 kernel/workqueue.c int cpu_going_down, cpumask_t *cpumask) cpumask 3828 kernel/workqueue.c cpumask_and(cpumask, cpumask_of_node(node), attrs->cpumask); cpumask 3830 kernel/workqueue.c cpumask_clear_cpu(cpu_going_down, cpumask); cpumask 3832 kernel/workqueue.c if (cpumask_empty(cpumask)) cpumask 3836 kernel/workqueue.c cpumask_and(cpumask, attrs->cpumask, wq_numa_possible_cpumask[node]); cpumask 3838 kernel/workqueue.c if (cpumask_empty(cpumask)) { cpumask 3844 kernel/workqueue.c return !cpumask_equal(cpumask, attrs->cpumask); cpumask 3847 kernel/workqueue.c cpumask_copy(cpumask, attrs->cpumask); cpumask 3918 kernel/workqueue.c cpumask_and(new_attrs->cpumask, new_attrs->cpumask, wq_unbound_cpumask); cpumask 3919 kernel/workqueue.c if (unlikely(cpumask_empty(new_attrs->cpumask))) cpumask 3920 kernel/workqueue.c cpumask_copy(new_attrs->cpumask, wq_unbound_cpumask); cpumask 3939 kernel/workqueue.c if (wq_calc_node_cpumask(new_attrs, node, -1, tmp_attrs->cpumask)) { cpumask 3951 kernel/workqueue.c cpumask_and(new_attrs->cpumask, new_attrs->cpumask, cpu_possible_mask); cpumask 4089 kernel/workqueue.c cpumask_t *cpumask; cpumask 4103 kernel/workqueue.c cpumask = target_attrs->cpumask; cpumask 4114 kernel/workqueue.c if (wq_calc_node_cpumask(wq->dfl_pwq->pool->attrs, node, cpu_off, cpumask)) { cpumask 4115 kernel/workqueue.c if (cpumask_equal(cpumask, pwq->pool->attrs->cpumask)) cpumask 4634 kernel/workqueue.c pr_cont(" cpus=%*pbl", nr_cpumask_bits, pool->attrs->cpumask); cpumask 4937 kernel/workqueue.c pool->attrs->cpumask) < 0); cpumask 4993 kernel/workqueue.c static cpumask_t cpumask; cpumask 4999 kernel/workqueue.c if (!cpumask_test_cpu(cpu, pool->attrs->cpumask)) cpumask 5002 kernel/workqueue.c cpumask_and(&cpumask, pool->attrs->cpumask, cpu_online_mask); cpumask 5006 kernel/workqueue.c WARN_ON_ONCE(set_cpus_allowed_ptr(worker->task, &cpumask) < 0); cpumask 5287 kernel/workqueue.c int workqueue_set_unbound_cpumask(cpumask_var_t cpumask) cpumask 5299 kernel/workqueue.c cpumask_and(cpumask, cpumask, cpu_possible_mask); cpumask 5300 kernel/workqueue.c if (!cpumask_empty(cpumask)) { cpumask 5307 kernel/workqueue.c cpumask_copy(wq_unbound_cpumask, cpumask); cpumask 5471 kernel/workqueue.c cpumask_pr_args(wq->unbound_attrs->cpumask)); cpumask 5490 kernel/workqueue.c ret = cpumask_parse(buf, attrs->cpumask); cpumask 5542 kernel/workqueue.c __ATTR(cpumask, 0644, wq_cpumask_show, wq_cpumask_store), cpumask 5568 kernel/workqueue.c cpumask_var_t cpumask; cpumask 5571 kernel/workqueue.c if (!zalloc_cpumask_var(&cpumask, GFP_KERNEL)) cpumask 5574 kernel/workqueue.c ret = cpumask_parse(buf, cpumask); cpumask 5576 kernel/workqueue.c ret = workqueue_set_unbound_cpumask(cpumask); cpumask 5578 kernel/workqueue.c free_cpumask_var(cpumask); cpumask 5583 kernel/workqueue.c __ATTR(cpumask, 0644, wq_unbound_cpumask_show, cpumask 5912 kernel/workqueue.c cpumask_copy(pool->attrs->cpumask, cpumask_of(cpu)); cpumask 95 lib/cpu_rmap.c const struct cpumask *mask, u16 dist) cpumask 156 lib/cpu_rmap.c const struct cpumask *affinity) cpumask 17 lib/cpumask.c unsigned int cpumask_next(int n, const struct cpumask *srcp) cpumask 34 lib/cpumask.c int cpumask_next_and(int n, const struct cpumask *src1p, cpumask 35 lib/cpumask.c const struct cpumask *src2p) cpumask 53 lib/cpumask.c int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) cpumask 77 lib/cpumask.c int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool wrap) cpumask 2633 mm/compaction.c const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); cpumask 2635 mm/compaction.c if (!cpumask_empty(cpumask)) cpumask 2636 mm/compaction.c set_cpus_allowed_ptr(tsk, cpumask); cpumask 2705 mm/compaction.c const struct cpumask *mask; cpumask 1771 mm/page_alloc.c const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); cpumask 1780 mm/page_alloc.c if (!cpumask_empty(cpumask)) cpumask 1781 mm/page_alloc.c set_cpus_allowed_ptr(current, cpumask); cpumask 5540 mm/page_alloc.c const struct cpumask *tmp = cpumask_of_node(0); cpumask 945 mm/slab.c const struct cpumask *mask = cpumask_of_node(node); cpumask 712 mm/swap.c static struct cpumask has_work; cpumask 3900 mm/vmscan.c const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); cpumask 3902 mm/vmscan.c if (!cpumask_empty(cpumask)) cpumask 3903 mm/vmscan.c set_cpus_allowed_ptr(tsk, cpumask); cpumask 4067 mm/vmscan.c const struct cpumask *mask; cpumask 1948 mm/vmstat.c const struct cpumask *node_cpus; cpumask 2438 net/core/dev.c int netif_set_xps_queue(struct net_device *dev, const struct cpumask *mask, cpumask 568 net/iucv/iucv.c cpumask_t cpumask; cpumask 572 net/iucv/iucv.c cpumask_copy(&cpumask, &iucv_irq_cpumask); cpumask 573 net/iucv/iucv.c cpumask_clear_cpu(cpumask_first(&iucv_irq_cpumask), &cpumask); cpumask 574 net/iucv/iucv.c for_each_cpu(cpu, &cpumask) cpumask 676 net/iucv/iucv.c cpumask_t cpumask; cpumask 681 net/iucv/iucv.c cpumask_copy(&cpumask, &iucv_buffer_cpumask); cpumask 682 net/iucv/iucv.c cpumask_clear_cpu(cpu, &cpumask); cpumask 683 net/iucv/iucv.c if (cpumask_empty(&cpumask)) cpumask 216 net/openvswitch/flow.h struct cpumask cpu_used_mask; cpumask 109 samples/bpf/cpustat_user.c cpu_set_t cpumask; cpumask 133 samples/bpf/cpustat_user.c CPU_ZERO(&cpumask); cpumask 134 samples/bpf/cpustat_user.c CPU_SET(i, &cpumask); cpumask 136 samples/bpf/cpustat_user.c sched_setaffinity(0, sizeof(cpumask), &cpumask); cpumask 232 samples/trace_events/trace-events-sample.h const char *string, const struct cpumask *mask), cpumask 73 tools/accounting/getdelays.c char cpumask[100+6*MAX_CPUS]; cpumask 323 tools/accounting/getdelays.c strncpy(cpumask, optarg, sizeof(cpumask)); cpumask 324 tools/accounting/getdelays.c cpumask[sizeof(cpumask) - 1] = '\0'; cpumask 326 tools/accounting/getdelays.c printf("cpumask %s maskset %d\n", cpumask, maskset); cpumask 401 tools/accounting/getdelays.c &cpumask, strlen(cpumask) + 1); cpumask 545 tools/accounting/getdelays.c &cpumask, strlen(cpumask) + 1); cpumask 173 tools/perf/builtin-ftrace.c char *cpumask; cpumask 182 tools/perf/builtin-ftrace.c cpumask = malloc(mask_size); cpumask 183 tools/perf/builtin-ftrace.c if (cpumask == NULL) { cpumask 188 tools/perf/builtin-ftrace.c cpu_map__snprint_mask(cpumap, cpumask, mask_size); cpumask 190 tools/perf/builtin-ftrace.c ret = write_tracing_file("tracing_cpumask", cpumask); cpumask 192 tools/perf/builtin-ftrace.c free(cpumask); cpumask 236 virt/kvm/kvm_main.c static inline bool kvm_kick_many_cpus(const struct cpumask *cpus, bool wait)