Home
last modified time | relevance | path

Searched refs:cpumask (Results 1 – 200 of 397) sorted by relevance

12

/linux-4.1.27/include/linux/
Dcpumask.h15 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; struct
88 extern const struct cpumask *const cpu_possible_mask;
89 extern const struct cpumask *const cpu_online_mask;
90 extern const struct cpumask *const cpu_present_mask;
91 extern const struct cpumask *const cpu_active_mask;
124 static inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first()
130 static inline unsigned int cpumask_next(int n, const struct cpumask *srcp) in cpumask_next()
135 static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp) in cpumask_next_zero()
141 const struct cpumask *srcp, in cpumask_next_and()
142 const struct cpumask *andp) in cpumask_next_and()
[all …]
Dstop_machine.h34 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg);
35 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg);
78 static inline int stop_cpus(const struct cpumask *cpumask, in stop_cpus() argument
81 if (cpumask_test_cpu(raw_smp_processor_id(), cpumask)) in stop_cpus()
86 static inline int try_stop_cpus(const struct cpumask *cpumask, in try_stop_cpus() argument
89 return stop_cpus(cpumask, fn, arg); in try_stop_cpus()
115 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus);
126 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus);
129 const struct cpumask *cpus);
134 const struct cpumask *cpus) in __stop_machine()
[all …]
Dcpu_cooling.h37 cpufreq_cooling_register(const struct cpumask *clip_cpus);
47 const struct cpumask *clip_cpus);
51 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
66 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register()
72 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
Dinterrupt.h227 extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask,
238 irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_set_affinity() argument
240 return __irq_set_affinity(irq, cpumask, false); in irq_set_affinity()
255 irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument
257 return __irq_set_affinity(irq, cpumask, true); in irq_force_affinity()
263 extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m);
270 static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) in irq_set_affinity()
275 static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument
288 const struct cpumask *m) in irq_set_affinity_hint()
Dpadata.h133 struct padata_cpumask cpumask; member
157 struct padata_cpumask cpumask; member
170 const struct cpumask *pcpumask,
171 const struct cpumask *cbcpumask);
177 cpumask_var_t cpumask);
Dsmp.h39 void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func,
96 void smp_call_function_many(const struct cpumask *mask,
99 int smp_call_function_any(const struct cpumask *mask,
145 smp_call_function_any(const struct cpumask *mask, smp_call_func_t func, in smp_call_function_any()
Dclockchips.h132 void (*broadcast)(const struct cpumask *mask);
142 const struct cpumask *cpumask; member
191 extern void tick_broadcast(const struct cpumask *mask);
Dcpuidle.h123 struct cpumask *cpumask; member
145 const struct cpumask *const coupled_cpus);
182 const struct cpumask *const coupled_cpus) in cpuidle_register()
Dtopology.h202 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask()
208 static inline const struct cpumask *cpu_cpu_mask(int cpu) in cpu_cpu_mask()
Dcpuset.h44 extern void cpuset_cpus_allowed(struct task_struct *p, struct cpumask *mask);
147 struct cpumask *mask) in cpuset_cpus_allowed()
Dirqdesc.h67 struct cpumask *percpu_enabled;
69 const struct cpumask *affinity_hint;
Dcpu_rmap.h43 const struct cpumask *affinity);
Dirq.h345 int (*irq_set_affinity)(struct irq_data *data, const struct cpumask *dest, bool force);
424 const struct cpumask *cpumask, bool force);
467 const struct cpumask *dest,
Dsched.h1035 static inline struct cpumask *sched_domain_span(struct sched_domain *sd) in sched_domain_span()
1049 typedef const struct cpumask *(*sched_domain_mask_f)(int cpu);
2120 extern int cpuset_cpumask_can_shrink(const struct cpumask *cur,
2121 const struct cpumask *trial);
2123 const struct cpumask *cs_cpus_allowed);
2126 const struct cpumask *new_mask);
2129 const struct cpumask *new_mask);
2132 const struct cpumask *new_mask) in do_set_cpus_allowed()
2136 const struct cpumask *new_mask) in set_cpus_allowed_ptr()
3015 extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
[all …]
/linux-4.1.27/kernel/
Dpadata.c40 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu()
42 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu()
58 cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash()
121 if (!cpumask_test_cpu(cb_cpu, pd->cpumask.cbcpu)) in padata_do_parallel()
177 num_cpus = cpumask_weight(pd->cpumask.pcpu); in padata_get_next()
351 const struct cpumask *pcpumask, in padata_setup_cpumasks()
352 const struct cpumask *cbcpumask) in padata_setup_cpumasks()
354 if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) in padata_setup_cpumasks()
357 cpumask_and(pd->cpumask.pcpu, pcpumask, cpu_online_mask); in padata_setup_cpumasks()
358 if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL)) { in padata_setup_cpumasks()
[all …]
Dstop_machine.c146 const struct cpumask *active_cpus;
330 static void queue_stop_cpus_work(const struct cpumask *cpumask, in queue_stop_cpus_work() argument
338 for_each_cpu(cpu, cpumask) { in queue_stop_cpus_work()
351 for_each_cpu(cpu, cpumask) in queue_stop_cpus_work()
356 static int __stop_cpus(const struct cpumask *cpumask, in __stop_cpus() argument
361 cpu_stop_init_done(&done, cpumask_weight(cpumask)); in __stop_cpus()
362 queue_stop_cpus_work(cpumask, fn, arg, &done); in __stop_cpus()
395 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in stop_cpus() argument
401 ret = __stop_cpus(cpumask, fn, arg); in stop_cpus()
424 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in try_stop_cpus() argument
[all …]
Dsmp.c27 cpumask_var_t cpumask; member
45 if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, in hotplug_cfd()
50 free_cpumask_var(cfd->cpumask); in hotplug_cfd()
62 free_cpumask_var(cfd->cpumask); in hotplug_cfd()
361 int smp_call_function_any(const struct cpumask *mask, in smp_call_function_any()
365 const struct cpumask *nodemask; in smp_call_function_any()
404 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many()
441 cpumask_and(cfd->cpumask, mask, cpu_online_mask); in smp_call_function_many()
442 cpumask_clear_cpu(this_cpu, cfd->cpumask); in smp_call_function_many()
445 if (unlikely(!cpumask_weight(cfd->cpumask))) in smp_call_function_many()
[all …]
Dworkqueue.c1677 set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); in worker_attach_to_pool()
3065 free_cpumask_var(attrs->cpumask); in free_workqueue_attrs()
3086 if (!alloc_cpumask_var(&attrs->cpumask, gfp_mask)) in alloc_workqueue_attrs()
3089 cpumask_copy(attrs->cpumask, cpu_possible_mask); in alloc_workqueue_attrs()
3100 cpumask_copy(to->cpumask, from->cpumask); in copy_workqueue_attrs()
3115 hash = jhash(cpumask_bits(attrs->cpumask), in wqattrs_hash()
3126 if (!cpumask_equal(a->cpumask, b->cpumask)) in wqattrs_equal()
3306 if (cpumask_subset(pool->attrs->cpumask, in get_unbound_pool()
3502 int cpu_going_down, cpumask_t *cpumask) in wq_calc_node_cpumask() argument
3508 cpumask_and(cpumask, cpumask_of_node(node), attrs->cpumask); in wq_calc_node_cpumask()
[all …]
Dcpu.c763 const struct cpumask *const cpu_possible_mask = to_cpumask(cpu_possible_bits);
767 const struct cpumask *const cpu_online_mask = to_cpumask(cpu_online_bits);
771 const struct cpumask *const cpu_present_mask = to_cpumask(cpu_present_bits);
775 const struct cpumask *const cpu_active_mask = to_cpumask(cpu_active_bits);
812 void init_cpu_present(const struct cpumask *src) in init_cpu_present()
817 void init_cpu_possible(const struct cpumask *src) in init_cpu_possible()
822 void init_cpu_online(const struct cpumask *src) in init_cpu_online()
Dup.c53 void on_each_cpu_mask(const struct cpumask *mask, in on_each_cpu_mask()
/linux-4.1.27/arch/m32r/kernel/
Dsmp.c80 static void send_IPI_mask(const struct cpumask *, int, int);
151 cpumask_t cpumask; in smp_flush_cache_all() local
155 cpumask_copy(&cpumask, cpu_online_mask); in smp_flush_cache_all()
156 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_flush_cache_all()
158 mask=cpumask_bits(&cpumask); in smp_flush_cache_all()
160 send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); in smp_flush_cache_all()
370 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
388 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
390 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
394 cpumask_and(&cpumask, &cpumask, cpu_online_mask); in flush_tlb_others()
[all …]
/linux-4.1.27/arch/x86/include/asm/uv/
Duv.h6 struct cpumask;
16 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask,
28 static inline const struct cpumask *
29 uv_flush_tlb_others(const struct cpumask *cpumask, struct mm_struct *mm, in uv_flush_tlb_others() argument
31 { return cpumask; } in uv_flush_tlb_others()
/linux-4.1.27/arch/x86/include/asm/
Dapic.h278 const struct cpumask *(*target_cpus)(void);
285 void (*vector_allocation_domain)(int cpu, struct cpumask *retmask,
286 const struct cpumask *mask);
301 int (*cpu_mask_to_apicid_and)(const struct cpumask *cpumask,
302 const struct cpumask *andmask,
306 void (*send_IPI_mask)(const struct cpumask *mask, int vector);
307 void (*send_IPI_mask_allbutself)(const struct cpumask *mask,
480 static inline const struct cpumask *default_target_cpus(void) in default_target_cpus()
489 static inline const struct cpumask *online_target_cpus(void) in online_target_cpus()
546 flat_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in flat_cpu_mask_to_apicid_and() argument
[all …]
Dipi.h122 extern void default_send_IPI_mask_sequence_phys(const struct cpumask *mask,
124 extern void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask,
149 extern void default_send_IPI_mask_sequence_logical(const struct cpumask *mask,
151 extern void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask,
153 extern void default_send_IPI_mask_logical(const struct cpumask *mask,
Dsmp.h40 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
45 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask()
50 static inline struct cpumask *cpu_llc_shared_mask(int cpu) in cpu_llc_shared_mask()
79 void (*send_call_func_ipi)(const struct cpumask *mask);
147 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
166 void native_send_call_func_ipi(const struct cpumask *mask);
Dtopology.h76 extern const struct cpumask *cpumask_of_node(int node);
79 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
120 extern const struct cpumask *cpu_coregroup_mask(int cpu);
Dtlbflush.h212 static inline void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
249 void native_flush_tlb_others(const struct cpumask *cpumask,
Dmsr.h225 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
226 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
254 static inline void rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, in rdmsr_on_cpus()
259 static inline void wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, in wrmsr_on_cpus()
Dx86_init.h190 struct cpumask;
200 const struct cpumask *mask,
Dhw_irq.h142 extern int assign_irq_vector(int, struct irq_cfg *, const struct cpumask *);
155 extern int apic_set_affinity(struct irq_data *data, const struct cpumask *mask,
Dpci.h135 static inline const struct cpumask *
/linux-4.1.27/drivers/cpuidle/
Dcpuidle-big_little.c149 struct cpumask *cpumask; in bl_idle_driver_init() local
152 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); in bl_idle_driver_init()
153 if (!cpumask) in bl_idle_driver_init()
158 cpumask_set_cpu(cpu, cpumask); in bl_idle_driver_init()
160 drv->cpumask = cpumask; in bl_idle_driver_init()
228 kfree(bl_idle_big_driver.cpumask); in bl_idle_init()
230 kfree(bl_idle_little_driver.cpumask); in bl_idle_init()
Ddt_idle_states.c99 const cpumask_t *cpumask) in idle_state_valid() argument
112 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid()
113 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid()
158 const cpumask_t *cpumask; in dt_init_idle_driver() local
169 cpumask = drv->cpumask ? : cpu_possible_mask; in dt_init_idle_driver()
170 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver()
180 if (!idle_state_valid(state_node, i, cpumask)) { in dt_init_idle_driver()
Ddriver.c50 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_unset_driver()
72 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_set_driver()
164 if (!drv->cpumask) in __cpuidle_driver_init()
165 drv->cpumask = (struct cpumask *)cpu_possible_mask; in __cpuidle_driver_init()
240 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_register_driver()
264 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_unregister_driver()
Dcpuidle.c526 for_each_cpu(cpu, drv->cpumask) { in cpuidle_unregister()
547 const struct cpumask *const coupled_cpus) in cpuidle_register()
558 for_each_cpu(cpu, drv->cpumask) { in cpuidle_register()
/linux-4.1.27/arch/x86/kernel/apic/
Dx2apic_cluster.c27 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
29 struct cpumask *cpus_in_cluster_ptr; in __x2apic_send_IPI_mask()
30 struct cpumask *ipi_mask_ptr; in __x2apic_send_IPI_mask()
77 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
83 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
99 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument
100 const struct cpumask *andmask, in x2apic_cpu_mask_to_apicid_and()
107 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
118 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
211 static const struct cpumask *x2apic_cluster_target_cpus(void) in x2apic_cluster_target_cpus()
[all …]
Dapic_flat_64.c65 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument
67 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask()
73 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument
75 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself()
234 static void physflat_send_IPI_mask(const struct cpumask *cpumask, int vector) in physflat_send_IPI_mask() argument
236 default_send_IPI_mask_sequence_phys(cpumask, vector); in physflat_send_IPI_mask()
239 static void physflat_send_IPI_mask_allbutself(const struct cpumask *cpumask, in physflat_send_IPI_mask_allbutself() argument
242 default_send_IPI_mask_allbutself_phys(cpumask, vector); in physflat_send_IPI_mask_allbutself()
Dapic_noop.c33 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask() argument
34 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask_allbutself() argument
86 static const struct cpumask *noop_target_cpus(void) in noop_target_cpus()
92 static void noop_vector_allocation_domain(int cpu, struct cpumask *retmask, in noop_vector_allocation_domain()
93 const struct cpumask *mask) in noop_vector_allocation_domain()
Dipi.c21 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys()
39 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys()
60 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical()
80 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical()
103 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument
105 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
Dx2apic_phys.c38 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
58 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
64 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
Dmsi.c79 msi_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force) in msi_set_affinity()
180 dmar_msi_set_affinity(struct irq_data *data, const struct cpumask *mask, in dmar_msi_set_affinity()
236 const struct cpumask *mask, bool force) in hpet_msi_set_affinity()
Dapic_numachip.c112 static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) in numachip_send_IPI_mask()
120 static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, in numachip_send_IPI_mask_allbutself()
Dx2apic_uv_x.c265 static void uv_send_IPI_mask(const struct cpumask *mask, int vector) in uv_send_IPI_mask()
273 static void uv_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in uv_send_IPI_mask_allbutself()
315 uv_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in uv_cpu_mask_to_apicid_and() argument
316 const struct cpumask *andmask, in uv_cpu_mask_to_apicid_and()
325 for_each_cpu_and(cpu, cpumask, andmask) { in uv_cpu_mask_to_apicid_and()
Dhtirq.c40 ht_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force) in ht_set_affinity()
/linux-4.1.27/arch/tile/include/asm/
Dsmp.h32 void send_IPI_many(const struct cpumask *mask, int tag);
75 static inline void arch_send_call_function_ipi_mask(struct cpumask *mask) in arch_send_call_function_ipi_mask()
102 extern struct cpumask cpu_lotar_map;
106 extern struct cpumask hash_for_home_map;
109 extern struct cpumask cpu_cacheable_map;
127 static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp, in __cpulist_parse_crop()
Dhomecache.h51 const struct cpumask *cache_cpumask,
54 const struct cpumask *tlb_cpumask,
61 extern void homecache_evict(const struct cpumask *mask);
Dtopology.h23 extern struct cpumask node_2_cpu_mask[];
39 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
/linux-4.1.27/arch/mn10300/mm/
Dtlb-smp.c52 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm,
94 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
105 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
106 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
108 cpumask_and(&tmp, &cpumask, cpu_online_mask); in flush_tlb_others()
109 BUG_ON(!cpumask_equal(&cpumask, &tmp)); in flush_tlb_others()
122 atomic_set_mask(cpumask.bits[0], &flush_cpumask.bits[0]); in flush_tlb_others()
/linux-4.1.27/kernel/time/
Dtick-common.c168 const struct cpumask *cpumask) in tick_setup_device() argument
206 if (!cpumask_equal(newdev->cpumask, cpumask)) in tick_setup_device()
207 irq_set_affinity(newdev->irq, cpumask); in tick_setup_device()
239 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_percpu()
241 if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) in tick_check_percpu()
247 if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) in tick_check_percpu()
269 !cpumask_equal(curdev->cpumask, newdev->cpumask); in tick_check_preferred()
296 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_new_device()
Dtick-broadcast.c54 struct cpumask *tick_get_broadcast_mask(void) in tick_get_broadcast_mask()
138 static void err_broadcast(const struct cpumask *mask) in err_broadcast()
258 static void tick_do_broadcast(struct cpumask *mask) in tick_do_broadcast()
502 struct cpumask *tick_get_broadcast_oneshot_mask(void) in tick_get_broadcast_oneshot_mask()
523 const struct cpumask *cpumask) in tick_broadcast_set_affinity() argument
528 if (cpumask_equal(bc->cpumask, cpumask)) in tick_broadcast_set_affinity()
531 bc->cpumask = cpumask; in tick_broadcast_set_affinity()
532 irq_set_affinity(bc->irq, bc->cpumask); in tick_broadcast_set_affinity()
823 static void tick_broadcast_init_next_event(struct cpumask *mask, in tick_broadcast_init_next_event()
Dtick-internal.h63 extern struct cpumask *tick_get_broadcast_mask(void);
124 extern struct cpumask *tick_get_broadcast_oneshot_mask(void);
Dtick-broadcast-hrtimer.c95 .cpumask = cpu_all_mask,
Dclockevents.c485 if (!dev->cpumask) { in clockevents_register_device()
487 dev->cpumask = cpumask_of(smp_processor_id()); in clockevents_register_device()
665 if (cpumask_test_cpu(cpu, dev->cpumask) && in tick_cleanup_dead_cpu()
666 cpumask_weight(dev->cpumask) == 1 && in tick_cleanup_dead_cpu()
/linux-4.1.27/arch/tile/kernel/
Dhardwall.c105 struct cpumask cpumask; /* cpus reserved */ member
202 static int check_rectangle(struct hardwall_info *r, struct cpumask *mask) in check_rectangle()
296 struct cpumask rect_cpus; in hardwall_protect_rectangle()
351 if (cpumask_test_cpu(cpu, &rect->cpumask)) in do_hardwall_trap()
477 return info && cpumask_test_cpu(cpu, &info->cpumask); in hardwall_ipi_valid()
493 struct cpumask mask; in hardwall_create()
502 if (copy_from_user(&mask, bits, min(sizeof(struct cpumask), size))) in hardwall_create()
510 if (size < sizeof(struct cpumask)) { in hardwall_create()
511 memset((char *)&mask + size, 0, sizeof(struct cpumask) - size); in hardwall_create()
512 } else if (size > sizeof(struct cpumask)) { in hardwall_create()
[all …]
Dsmpboot.c124 static __initdata struct cpumask init_affinity;
135 static struct cpumask cpu_started;
Dsetup.c847 struct cpumask node_2_cpu_mask[MAX_NUMNODES] __write_once;
855 static int __init cpu_to_bound_node(int cpu, struct cpumask* unbound_cpus) in cpu_to_bound_node()
865 struct cpumask *unbound_cpus) in node_neighbors()
889 struct cpumask unbound_cpus; in setup_numa_mapping()
1274 struct cpumask __write_once cpu_lotar_map;
1284 struct cpumask hash_for_home_map;
1296 struct cpumask __write_once cpu_cacheable_map;
1299 static __initdata struct cpumask disabled_map;
1325 struct cpumask hv_disabled_map, cpu_possible_init; in setup_cpu_maps()
1398 struct cpumask shared; in check_nohz_full_cpus()
Dsmp.c73 void send_IPI_many(const struct cpumask *mask, int tag) in send_IPI_many()
92 struct cpumask mask; in send_IPI_allbutself()
/linux-4.1.27/arch/hexagon/include/asm/
Dsmp.h36 extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg);
39 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/tile/mm/
Dmigrate.h33 const unsigned long *cpumask);
51 const struct cpumask *cache_cpumask,
52 const struct cpumask *tlb_cpumask,
Dhomecache.c64 static void hv_flush_update(const struct cpumask *cache_cpumask, in hv_flush_update()
65 struct cpumask *tlb_cpumask, in hv_flush_update()
69 struct cpumask mask; in hv_flush_update()
108 const struct cpumask *cache_cpumask_orig, in flush_remote()
111 const struct cpumask *tlb_cpumask_orig, in flush_remote()
115 struct cpumask cache_cpumask_copy, tlb_cpumask_copy; in flush_remote()
116 struct cpumask *cache_cpumask, *tlb_cpumask; in flush_remote()
238 void homecache_evict(const struct cpumask *mask) in homecache_evict()
Dinit.c203 static __initdata struct cpumask kdata_mask;
318 static struct cpumask __initdata ktext_mask;
412 const struct cpumask *my_cpu_mask = cpumask_of(smp_processor_id()); in kernel_physical_mapping_init()
413 struct cpumask kstripe_mask; in kernel_physical_mapping_init()
493 struct cpumask bad; in kernel_physical_mapping_init()
/linux-4.1.27/arch/ia64/kernel/
Dsmp.c169 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask()
259 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local
265 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask()
269 for_each_cpu(cpu, &cpumask) { in smp_flush_tlb_cpumask()
279 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask()
324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dirq.c76 bool is_affinity_mask_valid(const struct cpumask *cpumask) in is_affinity_mask_valid() argument
80 if (cpumask_weight(cpumask) != 1) in is_affinity_mask_valid()
/linux-4.1.27/arch/mn10300/kernel/
Dsmp.c107 static void send_IPI_mask(const cpumask_t *cpumask, int irq);
313 static void send_IPI_mask(const cpumask_t *cpumask, int irq) in send_IPI_mask() argument
319 if (cpumask_test_cpu(i, cpumask)) { in send_IPI_mask()
350 cpumask_t cpumask; in send_IPI_allbutself() local
352 cpumask_copy(&cpumask, cpu_online_mask); in send_IPI_allbutself()
353 cpumask_clear_cpu(smp_processor_id(), &cpumask); in send_IPI_allbutself()
354 send_IPI_mask(&cpumask, irq); in send_IPI_allbutself()
357 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
1067 static int hotplug_cpu_nmi_call_function(cpumask_t cpumask, in hotplug_cpu_nmi_call_function() argument
1084 nmi_call_func_mask_data.started = cpumask; in hotplug_cpu_nmi_call_function()
[all …]
/linux-4.1.27/Documentation/
Dpadata.txt17 const struct cpumask *pcpumask,
18 const struct cpumask *cbcpumask);
35 it is legal to supply a cpumask to padata that contains offline CPUs.
36 Once an offline CPU in the user supplied cpumask comes online, padata
47 padata cpumask contains no active CPU (flag not set).
57 cpumask_var_t cpumask);
68 PADATA_CPU_PARALLEL and cpumask specifies the new cpumask to use.
69 To simply add or remove one CPU from a certain cpumask the functions
73 If a user is interested in padata cpumask changes, he can register to
74 the padata cpumask change notifier:
[all …]
/linux-4.1.27/arch/arm64/include/asm/
Dsmp.h50 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int));
52 extern void (*__smp_cross_call)(const struct cpumask *, unsigned int);
69 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h25 const struct cpumask *cpu_coregroup_mask(int cpu);
/linux-4.1.27/arch/hexagon/kernel/
Dsmp.c105 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument
113 for_each_cpu(cpu, cpumask) { in send_ipi()
239 struct cpumask targets; in smp_send_stop()
250 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dtime.c116 static void broadcast(const struct cpumask *mask) in broadcast()
148 dummy_clock_dev->cpumask = cpumask_of(cpu); in setup_percpu_clockdev()
195 ce_dev->cpumask = cpu_all_mask; in time_init_deferred()
/linux-4.1.27/arch/x86/mm/
Dtlb.c133 void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
147 cpumask = uv_flush_tlb_others(cpumask, mm, start, end, cpu); in native_flush_tlb_others()
148 if (cpumask) in native_flush_tlb_others()
149 smp_call_function_many(cpumask, flush_tlb_func, in native_flush_tlb_others()
153 smp_call_function_many(cpumask, flush_tlb_func, &info, 1); in native_flush_tlb_others()
/linux-4.1.27/Documentation/accounting/
Dgetdelays.c72 char cpumask[100+6*MAX_CPUS]; variable
316 strncpy(cpumask, optarg, sizeof(cpumask)); in main()
317 cpumask[sizeof(cpumask) - 1] = '\0'; in main()
319 printf("cpumask %s maskset %d\n", cpumask, maskset); in main()
393 &cpumask, strlen(cpumask) + 1); in main()
534 &cpumask, strlen(cpumask) + 1); in main()
Dtaskstats.txt35 sends a register command and specifies a cpumask. Whenever a task exits on
36 one of the cpus in the cpumask, its per-pid statistics are sent to the
84 TASKSTATS_CMD_ATTR_REGISTER/DEREGISTER_CPUMASK and contain a cpumask in the
85 attribute payload. The cpumask is specified as an ascii string of
87 the cpumask would be "1-3,5,7-8". If userspace forgets to deregister interest
/linux-4.1.27/arch/blackfin/mach-common/
Dsmp.c189 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument
196 for_each_cpu(cpu, cpumask) { in send_ipi()
203 for_each_cpu(cpu, cpumask) in send_ipi()
212 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
224 void smp_send_msg(const struct cpumask *mask, unsigned long type) in smp_send_msg()
229 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
/linux-4.1.27/arch/arm/mach-vexpress/
Ddcscb.c43 unsigned int rst_hold, cpumask = (1 << cpu); in dcscb_cpu_powerup() local
46 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) in dcscb_cpu_powerup()
50 rst_hold &= ~(cpumask | (cpumask << 4)); in dcscb_cpu_powerup()
/linux-4.1.27/arch/powerpc/include/asm/
Dsmp.h96 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
101 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask()
139 static inline const struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
189 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dxics.h156 extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask,
159 #define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server) argument
Dirq.h71 int irq_choose_cpu(const struct cpumask *mask);
Dcputhreads.h42 static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads) in cpu_thread_mask_to_cores()
Dfadump.h194 struct cpumask cpu_online_mask;
/linux-4.1.27/arch/arm/include/asm/
Dsmp.h49 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int));
79 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
80 extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask);
Dtopology.h25 const struct cpumask *cpu_coregroup_mask(int cpu);
/linux-4.1.27/drivers/irqchip/
Dirq-metag.c44 const struct cpumask *cpumask, bool force);
193 const struct cpumask *cpumask, bool force) in metag_internal_irq_set_affinity() argument
204 cpu = cpumask_any_and(cpumask, cpu_online_mask); in metag_internal_irq_set_affinity()
Dirq-hip04.c145 const struct cpumask *mask_val, in hip04_irq_set_affinity()
228 u32 cpumask; in hip04_irq_dist_init() local
237 cpumask = hip04_get_cpumask(intc); in hip04_irq_dist_init()
238 cpumask |= cpumask << 16; in hip04_irq_dist_init()
240 writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3)); in hip04_irq_dist_init()
276 static void hip04_raise_softirq(const struct cpumask *mask, unsigned int irq) in hip04_raise_softirq()
Dirq-gic.c237 static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, in gic_set_affinity()
383 u32 cpumask; in gic_dist_init() local
392 cpumask = gic_get_cpumask(gic); in gic_dist_init()
393 cpumask |= cpumask << 8; in gic_dist_init()
394 cpumask |= cpumask << 16; in gic_dist_init()
396 writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4); in gic_dist_init()
632 static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq) in gic_raise_softirq()
Dirq-mips-gic.c386 static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask, in gic_set_affinity() argument
394 cpumask_and(&tmp, cpumask, cpu_online_mask); in gic_set_affinity()
409 cpumask_copy(d->affinity, cpumask); in gic_set_affinity()
Dirq-metag-ext.c505 const struct cpumask *cpumask, bool force) in meta_intc_set_affinity() argument
518 cpu = cpumask_any_and(cpumask, cpu_online_mask); in meta_intc_set_affinity()
/linux-4.1.27/Documentation/thermal/
Dcpu-cooling-api.txt20 struct cpumask *clip_cpus)
26 clip_cpus: cpumask of cpus where the frequency constraints will happen.
29 struct device_node *np, const struct cpumask *clip_cpus)
37 clip_cpus: cpumask of cpus where the frequency constraints will happen.
/linux-4.1.27/arch/ia64/include/asm/
Dirq.h30 bool is_affinity_mask_valid(const struct cpumask *cpumask);
Dsmp.h129 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/x86/xen/
Dsmp.h3 extern void xen_send_IPI_mask(const struct cpumask *mask,
5 extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
Dsmp.c564 static void __xen_send_IPI_mask(const struct cpumask *mask, in __xen_send_IPI_mask()
573 static void xen_smp_send_call_function_ipi(const struct cpumask *mask) in xen_smp_send_call_function_ipi()
626 void xen_send_IPI_mask(const struct cpumask *mask, in xen_send_IPI_mask()
651 void xen_send_IPI_mask_allbutself(const struct cpumask *mask, in xen_send_IPI_mask_allbutself()
/linux-4.1.27/arch/alpha/kernel/
Dsmp.c176 wait_for_txrdy (unsigned long cpumask) in wait_for_txrdy() argument
180 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy()
185 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy()
203 unsigned long cpumask; in send_secondary_console_msg() local
211 cpumask = (1UL << cpuid); in send_secondary_console_msg()
212 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg()
225 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg()
509 send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation) in send_ipi_message()
596 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dtime.c131 .cpumask = cpumask_of(cpu), in init_rtc_clockevent()
199 .cpumask = cpumask_of(cpu), in init_qemu_clockevent()
/linux-4.1.27/lib/
Dcpumask.c16 int cpumask_next_and(int n, const struct cpumask *src1p, in cpumask_next_and()
17 const struct cpumask *src2p) in cpumask_next_and()
34 int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) in cpumask_any_but()
Dcpu_rmap.c98 const struct cpumask *mask, u16 dist) in cpu_rmap_copy_neigh()
159 const struct cpumask *affinity) in cpu_rmap_update()
/linux-4.1.27/kernel/irq/
Dmanage.c164 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) in irq_copy_pending()
169 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) in irq_get_pending()
177 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) { } in irq_copy_pending()
179 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) { } in irq_get_pending()
182 int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, in irq_do_set_affinity()
202 int irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask, in irq_set_affinity_locked()
228 int __irq_set_affinity(unsigned int irq, const struct cpumask *mask, bool force) in __irq_set_affinity()
243 int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m) in irq_set_affinity_hint()
264 cpumask_var_t cpumask; in irq_affinity_notify() local
267 if (!desc || !alloc_cpumask_var(&cpumask, GFP_KERNEL)) in irq_affinity_notify()
[all …]
Dinternals.h113 extern int irq_select_affinity_usr(unsigned int irq, struct cpumask *mask);
118 const struct cpumask *dest, bool force);
/linux-4.1.27/arch/arm/kernel/
Dsmp.c446 static void (*__smp_cross_call)(const struct cpumask *, unsigned int);
448 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument
466 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call()
498 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
503 void arch_send_wakeup_ipi_mask(const struct cpumask *mask) in arch_send_wakeup_ipi_mask()
522 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
650 struct cpumask mask; in smp_send_stop()
Dtopology.c183 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
192 const struct cpumask *cpu_corepower_mask(int cpu) in cpu_corepower_mask()
/linux-4.1.27/arch/xtensa/include/asm/
Dsmp.h22 struct cpumask;
23 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/powerpc/sysdev/xics/
Dics-rtas.c122 const struct cpumask *cpumask, in ics_rtas_set_affinity() argument
141 irq_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_rtas_set_affinity()
144 __func__, cpumask_pr_args(cpumask), d->irq); in ics_rtas_set_affinity()
Dics-opal.c111 const struct cpumask *cpumask, in ics_opal_set_affinity() argument
132 wanted_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_opal_set_affinity()
135 __func__, cpumask_pr_args(cpumask), d->irq); in ics_opal_set_affinity()
Dxics-common.c272 int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, in xics_get_irq_server() argument
279 if (!cpumask_subset(cpu_possible_mask, cpumask)) { in xics_get_irq_server()
280 int server = cpumask_first_and(cpu_online_mask, cpumask); in xics_get_irq_server()
/linux-4.1.27/drivers/cpufreq/
Dacpi-cpufreq.c103 static void boost_set_msrs(bool enable, const struct cpumask *cpumask) in boost_set_msrs() argument
122 rdmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs()
124 for_each_cpu(cpu, cpumask) { in boost_set_msrs()
132 wrmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs()
256 const struct cpumask *mask;
330 static u32 get_cur_val(const struct cpumask *mask) in get_cur_val()
393 static unsigned int check_freqs(const struct cpumask *mask, unsigned int freq, in check_freqs()
525 const struct cpumask *cpumask; in boost_notify() local
527 cpumask = get_cpu_mask(cpu); in boost_notify()
539 boost_set_msrs(acpi_cpufreq_driver.boost_enabled, cpumask); in boost_notify()
[all …]
Dpowernow-k8.h40 struct cpumask *available_cores;
/linux-4.1.27/arch/blackfin/include/asm/
Dsmp.h22 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
51 void smp_timer_broadcast(const struct cpumask *mask);
/linux-4.1.27/arch/alpha/include/asm/
Dtopology.h25 extern struct cpumask node_to_cpumask_map[];
27 static const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
Dsmp.h48 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/drivers/sh/intc/
Dchip.c69 const struct cpumask *cpumask, in intc_set_affinity() argument
72 if (!cpumask_intersects(cpumask, cpu_online_mask)) in intc_set_affinity()
75 cpumask_copy(data->affinity, cpumask); in intc_set_affinity()
/linux-4.1.27/drivers/iommu/
Dirq_remapping.h30 struct cpumask;
61 int (*set_affinity)(struct irq_data *data, const struct cpumask *mask,
Dirq_remapping.c31 const struct cpumask *mask,
270 const struct cpumask *mask, bool force) in set_remapped_irq_affinity()
/linux-4.1.27/kernel/events/
Dhw_breakpoint.c133 static const struct cpumask *cpumask_of_bp(struct perf_event *bp) in cpumask_of_bp()
148 const struct cpumask *cpumask = cpumask_of_bp(bp); in fetch_bp_busy_slots() local
151 for_each_cpu(cpu, cpumask) { in fetch_bp_busy_slots()
206 const struct cpumask *cpumask = cpumask_of_bp(bp); in toggle_bp_slot() local
219 for_each_cpu(cpu, cpumask) in toggle_bp_slot()
/linux-4.1.27/crypto/
Dpcrypt.c76 struct pcrypt_cpumask *cpumask; in pcrypt_do_parallel() local
81 cpumask = rcu_dereference_bh(pcrypt->cb_cpumask); in pcrypt_do_parallel()
82 if (cpumask_test_cpu(cpu, cpumask->mask)) in pcrypt_do_parallel()
85 if (!cpumask_weight(cpumask->mask)) in pcrypt_do_parallel()
88 cpu_index = cpu % cpumask_weight(cpumask->mask); in pcrypt_do_parallel()
90 cpu = cpumask_first(cpumask->mask); in pcrypt_do_parallel()
92 cpu = cpumask_next(cpu, cpumask->mask); in pcrypt_do_parallel()
413 struct padata_cpumask *cpumask = (struct padata_cpumask *)data; in pcrypt_cpumask_change_notify() local
429 cpumask_copy(new_mask->mask, cpumask->cbcpu); in pcrypt_cpumask_change_notify()
/linux-4.1.27/arch/sh/include/asm/
Dsmp.h37 void smp_timer_broadcast(const struct cpumask *mask);
44 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
/linux-4.1.27/arch/powerpc/sysdev/
Dmpic.h40 const struct cpumask *cpumask, bool force);
/linux-4.1.27/arch/mips/include/asm/
Dsmp-ops.h26 void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action);
48 extern void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action);
Dsmp.h95 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/linux-4.1.27/arch/arc/include/asm/
Dsmp.h21 struct cpumask;
27 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/mips/pmcs-msp71xx/
Dmsp_irq_cic.c129 const struct cpumask *cpumask, bool force) in msp_cic_irq_set_affinity() argument
143 if (cpumask_test_cpu(cpu, cpumask)) in msp_cic_irq_set_affinity()
Dmsp_irq_per.c89 const struct cpumask *affinity, bool force) in msp_per_irq_set_affinity()
/linux-4.1.27/arch/arm64/kernel/
Dsmp.c475 void (*__smp_cross_call)(const struct cpumask *, unsigned int);
477 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument
491 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call()
522 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
625 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
Dirq.c64 const struct cpumask *affinity = d->affinity; in migrate_one_irq()
/linux-4.1.27/arch/s390/kernel/
Dsmp.c155 static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address) in pcpu_find_address()
388 static void smp_emergency_stop(cpumask_t *cpumask) in smp_emergency_stop() argument
394 for_each_cpu(cpu, cpumask) { in smp_emergency_stop()
403 for_each_cpu(cpu, cpumask) in smp_emergency_stop()
405 cpumask_clear_cpu(cpu, cpumask); in smp_emergency_stop()
406 if (cpumask_empty(cpumask)) in smp_emergency_stop()
417 cpumask_t cpumask; in smp_send_stop() local
425 cpumask_copy(&cpumask, cpu_online_mask); in smp_send_stop()
426 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_send_stop()
429 smp_emergency_stop(&cpumask); in smp_send_stop()
[all …]
Dtopology.c424 static const struct cpumask *cpu_thread_mask(int cpu) in cpu_thread_mask()
430 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
435 static const struct cpumask *cpu_book_mask(int cpu) in cpu_book_mask()
/linux-4.1.27/arch/arc/kernel/
Dsmp.c242 static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) in ipi_send_msg()
257 struct cpumask targets; in smp_send_stop()
268 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/linux-4.1.27/arch/sparc/kernel/
Dtime_32.c139 ce->cpumask = cpu_possible_mask; in setup_timer_ce()
197 int cpu = cpumask_first(evt->cpumask); in percpu_ce_setup()
217 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_next_event()
237 ce->cpumask = cpumask_of(cpu); in register_percpu_ce()
/linux-4.1.27/drivers/thermal/
Dcpu_cooling.c72 struct cpumask allowed_cpus;
325 const struct cpumask *clip_cpus) in __cpufreq_cooling_register()
423 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register()
444 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
/linux-4.1.27/arch/powerpc/platforms/cell/
Dspu_priv1_mmio.c81 const struct cpumask *spumask = cpumask_of_node(spu->node), in cpu_affinity_set()
82 *cpumask = cpumask_of_node(cpu_to_node(cpu)); in cpu_affinity_set() local
84 if (!cpumask_intersects(spumask, cpumask)) in cpu_affinity_set()
/linux-4.1.27/kernel/trace/
Dring_buffer.c324 for_each_cpu(cpu, buffer->cpumask)
498 cpumask_var_t cpumask; member
567 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_wait()
663 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_poll_wait()
1339 if (!alloc_cpumask_var(&buffer->cpumask, GFP_KERNEL)) in __ring_buffer_alloc()
1361 cpumask_copy(buffer->cpumask, cpu_online_mask); in __ring_buffer_alloc()
1363 cpumask_copy(buffer->cpumask, cpu_possible_mask); in __ring_buffer_alloc()
1399 free_cpumask_var(buffer->cpumask); in __ring_buffer_alloc()
1432 free_cpumask_var(buffer->cpumask); in ring_buffer_free()
1690 !cpumask_test_cpu(cpu_id, buffer->cpumask)) in ring_buffer_resize()
[all …]
/linux-4.1.27/arch/blackfin/mach-bf561/
Dsmp.c27 struct cpumask mask; in platform_init_cpus()
36 struct cpumask mask; in platform_prepare_cpus()
/linux-4.1.27/arch/x86/platform/uv/
Dtlb_uv.c413 cpumask_t *mask = bcp->uvhub_master->cpumask; in reset_with_ipi()
873 int uv_flush_send_and_wait(struct cpumask *flush_mask, struct bau_control *bcp, in uv_flush_send_and_wait()
1024 static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp, in set_distrib_bits()
1078 const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, in uv_flush_tlb_others() argument
1088 struct cpumask *flush_mask; in uv_flush_tlb_others()
1097 return cpumask; in uv_flush_tlb_others()
1108 return cpumask; in uv_flush_tlb_others()
1116 return cpumask; in uv_flush_tlb_others()
1125 flush_mask = (struct cpumask *)per_cpu(uv_flush_tlb_mask, cpu); in uv_flush_tlb_others()
1127 cpumask_andnot(flush_mask, cpumask, cpumask_of(cpu)); in uv_flush_tlb_others()
[all …]
Duv_irq.c31 static int uv_set_irq_affinity(struct irq_data *, const struct cpumask *, bool);
133 const struct cpumask *eligible_cpu = cpumask_of(cpu); in arch_enable_uv_irq()
198 uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask, in uv_set_irq_affinity()
Duv_time.c318 int ced_cpu = cpumask_first(ced->cpumask); in uv_rtc_next_event()
329 int ced_cpu = cpumask_first(evt->cpumask); in uv_rtc_timer_setup()
371 ced->cpumask = cpumask_of(smp_processor_id()); in uv_rtc_register_clockevents()
/linux-4.1.27/drivers/clocksource/
Dmips-gic-timer.c31 gic_write_cpu_compare(cnt, cpumask_first(evt->cpumask)); in gic_next_event()
68 cd->cpumask = cpumask_of(cpu); in gic_clockevent_cpu_init()
Ddummy_timer.c39 evt->cpumask = cpumask_of(cpu); in dummy_timer_setup()
Di8253.c181 i8253_clockevent.cpumask = cpumask_of(smp_processor_id()); in clockevent_i8253_init()
Dclps711x-timer.c95 clkevt->cpumask = cpumask_of(0); in _clps711x_clkevt_init()
Ddw_apb_timer.c121 cpumask_first(evt->cpumask), in apbt_set_mode()
234 dw_ced->ced.cpumask = cpumask_of(cpu); in dw_apb_clockevent_init()
/linux-4.1.27/arch/x86/lib/
Dmsr-smp.c97 static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, in __rwmsr_on_cpus()
125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in rdmsr_on_cpus()
139 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in wrmsr_on_cpus()
/linux-4.1.27/arch/x86/kernel/
Dvsmp_64.c218 static void fill_vector_allocation_domain(int cpu, struct cpumask *retmask, in fill_vector_allocation_domain()
219 const struct cpumask *mask) in fill_vector_allocation_domain()
/linux-4.1.27/arch/metag/include/asm/
Dsmp.h14 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
/linux-4.1.27/arch/parisc/include/asm/
Dsmp.h31 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dirq.h47 extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest);
/linux-4.1.27/arch/mips/include/asm/mach-loongson/
Dtopology.h8 #define cpumask_of_node(node) (&__node_data[(node)]->cpumask)
Dmmzone.h42 cpumask_t cpumask; member
/linux-4.1.27/arch/mips/netlogic/common/
Dsmp.c75 void nlm_send_ipi_mask(const struct cpumask *mask, unsigned int action) in nlm_send_ipi_mask()
189 cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask); in nlm_smp_setup()
/linux-4.1.27/arch/mips/bcm63xx/
Dirq.c35 static void (*internal_irq_unmask)(struct irq_data *d, const struct cpumask *m);
55 const struct cpumask *m) in enable_irq_for_cpu()
135 const struct cpumask *m) \
377 const struct cpumask *dest, in bcm63xx_internal_set_affinity()
/linux-4.1.27/kernel/sched/
Dcpupri.h25 struct task_struct *p, struct cpumask *lowest_mask);
Dcpudeadline.h24 struct cpumask *later_mask);
Dsched.h837 unsigned long cpumask[0]; /* iteration mask */ member
854 unsigned long cpumask[0]; member
857 static inline struct cpumask *sched_group_cpus(struct sched_group *sg) in sched_group_cpus()
859 return to_cpumask(sg->cpumask); in sched_group_cpus()
866 static inline struct cpumask *sched_group_mask(struct sched_group *sg) in sched_group_mask()
868 return to_cpumask(sg->sgc->cpumask); in sched_group_mask()
1194 const struct cpumask *newmask);
Dcore.c1308 const struct cpumask *nodemask = NULL; in select_fallback_rq()
4017 long sched_setaffinity(pid_t pid, const struct cpumask *in_mask) in sched_setaffinity()
4107 struct cpumask *new_mask) in get_user_cpu_mask()
4141 long sched_getaffinity(pid_t pid, struct cpumask *mask) in sched_getaffinity()
4650 int cpuset_cpumask_can_shrink(const struct cpumask *cur, in cpuset_cpumask_can_shrink()
4651 const struct cpumask *trial) in cpuset_cpumask_can_shrink()
4675 const struct cpumask *cs_cpus_allowed) in task_can_attach()
4756 void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed()
4788 int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_ptr()
5408 struct cpumask *groupmask) in sched_domain_debug_one()
[all …]
/linux-4.1.27/arch/sparc/include/asm/
Dsmp_64.h37 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology_64.h49 static inline const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
Dsmp_32.h97 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/mips/include/asm/mach-netlogic/
Dmulti-node.h59 cpumask_t cpumask; /* logical cpu mask for node */ member
/linux-4.1.27/arch/sh/kernel/
Dlocaltimer.c59 clk->cpumask = cpumask_of(cpu); in local_timer_setup()
Dtopology.c33 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
Dsmp.c275 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
288 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
/linux-4.1.27/arch/mips/kernel/
Dsmp-up.c21 static inline void up_send_ipi_mask(const struct cpumask *mask, in up_send_ipi_mask()
Dsmp-gic.c58 void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action) in gic_send_ipi_mask()
Dcevt-ds1287.c125 cd->cpumask = cpumask_of(0); in ds1287_clockevent_init()
Dcevt-gt641xx.c135 cd->cpumask = cpumask_of(0); in gt641xx_timer0_clockevent_init()
Dsmp-bmips.c292 static void bmips5000_send_ipi_mask(const struct cpumask *mask, in bmips5000_send_ipi_mask()
344 static void bmips43xx_send_ipi_mask(const struct cpumask *mask, in bmips43xx_send_ipi_mask()
/linux-4.1.27/arch/tile/lib/
DMakefile5 lib-y = cacheflush.o checksum.o cpumask.o delay.o uaccess.o \
/linux-4.1.27/arch/mn10300/include/asm/
Dsmp.h92 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.1.27/arch/ia64/hp/sim/
Dhpsim_irq.c27 hpsim_set_affinity_noop(struct irq_data *d, const struct cpumask *b, bool f) in hpsim_set_affinity_noop()
/linux-4.1.27/include/trace/events/
Dipi.h20 TP_PROTO(const struct cpumask *mask, const char *reason),
/linux-4.1.27/arch/metag/kernel/
Dtopology.c29 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
/linux-4.1.27/arch/powerpc/kernel/
Dsmp.c289 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
298 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
589 const struct cpumask *mask; in traverse_siblings_chip_id()
639 const struct cpumask *mask; in traverse_core_siblings()
Dirq.c430 const struct cpumask *map = cpu_online_mask; in migrate_irqs()
643 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu()
672 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu()
/linux-4.1.27/arch/parisc/kernel/
Dsmp.c202 send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op) in send_IPI_mask()
242 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/
Dlinux-cpu.c569 cpumask_t *cpumask; in cfs_cpt_bind() local
577 cpumask = cptab->ctb_cpumask; in cfs_cpt_bind()
580 cpumask = cptab->ctb_parts[cpt].cpt_cpumask; in cfs_cpt_bind()
584 if (cpumask_any_and(cpumask, cpu_online_mask) >= nr_cpu_ids) { in cfs_cpt_bind()
591 if (cpumask_test_cpu(i, cpumask)) in cfs_cpt_bind()
594 rc = set_cpus_allowed_ptr(current, cpumask); in cfs_cpt_bind()
/linux-4.1.27/arch/s390/include/asm/
Dsmp.h25 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h38 const struct cpumask *cpu_coregroup_mask(int cpu);
/linux-4.1.27/arch/xtensa/kernel/
Dsmp.c352 static void send_ipi_message(const struct cpumask *callmask, in send_ipi_message()
365 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
382 struct cpumask targets; in smp_send_stop()
/linux-4.1.27/arch/score/kernel/
Dtime.c97 score_clockevent.cpumask = cpumask_of(0); in time_init()
/linux-4.1.27/arch/blackfin/kernel/
Dtime-ts.c232 evt->cpumask = cpumask_of(0); in bfin_gptmr0_clockevent_init()
345 evt->cpumask = cpumask_of(cpu); in bfin_coretmr_clockevent_init()
/linux-4.1.27/arch/mips/loongson/loongson-3/
Dnuma.c236 cpumask_clear(&__node_data[(node)]->cpumask); in prom_meminit()
247 cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask); in prom_meminit()
/linux-4.1.27/Documentation/filesystems/
Dsysfs-pci.txt39 local_cpus nearby CPU mask (cpumask, ro)
55 cpumask - file contains a cpumask type
/linux-4.1.27/arch/um/kernel/
Dtime.c53 .cpumask = cpu_all_mask,
/linux-4.1.27/arch/arm/mach-omap1/
Dtimer32k.c162 clockevent_32k_timer.cpumask = cpumask_of(0); in omap_init_32k_timer()
/linux-4.1.27/arch/arm/mach-netx/
Dtime.c147 netx_clockevent.cpumask = cpumask_of(0); in netx_timer_init()
/linux-4.1.27/arch/arm/mach-lpc32xx/
Dtimer.c143 lpc32xx_clkevt.cpumask = cpumask_of(0); in lpc32xx_timer_init()
/linux-4.1.27/arch/openrisc/kernel/
Dtime.c128 clockevent_openrisc_timer.cpumask = cpumask_of(0); in openrisc_clockevent_init()
/linux-4.1.27/arch/mips/sibyte/sb1250/
Dsmp.c74 static inline void sb1250_send_ipi_mask(const struct cpumask *mask, in sb1250_send_ipi_mask()
/linux-4.1.27/arch/arm/mach-footbridge/
Ddc21285-timer.c124 ce->cpumask = cpumask_of(smp_processor_id()); in footbridge_timer_init()
/linux-4.1.27/net/iucv/
Diucv.c563 cpumask_t cpumask; in iucv_setmask_up() local
567 cpumask_copy(&cpumask, &iucv_irq_cpumask); in iucv_setmask_up()
568 cpumask_clear_cpu(cpumask_first(&iucv_irq_cpumask), &cpumask); in iucv_setmask_up()
569 for_each_cpu(cpu, &cpumask) in iucv_setmask_up()
663 cpumask_t cpumask; in iucv_cpu_notify() local
690 cpumask_copy(&cpumask, &iucv_buffer_cpumask); in iucv_cpu_notify()
691 cpumask_clear_cpu(cpu, &cpumask); in iucv_cpu_notify()
692 if (cpumask_empty(&cpumask)) in iucv_cpu_notify()
/linux-4.1.27/arch/mips/alchemy/common/
Dtime.c90 .cpumask = cpu_all_mask,
/linux-4.1.27/Documentation/cpuidle/
Ddriver.txt27 const struct cpumask *const coupled_cpus);
/linux-4.1.27/arch/avr32/kernel/
Dtime.c145 comparator.cpumask = cpumask_of(0); in time_init()
/linux-4.1.27/arch/mips/jz4740/
Dtime.c126 jz4740_clockevent.cpumask = cpumask_of(0); in plat_time_init()
/linux-4.1.27/arch/m68k/coldfire/
Dpit.c154 cf_pit_clockevent.cpumask = cpumask_of(smp_processor_id()); in hw_timer_init()
/linux-4.1.27/arch/arm/mach-w90x900/
Dtime.c135 nuc900_clockevent_device.cpumask = cpumask_of(0); in nuc900_clockevents_init()
/linux-4.1.27/arch/arm/plat-iop/
Dtime.c159 iop_clockevent.cpumask = cpumask_of(0); in iop_init_time()
/linux-4.1.27/arch/mips/paravirt/
Dparavirt-smp.c74 static void paravirt_send_ipi_mask(const struct cpumask *mask, unsigned int action) in paravirt_send_ipi_mask()
/linux-4.1.27/arch/mips/sibyte/bcm1480/
Dsmp.c86 static void bcm1480_send_ipi_mask(const struct cpumask *mask, in bcm1480_send_ipi_mask()
/linux-4.1.27/block/
Dblk-mq.c776 WARN_ON(!cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)); in __blk_mq_run_hw_queue()
888 next_cpu = cpumask_next(hctx->next_cpu, hctx->cpumask); in blk_mq_hctx_next_cpu()
890 next_cpu = cpumask_first(hctx->cpumask); in blk_mq_hctx_next_cpu()
909 if (cpumask_test_cpu(cpu, hctx->cpumask)) { in blk_mq_run_hw_queue()
1645 free_cpumask_var(hctx->cpumask); in blk_mq_free_hw_queues()
1781 cpumask_clear(hctx->cpumask); in blk_mq_map_swqueue()
1794 cpumask_set_cpu(i, hctx->cpumask); in blk_mq_map_swqueue()
1831 hctx->next_cpu = cpumask_first(hctx->cpumask); in blk_mq_map_swqueue()
1953 if (!zalloc_cpumask_var_node(&hctxs[i]->cpumask, GFP_KERNEL, in blk_mq_init_allocated_queue()
2025 free_cpumask_var(hctxs[i]->cpumask); in blk_mq_init_allocated_queue()
/linux-4.1.27/drivers/pci/
Dpci-sysfs.c83 const struct cpumask *mask; in pci_dev_show_local_cpu()
114 const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); in cpuaffinity_show() local
116 return cpumap_print_to_pagebuf(false, buf, cpumask); in cpuaffinity_show()
123 const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); in cpulistaffinity_show() local
125 return cpumap_print_to_pagebuf(true, buf, cpumask); in cpulistaffinity_show()

12