Home
last modified time | relevance | path

Searched refs:cpumask (Results 1 – 200 of 424) sorted by relevance

123

/linux-4.4.14/include/linux/
Dcpumask.h15 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; struct
88 extern const struct cpumask *const cpu_possible_mask;
89 extern const struct cpumask *const cpu_online_mask;
90 extern const struct cpumask *const cpu_present_mask;
91 extern const struct cpumask *const cpu_active_mask;
124 static inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first()
130 static inline unsigned int cpumask_next(int n, const struct cpumask *srcp) in cpumask_next()
135 static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp) in cpumask_next_zero()
141 const struct cpumask *srcp, in cpumask_next_and()
142 const struct cpumask *andp) in cpumask_next_and()
[all …]
Dstop_machine.h34 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg);
35 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg);
80 static inline int stop_cpus(const struct cpumask *cpumask, in stop_cpus() argument
83 if (cpumask_test_cpu(raw_smp_processor_id(), cpumask)) in stop_cpus()
88 static inline int try_stop_cpus(const struct cpumask *cpumask, in try_stop_cpus() argument
91 return stop_cpus(cpumask, fn, arg); in try_stop_cpus()
117 int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
120 const struct cpumask *cpus);
124 const struct cpumask *cpus) in stop_machine()
135 const struct cpumask *cpus) in stop_machine_from_inactive_cpu()
Dcpu_cooling.h31 typedef int (*get_static_t)(cpumask_t *cpumask, int interval,
40 cpufreq_cooling_register(const struct cpumask *clip_cpus);
43 cpufreq_power_cooling_register(const struct cpumask *clip_cpus,
54 const struct cpumask *clip_cpus);
58 const struct cpumask *clip_cpus,
64 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
71 const struct cpumask *clip_cpus, in of_cpufreq_power_cooling_register()
88 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register()
93 cpufreq_power_cooling_register(const struct cpumask *clip_cpus, in cpufreq_power_cooling_register()
101 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
[all …]
Dpm_opp.h137 int dev_pm_opp_of_cpumask_add_table(cpumask_var_t cpumask);
138 void dev_pm_opp_of_cpumask_remove_table(cpumask_var_t cpumask);
139 int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask);
140 int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask);
151 static inline int dev_pm_opp_of_cpumask_add_table(cpumask_var_t cpumask) in dev_pm_opp_of_cpumask_add_table() argument
156 static inline void dev_pm_opp_of_cpumask_remove_table(cpumask_var_t cpumask) in dev_pm_opp_of_cpumask_remove_table() argument
160 static inline int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask) in dev_pm_opp_of_get_sharing_cpus() argument
165 static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask) in dev_pm_opp_set_sharing_cpus() argument
Dinterrupt.h229 extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask,
240 irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_set_affinity() argument
242 return __irq_set_affinity(irq, cpumask, false); in irq_set_affinity()
257 irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument
259 return __irq_set_affinity(irq, cpumask, true); in irq_force_affinity()
265 extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m);
272 static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) in irq_set_affinity()
277 static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument
290 const struct cpumask *m) in irq_set_affinity_hint()
Dsmpboot.h42 cpumask_var_t cpumask; member
48 const struct cpumask *cpumask);
59 const struct cpumask *);
Dpadata.h133 struct padata_cpumask cpumask; member
157 struct padata_cpumask cpumask; member
170 const struct cpumask *pcpumask,
171 const struct cpumask *cbcpumask);
177 cpumask_var_t cpumask);
Dsmp.h39 void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func,
96 void smp_call_function_many(const struct cpumask *mask,
99 int smp_call_function_any(const struct cpumask *mask,
145 smp_call_function_any(const struct cpumask *mask, smp_call_func_t func, in smp_call_function_any()
Dclockchips.h118 void (*broadcast)(const struct cpumask *mask);
128 const struct cpumask *cpumask; member
203 extern void tick_broadcast(const struct cpumask *mask);
Dcpuidle.h122 struct cpumask *cpumask; member
144 const struct cpumask *const coupled_cpus);
177 const struct cpumask *const coupled_cpus) in cpuidle_register()
Dtick.h144 static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) in tick_nohz_full_add_cpus_to()
166 static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) { } in tick_nohz_full_add_cpus_to()
173 static inline const struct cpumask *housekeeping_cpumask(void) in housekeeping_cpumask()
Dtopology.h202 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask()
208 static inline const struct cpumask *cpu_cpu_mask(int cpu) in cpu_cpu_mask()
Dcpuset.h44 extern void cpuset_cpus_allowed(struct task_struct *p, struct cpumask *mask);
153 struct cpumask *mask) in cpuset_cpus_allowed()
Dirq.h359 int (*irq_set_affinity)(struct irq_data *data, const struct cpumask *dest, bool force);
439 const struct cpumask *cpumask, bool force);
487 const struct cpumask *dest,
672 static inline struct cpumask *irq_get_affinity_mask(int irq) in irq_get_affinity_mask()
679 static inline struct cpumask *irq_data_get_affinity_mask(struct irq_data *d) in irq_data_get_affinity_mask()
Dirqdesc.h65 struct cpumask *percpu_enabled;
67 const struct cpumask *affinity_hint;
Dsched.h1108 static inline struct cpumask *sched_domain_span(struct sched_domain *sd) in sched_domain_span()
1122 typedef const struct cpumask *(*sched_domain_mask_f)(int cpu);
1367 struct cpumask cpumask; member
2225 extern int cpuset_cpumask_can_shrink(const struct cpumask *cur,
2226 const struct cpumask *trial);
2228 const struct cpumask *cs_cpus_allowed);
2231 const struct cpumask *new_mask);
2234 const struct cpumask *new_mask);
2237 const struct cpumask *new_mask) in do_set_cpus_allowed()
2241 const struct cpumask *new_mask) in set_cpus_allowed_ptr()
[all …]
Dcpu_rmap.h43 const struct cpumask *affinity);
Dblk-mq.h30 cpumask_var_t cpumask; member
194 struct cpumask *blk_mq_tags_cpumask(struct blk_mq_tags *tags);
/linux-4.4.14/kernel/
Dpadata.c40 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu()
42 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu()
58 cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash()
121 if (!cpumask_test_cpu(cb_cpu, pd->cpumask.cbcpu)) in padata_do_parallel()
177 num_cpus = cpumask_weight(pd->cpumask.pcpu); in padata_get_next()
351 const struct cpumask *pcpumask, in padata_setup_cpumasks()
352 const struct cpumask *cbcpumask) in padata_setup_cpumasks()
354 if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) in padata_setup_cpumasks()
357 cpumask_and(pd->cpumask.pcpu, pcpumask, cpu_online_mask); in padata_setup_cpumasks()
358 if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL)) { in padata_setup_cpumasks()
[all …]
Dstop_machine.c151 const struct cpumask *active_cpus;
316 static void queue_stop_cpus_work(const struct cpumask *cpumask, in queue_stop_cpus_work() argument
329 for_each_cpu(cpu, cpumask) { in queue_stop_cpus_work()
339 static int __stop_cpus(const struct cpumask *cpumask, in __stop_cpus() argument
344 cpu_stop_init_done(&done, cpumask_weight(cpumask)); in __stop_cpus()
345 queue_stop_cpus_work(cpumask, fn, arg, &done); in __stop_cpus()
378 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in stop_cpus() argument
384 ret = __stop_cpus(cpumask, fn, arg); in stop_cpus()
407 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in try_stop_cpus() argument
414 ret = __stop_cpus(cpumask, fn, arg); in try_stop_cpus()
[all …]
Dsmpboot.c235 if (cpumask_test_cpu(cpu, cur->cpumask)) in smpboot_unpark_threads()
283 const struct cpumask *cpumask) in smpboot_register_percpu_thread_cpumask() argument
288 if (!alloc_cpumask_var(&plug_thread->cpumask, GFP_KERNEL)) in smpboot_register_percpu_thread_cpumask()
290 cpumask_copy(plug_thread->cpumask, cpumask); in smpboot_register_percpu_thread_cpumask()
298 free_cpumask_var(plug_thread->cpumask); in smpboot_register_percpu_thread_cpumask()
301 if (cpumask_test_cpu(cpu, cpumask)) in smpboot_register_percpu_thread_cpumask()
326 free_cpumask_var(plug_thread->cpumask); in smpboot_unregister_percpu_thread()
340 const struct cpumask *new) in smpboot_update_cpumask_percpu_thread()
342 struct cpumask *old = plug_thread->cpumask; in smpboot_update_cpumask_percpu_thread()
Dsmp.c27 cpumask_var_t cpumask; member
45 if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, in hotplug_cfd()
50 free_cpumask_var(cfd->cpumask); in hotplug_cfd()
62 free_cpumask_var(cfd->cpumask); in hotplug_cfd()
361 int smp_call_function_any(const struct cpumask *mask, in smp_call_function_any()
365 const struct cpumask *nodemask; in smp_call_function_any()
404 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many()
441 cpumask_and(cfd->cpumask, mask, cpu_online_mask); in smp_call_function_many()
442 cpumask_clear_cpu(this_cpu, cfd->cpumask); in smp_call_function_many()
445 if (unlikely(!cpumask_weight(cfd->cpumask))) in smp_call_function_many()
[all …]
Dworkqueue.c1672 set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); in worker_attach_to_pool()
1756 kthread_bind_mask(worker->task, pool->attrs->cpumask); in create_worker()
3028 free_cpumask_var(attrs->cpumask); in free_workqueue_attrs()
3049 if (!alloc_cpumask_var(&attrs->cpumask, gfp_mask)) in alloc_workqueue_attrs()
3052 cpumask_copy(attrs->cpumask, cpu_possible_mask); in alloc_workqueue_attrs()
3063 cpumask_copy(to->cpumask, from->cpumask); in copy_workqueue_attrs()
3078 hash = jhash(cpumask_bits(attrs->cpumask), in wqattrs_hash()
3089 if (!cpumask_equal(a->cpumask, b->cpumask)) in wqattrs_equal()
3256 if (cpumask_subset(attrs->cpumask, in get_unbound_pool()
3467 int cpu_going_down, cpumask_t *cpumask) in wq_calc_node_cpumask() argument
[all …]
Dcpu.c767 const struct cpumask *const cpu_possible_mask = to_cpumask(cpu_possible_bits);
771 const struct cpumask *const cpu_online_mask = to_cpumask(cpu_online_bits);
775 const struct cpumask *const cpu_present_mask = to_cpumask(cpu_present_bits);
779 const struct cpumask *const cpu_active_mask = to_cpumask(cpu_active_bits);
816 void init_cpu_present(const struct cpumask *src) in init_cpu_present()
821 void init_cpu_possible(const struct cpumask *src) in init_cpu_possible()
826 void init_cpu_online(const struct cpumask *src) in init_cpu_online()
Dup.c53 void on_each_cpu_mask(const struct cpumask *mask, in on_each_cpu_mask()
/linux-4.4.14/arch/m32r/kernel/
Dsmp.c80 static void send_IPI_mask(const struct cpumask *, int, int);
151 cpumask_t cpumask; in smp_flush_cache_all() local
155 cpumask_copy(&cpumask, cpu_online_mask); in smp_flush_cache_all()
156 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_flush_cache_all()
158 mask=cpumask_bits(&cpumask); in smp_flush_cache_all()
160 send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); in smp_flush_cache_all()
370 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
388 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
390 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
394 cpumask_and(&cpumask, &cpumask, cpu_online_mask); in flush_tlb_others()
[all …]
/linux-4.4.14/arch/x86/include/asm/uv/
Duv.h6 struct cpumask;
16 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask,
28 static inline const struct cpumask *
29 uv_flush_tlb_others(const struct cpumask *cpumask, struct mm_struct *mm, in uv_flush_tlb_others() argument
31 { return cpumask; } in uv_flush_tlb_others()
/linux-4.4.14/arch/x86/include/asm/
Dapic.h278 const struct cpumask *(*target_cpus)(void);
285 void (*vector_allocation_domain)(int cpu, struct cpumask *retmask,
286 const struct cpumask *mask);
301 int (*cpu_mask_to_apicid_and)(const struct cpumask *cpumask,
302 const struct cpumask *andmask,
306 void (*send_IPI_mask)(const struct cpumask *mask, int vector);
307 void (*send_IPI_mask_allbutself)(const struct cpumask *mask,
478 static inline const struct cpumask *default_target_cpus(void) in default_target_cpus()
487 static inline const struct cpumask *online_target_cpus(void) in online_target_cpus()
544 flat_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in flat_cpu_mask_to_apicid_and() argument
[all …]
Dipi.h122 extern void default_send_IPI_mask_sequence_phys(const struct cpumask *mask,
124 extern void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask,
149 extern void default_send_IPI_mask_sequence_logical(const struct cpumask *mask,
151 extern void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask,
153 extern void default_send_IPI_mask_logical(const struct cpumask *mask,
Dtopology.h76 extern const struct cpumask *cpumask_of_node(int node);
79 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
120 extern const struct cpumask *cpu_coregroup_mask(int cpu);
Dsmp.h40 static inline struct cpumask *cpu_llc_shared_mask(int cpu) in cpu_llc_shared_mask()
69 void (*send_call_func_ipi)(const struct cpumask *mask);
137 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
156 void native_send_call_func_ipi(const struct cpumask *mask);
Dtlbflush.h212 static inline void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
249 void native_flush_tlb_others(const struct cpumask *cpumask,
Dmsr.h253 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
254 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
282 static inline void rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, in rdmsr_on_cpus()
287 static inline void wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, in wrmsr_on_cpus()
Dirqdomain.h16 const struct cpumask *mask);
Dpci.h123 static inline const struct cpumask *
/linux-4.4.14/drivers/cpuidle/
Dcpuidle-big_little.c143 struct cpumask *cpumask; in bl_idle_driver_init() local
146 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); in bl_idle_driver_init()
147 if (!cpumask) in bl_idle_driver_init()
152 cpumask_set_cpu(cpu, cpumask); in bl_idle_driver_init()
154 drv->cpumask = cpumask; in bl_idle_driver_init()
222 kfree(bl_idle_big_driver.cpumask); in bl_idle_init()
224 kfree(bl_idle_little_driver.cpumask); in bl_idle_init()
Ddt_idle_states.c99 const cpumask_t *cpumask) in idle_state_valid() argument
112 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid()
113 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid()
158 const cpumask_t *cpumask; in dt_init_idle_driver() local
169 cpumask = drv->cpumask ? : cpu_possible_mask; in dt_init_idle_driver()
170 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver()
180 if (!idle_state_valid(state_node, i, cpumask)) { in dt_init_idle_driver()
Ddriver.c50 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_unset_driver()
72 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_set_driver()
164 if (!drv->cpumask) in __cpuidle_driver_init()
165 drv->cpumask = (struct cpumask *)cpu_possible_mask; in __cpuidle_driver_init()
244 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_register_driver()
268 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_unregister_driver()
Dcpuidle.c555 for_each_cpu(cpu, drv->cpumask) { in cpuidle_unregister()
576 const struct cpumask *const coupled_cpus) in cpuidle_register()
587 for_each_cpu(cpu, drv->cpumask) { in cpuidle_register()
/linux-4.4.14/arch/x86/kernel/apic/
Dx2apic_cluster.c27 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
29 struct cpumask *cpus_in_cluster_ptr; in __x2apic_send_IPI_mask()
30 struct cpumask *ipi_mask_ptr; in __x2apic_send_IPI_mask()
77 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
83 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
99 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument
100 const struct cpumask *andmask, in x2apic_cpu_mask_to_apicid_and()
107 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
118 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
211 static const struct cpumask *x2apic_cluster_target_cpus(void) in x2apic_cluster_target_cpus()
[all …]
Dapic_flat_64.c65 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument
67 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask()
73 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument
75 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself()
233 static void physflat_send_IPI_mask(const struct cpumask *cpumask, int vector) in physflat_send_IPI_mask() argument
235 default_send_IPI_mask_sequence_phys(cpumask, vector); in physflat_send_IPI_mask()
238 static void physflat_send_IPI_mask_allbutself(const struct cpumask *cpumask, in physflat_send_IPI_mask_allbutself() argument
241 default_send_IPI_mask_allbutself_phys(cpumask, vector); in physflat_send_IPI_mask_allbutself()
Dapic_noop.c33 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask() argument
34 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask_allbutself() argument
86 static const struct cpumask *noop_target_cpus(void) in noop_target_cpus()
92 static void noop_vector_allocation_domain(int cpu, struct cpumask *retmask, in noop_vector_allocation_domain()
93 const struct cpumask *mask) in noop_vector_allocation_domain()
Dipi.c21 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys()
39 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys()
60 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical()
80 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical()
103 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument
105 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
Dx2apic_phys.c40 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
60 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
66 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
Dx2apic_uv_x.c264 static void uv_send_IPI_mask(const struct cpumask *mask, int vector) in uv_send_IPI_mask()
272 static void uv_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in uv_send_IPI_mask_allbutself()
314 uv_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in uv_cpu_mask_to_apicid_and() argument
315 const struct cpumask *andmask, in uv_cpu_mask_to_apicid_and()
324 for_each_cpu_and(cpu, cpumask, andmask) { in uv_cpu_mask_to_apicid_and()
Dapic_numachip.c122 static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) in numachip_send_IPI_mask()
130 static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, in numachip_send_IPI_mask_allbutself()
Dvector.c104 const struct cpumask *mask) in __assign_irq_vector()
229 const struct cpumask *mask) in assign_irq_vector()
289 const struct cpumask *mask) in init_irq_alloc_info()
519 const struct cpumask *dest, bool force) in apic_set_affinity()
/linux-4.4.14/drivers/base/power/opp/
Dcpu.c123 int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask) in dev_pm_opp_set_sharing_cpus() argument
138 for_each_cpu(cpu, cpumask) { in dev_pm_opp_set_sharing_cpus()
164 void dev_pm_opp_of_cpumask_remove_table(cpumask_var_t cpumask) in dev_pm_opp_of_cpumask_remove_table() argument
169 WARN_ON(cpumask_empty(cpumask)); in dev_pm_opp_of_cpumask_remove_table()
171 for_each_cpu(cpu, cpumask) { in dev_pm_opp_of_cpumask_remove_table()
184 int dev_pm_opp_of_cpumask_add_table(cpumask_var_t cpumask) in dev_pm_opp_of_cpumask_add_table() argument
189 WARN_ON(cpumask_empty(cpumask)); in dev_pm_opp_of_cpumask_add_table()
191 for_each_cpu(cpu, cpumask) { in dev_pm_opp_of_cpumask_add_table()
205 dev_pm_opp_of_cpumask_remove_table(cpumask); in dev_pm_opp_of_cpumask_add_table()
220 int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask) in dev_pm_opp_of_get_sharing_cpus() argument
[all …]
/linux-4.4.14/arch/tile/include/asm/
Dsmp.h32 void send_IPI_many(const struct cpumask *mask, int tag);
75 static inline void arch_send_call_function_ipi_mask(struct cpumask *mask) in arch_send_call_function_ipi_mask()
102 extern struct cpumask cpu_lotar_map;
106 extern struct cpumask hash_for_home_map;
109 extern struct cpumask cpu_cacheable_map;
127 static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp, in __cpulist_parse_crop()
Dhomecache.h51 const struct cpumask *cache_cpumask,
54 const struct cpumask *tlb_cpumask,
61 extern void homecache_evict(const struct cpumask *mask);
Dtopology.h23 extern struct cpumask node_2_cpu_mask[];
39 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
/linux-4.4.14/arch/mn10300/mm/
Dtlb-smp.c52 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm,
94 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
105 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
106 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
108 cpumask_and(&tmp, &cpumask, cpu_online_mask); in flush_tlb_others()
109 BUG_ON(!cpumask_equal(&cpumask, &tmp)); in flush_tlb_others()
122 atomic_or(cpumask.bits[0], (atomic_t *)&flush_cpumask.bits[0]); in flush_tlb_others()
/linux-4.4.14/kernel/time/
Dtick-common.c179 const struct cpumask *cpumask) in tick_setup_device() argument
217 if (!cpumask_equal(newdev->cpumask, cpumask)) in tick_setup_device()
218 irq_set_affinity(newdev->irq, cpumask); in tick_setup_device()
250 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_percpu()
252 if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) in tick_check_percpu()
258 if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) in tick_check_percpu()
280 !cpumask_equal(curdev->cpumask, newdev->cpumask); in tick_check_preferred()
Dtick-broadcast.c54 struct cpumask *tick_get_broadcast_mask(void) in tick_get_broadcast_mask()
138 static void err_broadcast(const struct cpumask *mask) in err_broadcast()
259 static bool tick_do_broadcast(struct cpumask *mask) in tick_do_broadcast()
527 struct cpumask *tick_get_broadcast_oneshot_mask(void) in tick_get_broadcast_oneshot_mask()
548 const struct cpumask *cpumask) in tick_broadcast_set_affinity() argument
553 if (cpumask_equal(bc->cpumask, cpumask)) in tick_broadcast_set_affinity()
556 bc->cpumask = cpumask; in tick_broadcast_set_affinity()
557 irq_set_affinity(bc->irq, bc->cpumask); in tick_broadcast_set_affinity()
854 static void tick_broadcast_init_next_event(struct cpumask *mask, in tick_broadcast_init_next_event()
Dtick-internal.h74 extern struct cpumask *tick_get_broadcast_mask(void);
135 extern struct cpumask *tick_get_broadcast_oneshot_mask(void);
Dclockevents.c456 if (!dev->cpumask) { in clockevents_register_device()
458 dev->cpumask = cpumask_of(smp_processor_id()); in clockevents_register_device()
636 if (cpumask_test_cpu(cpu, dev->cpumask) && in tick_cleanup_dead_cpu()
637 cpumask_weight(dev->cpumask) == 1 && in tick_cleanup_dead_cpu()
Dtick-broadcast-hrtimer.c91 .cpumask = cpu_all_mask,
/linux-4.4.14/include/trace/events/
Dthermal.h82 TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load,
88 __bitmask(cpumask, num_possible_cpus())
97 __assign_bitmask(cpumask, cpumask_bits(cpus),
108 __get_bitmask(cpumask), __entry->freq,
114 TP_PROTO(const struct cpumask *cpus, unsigned int freq,
120 __bitmask(cpumask, num_possible_cpus())
127 __assign_bitmask(cpumask, cpumask_bits(cpus),
135 __get_bitmask(cpumask), __entry->freq, __entry->cdev_state,
Dipi.h20 TP_PROTO(const struct cpumask *mask, const char *reason),
/linux-4.4.14/arch/tile/kernel/
Dhardwall.c105 struct cpumask cpumask; /* cpus reserved */ member
202 static int check_rectangle(struct hardwall_info *r, struct cpumask *mask) in check_rectangle()
296 struct cpumask rect_cpus; in hardwall_protect_rectangle()
351 if (cpumask_test_cpu(cpu, &rect->cpumask)) in do_hardwall_trap()
477 return info && cpumask_test_cpu(cpu, &info->cpumask); in hardwall_ipi_valid()
493 struct cpumask mask; in hardwall_create()
502 if (copy_from_user(&mask, bits, min(sizeof(struct cpumask), size))) in hardwall_create()
510 if (size < sizeof(struct cpumask)) { in hardwall_create()
511 memset((char *)&mask + size, 0, sizeof(struct cpumask) - size); in hardwall_create()
512 } else if (size > sizeof(struct cpumask)) { in hardwall_create()
[all …]
Dsmpboot.c124 static __initdata struct cpumask init_affinity;
135 static struct cpumask cpu_started;
Dsetup.c847 struct cpumask node_2_cpu_mask[MAX_NUMNODES] __write_once;
855 static int __init cpu_to_bound_node(int cpu, struct cpumask* unbound_cpus) in cpu_to_bound_node()
865 struct cpumask *unbound_cpus) in node_neighbors()
889 struct cpumask unbound_cpus; in setup_numa_mapping()
1274 struct cpumask __write_once cpu_lotar_map;
1284 struct cpumask hash_for_home_map;
1296 struct cpumask __write_once cpu_cacheable_map;
1299 static __initdata struct cpumask disabled_map;
1325 struct cpumask hv_disabled_map, cpu_possible_init; in setup_cpu_maps()
1398 struct cpumask shared; in check_nohz_full_cpus()
Dsmp.c73 void send_IPI_many(const struct cpumask *mask, int tag) in send_IPI_many()
92 struct cpumask mask; in send_IPI_allbutself()
/linux-4.4.14/arch/hexagon/include/asm/
Dsmp.h36 extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg);
39 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/ia64/kernel/
Dsmp.c169 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask()
259 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local
265 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask()
269 for_each_cpu(cpu, &cpumask) { in smp_flush_tlb_cpumask()
279 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask()
324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dirq.c76 bool is_affinity_mask_valid(const struct cpumask *cpumask) in is_affinity_mask_valid() argument
80 if (cpumask_weight(cpumask) != 1) in is_affinity_mask_valid()
/linux-4.4.14/arch/tile/mm/
Dmigrate.h33 const unsigned long *cpumask);
51 const struct cpumask *cache_cpumask,
52 const struct cpumask *tlb_cpumask,
Dhomecache.c64 static void hv_flush_update(const struct cpumask *cache_cpumask, in hv_flush_update()
65 struct cpumask *tlb_cpumask, in hv_flush_update()
69 struct cpumask mask; in hv_flush_update()
108 const struct cpumask *cache_cpumask_orig, in flush_remote()
111 const struct cpumask *tlb_cpumask_orig, in flush_remote()
115 struct cpumask cache_cpumask_copy, tlb_cpumask_copy; in flush_remote()
116 struct cpumask *cache_cpumask, *tlb_cpumask; in flush_remote()
238 void homecache_evict(const struct cpumask *mask) in homecache_evict()
Dinit.c203 static __initdata struct cpumask kdata_mask;
318 static struct cpumask __initdata ktext_mask;
412 const struct cpumask *my_cpu_mask = cpumask_of(smp_processor_id()); in kernel_physical_mapping_init()
413 struct cpumask kstripe_mask; in kernel_physical_mapping_init()
493 struct cpumask bad; in kernel_physical_mapping_init()
/linux-4.4.14/arch/arm64/include/asm/
Dsmp.h46 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int));
48 extern void (*__smp_cross_call)(const struct cpumask *, unsigned int);
65 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h23 const struct cpumask *cpu_coregroup_mask(int cpu);
/linux-4.4.14/arch/mn10300/kernel/
Dsmp.c107 static void send_IPI_mask(const cpumask_t *cpumask, int irq);
313 static void send_IPI_mask(const cpumask_t *cpumask, int irq) in send_IPI_mask() argument
319 if (cpumask_test_cpu(i, cpumask)) { in send_IPI_mask()
350 cpumask_t cpumask; in send_IPI_allbutself() local
352 cpumask_copy(&cpumask, cpu_online_mask); in send_IPI_allbutself()
353 cpumask_clear_cpu(smp_processor_id(), &cpumask); in send_IPI_allbutself()
354 send_IPI_mask(&cpumask, irq); in send_IPI_allbutself()
357 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
1067 static int hotplug_cpu_nmi_call_function(cpumask_t cpumask, in hotplug_cpu_nmi_call_function() argument
1084 nmi_call_func_mask_data.started = cpumask; in hotplug_cpu_nmi_call_function()
[all …]
Dirq.c142 mn10300_cpupic_setaffinity(struct irq_data *d, const struct cpumask *mask, in mn10300_cpupic_setaffinity()
320 struct cpumask *mask = irq_data_get_affinity_mask(data); in migrate_irqs()
/linux-4.4.14/Documentation/
Dpadata.txt17 const struct cpumask *pcpumask,
18 const struct cpumask *cbcpumask);
35 it is legal to supply a cpumask to padata that contains offline CPUs.
36 Once an offline CPU in the user supplied cpumask comes online, padata
47 padata cpumask contains no active CPU (flag not set).
57 cpumask_var_t cpumask);
68 PADATA_CPU_PARALLEL and cpumask specifies the new cpumask to use.
69 To simply add or remove one CPU from a certain cpumask the functions
73 If a user is interested in padata cpumask changes, he can register to
74 the padata cpumask change notifier:
[all …]
/linux-4.4.14/arch/hexagon/kernel/
Dsmp.c105 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument
113 for_each_cpu(cpu, cpumask) { in send_ipi()
239 struct cpumask targets; in smp_send_stop()
250 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dtime.c102 static void broadcast(const struct cpumask *mask) in broadcast()
134 dummy_clock_dev->cpumask = cpumask_of(cpu); in setup_percpu_clockdev()
180 ce_dev->cpumask = cpu_all_mask; in time_init_deferred()
/linux-4.4.14/arch/x86/mm/
Dtlb.c131 void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
154 cpumask = uv_flush_tlb_others(cpumask, mm, start, end, cpu); in native_flush_tlb_others()
155 if (cpumask) in native_flush_tlb_others()
156 smp_call_function_many(cpumask, flush_tlb_func, in native_flush_tlb_others()
160 smp_call_function_many(cpumask, flush_tlb_func, &info, 1); in native_flush_tlb_others()
/linux-4.4.14/Documentation/accounting/
Dgetdelays.c72 char cpumask[100+6*MAX_CPUS]; variable
316 strncpy(cpumask, optarg, sizeof(cpumask)); in main()
317 cpumask[sizeof(cpumask) - 1] = '\0'; in main()
319 printf("cpumask %s maskset %d\n", cpumask, maskset); in main()
393 &cpumask, strlen(cpumask) + 1); in main()
534 &cpumask, strlen(cpumask) + 1); in main()
Dtaskstats.txt35 sends a register command and specifies a cpumask. Whenever a task exits on
36 one of the cpus in the cpumask, its per-pid statistics are sent to the
84 TASKSTATS_CMD_ATTR_REGISTER/DEREGISTER_CPUMASK and contain a cpumask in the
85 attribute payload. The cpumask is specified as an ascii string of
87 the cpumask would be "1-3,5,7-8". If userspace forgets to deregister interest
/linux-4.4.14/arch/blackfin/mach-common/
Dsmp.c189 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument
196 for_each_cpu(cpu, cpumask) { in send_ipi()
203 for_each_cpu(cpu, cpumask) in send_ipi()
212 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
224 void smp_send_msg(const struct cpumask *mask, unsigned long type) in smp_send_msg()
229 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
/linux-4.4.14/arch/powerpc/include/asm/
Dsmp.h96 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
101 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask()
139 static inline const struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
189 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dxics.h156 extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask,
159 #define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server) argument
Dirq.h71 int irq_choose_cpu(const struct cpumask *mask);
Dcputhreads.h42 static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads) in cpu_thread_mask_to_cores()
Dfadump.h194 struct cpumask cpu_online_mask;
/linux-4.4.14/arch/arm/mach-vexpress/
Ddcscb.c43 unsigned int rst_hold, cpumask = (1 << cpu); in dcscb_cpu_powerup() local
46 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) in dcscb_cpu_powerup()
50 rst_hold &= ~(cpumask | (cpumask << 4)); in dcscb_cpu_powerup()
/linux-4.4.14/arch/arm/include/asm/
Dsmp.h49 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int));
79 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
80 extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask);
Dtopology.h25 const struct cpumask *cpu_coregroup_mask(int cpu);
/linux-4.4.14/drivers/irqchip/
Dirq-metag.c44 const struct cpumask *cpumask, bool force);
193 const struct cpumask *cpumask, bool force) in metag_internal_irq_set_affinity() argument
204 cpu = cpumask_any_and(cpumask, cpu_online_mask); in metag_internal_irq_set_affinity()
Dirq-hip04.c145 const struct cpumask *mask_val, in hip04_irq_set_affinity()
231 u32 cpumask; in hip04_irq_dist_init() local
240 cpumask = hip04_get_cpumask(intc); in hip04_irq_dist_init()
241 cpumask |= cpumask << 16; in hip04_irq_dist_init()
243 writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3)); in hip04_irq_dist_init()
279 static void hip04_raise_softirq(const struct cpumask *mask, unsigned int irq) in hip04_raise_softirq()
Dirq-gic.c302 static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, in gic_set_affinity()
475 u32 cpumask; in gic_dist_init() local
484 cpumask = gic_get_cpumask(gic); in gic_dist_init()
485 cpumask |= cpumask << 8; in gic_dist_init()
486 cpumask |= cpumask << 16; in gic_dist_init()
488 writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4); in gic_dist_init()
769 static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq) in gic_raise_softirq()
Dirq-metag-ext.c504 const struct cpumask *cpumask, bool force) in meta_intc_set_affinity() argument
517 cpu = cpumask_any_and(cpumask, cpu_online_mask); in meta_intc_set_affinity()
Dirq-mips-gic.c433 static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask, in gic_set_affinity() argument
441 cpumask_and(&tmp, cpumask, cpu_online_mask); in gic_set_affinity()
456 cpumask_copy(irq_data_get_affinity_mask(d), cpumask); in gic_set_affinity()
/linux-4.4.14/drivers/thermal/
Dcpu_cooling.c97 struct cpumask allowed_cpus;
427 struct cpumask *cpumask = &cpufreq_device->allowed_cpus; in get_static_power() local
451 return cpufreq_device->plat_get_static_power(cpumask, tz->passive_delay, in get_static_power()
655 cpumask_t cpumask; in cpufreq_state2power() local
660 cpumask_and(&cpumask, &cpufreq_device->allowed_cpus, cpu_online_mask); in cpufreq_state2power()
661 num_cpus = cpumask_weight(&cpumask); in cpufreq_state2power()
787 const struct cpumask *clip_cpus, u32 capacitance, in __cpufreq_cooling_register()
925 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register()
946 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
977 cpufreq_power_cooling_register(const struct cpumask *clip_cpus, u32 capacitance, in cpufreq_power_cooling_register()
[all …]
/linux-4.4.14/arch/ia64/include/asm/
Dirq.h30 bool is_affinity_mask_valid(const struct cpumask *cpumask);
Dsmp.h129 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/crypto/
Dpcrypt.c77 struct pcrypt_cpumask *cpumask; in pcrypt_do_parallel() local
82 cpumask = rcu_dereference_bh(pcrypt->cb_cpumask); in pcrypt_do_parallel()
83 if (cpumask_test_cpu(cpu, cpumask->mask)) in pcrypt_do_parallel()
86 if (!cpumask_weight(cpumask->mask)) in pcrypt_do_parallel()
89 cpu_index = cpu % cpumask_weight(cpumask->mask); in pcrypt_do_parallel()
91 cpu = cpumask_first(cpumask->mask); in pcrypt_do_parallel()
93 cpu = cpumask_next(cpu, cpumask->mask); in pcrypt_do_parallel()
365 struct padata_cpumask *cpumask = (struct padata_cpumask *)data; in pcrypt_cpumask_change_notify() local
381 cpumask_copy(new_mask->mask, cpumask->cbcpu); in pcrypt_cpumask_change_notify()
/linux-4.4.14/arch/alpha/kernel/
Dsmp.c176 wait_for_txrdy (unsigned long cpumask) in wait_for_txrdy() argument
180 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy()
185 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy()
203 unsigned long cpumask; in send_secondary_console_msg() local
211 cpumask = (1UL << cpuid); in send_secondary_console_msg()
212 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg()
225 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg()
509 send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation) in send_ipi_message()
596 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dtime.c124 .cpumask = cpumask_of(cpu), in init_rtc_clockevent()
191 .cpumask = cpumask_of(cpu), in init_qemu_clockevent()
/linux-4.4.14/arch/x86/xen/
Dsmp.h3 extern void xen_send_IPI_mask(const struct cpumask *mask,
5 extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
Dsmp.c591 static void __xen_send_IPI_mask(const struct cpumask *mask, in __xen_send_IPI_mask()
600 static void xen_smp_send_call_function_ipi(const struct cpumask *mask) in xen_smp_send_call_function_ipi()
653 void xen_send_IPI_mask(const struct cpumask *mask, in xen_send_IPI_mask()
678 void xen_send_IPI_mask_allbutself(const struct cpumask *mask, in xen_send_IPI_mask_allbutself()
/linux-4.4.14/lib/
Dcpumask.c16 int cpumask_next_and(int n, const struct cpumask *src1p, in cpumask_next_and()
17 const struct cpumask *src2p) in cpumask_next_and()
34 int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) in cpumask_any_but()
Dcpu_rmap.c98 const struct cpumask *mask, u16 dist) in cpu_rmap_copy_neigh()
159 const struct cpumask *affinity) in cpu_rmap_update()
/linux-4.4.14/kernel/irq/
Dmanage.c166 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) in irq_copy_pending()
171 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) in irq_get_pending()
179 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) { } in irq_copy_pending()
181 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) { } in irq_get_pending()
184 int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, in irq_do_set_affinity()
204 int irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask, in irq_set_affinity_locked()
230 int __irq_set_affinity(unsigned int irq, const struct cpumask *mask, bool force) in __irq_set_affinity()
245 int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m) in irq_set_affinity_hint()
266 cpumask_var_t cpumask; in irq_affinity_notify() local
269 if (!desc || !alloc_cpumask_var(&cpumask, GFP_KERNEL)) in irq_affinity_notify()
[all …]
Dinternals.h108 extern int irq_select_affinity_usr(unsigned int irq, struct cpumask *mask);
113 const struct cpumask *dest, bool force);
Dcpuhotplug.c20 const struct cpumask *affinity = d->common->affinity; in migrate_one_irq()
/linux-4.4.14/arch/arm/kernel/
Dsmp.c464 static void (*__smp_cross_call)(const struct cpumask *, unsigned int);
466 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument
484 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call()
516 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
521 void arch_send_wakeup_ipi_mask(const struct cpumask *mask) in arch_send_wakeup_ipi_mask()
540 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
674 struct cpumask mask; in smp_send_stop()
Dtopology.c183 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
192 const struct cpumask *cpu_corepower_mask(int cpu) in cpu_corepower_mask()
Dirq.c127 const struct cpumask *affinity = irq_data_get_affinity_mask(d); in migrate_one_irq()
/linux-4.4.14/drivers/cpufreq/
Dacpi-cpufreq.c106 static void boost_set_msrs(bool enable, const struct cpumask *cpumask) in boost_set_msrs() argument
125 rdmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs()
127 for_each_cpu(cpu, cpumask) { in boost_set_msrs()
135 wrmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs()
262 const struct cpumask *mask;
337 get_cur_val(const struct cpumask *mask, struct acpi_cpufreq_data *data) in get_cur_val()
404 static unsigned int check_freqs(const struct cpumask *mask, unsigned int freq, in check_freqs()
536 const struct cpumask *cpumask; in boost_notify() local
538 cpumask = get_cpu_mask(cpu); in boost_notify()
550 boost_set_msrs(acpi_cpufreq_driver.boost_enabled, cpumask); in boost_notify()
[all …]
Dpowernow-k8.h40 struct cpumask *available_cores;
/linux-4.4.14/arch/powerpc/sysdev/xics/
Dics-rtas.c122 const struct cpumask *cpumask, in ics_rtas_set_affinity() argument
141 irq_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_rtas_set_affinity()
144 __func__, cpumask_pr_args(cpumask), d->irq); in ics_rtas_set_affinity()
Dics-opal.c111 const struct cpumask *cpumask, in ics_opal_set_affinity() argument
132 wanted_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_opal_set_affinity()
135 __func__, cpumask_pr_args(cpumask), d->irq); in ics_opal_set_affinity()
Dxics-common.c272 int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, in xics_get_irq_server() argument
279 if (!cpumask_subset(cpu_possible_mask, cpumask)) { in xics_get_irq_server()
280 int server = cpumask_first_and(cpu_online_mask, cpumask); in xics_get_irq_server()
/linux-4.4.14/arch/xtensa/include/asm/
Dsmp.h22 struct cpumask;
23 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/alpha/include/asm/
Dtopology.h25 extern struct cpumask node_to_cpumask_map[];
27 static const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
Dsmp.h48 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/blackfin/include/asm/
Dsmp.h22 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
51 void smp_timer_broadcast(const struct cpumask *mask);
/linux-4.4.14/drivers/sh/intc/
Dchip.c69 const struct cpumask *cpumask, in intc_set_affinity() argument
72 if (!cpumask_intersects(cpumask, cpu_online_mask)) in intc_set_affinity()
75 cpumask_copy(irq_data_get_affinity_mask(data), cpumask); in intc_set_affinity()
/linux-4.4.14/arch/sparc/kernel/
Dtime_32.c130 ce->cpumask = cpu_possible_mask; in setup_timer_ce()
187 int cpu = cpumask_first(evt->cpumask); in percpu_ce_shutdown()
195 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_periodic()
204 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_next_event()
226 ce->cpumask = cpumask_of(cpu); in register_percpu_ce()
/linux-4.4.14/kernel/events/
Dhw_breakpoint.c133 static const struct cpumask *cpumask_of_bp(struct perf_event *bp) in cpumask_of_bp()
148 const struct cpumask *cpumask = cpumask_of_bp(bp); in fetch_bp_busy_slots() local
151 for_each_cpu(cpu, cpumask) { in fetch_bp_busy_slots()
206 const struct cpumask *cpumask = cpumask_of_bp(bp); in toggle_bp_slot() local
219 for_each_cpu(cpu, cpumask) in toggle_bp_slot()
/linux-4.4.14/arch/sh/include/asm/
Dsmp.h37 void smp_timer_broadcast(const struct cpumask *mask);
44 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
/linux-4.4.14/arch/mips/include/asm/
Dsmp-ops.h26 void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action);
48 extern void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action);
Dsmp.h93 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/linux-4.4.14/arch/powerpc/sysdev/
Dmpic.h40 const struct cpumask *cpumask, bool force);
/linux-4.4.14/Documentation/thermal/
Dcpu-cooling-api.txt20 struct cpumask *clip_cpus)
26 clip_cpus: cpumask of cpus where the frequency constraints will happen.
29 struct device_node *np, const struct cpumask *clip_cpus)
37 clip_cpus: cpumask of cpus where the frequency constraints will happen.
40 const struct cpumask *clip_cpus, u32 capacitance,
54 struct device_node *np, const struct cpumask *clip_cpus, u32 capacitance,
171 int plat_get_static(cpumask_t *cpumask, int interval,
174 `cpumask` is the cpumask of the cpus involved in the calculation.
179 temperature of the cpus described by `cpumask` is left for
/linux-4.4.14/arch/mips/pmcs-msp71xx/
Dmsp_irq_cic.c130 const struct cpumask *cpumask, bool force) in msp_cic_irq_set_affinity() argument
144 if (cpumask_test_cpu(cpu, cpumask)) in msp_cic_irq_set_affinity()
Dmsp_irq_per.c89 const struct cpumask *affinity, bool force) in msp_per_irq_set_affinity()
/linux-4.4.14/arch/arc/include/asm/
Dsmp.h21 struct cpumask;
27 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/s390/kernel/
Dsmp.c157 static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address) in pcpu_find_address()
395 static void smp_emergency_stop(cpumask_t *cpumask) in smp_emergency_stop() argument
401 for_each_cpu(cpu, cpumask) { in smp_emergency_stop()
410 for_each_cpu(cpu, cpumask) in smp_emergency_stop()
412 cpumask_clear_cpu(cpu, cpumask); in smp_emergency_stop()
413 if (cpumask_empty(cpumask)) in smp_emergency_stop()
424 cpumask_t cpumask; in smp_send_stop() local
432 cpumask_copy(&cpumask, cpu_online_mask); in smp_send_stop()
433 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_send_stop()
436 smp_emergency_stop(&cpumask); in smp_send_stop()
[all …]
Dtopology.c429 static const struct cpumask *cpu_thread_mask(int cpu) in cpu_thread_mask()
435 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
440 static const struct cpumask *cpu_book_mask(int cpu) in cpu_book_mask()
/linux-4.4.14/drivers/clocksource/
Ddw_apb_timer.c119 cpumask_first(evt->cpumask)); in apbt_shutdown()
133 cpumask_first(evt->cpumask)); in apbt_set_oneshot()
166 cpumask_first(evt->cpumask)); in apbt_set_periodic()
190 cpumask_first(evt->cpumask)); in apbt_resume()
250 dw_ced->ced.cpumask = cpumask_of(cpu); in dw_apb_clockevent_init()
Dmips-gic-timer.c31 gic_write_cpu_compare(cnt, cpumask_first(evt->cpumask)); in gic_next_event()
62 cd->cpumask = cpumask_of(cpu); in gic_clockevent_cpu_init()
Ddummy_timer.c29 evt->cpumask = cpumask_of(cpu); in dummy_timer_setup()
Dnumachip.c75 ced->cpumask = cpumask_of(smp_processor_id()); in numachip_timer_each()
Dclps711x-timer.c89 clkevt->cpumask = cpumask_of(0); in _clps711x_clkevt_init()
/linux-4.4.14/arch/powerpc/platforms/cell/
Dspu_priv1_mmio.c81 const struct cpumask *spumask = cpumask_of_node(spu->node), in cpu_affinity_set()
82 *cpumask = cpumask_of_node(cpu_to_node(cpu)); in cpu_affinity_set() local
84 if (!cpumask_intersects(spumask, cpumask)) in cpu_affinity_set()
/linux-4.4.14/arch/s390/include/asm/
Dtopology.h41 const struct cpumask *cpu_coregroup_mask(int cpu);
69 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
Dsmp.h25 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/arc/kernel/
Dsmp.c250 static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) in ipi_send_msg()
265 struct cpumask targets; in smp_send_stop()
276 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
Dmcip.c213 idu_irq_set_affinity(struct irq_data *data, const struct cpumask *cpumask, in idu_irq_set_affinity() argument
220 if (!cpumask_and(&online, cpumask, cpu_online_mask)) in idu_irq_set_affinity()
/linux-4.4.14/arch/arm64/kernel/
Dsmp.c616 void (*__smp_cross_call)(const struct cpumask *, unsigned int);
618 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument
632 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call()
663 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
766 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
/linux-4.4.14/kernel/trace/
Dring_buffer.c272 for_each_cpu(cpu, buffer->cpumask)
474 cpumask_var_t cpumask; member
543 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_wait()
639 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_poll_wait()
1315 if (!alloc_cpumask_var(&buffer->cpumask, GFP_KERNEL)) in __ring_buffer_alloc()
1337 cpumask_copy(buffer->cpumask, cpu_online_mask); in __ring_buffer_alloc()
1339 cpumask_copy(buffer->cpumask, cpu_possible_mask); in __ring_buffer_alloc()
1375 free_cpumask_var(buffer->cpumask); in __ring_buffer_alloc()
1408 free_cpumask_var(buffer->cpumask); in ring_buffer_free()
1666 !cpumask_test_cpu(cpu_id, buffer->cpumask)) in ring_buffer_resize()
[all …]
/linux-4.4.14/arch/blackfin/mach-bf561/
Dsmp.c27 struct cpumask mask; in platform_init_cpus()
36 struct cpumask mask; in platform_prepare_cpus()
/linux-4.4.14/arch/x86/platform/uv/
Dtlb_uv.c413 cpumask_t *mask = bcp->uvhub_master->cpumask; in reset_with_ipi()
873 int uv_flush_send_and_wait(struct cpumask *flush_mask, struct bau_control *bcp, in uv_flush_send_and_wait()
1024 static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp, in set_distrib_bits()
1078 const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, in uv_flush_tlb_others() argument
1088 struct cpumask *flush_mask; in uv_flush_tlb_others()
1097 return cpumask; in uv_flush_tlb_others()
1108 return cpumask; in uv_flush_tlb_others()
1116 return cpumask; in uv_flush_tlb_others()
1125 flush_mask = (struct cpumask *)per_cpu(uv_flush_tlb_mask, cpu); in uv_flush_tlb_others()
1127 cpumask_andnot(flush_mask, cpumask, cpumask_of(cpu)); in uv_flush_tlb_others()
[all …]
Duv_time.c317 int ced_cpu = cpumask_first(ced->cpumask); in uv_rtc_next_event()
327 int ced_cpu = cpumask_first(evt->cpumask); in uv_rtc_shutdown()
360 ced->cpumask = cpumask_of(smp_processor_id()); in uv_rtc_register_clockevents()
/linux-4.4.14/arch/x86/lib/
Dmsr-smp.c97 static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, in __rwmsr_on_cpus()
125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in rdmsr_on_cpus()
139 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in wrmsr_on_cpus()
/linux-4.4.14/arch/x86/kernel/
Dvsmp_64.c218 static void fill_vector_allocation_domain(int cpu, struct cpumask *retmask, in fill_vector_allocation_domain()
219 const struct cpumask *mask) in fill_vector_allocation_domain()
/linux-4.4.14/arch/metag/include/asm/
Dsmp.h14 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology.h20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
/linux-4.4.14/block/
Dblk-mq.c733 WARN_ON(!cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)); in __blk_mq_run_hw_queue()
845 next_cpu = cpumask_next(hctx->next_cpu, hctx->cpumask); in blk_mq_hctx_next_cpu()
847 next_cpu = cpumask_first(hctx->cpumask); in blk_mq_hctx_next_cpu()
866 if (cpumask_test_cpu(cpu, hctx->cpumask)) { in blk_mq_run_hw_queue()
1673 free_cpumask_var(hctx->cpumask); in blk_mq_free_hw_queues()
1802 const struct cpumask *online_mask) in blk_mq_map_swqueue()
1815 cpumask_clear(hctx->cpumask); in blk_mq_map_swqueue()
1828 cpumask_set_cpu(i, hctx->cpumask); in blk_mq_map_swqueue()
1867 hctx->next_cpu = cpumask_first(hctx->cpumask); in blk_mq_map_swqueue()
1876 cpumask_set_cpu(i, hctx->tags->cpumask); in blk_mq_map_swqueue()
[all …]
Dblk-mq-cpumap.c35 const struct cpumask *online_mask) in blk_mq_update_queue_map()
Dblk-mq-tag.h47 cpumask_var_t cpumask; member
Dblk-mq.h52 const struct cpumask *online_mask);
/linux-4.4.14/arch/parisc/include/asm/
Dsmp.h31 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dirq.h47 extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest);
/linux-4.4.14/arch/mips/include/asm/mach-loongson64/
Dtopology.h8 #define cpumask_of_node(node) (&__node_data[(node)]->cpumask)
Dmmzone.h42 cpumask_t cpumask; member
/linux-4.4.14/arch/mips/bcm63xx/
Dirq.c35 static void (*internal_irq_unmask)(struct irq_data *d, const struct cpumask *m);
55 const struct cpumask *m) in enable_irq_for_cpu()
135 const struct cpumask *m) \
377 const struct cpumask *dest, in bcm63xx_internal_set_affinity()
/linux-4.4.14/kernel/sched/
Dcpupri.h25 struct task_struct *p, struct cpumask *lowest_mask);
Dcpudeadline.h25 struct cpumask *later_mask);
Dsched.h854 unsigned long cpumask[0]; /* iteration mask */ member
871 unsigned long cpumask[0]; member
874 static inline struct cpumask *sched_group_cpus(struct sched_group *sg) in sched_group_cpus()
876 return to_cpumask(sg->cpumask); in sched_group_cpus()
883 static inline struct cpumask *sched_group_mask(struct sched_group *sg) in sched_group_mask()
885 return to_cpumask(sg->sgc->cpumask); in sched_group_mask()
1202 const struct cpumask *newmask);
1258 extern void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask);
Dcore.c1161 void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_common()
1167 void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed()
1206 const struct cpumask *new_mask, bool check) in __set_cpus_allowed_ptr()
1261 int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_ptr()
1550 const struct cpumask *nodemask = NULL; in select_fallback_rq()
1656 const struct cpumask *new_mask, bool check) in __set_cpus_allowed_ptr()
4400 long sched_setaffinity(pid_t pid, const struct cpumask *in_mask) in sched_setaffinity()
4490 struct cpumask *new_mask) in get_user_cpu_mask()
4524 long sched_getaffinity(pid_t pid, struct cpumask *mask) in sched_getaffinity()
5043 int cpuset_cpumask_can_shrink(const struct cpumask *cur, in cpuset_cpumask_can_shrink()
[all …]
/linux-4.4.14/arch/mips/netlogic/common/
Dsmp.c75 void nlm_send_ipi_mask(const struct cpumask *mask, unsigned int action) in nlm_send_ipi_mask()
191 cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask); in nlm_smp_setup()
/linux-4.4.14/arch/sparc/include/asm/
Dsmp_64.h37 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
Dtopology_64.h52 static inline const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
Dsmp_32.h97 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/sh/kernel/
Dlocaltimer.c53 clk->cpumask = cpumask_of(cpu); in local_timer_setup()
Dtopology.c33 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
Dsmp.c275 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
288 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
/linux-4.4.14/arch/mips/include/asm/mach-netlogic/
Dmulti-node.h59 cpumask_t cpumask; /* logical cpu mask for node */ member
/linux-4.4.14/arch/mips/kernel/
Dsmp-up.c21 static inline void up_send_ipi_mask(const struct cpumask *mask, in up_send_ipi_mask()
Dsmp-gic.c60 void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action) in gic_send_ipi_mask()
Dcevt-ds1287.c132 cd->cpumask = cpumask_of(0); in ds1287_clockevent_init()
/linux-4.4.14/arch/tile/lib/
DMakefile5 lib-y = cacheflush.o checksum.o cpumask.o delay.o uaccess.o \
/linux-4.4.14/arch/mn10300/include/asm/
Dsmp.h92 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/linux-4.4.14/arch/metag/kernel/
Dtopology.c29 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
/linux-4.4.14/arch/ia64/hp/sim/
Dhpsim_irq.c27 hpsim_set_affinity_noop(struct irq_data *d, const struct cpumask *b, bool f) in hpsim_set_affinity_noop()
/linux-4.4.14/arch/powerpc/kernel/
Dsmp.c289 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
298 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
589 const struct cpumask *mask; in traverse_siblings_chip_id()
639 const struct cpumask *mask; in traverse_core_siblings()
Dirq.c430 const struct cpumask *map = cpu_online_mask; in migrate_irqs()
643 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu()
672 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu()
/linux-4.4.14/arch/parisc/kernel/
Dsmp.c202 send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op) in send_IPI_mask()
242 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/linux-4.4.14/drivers/staging/lustre/lustre/libcfs/linux/
Dlinux-cpu.c569 cpumask_t *cpumask; in cfs_cpt_bind() local
577 cpumask = cptab->ctb_cpumask; in cfs_cpt_bind()
580 cpumask = cptab->ctb_parts[cpt].cpt_cpumask; in cfs_cpt_bind()
584 if (cpumask_any_and(cpumask, cpu_online_mask) >= nr_cpu_ids) { in cfs_cpt_bind()
591 if (cpumask_test_cpu(i, cpumask)) in cfs_cpt_bind()
594 rc = set_cpus_allowed_ptr(current, cpumask); in cfs_cpt_bind()
/linux-4.4.14/arch/score/kernel/
Dtime.c86 score_clockevent.cpumask = cpumask_of(0); in time_init()
/linux-4.4.14/arch/xtensa/kernel/
Dsmp.c352 static void send_ipi_message(const struct cpumask *callmask, in send_ipi_message()
365 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
382 struct cpumask targets; in smp_send_stop()
/linux-4.4.14/mm/
Drmap.c621 if (cpumask_test_cpu(cpu, &tlb_ubc->cpumask)) in try_to_unmap_flush()
622 percpu_flush_tlb_batch_pages(&tlb_ubc->cpumask); in try_to_unmap_flush()
624 if (cpumask_any_but(&tlb_ubc->cpumask, cpu) < nr_cpu_ids) { in try_to_unmap_flush()
625 smp_call_function_many(&tlb_ubc->cpumask, in try_to_unmap_flush()
628 cpumask_clear(&tlb_ubc->cpumask); in try_to_unmap_flush()
648 cpumask_or(&tlb_ubc->cpumask, &tlb_ubc->cpumask, mm_cpumask(mm)); in set_tlb_ubc_flush_pending()
/linux-4.4.14/arch/openrisc/kernel/
Dtime.c104 clockevent_openrisc_timer.cpumask = cpumask_of(0); in openrisc_clockevent_init()
/linux-4.4.14/arch/blackfin/kernel/
Dtime-ts.c235 evt->cpumask = cpumask_of(0); in bfin_gptmr0_clockevent_init()
349 evt->cpumask = cpumask_of(cpu); in bfin_coretmr_clockevent_init()
/linux-4.4.14/arch/mips/loongson64/loongson-3/
Dnuma.c236 cpumask_clear(&__node_data[(node)]->cpumask); in prom_meminit()
247 cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask); in prom_meminit()
/linux-4.4.14/Documentation/filesystems/
Dsysfs-pci.txt39 local_cpus nearby CPU mask (cpumask, ro)
55 cpumask - file contains a cpumask type
/linux-4.4.14/arch/arm/mach-lpc32xx/
Dtimer.c131 lpc32xx_clkevt.cpumask = cpumask_of(0); in lpc32xx_timer_init()
/linux-4.4.14/arch/arm/mach-omap1/
Dtimer32k.c161 clockevent_32k_timer.cpumask = cpumask_of(0); in omap_init_32k_timer()
/linux-4.4.14/arch/um/kernel/
Dtime.c59 .cpumask = cpu_all_mask,
/linux-4.4.14/arch/mips/alchemy/common/
Dtime.c84 .cpumask = cpu_all_mask,
/linux-4.4.14/arch/arm/mach-footbridge/
Ddc21285-timer.c122 ce->cpumask = cpumask_of(smp_processor_id()); in footbridge_timer_init()
/linux-4.4.14/arch/mips/sibyte/sb1250/
Dsmp.c74 static inline void sb1250_send_ipi_mask(const struct cpumask *mask, in sb1250_send_ipi_mask()
/linux-4.4.14/arch/arm/mach-netx/
Dtime.c150 netx_clockevent.cpumask = cpumask_of(0); in netx_timer_init()
/linux-4.4.14/net/iucv/
Diucv.c563 cpumask_t cpumask; in iucv_setmask_up() local
567 cpumask_copy(&cpumask, &iucv_irq_cpumask); in iucv_setmask_up()
568 cpumask_clear_cpu(cpumask_first(&iucv_irq_cpumask), &cpumask); in iucv_setmask_up()
569 for_each_cpu(cpu, &cpumask) in iucv_setmask_up()
663 cpumask_t cpumask; in iucv_cpu_notify() local
690 cpumask_copy(&cpumask, &iucv_buffer_cpumask); in iucv_cpu_notify()
691 cpumask_clear_cpu(cpu, &cpumask); in iucv_cpu_notify()
692 if (cpumask_empty(&cpumask)) in iucv_cpu_notify()
/linux-4.4.14/arch/mips/sibyte/bcm1480/
Dsmp.c84 static void bcm1480_send_ipi_mask(const struct cpumask *mask, in bcm1480_send_ipi_mask()
/linux-4.4.14/Documentation/cpuidle/
Ddriver.txt27 const struct cpumask *const coupled_cpus);

123