/linux-4.1.27/include/linux/ |
D | cpumask.h | 15 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; struct 88 extern const struct cpumask *const cpu_possible_mask; 89 extern const struct cpumask *const cpu_online_mask; 90 extern const struct cpumask *const cpu_present_mask; 91 extern const struct cpumask *const cpu_active_mask; 124 static inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first() 130 static inline unsigned int cpumask_next(int n, const struct cpumask *srcp) in cpumask_next() 135 static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp) in cpumask_next_zero() 141 const struct cpumask *srcp, in cpumask_next_and() 142 const struct cpumask *andp) in cpumask_next_and() [all …]
|
D | stop_machine.h | 34 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); 35 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); 78 static inline int stop_cpus(const struct cpumask *cpumask, in stop_cpus() argument 81 if (cpumask_test_cpu(raw_smp_processor_id(), cpumask)) in stop_cpus() 86 static inline int try_stop_cpus(const struct cpumask *cpumask, in try_stop_cpus() argument 89 return stop_cpus(cpumask, fn, arg); in try_stop_cpus() 115 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus); 126 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus); 129 const struct cpumask *cpus); 134 const struct cpumask *cpus) in __stop_machine() [all …]
|
D | cpu_cooling.h | 37 cpufreq_cooling_register(const struct cpumask *clip_cpus); 47 const struct cpumask *clip_cpus); 51 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register() 66 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register() 72 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
|
D | interrupt.h | 227 extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask, 238 irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_set_affinity() argument 240 return __irq_set_affinity(irq, cpumask, false); in irq_set_affinity() 255 irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 257 return __irq_set_affinity(irq, cpumask, true); in irq_force_affinity() 263 extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m); 270 static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) in irq_set_affinity() 275 static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 288 const struct cpumask *m) in irq_set_affinity_hint()
|
D | padata.h | 133 struct padata_cpumask cpumask; member 157 struct padata_cpumask cpumask; member 170 const struct cpumask *pcpumask, 171 const struct cpumask *cbcpumask); 177 cpumask_var_t cpumask);
|
D | smp.h | 39 void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func, 96 void smp_call_function_many(const struct cpumask *mask, 99 int smp_call_function_any(const struct cpumask *mask, 145 smp_call_function_any(const struct cpumask *mask, smp_call_func_t func, in smp_call_function_any()
|
D | clockchips.h | 132 void (*broadcast)(const struct cpumask *mask); 142 const struct cpumask *cpumask; member 191 extern void tick_broadcast(const struct cpumask *mask);
|
D | cpuidle.h | 123 struct cpumask *cpumask; member 145 const struct cpumask *const coupled_cpus); 182 const struct cpumask *const coupled_cpus) in cpuidle_register()
|
D | topology.h | 202 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() 208 static inline const struct cpumask *cpu_cpu_mask(int cpu) in cpu_cpu_mask()
|
D | cpuset.h | 44 extern void cpuset_cpus_allowed(struct task_struct *p, struct cpumask *mask); 147 struct cpumask *mask) in cpuset_cpus_allowed()
|
D | irqdesc.h | 67 struct cpumask *percpu_enabled; 69 const struct cpumask *affinity_hint;
|
D | cpu_rmap.h | 43 const struct cpumask *affinity);
|
D | irq.h | 345 int (*irq_set_affinity)(struct irq_data *data, const struct cpumask *dest, bool force); 424 const struct cpumask *cpumask, bool force); 467 const struct cpumask *dest,
|
D | sched.h | 1035 static inline struct cpumask *sched_domain_span(struct sched_domain *sd) in sched_domain_span() 1049 typedef const struct cpumask *(*sched_domain_mask_f)(int cpu); 2120 extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, 2121 const struct cpumask *trial); 2123 const struct cpumask *cs_cpus_allowed); 2126 const struct cpumask *new_mask); 2129 const struct cpumask *new_mask); 2132 const struct cpumask *new_mask) in do_set_cpus_allowed() 2136 const struct cpumask *new_mask) in set_cpus_allowed_ptr() 3015 extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); [all …]
|
/linux-4.1.27/kernel/ |
D | padata.c | 40 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu() 42 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu() 58 cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash() 121 if (!cpumask_test_cpu(cb_cpu, pd->cpumask.cbcpu)) in padata_do_parallel() 177 num_cpus = cpumask_weight(pd->cpumask.pcpu); in padata_get_next() 351 const struct cpumask *pcpumask, in padata_setup_cpumasks() 352 const struct cpumask *cbcpumask) in padata_setup_cpumasks() 354 if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) in padata_setup_cpumasks() 357 cpumask_and(pd->cpumask.pcpu, pcpumask, cpu_online_mask); in padata_setup_cpumasks() 358 if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL)) { in padata_setup_cpumasks() [all …]
|
D | stop_machine.c | 146 const struct cpumask *active_cpus; 330 static void queue_stop_cpus_work(const struct cpumask *cpumask, in queue_stop_cpus_work() argument 338 for_each_cpu(cpu, cpumask) { in queue_stop_cpus_work() 351 for_each_cpu(cpu, cpumask) in queue_stop_cpus_work() 356 static int __stop_cpus(const struct cpumask *cpumask, in __stop_cpus() argument 361 cpu_stop_init_done(&done, cpumask_weight(cpumask)); in __stop_cpus() 362 queue_stop_cpus_work(cpumask, fn, arg, &done); in __stop_cpus() 395 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in stop_cpus() argument 401 ret = __stop_cpus(cpumask, fn, arg); in stop_cpus() 424 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg) in try_stop_cpus() argument [all …]
|
D | smp.c | 27 cpumask_var_t cpumask; member 45 if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, in hotplug_cfd() 50 free_cpumask_var(cfd->cpumask); in hotplug_cfd() 62 free_cpumask_var(cfd->cpumask); in hotplug_cfd() 361 int smp_call_function_any(const struct cpumask *mask, in smp_call_function_any() 365 const struct cpumask *nodemask; in smp_call_function_any() 404 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many() 441 cpumask_and(cfd->cpumask, mask, cpu_online_mask); in smp_call_function_many() 442 cpumask_clear_cpu(this_cpu, cfd->cpumask); in smp_call_function_many() 445 if (unlikely(!cpumask_weight(cfd->cpumask))) in smp_call_function_many() [all …]
|
D | workqueue.c | 1677 set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); in worker_attach_to_pool() 3065 free_cpumask_var(attrs->cpumask); in free_workqueue_attrs() 3086 if (!alloc_cpumask_var(&attrs->cpumask, gfp_mask)) in alloc_workqueue_attrs() 3089 cpumask_copy(attrs->cpumask, cpu_possible_mask); in alloc_workqueue_attrs() 3100 cpumask_copy(to->cpumask, from->cpumask); in copy_workqueue_attrs() 3115 hash = jhash(cpumask_bits(attrs->cpumask), in wqattrs_hash() 3126 if (!cpumask_equal(a->cpumask, b->cpumask)) in wqattrs_equal() 3306 if (cpumask_subset(pool->attrs->cpumask, in get_unbound_pool() 3502 int cpu_going_down, cpumask_t *cpumask) in wq_calc_node_cpumask() argument 3508 cpumask_and(cpumask, cpumask_of_node(node), attrs->cpumask); in wq_calc_node_cpumask() [all …]
|
D | cpu.c | 763 const struct cpumask *const cpu_possible_mask = to_cpumask(cpu_possible_bits); 767 const struct cpumask *const cpu_online_mask = to_cpumask(cpu_online_bits); 771 const struct cpumask *const cpu_present_mask = to_cpumask(cpu_present_bits); 775 const struct cpumask *const cpu_active_mask = to_cpumask(cpu_active_bits); 812 void init_cpu_present(const struct cpumask *src) in init_cpu_present() 817 void init_cpu_possible(const struct cpumask *src) in init_cpu_possible() 822 void init_cpu_online(const struct cpumask *src) in init_cpu_online()
|
D | up.c | 53 void on_each_cpu_mask(const struct cpumask *mask, in on_each_cpu_mask()
|
/linux-4.1.27/arch/m32r/kernel/ |
D | smp.c | 80 static void send_IPI_mask(const struct cpumask *, int, int); 151 cpumask_t cpumask; in smp_flush_cache_all() local 155 cpumask_copy(&cpumask, cpu_online_mask); in smp_flush_cache_all() 156 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_flush_cache_all() 158 mask=cpumask_bits(&cpumask); in smp_flush_cache_all() 160 send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); in smp_flush_cache_all() 370 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument 388 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others() 390 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others() 394 cpumask_and(&cpumask, &cpumask, cpu_online_mask); in flush_tlb_others() [all …]
|
/linux-4.1.27/arch/x86/include/asm/uv/ |
D | uv.h | 6 struct cpumask; 16 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, 28 static inline const struct cpumask * 29 uv_flush_tlb_others(const struct cpumask *cpumask, struct mm_struct *mm, in uv_flush_tlb_others() argument 31 { return cpumask; } in uv_flush_tlb_others()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | apic.h | 278 const struct cpumask *(*target_cpus)(void); 285 void (*vector_allocation_domain)(int cpu, struct cpumask *retmask, 286 const struct cpumask *mask); 301 int (*cpu_mask_to_apicid_and)(const struct cpumask *cpumask, 302 const struct cpumask *andmask, 306 void (*send_IPI_mask)(const struct cpumask *mask, int vector); 307 void (*send_IPI_mask_allbutself)(const struct cpumask *mask, 480 static inline const struct cpumask *default_target_cpus(void) in default_target_cpus() 489 static inline const struct cpumask *online_target_cpus(void) in online_target_cpus() 546 flat_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in flat_cpu_mask_to_apicid_and() argument [all …]
|
D | ipi.h | 122 extern void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, 124 extern void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, 149 extern void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, 151 extern void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, 153 extern void default_send_IPI_mask_logical(const struct cpumask *mask,
|
D | smp.h | 40 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 45 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask() 50 static inline struct cpumask *cpu_llc_shared_mask(int cpu) in cpu_llc_shared_mask() 79 void (*send_call_func_ipi)(const struct cpumask *mask); 147 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 166 void native_send_call_func_ipi(const struct cpumask *mask);
|
D | topology.h | 76 extern const struct cpumask *cpumask_of_node(int node); 79 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node() 120 extern const struct cpumask *cpu_coregroup_mask(int cpu);
|
D | tlbflush.h | 212 static inline void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument 249 void native_flush_tlb_others(const struct cpumask *cpumask,
|
D | msr.h | 225 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); 226 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); 254 static inline void rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, in rdmsr_on_cpus() 259 static inline void wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, in wrmsr_on_cpus()
|
D | x86_init.h | 190 struct cpumask; 200 const struct cpumask *mask,
|
D | hw_irq.h | 142 extern int assign_irq_vector(int, struct irq_cfg *, const struct cpumask *); 155 extern int apic_set_affinity(struct irq_data *data, const struct cpumask *mask,
|
D | pci.h | 135 static inline const struct cpumask *
|
/linux-4.1.27/drivers/cpuidle/ |
D | cpuidle-big_little.c | 149 struct cpumask *cpumask; in bl_idle_driver_init() local 152 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); in bl_idle_driver_init() 153 if (!cpumask) in bl_idle_driver_init() 158 cpumask_set_cpu(cpu, cpumask); in bl_idle_driver_init() 160 drv->cpumask = cpumask; in bl_idle_driver_init() 228 kfree(bl_idle_big_driver.cpumask); in bl_idle_init() 230 kfree(bl_idle_little_driver.cpumask); in bl_idle_init()
|
D | dt_idle_states.c | 99 const cpumask_t *cpumask) in idle_state_valid() argument 112 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid() 113 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid() 158 const cpumask_t *cpumask; in dt_init_idle_driver() local 169 cpumask = drv->cpumask ? : cpu_possible_mask; in dt_init_idle_driver() 170 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver() 180 if (!idle_state_valid(state_node, i, cpumask)) { in dt_init_idle_driver()
|
D | driver.c | 50 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_unset_driver() 72 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_set_driver() 164 if (!drv->cpumask) in __cpuidle_driver_init() 165 drv->cpumask = (struct cpumask *)cpu_possible_mask; in __cpuidle_driver_init() 240 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_register_driver() 264 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_unregister_driver()
|
D | cpuidle.c | 526 for_each_cpu(cpu, drv->cpumask) { in cpuidle_unregister() 547 const struct cpumask *const coupled_cpus) in cpuidle_register() 558 for_each_cpu(cpu, drv->cpumask) { in cpuidle_register()
|
/linux-4.1.27/arch/x86/kernel/apic/ |
D | x2apic_cluster.c | 27 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask() 29 struct cpumask *cpus_in_cluster_ptr; in __x2apic_send_IPI_mask() 30 struct cpumask *ipi_mask_ptr; in __x2apic_send_IPI_mask() 77 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask() 83 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself() 99 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument 100 const struct cpumask *andmask, in x2apic_cpu_mask_to_apicid_and() 107 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and() 118 for_each_cpu_and(i, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and() 211 static const struct cpumask *x2apic_cluster_target_cpus(void) in x2apic_cluster_target_cpus() [all …]
|
D | apic_flat_64.c | 65 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument 67 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask() 73 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument 75 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself() 234 static void physflat_send_IPI_mask(const struct cpumask *cpumask, int vector) in physflat_send_IPI_mask() argument 236 default_send_IPI_mask_sequence_phys(cpumask, vector); in physflat_send_IPI_mask() 239 static void physflat_send_IPI_mask_allbutself(const struct cpumask *cpumask, in physflat_send_IPI_mask_allbutself() argument 242 default_send_IPI_mask_allbutself_phys(cpumask, vector); in physflat_send_IPI_mask_allbutself()
|
D | apic_noop.c | 33 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask() argument 34 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask_allbutself() argument 86 static const struct cpumask *noop_target_cpus(void) in noop_target_cpus() 92 static void noop_vector_allocation_domain(int cpu, struct cpumask *retmask, in noop_vector_allocation_domain() 93 const struct cpumask *mask) in noop_vector_allocation_domain()
|
D | ipi.c | 21 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys() 39 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys() 60 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical() 80 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical() 103 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument 105 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
|
D | x2apic_phys.c | 38 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask() 58 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask() 64 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
|
D | msi.c | 79 msi_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force) in msi_set_affinity() 180 dmar_msi_set_affinity(struct irq_data *data, const struct cpumask *mask, in dmar_msi_set_affinity() 236 const struct cpumask *mask, bool force) in hpet_msi_set_affinity()
|
D | apic_numachip.c | 112 static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) in numachip_send_IPI_mask() 120 static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, in numachip_send_IPI_mask_allbutself()
|
D | x2apic_uv_x.c | 265 static void uv_send_IPI_mask(const struct cpumask *mask, int vector) in uv_send_IPI_mask() 273 static void uv_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in uv_send_IPI_mask_allbutself() 315 uv_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in uv_cpu_mask_to_apicid_and() argument 316 const struct cpumask *andmask, in uv_cpu_mask_to_apicid_and() 325 for_each_cpu_and(cpu, cpumask, andmask) { in uv_cpu_mask_to_apicid_and()
|
D | htirq.c | 40 ht_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force) in ht_set_affinity()
|
/linux-4.1.27/arch/tile/include/asm/ |
D | smp.h | 32 void send_IPI_many(const struct cpumask *mask, int tag); 75 static inline void arch_send_call_function_ipi_mask(struct cpumask *mask) in arch_send_call_function_ipi_mask() 102 extern struct cpumask cpu_lotar_map; 106 extern struct cpumask hash_for_home_map; 109 extern struct cpumask cpu_cacheable_map; 127 static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp, in __cpulist_parse_crop()
|
D | homecache.h | 51 const struct cpumask *cache_cpumask, 54 const struct cpumask *tlb_cpumask, 61 extern void homecache_evict(const struct cpumask *mask);
|
D | topology.h | 23 extern struct cpumask node_2_cpu_mask[]; 39 static inline const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
|
/linux-4.1.27/arch/mn10300/mm/ |
D | tlb-smp.c | 52 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, 94 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument 105 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others() 106 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others() 108 cpumask_and(&tmp, &cpumask, cpu_online_mask); in flush_tlb_others() 109 BUG_ON(!cpumask_equal(&cpumask, &tmp)); in flush_tlb_others() 122 atomic_set_mask(cpumask.bits[0], &flush_cpumask.bits[0]); in flush_tlb_others()
|
/linux-4.1.27/kernel/time/ |
D | tick-common.c | 168 const struct cpumask *cpumask) in tick_setup_device() argument 206 if (!cpumask_equal(newdev->cpumask, cpumask)) in tick_setup_device() 207 irq_set_affinity(newdev->irq, cpumask); in tick_setup_device() 239 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_percpu() 241 if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 247 if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 269 !cpumask_equal(curdev->cpumask, newdev->cpumask); in tick_check_preferred() 296 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_new_device()
|
D | tick-broadcast.c | 54 struct cpumask *tick_get_broadcast_mask(void) in tick_get_broadcast_mask() 138 static void err_broadcast(const struct cpumask *mask) in err_broadcast() 258 static void tick_do_broadcast(struct cpumask *mask) in tick_do_broadcast() 502 struct cpumask *tick_get_broadcast_oneshot_mask(void) in tick_get_broadcast_oneshot_mask() 523 const struct cpumask *cpumask) in tick_broadcast_set_affinity() argument 528 if (cpumask_equal(bc->cpumask, cpumask)) in tick_broadcast_set_affinity() 531 bc->cpumask = cpumask; in tick_broadcast_set_affinity() 532 irq_set_affinity(bc->irq, bc->cpumask); in tick_broadcast_set_affinity() 823 static void tick_broadcast_init_next_event(struct cpumask *mask, in tick_broadcast_init_next_event()
|
D | tick-internal.h | 63 extern struct cpumask *tick_get_broadcast_mask(void); 124 extern struct cpumask *tick_get_broadcast_oneshot_mask(void);
|
D | tick-broadcast-hrtimer.c | 95 .cpumask = cpu_all_mask,
|
D | clockevents.c | 485 if (!dev->cpumask) { in clockevents_register_device() 487 dev->cpumask = cpumask_of(smp_processor_id()); in clockevents_register_device() 665 if (cpumask_test_cpu(cpu, dev->cpumask) && in tick_cleanup_dead_cpu() 666 cpumask_weight(dev->cpumask) == 1 && in tick_cleanup_dead_cpu()
|
/linux-4.1.27/arch/tile/kernel/ |
D | hardwall.c | 105 struct cpumask cpumask; /* cpus reserved */ member 202 static int check_rectangle(struct hardwall_info *r, struct cpumask *mask) in check_rectangle() 296 struct cpumask rect_cpus; in hardwall_protect_rectangle() 351 if (cpumask_test_cpu(cpu, &rect->cpumask)) in do_hardwall_trap() 477 return info && cpumask_test_cpu(cpu, &info->cpumask); in hardwall_ipi_valid() 493 struct cpumask mask; in hardwall_create() 502 if (copy_from_user(&mask, bits, min(sizeof(struct cpumask), size))) in hardwall_create() 510 if (size < sizeof(struct cpumask)) { in hardwall_create() 511 memset((char *)&mask + size, 0, sizeof(struct cpumask) - size); in hardwall_create() 512 } else if (size > sizeof(struct cpumask)) { in hardwall_create() [all …]
|
D | smpboot.c | 124 static __initdata struct cpumask init_affinity; 135 static struct cpumask cpu_started;
|
D | setup.c | 847 struct cpumask node_2_cpu_mask[MAX_NUMNODES] __write_once; 855 static int __init cpu_to_bound_node(int cpu, struct cpumask* unbound_cpus) in cpu_to_bound_node() 865 struct cpumask *unbound_cpus) in node_neighbors() 889 struct cpumask unbound_cpus; in setup_numa_mapping() 1274 struct cpumask __write_once cpu_lotar_map; 1284 struct cpumask hash_for_home_map; 1296 struct cpumask __write_once cpu_cacheable_map; 1299 static __initdata struct cpumask disabled_map; 1325 struct cpumask hv_disabled_map, cpu_possible_init; in setup_cpu_maps() 1398 struct cpumask shared; in check_nohz_full_cpus()
|
D | smp.c | 73 void send_IPI_many(const struct cpumask *mask, int tag) in send_IPI_many() 92 struct cpumask mask; in send_IPI_allbutself()
|
/linux-4.1.27/arch/hexagon/include/asm/ |
D | smp.h | 36 extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg); 39 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/tile/mm/ |
D | migrate.h | 33 const unsigned long *cpumask); 51 const struct cpumask *cache_cpumask, 52 const struct cpumask *tlb_cpumask,
|
D | homecache.c | 64 static void hv_flush_update(const struct cpumask *cache_cpumask, in hv_flush_update() 65 struct cpumask *tlb_cpumask, in hv_flush_update() 69 struct cpumask mask; in hv_flush_update() 108 const struct cpumask *cache_cpumask_orig, in flush_remote() 111 const struct cpumask *tlb_cpumask_orig, in flush_remote() 115 struct cpumask cache_cpumask_copy, tlb_cpumask_copy; in flush_remote() 116 struct cpumask *cache_cpumask, *tlb_cpumask; in flush_remote() 238 void homecache_evict(const struct cpumask *mask) in homecache_evict()
|
D | init.c | 203 static __initdata struct cpumask kdata_mask; 318 static struct cpumask __initdata ktext_mask; 412 const struct cpumask *my_cpu_mask = cpumask_of(smp_processor_id()); in kernel_physical_mapping_init() 413 struct cpumask kstripe_mask; in kernel_physical_mapping_init() 493 struct cpumask bad; in kernel_physical_mapping_init()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | smp.c | 169 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask() 259 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local 265 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask() 269 for_each_cpu(cpu, &cpumask) { in smp_flush_tlb_cpumask() 279 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask() 324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
D | irq.c | 76 bool is_affinity_mask_valid(const struct cpumask *cpumask) in is_affinity_mask_valid() argument 80 if (cpumask_weight(cpumask) != 1) in is_affinity_mask_valid()
|
/linux-4.1.27/arch/mn10300/kernel/ |
D | smp.c | 107 static void send_IPI_mask(const cpumask_t *cpumask, int irq); 313 static void send_IPI_mask(const cpumask_t *cpumask, int irq) in send_IPI_mask() argument 319 if (cpumask_test_cpu(i, cpumask)) { in send_IPI_mask() 350 cpumask_t cpumask; in send_IPI_allbutself() local 352 cpumask_copy(&cpumask, cpu_online_mask); in send_IPI_allbutself() 353 cpumask_clear_cpu(smp_processor_id(), &cpumask); in send_IPI_allbutself() 354 send_IPI_mask(&cpumask, irq); in send_IPI_allbutself() 357 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 1067 static int hotplug_cpu_nmi_call_function(cpumask_t cpumask, in hotplug_cpu_nmi_call_function() argument 1084 nmi_call_func_mask_data.started = cpumask; in hotplug_cpu_nmi_call_function() [all …]
|
/linux-4.1.27/Documentation/ |
D | padata.txt | 17 const struct cpumask *pcpumask, 18 const struct cpumask *cbcpumask); 35 it is legal to supply a cpumask to padata that contains offline CPUs. 36 Once an offline CPU in the user supplied cpumask comes online, padata 47 padata cpumask contains no active CPU (flag not set). 57 cpumask_var_t cpumask); 68 PADATA_CPU_PARALLEL and cpumask specifies the new cpumask to use. 69 To simply add or remove one CPU from a certain cpumask the functions 73 If a user is interested in padata cpumask changes, he can register to 74 the padata cpumask change notifier: [all …]
|
/linux-4.1.27/arch/arm64/include/asm/ |
D | smp.h | 50 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); 52 extern void (*__smp_cross_call)(const struct cpumask *, unsigned int); 69 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | topology.h | 25 const struct cpumask *cpu_coregroup_mask(int cpu);
|
/linux-4.1.27/arch/hexagon/kernel/ |
D | smp.c | 105 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument 113 for_each_cpu(cpu, cpumask) { in send_ipi() 239 struct cpumask targets; in smp_send_stop() 250 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
D | time.c | 116 static void broadcast(const struct cpumask *mask) in broadcast() 148 dummy_clock_dev->cpumask = cpumask_of(cpu); in setup_percpu_clockdev() 195 ce_dev->cpumask = cpu_all_mask; in time_init_deferred()
|
/linux-4.1.27/arch/x86/mm/ |
D | tlb.c | 133 void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument 147 cpumask = uv_flush_tlb_others(cpumask, mm, start, end, cpu); in native_flush_tlb_others() 148 if (cpumask) in native_flush_tlb_others() 149 smp_call_function_many(cpumask, flush_tlb_func, in native_flush_tlb_others() 153 smp_call_function_many(cpumask, flush_tlb_func, &info, 1); in native_flush_tlb_others()
|
/linux-4.1.27/Documentation/accounting/ |
D | getdelays.c | 72 char cpumask[100+6*MAX_CPUS]; variable 316 strncpy(cpumask, optarg, sizeof(cpumask)); in main() 317 cpumask[sizeof(cpumask) - 1] = '\0'; in main() 319 printf("cpumask %s maskset %d\n", cpumask, maskset); in main() 393 &cpumask, strlen(cpumask) + 1); in main() 534 &cpumask, strlen(cpumask) + 1); in main()
|
D | taskstats.txt | 35 sends a register command and specifies a cpumask. Whenever a task exits on 36 one of the cpus in the cpumask, its per-pid statistics are sent to the 84 TASKSTATS_CMD_ATTR_REGISTER/DEREGISTER_CPUMASK and contain a cpumask in the 85 attribute payload. The cpumask is specified as an ascii string of 87 the cpumask would be "1-3,5,7-8". If userspace forgets to deregister interest
|
/linux-4.1.27/arch/blackfin/mach-common/ |
D | smp.c | 189 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument 196 for_each_cpu(cpu, cpumask) { in send_ipi() 203 for_each_cpu(cpu, cpumask) in send_ipi() 212 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 224 void smp_send_msg(const struct cpumask *mask, unsigned long type) in smp_send_msg() 229 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
|
/linux-4.1.27/arch/arm/mach-vexpress/ |
D | dcscb.c | 43 unsigned int rst_hold, cpumask = (1 << cpu); in dcscb_cpu_powerup() local 46 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) in dcscb_cpu_powerup() 50 rst_hold &= ~(cpumask | (cpumask << 4)); in dcscb_cpu_powerup()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | smp.h | 96 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 101 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask() 139 static inline const struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 189 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | xics.h | 156 extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, 159 #define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server) argument
|
D | irq.h | 71 int irq_choose_cpu(const struct cpumask *mask);
|
D | cputhreads.h | 42 static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads) in cpu_thread_mask_to_cores()
|
D | fadump.h | 194 struct cpumask cpu_online_mask;
|
/linux-4.1.27/arch/arm/include/asm/ |
D | smp.h | 49 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); 79 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); 80 extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask);
|
D | topology.h | 25 const struct cpumask *cpu_coregroup_mask(int cpu);
|
/linux-4.1.27/drivers/irqchip/ |
D | irq-metag.c | 44 const struct cpumask *cpumask, bool force); 193 const struct cpumask *cpumask, bool force) in metag_internal_irq_set_affinity() argument 204 cpu = cpumask_any_and(cpumask, cpu_online_mask); in metag_internal_irq_set_affinity()
|
D | irq-hip04.c | 145 const struct cpumask *mask_val, in hip04_irq_set_affinity() 228 u32 cpumask; in hip04_irq_dist_init() local 237 cpumask = hip04_get_cpumask(intc); in hip04_irq_dist_init() 238 cpumask |= cpumask << 16; in hip04_irq_dist_init() 240 writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3)); in hip04_irq_dist_init() 276 static void hip04_raise_softirq(const struct cpumask *mask, unsigned int irq) in hip04_raise_softirq()
|
D | irq-gic.c | 237 static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, in gic_set_affinity() 383 u32 cpumask; in gic_dist_init() local 392 cpumask = gic_get_cpumask(gic); in gic_dist_init() 393 cpumask |= cpumask << 8; in gic_dist_init() 394 cpumask |= cpumask << 16; in gic_dist_init() 396 writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4); in gic_dist_init() 632 static void gic_raise_softirq(const struct cpumask *mask, unsigned int irq) in gic_raise_softirq()
|
D | irq-mips-gic.c | 386 static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask, in gic_set_affinity() argument 394 cpumask_and(&tmp, cpumask, cpu_online_mask); in gic_set_affinity() 409 cpumask_copy(d->affinity, cpumask); in gic_set_affinity()
|
D | irq-metag-ext.c | 505 const struct cpumask *cpumask, bool force) in meta_intc_set_affinity() argument 518 cpu = cpumask_any_and(cpumask, cpu_online_mask); in meta_intc_set_affinity()
|
/linux-4.1.27/Documentation/thermal/ |
D | cpu-cooling-api.txt | 20 struct cpumask *clip_cpus) 26 clip_cpus: cpumask of cpus where the frequency constraints will happen. 29 struct device_node *np, const struct cpumask *clip_cpus) 37 clip_cpus: cpumask of cpus where the frequency constraints will happen.
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | irq.h | 30 bool is_affinity_mask_valid(const struct cpumask *cpumask);
|
D | smp.h | 129 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/x86/xen/ |
D | smp.h | 3 extern void xen_send_IPI_mask(const struct cpumask *mask, 5 extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
|
D | smp.c | 564 static void __xen_send_IPI_mask(const struct cpumask *mask, in __xen_send_IPI_mask() 573 static void xen_smp_send_call_function_ipi(const struct cpumask *mask) in xen_smp_send_call_function_ipi() 626 void xen_send_IPI_mask(const struct cpumask *mask, in xen_send_IPI_mask() 651 void xen_send_IPI_mask_allbutself(const struct cpumask *mask, in xen_send_IPI_mask_allbutself()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | smp.c | 176 wait_for_txrdy (unsigned long cpumask) in wait_for_txrdy() argument 180 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy() 185 if (!(hwrpb->txrdy & cpumask)) in wait_for_txrdy() 203 unsigned long cpumask; in send_secondary_console_msg() local 211 cpumask = (1UL << cpuid); in send_secondary_console_msg() 212 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg() 225 if (wait_for_txrdy(cpumask)) in send_secondary_console_msg() 509 send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation) in send_ipi_message() 596 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
D | time.c | 131 .cpumask = cpumask_of(cpu), in init_rtc_clockevent() 199 .cpumask = cpumask_of(cpu), in init_qemu_clockevent()
|
/linux-4.1.27/lib/ |
D | cpumask.c | 16 int cpumask_next_and(int n, const struct cpumask *src1p, in cpumask_next_and() 17 const struct cpumask *src2p) in cpumask_next_and() 34 int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) in cpumask_any_but()
|
D | cpu_rmap.c | 98 const struct cpumask *mask, u16 dist) in cpu_rmap_copy_neigh() 159 const struct cpumask *affinity) in cpu_rmap_update()
|
/linux-4.1.27/kernel/irq/ |
D | manage.c | 164 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) in irq_copy_pending() 169 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) in irq_get_pending() 177 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) { } in irq_copy_pending() 179 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) { } in irq_get_pending() 182 int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, in irq_do_set_affinity() 202 int irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask, in irq_set_affinity_locked() 228 int __irq_set_affinity(unsigned int irq, const struct cpumask *mask, bool force) in __irq_set_affinity() 243 int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m) in irq_set_affinity_hint() 264 cpumask_var_t cpumask; in irq_affinity_notify() local 267 if (!desc || !alloc_cpumask_var(&cpumask, GFP_KERNEL)) in irq_affinity_notify() [all …]
|
D | internals.h | 113 extern int irq_select_affinity_usr(unsigned int irq, struct cpumask *mask); 118 const struct cpumask *dest, bool force);
|
/linux-4.1.27/arch/arm/kernel/ |
D | smp.c | 446 static void (*__smp_cross_call)(const struct cpumask *, unsigned int); 448 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument 466 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call() 498 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 503 void arch_send_wakeup_ipi_mask(const struct cpumask *mask) in arch_send_wakeup_ipi_mask() 522 void tick_broadcast(const struct cpumask *mask) in tick_broadcast() 650 struct cpumask mask; in smp_send_stop()
|
D | topology.c | 183 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask() 192 const struct cpumask *cpu_corepower_mask(int cpu) in cpu_corepower_mask()
|
/linux-4.1.27/arch/xtensa/include/asm/ |
D | smp.h | 22 struct cpumask; 23 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/powerpc/sysdev/xics/ |
D | ics-rtas.c | 122 const struct cpumask *cpumask, in ics_rtas_set_affinity() argument 141 irq_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_rtas_set_affinity() 144 __func__, cpumask_pr_args(cpumask), d->irq); in ics_rtas_set_affinity()
|
D | ics-opal.c | 111 const struct cpumask *cpumask, in ics_opal_set_affinity() argument 132 wanted_server = xics_get_irq_server(d->irq, cpumask, 1); in ics_opal_set_affinity() 135 __func__, cpumask_pr_args(cpumask), d->irq); in ics_opal_set_affinity()
|
D | xics-common.c | 272 int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask, in xics_get_irq_server() argument 279 if (!cpumask_subset(cpu_possible_mask, cpumask)) { in xics_get_irq_server() 280 int server = cpumask_first_and(cpu_online_mask, cpumask); in xics_get_irq_server()
|
/linux-4.1.27/drivers/cpufreq/ |
D | acpi-cpufreq.c | 103 static void boost_set_msrs(bool enable, const struct cpumask *cpumask) in boost_set_msrs() argument 122 rdmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs() 124 for_each_cpu(cpu, cpumask) { in boost_set_msrs() 132 wrmsr_on_cpus(cpumask, msr_addr, msrs); in boost_set_msrs() 256 const struct cpumask *mask; 330 static u32 get_cur_val(const struct cpumask *mask) in get_cur_val() 393 static unsigned int check_freqs(const struct cpumask *mask, unsigned int freq, in check_freqs() 525 const struct cpumask *cpumask; in boost_notify() local 527 cpumask = get_cpu_mask(cpu); in boost_notify() 539 boost_set_msrs(acpi_cpufreq_driver.boost_enabled, cpumask); in boost_notify() [all …]
|
D | powernow-k8.h | 40 struct cpumask *available_cores;
|
/linux-4.1.27/arch/blackfin/include/asm/ |
D | smp.h | 22 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); 51 void smp_timer_broadcast(const struct cpumask *mask);
|
/linux-4.1.27/arch/alpha/include/asm/ |
D | topology.h | 25 extern struct cpumask node_to_cpumask_map[]; 27 static const struct cpumask *cpumask_of_node(int node) in cpumask_of_node()
|
D | smp.h | 48 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/drivers/sh/intc/ |
D | chip.c | 69 const struct cpumask *cpumask, in intc_set_affinity() argument 72 if (!cpumask_intersects(cpumask, cpu_online_mask)) in intc_set_affinity() 75 cpumask_copy(data->affinity, cpumask); in intc_set_affinity()
|
/linux-4.1.27/drivers/iommu/ |
D | irq_remapping.h | 30 struct cpumask; 61 int (*set_affinity)(struct irq_data *data, const struct cpumask *mask,
|
D | irq_remapping.c | 31 const struct cpumask *mask, 270 const struct cpumask *mask, bool force) in set_remapped_irq_affinity()
|
/linux-4.1.27/kernel/events/ |
D | hw_breakpoint.c | 133 static const struct cpumask *cpumask_of_bp(struct perf_event *bp) in cpumask_of_bp() 148 const struct cpumask *cpumask = cpumask_of_bp(bp); in fetch_bp_busy_slots() local 151 for_each_cpu(cpu, cpumask) { in fetch_bp_busy_slots() 206 const struct cpumask *cpumask = cpumask_of_bp(bp); in toggle_bp_slot() local 219 for_each_cpu(cpu, cpumask) in toggle_bp_slot()
|
/linux-4.1.27/crypto/ |
D | pcrypt.c | 76 struct pcrypt_cpumask *cpumask; in pcrypt_do_parallel() local 81 cpumask = rcu_dereference_bh(pcrypt->cb_cpumask); in pcrypt_do_parallel() 82 if (cpumask_test_cpu(cpu, cpumask->mask)) in pcrypt_do_parallel() 85 if (!cpumask_weight(cpumask->mask)) in pcrypt_do_parallel() 88 cpu_index = cpu % cpumask_weight(cpumask->mask); in pcrypt_do_parallel() 90 cpu = cpumask_first(cpumask->mask); in pcrypt_do_parallel() 92 cpu = cpumask_next(cpu, cpumask->mask); in pcrypt_do_parallel() 413 struct padata_cpumask *cpumask = (struct padata_cpumask *)data; in pcrypt_cpumask_change_notify() local 429 cpumask_copy(new_mask->mask, cpumask->cbcpu); in pcrypt_cpumask_change_notify()
|
/linux-4.1.27/arch/sh/include/asm/ |
D | smp.h | 37 void smp_timer_broadcast(const struct cpumask *mask); 44 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | topology.h | 20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
|
/linux-4.1.27/arch/powerpc/sysdev/ |
D | mpic.h | 40 const struct cpumask *cpumask, bool force);
|
/linux-4.1.27/arch/mips/include/asm/ |
D | smp-ops.h | 26 void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action); 48 extern void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action);
|
D | smp.h | 95 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/linux-4.1.27/arch/arc/include/asm/ |
D | smp.h | 21 struct cpumask; 27 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/mips/pmcs-msp71xx/ |
D | msp_irq_cic.c | 129 const struct cpumask *cpumask, bool force) in msp_cic_irq_set_affinity() argument 143 if (cpumask_test_cpu(cpu, cpumask)) in msp_cic_irq_set_affinity()
|
D | msp_irq_per.c | 89 const struct cpumask *affinity, bool force) in msp_per_irq_set_affinity()
|
/linux-4.1.27/arch/arm64/kernel/ |
D | smp.c | 475 void (*__smp_cross_call)(const struct cpumask *, unsigned int); 477 void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int)) in set_smp_cross_call() argument 491 static void smp_cross_call(const struct cpumask *target, unsigned int ipinr) in smp_cross_call() 522 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 625 void tick_broadcast(const struct cpumask *mask) in tick_broadcast()
|
D | irq.c | 64 const struct cpumask *affinity = d->affinity; in migrate_one_irq()
|
/linux-4.1.27/arch/s390/kernel/ |
D | smp.c | 155 static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address) in pcpu_find_address() 388 static void smp_emergency_stop(cpumask_t *cpumask) in smp_emergency_stop() argument 394 for_each_cpu(cpu, cpumask) { in smp_emergency_stop() 403 for_each_cpu(cpu, cpumask) in smp_emergency_stop() 405 cpumask_clear_cpu(cpu, cpumask); in smp_emergency_stop() 406 if (cpumask_empty(cpumask)) in smp_emergency_stop() 417 cpumask_t cpumask; in smp_send_stop() local 425 cpumask_copy(&cpumask, cpu_online_mask); in smp_send_stop() 426 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_send_stop() 429 smp_emergency_stop(&cpumask); in smp_send_stop() [all …]
|
D | topology.c | 424 static const struct cpumask *cpu_thread_mask(int cpu) in cpu_thread_mask() 430 const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask() 435 static const struct cpumask *cpu_book_mask(int cpu) in cpu_book_mask()
|
/linux-4.1.27/arch/arc/kernel/ |
D | smp.c | 242 static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) in ipi_send_msg() 257 struct cpumask targets; in smp_send_stop() 268 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | time_32.c | 139 ce->cpumask = cpu_possible_mask; in setup_timer_ce() 197 int cpu = cpumask_first(evt->cpumask); in percpu_ce_setup() 217 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_next_event() 237 ce->cpumask = cpumask_of(cpu); in register_percpu_ce()
|
/linux-4.1.27/drivers/thermal/ |
D | cpu_cooling.c | 72 struct cpumask allowed_cpus; 325 const struct cpumask *clip_cpus) in __cpufreq_cooling_register() 423 cpufreq_cooling_register(const struct cpumask *clip_cpus) in cpufreq_cooling_register() 444 const struct cpumask *clip_cpus) in of_cpufreq_cooling_register()
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
D | spu_priv1_mmio.c | 81 const struct cpumask *spumask = cpumask_of_node(spu->node), in cpu_affinity_set() 82 *cpumask = cpumask_of_node(cpu_to_node(cpu)); in cpu_affinity_set() local 84 if (!cpumask_intersects(spumask, cpumask)) in cpu_affinity_set()
|
/linux-4.1.27/kernel/trace/ |
D | ring_buffer.c | 324 for_each_cpu(cpu, buffer->cpumask) 498 cpumask_var_t cpumask; member 567 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_wait() 663 if (!cpumask_test_cpu(cpu, buffer->cpumask)) in ring_buffer_poll_wait() 1339 if (!alloc_cpumask_var(&buffer->cpumask, GFP_KERNEL)) in __ring_buffer_alloc() 1361 cpumask_copy(buffer->cpumask, cpu_online_mask); in __ring_buffer_alloc() 1363 cpumask_copy(buffer->cpumask, cpu_possible_mask); in __ring_buffer_alloc() 1399 free_cpumask_var(buffer->cpumask); in __ring_buffer_alloc() 1432 free_cpumask_var(buffer->cpumask); in ring_buffer_free() 1690 !cpumask_test_cpu(cpu_id, buffer->cpumask)) in ring_buffer_resize() [all …]
|
/linux-4.1.27/arch/blackfin/mach-bf561/ |
D | smp.c | 27 struct cpumask mask; in platform_init_cpus() 36 struct cpumask mask; in platform_prepare_cpus()
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | tlb_uv.c | 413 cpumask_t *mask = bcp->uvhub_master->cpumask; in reset_with_ipi() 873 int uv_flush_send_and_wait(struct cpumask *flush_mask, struct bau_control *bcp, in uv_flush_send_and_wait() 1024 static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp, in set_distrib_bits() 1078 const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, in uv_flush_tlb_others() argument 1088 struct cpumask *flush_mask; in uv_flush_tlb_others() 1097 return cpumask; in uv_flush_tlb_others() 1108 return cpumask; in uv_flush_tlb_others() 1116 return cpumask; in uv_flush_tlb_others() 1125 flush_mask = (struct cpumask *)per_cpu(uv_flush_tlb_mask, cpu); in uv_flush_tlb_others() 1127 cpumask_andnot(flush_mask, cpumask, cpumask_of(cpu)); in uv_flush_tlb_others() [all …]
|
D | uv_irq.c | 31 static int uv_set_irq_affinity(struct irq_data *, const struct cpumask *, bool); 133 const struct cpumask *eligible_cpu = cpumask_of(cpu); in arch_enable_uv_irq() 198 uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask, in uv_set_irq_affinity()
|
D | uv_time.c | 318 int ced_cpu = cpumask_first(ced->cpumask); in uv_rtc_next_event() 329 int ced_cpu = cpumask_first(evt->cpumask); in uv_rtc_timer_setup() 371 ced->cpumask = cpumask_of(smp_processor_id()); in uv_rtc_register_clockevents()
|
/linux-4.1.27/drivers/clocksource/ |
D | mips-gic-timer.c | 31 gic_write_cpu_compare(cnt, cpumask_first(evt->cpumask)); in gic_next_event() 68 cd->cpumask = cpumask_of(cpu); in gic_clockevent_cpu_init()
|
D | dummy_timer.c | 39 evt->cpumask = cpumask_of(cpu); in dummy_timer_setup()
|
D | i8253.c | 181 i8253_clockevent.cpumask = cpumask_of(smp_processor_id()); in clockevent_i8253_init()
|
D | clps711x-timer.c | 95 clkevt->cpumask = cpumask_of(0); in _clps711x_clkevt_init()
|
D | dw_apb_timer.c | 121 cpumask_first(evt->cpumask), in apbt_set_mode() 234 dw_ced->ced.cpumask = cpumask_of(cpu); in dw_apb_clockevent_init()
|
/linux-4.1.27/arch/x86/lib/ |
D | msr-smp.c | 97 static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, in __rwmsr_on_cpus() 125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in rdmsr_on_cpus() 139 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) in wrmsr_on_cpus()
|
/linux-4.1.27/arch/x86/kernel/ |
D | vsmp_64.c | 218 static void fill_vector_allocation_domain(int cpu, struct cpumask *retmask, in fill_vector_allocation_domain() 219 const struct cpumask *mask) in fill_vector_allocation_domain()
|
/linux-4.1.27/arch/metag/include/asm/ |
D | smp.h | 14 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | topology.h | 20 const struct cpumask *cpu_coregroup_mask(unsigned int cpu);
|
/linux-4.1.27/arch/parisc/include/asm/ |
D | smp.h | 31 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | irq.h | 47 extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest);
|
/linux-4.1.27/arch/mips/include/asm/mach-loongson/ |
D | topology.h | 8 #define cpumask_of_node(node) (&__node_data[(node)]->cpumask)
|
D | mmzone.h | 42 cpumask_t cpumask; member
|
/linux-4.1.27/arch/mips/netlogic/common/ |
D | smp.c | 75 void nlm_send_ipi_mask(const struct cpumask *mask, unsigned int action) in nlm_send_ipi_mask() 189 cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask); in nlm_smp_setup()
|
/linux-4.1.27/arch/mips/bcm63xx/ |
D | irq.c | 35 static void (*internal_irq_unmask)(struct irq_data *d, const struct cpumask *m); 55 const struct cpumask *m) in enable_irq_for_cpu() 135 const struct cpumask *m) \ 377 const struct cpumask *dest, in bcm63xx_internal_set_affinity()
|
/linux-4.1.27/kernel/sched/ |
D | cpupri.h | 25 struct task_struct *p, struct cpumask *lowest_mask);
|
D | cpudeadline.h | 24 struct cpumask *later_mask);
|
D | sched.h | 837 unsigned long cpumask[0]; /* iteration mask */ member 854 unsigned long cpumask[0]; member 857 static inline struct cpumask *sched_group_cpus(struct sched_group *sg) in sched_group_cpus() 859 return to_cpumask(sg->cpumask); in sched_group_cpus() 866 static inline struct cpumask *sched_group_mask(struct sched_group *sg) in sched_group_mask() 868 return to_cpumask(sg->sgc->cpumask); in sched_group_mask() 1194 const struct cpumask *newmask);
|
D | core.c | 1308 const struct cpumask *nodemask = NULL; in select_fallback_rq() 4017 long sched_setaffinity(pid_t pid, const struct cpumask *in_mask) in sched_setaffinity() 4107 struct cpumask *new_mask) in get_user_cpu_mask() 4141 long sched_getaffinity(pid_t pid, struct cpumask *mask) in sched_getaffinity() 4650 int cpuset_cpumask_can_shrink(const struct cpumask *cur, in cpuset_cpumask_can_shrink() 4651 const struct cpumask *trial) in cpuset_cpumask_can_shrink() 4675 const struct cpumask *cs_cpus_allowed) in task_can_attach() 4756 void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed() 4788 int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_ptr() 5408 struct cpumask *groupmask) in sched_domain_debug_one() [all …]
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | smp_64.h | 37 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | topology_64.h | 49 static inline const struct cpumask *cpu_coregroup_mask(int cpu) in cpu_coregroup_mask()
|
D | smp_32.h | 97 void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/mips/include/asm/mach-netlogic/ |
D | multi-node.h | 59 cpumask_t cpumask; /* logical cpu mask for node */ member
|
/linux-4.1.27/arch/sh/kernel/ |
D | localtimer.c | 59 clk->cpumask = cpumask_of(cpu); in local_timer_setup()
|
D | topology.c | 33 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
|
D | smp.c | 275 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 288 void smp_timer_broadcast(const struct cpumask *mask) in smp_timer_broadcast()
|
/linux-4.1.27/arch/mips/kernel/ |
D | smp-up.c | 21 static inline void up_send_ipi_mask(const struct cpumask *mask, in up_send_ipi_mask()
|
D | smp-gic.c | 58 void gic_send_ipi_mask(const struct cpumask *mask, unsigned int action) in gic_send_ipi_mask()
|
D | cevt-ds1287.c | 125 cd->cpumask = cpumask_of(0); in ds1287_clockevent_init()
|
D | cevt-gt641xx.c | 135 cd->cpumask = cpumask_of(0); in gt641xx_timer0_clockevent_init()
|
D | smp-bmips.c | 292 static void bmips5000_send_ipi_mask(const struct cpumask *mask, in bmips5000_send_ipi_mask() 344 static void bmips43xx_send_ipi_mask(const struct cpumask *mask, in bmips43xx_send_ipi_mask()
|
/linux-4.1.27/arch/tile/lib/ |
D | Makefile | 5 lib-y = cacheflush.o checksum.o cpumask.o delay.o uaccess.o \
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | smp.h | 92 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-4.1.27/arch/ia64/hp/sim/ |
D | hpsim_irq.c | 27 hpsim_set_affinity_noop(struct irq_data *d, const struct cpumask *b, bool f) in hpsim_set_affinity_noop()
|
/linux-4.1.27/include/trace/events/ |
D | ipi.h | 20 TP_PROTO(const struct cpumask *mask, const char *reason),
|
/linux-4.1.27/arch/metag/kernel/ |
D | topology.c | 29 const struct cpumask *cpu_coregroup_mask(unsigned int cpu) in cpu_coregroup_mask()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | smp.c | 289 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 298 void tick_broadcast(const struct cpumask *mask) in tick_broadcast() 589 const struct cpumask *mask; in traverse_siblings_chip_id() 639 const struct cpumask *mask; in traverse_core_siblings()
|
D | irq.c | 430 const struct cpumask *map = cpu_online_mask; in migrate_irqs() 643 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu() 672 int irq_choose_cpu(const struct cpumask *mask) in irq_choose_cpu()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | smp.c | 202 send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op) in send_IPI_mask() 242 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/ |
D | linux-cpu.c | 569 cpumask_t *cpumask; in cfs_cpt_bind() local 577 cpumask = cptab->ctb_cpumask; in cfs_cpt_bind() 580 cpumask = cptab->ctb_parts[cpt].cpt_cpumask; in cfs_cpt_bind() 584 if (cpumask_any_and(cpumask, cpu_online_mask) >= nr_cpu_ids) { in cfs_cpt_bind() 591 if (cpumask_test_cpu(i, cpumask)) in cfs_cpt_bind() 594 rc = set_cpus_allowed_ptr(current, cpumask); in cfs_cpt_bind()
|
/linux-4.1.27/arch/s390/include/asm/ |
D | smp.h | 25 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
D | topology.h | 38 const struct cpumask *cpu_coregroup_mask(int cpu);
|
/linux-4.1.27/arch/xtensa/kernel/ |
D | smp.c | 352 static void send_ipi_message(const struct cpumask *callmask, in send_ipi_message() 365 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 382 struct cpumask targets; in smp_send_stop()
|
/linux-4.1.27/arch/score/kernel/ |
D | time.c | 97 score_clockevent.cpumask = cpumask_of(0); in time_init()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | time-ts.c | 232 evt->cpumask = cpumask_of(0); in bfin_gptmr0_clockevent_init() 345 evt->cpumask = cpumask_of(cpu); in bfin_coretmr_clockevent_init()
|
/linux-4.1.27/arch/mips/loongson/loongson-3/ |
D | numa.c | 236 cpumask_clear(&__node_data[(node)]->cpumask); in prom_meminit() 247 cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask); in prom_meminit()
|
/linux-4.1.27/Documentation/filesystems/ |
D | sysfs-pci.txt | 39 local_cpus nearby CPU mask (cpumask, ro) 55 cpumask - file contains a cpumask type
|
/linux-4.1.27/arch/um/kernel/ |
D | time.c | 53 .cpumask = cpu_all_mask,
|
/linux-4.1.27/arch/arm/mach-omap1/ |
D | timer32k.c | 162 clockevent_32k_timer.cpumask = cpumask_of(0); in omap_init_32k_timer()
|
/linux-4.1.27/arch/arm/mach-netx/ |
D | time.c | 147 netx_clockevent.cpumask = cpumask_of(0); in netx_timer_init()
|
/linux-4.1.27/arch/arm/mach-lpc32xx/ |
D | timer.c | 143 lpc32xx_clkevt.cpumask = cpumask_of(0); in lpc32xx_timer_init()
|
/linux-4.1.27/arch/openrisc/kernel/ |
D | time.c | 128 clockevent_openrisc_timer.cpumask = cpumask_of(0); in openrisc_clockevent_init()
|
/linux-4.1.27/arch/mips/sibyte/sb1250/ |
D | smp.c | 74 static inline void sb1250_send_ipi_mask(const struct cpumask *mask, in sb1250_send_ipi_mask()
|
/linux-4.1.27/arch/arm/mach-footbridge/ |
D | dc21285-timer.c | 124 ce->cpumask = cpumask_of(smp_processor_id()); in footbridge_timer_init()
|
/linux-4.1.27/net/iucv/ |
D | iucv.c | 563 cpumask_t cpumask; in iucv_setmask_up() local 567 cpumask_copy(&cpumask, &iucv_irq_cpumask); in iucv_setmask_up() 568 cpumask_clear_cpu(cpumask_first(&iucv_irq_cpumask), &cpumask); in iucv_setmask_up() 569 for_each_cpu(cpu, &cpumask) in iucv_setmask_up() 663 cpumask_t cpumask; in iucv_cpu_notify() local 690 cpumask_copy(&cpumask, &iucv_buffer_cpumask); in iucv_cpu_notify() 691 cpumask_clear_cpu(cpu, &cpumask); in iucv_cpu_notify() 692 if (cpumask_empty(&cpumask)) in iucv_cpu_notify()
|
/linux-4.1.27/arch/mips/alchemy/common/ |
D | time.c | 90 .cpumask = cpu_all_mask,
|
/linux-4.1.27/Documentation/cpuidle/ |
D | driver.txt | 27 const struct cpumask *const coupled_cpus);
|
/linux-4.1.27/arch/avr32/kernel/ |
D | time.c | 145 comparator.cpumask = cpumask_of(0); in time_init()
|
/linux-4.1.27/arch/mips/jz4740/ |
D | time.c | 126 jz4740_clockevent.cpumask = cpumask_of(0); in plat_time_init()
|
/linux-4.1.27/arch/m68k/coldfire/ |
D | pit.c | 154 cf_pit_clockevent.cpumask = cpumask_of(smp_processor_id()); in hw_timer_init()
|
/linux-4.1.27/arch/arm/mach-w90x900/ |
D | time.c | 135 nuc900_clockevent_device.cpumask = cpumask_of(0); in nuc900_clockevents_init()
|
/linux-4.1.27/arch/arm/plat-iop/ |
D | time.c | 159 iop_clockevent.cpumask = cpumask_of(0); in iop_init_time()
|
/linux-4.1.27/arch/mips/paravirt/ |
D | paravirt-smp.c | 74 static void paravirt_send_ipi_mask(const struct cpumask *mask, unsigned int action) in paravirt_send_ipi_mask()
|
/linux-4.1.27/arch/mips/sibyte/bcm1480/ |
D | smp.c | 86 static void bcm1480_send_ipi_mask(const struct cpumask *mask, in bcm1480_send_ipi_mask()
|
/linux-4.1.27/block/ |
D | blk-mq.c | 776 WARN_ON(!cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)); in __blk_mq_run_hw_queue() 888 next_cpu = cpumask_next(hctx->next_cpu, hctx->cpumask); in blk_mq_hctx_next_cpu() 890 next_cpu = cpumask_first(hctx->cpumask); in blk_mq_hctx_next_cpu() 909 if (cpumask_test_cpu(cpu, hctx->cpumask)) { in blk_mq_run_hw_queue() 1645 free_cpumask_var(hctx->cpumask); in blk_mq_free_hw_queues() 1781 cpumask_clear(hctx->cpumask); in blk_mq_map_swqueue() 1794 cpumask_set_cpu(i, hctx->cpumask); in blk_mq_map_swqueue() 1831 hctx->next_cpu = cpumask_first(hctx->cpumask); in blk_mq_map_swqueue() 1953 if (!zalloc_cpumask_var_node(&hctxs[i]->cpumask, GFP_KERNEL, in blk_mq_init_allocated_queue() 2025 free_cpumask_var(hctxs[i]->cpumask); in blk_mq_init_allocated_queue()
|
/linux-4.1.27/drivers/pci/ |
D | pci-sysfs.c | 83 const struct cpumask *mask; in pci_dev_show_local_cpu() 114 const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); in cpuaffinity_show() local 116 return cpumap_print_to_pagebuf(false, buf, cpumask); in cpuaffinity_show() 123 const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev)); in cpulistaffinity_show() local 125 return cpumap_print_to_pagebuf(true, buf, cpumask); in cpulistaffinity_show()
|