Home
last modified time | relevance | path

Searched refs:cpus (Results 1 – 200 of 556) sorted by relevance

123

/linux-4.1.27/tools/perf/util/
Dcpumap.c11 struct cpu_map *cpus; in cpu_map__default_new() local
18 cpus = malloc(sizeof(*cpus) + nr_cpus * sizeof(int)); in cpu_map__default_new()
19 if (cpus != NULL) { in cpu_map__default_new()
22 cpus->map[i] = i; in cpu_map__default_new()
24 cpus->nr = nr_cpus; in cpu_map__default_new()
27 return cpus; in cpu_map__default_new()
33 struct cpu_map *cpus = malloc(sizeof(*cpus) + payload_size); in cpu_map__trim_new() local
35 if (cpus != NULL) { in cpu_map__trim_new()
36 cpus->nr = nr_cpus; in cpu_map__trim_new()
37 memcpy(cpus->map, tmp_cpus, payload_size); in cpu_map__trim_new()
[all …]
Drecord.c60 struct cpu_map *cpus; in perf_probe_api() local
63 cpus = cpu_map__new(NULL); in perf_probe_api()
64 if (!cpus) in perf_probe_api()
66 cpu = cpus->map[0]; in perf_probe_api()
67 cpu_map__delete(cpus); in perf_probe_api()
111 if (evlist->cpus->map[0] < 0) in perf_evlist__config()
216 if (!evlist || cpu_map__empty(evlist->cpus)) { in perf_evlist__can_select_event()
217 struct cpu_map *cpus = cpu_map__new(NULL); in perf_evlist__can_select_event() local
219 cpu = cpus ? cpus->map[0] : 0; in perf_evlist__can_select_event()
220 cpu_map__delete(cpus); in perf_evlist__can_select_event()
[all …]
Devlist.c35 void perf_evlist__init(struct perf_evlist *evlist, struct cpu_map *cpus, in perf_evlist__init() argument
43 perf_evlist__set_maps(evlist, cpus, threads); in perf_evlist__init()
117 cpu_map__delete(evlist->cpus); in perf_evlist__delete()
119 evlist->cpus = NULL; in perf_evlist__delete()
287 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__disable()
306 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable()
325 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__disable_event()
346 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable_event()
386 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable_event_thread()
402 bool per_cpu_mmaps = !cpu_map__empty(evlist->cpus); in perf_evlist__enable_event_idx()
[all …]
Dpython.c366 struct cpu_map *cpus; member
379 pcpus->cpus = cpu_map__new(cpustr); in pyrf_cpu_map__init()
380 if (pcpus->cpus == NULL) in pyrf_cpu_map__init()
387 cpu_map__delete(pcpus->cpus); in pyrf_cpu_map__delete()
395 return pcpus->cpus->nr; in pyrf_cpu_map__length()
402 if (i >= pcpus->cpus->nr) in pyrf_cpu_map__item()
405 return Py_BuildValue("i", pcpus->cpus->map[i]); in pyrf_cpu_map__item()
622 struct cpu_map *cpus = NULL; in pyrf_evsel__open() local
636 cpus = ((struct pyrf_cpu_map *)pcpus)->cpus; in pyrf_evsel__open()
643 if (perf_evsel__open(evsel, cpus, threads) < 0) { in pyrf_evsel__open()
[all …]
Devlist.h52 struct cpu_map *cpus; member
64 void perf_evlist__init(struct perf_evlist *evlist, struct cpu_map *cpus,
147 struct cpu_map *cpus, in perf_evlist__set_maps() argument
150 evlist->cpus = cpus; in perf_evlist__set_maps()
Dtop.c98 top->evlist->cpus->nr > 1 ? "s" : "", in perf_top__header_snprintf()
105 top->evlist->cpus->nr, in perf_top__header_snprintf()
106 top->evlist->cpus->nr > 1 ? "s" : ""); in perf_top__header_snprintf()
Dcpumap.h22 int cpu_map__build_socket_map(struct cpu_map *cpus, struct cpu_map **sockp);
23 int cpu_map__build_core_map(struct cpu_map *cpus, struct cpu_map **corep);
Devsel.h85 struct cpu_map *cpus; member
197 struct cpu_map *cpus);
200 int perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus,
Devsel.c1134 static int __perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus, in __perf_evsel__open() argument
1148 perf_evsel__alloc_fd(evsel, cpus->nr, nthreads) < 0) in __perf_evsel__open()
1180 for (cpu = 0; cpu < cpus->nr; cpu++) { in __perf_evsel__open()
1191 pid, cpus->map[cpu], group_fd, flags); in __perf_evsel__open()
1195 cpus->map[cpu], in __perf_evsel__open()
1295 int cpus[1]; member
1298 .cpus = { -1, },
1309 int perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus, in perf_evsel__open() argument
1312 if (cpus == NULL) { in perf_evsel__open()
1314 cpus = &empty_cpu_map.map; in perf_evsel__open()
[all …]
Dpmu.h23 struct cpu_map *cpus; member
Dsvghelper.h6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
Dpmu.c412 struct cpu_map *cpus; in pmu_cpumask() local
428 cpus = cpu_map__read(file); in pmu_cpumask()
430 return cpus; in pmu_cpumask()
464 pmu->cpus = pmu_cpumask(name); in pmu_lookup()
Dparse-events.c277 char *name, struct cpu_map *cpus) in __add_event() argument
287 evsel->cpus = cpus; in __add_event()
650 evsel = __add_event(list, idx, &attr, NULL, pmu->cpus); in parse_events_add_pmu()
667 pmu->cpus); in parse_events_add_pmu()
/linux-4.1.27/tools/perf/tests/
Dopen-syscall-all-cpus.c10 struct cpu_map *cpus; in test__open_syscall_event_on_all_cpus() local
22 cpus = cpu_map__new(NULL); in test__open_syscall_event_on_all_cpus()
23 if (cpus == NULL) { in test__open_syscall_event_on_all_cpus()
41 if (perf_evsel__open(evsel, cpus, threads) < 0) { in test__open_syscall_event_on_all_cpus()
48 for (cpu = 0; cpu < cpus->nr; ++cpu) { in test__open_syscall_event_on_all_cpus()
56 if (cpus->map[cpu] >= CPU_SETSIZE) { in test__open_syscall_event_on_all_cpus()
57 pr_debug("Ignoring CPU %d\n", cpus->map[cpu]); in test__open_syscall_event_on_all_cpus()
61 CPU_SET(cpus->map[cpu], &cpu_set); in test__open_syscall_event_on_all_cpus()
64 cpus->map[cpu], in test__open_syscall_event_on_all_cpus()
72 CPU_CLR(cpus->map[cpu], &cpu_set); in test__open_syscall_event_on_all_cpus()
[all …]
Dmmap-basic.c23 struct cpu_map *cpus; in test__basic_mmap() local
42 cpus = cpu_map__new(NULL); in test__basic_mmap()
43 if (cpus == NULL) { in test__basic_mmap()
49 CPU_SET(cpus->map[0], &cpu_set); in test__basic_mmap()
53 cpus->map[0], strerror_r(errno, sbuf, sizeof(sbuf))); in test__basic_mmap()
63 perf_evlist__set_maps(evlist, cpus, threads); in test__basic_mmap()
80 if (perf_evsel__open(evsels[i], cpus, threads) < 0) { in test__basic_mmap()
142 cpus = NULL; in test__basic_mmap()
145 cpu_map__delete(cpus); in test__basic_mmap()
Dkeep-tracking.c64 struct cpu_map *cpus = NULL; in test__keep_tracking() local
73 cpus = cpu_map__new(NULL); in test__keep_tracking()
74 CHECK_NOT_NULL__(cpus); in test__keep_tracking()
79 perf_evlist__set_maps(evlist, cpus, threads); in test__keep_tracking()
147 cpu_map__delete(cpus); in test__keep_tracking()
Dperf-time-to-tsc.c49 struct cpu_map *cpus = NULL; in test__perf_time_to_tsc() local
63 cpus = cpu_map__new(NULL); in test__perf_time_to_tsc()
64 CHECK_NOT_NULL__(cpus); in test__perf_time_to_tsc()
69 perf_evlist__set_maps(evlist, cpus, threads); in test__perf_time_to_tsc()
Dtask-exit.c61 evlist->cpus = cpu_map__dummy_new(); in test__task_exit()
63 if (!evlist->cpus || !evlist->threads) { in test__task_exit()
Dsw-clock.c53 evlist->cpus = cpu_map__dummy_new(); in __test__sw_clock_freq()
55 if (!evlist->cpus || !evlist->threads) { in __test__sw_clock_freq()
Dcode-reading.c403 struct cpu_map *cpus = NULL; in do_test_code_reading() local
463 cpus = cpu_map__new(NULL); in do_test_code_reading()
464 if (!cpus) { in do_test_code_reading()
478 perf_evlist__set_maps(evlist, cpus, threads); in do_test_code_reading()
542 cpu_map__delete(cpus); in do_test_code_reading()
Dswitch-tracking.c322 struct cpu_map *cpus = NULL; in test__switch_tracking() local
335 cpus = cpu_map__new(NULL); in test__switch_tracking()
336 if (!cpus) { in test__switch_tracking()
347 perf_evlist__set_maps(evlist, cpus, threads); in test__switch_tracking()
563 cpu_map__delete(cpus); in test__switch_tracking()
DBuild7 perf-y += open-syscall-all-cpus.o
/linux-4.1.27/tools/power/cpupower/utils/
Dcpufreq-info.c250 struct cpufreq_affected_cpus *cpus; in debug_output_one() local
273 cpus = cpufreq_get_related_cpus(cpu); in debug_output_one()
274 if (cpus) { in debug_output_one()
276 while (cpus->next) { in debug_output_one()
277 printf("%d ", cpus->cpu); in debug_output_one()
278 cpus = cpus->next; in debug_output_one()
280 printf("%d\n", cpus->cpu); in debug_output_one()
281 cpufreq_put_related_cpus(cpus); in debug_output_one()
284 cpus = cpufreq_get_affected_cpus(cpu); in debug_output_one()
285 if (cpus) { in debug_output_one()
[all …]
Dcpufreq-set.c295 struct cpufreq_affected_cpus *cpus; in cmd_freq_set() local
301 cpus = cpufreq_get_related_cpus(cpu); in cmd_freq_set()
302 if (!cpus) in cmd_freq_set()
304 while (cpus->next) { in cmd_freq_set()
305 bitmask_setbit(cpus_chosen, cpus->cpu); in cmd_freq_set()
306 cpus = cpus->next; in cmd_freq_set()
308 cpufreq_put_related_cpus(cpus); in cmd_freq_set()
/linux-4.1.27/include/linux/
Dstop_machine.h115 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus);
126 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus);
129 const struct cpumask *cpus);
134 const struct cpumask *cpus) in __stop_machine() argument
145 const struct cpumask *cpus) in stop_machine() argument
147 return __stop_machine(fn, data, cpus); in stop_machine()
151 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument
153 return __stop_machine(fn, data, cpus); in stop_machine_from_inactive_cpu()
Dqcom_scm.h16 extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus);
17 extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus);
Dcpufreq.h63 cpumask_var_t cpus; /* Online CPUs only */ member
140 return cpumask_weight(policy->cpus) > 1; in policy_is_shared()
/linux-4.1.27/tools/perf/python/
Dtwatch.py19 cpus = perf.cpu_map()
25 evsel.open(cpus = cpus, threads = threads);
26 evlist = perf.evlist(cpus, threads)
31 for cpu in cpus:
/linux-4.1.27/block/
Dblk-mq-cpumap.c37 cpumask_var_t cpus; in blk_mq_update_queue_map() local
39 if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) in blk_mq_update_queue_map()
42 cpumask_clear(cpus); in blk_mq_update_queue_map()
47 if (!cpumask_test_cpu(first_sibling, cpus)) in blk_mq_update_queue_map()
49 cpumask_set_cpu(i, cpus); in blk_mq_update_queue_map()
84 free_cpumask_var(cpus); in blk_mq_update_queue_map()
/linux-4.1.27/arch/mips/cavium-octeon/
Dsmp.c103 int cpus; in octeon_smp_setup() local
120 cpus = 1; in octeon_smp_setup()
123 set_cpu_possible(cpus, true); in octeon_smp_setup()
124 set_cpu_present(cpus, true); in octeon_smp_setup()
125 __cpu_number_map[id] = cpus; in octeon_smp_setup()
126 __cpu_logical_map[cpus] = id; in octeon_smp_setup()
127 cpus++; in octeon_smp_setup()
140 set_cpu_possible(cpus, true); in octeon_smp_setup()
141 __cpu_number_map[id] = cpus; in octeon_smp_setup()
142 __cpu_logical_map[cpus] = id; in octeon_smp_setup()
[all …]
/linux-4.1.27/arch/arm/mach-shmobile/
Dtimer.c40 struct device_node *np, *cpus; in shmobile_init_delay() local
46 cpus = of_find_node_by_path("/cpus"); in shmobile_init_delay()
47 if (!cpus) in shmobile_init_delay()
50 for_each_child_of_node(cpus, np) { in shmobile_init_delay()
68 of_node_put(cpus); in shmobile_init_delay()
Dpm-rcar-gen2.c58 struct device_node *np, *cpus; in rcar_gen2_pm_init() local
67 cpus = of_find_node_by_path("/cpus"); in rcar_gen2_pm_init()
68 if (!cpus) in rcar_gen2_pm_init()
71 for_each_child_of_node(cpus, np) { in rcar_gen2_pm_init()
Dplatsmp-apmu.c99 for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) { in apmu_parse_cfg()
100 id = apmu_config[k].cpus[bit]; in apmu_parse_cfg()
109 for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) { in apmu_parse_cfg()
110 id = apmu_config[k].cpus[bit]; in apmu_parse_cfg()
Dsmp-r8a7790.c42 .cpus = { 0, 1, 2, 3 },
46 .cpus = { 0x100, 0x0101, 0x102, 0x103 },
Dplatsmp-apmu.h21 int cpus[4]; member
Dsmp-r8a7791.c31 .cpus = { 0, 1 },
/linux-4.1.27/arch/x86/kernel/
Dtsc_sync.c125 int cpus = 2; in check_tsc_sync_source() local
149 while (atomic_read(&start_count) != cpus-1) in check_tsc_sync_source()
158 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source()
191 int cpus = 2; in check_tsc_sync_target() local
201 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target()
214 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target()
/linux-4.1.27/tools/power/cpupower/utils/helpers/
Dtopology.c67 int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF); in get_cpu_topology() local
69 cpu_top->core_info = malloc(sizeof(struct cpuid_core_info) * cpus); in get_cpu_topology()
73 for (cpu = 0; cpu < cpus; cpu++) { in get_cpu_topology()
88 qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info), in get_cpu_topology()
95 for(cpu = 1; cpu < cpus; cpu++) { in get_cpu_topology()
110 return cpus; in get_cpu_topology()
/linux-4.1.27/drivers/irqchip/
Dirq-bcm7038-l1.c46 struct bcm7038_l1_cpu *cpus[NR_CPUS]; member
127 cpu = intc->cpus[cpu_logical_map(smp_processor_id())]; in bcm7038_l1_irq_handle()
129 cpu = intc->cpus[0]; in bcm7038_l1_irq_handle()
159 intc->cpus[cpu_idx]->mask_cache[word] &= ~mask; in __bcm7038_l1_unmask()
160 l1_writel(mask, intc->cpus[cpu_idx]->map_base + in __bcm7038_l1_unmask()
170 intc->cpus[cpu_idx]->mask_cache[word] |= mask; in __bcm7038_l1_mask()
171 l1_writel(mask, intc->cpus[cpu_idx]->map_base + in __bcm7038_l1_mask()
209 was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] & in bcm7038_l1_set_affinity()
241 cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), in bcm7038_l1_init_one()
316 intc->cpus[0]->map_base, IRQS_PER_WORD * intc->n_words); in bcm7038_l1_of_init()
[all …]
/linux-4.1.27/drivers/lguest/
Dlguest_user.c119 cpu = &lg->cpus[cpu_id]; in read()
158 if (id >= ARRAY_SIZE(cpu->lg->cpus)) in lg_cpu_start()
163 cpu->lg = container_of(cpu, struct lguest, cpus[id]); in lg_cpu_start()
256 err = lg_cpu_start(&lg->cpus[0], 0, args[2]); in initialize()
278 free_page(lg->cpus[0].regs_page); in initialize()
319 cpu = &lg->cpus[cpu_id]; in write()
378 hrtimer_cancel(&lg->cpus[i].hrt); in close()
380 free_page(lg->cpus[i].regs_page); in close()
385 mmput(lg->cpus[i].mm); in close()
Dpage_tables.c1035 kill_guest(&lg->cpus[0], "Attempt to set pgd %u/%u", in guest_set_pgd()
1046 if (!allocate_switcher_mapping(&lg->cpus[0])) { in guest_set_pgd()
1047 kill_guest(&lg->cpus[0], in guest_set_pgd()
1058 guest_pagetable_clear_all(&lg->cpus[0]); in guest_set_pmd()
1075 struct lg_cpu *cpu = &lg->cpus[0]; in init_guest_pagetable()
/linux-4.1.27/tools/lib/traceevent/
Dplugin_function.c32 static int cpus = -1; variable
95 if (cpu > cpus) { in add_and_get_index()
107 for (i = cpus + 1; i <= cpu; i++) in add_and_get_index()
109 cpus = cpu; in add_and_get_index()
183 for (i = 0; i <= cpus; i++) { in PEVENT_PLUGIN_UNLOADER()
193 cpus = -1; in PEVENT_PLUGIN_UNLOADER()
Devent-parse.h474 int cpus; member
709 return pevent->cpus; in pevent_get_cpus()
712 static inline void pevent_set_cpus(struct pevent *pevent, int cpus) in pevent_set_cpus() argument
714 pevent->cpus = cpus; in pevent_set_cpus()
/linux-4.1.27/arch/arm/kernel/
Ddevtree.c77 struct device_node *cpu, *cpus; in arm_dt_init_cpu_maps() local
84 cpus = of_find_node_by_path("/cpus"); in arm_dt_init_cpu_maps()
86 if (!cpus) in arm_dt_init_cpu_maps()
89 for_each_child_of_node(cpus, cpu) { in arm_dt_init_cpu_maps()
160 set_smp_ops_by_method(cpus); in arm_dt_init_cpu_maps()
/linux-4.1.27/drivers/firmware/
Dqcom_scm.c415 int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) in qcom_scm_set_cold_boot_addr() argument
426 if (!cpus || (cpus && cpumask_empty(cpus))) in qcom_scm_set_cold_boot_addr()
429 for_each_cpu(cpu, cpus) { in qcom_scm_set_cold_boot_addr()
448 int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) in qcom_scm_set_warm_boot_addr() argument
458 for_each_cpu(cpu, cpus) { in qcom_scm_set_warm_boot_addr()
470 for_each_cpu(cpu, cpus) in qcom_scm_set_warm_boot_addr()
/linux-4.1.27/arch/arm/common/
Dmcpm_entry.c346 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; in __mcpm_cpu_going_down()
347 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_going_down()
360 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; in __mcpm_cpu_down()
361 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_down()
413 sync_cache_r(&c->cpus); in __mcpm_outbound_enter_critical()
421 cpustate = c->cpus[i].cpu; in __mcpm_outbound_enter_critical()
426 sync_cache_r(&c->cpus[i].cpu); in __mcpm_outbound_enter_critical()
469 mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN; in mcpm_sync_init()
475 mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP; in mcpm_sync_init()
/linux-4.1.27/tools/testing/selftests/rcutorture/bin/
Dkvm.sh48 cpus=0
96 --cpus)
97 checkarg --cpus "(number)" "$#" "$2" '^[0-9]*$' '^--'
98 cpus=$2
194 awk < $T/cfgcpu.sort > $T/cfgcpu.pack -v ncpus=$cpus '
281 -v ncpus=$cpus \
/linux-4.1.27/tools/virtio/virtio-trace/
Dtrace-agent.c62 s->cpus = get_total_cpus(); in agent_info_new()
66 for (i = 0; i < s->cpus; i++) in agent_info_new()
159 for (cpu = 0; cpu < s->cpus; cpu++) { in agent_info_init()
226 for (cpu = 0; cpu < s->cpus; cpu++) in agent_main_loop()
232 for (cpu = 0; cpu < s->cpus; cpu++) { in agent_main_loop()
248 for (i = 0; i < s->cpus; i++) { in agent_info_free()
Dtrace-agent.h20 int cpus; member
/linux-4.1.27/arch/ia64/kernel/
Dsmp.c295 cpumask_var_t cpus; in smp_flush_tlb_mm() local
304 if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) { in smp_flush_tlb_mm()
308 cpumask_copy(cpus, mm_cpumask(mm)); in smp_flush_tlb_mm()
309 smp_call_function_many(cpus, in smp_flush_tlb_mm()
311 free_cpumask_var(cpus); in smp_flush_tlb_mm()
/linux-4.1.27/arch/powerpc/platforms/powermac/
Dsmp.c635 struct device_node *cpus; in smp_core99_pfunc_tb_freeze() local
638 cpus = of_find_node_by_path("/cpus"); in smp_core99_pfunc_tb_freeze()
639 BUG_ON(cpus == NULL); in smp_core99_pfunc_tb_freeze()
642 pmf_call_function(cpus, "cpu-timebase", &args); in smp_core99_pfunc_tb_freeze()
643 of_node_put(cpus); in smp_core99_pfunc_tb_freeze()
713 struct device_node *cpus = in smp_core99_setup() local
715 if (cpus && in smp_core99_setup()
716 of_get_property(cpus, "platform-cpu-timebase", NULL)) { in smp_core99_setup()
769 struct device_node *cpus; in smp_core99_probe() local
775 for (cpus = NULL; (cpus = of_find_node_by_type(cpus, "cpu")) != NULL;) in smp_core99_probe()
Dfeature.c1052 struct device_node *cpus; in core99_reset_cpu() local
1062 cpus = of_find_node_by_path("/cpus"); in core99_reset_cpu()
1063 if (cpus == NULL) in core99_reset_cpu()
1065 for (np = cpus->child; np != NULL; np = np->sibling) { in core99_reset_cpu()
1075 of_node_put(cpus); in core99_reset_cpu()
1507 struct device_node *cpus; in g5_reset_cpu() local
1513 cpus = of_find_node_by_path("/cpus"); in g5_reset_cpu()
1514 if (cpus == NULL) in g5_reset_cpu()
1516 for (np = cpus->child; np != NULL; np = np->sibling) { in g5_reset_cpu()
1526 of_node_put(cpus); in g5_reset_cpu()
[all …]
/linux-4.1.27/arch/powerpc/kernel/
Drtas.c835 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument
837 if (!cpumask_empty(cpus)) { in rtas_cpu_state_change_mask()
838 cpumask_clear(cpus); in rtas_cpu_state_change_mask()
848 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument
854 if (cpumask_empty(cpus)) in rtas_cpu_state_change_mask()
857 for_each_cpu(cpu, cpus) { in rtas_cpu_state_change_mask()
875 cpumask_shift_right(cpus, cpus, cpu); in rtas_cpu_state_change_mask()
876 cpumask_shift_left(cpus, cpus, cpu); in rtas_cpu_state_change_mask()
880 cpumask_clear_cpu(cpu, cpus); in rtas_cpu_state_change_mask()
889 int rtas_online_cpus_mask(cpumask_var_t cpus) in rtas_online_cpus_mask() argument
[all …]
/linux-4.1.27/drivers/cpuidle/
Dcoupled.c429 cpumask_t cpus; in cpuidle_coupled_any_pokes_pending() local
432 cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); in cpuidle_coupled_any_pokes_pending()
433 ret = cpumask_and(&cpus, &cpuidle_coupled_poke_pending, &cpus); in cpuidle_coupled_any_pokes_pending()
616 cpumask_t cpus; in cpuidle_coupled_update_online_cpus() local
617 cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); in cpuidle_coupled_update_online_cpus()
618 coupled->online_count = cpumask_weight(&cpus); in cpuidle_coupled_update_online_cpus()
/linux-4.1.27/drivers/misc/sgi-gru/
Dgrutlbpurge.c356 int cpus, shift = 0, n; in gru_tgh_flush_init() local
358 cpus = uv_blade_nr_possible_cpus(gru->gs_blade_id); in gru_tgh_flush_init()
361 if (cpus) { in gru_tgh_flush_init()
362 n = 1 << fls(cpus - 1); in gru_tgh_flush_init()
375 gru->gs_tgh_first_remote = (cpus + (1 << shift) - 1) >> shift; in gru_tgh_flush_init()
Dgrulib.h146 int cpus; member
/linux-4.1.27/Documentation/devicetree/bindings/cpufreq/
Darm_big_little_dt.txt8 under node /cpus/cpu@x. Where x is the first cpu inside a cluster.
10 FIXME: Cpus should boot in the order specified in DT and all cpus for a cluster
24 cpus {
Dcpufreq-spear.txt17 /cpus/cpu@0.
21 cpus {
Dcpufreq-dt.txt8 under node /cpus/cpu@0.
27 cpus {
/linux-4.1.27/Documentation/devicetree/bindings/arm/
Dcpus.txt6 the "cpus" node, which in turn contains a number of subnodes (ie "cpu")
26 cpus and cpu node bindings definition
29 The ARM architecture, in accordance with the ePAPR, requires the cpus and cpu
32 - cpus node
36 The node name must be "cpus".
38 A cpus node must define the following properties:
102 * If cpus node's #address-cells property is set to 2
110 * If cpus node's #address-cells property is set to 1
243 cpus {
274 cpus {
[all …]
Dkirkwood.txt12 cpus/cpu@0 with three clocks, "cpu_clk", "ddrclk" and "powersave",
18 cpus {
Dtopology.txt47 child of the cpus node and provides a container where the actual topology
62 The cpu-map node's parent node must be the cpus node.
169 cpus {
388 cpus {
475 Documentation/devicetree/bindings/arm/cpus.txt
Dvexpress.txt56 Top-level standard "cpus" node is required. It must contain a node
59 cpus {
184 cpus {
Dal,alpine.txt25 cpus {
/linux-4.1.27/drivers/cpufreq/
Dcpufreq.c184 cpumask_setall(policy->cpus); in cpufreq_generic_init()
339 for_each_cpu(freqs->cpu, policy->cpus) in cpufreq_notify_transition()
685 return cpufreq_show_cpus(policy->cpus, buf); in show_affected_cpus()
882 for_each_cpu(j, policy->cpus) { in cpufreq_add_dev_symlink()
980 cpumask_set_cpu(cpu, policy->cpus); in cpufreq_add_policy_cpu()
1025 if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL)) in cpufreq_policy_alloc()
1041 free_cpumask_var(policy->cpus); in cpufreq_policy_alloc()
1075 free_cpumask_var(policy->cpus); in cpufreq_policy_free()
1158 cpumask_copy(policy->cpus, cpumask_of(cpu)); in __cpufreq_add_dev()
1172 cpumask_or(policy->related_cpus, policy->related_cpus, policy->cpus); in __cpufreq_add_dev()
[all …]
Dspeedstep-ich.c262 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_target()
295 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in speedstep_cpu_init()
297 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_cpu_init()
Dcpufreq_governor.c66 for_each_cpu(j, policy->cpus) { in dbs_check_cpu()
190 for_each_cpu(i, policy->cpus) in gov_queue_work()
205 for_each_cpu(i, policy->cpus) { in gov_cancel_work()
372 for_each_cpu(j, policy->cpus) { in cpufreq_governor_dbs()
Dp4-clockmod.c116 for_each_cpu(i, policy->cpus) in cpufreq_p4_target()
175 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in cpufreq_p4_cpu_init()
Dacpi-cpufreq.c460 cmd.mask = policy->cpus; in acpi_cpufreq_target()
694 cpumask_copy(policy->cpus, perf->shared_cpu_map); in acpi_cpufreq_cpu_init()
702 cpumask_copy(policy->cpus, cpu_core_mask(cpu)); in acpi_cpufreq_cpu_init()
706 cpumask_clear(policy->cpus); in acpi_cpufreq_cpu_init()
707 cpumask_set_cpu(cpu, policy->cpus); in acpi_cpufreq_cpu_init()
Dpowernv-cpufreq.c362 smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1); in powernv_cpufreq_target_index()
374 cpumask_set_cpu(base + i, policy->cpus); in powernv_cpufreq_cpu_init()
Dppc_cbe_cpufreq.c121 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in cbe_cpufreq_cpu_init()
Dspeedstep-centrino.c439 for_each_cpu(j, policy->cpus) { in centrino_target()
447 good_cpu = cpumask_any_and(policy->cpus, in centrino_target()
Dpowernow-k8.c1093 cpumask_copy(pol->cpus, cpu_core_mask(pol->cpu)); in powernowk8_cpu_init()
1094 data->available_cores = pol->cpus; in powernowk8_cpu_init()
1109 for_each_cpu(cpu, pol->cpus) in powernowk8_cpu_init()
1134 for_each_cpu(cpu, pol->cpus) in powernowk8_cpu_exit()
/linux-4.1.27/Documentation/ABI/testing/
Dsysfs-devices-system-xen_cpu16 Interface to online/offline Xen physical cpus
19 to online/offline physical cpus, except cpu0 due to several
Dsysfs-devices-system-cpu25 offline: cpus that are not online because they have been
26 HOTPLUGGED off or exceed the limit of cpus allowed by the
29 online: cpus that are online and being scheduled.
31 possible: cpus that have been allocated resources and can be
34 present: cpus that have been identified as being present in
253 shared_cpu_list: the list of logical cpus sharing the cache
255 shared_cpu_map: logical cpu mask containing the list of cpus sharing
/linux-4.1.27/samples/trace_events/
Dtrace-events-sample.h240 __bitmask( cpus, num_possible_cpus() )
249 __assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus());
296 __get_str(str), __get_bitmask(cpus))
/linux-4.1.27/kernel/
Dstop_machine.c550 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in __stop_machine() argument
556 .active_cpus = cpus, in __stop_machine()
583 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in stop_machine() argument
589 ret = __stop_machine(fn, data, cpus); in stop_machine()
618 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument
621 .active_cpus = cpus }; in stop_machine_from_inactive_cpu()
Dsmp.c669 cpumask_var_t cpus; in on_each_cpu_cond() local
674 if (likely(zalloc_cpumask_var(&cpus, (gfp_flags|__GFP_NOWARN)))) { in on_each_cpu_cond()
678 cpumask_set_cpu(cpu, cpus); in on_each_cpu_cond()
679 on_each_cpu_mask(cpus, func, info, wait); in on_each_cpu_cond()
681 free_cpumask_var(cpus); in on_each_cpu_cond()
/linux-4.1.27/Documentation/devicetree/bindings/arm/cpu-enable-method/
Dmarvell,berlin-smp7 be defined in the "cpus" node.
20 cpus {
Dal,alpine-smp8 "cpus" node.
21 cpus {
/linux-4.1.27/Documentation/devicetree/bindings/mips/img/
Dpistachio.txt10 A "cpus" node is required. Required properties:
22 cpus {
/linux-4.1.27/arch/s390/hypfs/
Dhypfs_diag.c120 __u8 cpus; member
127 __u8 cpus; member
154 return ((struct part_hdr *)hdr)->cpus; in part_hdr__rcpus()
256 __u8 cpus; member
263 __u8 cpus; member
280 return ((struct phys_hdr *)hdr)->cpus; in phys_hdr__cpus()
282 return ((struct x_phys_hdr *)hdr)->cpus; in phys_hdr__cpus()
/linux-4.1.27/Documentation/devicetree/bindings/arm/bcm/
Dbrcm,bcm11351-cpu-method.txt8 properties in the "cpus" device tree node:
19 cpus {
/linux-4.1.27/arch/powerpc/platforms/cell/
Dcpufreq_spudemand.c112 for_each_cpu(i, policy->cpus) { in spu_gov_govern()
129 for_each_cpu (i, policy->cpus) { in spu_gov_govern()
/linux-4.1.27/Documentation/ia64/
Dmca.txt7 the OS is in any state. Including when one of the cpus is already
22 to all the other cpus, the slaves.
24 * Slave cpus that receive the MCA interrupt call down into SAL, they
29 sends an unmaskable INIT event to the slave cpus that have not
68 tasks. But (and its a big but), the cpus that received the MCA
97 cpus. SAL picks one of the cpus as the monarch and the rest are
103 versions of SAL out there. Some drive all the cpus as monarchs. Some
126 all cpus.
/linux-4.1.27/Documentation/accounting/
Dtaskstats.txt36 one of the cpus in the cpumask, its per-pid statistics are sent to the
82 Commands to register/deregister interest in exit data from a set of cpus
86 comma-separated cpu ranges e.g. to listen to exit data from cpus 1,2,3,5,7,8
88 in cpus before closing the listening socket, the kernel cleans up its interest
165 extended and the number of cpus grows large.
172 - create more listeners and reduce the number of cpus being listened to by
175 of cpus to which it listens, especially if they are listening to just one cpu.
/linux-4.1.27/Documentation/thermal/
Dcpu-cooling-api.txt26 clip_cpus: cpumask of cpus where the frequency constraints will happen.
37 clip_cpus: cpumask of cpus where the frequency constraints will happen.
/linux-4.1.27/arch/ia64/mm/
Ddiscontig.c119 unsigned long pernodesize = 0, cpus; in compute_pernodesize() local
121 cpus = early_nr_cpus_node(node); in compute_pernodesize()
122 pernodesize += PERCPU_PAGE_SIZE * cpus; in compute_pernodesize()
270 int cpus = early_nr_cpus_node(node); in fill_pernode() local
278 pernode += PERCPU_PAGE_SIZE * cpus; in fill_pernode()
/linux-4.1.27/arch/powerpc/boot/dts/
Dmpc8536ds.dts18 cpus {
19 #cpus = <1>;
Dmpc8536ds_36b.dts18 cpus {
19 #cpus = <1>;
Diss4xx.dts22 dcr-parent = <&{/cpus/cpu@0}>;
28 cpus {
Damigaone.dts21 cpus {
22 #cpus = <1>;
Diss4xx-mpic.dts24 dcr-parent = <&{/cpus/cpu@0}>;
30 cpus {
Dsbc8548-pre.dtsi29 cpus {
Dp1020rdb-pc_camp_core0.dts33 cpus {
Dacadia.dts18 dcr-parent = <&{/cpus/cpu@0}>;
26 cpus {
Dep405.dts19 dcr-parent = <&{/cpus/cpu@0}>;
27 cpus {
Dmpc8572ds_camp_core0.dts23 cpus {
Dklondike.dts31 dcr-parent = <&{/cpus/cpu@0}>;
38 cpus {
Dwalnut.dts19 dcr-parent = <&{/cpus/cpu@0}>;
27 cpus {
Dgamecube.dts32 cpus {
Dwarp.dts19 dcr-parent = <&{/cpus/cpu@0}>;
26 cpus {
Dps3.dts54 cpus {
Dmpc8572ds_camp_core1.dts24 cpus {
Dp1020rdb-pc_camp_core1.dts29 cpus {
Dhotfoot.dts18 dcr-parent = <&{/cpus/cpu@0}>;
27 cpus {
/linux-4.1.27/Documentation/cgroups/
Dcpusets.txt28 2.2 Adding/removing cpus
171 - cpuset.cpus: list of CPUs in that cpuset
218 The cpus and mems files in the root (top_cpuset) cpuset are
219 read-only. The cpus file automatically tracks the value of
410 setting), it requests that all the CPUs in that cpusets allowed 'cpuset.cpus'
460 If two cpusets have partially overlapping 'cpuset.cpus' allowed, and only
560 3 : search cpus in a node [= system wide on non-NUMA system]
616 If a cpuset has its 'cpuset.cpus' modified, then each task in that cpuset
650 with non-empty cpus. But the moving of some (or all) tasks might fail if
687 /bin/echo 2-3 > cpuset.cpus
[all …]
/linux-4.1.27/scripts/gdb/
Dvmlinux-gdb.py30 import linux.cpus
/linux-4.1.27/Documentation/
Dcpu-hotplug.txt45 maxcpus=n Restrict boot time cpus to n. Say if you have 4 cpus, using
47 other cpus later online, read FAQ's for more info.
49 additional_cpus=n (*) Use this to limit hotpluggable cpus. This option sets
62 determine the number of potentially hot-pluggable cpus. The implementation
63 should only rely on this to count the # of cpus, but *MUST* not rely
65 BIOS doesn't mark such hot-pluggable cpus as disabled entries, one could
66 use this parameter "additional_cpus=x" to represent those cpus in the
69 possible_cpus=n [s390,x86_64] use this to set hotpluggable cpus.
118 If you merely need to avoid cpus going away, you could also use
177 Q: Does hot-add/hot-remove refer to physical add/remove of cpus?
DIRQ-affinity.txt11 IRQ affinity then the value will not change from the default of all cpus.
58 Here is an example of limiting that same irq (44) to cpus 1024 to 1031:
Dcputopology.txt75 above). [~cpu_online_mask + cpus >= NR_CPUS]
88 In this example, there are 64 CPUs in the system but cpus 32-63 exceed
/linux-4.1.27/Documentation/devicetree/bindings/mips/brcm/
Dbrcm,bmips.txt8 under the "cpus" node.
/linux-4.1.27/arch/mips/boot/dts/mti/
Dsead3.dts12 cpus {
/linux-4.1.27/arch/arm/boot/dts/
Dbcm47081.dtsi15 cpus {
Dbcm4708.dtsi15 cpus {
Ddra72x.dtsi15 cpus {
Daxm5516-amarillo.dts17 #include "axm5516-cpus.dtsi"
Dhip01-ca9x2.dts25 cpus {
Dstih41x.dtsi12 cpus {
Daxm5516-cpus.dtsi2 * arch/arm/boot/dts/axm5516-cpus.dtsi
13 cpus {
Dalphascale-asm9260.dtsi18 cpus {
Dvf500.dtsi15 cpus {
Dhisi-x5hd2-dkb.dts21 cpus {
Domap34xx.dtsi16 cpus {
Domap443x.dtsi14 cpus {
Dexynos4212.dtsi25 cpus {
Ddra74x.dtsi15 cpus {
Dxenvm-4.2.dts23 cpus {
Dexynos4412.dtsi25 cpus {
Domap4460.dtsi13 cpus {
Dk2l.dtsi12 cpus {
Dk2hk.dtsi12 cpus {
Dpxa2xx.dtsi25 cpus {
Dimx6dl.dtsi20 cpus {
Domap36xx.dtsi20 cpus {
Dstih418.dtsi13 cpus {
Ds3c2416.dtsi23 cpus {
Decx-2000.dts29 cpus {
Dbcm63138.dtsi20 cpus {
Dmoxart.dtsi15 cpus {
Dbcm7445.dtsi16 cpus {
Domap3-cm-t3x30.dtsi8 cpus {
Dk2e.dtsi12 cpus {
Dvt8500.dtsi14 cpus {
/linux-4.1.27/arch/cris/boot/dts/
Detraxfs.dtsi6 cpus {
/linux-4.1.27/arch/arc/boot/dts/
Dskeleton.dtsi22 cpus {
/linux-4.1.27/arch/mips/boot/dts/ralink/
Drt2880.dtsi6 cpus {
Drt3883.dtsi6 cpus {
Dmt7620a.dtsi6 cpus {
Drt3050.dtsi6 cpus {
/linux-4.1.27/arch/openrisc/boot/dts/
Dor1ksim.dts17 cpus {
/linux-4.1.27/arch/powerpc/include/asm/
Drtas.h356 extern int rtas_online_cpus_mask(cpumask_var_t cpus);
357 extern int rtas_offline_cpus_mask(cpumask_var_t cpus);
/linux-4.1.27/Documentation/devicetree/bindings/arm/msm/
Dqcom,idle-state.txt15 clocks. In addition to gating the clocks, QCOM cpus use this instruction as a
18 hierarchy to enter standby states, when all cpus are idle. An interrupt brings
51 voltages reduced, provided all cpus enter this state. Since the span of low
/linux-4.1.27/Documentation/cpu-freq/
Dcpu-drivers.txt125 SMP systems normally have same clock source for a group of cpus. For these the
127 routine must initialize policy->cpus with mask of all possible cpus (Online +
129 policy->related_cpus and will reset policy->cpus to carry only online cpus.
/linux-4.1.27/kernel/debug/kdb/
Dkdb_cmds26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
/linux-4.1.27/drivers/crypto/qat/qat_dh895xcc/
Dadf_isr.c117 unsigned int cpu, cpus = num_online_cpus(); in adf_request_irqs() local
131 cpu = ((accel_dev->accel_id * hw_data->num_banks) + i) % cpus; in adf_request_irqs()
Dadf_drv.c122 int cpus = num_online_cpus(); in adf_dev_configure() local
124 int instances = min(cpus, banks); in adf_dev_configure()
/linux-4.1.27/tools/perf/
Dbuiltin-stat.c168 return (evsel->cpus && !target.cpu_list) ? evsel->cpus : evsel_list->cpus; in perf_evsel__cpus()
401 struct cpu_map *cpus = perf_evsel__cpus(counter); in check_per_pkg() local
409 if (cpu_map__empty(cpus)) in check_per_pkg()
420 s = cpu_map__get_socket(cpus, cpu); in check_per_pkg()
603 const int ncpus = cpu_map__nr(evsel_list->cpus), in handle_initial_delay()
1517 if (cpu_map__build_socket_map(evsel_list->cpus, &aggr_map)) { in perf_stat_init_aggr_mode()
1524 if (cpu_map__build_core_map(evsel_list->cpus, &aggr_map)) { in perf_stat_init_aggr_mode()
/linux-4.1.27/scripts/dtc/
Dlivetree.c553 struct node *cpus, *bootcpu; in guess_boot_cpuid() local
556 cpus = get_node_by_path(tree, "/cpus"); in guess_boot_cpuid()
557 if (!cpus) in guess_boot_cpuid()
561 bootcpu = cpus->children; in guess_boot_cpuid()
/linux-4.1.27/Documentation/scheduler/
Dsched-stats.txt16 cpus on the machine, while domain0 is the most tightly focused domain,
17 sometimes balancing only between pairs of cpus. At this time, there
19 field in the domain stats is a bit map indicating which cpus are affected
65 The first field is a bit mask indicating what cpus this domain operates over.
/linux-4.1.27/drivers/bus/
Darm-cci.c127 cpumask_t cpus; member
758 cpu = cpumask_first(&cci_pmu->cpus); in cci_pmu_event_init()
786 cpumask_pr_args(&pmu->cpus)); in pmu_attr_cpumask_show()
839 if (!cpumask_test_and_clear_cpu(cpu, &pmu->cpus)) in cci_pmu_cpu_notifier()
848 cpumask_set_cpu(target, &pmu->cpus); in cci_pmu_cpu_notifier()
986 cpumask_set_cpu(smp_processor_id(), &pmu->cpus); in cci_pmu_probe()
/linux-4.1.27/arch/mips/boot/dts/brcm/
Dbcm6328.dtsi6 cpus {
Dbcm6368.dtsi6 cpus {
Dbcm3384_viper.dtsi14 cpus {
Dbcm3384_zephyr.dtsi13 cpus {
Dbcm7125.dtsi6 cpus {
Dbcm7360.dtsi6 cpus {
Dbcm7358.dtsi6 cpus {
/linux-4.1.27/arch/tile/kernel/
Dsetup.c887 int cpu, node, cpus, i, x, y; in setup_numa_mapping() local
923 cpus = cpu; in setup_numa_mapping()
937 for (cpu = 0; cpu < cpus; ++cpu) { in setup_numa_mapping()
1327 int cpus, i, rc; in setup_cpu_maps() local
1349 cpus = 1; /* this cpu */ in setup_cpu_maps()
1351 for (i = 0; cpus < setup_max_cpus; ++i) in setup_cpu_maps()
1353 ++cpus; in setup_cpu_maps()
/linux-4.1.27/arch/x86/include/asm/
Damd_nb.h46 atomic_t cpus; member
/linux-4.1.27/arch/arm/include/asm/
Dmcpm.h264 } cpus[MAX_CPUS_PER_CLUSTER]; member
/linux-4.1.27/drivers/edac/
Dcpc925_edac.c597 struct device_node *cpus; in cpc925_cpu_mask_disabled() local
607 cpus = of_find_node_by_path("/cpus"); in cpc925_cpu_mask_disabled()
608 if (cpus == NULL) { in cpc925_cpu_mask_disabled()
613 while ((cpunode = of_get_next_child(cpus, cpunode)) != NULL) { in cpc925_cpu_mask_disabled()
638 of_node_put(cpus); in cpc925_cpu_mask_disabled()
/linux-4.1.27/scripts/gdb/linux/
Dmodules.py16 from linux import cpus, utils
/linux-4.1.27/arch/c6x/boot/dts/
Dtms320c6457.dtsi6 cpus {
Dtms320c6474.dtsi6 cpus {
Dtms320c6455.dtsi6 cpus {
Dtms320c6472.dtsi6 cpus {
Dtms320c6678.dtsi6 cpus {
/linux-4.1.27/arch/arm/plat-samsung/include/plat/
Dcpu.h109 struct cpu_table *cpus, unsigned int cputab_size);
/linux-4.1.27/tools/power/x86/turbostat/
Dturbostat.c2740 } *cpus; in topology_probe() local
2752 cpus = calloc(1, (topo.max_cpu_num + 1) * sizeof(struct cpu_topology)); in topology_probe()
2753 if (cpus == NULL) in topology_probe()
2788 cpus[i].core_id = get_core_id(i); in topology_probe()
2789 if (cpus[i].core_id > max_core_id) in topology_probe()
2790 max_core_id = cpus[i].core_id; in topology_probe()
2792 cpus[i].physical_package_id = get_physical_package_id(i); in topology_probe()
2793 if (cpus[i].physical_package_id > max_package_id) in topology_probe()
2794 max_package_id = cpus[i].physical_package_id; in topology_probe()
2801 i, cpus[i].physical_package_id, cpus[i].core_id); in topology_probe()
[all …]
/linux-4.1.27/arch/s390/kernel/
Dtime.c712 atomic_t cpus; member
721 atomic_dec(&sync->cpus); in clock_sync_cpu()
770 while (atomic_read(&etr_sync->cpus) != 0) in etr_sync_clock()
833 atomic_set(&etr_sync.cpus, num_online_cpus() - 1); in etr_sync_clock_stop()
1542 while (atomic_read(&stp_sync->cpus) != 0) in stp_sync_clock()
1603 atomic_set(&stp_sync.cpus, num_online_cpus() - 1); in stp_work_fn()
Dcrash_dump.c524 int i, cpus = 0; in get_cpu_cnt() local
529 cpus++; in get_cpu_cnt()
531 return cpus; in get_cpu_cnt()
/linux-4.1.27/arch/mips/boot/dts/lantiq/
Ddanube.dtsi6 cpus {
/linux-4.1.27/Documentation/power/
Dsuspend-and-cpuhotplug.txt22 |tasks | | cpus | | | | cpus | |tasks|
70 Note down these cpus in | P
93 | Call _cpu_up() [for all those cpus in the frozen_cpus mask, in a loop]
/linux-4.1.27/kernel/power/
Dpower.h11 int cpus; member
/linux-4.1.27/arch/mn10300/kernel/
Dsmp.c395 int cpus, ret = 0; in smp_nmi_call_function() local
397 cpus = num_online_cpus() - 1; in smp_nmi_call_function()
398 if (cpus < 1) in smp_nmi_call_function()
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/
Dmce_amd.c596 atomic_inc(&b->cpus); in threshold_create_bank()
619 atomic_set(&b->cpus, 1); in threshold_create_bank()
708 if (!atomic_dec_and_test(&b->cpus)) { in threshold_remove_bank()
/linux-4.1.27/include/trace/events/
Dxen.h407 TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm,
409 TP_ARGS(cpus, mm, addr, end),
416 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
/linux-4.1.27/arch/arm64/boot/dts/sprd/
Dsc9836.dtsi15 cpus {
/linux-4.1.27/arch/xtensa/boot/dts/
Dxtfpga.dtsi16 cpus {
/linux-4.1.27/arch/arm64/boot/dts/arm/
Drtsm_ve-aemv8a.dts30 cpus {
/linux-4.1.27/kernel/sched/
Dfair.c146 unsigned int cpus = min_t(int, num_online_cpus(), 8); in get_update_sysctl_factor() local
154 factor = cpus; in get_update_sysctl_factor()
158 factor = 1 + ilog2(cpus); in get_update_sysctl_factor()
1135 int smt, cpu, cpus = 0; in update_numa_stats() local
1146 cpus++; in update_numa_stats()
1157 if (!cpus) in update_numa_stats()
1161 smt = DIV_ROUND_UP(SCHED_CAPACITY_SCALE * cpus, ns->compute_capacity); in update_numa_stats()
1162 capacity = cpus / smt; /* cores */ in update_numa_stats()
5427 struct cpumask *cpus; member
5599 for_each_cpu_and(cpu, env->dst_grpmask, env->cpus) { in can_migrate_task()
[all …]
/linux-4.1.27/arch/powerpc/boot/dts/fsl/
Dbsc9131si-pre.dtsi51 cpus {
Dmpc8536si-pre.dtsi56 cpus {
Dmpc8544si-pre.dtsi56 cpus {
Dp1021si-pre.dtsi55 cpus {
Dbsc9132si-pre.dtsi50 cpus {
Dmpc8572si-pre.dtsi57 cpus {
Dmpc8548si-pre.dtsi57 cpus {
Dc293si-pre.dtsi53 cpus {

123