/linux-4.1.27/tools/perf/util/ |
D | cpumap.c | 11 struct cpu_map *cpus; in cpu_map__default_new() local 18 cpus = malloc(sizeof(*cpus) + nr_cpus * sizeof(int)); in cpu_map__default_new() 19 if (cpus != NULL) { in cpu_map__default_new() 22 cpus->map[i] = i; in cpu_map__default_new() 24 cpus->nr = nr_cpus; in cpu_map__default_new() 27 return cpus; in cpu_map__default_new() 33 struct cpu_map *cpus = malloc(sizeof(*cpus) + payload_size); in cpu_map__trim_new() local 35 if (cpus != NULL) { in cpu_map__trim_new() 36 cpus->nr = nr_cpus; in cpu_map__trim_new() 37 memcpy(cpus->map, tmp_cpus, payload_size); in cpu_map__trim_new() [all …]
|
D | record.c | 60 struct cpu_map *cpus; in perf_probe_api() local 63 cpus = cpu_map__new(NULL); in perf_probe_api() 64 if (!cpus) in perf_probe_api() 66 cpu = cpus->map[0]; in perf_probe_api() 67 cpu_map__delete(cpus); in perf_probe_api() 111 if (evlist->cpus->map[0] < 0) in perf_evlist__config() 216 if (!evlist || cpu_map__empty(evlist->cpus)) { in perf_evlist__can_select_event() 217 struct cpu_map *cpus = cpu_map__new(NULL); in perf_evlist__can_select_event() local 219 cpu = cpus ? cpus->map[0] : 0; in perf_evlist__can_select_event() 220 cpu_map__delete(cpus); in perf_evlist__can_select_event() [all …]
|
D | evlist.c | 35 void perf_evlist__init(struct perf_evlist *evlist, struct cpu_map *cpus, in perf_evlist__init() argument 43 perf_evlist__set_maps(evlist, cpus, threads); in perf_evlist__init() 117 cpu_map__delete(evlist->cpus); in perf_evlist__delete() 119 evlist->cpus = NULL; in perf_evlist__delete() 287 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__disable() 306 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable() 325 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__disable_event() 346 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable_event() 386 int nr_cpus = cpu_map__nr(evlist->cpus); in perf_evlist__enable_event_thread() 402 bool per_cpu_mmaps = !cpu_map__empty(evlist->cpus); in perf_evlist__enable_event_idx() [all …]
|
D | python.c | 366 struct cpu_map *cpus; member 379 pcpus->cpus = cpu_map__new(cpustr); in pyrf_cpu_map__init() 380 if (pcpus->cpus == NULL) in pyrf_cpu_map__init() 387 cpu_map__delete(pcpus->cpus); in pyrf_cpu_map__delete() 395 return pcpus->cpus->nr; in pyrf_cpu_map__length() 402 if (i >= pcpus->cpus->nr) in pyrf_cpu_map__item() 405 return Py_BuildValue("i", pcpus->cpus->map[i]); in pyrf_cpu_map__item() 622 struct cpu_map *cpus = NULL; in pyrf_evsel__open() local 636 cpus = ((struct pyrf_cpu_map *)pcpus)->cpus; in pyrf_evsel__open() 643 if (perf_evsel__open(evsel, cpus, threads) < 0) { in pyrf_evsel__open() [all …]
|
D | evlist.h | 52 struct cpu_map *cpus; member 64 void perf_evlist__init(struct perf_evlist *evlist, struct cpu_map *cpus, 147 struct cpu_map *cpus, in perf_evlist__set_maps() argument 150 evlist->cpus = cpus; in perf_evlist__set_maps()
|
D | top.c | 98 top->evlist->cpus->nr > 1 ? "s" : "", in perf_top__header_snprintf() 105 top->evlist->cpus->nr, in perf_top__header_snprintf() 106 top->evlist->cpus->nr > 1 ? "s" : ""); in perf_top__header_snprintf()
|
D | cpumap.h | 22 int cpu_map__build_socket_map(struct cpu_map *cpus, struct cpu_map **sockp); 23 int cpu_map__build_core_map(struct cpu_map *cpus, struct cpu_map **corep);
|
D | evsel.h | 85 struct cpu_map *cpus; member 197 struct cpu_map *cpus); 200 int perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus,
|
D | evsel.c | 1134 static int __perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus, in __perf_evsel__open() argument 1148 perf_evsel__alloc_fd(evsel, cpus->nr, nthreads) < 0) in __perf_evsel__open() 1180 for (cpu = 0; cpu < cpus->nr; cpu++) { in __perf_evsel__open() 1191 pid, cpus->map[cpu], group_fd, flags); in __perf_evsel__open() 1195 cpus->map[cpu], in __perf_evsel__open() 1295 int cpus[1]; member 1298 .cpus = { -1, }, 1309 int perf_evsel__open(struct perf_evsel *evsel, struct cpu_map *cpus, in perf_evsel__open() argument 1312 if (cpus == NULL) { in perf_evsel__open() 1314 cpus = &empty_cpu_map.map; in perf_evsel__open() [all …]
|
D | pmu.h | 23 struct cpu_map *cpus; member
|
D | svghelper.h | 6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
|
D | pmu.c | 412 struct cpu_map *cpus; in pmu_cpumask() local 428 cpus = cpu_map__read(file); in pmu_cpumask() 430 return cpus; in pmu_cpumask() 464 pmu->cpus = pmu_cpumask(name); in pmu_lookup()
|
D | parse-events.c | 277 char *name, struct cpu_map *cpus) in __add_event() argument 287 evsel->cpus = cpus; in __add_event() 650 evsel = __add_event(list, idx, &attr, NULL, pmu->cpus); in parse_events_add_pmu() 667 pmu->cpus); in parse_events_add_pmu()
|
/linux-4.1.27/tools/perf/tests/ |
D | open-syscall-all-cpus.c | 10 struct cpu_map *cpus; in test__open_syscall_event_on_all_cpus() local 22 cpus = cpu_map__new(NULL); in test__open_syscall_event_on_all_cpus() 23 if (cpus == NULL) { in test__open_syscall_event_on_all_cpus() 41 if (perf_evsel__open(evsel, cpus, threads) < 0) { in test__open_syscall_event_on_all_cpus() 48 for (cpu = 0; cpu < cpus->nr; ++cpu) { in test__open_syscall_event_on_all_cpus() 56 if (cpus->map[cpu] >= CPU_SETSIZE) { in test__open_syscall_event_on_all_cpus() 57 pr_debug("Ignoring CPU %d\n", cpus->map[cpu]); in test__open_syscall_event_on_all_cpus() 61 CPU_SET(cpus->map[cpu], &cpu_set); in test__open_syscall_event_on_all_cpus() 64 cpus->map[cpu], in test__open_syscall_event_on_all_cpus() 72 CPU_CLR(cpus->map[cpu], &cpu_set); in test__open_syscall_event_on_all_cpus() [all …]
|
D | mmap-basic.c | 23 struct cpu_map *cpus; in test__basic_mmap() local 42 cpus = cpu_map__new(NULL); in test__basic_mmap() 43 if (cpus == NULL) { in test__basic_mmap() 49 CPU_SET(cpus->map[0], &cpu_set); in test__basic_mmap() 53 cpus->map[0], strerror_r(errno, sbuf, sizeof(sbuf))); in test__basic_mmap() 63 perf_evlist__set_maps(evlist, cpus, threads); in test__basic_mmap() 80 if (perf_evsel__open(evsels[i], cpus, threads) < 0) { in test__basic_mmap() 142 cpus = NULL; in test__basic_mmap() 145 cpu_map__delete(cpus); in test__basic_mmap()
|
D | keep-tracking.c | 64 struct cpu_map *cpus = NULL; in test__keep_tracking() local 73 cpus = cpu_map__new(NULL); in test__keep_tracking() 74 CHECK_NOT_NULL__(cpus); in test__keep_tracking() 79 perf_evlist__set_maps(evlist, cpus, threads); in test__keep_tracking() 147 cpu_map__delete(cpus); in test__keep_tracking()
|
D | perf-time-to-tsc.c | 49 struct cpu_map *cpus = NULL; in test__perf_time_to_tsc() local 63 cpus = cpu_map__new(NULL); in test__perf_time_to_tsc() 64 CHECK_NOT_NULL__(cpus); in test__perf_time_to_tsc() 69 perf_evlist__set_maps(evlist, cpus, threads); in test__perf_time_to_tsc()
|
D | task-exit.c | 61 evlist->cpus = cpu_map__dummy_new(); in test__task_exit() 63 if (!evlist->cpus || !evlist->threads) { in test__task_exit()
|
D | sw-clock.c | 53 evlist->cpus = cpu_map__dummy_new(); in __test__sw_clock_freq() 55 if (!evlist->cpus || !evlist->threads) { in __test__sw_clock_freq()
|
D | code-reading.c | 403 struct cpu_map *cpus = NULL; in do_test_code_reading() local 463 cpus = cpu_map__new(NULL); in do_test_code_reading() 464 if (!cpus) { in do_test_code_reading() 478 perf_evlist__set_maps(evlist, cpus, threads); in do_test_code_reading() 542 cpu_map__delete(cpus); in do_test_code_reading()
|
D | switch-tracking.c | 322 struct cpu_map *cpus = NULL; in test__switch_tracking() local 335 cpus = cpu_map__new(NULL); in test__switch_tracking() 336 if (!cpus) { in test__switch_tracking() 347 perf_evlist__set_maps(evlist, cpus, threads); in test__switch_tracking() 563 cpu_map__delete(cpus); in test__switch_tracking()
|
D | Build | 7 perf-y += open-syscall-all-cpus.o
|
/linux-4.1.27/tools/power/cpupower/utils/ |
D | cpufreq-info.c | 250 struct cpufreq_affected_cpus *cpus; in debug_output_one() local 273 cpus = cpufreq_get_related_cpus(cpu); in debug_output_one() 274 if (cpus) { in debug_output_one() 276 while (cpus->next) { in debug_output_one() 277 printf("%d ", cpus->cpu); in debug_output_one() 278 cpus = cpus->next; in debug_output_one() 280 printf("%d\n", cpus->cpu); in debug_output_one() 281 cpufreq_put_related_cpus(cpus); in debug_output_one() 284 cpus = cpufreq_get_affected_cpus(cpu); in debug_output_one() 285 if (cpus) { in debug_output_one() [all …]
|
D | cpufreq-set.c | 295 struct cpufreq_affected_cpus *cpus; in cmd_freq_set() local 301 cpus = cpufreq_get_related_cpus(cpu); in cmd_freq_set() 302 if (!cpus) in cmd_freq_set() 304 while (cpus->next) { in cmd_freq_set() 305 bitmask_setbit(cpus_chosen, cpus->cpu); in cmd_freq_set() 306 cpus = cpus->next; in cmd_freq_set() 308 cpufreq_put_related_cpus(cpus); in cmd_freq_set()
|
/linux-4.1.27/include/linux/ |
D | stop_machine.h | 115 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus); 126 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus); 129 const struct cpumask *cpus); 134 const struct cpumask *cpus) in __stop_machine() argument 145 const struct cpumask *cpus) in stop_machine() argument 147 return __stop_machine(fn, data, cpus); in stop_machine() 151 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument 153 return __stop_machine(fn, data, cpus); in stop_machine_from_inactive_cpu()
|
D | qcom_scm.h | 16 extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus); 17 extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus);
|
D | cpufreq.h | 63 cpumask_var_t cpus; /* Online CPUs only */ member 140 return cpumask_weight(policy->cpus) > 1; in policy_is_shared()
|
/linux-4.1.27/tools/perf/python/ |
D | twatch.py | 19 cpus = perf.cpu_map() 25 evsel.open(cpus = cpus, threads = threads); 26 evlist = perf.evlist(cpus, threads) 31 for cpu in cpus:
|
/linux-4.1.27/block/ |
D | blk-mq-cpumap.c | 37 cpumask_var_t cpus; in blk_mq_update_queue_map() local 39 if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) in blk_mq_update_queue_map() 42 cpumask_clear(cpus); in blk_mq_update_queue_map() 47 if (!cpumask_test_cpu(first_sibling, cpus)) in blk_mq_update_queue_map() 49 cpumask_set_cpu(i, cpus); in blk_mq_update_queue_map() 84 free_cpumask_var(cpus); in blk_mq_update_queue_map()
|
/linux-4.1.27/arch/mips/cavium-octeon/ |
D | smp.c | 103 int cpus; in octeon_smp_setup() local 120 cpus = 1; in octeon_smp_setup() 123 set_cpu_possible(cpus, true); in octeon_smp_setup() 124 set_cpu_present(cpus, true); in octeon_smp_setup() 125 __cpu_number_map[id] = cpus; in octeon_smp_setup() 126 __cpu_logical_map[cpus] = id; in octeon_smp_setup() 127 cpus++; in octeon_smp_setup() 140 set_cpu_possible(cpus, true); in octeon_smp_setup() 141 __cpu_number_map[id] = cpus; in octeon_smp_setup() 142 __cpu_logical_map[cpus] = id; in octeon_smp_setup() [all …]
|
/linux-4.1.27/arch/arm/mach-shmobile/ |
D | timer.c | 40 struct device_node *np, *cpus; in shmobile_init_delay() local 46 cpus = of_find_node_by_path("/cpus"); in shmobile_init_delay() 47 if (!cpus) in shmobile_init_delay() 50 for_each_child_of_node(cpus, np) { in shmobile_init_delay() 68 of_node_put(cpus); in shmobile_init_delay()
|
D | pm-rcar-gen2.c | 58 struct device_node *np, *cpus; in rcar_gen2_pm_init() local 67 cpus = of_find_node_by_path("/cpus"); in rcar_gen2_pm_init() 68 if (!cpus) in rcar_gen2_pm_init() 71 for_each_child_of_node(cpus, np) { in rcar_gen2_pm_init()
|
D | platsmp-apmu.c | 99 for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) { in apmu_parse_cfg() 100 id = apmu_config[k].cpus[bit]; in apmu_parse_cfg() 109 for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) { in apmu_parse_cfg() 110 id = apmu_config[k].cpus[bit]; in apmu_parse_cfg()
|
D | smp-r8a7790.c | 42 .cpus = { 0, 1, 2, 3 }, 46 .cpus = { 0x100, 0x0101, 0x102, 0x103 },
|
D | platsmp-apmu.h | 21 int cpus[4]; member
|
D | smp-r8a7791.c | 31 .cpus = { 0, 1 },
|
/linux-4.1.27/arch/x86/kernel/ |
D | tsc_sync.c | 125 int cpus = 2; in check_tsc_sync_source() local 149 while (atomic_read(&start_count) != cpus-1) in check_tsc_sync_source() 158 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source() 191 int cpus = 2; in check_tsc_sync_target() local 201 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target() 214 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target()
|
/linux-4.1.27/tools/power/cpupower/utils/helpers/ |
D | topology.c | 67 int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF); in get_cpu_topology() local 69 cpu_top->core_info = malloc(sizeof(struct cpuid_core_info) * cpus); in get_cpu_topology() 73 for (cpu = 0; cpu < cpus; cpu++) { in get_cpu_topology() 88 qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info), in get_cpu_topology() 95 for(cpu = 1; cpu < cpus; cpu++) { in get_cpu_topology() 110 return cpus; in get_cpu_topology()
|
/linux-4.1.27/drivers/irqchip/ |
D | irq-bcm7038-l1.c | 46 struct bcm7038_l1_cpu *cpus[NR_CPUS]; member 127 cpu = intc->cpus[cpu_logical_map(smp_processor_id())]; in bcm7038_l1_irq_handle() 129 cpu = intc->cpus[0]; in bcm7038_l1_irq_handle() 159 intc->cpus[cpu_idx]->mask_cache[word] &= ~mask; in __bcm7038_l1_unmask() 160 l1_writel(mask, intc->cpus[cpu_idx]->map_base + in __bcm7038_l1_unmask() 170 intc->cpus[cpu_idx]->mask_cache[word] |= mask; in __bcm7038_l1_mask() 171 l1_writel(mask, intc->cpus[cpu_idx]->map_base + in __bcm7038_l1_mask() 209 was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] & in bcm7038_l1_set_affinity() 241 cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), in bcm7038_l1_init_one() 316 intc->cpus[0]->map_base, IRQS_PER_WORD * intc->n_words); in bcm7038_l1_of_init() [all …]
|
/linux-4.1.27/drivers/lguest/ |
D | lguest_user.c | 119 cpu = &lg->cpus[cpu_id]; in read() 158 if (id >= ARRAY_SIZE(cpu->lg->cpus)) in lg_cpu_start() 163 cpu->lg = container_of(cpu, struct lguest, cpus[id]); in lg_cpu_start() 256 err = lg_cpu_start(&lg->cpus[0], 0, args[2]); in initialize() 278 free_page(lg->cpus[0].regs_page); in initialize() 319 cpu = &lg->cpus[cpu_id]; in write() 378 hrtimer_cancel(&lg->cpus[i].hrt); in close() 380 free_page(lg->cpus[i].regs_page); in close() 385 mmput(lg->cpus[i].mm); in close()
|
D | page_tables.c | 1035 kill_guest(&lg->cpus[0], "Attempt to set pgd %u/%u", in guest_set_pgd() 1046 if (!allocate_switcher_mapping(&lg->cpus[0])) { in guest_set_pgd() 1047 kill_guest(&lg->cpus[0], in guest_set_pgd() 1058 guest_pagetable_clear_all(&lg->cpus[0]); in guest_set_pmd() 1075 struct lg_cpu *cpu = &lg->cpus[0]; in init_guest_pagetable()
|
/linux-4.1.27/tools/lib/traceevent/ |
D | plugin_function.c | 32 static int cpus = -1; variable 95 if (cpu > cpus) { in add_and_get_index() 107 for (i = cpus + 1; i <= cpu; i++) in add_and_get_index() 109 cpus = cpu; in add_and_get_index() 183 for (i = 0; i <= cpus; i++) { in PEVENT_PLUGIN_UNLOADER() 193 cpus = -1; in PEVENT_PLUGIN_UNLOADER()
|
D | event-parse.h | 474 int cpus; member 709 return pevent->cpus; in pevent_get_cpus() 712 static inline void pevent_set_cpus(struct pevent *pevent, int cpus) in pevent_set_cpus() argument 714 pevent->cpus = cpus; in pevent_set_cpus()
|
/linux-4.1.27/arch/arm/kernel/ |
D | devtree.c | 77 struct device_node *cpu, *cpus; in arm_dt_init_cpu_maps() local 84 cpus = of_find_node_by_path("/cpus"); in arm_dt_init_cpu_maps() 86 if (!cpus) in arm_dt_init_cpu_maps() 89 for_each_child_of_node(cpus, cpu) { in arm_dt_init_cpu_maps() 160 set_smp_ops_by_method(cpus); in arm_dt_init_cpu_maps()
|
/linux-4.1.27/drivers/firmware/ |
D | qcom_scm.c | 415 int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) in qcom_scm_set_cold_boot_addr() argument 426 if (!cpus || (cpus && cpumask_empty(cpus))) in qcom_scm_set_cold_boot_addr() 429 for_each_cpu(cpu, cpus) { in qcom_scm_set_cold_boot_addr() 448 int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) in qcom_scm_set_warm_boot_addr() argument 458 for_each_cpu(cpu, cpus) { in qcom_scm_set_warm_boot_addr() 470 for_each_cpu(cpu, cpus) in qcom_scm_set_warm_boot_addr()
|
/linux-4.1.27/arch/arm/common/ |
D | mcpm_entry.c | 346 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; in __mcpm_cpu_going_down() 347 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_going_down() 360 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; in __mcpm_cpu_down() 361 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_down() 413 sync_cache_r(&c->cpus); in __mcpm_outbound_enter_critical() 421 cpustate = c->cpus[i].cpu; in __mcpm_outbound_enter_critical() 426 sync_cache_r(&c->cpus[i].cpu); in __mcpm_outbound_enter_critical() 469 mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN; in mcpm_sync_init() 475 mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP; in mcpm_sync_init()
|
/linux-4.1.27/tools/testing/selftests/rcutorture/bin/ |
D | kvm.sh | 48 cpus=0 96 --cpus) 97 checkarg --cpus "(number)" "$#" "$2" '^[0-9]*$' '^--' 98 cpus=$2 194 awk < $T/cfgcpu.sort > $T/cfgcpu.pack -v ncpus=$cpus ' 281 -v ncpus=$cpus \
|
/linux-4.1.27/tools/virtio/virtio-trace/ |
D | trace-agent.c | 62 s->cpus = get_total_cpus(); in agent_info_new() 66 for (i = 0; i < s->cpus; i++) in agent_info_new() 159 for (cpu = 0; cpu < s->cpus; cpu++) { in agent_info_init() 226 for (cpu = 0; cpu < s->cpus; cpu++) in agent_main_loop() 232 for (cpu = 0; cpu < s->cpus; cpu++) { in agent_main_loop() 248 for (i = 0; i < s->cpus; i++) { in agent_info_free()
|
D | trace-agent.h | 20 int cpus; member
|
/linux-4.1.27/arch/ia64/kernel/ |
D | smp.c | 295 cpumask_var_t cpus; in smp_flush_tlb_mm() local 304 if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) { in smp_flush_tlb_mm() 308 cpumask_copy(cpus, mm_cpumask(mm)); in smp_flush_tlb_mm() 309 smp_call_function_many(cpus, in smp_flush_tlb_mm() 311 free_cpumask_var(cpus); in smp_flush_tlb_mm()
|
/linux-4.1.27/arch/powerpc/platforms/powermac/ |
D | smp.c | 635 struct device_node *cpus; in smp_core99_pfunc_tb_freeze() local 638 cpus = of_find_node_by_path("/cpus"); in smp_core99_pfunc_tb_freeze() 639 BUG_ON(cpus == NULL); in smp_core99_pfunc_tb_freeze() 642 pmf_call_function(cpus, "cpu-timebase", &args); in smp_core99_pfunc_tb_freeze() 643 of_node_put(cpus); in smp_core99_pfunc_tb_freeze() 713 struct device_node *cpus = in smp_core99_setup() local 715 if (cpus && in smp_core99_setup() 716 of_get_property(cpus, "platform-cpu-timebase", NULL)) { in smp_core99_setup() 769 struct device_node *cpus; in smp_core99_probe() local 775 for (cpus = NULL; (cpus = of_find_node_by_type(cpus, "cpu")) != NULL;) in smp_core99_probe()
|
D | feature.c | 1052 struct device_node *cpus; in core99_reset_cpu() local 1062 cpus = of_find_node_by_path("/cpus"); in core99_reset_cpu() 1063 if (cpus == NULL) in core99_reset_cpu() 1065 for (np = cpus->child; np != NULL; np = np->sibling) { in core99_reset_cpu() 1075 of_node_put(cpus); in core99_reset_cpu() 1507 struct device_node *cpus; in g5_reset_cpu() local 1513 cpus = of_find_node_by_path("/cpus"); in g5_reset_cpu() 1514 if (cpus == NULL) in g5_reset_cpu() 1516 for (np = cpus->child; np != NULL; np = np->sibling) { in g5_reset_cpu() 1526 of_node_put(cpus); in g5_reset_cpu() [all …]
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | rtas.c | 835 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument 837 if (!cpumask_empty(cpus)) { in rtas_cpu_state_change_mask() 838 cpumask_clear(cpus); in rtas_cpu_state_change_mask() 848 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument 854 if (cpumask_empty(cpus)) in rtas_cpu_state_change_mask() 857 for_each_cpu(cpu, cpus) { in rtas_cpu_state_change_mask() 875 cpumask_shift_right(cpus, cpus, cpu); in rtas_cpu_state_change_mask() 876 cpumask_shift_left(cpus, cpus, cpu); in rtas_cpu_state_change_mask() 880 cpumask_clear_cpu(cpu, cpus); in rtas_cpu_state_change_mask() 889 int rtas_online_cpus_mask(cpumask_var_t cpus) in rtas_online_cpus_mask() argument [all …]
|
/linux-4.1.27/drivers/cpuidle/ |
D | coupled.c | 429 cpumask_t cpus; in cpuidle_coupled_any_pokes_pending() local 432 cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); in cpuidle_coupled_any_pokes_pending() 433 ret = cpumask_and(&cpus, &cpuidle_coupled_poke_pending, &cpus); in cpuidle_coupled_any_pokes_pending() 616 cpumask_t cpus; in cpuidle_coupled_update_online_cpus() local 617 cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); in cpuidle_coupled_update_online_cpus() 618 coupled->online_count = cpumask_weight(&cpus); in cpuidle_coupled_update_online_cpus()
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
D | grutlbpurge.c | 356 int cpus, shift = 0, n; in gru_tgh_flush_init() local 358 cpus = uv_blade_nr_possible_cpus(gru->gs_blade_id); in gru_tgh_flush_init() 361 if (cpus) { in gru_tgh_flush_init() 362 n = 1 << fls(cpus - 1); in gru_tgh_flush_init() 375 gru->gs_tgh_first_remote = (cpus + (1 << shift) - 1) >> shift; in gru_tgh_flush_init()
|
D | grulib.h | 146 int cpus; member
|
/linux-4.1.27/Documentation/devicetree/bindings/cpufreq/ |
D | arm_big_little_dt.txt | 8 under node /cpus/cpu@x. Where x is the first cpu inside a cluster. 10 FIXME: Cpus should boot in the order specified in DT and all cpus for a cluster 24 cpus {
|
D | cpufreq-spear.txt | 17 /cpus/cpu@0. 21 cpus {
|
D | cpufreq-dt.txt | 8 under node /cpus/cpu@0. 27 cpus {
|
/linux-4.1.27/Documentation/devicetree/bindings/arm/ |
D | cpus.txt | 6 the "cpus" node, which in turn contains a number of subnodes (ie "cpu") 26 cpus and cpu node bindings definition 29 The ARM architecture, in accordance with the ePAPR, requires the cpus and cpu 32 - cpus node 36 The node name must be "cpus". 38 A cpus node must define the following properties: 102 * If cpus node's #address-cells property is set to 2 110 * If cpus node's #address-cells property is set to 1 243 cpus { 274 cpus { [all …]
|
D | kirkwood.txt | 12 cpus/cpu@0 with three clocks, "cpu_clk", "ddrclk" and "powersave", 18 cpus {
|
D | topology.txt | 47 child of the cpus node and provides a container where the actual topology 62 The cpu-map node's parent node must be the cpus node. 169 cpus { 388 cpus { 475 Documentation/devicetree/bindings/arm/cpus.txt
|
D | vexpress.txt | 56 Top-level standard "cpus" node is required. It must contain a node 59 cpus { 184 cpus {
|
D | al,alpine.txt | 25 cpus {
|
/linux-4.1.27/drivers/cpufreq/ |
D | cpufreq.c | 184 cpumask_setall(policy->cpus); in cpufreq_generic_init() 339 for_each_cpu(freqs->cpu, policy->cpus) in cpufreq_notify_transition() 685 return cpufreq_show_cpus(policy->cpus, buf); in show_affected_cpus() 882 for_each_cpu(j, policy->cpus) { in cpufreq_add_dev_symlink() 980 cpumask_set_cpu(cpu, policy->cpus); in cpufreq_add_policy_cpu() 1025 if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL)) in cpufreq_policy_alloc() 1041 free_cpumask_var(policy->cpus); in cpufreq_policy_alloc() 1075 free_cpumask_var(policy->cpus); in cpufreq_policy_free() 1158 cpumask_copy(policy->cpus, cpumask_of(cpu)); in __cpufreq_add_dev() 1172 cpumask_or(policy->related_cpus, policy->related_cpus, policy->cpus); in __cpufreq_add_dev() [all …]
|
D | speedstep-ich.c | 262 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_target() 295 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in speedstep_cpu_init() 297 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_cpu_init()
|
D | cpufreq_governor.c | 66 for_each_cpu(j, policy->cpus) { in dbs_check_cpu() 190 for_each_cpu(i, policy->cpus) in gov_queue_work() 205 for_each_cpu(i, policy->cpus) { in gov_cancel_work() 372 for_each_cpu(j, policy->cpus) { in cpufreq_governor_dbs()
|
D | p4-clockmod.c | 116 for_each_cpu(i, policy->cpus) in cpufreq_p4_target() 175 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in cpufreq_p4_cpu_init()
|
D | acpi-cpufreq.c | 460 cmd.mask = policy->cpus; in acpi_cpufreq_target() 694 cpumask_copy(policy->cpus, perf->shared_cpu_map); in acpi_cpufreq_cpu_init() 702 cpumask_copy(policy->cpus, cpu_core_mask(cpu)); in acpi_cpufreq_cpu_init() 706 cpumask_clear(policy->cpus); in acpi_cpufreq_cpu_init() 707 cpumask_set_cpu(cpu, policy->cpus); in acpi_cpufreq_cpu_init()
|
D | powernv-cpufreq.c | 362 smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1); in powernv_cpufreq_target_index() 374 cpumask_set_cpu(base + i, policy->cpus); in powernv_cpufreq_cpu_init()
|
D | ppc_cbe_cpufreq.c | 121 cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); in cbe_cpufreq_cpu_init()
|
D | speedstep-centrino.c | 439 for_each_cpu(j, policy->cpus) { in centrino_target() 447 good_cpu = cpumask_any_and(policy->cpus, in centrino_target()
|
D | powernow-k8.c | 1093 cpumask_copy(pol->cpus, cpu_core_mask(pol->cpu)); in powernowk8_cpu_init() 1094 data->available_cores = pol->cpus; in powernowk8_cpu_init() 1109 for_each_cpu(cpu, pol->cpus) in powernowk8_cpu_init() 1134 for_each_cpu(cpu, pol->cpus) in powernowk8_cpu_exit()
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-devices-system-xen_cpu | 16 Interface to online/offline Xen physical cpus 19 to online/offline physical cpus, except cpu0 due to several
|
D | sysfs-devices-system-cpu | 25 offline: cpus that are not online because they have been 26 HOTPLUGGED off or exceed the limit of cpus allowed by the 29 online: cpus that are online and being scheduled. 31 possible: cpus that have been allocated resources and can be 34 present: cpus that have been identified as being present in 253 shared_cpu_list: the list of logical cpus sharing the cache 255 shared_cpu_map: logical cpu mask containing the list of cpus sharing
|
/linux-4.1.27/samples/trace_events/ |
D | trace-events-sample.h | 240 __bitmask( cpus, num_possible_cpus() ) 249 __assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus()); 296 __get_str(str), __get_bitmask(cpus))
|
/linux-4.1.27/kernel/ |
D | stop_machine.c | 550 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in __stop_machine() argument 556 .active_cpus = cpus, in __stop_machine() 583 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in stop_machine() argument 589 ret = __stop_machine(fn, data, cpus); in stop_machine() 618 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument 621 .active_cpus = cpus }; in stop_machine_from_inactive_cpu()
|
D | smp.c | 669 cpumask_var_t cpus; in on_each_cpu_cond() local 674 if (likely(zalloc_cpumask_var(&cpus, (gfp_flags|__GFP_NOWARN)))) { in on_each_cpu_cond() 678 cpumask_set_cpu(cpu, cpus); in on_each_cpu_cond() 679 on_each_cpu_mask(cpus, func, info, wait); in on_each_cpu_cond() 681 free_cpumask_var(cpus); in on_each_cpu_cond()
|
/linux-4.1.27/Documentation/devicetree/bindings/arm/cpu-enable-method/ |
D | marvell,berlin-smp | 7 be defined in the "cpus" node. 20 cpus {
|
D | al,alpine-smp | 8 "cpus" node. 21 cpus {
|
/linux-4.1.27/Documentation/devicetree/bindings/mips/img/ |
D | pistachio.txt | 10 A "cpus" node is required. Required properties: 22 cpus {
|
/linux-4.1.27/arch/s390/hypfs/ |
D | hypfs_diag.c | 120 __u8 cpus; member 127 __u8 cpus; member 154 return ((struct part_hdr *)hdr)->cpus; in part_hdr__rcpus() 256 __u8 cpus; member 263 __u8 cpus; member 280 return ((struct phys_hdr *)hdr)->cpus; in phys_hdr__cpus() 282 return ((struct x_phys_hdr *)hdr)->cpus; in phys_hdr__cpus()
|
/linux-4.1.27/Documentation/devicetree/bindings/arm/bcm/ |
D | brcm,bcm11351-cpu-method.txt | 8 properties in the "cpus" device tree node: 19 cpus {
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
D | cpufreq_spudemand.c | 112 for_each_cpu(i, policy->cpus) { in spu_gov_govern() 129 for_each_cpu (i, policy->cpus) { in spu_gov_govern()
|
/linux-4.1.27/Documentation/ia64/ |
D | mca.txt | 7 the OS is in any state. Including when one of the cpus is already 22 to all the other cpus, the slaves. 24 * Slave cpus that receive the MCA interrupt call down into SAL, they 29 sends an unmaskable INIT event to the slave cpus that have not 68 tasks. But (and its a big but), the cpus that received the MCA 97 cpus. SAL picks one of the cpus as the monarch and the rest are 103 versions of SAL out there. Some drive all the cpus as monarchs. Some 126 all cpus.
|
/linux-4.1.27/Documentation/accounting/ |
D | taskstats.txt | 36 one of the cpus in the cpumask, its per-pid statistics are sent to the 82 Commands to register/deregister interest in exit data from a set of cpus 86 comma-separated cpu ranges e.g. to listen to exit data from cpus 1,2,3,5,7,8 88 in cpus before closing the listening socket, the kernel cleans up its interest 165 extended and the number of cpus grows large. 172 - create more listeners and reduce the number of cpus being listened to by 175 of cpus to which it listens, especially if they are listening to just one cpu.
|
/linux-4.1.27/Documentation/thermal/ |
D | cpu-cooling-api.txt | 26 clip_cpus: cpumask of cpus where the frequency constraints will happen. 37 clip_cpus: cpumask of cpus where the frequency constraints will happen.
|
/linux-4.1.27/arch/ia64/mm/ |
D | discontig.c | 119 unsigned long pernodesize = 0, cpus; in compute_pernodesize() local 121 cpus = early_nr_cpus_node(node); in compute_pernodesize() 122 pernodesize += PERCPU_PAGE_SIZE * cpus; in compute_pernodesize() 270 int cpus = early_nr_cpus_node(node); in fill_pernode() local 278 pernode += PERCPU_PAGE_SIZE * cpus; in fill_pernode()
|
/linux-4.1.27/arch/powerpc/boot/dts/ |
D | mpc8536ds.dts | 18 cpus { 19 #cpus = <1>;
|
D | mpc8536ds_36b.dts | 18 cpus { 19 #cpus = <1>;
|
D | iss4xx.dts | 22 dcr-parent = <&{/cpus/cpu@0}>; 28 cpus {
|
D | amigaone.dts | 21 cpus { 22 #cpus = <1>;
|
D | iss4xx-mpic.dts | 24 dcr-parent = <&{/cpus/cpu@0}>; 30 cpus {
|
D | sbc8548-pre.dtsi | 29 cpus {
|
D | p1020rdb-pc_camp_core0.dts | 33 cpus {
|
D | acadia.dts | 18 dcr-parent = <&{/cpus/cpu@0}>; 26 cpus {
|
D | ep405.dts | 19 dcr-parent = <&{/cpus/cpu@0}>; 27 cpus {
|
D | mpc8572ds_camp_core0.dts | 23 cpus {
|
D | klondike.dts | 31 dcr-parent = <&{/cpus/cpu@0}>; 38 cpus {
|
D | walnut.dts | 19 dcr-parent = <&{/cpus/cpu@0}>; 27 cpus {
|
D | gamecube.dts | 32 cpus {
|
D | warp.dts | 19 dcr-parent = <&{/cpus/cpu@0}>; 26 cpus {
|
D | ps3.dts | 54 cpus {
|
D | mpc8572ds_camp_core1.dts | 24 cpus {
|
D | p1020rdb-pc_camp_core1.dts | 29 cpus {
|
D | hotfoot.dts | 18 dcr-parent = <&{/cpus/cpu@0}>; 27 cpus {
|
/linux-4.1.27/Documentation/cgroups/ |
D | cpusets.txt | 28 2.2 Adding/removing cpus 171 - cpuset.cpus: list of CPUs in that cpuset 218 The cpus and mems files in the root (top_cpuset) cpuset are 219 read-only. The cpus file automatically tracks the value of 410 setting), it requests that all the CPUs in that cpusets allowed 'cpuset.cpus' 460 If two cpusets have partially overlapping 'cpuset.cpus' allowed, and only 560 3 : search cpus in a node [= system wide on non-NUMA system] 616 If a cpuset has its 'cpuset.cpus' modified, then each task in that cpuset 650 with non-empty cpus. But the moving of some (or all) tasks might fail if 687 /bin/echo 2-3 > cpuset.cpus [all …]
|
/linux-4.1.27/scripts/gdb/ |
D | vmlinux-gdb.py | 30 import linux.cpus
|
/linux-4.1.27/Documentation/ |
D | cpu-hotplug.txt | 45 maxcpus=n Restrict boot time cpus to n. Say if you have 4 cpus, using 47 other cpus later online, read FAQ's for more info. 49 additional_cpus=n (*) Use this to limit hotpluggable cpus. This option sets 62 determine the number of potentially hot-pluggable cpus. The implementation 63 should only rely on this to count the # of cpus, but *MUST* not rely 65 BIOS doesn't mark such hot-pluggable cpus as disabled entries, one could 66 use this parameter "additional_cpus=x" to represent those cpus in the 69 possible_cpus=n [s390,x86_64] use this to set hotpluggable cpus. 118 If you merely need to avoid cpus going away, you could also use 177 Q: Does hot-add/hot-remove refer to physical add/remove of cpus?
|
D | IRQ-affinity.txt | 11 IRQ affinity then the value will not change from the default of all cpus. 58 Here is an example of limiting that same irq (44) to cpus 1024 to 1031:
|
D | cputopology.txt | 75 above). [~cpu_online_mask + cpus >= NR_CPUS] 88 In this example, there are 64 CPUs in the system but cpus 32-63 exceed
|
/linux-4.1.27/Documentation/devicetree/bindings/mips/brcm/ |
D | brcm,bmips.txt | 8 under the "cpus" node.
|
/linux-4.1.27/arch/mips/boot/dts/mti/ |
D | sead3.dts | 12 cpus {
|
/linux-4.1.27/arch/arm/boot/dts/ |
D | bcm47081.dtsi | 15 cpus {
|
D | bcm4708.dtsi | 15 cpus {
|
D | dra72x.dtsi | 15 cpus {
|
D | axm5516-amarillo.dts | 17 #include "axm5516-cpus.dtsi"
|
D | hip01-ca9x2.dts | 25 cpus {
|
D | stih41x.dtsi | 12 cpus {
|
D | axm5516-cpus.dtsi | 2 * arch/arm/boot/dts/axm5516-cpus.dtsi 13 cpus {
|
D | alphascale-asm9260.dtsi | 18 cpus {
|
D | vf500.dtsi | 15 cpus {
|
D | hisi-x5hd2-dkb.dts | 21 cpus {
|
D | omap34xx.dtsi | 16 cpus {
|
D | omap443x.dtsi | 14 cpus {
|
D | exynos4212.dtsi | 25 cpus {
|
D | dra74x.dtsi | 15 cpus {
|
D | xenvm-4.2.dts | 23 cpus {
|
D | exynos4412.dtsi | 25 cpus {
|
D | omap4460.dtsi | 13 cpus {
|
D | k2l.dtsi | 12 cpus {
|
D | k2hk.dtsi | 12 cpus {
|
D | pxa2xx.dtsi | 25 cpus {
|
D | imx6dl.dtsi | 20 cpus {
|
D | omap36xx.dtsi | 20 cpus {
|
D | stih418.dtsi | 13 cpus {
|
D | s3c2416.dtsi | 23 cpus {
|
D | ecx-2000.dts | 29 cpus {
|
D | bcm63138.dtsi | 20 cpus {
|
D | moxart.dtsi | 15 cpus {
|
D | bcm7445.dtsi | 16 cpus {
|
D | omap3-cm-t3x30.dtsi | 8 cpus {
|
D | k2e.dtsi | 12 cpus {
|
D | vt8500.dtsi | 14 cpus {
|
/linux-4.1.27/arch/cris/boot/dts/ |
D | etraxfs.dtsi | 6 cpus {
|
/linux-4.1.27/arch/arc/boot/dts/ |
D | skeleton.dtsi | 22 cpus {
|
/linux-4.1.27/arch/mips/boot/dts/ralink/ |
D | rt2880.dtsi | 6 cpus {
|
D | rt3883.dtsi | 6 cpus {
|
D | mt7620a.dtsi | 6 cpus {
|
D | rt3050.dtsi | 6 cpus {
|
/linux-4.1.27/arch/openrisc/boot/dts/ |
D | or1ksim.dts | 17 cpus {
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | rtas.h | 356 extern int rtas_online_cpus_mask(cpumask_var_t cpus); 357 extern int rtas_offline_cpus_mask(cpumask_var_t cpus);
|
/linux-4.1.27/Documentation/devicetree/bindings/arm/msm/ |
D | qcom,idle-state.txt | 15 clocks. In addition to gating the clocks, QCOM cpus use this instruction as a 18 hierarchy to enter standby states, when all cpus are idle. An interrupt brings 51 voltages reduced, provided all cpus enter this state. Since the span of low
|
/linux-4.1.27/Documentation/cpu-freq/ |
D | cpu-drivers.txt | 125 SMP systems normally have same clock source for a group of cpus. For these the 127 routine must initialize policy->cpus with mask of all possible cpus (Online + 129 policy->related_cpus and will reset policy->cpus to carry only online cpus.
|
/linux-4.1.27/kernel/debug/kdb/ |
D | kdb_cmds | 26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
|
/linux-4.1.27/drivers/crypto/qat/qat_dh895xcc/ |
D | adf_isr.c | 117 unsigned int cpu, cpus = num_online_cpus(); in adf_request_irqs() local 131 cpu = ((accel_dev->accel_id * hw_data->num_banks) + i) % cpus; in adf_request_irqs()
|
D | adf_drv.c | 122 int cpus = num_online_cpus(); in adf_dev_configure() local 124 int instances = min(cpus, banks); in adf_dev_configure()
|
/linux-4.1.27/tools/perf/ |
D | builtin-stat.c | 168 return (evsel->cpus && !target.cpu_list) ? evsel->cpus : evsel_list->cpus; in perf_evsel__cpus() 401 struct cpu_map *cpus = perf_evsel__cpus(counter); in check_per_pkg() local 409 if (cpu_map__empty(cpus)) in check_per_pkg() 420 s = cpu_map__get_socket(cpus, cpu); in check_per_pkg() 603 const int ncpus = cpu_map__nr(evsel_list->cpus), in handle_initial_delay() 1517 if (cpu_map__build_socket_map(evsel_list->cpus, &aggr_map)) { in perf_stat_init_aggr_mode() 1524 if (cpu_map__build_core_map(evsel_list->cpus, &aggr_map)) { in perf_stat_init_aggr_mode()
|
/linux-4.1.27/scripts/dtc/ |
D | livetree.c | 553 struct node *cpus, *bootcpu; in guess_boot_cpuid() local 556 cpus = get_node_by_path(tree, "/cpus"); in guess_boot_cpuid() 557 if (!cpus) in guess_boot_cpuid() 561 bootcpu = cpus->children; in guess_boot_cpuid()
|
/linux-4.1.27/Documentation/scheduler/ |
D | sched-stats.txt | 16 cpus on the machine, while domain0 is the most tightly focused domain, 17 sometimes balancing only between pairs of cpus. At this time, there 19 field in the domain stats is a bit map indicating which cpus are affected 65 The first field is a bit mask indicating what cpus this domain operates over.
|
/linux-4.1.27/drivers/bus/ |
D | arm-cci.c | 127 cpumask_t cpus; member 758 cpu = cpumask_first(&cci_pmu->cpus); in cci_pmu_event_init() 786 cpumask_pr_args(&pmu->cpus)); in pmu_attr_cpumask_show() 839 if (!cpumask_test_and_clear_cpu(cpu, &pmu->cpus)) in cci_pmu_cpu_notifier() 848 cpumask_set_cpu(target, &pmu->cpus); in cci_pmu_cpu_notifier() 986 cpumask_set_cpu(smp_processor_id(), &pmu->cpus); in cci_pmu_probe()
|
/linux-4.1.27/arch/mips/boot/dts/brcm/ |
D | bcm6328.dtsi | 6 cpus {
|
D | bcm6368.dtsi | 6 cpus {
|
D | bcm3384_viper.dtsi | 14 cpus {
|
D | bcm3384_zephyr.dtsi | 13 cpus {
|
D | bcm7125.dtsi | 6 cpus {
|
D | bcm7360.dtsi | 6 cpus {
|
D | bcm7358.dtsi | 6 cpus {
|
/linux-4.1.27/arch/tile/kernel/ |
D | setup.c | 887 int cpu, node, cpus, i, x, y; in setup_numa_mapping() local 923 cpus = cpu; in setup_numa_mapping() 937 for (cpu = 0; cpu < cpus; ++cpu) { in setup_numa_mapping() 1327 int cpus, i, rc; in setup_cpu_maps() local 1349 cpus = 1; /* this cpu */ in setup_cpu_maps() 1351 for (i = 0; cpus < setup_max_cpus; ++i) in setup_cpu_maps() 1353 ++cpus; in setup_cpu_maps()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | amd_nb.h | 46 atomic_t cpus; member
|
/linux-4.1.27/arch/arm/include/asm/ |
D | mcpm.h | 264 } cpus[MAX_CPUS_PER_CLUSTER]; member
|
/linux-4.1.27/drivers/edac/ |
D | cpc925_edac.c | 597 struct device_node *cpus; in cpc925_cpu_mask_disabled() local 607 cpus = of_find_node_by_path("/cpus"); in cpc925_cpu_mask_disabled() 608 if (cpus == NULL) { in cpc925_cpu_mask_disabled() 613 while ((cpunode = of_get_next_child(cpus, cpunode)) != NULL) { in cpc925_cpu_mask_disabled() 638 of_node_put(cpus); in cpc925_cpu_mask_disabled()
|
/linux-4.1.27/scripts/gdb/linux/ |
D | modules.py | 16 from linux import cpus, utils
|
/linux-4.1.27/arch/c6x/boot/dts/ |
D | tms320c6457.dtsi | 6 cpus {
|
D | tms320c6474.dtsi | 6 cpus {
|
D | tms320c6455.dtsi | 6 cpus {
|
D | tms320c6472.dtsi | 6 cpus {
|
D | tms320c6678.dtsi | 6 cpus {
|
/linux-4.1.27/arch/arm/plat-samsung/include/plat/ |
D | cpu.h | 109 struct cpu_table *cpus, unsigned int cputab_size);
|
/linux-4.1.27/tools/power/x86/turbostat/ |
D | turbostat.c | 2740 } *cpus; in topology_probe() local 2752 cpus = calloc(1, (topo.max_cpu_num + 1) * sizeof(struct cpu_topology)); in topology_probe() 2753 if (cpus == NULL) in topology_probe() 2788 cpus[i].core_id = get_core_id(i); in topology_probe() 2789 if (cpus[i].core_id > max_core_id) in topology_probe() 2790 max_core_id = cpus[i].core_id; in topology_probe() 2792 cpus[i].physical_package_id = get_physical_package_id(i); in topology_probe() 2793 if (cpus[i].physical_package_id > max_package_id) in topology_probe() 2794 max_package_id = cpus[i].physical_package_id; in topology_probe() 2801 i, cpus[i].physical_package_id, cpus[i].core_id); in topology_probe() [all …]
|
/linux-4.1.27/arch/s390/kernel/ |
D | time.c | 712 atomic_t cpus; member 721 atomic_dec(&sync->cpus); in clock_sync_cpu() 770 while (atomic_read(&etr_sync->cpus) != 0) in etr_sync_clock() 833 atomic_set(&etr_sync.cpus, num_online_cpus() - 1); in etr_sync_clock_stop() 1542 while (atomic_read(&stp_sync->cpus) != 0) in stp_sync_clock() 1603 atomic_set(&stp_sync.cpus, num_online_cpus() - 1); in stp_work_fn()
|
D | crash_dump.c | 524 int i, cpus = 0; in get_cpu_cnt() local 529 cpus++; in get_cpu_cnt() 531 return cpus; in get_cpu_cnt()
|
/linux-4.1.27/arch/mips/boot/dts/lantiq/ |
D | danube.dtsi | 6 cpus {
|
/linux-4.1.27/Documentation/power/ |
D | suspend-and-cpuhotplug.txt | 22 |tasks | | cpus | | | | cpus | |tasks| 70 Note down these cpus in | P 93 | Call _cpu_up() [for all those cpus in the frozen_cpus mask, in a loop]
|
/linux-4.1.27/kernel/power/ |
D | power.h | 11 int cpus; member
|
/linux-4.1.27/arch/mn10300/kernel/ |
D | smp.c | 395 int cpus, ret = 0; in smp_nmi_call_function() local 397 cpus = num_online_cpus() - 1; in smp_nmi_call_function() 398 if (cpus < 1) in smp_nmi_call_function()
|
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/ |
D | mce_amd.c | 596 atomic_inc(&b->cpus); in threshold_create_bank() 619 atomic_set(&b->cpus, 1); in threshold_create_bank() 708 if (!atomic_dec_and_test(&b->cpus)) { in threshold_remove_bank()
|
/linux-4.1.27/include/trace/events/ |
D | xen.h | 407 TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm, 409 TP_ARGS(cpus, mm, addr, end), 416 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
|
/linux-4.1.27/arch/arm64/boot/dts/sprd/ |
D | sc9836.dtsi | 15 cpus {
|
/linux-4.1.27/arch/xtensa/boot/dts/ |
D | xtfpga.dtsi | 16 cpus {
|
/linux-4.1.27/arch/arm64/boot/dts/arm/ |
D | rtsm_ve-aemv8a.dts | 30 cpus {
|
/linux-4.1.27/kernel/sched/ |
D | fair.c | 146 unsigned int cpus = min_t(int, num_online_cpus(), 8); in get_update_sysctl_factor() local 154 factor = cpus; in get_update_sysctl_factor() 158 factor = 1 + ilog2(cpus); in get_update_sysctl_factor() 1135 int smt, cpu, cpus = 0; in update_numa_stats() local 1146 cpus++; in update_numa_stats() 1157 if (!cpus) in update_numa_stats() 1161 smt = DIV_ROUND_UP(SCHED_CAPACITY_SCALE * cpus, ns->compute_capacity); in update_numa_stats() 1162 capacity = cpus / smt; /* cores */ in update_numa_stats() 5427 struct cpumask *cpus; member 5599 for_each_cpu_and(cpu, env->dst_grpmask, env->cpus) { in can_migrate_task() [all …]
|
/linux-4.1.27/arch/powerpc/boot/dts/fsl/ |
D | bsc9131si-pre.dtsi | 51 cpus {
|
D | mpc8536si-pre.dtsi | 56 cpus {
|
D | mpc8544si-pre.dtsi | 56 cpus {
|
D | p1021si-pre.dtsi | 55 cpus {
|
D | bsc9132si-pre.dtsi | 50 cpus {
|
D | mpc8572si-pre.dtsi | 57 cpus {
|
D | mpc8548si-pre.dtsi | 57 cpus {
|
D | c293si-pre.dtsi | 53 cpus {
|