cpus 36 arch/arm/common/mcpm_entry.c mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; cpus 37 arch/arm/common/mcpm_entry.c sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); cpus 50 arch/arm/common/mcpm_entry.c mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; cpus 51 arch/arm/common/mcpm_entry.c sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); cpus 103 arch/arm/common/mcpm_entry.c sync_cache_r(&c->cpus); cpus 111 arch/arm/common/mcpm_entry.c cpustate = c->cpus[i].cpu; cpus 116 arch/arm/common/mcpm_entry.c sync_cache_r(&c->cpus[i].cpu); cpus 439 arch/arm/common/mcpm_entry.c mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN; cpus 445 arch/arm/common/mcpm_entry.c mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP; cpus 288 arch/arm/include/asm/mcpm.h } cpus[MAX_CPUS_PER_CLUSTER]; cpus 74 arch/arm/kernel/devtree.c struct device_node *cpu, *cpus; cpus 81 arch/arm/kernel/devtree.c cpus = of_find_node_by_path("/cpus"); cpus 83 arch/arm/kernel/devtree.c if (!cpus) cpus 167 arch/arm/kernel/devtree.c set_smp_ops_by_method(cpus); cpus 761 arch/arm/kernel/smp.c struct cpumask *cpus = freq->policy->cpus; cpus 762 arch/arm/kernel/smp.c int cpu, first = cpumask_first(cpus); cpus 769 arch/arm/kernel/smp.c for_each_cpu(cpu, cpus) { cpus 789 arch/arm/kernel/smp.c for_each_cpu(cpu, cpus) cpus 105 arch/arm/plat-samsung/include/plat/cpu.h struct cpu_table *cpus, unsigned int cputab_size); cpus 1097 arch/arm64/kernel/cpufeature.c static const struct midr_range cpus[] = { cpus 1104 arch/arm64/kernel/cpufeature.c return is_midr_in_range_list(read_cpuid_id(), cpus); cpus 295 arch/ia64/kernel/smp.c cpumask_var_t cpus; cpus 304 arch/ia64/kernel/smp.c if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) { cpus 308 arch/ia64/kernel/smp.c cpumask_copy(cpus, mm_cpumask(mm)); cpus 309 arch/ia64/kernel/smp.c smp_call_function_many(cpus, cpus 311 arch/ia64/kernel/smp.c free_cpumask_var(cpus); cpus 115 arch/ia64/mm/discontig.c unsigned long pernodesize = 0, cpus; cpus 117 arch/ia64/mm/discontig.c cpus = early_nr_cpus_node(node); cpus 118 arch/ia64/mm/discontig.c pernodesize += PERCPU_PAGE_SIZE * cpus; cpus 263 arch/ia64/mm/discontig.c int cpus = early_nr_cpus_node(node); cpus 270 arch/ia64/mm/discontig.c pernode += PERCPU_PAGE_SIZE * cpus; cpus 142 arch/mips/cavium-octeon/smp.c int cpus; cpus 161 arch/mips/cavium-octeon/smp.c cpus = 1; cpus 164 arch/mips/cavium-octeon/smp.c set_cpu_possible(cpus, true); cpus 165 arch/mips/cavium-octeon/smp.c set_cpu_present(cpus, true); cpus 166 arch/mips/cavium-octeon/smp.c __cpu_number_map[id] = cpus; cpus 167 arch/mips/cavium-octeon/smp.c __cpu_logical_map[cpus] = id; cpus 168 arch/mips/cavium-octeon/smp.c cpus++; cpus 181 arch/mips/cavium-octeon/smp.c set_cpu_possible(cpus, true); cpus 182 arch/mips/cavium-octeon/smp.c __cpu_number_map[id] = cpus; cpus 183 arch/mips/cavium-octeon/smp.c __cpu_logical_map[cpus] = id; cpus 184 arch/mips/cavium-octeon/smp.c cpus++; cpus 1023 arch/nds32/kernel/perf_event_cpu.c static DEVICE_ATTR(cpus, 0444, nds32_pmu_cpumask_show, NULL); cpus 371 arch/powerpc/include/asm/rtas.h extern int rtas_online_cpus_mask(cpumask_var_t cpus); cpus 372 arch/powerpc/include/asm/rtas.h extern int rtas_offline_cpus_mask(cpumask_var_t cpus); cpus 852 arch/powerpc/kernel/rtas.c cpumask_var_t cpus) cpus 854 arch/powerpc/kernel/rtas.c if (!cpumask_empty(cpus)) { cpus 855 arch/powerpc/kernel/rtas.c cpumask_clear(cpus); cpus 865 arch/powerpc/kernel/rtas.c cpumask_var_t cpus) cpus 871 arch/powerpc/kernel/rtas.c if (cpumask_empty(cpus)) cpus 874 arch/powerpc/kernel/rtas.c for_each_cpu(cpu, cpus) { cpus 894 arch/powerpc/kernel/rtas.c cpumask_shift_right(cpus, cpus, cpu); cpus 895 arch/powerpc/kernel/rtas.c cpumask_shift_left(cpus, cpus, cpu); cpus 899 arch/powerpc/kernel/rtas.c cpumask_clear_cpu(cpu, cpus); cpus 909 arch/powerpc/kernel/rtas.c int rtas_online_cpus_mask(cpumask_var_t cpus) cpus 913 arch/powerpc/kernel/rtas.c ret = rtas_cpu_state_change_mask(UP, cpus); cpus 922 arch/powerpc/kernel/rtas.c cpumask_copy(tmp_mask, cpus); cpus 930 arch/powerpc/kernel/rtas.c int rtas_offline_cpus_mask(cpumask_var_t cpus) cpus 932 arch/powerpc/kernel/rtas.c return rtas_cpu_state_change_mask(DOWN, cpus); cpus 355 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpus 358 arch/powerpc/mm/book3s64/radix_tlb.c on_each_cpu_mask(cpus, do_tlbiel_pid, &t, 1); cpus 494 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpus 496 arch/powerpc/mm/book3s64/radix_tlb.c on_each_cpu_mask(cpus, do_tlbiel_va, &t, 1); cpus 545 arch/powerpc/mm/book3s64/radix_tlb.c struct cpumask *cpus = mm_cpumask(mm); cpus 550 arch/powerpc/mm/book3s64/radix_tlb.c on_each_cpu_mask(cpus, do_tlbiel_va_range, &t, 1); cpus 94 arch/powerpc/platforms/cell/cpufreq_spudemand.c for_each_cpu(i, policy->cpus) { cpus 117 arch/powerpc/platforms/cell/cpufreq_spudemand.c for_each_cpu (i, policy->cpus) { cpus 632 arch/powerpc/platforms/powermac/smp.c struct device_node *cpus; cpus 635 arch/powerpc/platforms/powermac/smp.c cpus = of_find_node_by_path("/cpus"); cpus 636 arch/powerpc/platforms/powermac/smp.c BUG_ON(cpus == NULL); cpus 639 arch/powerpc/platforms/powermac/smp.c pmf_call_function(cpus, "cpu-timebase", &args); cpus 640 arch/powerpc/platforms/powermac/smp.c of_node_put(cpus); cpus 710 arch/powerpc/platforms/powermac/smp.c struct device_node *cpus = cpus 712 arch/powerpc/platforms/powermac/smp.c if (cpus && cpus 713 arch/powerpc/platforms/powermac/smp.c of_get_property(cpus, "platform-cpu-timebase", NULL)) { cpus 766 arch/powerpc/platforms/powermac/smp.c struct device_node *cpus; cpus 772 arch/powerpc/platforms/powermac/smp.c for_each_node_by_type(cpus, "cpu") cpus 92 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_part_hdr *)hdr)->cpus; cpus 174 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_phys_hdr *)hdr)->cpus; cpus 176 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_x_phys_hdr *)hdr)->cpus; cpus 131 arch/s390/include/asm/diag.h __u8 cpus; cpus 138 arch/s390/include/asm/diag.h __u8 cpus; cpus 190 arch/s390/include/asm/diag.h __u8 cpus; cpus 197 arch/s390/include/asm/diag.h __u8 cpus; cpus 224 arch/s390/include/asm/diag.h struct diag204_x_cpu_info cpus[]; cpus 229 arch/s390/include/asm/diag.h struct diag204_x_phys_cpu cpus[]; cpus 536 arch/s390/kernel/crash_dump.c int cpus = 0; cpus 540 arch/s390/kernel/crash_dump.c cpus++; cpus 541 arch/s390/kernel/crash_dump.c return cpus; cpus 234 arch/s390/kernel/sthyi.c for (i = 0; i < block->hdr.cpus; i++) { cpus 235 arch/s390/kernel/sthyi.c switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) { cpus 237 arch/s390/kernel/sthyi.c if (block->cpus[i].weight == DED_WEIGHT) cpus 243 arch/s390/kernel/sthyi.c if (block->cpus[i].weight == DED_WEIGHT) cpus 263 arch/s390/kernel/sthyi.c if (!(block->cpus[i].cflag & DIAG204_CPU_ONLINE)) cpus 266 arch/s390/kernel/sthyi.c switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) { cpus 269 arch/s390/kernel/sthyi.c if (block->cpus[i].cur_weight < DED_WEIGHT) cpus 270 arch/s390/kernel/sthyi.c weight_cp |= block->cpus[i].cur_weight; cpus 274 arch/s390/kernel/sthyi.c if (block->cpus[i].cur_weight < DED_WEIGHT) cpus 275 arch/s390/kernel/sthyi.c weight_ifl |= block->cpus[i].cur_weight; cpus 284 arch/s390/kernel/sthyi.c capped |= block->cpus[i].cflag & DIAG204_CPU_CAPPED; cpus 285 arch/s390/kernel/sthyi.c cpu_inf->lpar_cap |= block->cpus[i].cpu_type_cap; cpus 286 arch/s390/kernel/sthyi.c cpu_inf->lpar_grp_cap |= block->cpus[i].group_cpu_type_cap; cpus 288 arch/s390/kernel/sthyi.c if (block->cpus[i].weight == DED_WEIGHT) cpus 300 arch/s390/kernel/sthyi.c return (struct diag204_x_part_block *)&block->cpus[i]; cpus 490 arch/s390/kernel/time.c atomic_t cpus; cpus 604 arch/s390/kernel/time.c while (atomic_read(&sync->cpus) != 0) cpus 625 arch/s390/kernel/time.c atomic_dec(&sync->cpus); cpus 671 arch/s390/kernel/time.c atomic_set(&stp_sync.cpus, num_online_cpus() - 1); cpus 808 arch/s390/kvm/priv.c int cpus = 0; cpus 811 arch/s390/kvm/priv.c cpus = atomic_read(&vcpu->kvm->online_vcpus); cpus 822 arch/s390/kvm/priv.c mem->vm[0].cpus_total = cpus; cpus 823 arch/s390/kvm/priv.c mem->vm[0].cpus_configured = cpus; cpus 659 arch/sparc/kernel/time_64.c for_each_cpu(cpu, freq->policy->cpus) { cpus 20 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus, cpus 55 arch/x86/hyperv/mmu.c static void hyperv_flush_tlb_others(const struct cpumask *cpus, cpus 64 arch/x86/hyperv/mmu.c trace_hyperv_mmu_flush_tlb_others(cpus, info); cpus 69 arch/x86/hyperv/mmu.c if (cpumask_empty(cpus)) cpus 98 arch/x86/hyperv/mmu.c if (cpumask_equal(cpus, cpu_present_mask)) { cpus 112 arch/x86/hyperv/mmu.c if (hv_cpu_number_to_vp_number(cpumask_last(cpus)) >= 64) cpus 115 arch/x86/hyperv/mmu.c for_each_cpu(cpu, cpus) { cpus 153 arch/x86/hyperv/mmu.c status = hyperv_flush_tlb_others_ex(cpus, info); cpus 161 arch/x86/hyperv/mmu.c native_flush_tlb_others(cpus, info); cpus 164 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus, cpus 196 arch/x86/hyperv/mmu.c nr_bank = cpumask_to_vpset(&(flush->hv_vp_set), cpus); cpus 60 arch/x86/include/asm/amd_nb.h refcount_t cpus; cpus 209 arch/x86/include/asm/paravirt_types.h void (*flush_tlb_others)(const struct cpumask *cpus, cpus 12 arch/x86/include/asm/trace/hyperv.h TP_PROTO(const struct cpumask *cpus, cpus 14 arch/x86/include/asm/trace/hyperv.h TP_ARGS(cpus, info), cpus 21 arch/x86/include/asm/trace/hyperv.h TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); cpus 60 arch/x86/include/asm/trace/hyperv.h TP_PROTO(const struct cpumask *cpus, cpus 62 arch/x86/include/asm/trace/hyperv.h TP_ARGS(cpus, vector), cpus 67 arch/x86/include/asm/trace/hyperv.h TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); cpus 1326 arch/x86/kernel/cpu/mce/amd.c refcount_inc(&b->cpus); cpus 1347 arch/x86/kernel/cpu/mce/amd.c refcount_set(&b->cpus, 1); cpus 1417 arch/x86/kernel/cpu/mce/amd.c if (!refcount_dec_and_test(&b->cpus)) { cpus 962 arch/x86/kernel/cpu/mce/core.c int cpus = num_online_cpus(); cpus 978 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_callin) != cpus) { cpus 1043 arch/x86/kernel/cpu/mce/core.c int cpus = num_online_cpus(); cpus 1049 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_executing) <= cpus) { cpus 312 arch/x86/kernel/tsc_sync.c int cpus = 2; cpus 334 arch/x86/kernel/tsc_sync.c while (atomic_read(&start_count) != cpus - 1) { cpus 349 arch/x86/kernel/tsc_sync.c while (atomic_read(&stop_count) != cpus-1) cpus 405 arch/x86/kernel/tsc_sync.c int cpus = 2; cpus 431 arch/x86/kernel/tsc_sync.c while (atomic_read(&start_count) != cpus) cpus 449 arch/x86/kernel/tsc_sync.c while (atomic_read(&stop_count) != cpus) cpus 7113 arch/x86/kvm/x86.c for_each_cpu(cpu, freq->policy->cpus) cpus 2196 arch/x86/platform/uv/tlb_uv.c int cpus; cpus 2225 arch/x86/platform/uv/tlb_uv.c cpus = uv_blade_nr_possible_cpus(uvhub); cpus 2226 arch/x86/platform/uv/tlb_uv.c if (cpus && (uv_blade_to_pnode(uvhub) < uv_base_pnode)) cpus 1348 arch/x86/xen/mmu_pv.c static void xen_flush_tlb_others(const struct cpumask *cpus, cpus 1359 arch/x86/xen/mmu_pv.c trace_xen_mmu_flush_tlb_others(cpus, info->mm, info->start, info->end); cpus 1361 arch/x86/xen/mmu_pv.c if (cpumask_empty(cpus)) cpus 1369 arch/x86/xen/mmu_pv.c cpumask_and(to_cpumask(args->mask), cpus, cpu_online_mask); cpus 26 drivers/base/arch_topology.c void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpus 34 drivers/base/arch_topology.c for_each_cpu(i, cpus) cpus 1466 drivers/base/power/domain.c cpumask_set_cpu(cpu, genpd->cpus); cpus 1468 drivers/base/power/domain.c cpumask_clear_cpu(cpu, genpd->cpus); cpus 1820 drivers/base/power/domain.c !zalloc_cpumask_var(&genpd->cpus, GFP_KERNEL)) cpus 1828 drivers/base/power/domain.c free_cpumask_var(genpd->cpus); cpus 1877 drivers/base/power/domain.c free_cpumask_var(genpd->cpus); cpus 273 drivers/base/power/domain_governor.c for_each_cpu_and(cpu, genpd->cpus, cpu_online_mask) { cpus 55 drivers/clk/sunxi/clk-sun9i-cpus.c struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw); cpus 60 drivers/clk/sunxi/clk-sun9i-cpus.c reg = readl(cpus->reg); cpus 155 drivers/clk/sunxi/clk-sun9i-cpus.c struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw); cpus 162 drivers/clk/sunxi/clk-sun9i-cpus.c reg = readl(cpus->reg); cpus 170 drivers/clk/sunxi/clk-sun9i-cpus.c writel(reg, cpus->reg); cpus 188 drivers/clk/sunxi/clk-sun9i-cpus.c struct sun9i_a80_cpus_clk *cpus; cpus 193 drivers/clk/sunxi/clk-sun9i-cpus.c cpus = kzalloc(sizeof(*cpus), GFP_KERNEL); cpus 194 drivers/clk/sunxi/clk-sun9i-cpus.c if (!cpus) cpus 197 drivers/clk/sunxi/clk-sun9i-cpus.c cpus->reg = of_io_request_and_map(node, 0, of_node_full_name(node)); cpus 198 drivers/clk/sunxi/clk-sun9i-cpus.c if (IS_ERR(cpus->reg)) cpus 211 drivers/clk/sunxi/clk-sun9i-cpus.c mux->reg = cpus->reg; cpus 219 drivers/clk/sunxi/clk-sun9i-cpus.c &cpus->hw, &sun9i_a80_cpus_clk_ops, cpus 235 drivers/clk/sunxi/clk-sun9i-cpus.c iounmap(cpus->reg); cpus 239 drivers/clk/sunxi/clk-sun9i-cpus.c kfree(cpus); cpus 437 drivers/cpufreq/acpi-cpufreq.c cpumask_of(policy->cpu) : policy->cpus; cpus 680 drivers/cpufreq/acpi-cpufreq.c cpumask_copy(policy->cpus, perf->shared_cpu_map); cpus 688 drivers/cpufreq/acpi-cpufreq.c cpumask_copy(policy->cpus, topology_core_cpumask(cpu)); cpus 692 drivers/cpufreq/acpi-cpufreq.c cpumask_clear(policy->cpus); cpus 693 drivers/cpufreq/acpi-cpufreq.c cpumask_set_cpu(cpu, policy->cpus); cpus 473 drivers/cpufreq/arm_big_little.c cpumask_copy(policy->cpus, topology_core_cpumask(policy->cpu)); cpus 475 drivers/cpufreq/arm_big_little.c for_each_cpu(cpu, policy->cpus) cpus 482 drivers/cpufreq/arm_big_little.c ret = get_cluster_clk_and_freq_table(cpu_dev, policy->cpus); cpus 490 drivers/cpufreq/arm_big_little.c dev_pm_opp_of_register_em(policy->cpus); cpus 128 drivers/cpufreq/armada-8k-cpufreq.c struct cpumask cpus; cpus 141 drivers/cpufreq/armada-8k-cpufreq.c cpumask_copy(&cpus, cpu_possible_mask); cpus 148 drivers/cpufreq/armada-8k-cpufreq.c for_each_cpu(cpu, &cpus) { cpus 178 drivers/cpufreq/armada-8k-cpufreq.c cpumask_andnot(&cpus, &cpus, &shared_cpus); cpus 592 drivers/cpufreq/brcmstb-avs-cpufreq.c cpumask_setall(policy->cpus); cpus 325 drivers/cpufreq/cppc_cpufreq.c cpumask_copy(policy->cpus, cpu->shared_cpu_map); cpus 327 drivers/cpufreq/cppc_cpufreq.c for_each_cpu(i, policy->cpus) { cpus 174 drivers/cpufreq/cpufreq-dt.c ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, policy->cpus); cpus 184 drivers/cpufreq/cpufreq-dt.c if (dev_pm_opp_get_sharing_cpus(cpu_dev, policy->cpus)) cpus 222 drivers/cpufreq/cpufreq-dt.c if (!dev_pm_opp_of_cpumask_add_table(policy->cpus)) cpus 237 drivers/cpufreq/cpufreq-dt.c cpumask_setall(policy->cpus); cpus 243 drivers/cpufreq/cpufreq-dt.c ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus); cpus 278 drivers/cpufreq/cpufreq-dt.c dev_pm_opp_of_register_em(policy->cpus); cpus 286 drivers/cpufreq/cpufreq-dt.c dev_pm_opp_of_cpumask_remove_table(policy->cpus); cpus 152 drivers/cpufreq/cpufreq.c __weak void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpus 176 drivers/cpufreq/cpufreq.c cpumask_setall(policy->cpus); cpus 184 drivers/cpufreq/cpufreq.c return policy && cpumask_test_cpu(cpu, policy->cpus) ? policy : NULL; cpus 376 drivers/cpufreq/cpufreq.c cpumask_pr_args(policy->cpus)); cpus 378 drivers/cpufreq/cpufreq.c for_each_cpu(cpu, policy->cpus) cpus 855 drivers/cpufreq/cpufreq.c return cpufreq_show_cpus(policy->cpus, buf); cpus 1095 drivers/cpufreq/cpufreq.c if (cpumask_test_cpu(cpu, policy->cpus)) cpus 1102 drivers/cpufreq/cpufreq.c cpumask_set_cpu(cpu, policy->cpus); cpus 1187 drivers/cpufreq/cpufreq.c if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL)) cpus 1218 drivers/cpufreq/cpufreq.c ret, cpumask_pr_args(policy->cpus)); cpus 1226 drivers/cpufreq/cpufreq.c ret, cpumask_pr_args(policy->cpus)); cpus 1250 drivers/cpufreq/cpufreq.c free_cpumask_var(policy->cpus); cpus 1294 drivers/cpufreq/cpufreq.c free_cpumask_var(policy->cpus); cpus 1337 drivers/cpufreq/cpufreq.c cpumask_copy(policy->cpus, policy->related_cpus); cpus 1339 drivers/cpufreq/cpufreq.c cpumask_copy(policy->cpus, cpumask_of(cpu)); cpus 1357 drivers/cpufreq/cpufreq.c cpumask_copy(policy->related_cpus, policy->cpus); cpus 1365 drivers/cpufreq/cpufreq.c cpumask_and(policy->cpus, policy->cpus, cpu_online_mask); cpus 1551 drivers/cpufreq/cpufreq.c cpumask_clear_cpu(cpu, policy->cpus); cpus 1561 drivers/cpufreq/cpufreq.c policy->cpu = cpumask_any(policy->cpus); cpus 102 drivers/cpufreq/cpufreq_governor.c for_each_cpu(j, policy_dbs->policy->cpus) { cpus 137 drivers/cpufreq/cpufreq_governor.c for_each_cpu(j, policy->cpus) { cpus 331 drivers/cpufreq/cpufreq_governor.c for_each_cpu(cpu, policy->cpus) { cpus 343 drivers/cpufreq/cpufreq_governor.c for_each_cpu(i, policy->cpus) cpus 523 drivers/cpufreq/cpufreq_governor.c for_each_cpu(j, policy->cpus) { cpus 439 drivers/cpufreq/cpufreq_ondemand.c cpumask_or(&done, &done, policy->cpus); cpus 196 drivers/cpufreq/imx6q-cpufreq.c dev_pm_opp_of_register_em(policy->cpus); cpus 36 drivers/cpufreq/mediatek-cpufreq.c struct cpumask cpus; cpus 54 drivers/cpufreq/mediatek-cpufreq.c if (cpumask_test_cpu(cpu, &info->cpus)) cpus 358 drivers/cpufreq/mediatek-cpufreq.c ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, &info->cpus); cpus 365 drivers/cpufreq/mediatek-cpufreq.c ret = dev_pm_opp_of_cpumask_add_table(&info->cpus); cpus 397 drivers/cpufreq/mediatek-cpufreq.c dev_pm_opp_of_cpumask_remove_table(&info->cpus); cpus 423 drivers/cpufreq/mediatek-cpufreq.c dev_pm_opp_of_cpumask_remove_table(&info->cpus); cpus 446 drivers/cpufreq/mediatek-cpufreq.c cpumask_copy(policy->cpus, &info->cpus); cpus 451 drivers/cpufreq/mediatek-cpufreq.c dev_pm_opp_of_register_em(policy->cpus); cpus 134 drivers/cpufreq/omap-cpufreq.c dev_pm_opp_of_register_em(policy->cpus); cpus 111 drivers/cpufreq/p4-clockmod.c for_each_cpu(i, policy->cpus) cpus 162 drivers/cpufreq/p4-clockmod.c cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu)); cpus 1074 drivers/cpufreq/powernow-k8.c cpumask_copy(pol->cpus, topology_core_cpumask(pol->cpu)); cpus 1075 drivers/cpufreq/powernow-k8.c data->available_cores = pol->cpus; cpus 1082 drivers/cpufreq/powernow-k8.c for_each_cpu(cpu, pol->cpus) cpus 1107 drivers/cpufreq/powernow-k8.c for_each_cpu(cpu, pol->cpus) cpus 685 drivers/cpufreq/powernv-cpufreq.c if (!cpumask_test_cpu(raw_smp_processor_id(), policy->cpus)) { cpus 687 drivers/cpufreq/powernv-cpufreq.c add_timer_on(&gpstates->timer, cpumask_first(policy->cpus)); cpus 822 drivers/cpufreq/powernv-cpufreq.c smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1); cpus 835 drivers/cpufreq/powernv-cpufreq.c cpumask_set_cpu(base + i, policy->cpus); cpus 924 drivers/cpufreq/powernv-cpufreq.c cpumask_andnot(&mask, &mask, policy.cpus); cpus 109 drivers/cpufreq/ppc_cbe_cpufreq.c cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu)); cpus 147 drivers/cpufreq/qcom-cpufreq-hw.c dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus); cpus 219 drivers/cpufreq/qcom-cpufreq-hw.c qcom_get_related_cpus(index, policy->cpus); cpus 220 drivers/cpufreq/qcom-cpufreq-hw.c if (!cpumask_weight(policy->cpus)) { cpus 241 drivers/cpufreq/qcom-cpufreq-hw.c dev_pm_opp_of_register_em(policy->cpus); cpus 93 drivers/cpufreq/qoriq-cpufreq.c struct cpumask *dstp = policy->cpus; cpus 154 drivers/cpufreq/scmi-cpufreq.c ret = scmi_get_sharing_cpus(cpu_dev, policy->cpus); cpus 160 drivers/cpufreq/scmi-cpufreq.c ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus); cpus 203 drivers/cpufreq/scmi-cpufreq.c em_register_perf_domain(policy->cpus, nr_opp, &em_cb); cpus 117 drivers/cpufreq/scpi-cpufreq.c ret = scpi_get_sharing_cpus(cpu_dev, policy->cpus); cpus 123 drivers/cpufreq/scpi-cpufreq.c ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus); cpus 172 drivers/cpufreq/scpi-cpufreq.c dev_pm_opp_of_register_em(policy->cpus); cpus 440 drivers/cpufreq/speedstep-centrino.c for_each_cpu(j, policy->cpus) { cpus 448 drivers/cpufreq/speedstep-centrino.c good_cpu = cpumask_any_and(policy->cpus, cpus 263 drivers/cpufreq/speedstep-ich.c policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); cpus 296 drivers/cpufreq/speedstep-ich.c cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu)); cpus 298 drivers/cpufreq/speedstep-ich.c policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); cpus 21 drivers/cpufreq/tegra186-cpufreq.c int cpus[4]; cpus 30 drivers/cpufreq/tegra186-cpufreq.c .cpus = { 1, 2, NO_CPU, NO_CPU }, cpus 36 drivers/cpufreq/tegra186-cpufreq.c .cpus = { 0, 3, 4, 5 }, cpus 64 drivers/cpufreq/tegra186-cpufreq.c for (core = 0; core < ARRAY_SIZE(info->cpus); core++) { cpus 65 drivers/cpufreq/tegra186-cpufreq.c if (info->cpus[core] == policy->cpu) cpus 68 drivers/cpufreq/tegra186-cpufreq.c if (core == ARRAY_SIZE(info->cpus)) cpus 441 drivers/cpuidle/coupled.c cpumask_t cpus; cpus 444 drivers/cpuidle/coupled.c cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); cpus 445 drivers/cpuidle/coupled.c ret = cpumask_and(&cpus, &cpuidle_coupled_poke_pending, &cpus); cpus 628 drivers/cpuidle/coupled.c cpumask_t cpus; cpus 629 drivers/cpuidle/coupled.c cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus); cpus 630 drivers/cpuidle/coupled.c coupled->online_count = cpumask_weight(&cpus); cpus 399 drivers/crypto/caam/qi.c const cpumask_t *cpus = qman_affine_cpus(); cpus 431 drivers/crypto/caam/qi.c if (!cpumask_test_cpu(*cpu, cpus)) { cpus 434 drivers/crypto/caam/qi.c *pcpu = cpumask_next(*pcpu, cpus); cpus 436 drivers/crypto/caam/qi.c *pcpu = cpumask_first(cpus); cpus 507 drivers/crypto/caam/qi.c const cpumask_t *cpus = qman_affine_cpus(); cpus 509 drivers/crypto/caam/qi.c for_each_cpu(i, cpus) { cpus 684 drivers/crypto/caam/qi.c const cpumask_t *cpus = qman_affine_cpus(); cpus 687 drivers/crypto/caam/qi.c for_each_cpu(i, cpus) { cpus 701 drivers/crypto/caam/qi.c const cpumask_t *cpus = qman_affine_cpus(); cpus 703 drivers/crypto/caam/qi.c for_each_cpu(i, cpus) cpus 712 drivers/crypto/caam/qi.c const cpumask_t *cpus = qman_affine_cpus(); cpus 736 drivers/crypto/caam/qi.c for_each_cpu(i, cpus) { cpus 176 drivers/crypto/qat/qat_common/adf_isr.c unsigned int cpu, cpus = num_online_cpus(); cpus 191 drivers/crypto/qat/qat_common/adf_isr.c i) % cpus; cpus 162 drivers/crypto/qat/qat_common/qat_crypto.c int cpus = num_online_cpus(); cpus 164 drivers/crypto/qat/qat_common/qat_crypto.c int instances = min(cpus, banks); cpus 77 drivers/firmware/psci/psci_checker.c static unsigned int down_and_up_cpus(const struct cpumask *cpus, cpus 86 drivers/firmware/psci/psci_checker.c for_each_cpu(cpu, cpus) { cpus 336 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) cpus 347 drivers/firmware/qcom_scm-32.c if (!cpus || (cpus && cpumask_empty(cpus))) cpus 350 drivers/firmware/qcom_scm-32.c for_each_cpu(cpu, cpus) { cpus 370 drivers/firmware/qcom_scm-32.c const cpumask_t *cpus) cpus 384 drivers/firmware/qcom_scm-32.c for_each_cpu(cpu, cpus) { cpus 399 drivers/firmware/qcom_scm-32.c for_each_cpu(cpu, cpus) cpus 167 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) cpus 182 drivers/firmware/qcom_scm-64.c const cpumask_t *cpus) cpus 98 drivers/firmware/qcom_scm.c int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) cpus 100 drivers/firmware/qcom_scm.c return __qcom_scm_set_cold_boot_addr(entry, cpus); cpus 112 drivers/firmware/qcom_scm.c int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) cpus 114 drivers/firmware/qcom_scm.c return __qcom_scm_set_warm_boot_addr(__scm->dev, entry, cpus); cpus 19 drivers/firmware/qcom_scm.h const cpumask_t *cpus); cpus 20 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus); cpus 1109 drivers/gpu/drm/vc4/vc4_hdmi.c dai_link->cpus = &hdmi->audio.cpu; cpus 1120 drivers/gpu/drm/vc4/vc4_hdmi.c dai_link->cpus->dai_name = dev_name(dev); cpus 81 drivers/irqchip/irq-bcm6345-l1.c struct bcm6345_l1_cpu *cpus[NR_CPUS]; cpus 124 drivers/irqchip/irq-bcm6345-l1.c cpu = intc->cpus[cpu_logical_map(smp_processor_id())]; cpus 126 drivers/irqchip/irq-bcm6345-l1.c cpu = intc->cpus[0]; cpus 159 drivers/irqchip/irq-bcm6345-l1.c intc->cpus[cpu_idx]->enable_cache[word] |= mask; cpus 160 drivers/irqchip/irq-bcm6345-l1.c __raw_writel(intc->cpus[cpu_idx]->enable_cache[word], cpus 161 drivers/irqchip/irq-bcm6345-l1.c intc->cpus[cpu_idx]->map_base + reg_enable(intc, word)); cpus 171 drivers/irqchip/irq-bcm6345-l1.c intc->cpus[cpu_idx]->enable_cache[word] &= ~mask; cpus 172 drivers/irqchip/irq-bcm6345-l1.c __raw_writel(intc->cpus[cpu_idx]->enable_cache[word], cpus 173 drivers/irqchip/irq-bcm6345-l1.c intc->cpus[cpu_idx]->map_base + reg_enable(intc, word)); cpus 220 drivers/irqchip/irq-bcm6345-l1.c enabled = intc->cpus[old_cpu]->enable_cache[word] & mask; cpus 255 drivers/irqchip/irq-bcm6345-l1.c cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), cpus 340 drivers/irqchip/irq-bcm6345-l1.c struct bcm6345_l1_cpu *cpu = intc->cpus[idx]; cpus 350 drivers/irqchip/irq-bcm6345-l1.c struct bcm6345_l1_cpu *cpu = intc->cpus[idx]; cpus 41 drivers/irqchip/irq-bcm7038-l1.c struct bcm7038_l1_cpu *cpus[NR_CPUS]; cpus 122 drivers/irqchip/irq-bcm7038-l1.c cpu = intc->cpus[cpu_logical_map(smp_processor_id())]; cpus 124 drivers/irqchip/irq-bcm7038-l1.c cpu = intc->cpus[0]; cpus 154 drivers/irqchip/irq-bcm7038-l1.c intc->cpus[cpu_idx]->mask_cache[word] &= ~mask; cpus 155 drivers/irqchip/irq-bcm7038-l1.c l1_writel(mask, intc->cpus[cpu_idx]->map_base + cpus 165 drivers/irqchip/irq-bcm7038-l1.c intc->cpus[cpu_idx]->mask_cache[word] |= mask; cpus 166 drivers/irqchip/irq-bcm7038-l1.c l1_writel(mask, intc->cpus[cpu_idx]->map_base + cpus 204 drivers/irqchip/irq-bcm7038-l1.c was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] & cpus 265 drivers/irqchip/irq-bcm7038-l1.c cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), cpus 354 drivers/irqchip/irq-bcm7038-l1.c struct bcm7038_l1_cpu *cpu = intc->cpus[idx]; cpus 44 drivers/leds/trigger/ledtrig-activity.c int cpus; cpus 56 drivers/leds/trigger/ledtrig-activity.c cpus = 0; cpus 65 drivers/leds/trigger/ledtrig-activity.c cpus++; cpus 73 drivers/leds/trigger/ledtrig-activity.c curr_boot = ktime_get_boottime_ns() * cpus; cpus 130 drivers/leds/trigger/ledtrig-activity.c target = (cpus > 1) ? (100 / cpus) : 50; cpus 168 drivers/misc/sgi-gru/grufile.c info.cpus = num_online_cpus(); cpus 146 drivers/misc/sgi-gru/grulib.h int cpus; cpus 297 drivers/misc/sgi-gru/grutlbpurge.c int cpus, shift = 0, n; cpus 299 drivers/misc/sgi-gru/grutlbpurge.c cpus = uv_blade_nr_possible_cpus(gru->gs_blade_id); cpus 302 drivers/misc/sgi-gru/grutlbpurge.c if (cpus) { cpus 303 drivers/misc/sgi-gru/grutlbpurge.c n = 1 << fls(cpus - 1); cpus 316 drivers/misc/sgi-gru/grutlbpurge.c gru->gs_tgh_first_remote = (cpus + (1 << shift) - 1) >> shift; cpus 6317 drivers/net/ethernet/broadcom/bnx2.c int cpus = netif_get_num_default_rss_queues(); cpus 6321 drivers/net/ethernet/broadcom/bnx2.c msix_vecs = max(cpus + 1, bp->num_req_tx_rings); cpus 6323 drivers/net/ethernet/broadcom/bnx2.c msix_vecs = max(cpus, bp->num_req_rx_rings); cpus 699 drivers/net/ethernet/chelsio/cxgb3/common.h const u8 * cpus, const u16 *rspq); cpus 574 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c u8 cpus[SGE_QSETS + 1]; cpus 578 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cpus[i] = i; cpus 579 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c cpus[SGE_QSETS] = 0xff; /* terminator */ cpus 589 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c V_RRCPLCPUSIZE(6) | F_HASHTOEPLITZ, cpus, rspq_map); cpus 2535 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c const u8 * cpus, const u16 *rspq) cpus 2539 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c if (cpus) cpus 2544 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c val |= (cpus[cpu_idx++] & 0x3f) << (8 * j); cpus 2545 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c if (cpus[cpu_idx] == 0xff) cpus 779 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c const cpumask_t *cpus = qman_affine_cpus(); cpus 783 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c for_each_cpu_and(cpu, cpus, cpu_online_mask) { cpus 553 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c const cpumask_t *cpus = qman_affine_cpus(); cpus 572 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c for_each_cpu_and(cpu, cpus, cpu_online_mask) { cpus 589 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c for_each_cpu_and(cpu, cpus, cpu_online_mask) { cpus 1045 drivers/net/ethernet/freescale/enetc/enetc.c int cpus = num_online_cpus(); cpus 1054 drivers/net/ethernet/freescale/enetc/enetc.c priv->num_rx_rings = min_t(int, cpus, si->num_rx_rings); cpus 1056 drivers/net/ethernet/freescale/enetc/enetc.c priv->bdr_int_num = cpus; cpus 10928 drivers/net/ethernet/intel/i40e/i40e_main.c int cpus, extra_vectors; cpus 10968 drivers/net/ethernet/intel/i40e/i40e_main.c cpus = num_online_cpus(); cpus 10969 drivers/net/ethernet/intel/i40e/i40e_main.c pf->num_lan_msix = min_t(int, cpus, vectors_left / 2); cpus 11035 drivers/net/ethernet/intel/i40e/i40e_main.c extra_vectors = min_t(int, cpus - pf->num_lan_msix, vectors_left); cpus 1446 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c int nrxqs, cpu, cpus = num_possible_cpus(); cpus 1449 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c nrxqs = port->nrxqs / cpus; cpus 1460 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c return port->first_rxq + ((rxq * nrxqs + rxq / cpus) % port->nrxqs); cpus 1103 drivers/opp/of.c void dev_pm_opp_of_register_em(struct cpumask *cpus) cpus 1106 drivers/opp/of.c int ret, nr_opp, cpu = cpumask_first(cpus); cpus 1135 drivers/opp/of.c em_register_perf_domain(cpus, nr_opp, &em_cb); cpus 497 drivers/perf/arm_pmu.c static DEVICE_ATTR(cpus, S_IRUGO, armpmu_cpumask_show, NULL); cpus 435 drivers/thermal/cpu_cooling.c num_cpus = cpumask_weight(cpufreq_cdev->policy->cpus); cpus 892 fs/btrfs/compression.c int cpus = num_online_cpus(); cpus 916 fs/btrfs/compression.c if (atomic_read(total_ws) > cpus) { cpus 921 fs/btrfs/compression.c if (atomic_read(total_ws) > cpus && !*free_ws) cpus 132 include/asm-generic/mshyperv.h const struct cpumask *cpus) cpus 151 include/asm-generic/mshyperv.h for_each_cpu(cpu, cpus) { cpus 53 include/linux/cpufreq.h cpumask_var_t cpus; /* Online CPUs only */ cpus 196 include/linux/cpufreq.h return cpumask_empty(policy->cpus); cpus 201 include/linux/cpufreq.h return cpumask_weight(policy->cpus) > 1; cpus 991 include/linux/cpufreq.h extern void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq, cpus 40 include/linux/energy_model.h unsigned long cpus[0]; cpus 91 include/linux/energy_model.h cpu = cpumask_first(to_cpumask(pd->cpus)); cpus 667 include/linux/netdevice.h u16 cpus[0]; cpus 112 include/linux/pm_domain.h cpumask_var_t cpus; /* A cpumask of the attached CPUs */ cpus 350 include/linux/pm_opp.h void dev_pm_opp_of_register_em(struct cpumask *cpus); cpus 390 include/linux/pm_opp.h static inline void dev_pm_opp_of_register_em(struct cpumask *cpus) cpus 38 include/linux/qcom_scm.h extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus); cpus 39 include/linux/qcom_scm.h extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus); cpus 68 include/linux/qcom_scm.h int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) cpus 73 include/linux/qcom_scm.h int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) cpus 125 include/linux/stop_machine.h int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); cpus 136 include/linux/stop_machine.h int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); cpus 139 include/linux/stop_machine.h const struct cpumask *cpus); cpus 143 include/linux/stop_machine.h const struct cpumask *cpus) cpus 154 include/linux/stop_machine.h const struct cpumask *cpus) cpus 156 include/linux/stop_machine.h return stop_machine_cpuslocked(fn, data, cpus); cpus 160 include/linux/stop_machine.h const struct cpumask *cpus) cpus 162 include/linux/stop_machine.h return stop_machine(fn, data, cpus); cpus 46 include/sound/simple_card_utils.h struct snd_soc_dai_link_component cpus; /* single cpu */ cpus 85 include/sound/simple_card_utils.h asoc_simple_parse_clk(dev, node, simple_dai, dai_link->cpus) cpus 101 include/sound/simple_card_utils.h asoc_simple_parse_dai(node, dai_link->cpus, is_single_link) cpus 770 include/sound/soc.h struct snd_soc_dai_link_component *cpus; cpus 917 include/sound/soc.h .cpus = cpu, \ cpus 96 include/trace/events/thermal.h TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load, cpus 99 include/trace/events/thermal.h TP_ARGS(cpus, freq, load, load_len, dynamic_power), cpus 110 include/trace/events/thermal.h __assign_bitmask(cpumask, cpumask_bits(cpus), cpus 126 include/trace/events/thermal.h TP_PROTO(const struct cpumask *cpus, unsigned int freq, cpus 129 include/trace/events/thermal.h TP_ARGS(cpus, freq, cdev_state, power), cpus 139 include/trace/events/thermal.h __assign_bitmask(cpumask, cpumask_bits(cpus), cpus 370 include/trace/events/xen.h TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm, cpus 372 include/trace/events/xen.h TP_ARGS(cpus, mm, addr, end), cpus 379 include/trace/events/xen.h TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); cpus 63 kernel/power/energy_model.c debugfs_create_file("cpus", 0444, d, pd->cpus, &em_debug_cpus_fops); cpus 158 kernel/power/energy_model.c cpumask_copy(to_cpumask(pd->cpus), span); cpus 12 kernel/power/power.h int cpus; cpus 74 kernel/sched/cpufreq.c return cpumask_test_cpu(smp_processor_id(), policy->cpus) || cpus 130 kernel/sched/cpufreq_schedutil.c for_each_cpu(cpu, policy->cpus) cpus 506 kernel/sched/cpufreq_schedutil.c for_each_cpu(j, policy->cpus) { cpus 846 kernel/sched/cpufreq_schedutil.c for_each_cpu(cpu, policy->cpus) { cpus 854 kernel/sched/cpufreq_schedutil.c for_each_cpu(cpu, policy->cpus) { cpus 870 kernel/sched/cpufreq_schedutil.c for_each_cpu(cpu, policy->cpus) cpus 57 kernel/sched/deadline.c int cpus = 0; cpus 62 kernel/sched/deadline.c cpus++; cpus 64 kernel/sched/deadline.c return cpus; cpus 2559 kernel/sched/deadline.c int cpus, err = -1; cpus 2574 kernel/sched/deadline.c cpus = dl_bw_cpus(task_cpu(p)); cpus 2576 kernel/sched/deadline.c !__dl_overflow(dl_b, cpus, 0, new_bw)) { cpus 2578 kernel/sched/deadline.c __dl_sub(dl_b, p->dl.dl_bw, cpus); cpus 2579 kernel/sched/deadline.c __dl_add(dl_b, new_bw, cpus); cpus 2582 kernel/sched/deadline.c !__dl_overflow(dl_b, cpus, p->dl.dl_bw, new_bw)) { cpus 2590 kernel/sched/deadline.c __dl_sub(dl_b, p->dl.dl_bw, cpus); cpus 2591 kernel/sched/deadline.c __dl_add(dl_b, new_bw, cpus); cpus 2721 kernel/sched/deadline.c int cpus, ret; cpus 2729 kernel/sched/deadline.c cpus = dl_bw_cpus(dest_cpu); cpus 2730 kernel/sched/deadline.c overflow = __dl_overflow(dl_b, cpus, 0, p->dl.dl_bw); cpus 2740 kernel/sched/deadline.c __dl_add(dl_b, p->dl.dl_bw, cpus); cpus 2775 kernel/sched/deadline.c int cpus; cpus 2780 kernel/sched/deadline.c cpus = dl_bw_cpus(cpu); cpus 2781 kernel/sched/deadline.c overflow = __dl_overflow(dl_b, cpus, 0, 0); cpus 150 kernel/sched/fair.c unsigned int cpus = min_t(unsigned int, num_online_cpus(), 8); cpus 158 kernel/sched/fair.c factor = cpus; cpus 162 kernel/sched/fair.c factor = 1 + ilog2(cpus); cpus 5885 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask); cpus 5894 kernel/sched/fair.c cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); cpus 5896 kernel/sched/fair.c for_each_cpu_wrap(core, cpus, target) { cpus 5900 kernel/sched/fair.c __cpumask_clear_cpu(cpu, cpus); cpus 5960 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask); cpus 5992 kernel/sched/fair.c cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); cpus 5994 kernel/sched/fair.c for_each_cpu_wrap(cpu, cpus, target) { cpus 7136 kernel/sched/fair.c struct cpumask *cpus; cpus 7281 kernel/sched/fair.c for_each_cpu_and(cpu, env->dst_grpmask, env->cpus) { cpus 8061 kernel/sched/fair.c for_each_cpu_and(i, sched_group_span(group), env->cpus) { cpus 8643 kernel/sched/fair.c for_each_cpu_and(i, sched_group_span(group), env->cpus) { cpus 8796 kernel/sched/fair.c if (!cpumask_test_cpu(env->dst_cpu, env->cpus)) cpus 8807 kernel/sched/fair.c for_each_cpu_and(cpu, group_balance_mask(sg), env->cpus) { cpus 8838 kernel/sched/fair.c struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask); cpus 8847 kernel/sched/fair.c .cpus = cpus, cpus 8852 kernel/sched/fair.c cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask); cpus 8946 kernel/sched/fair.c __cpumask_clear_cpu(env.dst_cpu, env.cpus); cpus 8973 kernel/sched/fair.c __cpumask_clear_cpu(cpu_of(busiest), cpus); cpus 8982 kernel/sched/fair.c if (!cpumask_subset(cpus, env.dst_grpmask)) { cpus 294 kernel/sched/sched.h void __dl_sub(struct dl_bw *dl_b, u64 tsk_bw, int cpus) cpus 297 kernel/sched/sched.h __dl_update(dl_b, (s32)tsk_bw / cpus); cpus 301 kernel/sched/sched.h void __dl_add(struct dl_bw *dl_b, u64 tsk_bw, int cpus) cpus 304 kernel/sched/sched.h __dl_update(dl_b, -((s32)tsk_bw / cpus)); cpus 308 kernel/sched/sched.h bool __dl_overflow(struct dl_bw *dl_b, int cpus, u64 old_bw, u64 new_bw) cpus 311 kernel/sched/sched.h dl_b->bw * cpus < dl_b->total_bw - old_bw + new_bw; cpus 1281 kernel/sched/sched.h extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu); cpus 1286 kernel/sched/sched.h static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu) cpus 2460 kernel/sched/sched.h #define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus))) cpus 1736 kernel/sched/topology.c int sched_numa_find_closest(const struct cpumask *cpus, int cpu) cpus 1741 kernel/sched/topology.c cpu = cpumask_any_and(cpus, sched_domains_numa_masks[i][j]); cpus 687 kernel/smp.c cpumask_var_t cpus; cpus 692 kernel/smp.c if (likely(zalloc_cpumask_var(&cpus, (gfp_flags|__GFP_NOWARN)))) { cpus 696 kernel/smp.c __cpumask_set_cpu(cpu, cpus); cpus 697 kernel/smp.c on_each_cpu_mask(cpus, func, info, wait); cpus 699 kernel/smp.c free_cpumask_var(cpus); cpus 596 kernel/stop_machine.c const struct cpumask *cpus) cpus 602 kernel/stop_machine.c .active_cpus = cpus, cpus 631 kernel/stop_machine.c int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus) cpus 637 kernel/stop_machine.c ret = stop_machine_cpuslocked(fn, data, cpus); cpus 666 kernel/stop_machine.c const struct cpumask *cpus) cpus 669 kernel/stop_machine.c .active_cpus = cpus }; cpus 487 kernel/trace/ring_buffer.c int cpus; cpus 1407 kernel/trace/ring_buffer.c buffer->cpus = nr_cpu_ids; cpus 4502 mm/slub.c DECLARE_BITMAP(cpus, NR_CPUS); cpus 4578 mm/slub.c to_cpumask(l->cpus)); cpus 4608 mm/slub.c cpumask_clear(to_cpumask(l->cpus)); cpus 4609 mm/slub.c cpumask_set_cpu(track->cpu, to_cpumask(l->cpus)); cpus 4692 mm/slub.c !cpumask_empty(to_cpumask(l->cpus)) && cpus 4696 mm/slub.c cpumask_pr_args(to_cpumask(l->cpus))); cpus 4031 net/core/dev.c tcpu = map->cpus[reciprocal_scale(hash, map->len)]; cpus 699 net/core/net-sysfs.c cpumask_set_cpu(map->cpus[i], mask); cpus 738 net/core/net-sysfs.c map->cpus[i++] = cpu; cpus 241 samples/trace_events/trace-events-sample.h __bitmask( cpus, num_possible_cpus() ) cpus 250 samples/trace_events/trace-events-sample.h __assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus()); cpus 297 samples/trace_events/trace-events-sample.h __get_str(str), __get_bitmask(cpus)) cpus 626 scripts/dtc/livetree.c struct node *cpus, *bootcpu; cpus 629 scripts/dtc/livetree.c cpus = get_node_by_path(tree, "/cpus"); cpus 630 scripts/dtc/livetree.c if (!cpus) cpus 634 scripts/dtc/livetree.c bootcpu = cpus->children; cpus 510 sound/soc/atmel/atmel-classd.c dai_link->cpus = &comp[0]; cpus 521 sound/soc/atmel/atmel-classd.c dai_link->cpus->dai_name = dev_name(dev); cpus 521 sound/soc/atmel/atmel-pdmic.c dai_link->cpus = &comp[0]; cpus 532 sound/soc/atmel/atmel-pdmic.c dai_link->cpus->dai_name = dev_name(dev); cpus 115 sound/soc/atmel/atmel_wm8904.c dailink->cpus->of_node = cpu_np; cpus 144 sound/soc/atmel/atmel_wm8904.c id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc"); cpus 170 sound/soc/atmel/atmel_wm8904.c id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc"); cpus 96 sound/soc/atmel/mikroe-proto.c dai->cpus = &comp[0]; cpus 120 sound/soc/atmel/mikroe-proto.c dai->cpus->of_node = cpu_np; cpus 212 sound/soc/atmel/sam9g20_wm8731.c at91sam9g20ek_dai.cpus->dai_name = NULL; cpus 219 sound/soc/atmel/sam9g20_wm8731.c at91sam9g20ek_dai.cpus->of_node = cpu_np; cpus 106 sound/soc/atmel/sam9x5_wm8731.c dai->cpus = &comp[0]; cpus 147 sound/soc/atmel/sam9x5_wm8731.c dai->cpus->of_node = cpu_np; cpus 339 sound/soc/atmel/tse850-pcm5142.c dailink->cpus->of_node = cpu_np; cpus 136 sound/soc/fsl/eukrea-tlv320.c eukrea_tlv320_dai.cpus->of_node = ssi_np; cpus 139 sound/soc/fsl/eukrea-tlv320.c eukrea_tlv320_dai.cpus->dai_name = "imx-ssi.0"; cpus 634 sound/soc/fsl/fsl-asoc-card.c priv->dai_link[0].cpus->of_node = cpu_np; cpus 665 sound/soc/fsl/fsl-asoc-card.c priv->dai_link[1].cpus->of_node = asrc_np; cpus 671 sound/soc/fsl/fsl-asoc-card.c priv->dai_link[2].cpus->of_node = cpu_np; cpus 243 sound/soc/fsl/imx-audmix.c priv->dai[i].cpus = &dlc[0]; cpus 255 sound/soc/fsl/imx-audmix.c priv->dai[i].cpus->of_node = args.np; cpus 256 sound/soc/fsl/imx-audmix.c priv->dai[i].cpus->dai_name = dev_name(&cpu_pdev->dev); cpus 272 sound/soc/fsl/imx-audmix.c priv->dai[num_dai + i].cpus = &dlc[3]; cpus 283 sound/soc/fsl/imx-audmix.c priv->dai[num_dai + i].cpus->of_node = audmix_np; cpus 284 sound/soc/fsl/imx-audmix.c priv->dai[num_dai + i].cpus->dai_name = be_name; cpus 161 sound/soc/fsl/imx-es8328.c data->dai.cpus = &comp[0]; cpus 173 sound/soc/fsl/imx-es8328.c data->dai.cpus->of_node = ssi_np; cpus 140 sound/soc/fsl/imx-sgtl5000.c data->dai.cpus = &comp[0]; cpus 152 sound/soc/fsl/imx-sgtl5000.c data->dai.cpus->of_node = ssi_np; cpus 35 sound/soc/fsl/imx-spdif.c data->dai.cpus = &comp[0]; cpus 47 sound/soc/fsl/imx-spdif.c data->dai.cpus->of_node = spdif_np; cpus 216 sound/soc/fsl/mpc8610_hpcd.c machine_data->dai[0].cpus = &comp[0]; cpus 224 sound/soc/fsl/mpc8610_hpcd.c machine_data->dai[1].cpus = &comp[3]; cpus 232 sound/soc/fsl/mpc8610_hpcd.c machine_data->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev); cpus 226 sound/soc/fsl/p1022_ds.c mdata->dai[0].cpus = &comp[0]; cpus 234 sound/soc/fsl/p1022_ds.c mdata->dai[1].cpus = &comp[3]; cpus 243 sound/soc/fsl/p1022_ds.c mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev); cpus 229 sound/soc/fsl/p1022_rdk.c mdata->dai[0].cpus = &comp[0]; cpus 237 sound/soc/fsl/p1022_rdk.c mdata->dai[1].cpus = &comp[3]; cpus 245 sound/soc/fsl/p1022_rdk.c mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev); cpus 216 sound/soc/generic/audio-graph-card.c struct snd_soc_dai_link_component *cpus = dai_link->cpus; cpus 257 sound/soc/generic/audio-graph-card.c cpus->dai_name); cpus 267 sound/soc/generic/audio-graph-card.c cpus->of_node = NULL; cpus 268 sound/soc/generic/audio-graph-card.c cpus->dai_name = "snd-soc-dummy-dai"; cpus 269 sound/soc/generic/audio-graph-card.c cpus->name = "snd-soc-dummy"; cpus 393 sound/soc/generic/audio-graph-card.c dai_link->cpus->dai_name, cpus 358 sound/soc/generic/simple-card-utils.c dai_link->platforms->of_node = dai_link->cpus->of_node; cpus 382 sound/soc/generic/simple-card-utils.c dai_link->cpus->dai_name = NULL; cpus 392 sound/soc/generic/simple-card-utils.c of_node_put(dai_link->cpus->of_node); cpus 582 sound/soc/generic/simple-card-utils.c dai_link[i].cpus = &dai_props[i].cpus; cpus 125 sound/soc/generic/simple-card.c struct snd_soc_dai_link_component *cpus = dai_link->cpus; cpus 174 sound/soc/generic/simple-card.c cpus->dai_name); cpus 183 sound/soc/generic/simple-card.c cpus->of_node = NULL; cpus 184 sound/soc/generic/simple-card.c cpus->dai_name = "snd-soc-dummy-dai"; cpus 185 sound/soc/generic/simple-card.c cpus->name = "snd-soc-dummy"; cpus 327 sound/soc/generic/simple-card.c dai_link->cpus->dai_name, cpus 653 sound/soc/generic/simple-card.c struct snd_soc_dai_link_component *cpus; cpus 679 sound/soc/generic/simple-card.c cpus = dai_link->cpus; cpus 680 sound/soc/generic/simple-card.c cpus->dai_name = cinfo->cpu_dai.name; cpus 684 sound/soc/intel/boards/bxt_da7219_max98357a.c broxton_dais[i].cpus->dai_name = "SSP1 Pin"; cpus 690 sound/soc/intel/boards/bxt_da7219_max98357a.c broxton_dais[i].cpus->dai_name = "SSP2 Pin"; cpus 703 sound/soc/intel/boards/bxt_da7219_max98357a.c broxton_dais[i].cpus->dai_name = "SSP1 Pin"; cpus 709 sound/soc/intel/boards/bxt_da7219_max98357a.c broxton_dais[i].cpus->dai_name = "SSP0 Pin"; cpus 602 sound/soc/intel/boards/bxt_rt298.c card->dai_link[i].cpus->dai_name = "SSP5 Pin"; cpus 606 sound/soc/intel/boards/bxt_rt298.c card->dai_link[i].cpus->dai_name = "SSP2 Pin"; cpus 531 sound/soc/intel/boards/bytcht_es8316.c byt_cht_es8316_dais[dai_index].cpus->dai_name = "ssp0-port"; cpus 1299 sound/soc/intel/boards/bytcr_rt5640.c byt_rt5640_dais[dai_index].cpus->dai_name = cpus 1061 sound/soc/intel/boards/bytcr_rt5651.c byt_rt5651_dais[dai_index].cpus->dai_name = cpus 663 sound/soc/intel/boards/cht_bsw_rt5645.c cht_dailink[dai_index].cpus->dai_name = cpus 417 sound/soc/intel/boards/sof_rt5682.c struct snd_soc_dai_link_component *cpus; cpus 423 sound/soc/intel/boards/sof_rt5682.c cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component) * cpus 425 sound/soc/intel/boards/sof_rt5682.c if (!links || !cpus) cpus 445 sound/soc/intel/boards/sof_rt5682.c links[id].cpus = &cpus[id]; cpus 448 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, cpus 451 sound/soc/intel/boards/sof_rt5682.c if (!links[id].cpus->dai_name) cpus 464 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, cpus 467 sound/soc/intel/boards/sof_rt5682.c if (!links[id].cpus->dai_name) cpus 476 sound/soc/intel/boards/sof_rt5682.c links[id].cpus = &cpus[id]; cpus 477 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = "DMIC01 Pin"; cpus 482 sound/soc/intel/boards/sof_rt5682.c links[id + 1].cpus = &cpus[id + 1]; cpus 483 sound/soc/intel/boards/sof_rt5682.c links[id + 1].cpus->dai_name = "DMIC16k Pin"; cpus 516 sound/soc/intel/boards/sof_rt5682.c links[id].cpus = &cpus[id]; cpus 518 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, cpus 520 sound/soc/intel/boards/sof_rt5682.c if (!links[id].cpus->dai_name) cpus 557 sound/soc/intel/boards/sof_rt5682.c links[id].cpus = &cpus[id]; cpus 560 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, cpus 563 sound/soc/intel/boards/sof_rt5682.c if (!links[id].cpus->dai_name) cpus 567 sound/soc/intel/boards/sof_rt5682.c links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, cpus 570 sound/soc/intel/boards/sof_rt5682.c if (!links[id].cpus->dai_name) cpus 1090 sound/soc/intel/skylake/skl-pcm.c dai_link->cpus->dai_name); cpus 111 sound/soc/kirkwood/armada-370-db.c a370db_dai[0].cpus->of_node = cpus 114 sound/soc/kirkwood/armada-370-db.c a370db_dai[0].platforms->of_node = a370db_dai[0].cpus->of_node; cpus 120 sound/soc/kirkwood/armada-370-db.c a370db_dai[1].cpus->of_node = a370db_dai[0].cpus->of_node; cpus 121 sound/soc/kirkwood/armada-370-db.c a370db_dai[1].platforms->of_node = a370db_dai[0].cpus->of_node; cpus 127 sound/soc/kirkwood/armada-370-db.c a370db_dai[2].cpus->of_node = a370db_dai[0].cpus->of_node; cpus 128 sound/soc/kirkwood/armada-370-db.c a370db_dai[2].platforms->of_node = a370db_dai[0].cpus->of_node; cpus 119 sound/soc/meson/axg-card.c if (link->cpus) cpus 120 sound/soc/meson/axg-card.c of_node_put(link->cpus->of_node); cpus 277 sound/soc/meson/axg-card.c lb->cpus = &dlc[0]; cpus 283 sound/soc/meson/axg-card.c lb->cpus->of_node = pad->cpus->of_node; cpus 284 sound/soc/meson/axg-card.c lb->cpus->dai_name = "TDM Loopback"; cpus 299 sound/soc/meson/axg-card.c of_node_get(lb->cpus->of_node); cpus 441 sound/soc/meson/axg-card.c link->dai_fmt = axg_card_parse_daifmt(node, link->cpus->of_node); cpus 569 sound/soc/meson/axg-card.c dai_link->cpus = cpu; cpus 572 sound/soc/meson/axg-card.c ret = axg_card_parse_dai(card, np, &dai_link->cpus->of_node, cpus 573 sound/soc/meson/axg-card.c &dai_link->cpus->dai_name); cpus 577 sound/soc/meson/axg-card.c if (axg_card_cpu_is_playback_fe(dai_link->cpus->of_node)) cpus 579 sound/soc/meson/axg-card.c else if (axg_card_cpu_is_capture_fe(dai_link->cpus->of_node)) cpus 587 sound/soc/meson/axg-card.c if (axg_card_cpu_is_tdm_iface(dai_link->cpus->of_node)) cpus 589 sound/soc/meson/axg-card.c else if (axg_card_cpu_is_codec(dai_link->cpus->of_node)) cpus 127 sound/soc/mxs/mxs-sgtl5000.c mxs_sgtl5000_dai[i].cpus->dai_name = NULL; cpus 128 sound/soc/mxs/mxs-sgtl5000.c mxs_sgtl5000_dai[i].cpus->of_node = saif_np[i]; cpus 158 sound/soc/qcom/apq8016_sbc.c link->cpus = &dlc[0]; cpus 173 sound/soc/qcom/apq8016_sbc.c link->cpus->of_node = of_parse_phandle(cpu, "sound-dai", 0); cpus 174 sound/soc/qcom/apq8016_sbc.c if (!link->cpus->of_node) { cpus 180 sound/soc/qcom/apq8016_sbc.c ret = snd_soc_of_get_dai_name(cpu, &link->cpus->dai_name); cpus 193 sound/soc/qcom/apq8016_sbc.c link->platforms->of_node = link->cpus->of_node; cpus 50 sound/soc/qcom/common.c link->cpus = &dlc[0]; cpus 78 sound/soc/qcom/common.c link->cpus->of_node = args.np; cpus 81 sound/soc/qcom/common.c ret = snd_soc_of_get_dai_name(cpu, &link->cpus->dai_name); cpus 112 sound/soc/qcom/common.c link->platforms->of_node = link->cpus->of_node; cpus 73 sound/soc/qcom/storm.c dai_link->cpus->of_node = of_parse_phandle(np, "cpu", 0); cpus 74 sound/soc/qcom/storm.c if (!dai_link->cpus->of_node) { cpus 78 sound/soc/qcom/storm.c dai_link->platforms->of_node = dai_link->cpus->of_node; cpus 232 sound/soc/rockchip/rk3288_hdmi_analog.c rk_dailink.cpus->of_node = of_parse_phandle(np, "rockchip,i2s-controller", cpus 234 sound/soc/rockchip/rk3288_hdmi_analog.c if (!rk_dailink.cpus->of_node) { cpus 240 sound/soc/rockchip/rk3288_hdmi_analog.c rk_dailink.platforms->of_node = rk_dailink.cpus->of_node; cpus 514 sound/soc/rockchip/rk3399_gru_sound.c dai->cpus->of_node = np_cpu; cpus 233 sound/soc/rockchip/rockchip_max98090.c rk_dailink.cpus->of_node = of_parse_phandle(np, cpus 235 sound/soc/rockchip/rockchip_max98090.c if (!rk_dailink.cpus->of_node) { cpus 241 sound/soc/rockchip/rockchip_max98090.c rk_dailink.platforms->of_node = rk_dailink.cpus->of_node; cpus 172 sound/soc/rockchip/rockchip_rt5645.c rk_dailink.cpus->of_node = of_parse_phandle(np, cpus 174 sound/soc/rockchip/rockchip_rt5645.c if (!rk_dailink.cpus->of_node) { cpus 181 sound/soc/rockchip/rockchip_rt5645.c rk_dailink.platforms->of_node = rk_dailink.cpus->of_node; cpus 200 sound/soc/rockchip/rockchip_rt5645.c of_node_put(rk_dailink.cpus->of_node); cpus 201 sound/soc/rockchip/rockchip_rt5645.c rk_dailink.cpus->of_node = NULL; cpus 211 sound/soc/rockchip/rockchip_rt5645.c of_node_put(rk_dailink.cpus->of_node); cpus 212 sound/soc/rockchip/rockchip_rt5645.c rk_dailink.cpus->of_node = NULL; cpus 84 sound/soc/samsung/arndale_rt5631.c of_node_put(dai_link->cpus->of_node); cpus 98 sound/soc/samsung/arndale_rt5631.c if (!arndale_rt5631_dai[n].cpus->dai_name) { cpus 99 sound/soc/samsung/arndale_rt5631.c arndale_rt5631_dai[n].cpus->of_node = of_parse_phandle(np, cpus 102 sound/soc/samsung/arndale_rt5631.c if (!arndale_rt5631_dai[n].cpus->of_node) { cpus 110 sound/soc/samsung/arndale_rt5631.c arndale_rt5631_dai[n].cpus->of_node; cpus 274 sound/soc/samsung/odroid.c ret = snd_soc_get_dai_name(&args, &link->cpus->dai_name); cpus 160 sound/soc/samsung/smdk_wm8994.c smdk_dai[0].cpus->dai_name = NULL; cpus 161 sound/soc/samsung/smdk_wm8994.c smdk_dai[0].cpus->of_node = of_parse_phandle(np, cpus 163 sound/soc/samsung/smdk_wm8994.c if (!smdk_dai[0].cpus->of_node) { cpus 170 sound/soc/samsung/smdk_wm8994.c smdk_dai[0].platforms->of_node = smdk_dai[0].cpus->of_node; cpus 149 sound/soc/samsung/snow.c link->cpus = links_cpus; cpus 166 sound/soc/samsung/snow.c link->cpus->of_node = of_parse_phandle(cpu, "sound-dai", 0); cpus 169 sound/soc/samsung/snow.c if (!link->cpus->of_node) { cpus 179 sound/soc/samsung/snow.c of_node_put(link->cpus->of_node); cpus 184 sound/soc/samsung/snow.c priv->clk_i2s_bus = of_clk_get_by_name(link->cpus->of_node, cpus 188 sound/soc/samsung/snow.c of_node_put(link->cpus->of_node); cpus 194 sound/soc/samsung/snow.c link->cpus->of_node = of_parse_phandle(dev->of_node, cpus 196 sound/soc/samsung/snow.c if (!link->cpus->of_node) { cpus 204 sound/soc/samsung/snow.c of_node_put(link->cpus->of_node); cpus 210 sound/soc/samsung/snow.c link->platforms->of_node = link->cpus->of_node; cpus 231 sound/soc/samsung/snow.c of_node_put(link->cpus->of_node); cpus 576 sound/soc/samsung/tm2_wm5110.c dai_link->cpus->name = NULL; cpus 583 sound/soc/samsung/tm2_wm5110.c dai_link->cpus->of_node = cpu_dai_node[dai_index]; cpus 99 sound/soc/sirf/sirf-audio.c sirf_audio_dai_link[0].cpus->of_node = cpus 899 sound/soc/soc-core.c rtd->cpu_dai = snd_soc_find_dai(dai_link->cpus); cpus 902 sound/soc/soc-core.c dai_link->cpus->dai_name); cpus 1320 sound/soc/soc-core.c if (link->cpus->name && link->cpus->of_node) { cpus 1331 sound/soc/soc-core.c if ((link->cpus->of_node || link->cpus->name) && cpus 1332 sound/soc/soc-core.c !soc_find_component(link->cpus)) cpus 1339 sound/soc/soc-core.c if (!link->cpus->dai_name && cpus 1340 sound/soc/soc-core.c !(link->cpus->name || link->cpus->of_node)) { cpus 556 sound/soc/soc-topology.c kfree(link->cpus->dai_name); cpus 1908 sound/soc/soc-topology.c link->cpus = &dlc[0]; cpus 1927 sound/soc/soc-topology.c link->cpus->dai_name = kstrdup(pcm->dai_name, GFP_KERNEL); cpus 1962 sound/soc/soc-topology.c kfree(link->cpus->dai_name); cpus 41 sound/soc/sof/nocodec.c links[i].cpus = &dlc[0]; cpus 51 sound/soc/sof/nocodec.c links[i].cpus->dai_name = ops->drv[i].name; cpus 2427 sound/soc/sof/topology.c dai->cpu_dai_name = link->cpus->dai_name; cpus 2700 sound/soc/sof/topology.c sof_dai->cpu_dai_name = link->cpus->dai_name; cpus 2756 sound/soc/sof/topology.c dai = snd_soc_find_dai(link->cpus); cpus 2759 sound/soc/sof/topology.c link->cpus->dai_name, __func__); cpus 2934 sound/soc/sof/topology.c dai = snd_soc_find_dai(link->cpus); cpus 2937 sound/soc/sof/topology.c link->cpus->dai_name, __func__); cpus 1304 sound/soc/sunxi/sun4i-codec.c link->cpus = &dlc[0]; cpus 1315 sound/soc/sunxi/sun4i-codec.c link->cpus->dai_name = dev_name(dev); cpus 189 sound/soc/tegra/tegra_alc5632.c tegra_alc5632_dai.cpus->of_node = of_parse_phandle(np, cpus 191 sound/soc/tegra/tegra_alc5632.c if (!tegra_alc5632_dai.cpus->of_node) { cpus 198 sound/soc/tegra/tegra_alc5632.c tegra_alc5632_dai.platforms->of_node = tegra_alc5632_dai.cpus->of_node; cpus 216 sound/soc/tegra/tegra_alc5632.c of_node_put(tegra_alc5632_dai.cpus->of_node); cpus 217 sound/soc/tegra/tegra_alc5632.c tegra_alc5632_dai.cpus->of_node = NULL; cpus 235 sound/soc/tegra/tegra_alc5632.c of_node_put(tegra_alc5632_dai.cpus->of_node); cpus 236 sound/soc/tegra/tegra_alc5632.c tegra_alc5632_dai.cpus->of_node = NULL; cpus 236 sound/soc/tegra/tegra_max98090.c tegra_max98090_dai.cpus->of_node = of_parse_phandle(np, cpus 238 sound/soc/tegra/tegra_max98090.c if (!tegra_max98090_dai.cpus->of_node) { cpus 245 sound/soc/tegra/tegra_max98090.c tegra_max98090_dai.platforms->of_node = tegra_max98090_dai.cpus->of_node; cpus 182 sound/soc/tegra/tegra_rt5640.c tegra_rt5640_dai.cpus->of_node = of_parse_phandle(np, cpus 184 sound/soc/tegra/tegra_rt5640.c if (!tegra_rt5640_dai.cpus->of_node) { cpus 191 sound/soc/tegra/tegra_rt5640.c tegra_rt5640_dai.platforms->of_node = tegra_rt5640_dai.cpus->of_node; cpus 255 sound/soc/tegra/tegra_rt5677.c tegra_rt5677_dai.cpus->of_node = of_parse_phandle(np, cpus 257 sound/soc/tegra/tegra_rt5677.c if (!tegra_rt5677_dai.cpus->of_node) { cpus 263 sound/soc/tegra/tegra_rt5677.c tegra_rt5677_dai.platforms->of_node = tegra_rt5677_dai.cpus->of_node; cpus 281 sound/soc/tegra/tegra_rt5677.c of_node_put(tegra_rt5677_dai.cpus->of_node); cpus 282 sound/soc/tegra/tegra_rt5677.c tegra_rt5677_dai.cpus->of_node = NULL; cpus 303 sound/soc/tegra/tegra_rt5677.c of_node_put(tegra_rt5677_dai.cpus->of_node); cpus 304 sound/soc/tegra/tegra_rt5677.c tegra_rt5677_dai.cpus->of_node = NULL; cpus 140 sound/soc/tegra/tegra_sgtl5000.c tegra_sgtl5000_dai.cpus->of_node = of_parse_phandle(np, cpus 142 sound/soc/tegra/tegra_sgtl5000.c if (!tegra_sgtl5000_dai.cpus->of_node) { cpus 149 sound/soc/tegra/tegra_sgtl5000.c tegra_sgtl5000_dai.platforms->of_node = tegra_sgtl5000_dai.cpus->of_node; cpus 167 sound/soc/tegra/tegra_sgtl5000.c of_node_put(tegra_sgtl5000_dai.cpus->of_node); cpus 168 sound/soc/tegra/tegra_sgtl5000.c tegra_sgtl5000_dai.cpus->of_node = NULL; cpus 187 sound/soc/tegra/tegra_sgtl5000.c of_node_put(tegra_sgtl5000_dai.cpus->of_node); cpus 188 sound/soc/tegra/tegra_sgtl5000.c tegra_sgtl5000_dai.cpus->of_node = NULL; cpus 145 sound/soc/tegra/tegra_wm8753.c tegra_wm8753_dai.cpus->of_node = of_parse_phandle(np, cpus 147 sound/soc/tegra/tegra_wm8753.c if (!tegra_wm8753_dai.cpus->of_node) { cpus 154 sound/soc/tegra/tegra_wm8753.c tegra_wm8753_dai.platforms->of_node = tegra_wm8753_dai.cpus->of_node; cpus 319 sound/soc/tegra/tegra_wm8903.c tegra_wm8903_dai.cpus->of_node = of_parse_phandle(np, cpus 321 sound/soc/tegra/tegra_wm8903.c if (!tegra_wm8903_dai.cpus->of_node) { cpus 328 sound/soc/tegra/tegra_wm8903.c tegra_wm8903_dai.platforms->of_node = tegra_wm8903_dai.cpus->of_node; cpus 99 sound/soc/tegra/tegra_wm9712.c tegra_wm9712_dai.cpus->of_node = of_parse_phandle(np, cpus 101 sound/soc/tegra/tegra_wm9712.c if (!tegra_wm9712_dai.cpus->of_node) { cpus 108 sound/soc/tegra/tegra_wm9712.c tegra_wm9712_dai.platforms->of_node = tegra_wm9712_dai.cpus->of_node; cpus 132 sound/soc/tegra/trimslice.c trimslice_tlv320aic23_dai.cpus->of_node = of_parse_phandle(np, cpus 134 sound/soc/tegra/trimslice.c if (!trimslice_tlv320aic23_dai.cpus->of_node) { cpus 142 sound/soc/tegra/trimslice.c trimslice_tlv320aic23_dai.cpus->of_node; cpus 407 sound/soc/ti/davinci-evm.c dai->cpus->of_node = of_parse_phandle(np, "ti,mcasp-controller", 0); cpus 408 sound/soc/ti/davinci-evm.c if (!dai->cpus->of_node) cpus 411 sound/soc/ti/davinci-evm.c dai->platforms->of_node = dai->cpus->of_node; cpus 256 sound/soc/ti/omap-abe-twl6040.c priv->dai_links[0].cpus = link0_cpus; cpus 258 sound/soc/ti/omap-abe-twl6040.c priv->dai_links[0].cpus->of_node = dai_node; cpus 272 sound/soc/ti/omap-abe-twl6040.c priv->dai_links[1].cpus = link1_cpus; cpus 274 sound/soc/ti/omap-abe-twl6040.c priv->dai_links[1].cpus->of_node = dai_node; cpus 370 sound/soc/ti/omap-hdmi.c card->dai_link->cpus = &compnent[0]; cpus 379 sound/soc/ti/omap-hdmi.c card->dai_link->cpus->dai_name = dev_name(ad->dssdev); cpus 265 sound/soc/ti/omap-twl4030.c omap_twl4030_dai_links[0].cpus->dai_name = NULL; cpus 266 sound/soc/ti/omap-twl4030.c omap_twl4030_dai_links[0].cpus->of_node = dai_node; cpus 275 sound/soc/ti/omap-twl4030.c omap_twl4030_dai_links[1].cpus->dai_name = NULL; cpus 276 sound/soc/ti/omap-twl4030.c omap_twl4030_dai_links[1].cpus->of_node = dai_node; cpus 380 sound/soc/ti/rx51.c rx51_dai[0].cpus->dai_name = NULL; cpus 382 sound/soc/ti/rx51.c rx51_dai[0].cpus->of_node = dai_node; cpus 67 sound/soc/ux500/mop500.c of_node_put(mop500_dai_links[i].cpus->of_node); cpus 89 sound/soc/ux500/mop500.c mop500_dai_links[i].cpus->of_node = msp_np[i]; cpus 90 sound/soc/ux500/mop500.c mop500_dai_links[i].cpus->dai_name = NULL; cpus 447 tools/bpf/bpftool/common.c int cpus = libbpf_num_possible_cpus(); cpus 449 tools/bpf/bpftool/common.c if (cpus < 0) { cpus 450 tools/bpf/bpftool/common.c p_err("Can't get # of possible cpus: %s", strerror(-cpus)); cpus 453 tools/bpf/bpftool/common.c return cpus; cpus 198 tools/bpf/bpftool/map_perf_ring.c opts.cpus = &ctx.cpu; cpus 5309 tools/lib/bpf/libbpf.c int *cpus; cpus 5458 tools/lib/bpf/libbpf.c p.cpus = opts->cpus; cpus 5544 tools/lib/bpf/libbpf.c cpu = p->cpu_cnt > 0 ? p->cpus[i] : i; cpus 5988 tools/lib/bpf/libbpf.c static int cpus; cpus 5992 tools/lib/bpf/libbpf.c tmp_cpus = READ_ONCE(cpus); cpus 6007 tools/lib/bpf/libbpf.c WRITE_ONCE(cpus, tmp_cpus); cpus 181 tools/lib/traceevent/event-parse-api.c return tep->cpus; cpus 191 tools/lib/traceevent/event-parse-api.c void tep_set_cpus(struct tep_handle *tep, int cpus) cpus 194 tools/lib/traceevent/event-parse-api.c tep->cpus = cpus; cpus 33 tools/lib/traceevent/event-parse-local.h int cpus; cpus 557 tools/lib/traceevent/event-parse.h void tep_set_cpus(struct tep_handle *tep, int cpus); cpus 33 tools/lib/traceevent/plugins/plugin_function.c static int cpus = -1; cpus 96 tools/lib/traceevent/plugins/plugin_function.c if (cpu > cpus) { cpus 108 tools/lib/traceevent/plugins/plugin_function.c for (i = cpus + 1; i <= cpu; i++) cpus 110 tools/lib/traceevent/plugins/plugin_function.c cpus = cpu; cpus 184 tools/lib/traceevent/plugins/plugin_function.c for (i = 0; i <= cpus; i++) { cpus 194 tools/lib/traceevent/plugins/plugin_function.c cpus = -1; cpus 162 tools/perf/arch/arm/util/cs-etm.c struct perf_cpu_map *event_cpus = evsel->evlist->core.cpus; cpus 260 tools/perf/arch/arm/util/cs-etm.c struct perf_cpu_map *cpus = evlist->core.cpus; cpus 403 tools/perf/arch/arm/util/cs-etm.c if (!perf_cpu_map__empty(cpus)) { cpus 427 tools/perf/arch/arm/util/cs-etm.c if (!perf_cpu_map__empty(cpus)) cpus 496 tools/perf/arch/arm/util/cs-etm.c struct perf_cpu_map *event_cpus = evlist->core.cpus; cpus 643 tools/perf/arch/arm/util/cs-etm.c struct perf_cpu_map *event_cpus = session->evlist->core.cpus; cpus 22 tools/perf/arch/arm64/util/header.c struct perf_cpu_map *cpus; cpus 25 tools/perf/arch/arm64/util/header.c if (!sysfs || !pmu || !pmu->cpus) cpus 33 tools/perf/arch/arm64/util/header.c cpus = perf_cpu_map__get(pmu->cpus); cpus 34 tools/perf/arch/arm64/util/header.c for (cpu = 0; cpu < perf_cpu_map__nr(cpus); cpu++) { cpus 36 tools/perf/arch/arm64/util/header.c sysfs, cpus->map[cpu]); cpus 66 tools/perf/arch/arm64/util/header.c perf_cpu_map__put(cpus); cpus 15 tools/perf/arch/nds32/util/header.c struct cpu_map *cpus; cpus 18 tools/perf/arch/nds32/util/header.c if (!sysfs || !pmu || !pmu->cpus) cpus 25 tools/perf/arch/nds32/util/header.c cpus = cpu_map__get(pmu->cpus); cpus 26 tools/perf/arch/nds32/util/header.c sprintf(buf, "0x%x", cpus->nr - 1); cpus 27 tools/perf/arch/nds32/util/header.c cpu_map__put(cpus); cpus 59 tools/perf/arch/x86/tests/perf-time-to-tsc.c struct perf_cpu_map *cpus = NULL; cpus 74 tools/perf/arch/x86/tests/perf-time-to-tsc.c cpus = perf_cpu_map__new(NULL); cpus 75 tools/perf/arch/x86/tests/perf-time-to-tsc.c CHECK_NOT_NULL__(cpus); cpus 80 tools/perf/arch/x86/tests/perf-time-to-tsc.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 113 tools/perf/arch/x86/util/intel-bts.c const struct perf_cpu_map *cpus = evlist->core.cpus; cpus 140 tools/perf/arch/x86/util/intel-bts.c if (opts->full_auxtrace && !perf_cpu_map__empty(cpus)) { cpus 221 tools/perf/arch/x86/util/intel-bts.c if (!perf_cpu_map__empty(cpus)) cpus 371 tools/perf/arch/x86/util/intel-pt.c per_cpu_mmaps = !perf_cpu_map__empty(session->evlist->core.cpus); cpus 583 tools/perf/arch/x86/util/intel-pt.c const struct perf_cpu_map *cpus = evlist->core.cpus; cpus 708 tools/perf/arch/x86/util/intel-pt.c if (have_timing_info && !perf_cpu_map__empty(cpus)) { cpus 766 tools/perf/arch/x86/util/intel-pt.c if (!perf_cpu_map__empty(cpus)) cpus 790 tools/perf/arch/x86/util/intel-pt.c if (!perf_cpu_map__empty(cpus)) { cpus 802 tools/perf/arch/x86/util/intel-pt.c if (!ptr->have_sched_switch && !perf_cpu_map__empty(cpus) && cpus 198 tools/perf/builtin-ftrace.c struct perf_cpu_map *cpumap = ftrace->evlist->core.cpus; cpus 770 tools/perf/builtin-record.c if (evsel__open(pos, pos->core.cpus, pos->core.threads) < 0) { cpus 1316 tools/perf/builtin-record.c err = perf_event__synthesize_cpu_map(&rec->tool, rec->evlist->core.cpus, cpus 172 tools/perf/builtin-sched.c struct perf_cpu_map *cpus; cpus 1614 tools/perf/builtin-sched.c if (sched->map.cpus && !cpu_map__has(sched->map.cpus, cpu)) cpus 1636 tools/perf/builtin-sched.c if (sched->map.cpus && !cpu_map__has(sched->map.cpus, this_cpu)) cpus 3198 tools/perf/builtin-sched.c sched->map.cpus = map; cpus 1645 tools/perf/builtin-script.c struct perf_cpu_map *cpus; cpus 1942 tools/perf/builtin-script.c counter->core.cpus->map[cpu], cpus 3280 tools/perf/builtin-script.c if (!script->cpus || !script->threads) cpus 3286 tools/perf/builtin-script.c perf_evlist__set_maps(&evlist->core, script->cpus, script->threads); cpus 3321 tools/perf/builtin-script.c if (script->cpus) { cpus 3326 tools/perf/builtin-script.c script->cpus = cpu_map__new_data(&event->cpu_map.data); cpus 3327 tools/perf/builtin-script.c if (!script->cpus) cpus 171 tools/perf/builtin-stat.c struct perf_cpu_map *cpus; cpus 879 tools/perf/builtin-stat.c if (cpu_map__build_socket_map(evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 886 tools/perf/builtin-stat.c if (cpu_map__build_die_map(evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 893 tools/perf/builtin-stat.c if (cpu_map__build_core_map(evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 901 tools/perf/builtin-stat.c if (cpu_map__build_core_map(evsel_list->core.cpus, cpus 921 tools/perf/builtin-stat.c nr = perf_cpu_map__max(evsel_list->core.cpus); cpus 1011 tools/perf/builtin-stat.c static int perf_env__build_socket_map(struct perf_env *env, struct perf_cpu_map *cpus, cpus 1014 tools/perf/builtin-stat.c return cpu_map__build_map(cpus, sockp, perf_env__get_socket, env); cpus 1017 tools/perf/builtin-stat.c static int perf_env__build_die_map(struct perf_env *env, struct perf_cpu_map *cpus, cpus 1020 tools/perf/builtin-stat.c return cpu_map__build_map(cpus, diep, perf_env__get_die, env); cpus 1023 tools/perf/builtin-stat.c static int perf_env__build_core_map(struct perf_env *env, struct perf_cpu_map *cpus, cpus 1026 tools/perf/builtin-stat.c return cpu_map__build_map(cpus, corep, perf_env__get_core, env); cpus 1052 tools/perf/builtin-stat.c if (perf_env__build_socket_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 1059 tools/perf/builtin-stat.c if (perf_env__build_die_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 1066 tools/perf/builtin-stat.c if (perf_env__build_core_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) { cpus 1487 tools/perf/builtin-stat.c if (perf_cpu_map__empty(st->cpus)) { cpus 1506 tools/perf/builtin-stat.c if (!st->cpus || !st->threads) cpus 1512 tools/perf/builtin-stat.c perf_evlist__set_maps(&evsel_list->core, st->cpus, st->threads); cpus 1546 tools/perf/builtin-stat.c struct perf_cpu_map *cpus; cpus 1548 tools/perf/builtin-stat.c if (st->cpus) { cpus 1553 tools/perf/builtin-stat.c cpus = cpu_map__new_data(&event->cpu_map.data); cpus 1554 tools/perf/builtin-stat.c if (!cpus) cpus 1557 tools/perf/builtin-stat.c st->cpus = cpus; cpus 1017 tools/perf/builtin-top.c if (evsel__open(counter, top->evlist->core.cpus, cpus 15 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(int)); cpus 17 tools/perf/lib/cpumap.c if (cpus != NULL) { cpus 18 tools/perf/lib/cpumap.c cpus->nr = 1; cpus 19 tools/perf/lib/cpumap.c cpus->map[0] = -1; cpus 20 tools/perf/lib/cpumap.c refcount_set(&cpus->refcnt, 1); cpus 23 tools/perf/lib/cpumap.c return cpus; cpus 50 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus; cpus 57 tools/perf/lib/cpumap.c cpus = malloc(sizeof(*cpus) + nr_cpus * sizeof(int)); cpus 58 tools/perf/lib/cpumap.c if (cpus != NULL) { cpus 62 tools/perf/lib/cpumap.c cpus->map[i] = i; cpus 64 tools/perf/lib/cpumap.c cpus->nr = nr_cpus; cpus 65 tools/perf/lib/cpumap.c refcount_set(&cpus->refcnt, 1); cpus 68 tools/perf/lib/cpumap.c return cpus; cpus 74 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + payload_size); cpus 76 tools/perf/lib/cpumap.c if (cpus != NULL) { cpus 77 tools/perf/lib/cpumap.c cpus->nr = nr_cpus; cpus 78 tools/perf/lib/cpumap.c memcpy(cpus->map, tmp_cpus, payload_size); cpus 79 tools/perf/lib/cpumap.c refcount_set(&cpus->refcnt, 1); cpus 82 tools/perf/lib/cpumap.c return cpus; cpus 87 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus = NULL; cpus 135 tools/perf/lib/cpumap.c cpus = cpu_map__trim_new(nr_cpus, tmp_cpus); cpus 137 tools/perf/lib/cpumap.c cpus = cpu_map__default_new(); cpus 140 tools/perf/lib/cpumap.c return cpus; cpus 145 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus = NULL; cpus 152 tools/perf/lib/cpumap.c cpus = perf_cpu_map__read(onlnf); cpus 154 tools/perf/lib/cpumap.c return cpus; cpus 159 tools/perf/lib/cpumap.c struct perf_cpu_map *cpus = NULL; cpus 223 tools/perf/lib/cpumap.c cpus = cpu_map__trim_new(nr_cpus, tmp_cpus); cpus 225 tools/perf/lib/cpumap.c cpus = cpu_map__default_new(); cpus 227 tools/perf/lib/cpumap.c cpus = perf_cpu_map__dummy_new(); cpus 231 tools/perf/lib/cpumap.c return cpus; cpus 234 tools/perf/lib/cpumap.c int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx) cpus 236 tools/perf/lib/cpumap.c if (idx < cpus->nr) cpus 237 tools/perf/lib/cpumap.c return cpus->map[idx]; cpus 242 tools/perf/lib/cpumap.c int perf_cpu_map__nr(const struct perf_cpu_map *cpus) cpus 244 tools/perf/lib/cpumap.c return cpus ? cpus->nr : 1; cpus 252 tools/perf/lib/cpumap.c int perf_cpu_map__idx(struct perf_cpu_map *cpus, int cpu) cpus 256 tools/perf/lib/cpumap.c for (i = 0; i < cpus->nr; ++i) { cpus 257 tools/perf/lib/cpumap.c if (cpus->map[i] == cpu) cpus 40 tools/perf/lib/evlist.c perf_cpu_map__put(evsel->cpus); cpus 41 tools/perf/lib/evlist.c evsel->cpus = perf_cpu_map__get(evlist->cpus); cpus 42 tools/perf/lib/evlist.c } else if (evsel->cpus != evsel->own_cpus) { cpus 43 tools/perf/lib/evlist.c perf_cpu_map__put(evsel->cpus); cpus 44 tools/perf/lib/evlist.c evsel->cpus = perf_cpu_map__get(evsel->own_cpus); cpus 110 tools/perf/lib/evlist.c struct perf_cpu_map *cpus, cpus 120 tools/perf/lib/evlist.c if (cpus != evlist->cpus) { cpus 121 tools/perf/lib/evlist.c perf_cpu_map__put(evlist->cpus); cpus 122 tools/perf/lib/evlist.c evlist->cpus = perf_cpu_map__get(cpus); cpus 139 tools/perf/lib/evlist.c err = perf_evsel__open(evsel, evsel->cpus, evsel->threads); cpus 248 tools/perf/lib/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->cpus); cpus 66 tools/perf/lib/evsel.c int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus, cpus 71 tools/perf/lib/evsel.c if (cpus == NULL) { cpus 80 tools/perf/lib/evsel.c cpus = empty_cpu_map; cpus 96 tools/perf/lib/evsel.c perf_evsel__alloc_fd(evsel, cpus->nr, threads->nr) < 0) cpus 99 tools/perf/lib/evsel.c for (cpu = 0; cpu < cpus->nr; cpu++) { cpus 105 tools/perf/lib/evsel.c cpus->map[cpu], -1, 0); cpus 221 tools/perf/lib/evsel.c return evsel->cpus; cpus 17 tools/perf/lib/include/internal/cpumap.h int perf_cpu_map__idx(struct perf_cpu_map *cpus, int cpu); cpus 19 tools/perf/lib/include/internal/evlist.h struct perf_cpu_map *cpus; cpus 40 tools/perf/lib/include/internal/evsel.h struct perf_cpu_map *cpus; cpus 16 tools/perf/lib/include/perf/cpumap.h LIBPERF_API int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx); cpus 17 tools/perf/lib/include/perf/cpumap.h LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus); cpus 21 tools/perf/lib/include/perf/cpumap.h #define perf_cpu_map__for_each_cpu(cpu, idx, cpus) \ cpus 22 tools/perf/lib/include/perf/cpumap.h for ((idx) = 0, (cpu) = perf_cpu_map__cpu(cpus, idx); \ cpus 23 tools/perf/lib/include/perf/cpumap.h (idx) < perf_cpu_map__nr(cpus); \ cpus 24 tools/perf/lib/include/perf/cpumap.h (idx)++, (cpu) = perf_cpu_map__cpu(cpus, idx)) cpus 159 tools/perf/lib/include/perf/event.h struct perf_record_cpu_map_data cpus; cpus 32 tools/perf/lib/include/perf/evlist.h struct perf_cpu_map *cpus, cpus 28 tools/perf/lib/include/perf/evsel.h LIBPERF_API int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus, cpus 15 tools/perf/lib/tests/test-cpumap.c struct perf_cpu_map *cpus; cpus 21 tools/perf/lib/tests/test-cpumap.c cpus = perf_cpu_map__dummy_new(); cpus 22 tools/perf/lib/tests/test-cpumap.c if (!cpus) cpus 25 tools/perf/lib/tests/test-cpumap.c perf_cpu_map__get(cpus); cpus 26 tools/perf/lib/tests/test-cpumap.c perf_cpu_map__put(cpus); cpus 27 tools/perf/lib/tests/test-cpumap.c perf_cpu_map__put(cpus); cpus 19 tools/perf/lib/tests/test-evlist.c struct perf_cpu_map *cpus; cpus 32 tools/perf/lib/tests/test-evlist.c cpus = perf_cpu_map__new(NULL); cpus 33 tools/perf/lib/tests/test-evlist.c __T("failed to create cpus", cpus); cpus 48 tools/perf/lib/tests/test-evlist.c perf_evlist__set_maps(evlist, cpus, NULL); cpus 54 tools/perf/lib/tests/test-evlist.c cpus = perf_evsel__cpus(evsel); cpus 56 tools/perf/lib/tests/test-evlist.c perf_cpu_map__for_each_cpu(cpu, tmp, cpus) { cpus 67 tools/perf/lib/tests/test-evlist.c perf_cpu_map__put(cpus); cpus 18 tools/perf/lib/tests/test-evsel.c struct perf_cpu_map *cpus; cpus 26 tools/perf/lib/tests/test-evsel.c cpus = perf_cpu_map__new(NULL); cpus 27 tools/perf/lib/tests/test-evsel.c __T("failed to create cpus", cpus); cpus 32 tools/perf/lib/tests/test-evsel.c err = perf_evsel__open(evsel, cpus, NULL); cpus 35 tools/perf/lib/tests/test-evsel.c perf_cpu_map__for_each_cpu(cpu, tmp, cpus) { cpus 45 tools/perf/lib/tests/test-evsel.c perf_cpu_map__put(cpus); cpus 568 tools/perf/tests/code-reading.c struct perf_cpu_map *cpus = NULL; cpus 628 tools/perf/tests/code-reading.c cpus = perf_cpu_map__new(NULL); cpus 629 tools/perf/tests/code-reading.c if (!cpus) { cpus 643 tools/perf/tests/code-reading.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 670 tools/perf/tests/code-reading.c perf_cpu_map__get(cpus); cpus 720 tools/perf/tests/code-reading.c perf_cpu_map__put(cpus); cpus 54 tools/perf/tests/cpumap.c struct cpu_map_entries *cpus; cpus 62 tools/perf/tests/cpumap.c cpus = (struct cpu_map_entries *)data->data; cpus 64 tools/perf/tests/cpumap.c TEST_ASSERT_VAL("wrong nr", cpus->nr == 2); cpus 65 tools/perf/tests/cpumap.c TEST_ASSERT_VAL("wrong cpu", cpus->cpu[0] == 1); cpus 66 tools/perf/tests/cpumap.c TEST_ASSERT_VAL("wrong cpu", cpus->cpu[1] == 256); cpus 80 tools/perf/tests/cpumap.c struct perf_cpu_map *cpus; cpus 83 tools/perf/tests/cpumap.c cpus = perf_cpu_map__new("0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19"); cpus 86 tools/perf/tests/cpumap.c !perf_event__synthesize_cpu_map(NULL, cpus, process_event_mask, NULL)); cpus 88 tools/perf/tests/cpumap.c perf_cpu_map__put(cpus); cpus 91 tools/perf/tests/cpumap.c cpus = perf_cpu_map__new("1,256"); cpus 94 tools/perf/tests/cpumap.c !perf_event__synthesize_cpu_map(NULL, cpus, process_event_cpus, NULL)); cpus 96 tools/perf/tests/cpumap.c perf_cpu_map__put(cpus); cpus 115 tools/perf/tests/event-times.c struct perf_cpu_map *cpus; cpus 120 tools/perf/tests/event-times.c cpus = perf_cpu_map__new("0"); cpus 121 tools/perf/tests/event-times.c if (cpus == NULL) { cpus 128 tools/perf/tests/event-times.c err = perf_evsel__open_per_cpu(evsel, cpus); cpus 137 tools/perf/tests/event-times.c perf_cpu_map__put(cpus); cpus 144 tools/perf/tests/event-times.c struct perf_cpu_map *cpus; cpus 149 tools/perf/tests/event-times.c cpus = perf_cpu_map__new("0"); cpus 150 tools/perf/tests/event-times.c if (cpus == NULL) { cpus 155 tools/perf/tests/event-times.c err = perf_evsel__open_per_cpu(evsel, cpus); cpus 159 tools/perf/tests/event-times.c perf_cpu_map__put(cpus); cpus 74 tools/perf/tests/event_update.c map = cpu_map__new_data(&ev_data->cpus); cpus 74 tools/perf/tests/keep-tracking.c struct perf_cpu_map *cpus = NULL; cpus 83 tools/perf/tests/keep-tracking.c cpus = perf_cpu_map__new(NULL); cpus 84 tools/perf/tests/keep-tracking.c CHECK_NOT_NULL__(cpus); cpus 89 tools/perf/tests/keep-tracking.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 157 tools/perf/tests/keep-tracking.c perf_cpu_map__put(cpus); cpus 36 tools/perf/tests/mmap-basic.c struct perf_cpu_map *cpus; cpus 54 tools/perf/tests/mmap-basic.c cpus = perf_cpu_map__new(NULL); cpus 55 tools/perf/tests/mmap-basic.c if (cpus == NULL) { cpus 61 tools/perf/tests/mmap-basic.c CPU_SET(cpus->map[0], &cpu_set); cpus 65 tools/perf/tests/mmap-basic.c cpus->map[0], str_error_r(errno, sbuf, sizeof(sbuf))); cpus 75 tools/perf/tests/mmap-basic.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 92 tools/perf/tests/mmap-basic.c if (evsel__open(evsels[i], cpus, threads) < 0) { cpus 160 tools/perf/tests/mmap-basic.c cpus = NULL; cpus 163 tools/perf/tests/mmap-basic.c perf_cpu_map__put(cpus); cpus 26 tools/perf/tests/openat-syscall-all-cpus.c struct perf_cpu_map *cpus; cpus 39 tools/perf/tests/openat-syscall-all-cpus.c cpus = perf_cpu_map__new(NULL); cpus 40 tools/perf/tests/openat-syscall-all-cpus.c if (cpus == NULL) { cpus 54 tools/perf/tests/openat-syscall-all-cpus.c if (evsel__open(evsel, cpus, threads) < 0) { cpus 61 tools/perf/tests/openat-syscall-all-cpus.c for (cpu = 0; cpu < cpus->nr; ++cpu) { cpus 69 tools/perf/tests/openat-syscall-all-cpus.c if (cpus->map[cpu] >= CPU_SETSIZE) { cpus 70 tools/perf/tests/openat-syscall-all-cpus.c pr_debug("Ignoring CPU %d\n", cpus->map[cpu]); cpus 74 tools/perf/tests/openat-syscall-all-cpus.c CPU_SET(cpus->map[cpu], &cpu_set); cpus 77 tools/perf/tests/openat-syscall-all-cpus.c cpus->map[cpu], cpus 85 tools/perf/tests/openat-syscall-all-cpus.c CPU_CLR(cpus->map[cpu], &cpu_set); cpus 93 tools/perf/tests/openat-syscall-all-cpus.c if (perf_evsel__alloc_counts(evsel, cpus->nr, 1) < 0) { cpus 94 tools/perf/tests/openat-syscall-all-cpus.c pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr); cpus 100 tools/perf/tests/openat-syscall-all-cpus.c for (cpu = 0; cpu < cpus->nr; ++cpu) { cpus 103 tools/perf/tests/openat-syscall-all-cpus.c if (cpus->map[cpu] >= CPU_SETSIZE) cpus 115 tools/perf/tests/openat-syscall-all-cpus.c expected, cpus->map[cpu], perf_counts(evsel->counts, cpu, 0)->val); cpus 126 tools/perf/tests/openat-syscall-all-cpus.c perf_cpu_map__put(cpus); cpus 44 tools/perf/tests/sw-clock.c struct perf_cpu_map *cpus; cpus 63 tools/perf/tests/sw-clock.c cpus = perf_cpu_map__dummy_new(); cpus 65 tools/perf/tests/sw-clock.c if (!cpus || !threads) { cpus 71 tools/perf/tests/sw-clock.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 73 tools/perf/tests/sw-clock.c cpus = NULL; cpus 132 tools/perf/tests/sw-clock.c perf_cpu_map__put(cpus); cpus 336 tools/perf/tests/switch-tracking.c struct perf_cpu_map *cpus = NULL; cpus 349 tools/perf/tests/switch-tracking.c cpus = perf_cpu_map__new(NULL); cpus 350 tools/perf/tests/switch-tracking.c if (!cpus) { cpus 361 tools/perf/tests/switch-tracking.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 577 tools/perf/tests/switch-tracking.c perf_cpu_map__put(cpus); cpus 53 tools/perf/tests/task-exit.c struct perf_cpu_map *cpus; cpus 72 tools/perf/tests/task-exit.c cpus = perf_cpu_map__dummy_new(); cpus 74 tools/perf/tests/task-exit.c if (!cpus || !threads) { cpus 80 tools/perf/tests/task-exit.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 82 tools/perf/tests/task-exit.c cpus = NULL; cpus 152 tools/perf/tests/task-exit.c perf_cpu_map__put(cpus); cpus 136 tools/perf/util/auxtrace.c mp->cpu = evlist->core.cpus->map[idx]; cpus 21 tools/perf/util/cpumap.c static struct perf_cpu_map *cpu_map__from_entries(struct cpu_map_entries *cpus) cpus 25 tools/perf/util/cpumap.c map = perf_cpu_map__empty_new(cpus->nr); cpus 29 tools/perf/util/cpumap.c for (i = 0; i < cpus->nr; i++) { cpus 35 tools/perf/util/cpumap.c if (cpus->cpu[i] == (u16) -1) cpus 38 tools/perf/util/cpumap.c map->map[i] = (int) cpus->cpu[i]; cpus 83 tools/perf/util/cpumap.c struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(int) * nr); cpus 85 tools/perf/util/cpumap.c if (cpus != NULL) { cpus 88 tools/perf/util/cpumap.c cpus->nr = nr; cpus 90 tools/perf/util/cpumap.c cpus->map[i] = -1; cpus 92 tools/perf/util/cpumap.c refcount_set(&cpus->refcnt, 1); cpus 95 tools/perf/util/cpumap.c return cpus; cpus 131 tools/perf/util/cpumap.c int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res, cpus 136 tools/perf/util/cpumap.c int nr = cpus->nr; cpus 145 tools/perf/util/cpumap.c s1 = f(cpus, cpu, data); cpus 238 tools/perf/util/cpumap.c int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp) cpus 240 tools/perf/util/cpumap.c return cpu_map__build_map(cpus, sockp, cpu_map__get_socket, NULL); cpus 243 tools/perf/util/cpumap.c int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep) cpus 245 tools/perf/util/cpumap.c return cpu_map__build_map(cpus, diep, cpu_map__get_die, NULL); cpus 248 tools/perf/util/cpumap.c int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep) cpus 250 tools/perf/util/cpumap.c return cpu_map__build_map(cpus, corep, cpu_map__get_core, NULL); cpus 460 tools/perf/util/cpumap.c bool cpu_map__has(struct perf_cpu_map *cpus, int cpu) cpus 462 tools/perf/util/cpumap.c return perf_cpu_map__idx(cpus, cpu) != -1; cpus 465 tools/perf/util/cpumap.c int cpu_map__cpu(struct perf_cpu_map *cpus, int idx) cpus 467 tools/perf/util/cpumap.c return cpus->map[idx]; cpus 23 tools/perf/util/cpumap.h int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp); cpus 24 tools/perf/util/cpumap.h int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep); cpus 25 tools/perf/util/cpumap.h int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep); cpus 57 tools/perf/util/cpumap.h int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res, cpus 61 tools/perf/util/cpumap.h int cpu_map__cpu(struct perf_cpu_map *cpus, int idx); cpus 62 tools/perf/util/cpumap.h bool cpu_map__has(struct perf_cpu_map *cpus, int cpu); cpus 282 tools/perf/util/cputopo.c node->cpus = buf; cpus 350 tools/perf/util/cputopo.c zfree(&tp->nodes[i].cpus); cpus 17 tools/perf/util/cputopo.h char *cpus; cpus 297 tools/perf/util/event.c struct perf_cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data); cpus 302 tools/perf/util/event.c if (cpus) cpus 303 tools/perf/util/event.c ret += cpu_map__fprintf(cpus, fp); cpus 307 tools/perf/util/event.c perf_cpu_map__put(cpus); cpus 55 tools/perf/util/evlist.c void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus, cpus 59 tools/perf/util/evlist.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 151 tools/perf/util/evlist.c perf_cpu_map__put(evlist->core.cpus); cpus 153 tools/perf/util/evlist.c evlist->core.cpus = NULL; cpus 377 tools/perf/util/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->core.cpus); cpus 393 tools/perf/util/evlist.c bool per_cpu_mmaps = !perf_cpu_map__empty(evlist->core.cpus); cpus 432 tools/perf/util/evlist.c if (evlist->core.cpus && cpu >= 0) cpus 433 tools/perf/util/evlist.c sid->cpu = evlist->core.cpus->map[cpu]; cpus 600 tools/perf/util/evlist.c evlist->core.nr_mmaps = perf_cpu_map__nr(evlist->core.cpus); cpus 601 tools/perf/util/evlist.c if (perf_cpu_map__empty(evlist->core.cpus)) cpus 639 tools/perf/util/evlist.c int evlist_cpu = cpu_map__cpu(evlist->core.cpus, cpu_idx); cpus 666 tools/perf/util/evlist.c cpu = perf_cpu_map__idx(evsel->core.cpus, evlist_cpu); cpus 715 tools/perf/util/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->core.cpus); cpus 894 tools/perf/util/evlist.c const struct perf_cpu_map *cpus = evlist->core.cpus; cpus 922 tools/perf/util/evlist.c perf_evsel__alloc_id(&evsel->core, perf_cpu_map__nr(cpus), threads->nr) < 0) cpus 926 tools/perf/util/evlist.c if (perf_cpu_map__empty(cpus)) cpus 940 tools/perf/util/evlist.c struct perf_cpu_map *cpus; cpus 968 tools/perf/util/evlist.c cpus = perf_cpu_map__dummy_new(); cpus 970 tools/perf/util/evlist.c cpus = perf_cpu_map__new(target->cpu_list); cpus 972 tools/perf/util/evlist.c if (!cpus) cpus 977 tools/perf/util/evlist.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 1211 tools/perf/util/evlist.c struct perf_cpu_map *cpus; cpus 1224 tools/perf/util/evlist.c cpus = perf_cpu_map__new(NULL); cpus 1225 tools/perf/util/evlist.c if (!cpus) cpus 1232 tools/perf/util/evlist.c perf_evlist__set_maps(&evlist->core, cpus, threads); cpus 1236 tools/perf/util/evlist.c perf_cpu_map__put(cpus); cpus 1249 tools/perf/util/evlist.c if (evlist->core.threads == NULL && evlist->core.cpus == NULL) { cpus 1258 tools/perf/util/evlist.c err = evsel__open(evsel, evsel->core.cpus, evsel->core.threads); cpus 1766 tools/perf/util/evlist.c if (evsel__open(counter, evlist->core.cpus, cpus 87 tools/perf/util/evlist.h void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus, cpus 1252 tools/perf/util/evsel.c perf_cpu_map__put(evsel->core.cpus); cpus 1572 tools/perf/util/evsel.c int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus, cpus 1584 tools/perf/util/evsel.c if (cpus == NULL) { cpus 1593 tools/perf/util/evsel.c cpus = empty_cpu_map; cpus 1614 tools/perf/util/evsel.c perf_evsel__alloc_fd(&evsel->core, cpus->nr, nthreads) < 0) cpus 1650 tools/perf/util/evsel.c for (cpu = 0; cpu < cpus->nr; cpu++) { cpus 1662 tools/perf/util/evsel.c fd = perf_event_open(evsel, pid, cpus->map[cpu], cpus 1670 tools/perf/util/evsel.c if (ignore_missing_thread(evsel, cpus->nr, cpu, threads, thread, err)) { cpus 1832 tools/perf/util/evsel.c struct perf_cpu_map *cpus) cpus 1834 tools/perf/util/evsel.c return evsel__open(evsel, cpus, NULL); cpus 2539 tools/perf/util/evsel.c struct perf_cpu_map *cpus = evsel->core.cpus; cpus 2542 tools/perf/util/evsel.c if (perf_evsel__alloc_id(&evsel->core, cpus->nr, threads->nr)) cpus 225 tools/perf/util/evsel.h struct perf_cpu_map *cpus); cpus 228 tools/perf/util/evsel.h int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus, cpus 701 tools/perf/util/header.c ret = do_write_string(ff, n->cpus); cpus 3705 tools/perf/util/header.c map = cpu_map__new_data(&ev_cpus->cpus); cpus 3791 tools/perf/util/header.c map = cpu_map__new_data(&ev_cpus->cpus); cpus 327 tools/perf/util/parse-events.c struct perf_cpu_map *cpus = pmu ? pmu->cpus : cpus 337 tools/perf/util/parse-events.c evsel->core.cpus = perf_cpu_map__get(cpus); cpus 338 tools/perf/util/parse-events.c evsel->core.own_cpus = perf_cpu_map__get(cpus); cpus 580 tools/perf/util/pmu.c struct perf_cpu_map *cpus; cpus 586 tools/perf/util/pmu.c cpus = perf_cpu_map__read(file); cpus 588 tools/perf/util/pmu.c return cpus; cpus 601 tools/perf/util/pmu.c struct perf_cpu_map *cpus; cpus 615 tools/perf/util/pmu.c cpus = __pmu_cpumask(path); cpus 616 tools/perf/util/pmu.c if (cpus) cpus 617 tools/perf/util/pmu.c return cpus; cpus 626 tools/perf/util/pmu.c struct perf_cpu_map *cpus; cpus 630 tools/perf/util/pmu.c cpus = __pmu_cpumask(path); cpus 631 tools/perf/util/pmu.c perf_cpu_map__put(cpus); cpus 633 tools/perf/util/pmu.c return !!cpus; cpus 841 tools/perf/util/pmu.c pmu->cpus = pmu_cpumask(name); cpus 31 tools/perf/util/pmu.h struct perf_cpu_map *cpus; cpus 550 tools/perf/util/python.c struct perf_cpu_map *cpus; cpus 563 tools/perf/util/python.c pcpus->cpus = perf_cpu_map__new(cpustr); cpus 564 tools/perf/util/python.c if (pcpus->cpus == NULL) cpus 571 tools/perf/util/python.c perf_cpu_map__put(pcpus->cpus); cpus 579 tools/perf/util/python.c return pcpus->cpus->nr; cpus 586 tools/perf/util/python.c if (i >= pcpus->cpus->nr) cpus 589 tools/perf/util/python.c return Py_BuildValue("i", pcpus->cpus->map[i]); cpus 810 tools/perf/util/python.c struct perf_cpu_map *cpus = NULL; cpus 824 tools/perf/util/python.c cpus = ((struct pyrf_cpu_map *)pcpus)->cpus; cpus 831 tools/perf/util/python.c if (evsel__open(evsel, cpus, threads) < 0) { cpus 879 tools/perf/util/python.c struct perf_cpu_map *cpus; cpus 886 tools/perf/util/python.c cpus = ((struct pyrf_cpu_map *)pcpus)->cpus; cpus 887 tools/perf/util/python.c evlist__init(&pevlist->evlist, cpus, threads); cpus 67 tools/perf/util/record.c struct perf_cpu_map *cpus; cpus 70 tools/perf/util/record.c cpus = perf_cpu_map__new(NULL); cpus 71 tools/perf/util/record.c if (!cpus) cpus 73 tools/perf/util/record.c cpu = cpus->map[0]; cpus 74 tools/perf/util/record.c perf_cpu_map__put(cpus); cpus 122 tools/perf/util/record.c struct perf_cpu_map *cpus; cpus 125 tools/perf/util/record.c cpus = perf_cpu_map__new(NULL); cpus 126 tools/perf/util/record.c if (!cpus) cpus 128 tools/perf/util/record.c cpu = cpus->map[0]; cpus 129 tools/perf/util/record.c perf_cpu_map__put(cpus); cpus 154 tools/perf/util/record.c if (evlist->core.cpus->map[0] < 0) cpus 281 tools/perf/util/record.c if (!evlist || perf_cpu_map__empty(evlist->core.cpus)) { cpus 282 tools/perf/util/record.c struct perf_cpu_map *cpus = perf_cpu_map__new(NULL); cpus 284 tools/perf/util/record.c cpu = cpus ? cpus->map[0] : 0; cpus 285 tools/perf/util/record.c perf_cpu_map__put(cpus); cpus 287 tools/perf/util/record.c cpu = evlist->core.cpus->map[0]; cpus 1396 tools/perf/util/scripting-engines/trace-event-python.c struct perf_cpu_map *cpus = counter->core.cpus; cpus 1406 tools/perf/util/scripting-engines/trace-event-python.c for (cpu = 0; cpu < cpus->nr; cpu++) { cpus 1407 tools/perf/util/scripting-engines/trace-event-python.c process_stat(counter, cpus->map[cpu], cpus 849 tools/perf/util/session.c struct cpu_map_entries *cpus; cpus 857 tools/perf/util/session.c cpus = (struct cpu_map_entries *)data->data; cpus 859 tools/perf/util/session.c cpus->nr = bswap_16(cpus->nr); cpus 861 tools/perf/util/session.c for (i = 0; i < cpus->nr; i++) cpus 862 tools/perf/util/session.c cpus->cpu[i] = bswap_16(cpus->cpu[i]); cpus 331 tools/perf/util/stat-display.c if (config->aggr_get_id(config, evlist->core.cpus, cpu2) == id) cpus 504 tools/perf/util/stat-display.c s2 = config->aggr_get_id(config, evlist->core.cpus, cpu); cpus 749 tools/perf/util/stat-display.c int ncpus = perf_cpu_map__nr(counter->core.cpus); cpus 872 tools/perf/util/stat-display.c nrcpus = evlist->core.cpus->nr; cpus 242 tools/perf/util/stat.c struct perf_cpu_map *cpus = evsel__cpus(counter); cpus 250 tools/perf/util/stat.c if (perf_cpu_map__empty(cpus)) cpus 272 tools/perf/util/stat.c s = cpu_map__get_socket(cpus, cpu, NULL); cpus 85 tools/perf/util/svghelper.c void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end) cpus 107 tools/perf/util/svghelper.c total_height = (1 + rows + cpu2slot(cpus)) * SLOT_MULT; cpus 9 tools/perf/util/svghelper.h void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end); cpus 918 tools/perf/util/synthetic-events.c static void synthesize_cpus(struct cpu_map_entries *cpus, cpus 923 tools/perf/util/synthetic-events.c cpus->nr = map->nr; cpus 926 tools/perf/util/synthetic-events.c cpus->cpu[i] = map->map[i]; cpus 1798 tools/perf/util/synthetic-events.c err = perf_event__synthesize_cpu_map(tool, evlist->core.cpus, process, NULL); cpus 35 tools/perf/util/synthetic-events.h int perf_event__synthesize_cpu_map(struct perf_tool *tool, struct perf_cpu_map *cpus, perf_event__handler_t process, struct machine *machine); cpus 98 tools/perf/util/top.c top->evlist->core.cpus->nr > 1 ? "s" : "", cpus 105 tools/perf/util/top.c top->evlist->core.cpus->nr, cpus 106 tools/perf/util/top.c top->evlist->core.cpus->nr > 1 ? "s" : ""); cpus 133 tools/power/cpupower/lib/cpupower.c int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF); cpus 135 tools/power/cpupower/lib/cpupower.c cpu_top->core_info = malloc(sizeof(struct cpuid_core_info) * cpus); cpus 139 tools/power/cpupower/lib/cpupower.c for (cpu = 0; cpu < cpus; cpu++) { cpus 160 tools/power/cpupower/lib/cpupower.c qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info), cpus 167 tools/power/cpupower/lib/cpupower.c for(cpu = 1; cpu < cpus; cpu++) { cpus 185 tools/power/cpupower/lib/cpupower.c return cpus; cpus 393 tools/power/cpupower/utils/cpufreq-info.c struct cpufreq_affected_cpus *cpus = cpufreq_get_affected_cpus(cpu); cpus 396 tools/power/cpupower/utils/cpufreq-info.c if (!cpus) { cpus 401 tools/power/cpupower/utils/cpufreq-info.c while (cpus->next) { cpus 402 tools/power/cpupower/utils/cpufreq-info.c printf("%d ", cpus->cpu); cpus 403 tools/power/cpupower/utils/cpufreq-info.c cpus = cpus->next; cpus 405 tools/power/cpupower/utils/cpufreq-info.c printf("%d\n", cpus->cpu); cpus 406 tools/power/cpupower/utils/cpufreq-info.c cpufreq_put_affected_cpus(cpus); cpus 414 tools/power/cpupower/utils/cpufreq-info.c struct cpufreq_affected_cpus *cpus = cpufreq_get_related_cpus(cpu); cpus 417 tools/power/cpupower/utils/cpufreq-info.c if (!cpus) { cpus 422 tools/power/cpupower/utils/cpufreq-info.c while (cpus->next) { cpus 423 tools/power/cpupower/utils/cpufreq-info.c printf("%d ", cpus->cpu); cpus 424 tools/power/cpupower/utils/cpufreq-info.c cpus = cpus->next; cpus 426 tools/power/cpupower/utils/cpufreq-info.c printf("%d\n", cpus->cpu); cpus 427 tools/power/cpupower/utils/cpufreq-info.c cpufreq_put_related_cpus(cpus); cpus 295 tools/power/cpupower/utils/cpufreq-set.c struct cpufreq_affected_cpus *cpus; cpus 301 tools/power/cpupower/utils/cpufreq-set.c cpus = cpufreq_get_related_cpus(cpu); cpus 302 tools/power/cpupower/utils/cpufreq-set.c if (!cpus) cpus 304 tools/power/cpupower/utils/cpufreq-set.c while (cpus->next) { cpus 305 tools/power/cpupower/utils/cpufreq-set.c bitmask_setbit(cpus_chosen, cpus->cpu); cpus 306 tools/power/cpupower/utils/cpufreq-set.c cpus = cpus->next; cpus 309 tools/power/cpupower/utils/cpufreq-set.c bitmask_setbit(cpus_chosen, cpus->cpu); cpus 310 tools/power/cpupower/utils/cpufreq-set.c cpufreq_put_related_cpus(cpus); cpus 286 tools/power/x86/turbostat/turbostat.c } *cpus; cpus 949 tools/power/x86/turbostat/turbostat.c outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].die_id); cpus 957 tools/power/x86/turbostat/turbostat.c cpus[t->cpu_id].physical_node_id); cpus 2516 tools/power/x86/turbostat/turbostat.c if (cpus[i].put_ids) cpus 2517 tools/power/x86/turbostat/turbostat.c CPU_FREE(cpus[i].put_ids); cpus 2519 tools/power/x86/turbostat/turbostat.c free(cpus); cpus 2578 tools/power/x86/turbostat/turbostat.c cpus[cpu].logical_node_id = -1; cpus 2584 tools/power/x86/turbostat/turbostat.c if (cpus[cpu].physical_package_id != pkg) cpus 2587 tools/power/x86/turbostat/turbostat.c if (cpus[cpu].logical_node_id != -1) cpus 2589 tools/power/x86/turbostat/turbostat.c cpus[cpu].logical_node_id = lnode; cpus 2590 tools/power/x86/turbostat/turbostat.c node = cpus[cpu].physical_node_id; cpus 2597 tools/power/x86/turbostat/turbostat.c if ((cpus[cpux].physical_package_id == pkg) && cpus 2598 tools/power/x86/turbostat/turbostat.c (cpus[cpux].physical_node_id == node)) { cpus 2599 tools/power/x86/turbostat/turbostat.c cpus[cpux].logical_node_id = lnode; cpus 2665 tools/power/x86/turbostat/turbostat.c (cpus[so].thread_id < 0)) cpus 2666 tools/power/x86/turbostat/turbostat.c cpus[so].thread_id = cpus 2801 tools/power/x86/turbostat/turbostat.c cpus[cpu].thread_id = -1; cpus 4975 tools/power/x86/turbostat/turbostat.c cpus = calloc(1, (topo.max_cpu_num + 1) * sizeof(struct cpu_topology)); cpus 4976 tools/power/x86/turbostat/turbostat.c if (cpus == NULL) cpus 5022 tools/power/x86/turbostat/turbostat.c cpus[i].logical_cpu_id = i; cpus 5025 tools/power/x86/turbostat/turbostat.c cpus[i].physical_package_id = get_physical_package_id(i); cpus 5026 tools/power/x86/turbostat/turbostat.c if (cpus[i].physical_package_id > max_package_id) cpus 5027 tools/power/x86/turbostat/turbostat.c max_package_id = cpus[i].physical_package_id; cpus 5030 tools/power/x86/turbostat/turbostat.c cpus[i].die_id = get_die_id(i); cpus 5031 tools/power/x86/turbostat/turbostat.c if (cpus[i].die_id > max_die_id) cpus 5032 tools/power/x86/turbostat/turbostat.c max_die_id = cpus[i].die_id; cpus 5035 tools/power/x86/turbostat/turbostat.c cpus[i].physical_node_id = get_physical_node_id(&cpus[i]); cpus 5036 tools/power/x86/turbostat/turbostat.c if (cpus[i].physical_node_id > topo.max_node_num) cpus 5037 tools/power/x86/turbostat/turbostat.c topo.max_node_num = cpus[i].physical_node_id; cpus 5040 tools/power/x86/turbostat/turbostat.c cpus[i].physical_core_id = get_core_id(i); cpus 5041 tools/power/x86/turbostat/turbostat.c if (cpus[i].physical_core_id > max_core_id) cpus 5042 tools/power/x86/turbostat/turbostat.c max_core_id = cpus[i].physical_core_id; cpus 5045 tools/power/x86/turbostat/turbostat.c siblings = get_thread_siblings(&cpus[i]); cpus 5048 tools/power/x86/turbostat/turbostat.c if (cpus[i].thread_id == 0) cpus 5091 tools/power/x86/turbostat/turbostat.c i, cpus[i].physical_package_id, cpus[i].die_id, cpus 5092 tools/power/x86/turbostat/turbostat.c cpus[i].physical_node_id, cpus 5093 tools/power/x86/turbostat/turbostat.c cpus[i].logical_node_id, cpus 5094 tools/power/x86/turbostat/turbostat.c cpus[i].physical_core_id, cpus 5095 tools/power/x86/turbostat/turbostat.c cpus[i].thread_id); cpus 5142 tools/power/x86/turbostat/turbostat.c int pkg_id = cpus[cpu_id].physical_package_id; cpus 5143 tools/power/x86/turbostat/turbostat.c int node_id = cpus[cpu_id].logical_node_id; cpus 5144 tools/power/x86/turbostat/turbostat.c int core_id = cpus[cpu_id].physical_core_id; cpus 5145 tools/power/x86/turbostat/turbostat.c int thread_id = cpus[cpu_id].thread_id; cpus 233 tools/testing/selftests/net/reuseport_bpf_cpu.c int *rcv_fd, cpus; cpus 235 tools/testing/selftests/net/reuseport_bpf_cpu.c cpus = sysconf(_SC_NPROCESSORS_ONLN); cpus 236 tools/testing/selftests/net/reuseport_bpf_cpu.c if (cpus <= 0) cpus 239 tools/testing/selftests/net/reuseport_bpf_cpu.c rcv_fd = calloc(cpus, sizeof(int)); cpus 244 tools/testing/selftests/net/reuseport_bpf_cpu.c test(rcv_fd, cpus, AF_INET, SOCK_DGRAM); cpus 247 tools/testing/selftests/net/reuseport_bpf_cpu.c test(rcv_fd, cpus, AF_INET6, SOCK_DGRAM); cpus 250 tools/testing/selftests/net/reuseport_bpf_cpu.c test(rcv_fd, cpus, AF_INET, SOCK_STREAM); cpus 253 tools/testing/selftests/net/reuseport_bpf_cpu.c test(rcv_fd, cpus, AF_INET6, SOCK_STREAM); cpus 60 tools/virtio/virtio-trace/trace-agent.c s->cpus = get_total_cpus(); cpus 64 tools/virtio/virtio-trace/trace-agent.c for (i = 0; i < s->cpus; i++) cpus 157 tools/virtio/virtio-trace/trace-agent.c for (cpu = 0; cpu < s->cpus; cpu++) { cpus 224 tools/virtio/virtio-trace/trace-agent.c for (cpu = 0; cpu < s->cpus; cpu++) cpus 230 tools/virtio/virtio-trace/trace-agent.c for (cpu = 0; cpu < s->cpus; cpu++) { cpus 246 tools/virtio/virtio-trace/trace-agent.c for (i = 0; i < s->cpus; i++) { cpus 21 tools/virtio/virtio-trace/trace-agent.h int cpus; cpus 236 virt/kvm/kvm_main.c static inline bool kvm_kick_many_cpus(const struct cpumask *cpus, bool wait) cpus 238 virt/kvm/kvm_main.c if (unlikely(!cpus)) cpus 239 virt/kvm/kvm_main.c cpus = cpu_online_mask; cpus 241 virt/kvm/kvm_main.c if (cpumask_empty(cpus)) cpus 244 virt/kvm/kvm_main.c smp_call_function_many(cpus, ack_flush, NULL, wait); cpus 280 virt/kvm/kvm_main.c cpumask_var_t cpus; cpus 283 virt/kvm/kvm_main.c zalloc_cpumask_var(&cpus, GFP_ATOMIC); cpus 285 virt/kvm/kvm_main.c called = kvm_make_vcpus_request_mask(kvm, req, NULL, cpus); cpus 287 virt/kvm/kvm_main.c free_cpumask_var(cpus);