cpus               36 arch/arm/common/mcpm_entry.c 	mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN;
cpus               37 arch/arm/common/mcpm_entry.c 	sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
cpus               50 arch/arm/common/mcpm_entry.c 	mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN;
cpus               51 arch/arm/common/mcpm_entry.c 	sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
cpus              103 arch/arm/common/mcpm_entry.c 	sync_cache_r(&c->cpus);
cpus              111 arch/arm/common/mcpm_entry.c 			cpustate = c->cpus[i].cpu;
cpus              116 arch/arm/common/mcpm_entry.c 			sync_cache_r(&c->cpus[i].cpu);
cpus              439 arch/arm/common/mcpm_entry.c 			mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN;
cpus              445 arch/arm/common/mcpm_entry.c 		mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP;
cpus              288 arch/arm/include/asm/mcpm.h 	} cpus[MAX_CPUS_PER_CLUSTER];
cpus               74 arch/arm/kernel/devtree.c 	struct device_node *cpu, *cpus;
cpus               81 arch/arm/kernel/devtree.c 	cpus = of_find_node_by_path("/cpus");
cpus               83 arch/arm/kernel/devtree.c 	if (!cpus)
cpus              167 arch/arm/kernel/devtree.c 		set_smp_ops_by_method(cpus);
cpus              761 arch/arm/kernel/smp.c 	struct cpumask *cpus = freq->policy->cpus;
cpus              762 arch/arm/kernel/smp.c 	int cpu, first = cpumask_first(cpus);
cpus              769 arch/arm/kernel/smp.c 		for_each_cpu(cpu, cpus) {
cpus              789 arch/arm/kernel/smp.c 		for_each_cpu(cpu, cpus)
cpus              105 arch/arm/plat-samsung/include/plat/cpu.h 			 struct cpu_table *cpus, unsigned int cputab_size);
cpus             1097 arch/arm64/kernel/cpufeature.c 	static const struct midr_range cpus[] = {
cpus             1104 arch/arm64/kernel/cpufeature.c 	return is_midr_in_range_list(read_cpuid_id(), cpus);
cpus              295 arch/ia64/kernel/smp.c 	cpumask_var_t cpus;
cpus              304 arch/ia64/kernel/smp.c 	if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) {
cpus              308 arch/ia64/kernel/smp.c 		cpumask_copy(cpus, mm_cpumask(mm));
cpus              309 arch/ia64/kernel/smp.c 		smp_call_function_many(cpus,
cpus              311 arch/ia64/kernel/smp.c 		free_cpumask_var(cpus);
cpus              115 arch/ia64/mm/discontig.c 	unsigned long pernodesize = 0, cpus;
cpus              117 arch/ia64/mm/discontig.c 	cpus = early_nr_cpus_node(node);
cpus              118 arch/ia64/mm/discontig.c 	pernodesize += PERCPU_PAGE_SIZE * cpus;
cpus              263 arch/ia64/mm/discontig.c 	int cpus = early_nr_cpus_node(node);
cpus              270 arch/ia64/mm/discontig.c 	pernode += PERCPU_PAGE_SIZE * cpus;
cpus              142 arch/mips/cavium-octeon/smp.c 	int cpus;
cpus              161 arch/mips/cavium-octeon/smp.c 	cpus = 1;
cpus              164 arch/mips/cavium-octeon/smp.c 			set_cpu_possible(cpus, true);
cpus              165 arch/mips/cavium-octeon/smp.c 			set_cpu_present(cpus, true);
cpus              166 arch/mips/cavium-octeon/smp.c 			__cpu_number_map[id] = cpus;
cpus              167 arch/mips/cavium-octeon/smp.c 			__cpu_logical_map[cpus] = id;
cpus              168 arch/mips/cavium-octeon/smp.c 			cpus++;
cpus              181 arch/mips/cavium-octeon/smp.c 			set_cpu_possible(cpus, true);
cpus              182 arch/mips/cavium-octeon/smp.c 			__cpu_number_map[id] = cpus;
cpus              183 arch/mips/cavium-octeon/smp.c 			__cpu_logical_map[cpus] = id;
cpus              184 arch/mips/cavium-octeon/smp.c 			cpus++;
cpus             1023 arch/nds32/kernel/perf_event_cpu.c static DEVICE_ATTR(cpus, 0444, nds32_pmu_cpumask_show, NULL);
cpus              371 arch/powerpc/include/asm/rtas.h extern int rtas_online_cpus_mask(cpumask_var_t cpus);
cpus              372 arch/powerpc/include/asm/rtas.h extern int rtas_offline_cpus_mask(cpumask_var_t cpus);
cpus              852 arch/powerpc/kernel/rtas.c 				cpumask_var_t cpus)
cpus              854 arch/powerpc/kernel/rtas.c 	if (!cpumask_empty(cpus)) {
cpus              855 arch/powerpc/kernel/rtas.c 		cpumask_clear(cpus);
cpus              865 arch/powerpc/kernel/rtas.c 				cpumask_var_t cpus)
cpus              871 arch/powerpc/kernel/rtas.c 	if (cpumask_empty(cpus))
cpus              874 arch/powerpc/kernel/rtas.c 	for_each_cpu(cpu, cpus) {
cpus              894 arch/powerpc/kernel/rtas.c 				cpumask_shift_right(cpus, cpus, cpu);
cpus              895 arch/powerpc/kernel/rtas.c 				cpumask_shift_left(cpus, cpus, cpu);
cpus              899 arch/powerpc/kernel/rtas.c 				cpumask_clear_cpu(cpu, cpus);
cpus              909 arch/powerpc/kernel/rtas.c int rtas_online_cpus_mask(cpumask_var_t cpus)
cpus              913 arch/powerpc/kernel/rtas.c 	ret = rtas_cpu_state_change_mask(UP, cpus);
cpus              922 arch/powerpc/kernel/rtas.c 		cpumask_copy(tmp_mask, cpus);
cpus              930 arch/powerpc/kernel/rtas.c int rtas_offline_cpus_mask(cpumask_var_t cpus)
cpus              932 arch/powerpc/kernel/rtas.c 	return rtas_cpu_state_change_mask(DOWN, cpus);
cpus              355 arch/powerpc/mm/book3s64/radix_tlb.c 	struct cpumask *cpus = mm_cpumask(mm);
cpus              358 arch/powerpc/mm/book3s64/radix_tlb.c 	on_each_cpu_mask(cpus, do_tlbiel_pid, &t, 1);
cpus              494 arch/powerpc/mm/book3s64/radix_tlb.c 	struct cpumask *cpus = mm_cpumask(mm);
cpus              496 arch/powerpc/mm/book3s64/radix_tlb.c 	on_each_cpu_mask(cpus, do_tlbiel_va, &t, 1);
cpus              545 arch/powerpc/mm/book3s64/radix_tlb.c 	struct cpumask *cpus = mm_cpumask(mm);
cpus              550 arch/powerpc/mm/book3s64/radix_tlb.c 	on_each_cpu_mask(cpus, do_tlbiel_va_range, &t, 1);
cpus               94 arch/powerpc/platforms/cell/cpufreq_spudemand.c 	for_each_cpu(i, policy->cpus) {
cpus              117 arch/powerpc/platforms/cell/cpufreq_spudemand.c 	for_each_cpu (i, policy->cpus) {
cpus              632 arch/powerpc/platforms/powermac/smp.c 	struct device_node *cpus;
cpus              635 arch/powerpc/platforms/powermac/smp.c 	cpus = of_find_node_by_path("/cpus");
cpus              636 arch/powerpc/platforms/powermac/smp.c 	BUG_ON(cpus == NULL);
cpus              639 arch/powerpc/platforms/powermac/smp.c 	pmf_call_function(cpus, "cpu-timebase", &args);
cpus              640 arch/powerpc/platforms/powermac/smp.c 	of_node_put(cpus);
cpus              710 arch/powerpc/platforms/powermac/smp.c 		struct device_node *cpus =
cpus              712 arch/powerpc/platforms/powermac/smp.c 		if (cpus &&
cpus              713 arch/powerpc/platforms/powermac/smp.c 		    of_get_property(cpus, "platform-cpu-timebase", NULL)) {
cpus              766 arch/powerpc/platforms/powermac/smp.c 	struct device_node *cpus;
cpus              772 arch/powerpc/platforms/powermac/smp.c 	for_each_node_by_type(cpus, "cpu")
cpus               92 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_part_hdr *)hdr)->cpus;
cpus              174 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_phys_hdr *)hdr)->cpus;
cpus              176 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_x_phys_hdr *)hdr)->cpus;
cpus              131 arch/s390/include/asm/diag.h 	__u8 cpus;
cpus              138 arch/s390/include/asm/diag.h 	__u8  cpus;
cpus              190 arch/s390/include/asm/diag.h 	__u8 cpus;
cpus              197 arch/s390/include/asm/diag.h 	__u8 cpus;
cpus              224 arch/s390/include/asm/diag.h 	struct diag204_x_cpu_info cpus[];
cpus              229 arch/s390/include/asm/diag.h 	struct diag204_x_phys_cpu cpus[];
cpus              536 arch/s390/kernel/crash_dump.c 	int cpus = 0;
cpus              540 arch/s390/kernel/crash_dump.c 			cpus++;
cpus              541 arch/s390/kernel/crash_dump.c 	return cpus;
cpus              234 arch/s390/kernel/sthyi.c 	for (i = 0; i < block->hdr.cpus; i++) {
cpus              235 arch/s390/kernel/sthyi.c 		switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
cpus              237 arch/s390/kernel/sthyi.c 			if (block->cpus[i].weight == DED_WEIGHT)
cpus              243 arch/s390/kernel/sthyi.c 			if (block->cpus[i].weight == DED_WEIGHT)
cpus              263 arch/s390/kernel/sthyi.c 		if (!(block->cpus[i].cflag & DIAG204_CPU_ONLINE))
cpus              266 arch/s390/kernel/sthyi.c 		switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
cpus              269 arch/s390/kernel/sthyi.c 			if (block->cpus[i].cur_weight < DED_WEIGHT)
cpus              270 arch/s390/kernel/sthyi.c 				weight_cp |= block->cpus[i].cur_weight;
cpus              274 arch/s390/kernel/sthyi.c 			if (block->cpus[i].cur_weight < DED_WEIGHT)
cpus              275 arch/s390/kernel/sthyi.c 				weight_ifl |= block->cpus[i].cur_weight;
cpus              284 arch/s390/kernel/sthyi.c 		capped |= block->cpus[i].cflag & DIAG204_CPU_CAPPED;
cpus              285 arch/s390/kernel/sthyi.c 		cpu_inf->lpar_cap |= block->cpus[i].cpu_type_cap;
cpus              286 arch/s390/kernel/sthyi.c 		cpu_inf->lpar_grp_cap |= block->cpus[i].group_cpu_type_cap;
cpus              288 arch/s390/kernel/sthyi.c 		if (block->cpus[i].weight == DED_WEIGHT)
cpus              300 arch/s390/kernel/sthyi.c 	return (struct diag204_x_part_block *)&block->cpus[i];
cpus              490 arch/s390/kernel/time.c 	atomic_t cpus;
cpus              604 arch/s390/kernel/time.c 		while (atomic_read(&sync->cpus) != 0)
cpus              625 arch/s390/kernel/time.c 		atomic_dec(&sync->cpus);
cpus              671 arch/s390/kernel/time.c 	atomic_set(&stp_sync.cpus, num_online_cpus() - 1);
cpus              808 arch/s390/kvm/priv.c 	int cpus = 0;
cpus              811 arch/s390/kvm/priv.c 	cpus = atomic_read(&vcpu->kvm->online_vcpus);
cpus              822 arch/s390/kvm/priv.c 	mem->vm[0].cpus_total = cpus;
cpus              823 arch/s390/kvm/priv.c 	mem->vm[0].cpus_configured = cpus;
cpus              659 arch/sparc/kernel/time_64.c 	for_each_cpu(cpu, freq->policy->cpus) {
cpus               20 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
cpus               55 arch/x86/hyperv/mmu.c static void hyperv_flush_tlb_others(const struct cpumask *cpus,
cpus               64 arch/x86/hyperv/mmu.c 	trace_hyperv_mmu_flush_tlb_others(cpus, info);
cpus               69 arch/x86/hyperv/mmu.c 	if (cpumask_empty(cpus))
cpus               98 arch/x86/hyperv/mmu.c 	if (cpumask_equal(cpus, cpu_present_mask)) {
cpus              112 arch/x86/hyperv/mmu.c 		if (hv_cpu_number_to_vp_number(cpumask_last(cpus)) >= 64)
cpus              115 arch/x86/hyperv/mmu.c 		for_each_cpu(cpu, cpus) {
cpus              153 arch/x86/hyperv/mmu.c 	status = hyperv_flush_tlb_others_ex(cpus, info);
cpus              161 arch/x86/hyperv/mmu.c 	native_flush_tlb_others(cpus, info);
cpus              164 arch/x86/hyperv/mmu.c static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
cpus              196 arch/x86/hyperv/mmu.c 	nr_bank = cpumask_to_vpset(&(flush->hv_vp_set), cpus);
cpus               60 arch/x86/include/asm/amd_nb.h 	refcount_t		cpus;
cpus              209 arch/x86/include/asm/paravirt_types.h 	void (*flush_tlb_others)(const struct cpumask *cpus,
cpus               12 arch/x86/include/asm/trace/hyperv.h 	    TP_PROTO(const struct cpumask *cpus,
cpus               14 arch/x86/include/asm/trace/hyperv.h 	    TP_ARGS(cpus, info),
cpus               21 arch/x86/include/asm/trace/hyperv.h 	    TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
cpus               60 arch/x86/include/asm/trace/hyperv.h 	    TP_PROTO(const struct cpumask *cpus,
cpus               62 arch/x86/include/asm/trace/hyperv.h 	    TP_ARGS(cpus, vector),
cpus               67 arch/x86/include/asm/trace/hyperv.h 	    TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
cpus             1326 arch/x86/kernel/cpu/mce/amd.c 			refcount_inc(&b->cpus);
cpus             1347 arch/x86/kernel/cpu/mce/amd.c 		refcount_set(&b->cpus, 1);
cpus             1417 arch/x86/kernel/cpu/mce/amd.c 		if (!refcount_dec_and_test(&b->cpus)) {
cpus              962 arch/x86/kernel/cpu/mce/core.c 	int cpus = num_online_cpus();
cpus              978 arch/x86/kernel/cpu/mce/core.c 	while (atomic_read(&mce_callin) != cpus) {
cpus             1043 arch/x86/kernel/cpu/mce/core.c 		int cpus = num_online_cpus();
cpus             1049 arch/x86/kernel/cpu/mce/core.c 		while (atomic_read(&mce_executing) <= cpus) {
cpus              312 arch/x86/kernel/tsc_sync.c 	int cpus = 2;
cpus              334 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&start_count) != cpus - 1) {
cpus              349 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&stop_count) != cpus-1)
cpus              405 arch/x86/kernel/tsc_sync.c 	int cpus = 2;
cpus              431 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&start_count) != cpus)
cpus              449 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&stop_count) != cpus)
cpus             7113 arch/x86/kvm/x86.c 	for_each_cpu(cpu, freq->policy->cpus)
cpus             2196 arch/x86/platform/uv/tlb_uv.c 	int cpus;
cpus             2225 arch/x86/platform/uv/tlb_uv.c 		cpus = uv_blade_nr_possible_cpus(uvhub);
cpus             2226 arch/x86/platform/uv/tlb_uv.c 		if (cpus && (uv_blade_to_pnode(uvhub) < uv_base_pnode))
cpus             1348 arch/x86/xen/mmu_pv.c static void xen_flush_tlb_others(const struct cpumask *cpus,
cpus             1359 arch/x86/xen/mmu_pv.c 	trace_xen_mmu_flush_tlb_others(cpus, info->mm, info->start, info->end);
cpus             1361 arch/x86/xen/mmu_pv.c 	if (cpumask_empty(cpus))
cpus             1369 arch/x86/xen/mmu_pv.c 	cpumask_and(to_cpumask(args->mask), cpus, cpu_online_mask);
cpus               26 drivers/base/arch_topology.c void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq,
cpus               34 drivers/base/arch_topology.c 	for_each_cpu(i, cpus)
cpus             1466 drivers/base/power/domain.c 		cpumask_set_cpu(cpu, genpd->cpus);
cpus             1468 drivers/base/power/domain.c 		cpumask_clear_cpu(cpu, genpd->cpus);
cpus             1820 drivers/base/power/domain.c 	    !zalloc_cpumask_var(&genpd->cpus, GFP_KERNEL))
cpus             1828 drivers/base/power/domain.c 				free_cpumask_var(genpd->cpus);
cpus             1877 drivers/base/power/domain.c 		free_cpumask_var(genpd->cpus);
cpus              273 drivers/base/power/domain_governor.c 	for_each_cpu_and(cpu, genpd->cpus, cpu_online_mask) {
cpus               55 drivers/clk/sunxi/clk-sun9i-cpus.c 	struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw);
cpus               60 drivers/clk/sunxi/clk-sun9i-cpus.c 	reg = readl(cpus->reg);
cpus              155 drivers/clk/sunxi/clk-sun9i-cpus.c 	struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw);
cpus              162 drivers/clk/sunxi/clk-sun9i-cpus.c 	reg = readl(cpus->reg);
cpus              170 drivers/clk/sunxi/clk-sun9i-cpus.c 	writel(reg, cpus->reg);
cpus              188 drivers/clk/sunxi/clk-sun9i-cpus.c 	struct sun9i_a80_cpus_clk *cpus;
cpus              193 drivers/clk/sunxi/clk-sun9i-cpus.c 	cpus = kzalloc(sizeof(*cpus), GFP_KERNEL);
cpus              194 drivers/clk/sunxi/clk-sun9i-cpus.c 	if (!cpus)
cpus              197 drivers/clk/sunxi/clk-sun9i-cpus.c 	cpus->reg = of_io_request_and_map(node, 0, of_node_full_name(node));
cpus              198 drivers/clk/sunxi/clk-sun9i-cpus.c 	if (IS_ERR(cpus->reg))
cpus              211 drivers/clk/sunxi/clk-sun9i-cpus.c 	mux->reg = cpus->reg;
cpus              219 drivers/clk/sunxi/clk-sun9i-cpus.c 				     &cpus->hw, &sun9i_a80_cpus_clk_ops,
cpus              235 drivers/clk/sunxi/clk-sun9i-cpus.c 	iounmap(cpus->reg);
cpus              239 drivers/clk/sunxi/clk-sun9i-cpus.c 	kfree(cpus);
cpus              437 drivers/cpufreq/acpi-cpufreq.c 		cpumask_of(policy->cpu) : policy->cpus;
cpus              680 drivers/cpufreq/acpi-cpufreq.c 		cpumask_copy(policy->cpus, perf->shared_cpu_map);
cpus              688 drivers/cpufreq/acpi-cpufreq.c 		cpumask_copy(policy->cpus, topology_core_cpumask(cpu));
cpus              692 drivers/cpufreq/acpi-cpufreq.c 		cpumask_clear(policy->cpus);
cpus              693 drivers/cpufreq/acpi-cpufreq.c 		cpumask_set_cpu(cpu, policy->cpus);
cpus              473 drivers/cpufreq/arm_big_little.c 		cpumask_copy(policy->cpus, topology_core_cpumask(policy->cpu));
cpus              475 drivers/cpufreq/arm_big_little.c 		for_each_cpu(cpu, policy->cpus)
cpus              482 drivers/cpufreq/arm_big_little.c 	ret = get_cluster_clk_and_freq_table(cpu_dev, policy->cpus);
cpus              490 drivers/cpufreq/arm_big_little.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus              128 drivers/cpufreq/armada-8k-cpufreq.c 	struct cpumask cpus;
cpus              141 drivers/cpufreq/armada-8k-cpufreq.c 	cpumask_copy(&cpus, cpu_possible_mask);
cpus              148 drivers/cpufreq/armada-8k-cpufreq.c 	for_each_cpu(cpu, &cpus) {
cpus              178 drivers/cpufreq/armada-8k-cpufreq.c 		cpumask_andnot(&cpus, &cpus, &shared_cpus);
cpus              592 drivers/cpufreq/brcmstb-avs-cpufreq.c 	cpumask_setall(policy->cpus);
cpus              325 drivers/cpufreq/cppc_cpufreq.c 		cpumask_copy(policy->cpus, cpu->shared_cpu_map);
cpus              327 drivers/cpufreq/cppc_cpufreq.c 		for_each_cpu(i, policy->cpus) {
cpus              174 drivers/cpufreq/cpufreq-dt.c 	ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, policy->cpus);
cpus              184 drivers/cpufreq/cpufreq-dt.c 		if (dev_pm_opp_get_sharing_cpus(cpu_dev, policy->cpus))
cpus              222 drivers/cpufreq/cpufreq-dt.c 	if (!dev_pm_opp_of_cpumask_add_table(policy->cpus))
cpus              237 drivers/cpufreq/cpufreq-dt.c 		cpumask_setall(policy->cpus);
cpus              243 drivers/cpufreq/cpufreq-dt.c 		ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
cpus              278 drivers/cpufreq/cpufreq-dt.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus              286 drivers/cpufreq/cpufreq-dt.c 		dev_pm_opp_of_cpumask_remove_table(policy->cpus);
cpus              152 drivers/cpufreq/cpufreq.c __weak void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq,
cpus              176 drivers/cpufreq/cpufreq.c 	cpumask_setall(policy->cpus);
cpus              184 drivers/cpufreq/cpufreq.c 	return policy && cpumask_test_cpu(cpu, policy->cpus) ? policy : NULL;
cpus              376 drivers/cpufreq/cpufreq.c 			 cpumask_pr_args(policy->cpus));
cpus              378 drivers/cpufreq/cpufreq.c 		for_each_cpu(cpu, policy->cpus)
cpus              855 drivers/cpufreq/cpufreq.c 	return cpufreq_show_cpus(policy->cpus, buf);
cpus             1095 drivers/cpufreq/cpufreq.c 	if (cpumask_test_cpu(cpu, policy->cpus))
cpus             1102 drivers/cpufreq/cpufreq.c 	cpumask_set_cpu(cpu, policy->cpus);
cpus             1187 drivers/cpufreq/cpufreq.c 	if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL))
cpus             1218 drivers/cpufreq/cpufreq.c 			ret, cpumask_pr_args(policy->cpus));
cpus             1226 drivers/cpufreq/cpufreq.c 			ret, cpumask_pr_args(policy->cpus));
cpus             1250 drivers/cpufreq/cpufreq.c 	free_cpumask_var(policy->cpus);
cpus             1294 drivers/cpufreq/cpufreq.c 	free_cpumask_var(policy->cpus);
cpus             1337 drivers/cpufreq/cpufreq.c 		cpumask_copy(policy->cpus, policy->related_cpus);
cpus             1339 drivers/cpufreq/cpufreq.c 		cpumask_copy(policy->cpus, cpumask_of(cpu));
cpus             1357 drivers/cpufreq/cpufreq.c 		cpumask_copy(policy->related_cpus, policy->cpus);
cpus             1365 drivers/cpufreq/cpufreq.c 	cpumask_and(policy->cpus, policy->cpus, cpu_online_mask);
cpus             1551 drivers/cpufreq/cpufreq.c 	cpumask_clear_cpu(cpu, policy->cpus);
cpus             1561 drivers/cpufreq/cpufreq.c 		policy->cpu = cpumask_any(policy->cpus);
cpus              102 drivers/cpufreq/cpufreq_governor.c 		for_each_cpu(j, policy_dbs->policy->cpus) {
cpus              137 drivers/cpufreq/cpufreq_governor.c 	for_each_cpu(j, policy->cpus) {
cpus              331 drivers/cpufreq/cpufreq_governor.c 	for_each_cpu(cpu, policy->cpus) {
cpus              343 drivers/cpufreq/cpufreq_governor.c 	for_each_cpu(i, policy->cpus)
cpus              523 drivers/cpufreq/cpufreq_governor.c 	for_each_cpu(j, policy->cpus) {
cpus              439 drivers/cpufreq/cpufreq_ondemand.c 		cpumask_or(&done, &done, policy->cpus);
cpus              196 drivers/cpufreq/imx6q-cpufreq.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus               36 drivers/cpufreq/mediatek-cpufreq.c 	struct cpumask cpus;
cpus               54 drivers/cpufreq/mediatek-cpufreq.c 		if (cpumask_test_cpu(cpu, &info->cpus))
cpus              358 drivers/cpufreq/mediatek-cpufreq.c 	ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, &info->cpus);
cpus              365 drivers/cpufreq/mediatek-cpufreq.c 	ret = dev_pm_opp_of_cpumask_add_table(&info->cpus);
cpus              397 drivers/cpufreq/mediatek-cpufreq.c 	dev_pm_opp_of_cpumask_remove_table(&info->cpus);
cpus              423 drivers/cpufreq/mediatek-cpufreq.c 	dev_pm_opp_of_cpumask_remove_table(&info->cpus);
cpus              446 drivers/cpufreq/mediatek-cpufreq.c 	cpumask_copy(policy->cpus, &info->cpus);
cpus              451 drivers/cpufreq/mediatek-cpufreq.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus              134 drivers/cpufreq/omap-cpufreq.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus              111 drivers/cpufreq/p4-clockmod.c 	for_each_cpu(i, policy->cpus)
cpus              162 drivers/cpufreq/p4-clockmod.c 	cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
cpus             1074 drivers/cpufreq/powernow-k8.c 	cpumask_copy(pol->cpus, topology_core_cpumask(pol->cpu));
cpus             1075 drivers/cpufreq/powernow-k8.c 	data->available_cores = pol->cpus;
cpus             1082 drivers/cpufreq/powernow-k8.c 	for_each_cpu(cpu, pol->cpus)
cpus             1107 drivers/cpufreq/powernow-k8.c 	for_each_cpu(cpu, pol->cpus)
cpus              685 drivers/cpufreq/powernv-cpufreq.c 	if (!cpumask_test_cpu(raw_smp_processor_id(), policy->cpus)) {
cpus              687 drivers/cpufreq/powernv-cpufreq.c 		add_timer_on(&gpstates->timer, cpumask_first(policy->cpus));
cpus              822 drivers/cpufreq/powernv-cpufreq.c 	smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1);
cpus              835 drivers/cpufreq/powernv-cpufreq.c 		cpumask_set_cpu(base + i, policy->cpus);
cpus              924 drivers/cpufreq/powernv-cpufreq.c 		cpumask_andnot(&mask, &mask, policy.cpus);
cpus              109 drivers/cpufreq/ppc_cbe_cpufreq.c 	cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu));
cpus              147 drivers/cpufreq/qcom-cpufreq-hw.c 	dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
cpus              219 drivers/cpufreq/qcom-cpufreq-hw.c 	qcom_get_related_cpus(index, policy->cpus);
cpus              220 drivers/cpufreq/qcom-cpufreq-hw.c 	if (!cpumask_weight(policy->cpus)) {
cpus              241 drivers/cpufreq/qcom-cpufreq-hw.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus               93 drivers/cpufreq/qoriq-cpufreq.c 	struct cpumask *dstp = policy->cpus;
cpus              154 drivers/cpufreq/scmi-cpufreq.c 	ret = scmi_get_sharing_cpus(cpu_dev, policy->cpus);
cpus              160 drivers/cpufreq/scmi-cpufreq.c 	ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
cpus              203 drivers/cpufreq/scmi-cpufreq.c 	em_register_perf_domain(policy->cpus, nr_opp, &em_cb);
cpus              117 drivers/cpufreq/scpi-cpufreq.c 	ret = scpi_get_sharing_cpus(cpu_dev, policy->cpus);
cpus              123 drivers/cpufreq/scpi-cpufreq.c 	ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
cpus              172 drivers/cpufreq/scpi-cpufreq.c 	dev_pm_opp_of_register_em(policy->cpus);
cpus              440 drivers/cpufreq/speedstep-centrino.c 	for_each_cpu(j, policy->cpus) {
cpus              448 drivers/cpufreq/speedstep-centrino.c 			good_cpu = cpumask_any_and(policy->cpus,
cpus              263 drivers/cpufreq/speedstep-ich.c 	policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
cpus              296 drivers/cpufreq/speedstep-ich.c 	cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
cpus              298 drivers/cpufreq/speedstep-ich.c 	policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
cpus               21 drivers/cpufreq/tegra186-cpufreq.c 	int cpus[4];
cpus               30 drivers/cpufreq/tegra186-cpufreq.c 		.cpus = { 1, 2, NO_CPU, NO_CPU },
cpus               36 drivers/cpufreq/tegra186-cpufreq.c 		.cpus = { 0, 3, 4, 5 },
cpus               64 drivers/cpufreq/tegra186-cpufreq.c 		for (core = 0; core < ARRAY_SIZE(info->cpus); core++) {
cpus               65 drivers/cpufreq/tegra186-cpufreq.c 			if (info->cpus[core] == policy->cpu)
cpus               68 drivers/cpufreq/tegra186-cpufreq.c 		if (core == ARRAY_SIZE(info->cpus))
cpus              441 drivers/cpuidle/coupled.c 	cpumask_t cpus;
cpus              444 drivers/cpuidle/coupled.c 	cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus);
cpus              445 drivers/cpuidle/coupled.c 	ret = cpumask_and(&cpus, &cpuidle_coupled_poke_pending, &cpus);
cpus              628 drivers/cpuidle/coupled.c 	cpumask_t cpus;
cpus              629 drivers/cpuidle/coupled.c 	cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus);
cpus              630 drivers/cpuidle/coupled.c 	coupled->online_count = cpumask_weight(&cpus);
cpus              399 drivers/crypto/caam/qi.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              431 drivers/crypto/caam/qi.c 	if (!cpumask_test_cpu(*cpu, cpus)) {
cpus              434 drivers/crypto/caam/qi.c 		*pcpu = cpumask_next(*pcpu, cpus);
cpus              436 drivers/crypto/caam/qi.c 			*pcpu = cpumask_first(cpus);
cpus              507 drivers/crypto/caam/qi.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              509 drivers/crypto/caam/qi.c 	for_each_cpu(i, cpus) {
cpus              684 drivers/crypto/caam/qi.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              687 drivers/crypto/caam/qi.c 	for_each_cpu(i, cpus) {
cpus              701 drivers/crypto/caam/qi.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              703 drivers/crypto/caam/qi.c 	for_each_cpu(i, cpus)
cpus              712 drivers/crypto/caam/qi.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              736 drivers/crypto/caam/qi.c 	for_each_cpu(i, cpus) {
cpus              176 drivers/crypto/qat/qat_common/adf_isr.c 			unsigned int cpu, cpus = num_online_cpus();
cpus              191 drivers/crypto/qat/qat_common/adf_isr.c 			       i) % cpus;
cpus              162 drivers/crypto/qat/qat_common/qat_crypto.c 	int cpus = num_online_cpus();
cpus              164 drivers/crypto/qat/qat_common/qat_crypto.c 	int instances = min(cpus, banks);
cpus               77 drivers/firmware/psci/psci_checker.c static unsigned int down_and_up_cpus(const struct cpumask *cpus,
cpus               86 drivers/firmware/psci/psci_checker.c 	for_each_cpu(cpu, cpus) {
cpus              336 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
cpus              347 drivers/firmware/qcom_scm-32.c 	if (!cpus || (cpus && cpumask_empty(cpus)))
cpus              350 drivers/firmware/qcom_scm-32.c 	for_each_cpu(cpu, cpus) {
cpus              370 drivers/firmware/qcom_scm-32.c 				  const cpumask_t *cpus)
cpus              384 drivers/firmware/qcom_scm-32.c 	for_each_cpu(cpu, cpus) {
cpus              399 drivers/firmware/qcom_scm-32.c 		for_each_cpu(cpu, cpus)
cpus              167 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
cpus              182 drivers/firmware/qcom_scm-64.c 				  const cpumask_t *cpus)
cpus               98 drivers/firmware/qcom_scm.c int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
cpus              100 drivers/firmware/qcom_scm.c 	return __qcom_scm_set_cold_boot_addr(entry, cpus);
cpus              112 drivers/firmware/qcom_scm.c int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus)
cpus              114 drivers/firmware/qcom_scm.c 	return __qcom_scm_set_warm_boot_addr(__scm->dev, entry, cpus);
cpus               19 drivers/firmware/qcom_scm.h 		const cpumask_t *cpus);
cpus               20 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus);
cpus             1109 drivers/gpu/drm/vc4/vc4_hdmi.c 	dai_link->cpus		= &hdmi->audio.cpu;
cpus             1120 drivers/gpu/drm/vc4/vc4_hdmi.c 	dai_link->cpus->dai_name = dev_name(dev);
cpus               81 drivers/irqchip/irq-bcm6345-l1.c 	struct bcm6345_l1_cpu	*cpus[NR_CPUS];
cpus              124 drivers/irqchip/irq-bcm6345-l1.c 	cpu = intc->cpus[cpu_logical_map(smp_processor_id())];
cpus              126 drivers/irqchip/irq-bcm6345-l1.c 	cpu = intc->cpus[0];
cpus              159 drivers/irqchip/irq-bcm6345-l1.c 	intc->cpus[cpu_idx]->enable_cache[word] |= mask;
cpus              160 drivers/irqchip/irq-bcm6345-l1.c 	__raw_writel(intc->cpus[cpu_idx]->enable_cache[word],
cpus              161 drivers/irqchip/irq-bcm6345-l1.c 		intc->cpus[cpu_idx]->map_base + reg_enable(intc, word));
cpus              171 drivers/irqchip/irq-bcm6345-l1.c 	intc->cpus[cpu_idx]->enable_cache[word] &= ~mask;
cpus              172 drivers/irqchip/irq-bcm6345-l1.c 	__raw_writel(intc->cpus[cpu_idx]->enable_cache[word],
cpus              173 drivers/irqchip/irq-bcm6345-l1.c 		intc->cpus[cpu_idx]->map_base + reg_enable(intc, word));
cpus              220 drivers/irqchip/irq-bcm6345-l1.c 		enabled = intc->cpus[old_cpu]->enable_cache[word] & mask;
cpus              255 drivers/irqchip/irq-bcm6345-l1.c 	cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32),
cpus              340 drivers/irqchip/irq-bcm6345-l1.c 		struct bcm6345_l1_cpu *cpu = intc->cpus[idx];
cpus              350 drivers/irqchip/irq-bcm6345-l1.c 		struct bcm6345_l1_cpu *cpu = intc->cpus[idx];
cpus               41 drivers/irqchip/irq-bcm7038-l1.c 	struct bcm7038_l1_cpu	*cpus[NR_CPUS];
cpus              122 drivers/irqchip/irq-bcm7038-l1.c 	cpu = intc->cpus[cpu_logical_map(smp_processor_id())];
cpus              124 drivers/irqchip/irq-bcm7038-l1.c 	cpu = intc->cpus[0];
cpus              154 drivers/irqchip/irq-bcm7038-l1.c 	intc->cpus[cpu_idx]->mask_cache[word] &= ~mask;
cpus              155 drivers/irqchip/irq-bcm7038-l1.c 	l1_writel(mask, intc->cpus[cpu_idx]->map_base +
cpus              165 drivers/irqchip/irq-bcm7038-l1.c 	intc->cpus[cpu_idx]->mask_cache[word] |= mask;
cpus              166 drivers/irqchip/irq-bcm7038-l1.c 	l1_writel(mask, intc->cpus[cpu_idx]->map_base +
cpus              204 drivers/irqchip/irq-bcm7038-l1.c 	was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] &
cpus              265 drivers/irqchip/irq-bcm7038-l1.c 	cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32),
cpus              354 drivers/irqchip/irq-bcm7038-l1.c 		struct bcm7038_l1_cpu *cpu = intc->cpus[idx];
cpus               44 drivers/leds/trigger/ledtrig-activity.c 	int cpus;
cpus               56 drivers/leds/trigger/ledtrig-activity.c 	cpus = 0;
cpus               65 drivers/leds/trigger/ledtrig-activity.c 		cpus++;
cpus               73 drivers/leds/trigger/ledtrig-activity.c 	curr_boot = ktime_get_boottime_ns() * cpus;
cpus              130 drivers/leds/trigger/ledtrig-activity.c 	target = (cpus > 1) ? (100 / cpus) : 50;
cpus              168 drivers/misc/sgi-gru/grufile.c 	info.cpus = num_online_cpus();
cpus              146 drivers/misc/sgi-gru/grulib.h 	int		cpus;
cpus              297 drivers/misc/sgi-gru/grutlbpurge.c 	int cpus, shift = 0, n;
cpus              299 drivers/misc/sgi-gru/grutlbpurge.c 	cpus = uv_blade_nr_possible_cpus(gru->gs_blade_id);
cpus              302 drivers/misc/sgi-gru/grutlbpurge.c 	if (cpus) {
cpus              303 drivers/misc/sgi-gru/grutlbpurge.c 		n = 1 << fls(cpus - 1);
cpus              316 drivers/misc/sgi-gru/grutlbpurge.c 	gru->gs_tgh_first_remote = (cpus + (1 << shift) - 1) >> shift;
cpus             6317 drivers/net/ethernet/broadcom/bnx2.c 	int cpus = netif_get_num_default_rss_queues();
cpus             6321 drivers/net/ethernet/broadcom/bnx2.c 		msix_vecs = max(cpus + 1, bp->num_req_tx_rings);
cpus             6323 drivers/net/ethernet/broadcom/bnx2.c 		msix_vecs = max(cpus, bp->num_req_rx_rings);
cpus              699 drivers/net/ethernet/chelsio/cxgb3/common.h 		   const u8 * cpus, const u16 *rspq);
cpus              574 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	u8 cpus[SGE_QSETS + 1];
cpus              578 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		cpus[i] = i;
cpus              579 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	cpus[SGE_QSETS] = 0xff;	/* terminator */
cpus              589 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		      V_RRCPLCPUSIZE(6) | F_HASHTOEPLITZ, cpus, rspq_map);
cpus             2535 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 		   const u8 * cpus, const u16 *rspq)
cpus             2539 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	if (cpus)
cpus             2544 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 				val |= (cpus[cpu_idx++] & 0x3f) << (8 * j);
cpus             2545 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 				if (cpus[cpu_idx] == 0xff)
cpus              779 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              783 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	for_each_cpu_and(cpu, cpus, cpu_online_mask) {
cpus              553 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c 	const cpumask_t *cpus = qman_affine_cpus();
cpus              572 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c 	for_each_cpu_and(cpu, cpus, cpu_online_mask) {
cpus              589 drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c 	for_each_cpu_and(cpu, cpus, cpu_online_mask) {
cpus             1045 drivers/net/ethernet/freescale/enetc/enetc.c 	int cpus = num_online_cpus();
cpus             1054 drivers/net/ethernet/freescale/enetc/enetc.c 	priv->num_rx_rings = min_t(int, cpus, si->num_rx_rings);
cpus             1056 drivers/net/ethernet/freescale/enetc/enetc.c 	priv->bdr_int_num = cpus;
cpus             10928 drivers/net/ethernet/intel/i40e/i40e_main.c 	int cpus, extra_vectors;
cpus             10968 drivers/net/ethernet/intel/i40e/i40e_main.c 	cpus = num_online_cpus();
cpus             10969 drivers/net/ethernet/intel/i40e/i40e_main.c 	pf->num_lan_msix = min_t(int, cpus, vectors_left / 2);
cpus             11035 drivers/net/ethernet/intel/i40e/i40e_main.c 	extra_vectors = min_t(int, cpus - pf->num_lan_msix, vectors_left);
cpus             1446 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c 	int nrxqs, cpu, cpus = num_possible_cpus();
cpus             1449 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c 	nrxqs = port->nrxqs / cpus;
cpus             1460 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c 	return port->first_rxq + ((rxq * nrxqs + rxq / cpus) % port->nrxqs);
cpus             1103 drivers/opp/of.c void dev_pm_opp_of_register_em(struct cpumask *cpus)
cpus             1106 drivers/opp/of.c 	int ret, nr_opp, cpu = cpumask_first(cpus);
cpus             1135 drivers/opp/of.c 	em_register_perf_domain(cpus, nr_opp, &em_cb);
cpus              497 drivers/perf/arm_pmu.c static DEVICE_ATTR(cpus, S_IRUGO, armpmu_cpumask_show, NULL);
cpus              435 drivers/thermal/cpu_cooling.c 	num_cpus = cpumask_weight(cpufreq_cdev->policy->cpus);
cpus              892 fs/btrfs/compression.c 	int cpus = num_online_cpus();
cpus              916 fs/btrfs/compression.c 	if (atomic_read(total_ws) > cpus) {
cpus              921 fs/btrfs/compression.c 		if (atomic_read(total_ws) > cpus && !*free_ws)
cpus              132 include/asm-generic/mshyperv.h 				    const struct cpumask *cpus)
cpus              151 include/asm-generic/mshyperv.h 	for_each_cpu(cpu, cpus) {
cpus               53 include/linux/cpufreq.h 	cpumask_var_t		cpus;	/* Online CPUs only */
cpus              196 include/linux/cpufreq.h 	return cpumask_empty(policy->cpus);
cpus              201 include/linux/cpufreq.h 	return cpumask_weight(policy->cpus) > 1;
cpus              991 include/linux/cpufreq.h extern void arch_set_freq_scale(struct cpumask *cpus, unsigned long cur_freq,
cpus               40 include/linux/energy_model.h 	unsigned long cpus[0];
cpus               91 include/linux/energy_model.h 	cpu = cpumask_first(to_cpumask(pd->cpus));
cpus              667 include/linux/netdevice.h 	u16 cpus[0];
cpus              112 include/linux/pm_domain.h 	cpumask_var_t cpus;		/* A cpumask of the attached CPUs */
cpus              350 include/linux/pm_opp.h void dev_pm_opp_of_register_em(struct cpumask *cpus);
cpus              390 include/linux/pm_opp.h static inline void dev_pm_opp_of_register_em(struct cpumask *cpus)
cpus               38 include/linux/qcom_scm.h extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus);
cpus               39 include/linux/qcom_scm.h extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus);
cpus               68 include/linux/qcom_scm.h int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus)
cpus               73 include/linux/qcom_scm.h int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus)
cpus              125 include/linux/stop_machine.h int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
cpus              136 include/linux/stop_machine.h int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
cpus              139 include/linux/stop_machine.h 				   const struct cpumask *cpus);
cpus              143 include/linux/stop_machine.h 					  const struct cpumask *cpus)
cpus              154 include/linux/stop_machine.h 			       const struct cpumask *cpus)
cpus              156 include/linux/stop_machine.h 	return stop_machine_cpuslocked(fn, data, cpus);
cpus              160 include/linux/stop_machine.h 						 const struct cpumask *cpus)
cpus              162 include/linux/stop_machine.h 	return stop_machine(fn, data, cpus);
cpus               46 include/sound/simple_card_utils.h 		struct snd_soc_dai_link_component cpus;   /* single cpu */
cpus               85 include/sound/simple_card_utils.h 	asoc_simple_parse_clk(dev, node, simple_dai, dai_link->cpus)
cpus              101 include/sound/simple_card_utils.h 	asoc_simple_parse_dai(node, dai_link->cpus, is_single_link)
cpus              770 include/sound/soc.h 	struct snd_soc_dai_link_component *cpus;
cpus              917 include/sound/soc.h 	.cpus		= cpu,				\
cpus               96 include/trace/events/thermal.h 	TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load,
cpus               99 include/trace/events/thermal.h 	TP_ARGS(cpus, freq, load, load_len, dynamic_power),
cpus              110 include/trace/events/thermal.h 		__assign_bitmask(cpumask, cpumask_bits(cpus),
cpus              126 include/trace/events/thermal.h 	TP_PROTO(const struct cpumask *cpus, unsigned int freq,
cpus              129 include/trace/events/thermal.h 	TP_ARGS(cpus, freq, cdev_state, power),
cpus              139 include/trace/events/thermal.h 		__assign_bitmask(cpumask, cpumask_bits(cpus),
cpus              370 include/trace/events/xen.h 	    TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm,
cpus              372 include/trace/events/xen.h 	    TP_ARGS(cpus, mm, addr, end),
cpus              379 include/trace/events/xen.h 	    TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
cpus               63 kernel/power/energy_model.c 	debugfs_create_file("cpus", 0444, d, pd->cpus, &em_debug_cpus_fops);
cpus              158 kernel/power/energy_model.c 	cpumask_copy(to_cpumask(pd->cpus), span);
cpus               12 kernel/power/power.h 	int			cpus;
cpus               74 kernel/sched/cpufreq.c 	return cpumask_test_cpu(smp_processor_id(), policy->cpus) ||
cpus              130 kernel/sched/cpufreq_schedutil.c 		for_each_cpu(cpu, policy->cpus)
cpus              506 kernel/sched/cpufreq_schedutil.c 	for_each_cpu(j, policy->cpus) {
cpus              846 kernel/sched/cpufreq_schedutil.c 	for_each_cpu(cpu, policy->cpus) {
cpus              854 kernel/sched/cpufreq_schedutil.c 	for_each_cpu(cpu, policy->cpus) {
cpus              870 kernel/sched/cpufreq_schedutil.c 	for_each_cpu(cpu, policy->cpus)
cpus               57 kernel/sched/deadline.c 	int cpus = 0;
cpus               62 kernel/sched/deadline.c 		cpus++;
cpus               64 kernel/sched/deadline.c 	return cpus;
cpus             2559 kernel/sched/deadline.c 	int cpus, err = -1;
cpus             2574 kernel/sched/deadline.c 	cpus = dl_bw_cpus(task_cpu(p));
cpus             2576 kernel/sched/deadline.c 	    !__dl_overflow(dl_b, cpus, 0, new_bw)) {
cpus             2578 kernel/sched/deadline.c 			__dl_sub(dl_b, p->dl.dl_bw, cpus);
cpus             2579 kernel/sched/deadline.c 		__dl_add(dl_b, new_bw, cpus);
cpus             2582 kernel/sched/deadline.c 		   !__dl_overflow(dl_b, cpus, p->dl.dl_bw, new_bw)) {
cpus             2590 kernel/sched/deadline.c 		__dl_sub(dl_b, p->dl.dl_bw, cpus);
cpus             2591 kernel/sched/deadline.c 		__dl_add(dl_b, new_bw, cpus);
cpus             2721 kernel/sched/deadline.c 	int cpus, ret;
cpus             2729 kernel/sched/deadline.c 	cpus = dl_bw_cpus(dest_cpu);
cpus             2730 kernel/sched/deadline.c 	overflow = __dl_overflow(dl_b, cpus, 0, p->dl.dl_bw);
cpus             2740 kernel/sched/deadline.c 		__dl_add(dl_b, p->dl.dl_bw, cpus);
cpus             2775 kernel/sched/deadline.c 	int cpus;
cpus             2780 kernel/sched/deadline.c 	cpus = dl_bw_cpus(cpu);
cpus             2781 kernel/sched/deadline.c 	overflow = __dl_overflow(dl_b, cpus, 0, 0);
cpus              150 kernel/sched/fair.c 	unsigned int cpus = min_t(unsigned int, num_online_cpus(), 8);
cpus              158 kernel/sched/fair.c 		factor = cpus;
cpus              162 kernel/sched/fair.c 		factor = 1 + ilog2(cpus);
cpus             5885 kernel/sched/fair.c 	struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask);
cpus             5894 kernel/sched/fair.c 	cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
cpus             5896 kernel/sched/fair.c 	for_each_cpu_wrap(core, cpus, target) {
cpus             5900 kernel/sched/fair.c 			__cpumask_clear_cpu(cpu, cpus);
cpus             5960 kernel/sched/fair.c 	struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask);
cpus             5992 kernel/sched/fair.c 	cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
cpus             5994 kernel/sched/fair.c 	for_each_cpu_wrap(cpu, cpus, target) {
cpus             7136 kernel/sched/fair.c 	struct cpumask		*cpus;
cpus             7281 kernel/sched/fair.c 		for_each_cpu_and(cpu, env->dst_grpmask, env->cpus) {
cpus             8061 kernel/sched/fair.c 	for_each_cpu_and(i, sched_group_span(group), env->cpus) {
cpus             8643 kernel/sched/fair.c 	for_each_cpu_and(i, sched_group_span(group), env->cpus) {
cpus             8796 kernel/sched/fair.c 	if (!cpumask_test_cpu(env->dst_cpu, env->cpus))
cpus             8807 kernel/sched/fair.c 	for_each_cpu_and(cpu, group_balance_mask(sg), env->cpus) {
cpus             8838 kernel/sched/fair.c 	struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask);
cpus             8847 kernel/sched/fair.c 		.cpus		= cpus,
cpus             8852 kernel/sched/fair.c 	cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
cpus             8946 kernel/sched/fair.c 			__cpumask_clear_cpu(env.dst_cpu, env.cpus);
cpus             8973 kernel/sched/fair.c 			__cpumask_clear_cpu(cpu_of(busiest), cpus);
cpus             8982 kernel/sched/fair.c 			if (!cpumask_subset(cpus, env.dst_grpmask)) {
cpus              294 kernel/sched/sched.h void __dl_sub(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
cpus              297 kernel/sched/sched.h 	__dl_update(dl_b, (s32)tsk_bw / cpus);
cpus              301 kernel/sched/sched.h void __dl_add(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
cpus              304 kernel/sched/sched.h 	__dl_update(dl_b, -((s32)tsk_bw / cpus));
cpus              308 kernel/sched/sched.h bool __dl_overflow(struct dl_bw *dl_b, int cpus, u64 old_bw, u64 new_bw)
cpus              311 kernel/sched/sched.h 	       dl_b->bw * cpus < dl_b->total_bw - old_bw + new_bw;
cpus             1281 kernel/sched/sched.h extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu);
cpus             1286 kernel/sched/sched.h static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
cpus             2460 kernel/sched/sched.h #define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus)))
cpus             1736 kernel/sched/topology.c int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
cpus             1741 kernel/sched/topology.c 		cpu = cpumask_any_and(cpus, sched_domains_numa_masks[i][j]);
cpus              687 kernel/smp.c   	cpumask_var_t cpus;
cpus              692 kernel/smp.c   	if (likely(zalloc_cpumask_var(&cpus, (gfp_flags|__GFP_NOWARN)))) {
cpus              696 kernel/smp.c   				__cpumask_set_cpu(cpu, cpus);
cpus              697 kernel/smp.c   		on_each_cpu_mask(cpus, func, info, wait);
cpus              699 kernel/smp.c   		free_cpumask_var(cpus);
cpus              596 kernel/stop_machine.c 			    const struct cpumask *cpus)
cpus              602 kernel/stop_machine.c 		.active_cpus = cpus,
cpus              631 kernel/stop_machine.c int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus)
cpus              637 kernel/stop_machine.c 	ret = stop_machine_cpuslocked(fn, data, cpus);
cpus              666 kernel/stop_machine.c 				  const struct cpumask *cpus)
cpus              669 kernel/stop_machine.c 					    .active_cpus = cpus };
cpus              487 kernel/trace/ring_buffer.c 	int				cpus;
cpus             1407 kernel/trace/ring_buffer.c 	buffer->cpus = nr_cpu_ids;
cpus             4502 mm/slub.c      	DECLARE_BITMAP(cpus, NR_CPUS);
cpus             4578 mm/slub.c      						to_cpumask(l->cpus));
cpus             4608 mm/slub.c      	cpumask_clear(to_cpumask(l->cpus));
cpus             4609 mm/slub.c      	cpumask_set_cpu(track->cpu, to_cpumask(l->cpus));
cpus             4692 mm/slub.c      				!cpumask_empty(to_cpumask(l->cpus)) &&
cpus             4696 mm/slub.c      					 cpumask_pr_args(to_cpumask(l->cpus)));
cpus             4031 net/core/dev.c 		tcpu = map->cpus[reciprocal_scale(hash, map->len)];
cpus              699 net/core/net-sysfs.c 			cpumask_set_cpu(map->cpus[i], mask);
cpus              738 net/core/net-sysfs.c 		map->cpus[i++] = cpu;
cpus              241 samples/trace_events/trace-events-sample.h 		__bitmask(	cpus,	num_possible_cpus()	)
cpus              250 samples/trace_events/trace-events-sample.h 		__assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus());
cpus              297 samples/trace_events/trace-events-sample.h 		  __get_str(str), __get_bitmask(cpus))
cpus              626 scripts/dtc/livetree.c 	struct node *cpus, *bootcpu;
cpus              629 scripts/dtc/livetree.c 	cpus = get_node_by_path(tree, "/cpus");
cpus              630 scripts/dtc/livetree.c 	if (!cpus)
cpus              634 scripts/dtc/livetree.c 	bootcpu = cpus->children;
cpus              510 sound/soc/atmel/atmel-classd.c 	dai_link->cpus		= &comp[0];
cpus              521 sound/soc/atmel/atmel-classd.c 	dai_link->cpus->dai_name	= dev_name(dev);
cpus              521 sound/soc/atmel/atmel-pdmic.c 	dai_link->cpus		= &comp[0];
cpus              532 sound/soc/atmel/atmel-pdmic.c 	dai_link->cpus->dai_name	= dev_name(dev);
cpus              115 sound/soc/atmel/atmel_wm8904.c 	dailink->cpus->of_node = cpu_np;
cpus              144 sound/soc/atmel/atmel_wm8904.c 	id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc");
cpus              170 sound/soc/atmel/atmel_wm8904.c 	id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc");
cpus               96 sound/soc/atmel/mikroe-proto.c 	dai->cpus = &comp[0];
cpus              120 sound/soc/atmel/mikroe-proto.c 	dai->cpus->of_node = cpu_np;
cpus              212 sound/soc/atmel/sam9g20_wm8731.c 	at91sam9g20ek_dai.cpus->dai_name = NULL;
cpus              219 sound/soc/atmel/sam9g20_wm8731.c 	at91sam9g20ek_dai.cpus->of_node = cpu_np;
cpus              106 sound/soc/atmel/sam9x5_wm8731.c 	dai->cpus = &comp[0];
cpus              147 sound/soc/atmel/sam9x5_wm8731.c 	dai->cpus->of_node = cpu_np;
cpus              339 sound/soc/atmel/tse850-pcm5142.c 	dailink->cpus->of_node = cpu_np;
cpus              136 sound/soc/fsl/eukrea-tlv320.c 		eukrea_tlv320_dai.cpus->of_node = ssi_np;
cpus              139 sound/soc/fsl/eukrea-tlv320.c 		eukrea_tlv320_dai.cpus->dai_name = "imx-ssi.0";
cpus              634 sound/soc/fsl/fsl-asoc-card.c 	priv->dai_link[0].cpus->of_node = cpu_np;
cpus              665 sound/soc/fsl/fsl-asoc-card.c 		priv->dai_link[1].cpus->of_node = asrc_np;
cpus              671 sound/soc/fsl/fsl-asoc-card.c 		priv->dai_link[2].cpus->of_node = cpu_np;
cpus              243 sound/soc/fsl/imx-audmix.c 		priv->dai[i].cpus = &dlc[0];
cpus              255 sound/soc/fsl/imx-audmix.c 		priv->dai[i].cpus->of_node = args.np;
cpus              256 sound/soc/fsl/imx-audmix.c 		priv->dai[i].cpus->dai_name = dev_name(&cpu_pdev->dev);
cpus              272 sound/soc/fsl/imx-audmix.c 		priv->dai[num_dai + i].cpus = &dlc[3];
cpus              283 sound/soc/fsl/imx-audmix.c 		priv->dai[num_dai + i].cpus->of_node = audmix_np;
cpus              284 sound/soc/fsl/imx-audmix.c 		priv->dai[num_dai + i].cpus->dai_name = be_name;
cpus              161 sound/soc/fsl/imx-es8328.c 	data->dai.cpus		= &comp[0];
cpus              173 sound/soc/fsl/imx-es8328.c 	data->dai.cpus->of_node = ssi_np;
cpus              140 sound/soc/fsl/imx-sgtl5000.c 	data->dai.cpus		= &comp[0];
cpus              152 sound/soc/fsl/imx-sgtl5000.c 	data->dai.cpus->of_node = ssi_np;
cpus               35 sound/soc/fsl/imx-spdif.c 	data->dai.cpus		= &comp[0];
cpus               47 sound/soc/fsl/imx-spdif.c 	data->dai.cpus->of_node = spdif_np;
cpus              216 sound/soc/fsl/mpc8610_hpcd.c 	machine_data->dai[0].cpus	= &comp[0];
cpus              224 sound/soc/fsl/mpc8610_hpcd.c 	machine_data->dai[1].cpus	= &comp[3];
cpus              232 sound/soc/fsl/mpc8610_hpcd.c 	machine_data->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev);
cpus              226 sound/soc/fsl/p1022_ds.c 	mdata->dai[0].cpus	= &comp[0];
cpus              234 sound/soc/fsl/p1022_ds.c 	mdata->dai[1].cpus	= &comp[3];
cpus              243 sound/soc/fsl/p1022_ds.c 	mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev);
cpus              229 sound/soc/fsl/p1022_rdk.c 	mdata->dai[0].cpus	= &comp[0];
cpus              237 sound/soc/fsl/p1022_rdk.c 	mdata->dai[1].cpus	= &comp[3];
cpus              245 sound/soc/fsl/p1022_rdk.c 	mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev);
cpus              216 sound/soc/generic/audio-graph-card.c 	struct snd_soc_dai_link_component *cpus = dai_link->cpus;
cpus              257 sound/soc/generic/audio-graph-card.c 						   cpus->dai_name);
cpus              267 sound/soc/generic/audio-graph-card.c 		cpus->of_node		= NULL;
cpus              268 sound/soc/generic/audio-graph-card.c 		cpus->dai_name		= "snd-soc-dummy-dai";
cpus              269 sound/soc/generic/audio-graph-card.c 		cpus->name		= "snd-soc-dummy";
cpus              393 sound/soc/generic/audio-graph-card.c 					   dai_link->cpus->dai_name,
cpus              358 sound/soc/generic/simple-card-utils.c 		dai_link->platforms->of_node = dai_link->cpus->of_node;
cpus              382 sound/soc/generic/simple-card-utils.c 		dai_link->cpus->dai_name = NULL;
cpus              392 sound/soc/generic/simple-card-utils.c 		of_node_put(dai_link->cpus->of_node);
cpus              582 sound/soc/generic/simple-card-utils.c 		dai_link[i].cpus		= &dai_props[i].cpus;
cpus              125 sound/soc/generic/simple-card.c 	struct snd_soc_dai_link_component *cpus = dai_link->cpus;
cpus              174 sound/soc/generic/simple-card.c 						   cpus->dai_name);
cpus              183 sound/soc/generic/simple-card.c 		cpus->of_node		= NULL;
cpus              184 sound/soc/generic/simple-card.c 		cpus->dai_name		= "snd-soc-dummy-dai";
cpus              185 sound/soc/generic/simple-card.c 		cpus->name		= "snd-soc-dummy";
cpus              327 sound/soc/generic/simple-card.c 					   dai_link->cpus->dai_name,
cpus              653 sound/soc/generic/simple-card.c 		struct snd_soc_dai_link_component *cpus;
cpus              679 sound/soc/generic/simple-card.c 		cpus			= dai_link->cpus;
cpus              680 sound/soc/generic/simple-card.c 		cpus->dai_name		= cinfo->cpu_dai.name;
cpus              684 sound/soc/intel/boards/bxt_da7219_max98357a.c 				broxton_dais[i].cpus->dai_name = "SSP1 Pin";
cpus              690 sound/soc/intel/boards/bxt_da7219_max98357a.c 				broxton_dais[i].cpus->dai_name = "SSP2 Pin";
cpus              703 sound/soc/intel/boards/bxt_da7219_max98357a.c 				broxton_dais[i].cpus->dai_name = "SSP1 Pin";
cpus              709 sound/soc/intel/boards/bxt_da7219_max98357a.c 				broxton_dais[i].cpus->dai_name = "SSP0 Pin";
cpus              602 sound/soc/intel/boards/bxt_rt298.c 				card->dai_link[i].cpus->dai_name = "SSP5 Pin";
cpus              606 sound/soc/intel/boards/bxt_rt298.c 				card->dai_link[i].cpus->dai_name = "SSP2 Pin";
cpus              531 sound/soc/intel/boards/bytcht_es8316.c 		byt_cht_es8316_dais[dai_index].cpus->dai_name = "ssp0-port";
cpus             1299 sound/soc/intel/boards/bytcr_rt5640.c 		byt_rt5640_dais[dai_index].cpus->dai_name =
cpus             1061 sound/soc/intel/boards/bytcr_rt5651.c 		byt_rt5651_dais[dai_index].cpus->dai_name =
cpus              663 sound/soc/intel/boards/cht_bsw_rt5645.c 		cht_dailink[dai_index].cpus->dai_name =
cpus              417 sound/soc/intel/boards/sof_rt5682.c 	struct snd_soc_dai_link_component *cpus;
cpus              423 sound/soc/intel/boards/sof_rt5682.c 	cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component) *
cpus              425 sound/soc/intel/boards/sof_rt5682.c 	if (!links || !cpus)
cpus              445 sound/soc/intel/boards/sof_rt5682.c 	links[id].cpus = &cpus[id];
cpus              448 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
cpus              451 sound/soc/intel/boards/sof_rt5682.c 		if (!links[id].cpus->dai_name)
cpus              464 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
cpus              467 sound/soc/intel/boards/sof_rt5682.c 		if (!links[id].cpus->dai_name)
cpus              476 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus = &cpus[id];
cpus              477 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus->dai_name = "DMIC01 Pin";
cpus              482 sound/soc/intel/boards/sof_rt5682.c 			links[id + 1].cpus = &cpus[id + 1];
cpus              483 sound/soc/intel/boards/sof_rt5682.c 			links[id + 1].cpus->dai_name = "DMIC16k Pin";
cpus              516 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus = &cpus[id];
cpus              518 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
cpus              520 sound/soc/intel/boards/sof_rt5682.c 		if (!links[id].cpus->dai_name)
cpus              557 sound/soc/intel/boards/sof_rt5682.c 		links[id].cpus = &cpus[id];
cpus              560 sound/soc/intel/boards/sof_rt5682.c 			links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
cpus              563 sound/soc/intel/boards/sof_rt5682.c 			if (!links[id].cpus->dai_name)
cpus              567 sound/soc/intel/boards/sof_rt5682.c 			links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
cpus              570 sound/soc/intel/boards/sof_rt5682.c 			if (!links[id].cpus->dai_name)
cpus             1090 sound/soc/intel/skylake/skl-pcm.c 					dai_link->cpus->dai_name);
cpus              111 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[0].cpus->of_node =
cpus              114 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[0].platforms->of_node = a370db_dai[0].cpus->of_node;
cpus              120 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[1].cpus->of_node = a370db_dai[0].cpus->of_node;
cpus              121 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[1].platforms->of_node = a370db_dai[0].cpus->of_node;
cpus              127 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[2].cpus->of_node = a370db_dai[0].cpus->of_node;
cpus              128 sound/soc/kirkwood/armada-370-db.c 	a370db_dai[2].platforms->of_node = a370db_dai[0].cpus->of_node;
cpus              119 sound/soc/meson/axg-card.c 			if (link->cpus)
cpus              120 sound/soc/meson/axg-card.c 				of_node_put(link->cpus->of_node);
cpus              277 sound/soc/meson/axg-card.c 	lb->cpus = &dlc[0];
cpus              283 sound/soc/meson/axg-card.c 	lb->cpus->of_node = pad->cpus->of_node;
cpus              284 sound/soc/meson/axg-card.c 	lb->cpus->dai_name = "TDM Loopback";
cpus              299 sound/soc/meson/axg-card.c 	of_node_get(lb->cpus->of_node);
cpus              441 sound/soc/meson/axg-card.c 	link->dai_fmt = axg_card_parse_daifmt(node, link->cpus->of_node);
cpus              569 sound/soc/meson/axg-card.c 	dai_link->cpus = cpu;
cpus              572 sound/soc/meson/axg-card.c 	ret = axg_card_parse_dai(card, np, &dai_link->cpus->of_node,
cpus              573 sound/soc/meson/axg-card.c 				 &dai_link->cpus->dai_name);
cpus              577 sound/soc/meson/axg-card.c 	if (axg_card_cpu_is_playback_fe(dai_link->cpus->of_node))
cpus              579 sound/soc/meson/axg-card.c 	else if (axg_card_cpu_is_capture_fe(dai_link->cpus->of_node))
cpus              587 sound/soc/meson/axg-card.c 	if (axg_card_cpu_is_tdm_iface(dai_link->cpus->of_node))
cpus              589 sound/soc/meson/axg-card.c 	else if (axg_card_cpu_is_codec(dai_link->cpus->of_node))
cpus              127 sound/soc/mxs/mxs-sgtl5000.c 		mxs_sgtl5000_dai[i].cpus->dai_name = NULL;
cpus              128 sound/soc/mxs/mxs-sgtl5000.c 		mxs_sgtl5000_dai[i].cpus->of_node = saif_np[i];
cpus              158 sound/soc/qcom/apq8016_sbc.c 		link->cpus	= &dlc[0];
cpus              173 sound/soc/qcom/apq8016_sbc.c 		link->cpus->of_node = of_parse_phandle(cpu, "sound-dai", 0);
cpus              174 sound/soc/qcom/apq8016_sbc.c 		if (!link->cpus->of_node) {
cpus              180 sound/soc/qcom/apq8016_sbc.c 		ret = snd_soc_of_get_dai_name(cpu, &link->cpus->dai_name);
cpus              193 sound/soc/qcom/apq8016_sbc.c 		link->platforms->of_node = link->cpus->of_node;
cpus               50 sound/soc/qcom/common.c 		link->cpus	= &dlc[0];
cpus               78 sound/soc/qcom/common.c 		link->cpus->of_node = args.np;
cpus               81 sound/soc/qcom/common.c 		ret = snd_soc_of_get_dai_name(cpu, &link->cpus->dai_name);
cpus              112 sound/soc/qcom/common.c 			link->platforms->of_node = link->cpus->of_node;
cpus               73 sound/soc/qcom/storm.c 	dai_link->cpus->of_node = of_parse_phandle(np, "cpu", 0);
cpus               74 sound/soc/qcom/storm.c 	if (!dai_link->cpus->of_node) {
cpus               78 sound/soc/qcom/storm.c 	dai_link->platforms->of_node = dai_link->cpus->of_node;
cpus              232 sound/soc/rockchip/rk3288_hdmi_analog.c 	rk_dailink.cpus->of_node = of_parse_phandle(np, "rockchip,i2s-controller",
cpus              234 sound/soc/rockchip/rk3288_hdmi_analog.c 	if (!rk_dailink.cpus->of_node) {
cpus              240 sound/soc/rockchip/rk3288_hdmi_analog.c 	rk_dailink.platforms->of_node = rk_dailink.cpus->of_node;
cpus              514 sound/soc/rockchip/rk3399_gru_sound.c 		dai->cpus->of_node = np_cpu;
cpus              233 sound/soc/rockchip/rockchip_max98090.c 	rk_dailink.cpus->of_node = of_parse_phandle(np,
cpus              235 sound/soc/rockchip/rockchip_max98090.c 	if (!rk_dailink.cpus->of_node) {
cpus              241 sound/soc/rockchip/rockchip_max98090.c 	rk_dailink.platforms->of_node = rk_dailink.cpus->of_node;
cpus              172 sound/soc/rockchip/rockchip_rt5645.c 	rk_dailink.cpus->of_node = of_parse_phandle(np,
cpus              174 sound/soc/rockchip/rockchip_rt5645.c 	if (!rk_dailink.cpus->of_node) {
cpus              181 sound/soc/rockchip/rockchip_rt5645.c 	rk_dailink.platforms->of_node = rk_dailink.cpus->of_node;
cpus              200 sound/soc/rockchip/rockchip_rt5645.c 	of_node_put(rk_dailink.cpus->of_node);
cpus              201 sound/soc/rockchip/rockchip_rt5645.c 	rk_dailink.cpus->of_node = NULL;
cpus              211 sound/soc/rockchip/rockchip_rt5645.c 	of_node_put(rk_dailink.cpus->of_node);
cpus              212 sound/soc/rockchip/rockchip_rt5645.c 	rk_dailink.cpus->of_node = NULL;
cpus               84 sound/soc/samsung/arndale_rt5631.c 		of_node_put(dai_link->cpus->of_node);
cpus               98 sound/soc/samsung/arndale_rt5631.c 		if (!arndale_rt5631_dai[n].cpus->dai_name) {
cpus               99 sound/soc/samsung/arndale_rt5631.c 			arndale_rt5631_dai[n].cpus->of_node = of_parse_phandle(np,
cpus              102 sound/soc/samsung/arndale_rt5631.c 			if (!arndale_rt5631_dai[n].cpus->of_node) {
cpus              110 sound/soc/samsung/arndale_rt5631.c 					arndale_rt5631_dai[n].cpus->of_node;
cpus              274 sound/soc/samsung/odroid.c 		ret = snd_soc_get_dai_name(&args, &link->cpus->dai_name);
cpus              160 sound/soc/samsung/smdk_wm8994.c 		smdk_dai[0].cpus->dai_name = NULL;
cpus              161 sound/soc/samsung/smdk_wm8994.c 		smdk_dai[0].cpus->of_node = of_parse_phandle(np,
cpus              163 sound/soc/samsung/smdk_wm8994.c 		if (!smdk_dai[0].cpus->of_node) {
cpus              170 sound/soc/samsung/smdk_wm8994.c 		smdk_dai[0].platforms->of_node = smdk_dai[0].cpus->of_node;
cpus              149 sound/soc/samsung/snow.c 	link->cpus = links_cpus;
cpus              166 sound/soc/samsung/snow.c 		link->cpus->of_node = of_parse_phandle(cpu, "sound-dai", 0);
cpus              169 sound/soc/samsung/snow.c 		if (!link->cpus->of_node) {
cpus              179 sound/soc/samsung/snow.c 			of_node_put(link->cpus->of_node);
cpus              184 sound/soc/samsung/snow.c 		priv->clk_i2s_bus = of_clk_get_by_name(link->cpus->of_node,
cpus              188 sound/soc/samsung/snow.c 			of_node_put(link->cpus->of_node);
cpus              194 sound/soc/samsung/snow.c 		link->cpus->of_node = of_parse_phandle(dev->of_node,
cpus              196 sound/soc/samsung/snow.c 		if (!link->cpus->of_node) {
cpus              204 sound/soc/samsung/snow.c 			of_node_put(link->cpus->of_node);
cpus              210 sound/soc/samsung/snow.c 	link->platforms->of_node = link->cpus->of_node;
cpus              231 sound/soc/samsung/snow.c 	of_node_put(link->cpus->of_node);
cpus              576 sound/soc/samsung/tm2_wm5110.c 		dai_link->cpus->name = NULL;
cpus              583 sound/soc/samsung/tm2_wm5110.c 		dai_link->cpus->of_node = cpu_dai_node[dai_index];
cpus               99 sound/soc/sirf/sirf-audio.c 	sirf_audio_dai_link[0].cpus->of_node =
cpus              899 sound/soc/soc-core.c 	rtd->cpu_dai = snd_soc_find_dai(dai_link->cpus);
cpus              902 sound/soc/soc-core.c 			 dai_link->cpus->dai_name);
cpus             1320 sound/soc/soc-core.c 	if (link->cpus->name && link->cpus->of_node) {
cpus             1331 sound/soc/soc-core.c 	if ((link->cpus->of_node || link->cpus->name) &&
cpus             1332 sound/soc/soc-core.c 	    !soc_find_component(link->cpus))
cpus             1339 sound/soc/soc-core.c 	if (!link->cpus->dai_name &&
cpus             1340 sound/soc/soc-core.c 	    !(link->cpus->name || link->cpus->of_node)) {
cpus              556 sound/soc/soc-topology.c 	kfree(link->cpus->dai_name);
cpus             1908 sound/soc/soc-topology.c 	link->cpus	= &dlc[0];
cpus             1927 sound/soc/soc-topology.c 		link->cpus->dai_name = kstrdup(pcm->dai_name, GFP_KERNEL);
cpus             1962 sound/soc/soc-topology.c 	kfree(link->cpus->dai_name);
cpus               41 sound/soc/sof/nocodec.c 		links[i].cpus = &dlc[0];
cpus               51 sound/soc/sof/nocodec.c 		links[i].cpus->dai_name = ops->drv[i].name;
cpus             2427 sound/soc/sof/topology.c 			dai->cpu_dai_name = link->cpus->dai_name;
cpus             2700 sound/soc/sof/topology.c 			sof_dai->cpu_dai_name = link->cpus->dai_name;
cpus             2756 sound/soc/sof/topology.c 	dai = snd_soc_find_dai(link->cpus);
cpus             2759 sound/soc/sof/topology.c 			link->cpus->dai_name, __func__);
cpus             2934 sound/soc/sof/topology.c 	dai = snd_soc_find_dai(link->cpus);
cpus             2937 sound/soc/sof/topology.c 			link->cpus->dai_name, __func__);
cpus             1304 sound/soc/sunxi/sun4i-codec.c 	link->cpus	= &dlc[0];
cpus             1315 sound/soc/sunxi/sun4i-codec.c 	link->cpus->dai_name	= dev_name(dev);
cpus              189 sound/soc/tegra/tegra_alc5632.c 	tegra_alc5632_dai.cpus->of_node = of_parse_phandle(np,
cpus              191 sound/soc/tegra/tegra_alc5632.c 	if (!tegra_alc5632_dai.cpus->of_node) {
cpus              198 sound/soc/tegra/tegra_alc5632.c 	tegra_alc5632_dai.platforms->of_node = tegra_alc5632_dai.cpus->of_node;
cpus              216 sound/soc/tegra/tegra_alc5632.c 	of_node_put(tegra_alc5632_dai.cpus->of_node);
cpus              217 sound/soc/tegra/tegra_alc5632.c 	tegra_alc5632_dai.cpus->of_node = NULL;
cpus              235 sound/soc/tegra/tegra_alc5632.c 	of_node_put(tegra_alc5632_dai.cpus->of_node);
cpus              236 sound/soc/tegra/tegra_alc5632.c 	tegra_alc5632_dai.cpus->of_node = NULL;
cpus              236 sound/soc/tegra/tegra_max98090.c 	tegra_max98090_dai.cpus->of_node = of_parse_phandle(np,
cpus              238 sound/soc/tegra/tegra_max98090.c 	if (!tegra_max98090_dai.cpus->of_node) {
cpus              245 sound/soc/tegra/tegra_max98090.c 	tegra_max98090_dai.platforms->of_node = tegra_max98090_dai.cpus->of_node;
cpus              182 sound/soc/tegra/tegra_rt5640.c 	tegra_rt5640_dai.cpus->of_node = of_parse_phandle(np,
cpus              184 sound/soc/tegra/tegra_rt5640.c 	if (!tegra_rt5640_dai.cpus->of_node) {
cpus              191 sound/soc/tegra/tegra_rt5640.c 	tegra_rt5640_dai.platforms->of_node = tegra_rt5640_dai.cpus->of_node;
cpus              255 sound/soc/tegra/tegra_rt5677.c 	tegra_rt5677_dai.cpus->of_node = of_parse_phandle(np,
cpus              257 sound/soc/tegra/tegra_rt5677.c 	if (!tegra_rt5677_dai.cpus->of_node) {
cpus              263 sound/soc/tegra/tegra_rt5677.c 	tegra_rt5677_dai.platforms->of_node = tegra_rt5677_dai.cpus->of_node;
cpus              281 sound/soc/tegra/tegra_rt5677.c 	of_node_put(tegra_rt5677_dai.cpus->of_node);
cpus              282 sound/soc/tegra/tegra_rt5677.c 	tegra_rt5677_dai.cpus->of_node = NULL;
cpus              303 sound/soc/tegra/tegra_rt5677.c 	of_node_put(tegra_rt5677_dai.cpus->of_node);
cpus              304 sound/soc/tegra/tegra_rt5677.c 	tegra_rt5677_dai.cpus->of_node = NULL;
cpus              140 sound/soc/tegra/tegra_sgtl5000.c 	tegra_sgtl5000_dai.cpus->of_node = of_parse_phandle(np,
cpus              142 sound/soc/tegra/tegra_sgtl5000.c 	if (!tegra_sgtl5000_dai.cpus->of_node) {
cpus              149 sound/soc/tegra/tegra_sgtl5000.c 	tegra_sgtl5000_dai.platforms->of_node = tegra_sgtl5000_dai.cpus->of_node;
cpus              167 sound/soc/tegra/tegra_sgtl5000.c 	of_node_put(tegra_sgtl5000_dai.cpus->of_node);
cpus              168 sound/soc/tegra/tegra_sgtl5000.c 	tegra_sgtl5000_dai.cpus->of_node = NULL;
cpus              187 sound/soc/tegra/tegra_sgtl5000.c 	of_node_put(tegra_sgtl5000_dai.cpus->of_node);
cpus              188 sound/soc/tegra/tegra_sgtl5000.c 	tegra_sgtl5000_dai.cpus->of_node = NULL;
cpus              145 sound/soc/tegra/tegra_wm8753.c 	tegra_wm8753_dai.cpus->of_node = of_parse_phandle(np,
cpus              147 sound/soc/tegra/tegra_wm8753.c 	if (!tegra_wm8753_dai.cpus->of_node) {
cpus              154 sound/soc/tegra/tegra_wm8753.c 	tegra_wm8753_dai.platforms->of_node = tegra_wm8753_dai.cpus->of_node;
cpus              319 sound/soc/tegra/tegra_wm8903.c 	tegra_wm8903_dai.cpus->of_node = of_parse_phandle(np,
cpus              321 sound/soc/tegra/tegra_wm8903.c 	if (!tegra_wm8903_dai.cpus->of_node) {
cpus              328 sound/soc/tegra/tegra_wm8903.c 	tegra_wm8903_dai.platforms->of_node = tegra_wm8903_dai.cpus->of_node;
cpus               99 sound/soc/tegra/tegra_wm9712.c 	tegra_wm9712_dai.cpus->of_node = of_parse_phandle(np,
cpus              101 sound/soc/tegra/tegra_wm9712.c 	if (!tegra_wm9712_dai.cpus->of_node) {
cpus              108 sound/soc/tegra/tegra_wm9712.c 	tegra_wm9712_dai.platforms->of_node = tegra_wm9712_dai.cpus->of_node;
cpus              132 sound/soc/tegra/trimslice.c 	trimslice_tlv320aic23_dai.cpus->of_node = of_parse_phandle(np,
cpus              134 sound/soc/tegra/trimslice.c 	if (!trimslice_tlv320aic23_dai.cpus->of_node) {
cpus              142 sound/soc/tegra/trimslice.c 			trimslice_tlv320aic23_dai.cpus->of_node;
cpus              407 sound/soc/ti/davinci-evm.c 	dai->cpus->of_node = of_parse_phandle(np, "ti,mcasp-controller", 0);
cpus              408 sound/soc/ti/davinci-evm.c 	if (!dai->cpus->of_node)
cpus              411 sound/soc/ti/davinci-evm.c 	dai->platforms->of_node = dai->cpus->of_node;
cpus              256 sound/soc/ti/omap-abe-twl6040.c 	priv->dai_links[0].cpus = link0_cpus;
cpus              258 sound/soc/ti/omap-abe-twl6040.c 	priv->dai_links[0].cpus->of_node = dai_node;
cpus              272 sound/soc/ti/omap-abe-twl6040.c 		priv->dai_links[1].cpus = link1_cpus;
cpus              274 sound/soc/ti/omap-abe-twl6040.c 		priv->dai_links[1].cpus->of_node = dai_node;
cpus              370 sound/soc/ti/omap-hdmi.c 	card->dai_link->cpus		= &compnent[0];
cpus              379 sound/soc/ti/omap-hdmi.c 	card->dai_link->cpus->dai_name = dev_name(ad->dssdev);
cpus              265 sound/soc/ti/omap-twl4030.c 		omap_twl4030_dai_links[0].cpus->dai_name  = NULL;
cpus              266 sound/soc/ti/omap-twl4030.c 		omap_twl4030_dai_links[0].cpus->of_node = dai_node;
cpus              275 sound/soc/ti/omap-twl4030.c 			omap_twl4030_dai_links[1].cpus->dai_name  = NULL;
cpus              276 sound/soc/ti/omap-twl4030.c 			omap_twl4030_dai_links[1].cpus->of_node = dai_node;
cpus              380 sound/soc/ti/rx51.c 		rx51_dai[0].cpus->dai_name = NULL;
cpus              382 sound/soc/ti/rx51.c 		rx51_dai[0].cpus->of_node = dai_node;
cpus               67 sound/soc/ux500/mop500.c 		of_node_put(mop500_dai_links[i].cpus->of_node);
cpus               89 sound/soc/ux500/mop500.c 		mop500_dai_links[i].cpus->of_node = msp_np[i];
cpus               90 sound/soc/ux500/mop500.c 		mop500_dai_links[i].cpus->dai_name = NULL;
cpus              447 tools/bpf/bpftool/common.c 	int cpus = libbpf_num_possible_cpus();
cpus              449 tools/bpf/bpftool/common.c 	if (cpus < 0) {
cpus              450 tools/bpf/bpftool/common.c 		p_err("Can't get # of possible cpus: %s", strerror(-cpus));
cpus              453 tools/bpf/bpftool/common.c 	return cpus;
cpus              198 tools/bpf/bpftool/map_perf_ring.c 	opts.cpus = &ctx.cpu;
cpus             5309 tools/lib/bpf/libbpf.c 	int *cpus;
cpus             5458 tools/lib/bpf/libbpf.c 	p.cpus = opts->cpus;
cpus             5544 tools/lib/bpf/libbpf.c 		cpu = p->cpu_cnt > 0 ? p->cpus[i] : i;
cpus             5988 tools/lib/bpf/libbpf.c 	static int cpus;
cpus             5992 tools/lib/bpf/libbpf.c 	tmp_cpus = READ_ONCE(cpus);
cpus             6007 tools/lib/bpf/libbpf.c 	WRITE_ONCE(cpus, tmp_cpus);
cpus              181 tools/lib/traceevent/event-parse-api.c 		return tep->cpus;
cpus              191 tools/lib/traceevent/event-parse-api.c void tep_set_cpus(struct tep_handle *tep, int cpus)
cpus              194 tools/lib/traceevent/event-parse-api.c 		tep->cpus = cpus;
cpus               33 tools/lib/traceevent/event-parse-local.h 	int cpus;
cpus              557 tools/lib/traceevent/event-parse.h void tep_set_cpus(struct tep_handle *tep, int cpus);
cpus               33 tools/lib/traceevent/plugins/plugin_function.c static int cpus = -1;
cpus               96 tools/lib/traceevent/plugins/plugin_function.c 	if (cpu > cpus) {
cpus              108 tools/lib/traceevent/plugins/plugin_function.c 		for (i = cpus + 1; i <= cpu; i++)
cpus              110 tools/lib/traceevent/plugins/plugin_function.c 		cpus = cpu;
cpus              184 tools/lib/traceevent/plugins/plugin_function.c 	for (i = 0; i <= cpus; i++) {
cpus              194 tools/lib/traceevent/plugins/plugin_function.c 	cpus = -1;
cpus              162 tools/perf/arch/arm/util/cs-etm.c 	struct perf_cpu_map *event_cpus = evsel->evlist->core.cpus;
cpus              260 tools/perf/arch/arm/util/cs-etm.c 	struct perf_cpu_map *cpus = evlist->core.cpus;
cpus              403 tools/perf/arch/arm/util/cs-etm.c 	if (!perf_cpu_map__empty(cpus)) {
cpus              427 tools/perf/arch/arm/util/cs-etm.c 		if (!perf_cpu_map__empty(cpus))
cpus              496 tools/perf/arch/arm/util/cs-etm.c 	struct perf_cpu_map *event_cpus = evlist->core.cpus;
cpus              643 tools/perf/arch/arm/util/cs-etm.c 	struct perf_cpu_map *event_cpus = session->evlist->core.cpus;
cpus               22 tools/perf/arch/arm64/util/header.c 	struct perf_cpu_map *cpus;
cpus               25 tools/perf/arch/arm64/util/header.c 	if (!sysfs || !pmu || !pmu->cpus)
cpus               33 tools/perf/arch/arm64/util/header.c 	cpus = perf_cpu_map__get(pmu->cpus);
cpus               34 tools/perf/arch/arm64/util/header.c 	for (cpu = 0; cpu < perf_cpu_map__nr(cpus); cpu++) {
cpus               36 tools/perf/arch/arm64/util/header.c 				sysfs, cpus->map[cpu]);
cpus               66 tools/perf/arch/arm64/util/header.c 	perf_cpu_map__put(cpus);
cpus               15 tools/perf/arch/nds32/util/header.c 	struct cpu_map *cpus;
cpus               18 tools/perf/arch/nds32/util/header.c 	if (!sysfs || !pmu || !pmu->cpus)
cpus               25 tools/perf/arch/nds32/util/header.c 	cpus = cpu_map__get(pmu->cpus);
cpus               26 tools/perf/arch/nds32/util/header.c 	sprintf(buf, "0x%x", cpus->nr - 1);
cpus               27 tools/perf/arch/nds32/util/header.c 	cpu_map__put(cpus);
cpus               59 tools/perf/arch/x86/tests/perf-time-to-tsc.c 	struct perf_cpu_map *cpus = NULL;
cpus               74 tools/perf/arch/x86/tests/perf-time-to-tsc.c 	cpus = perf_cpu_map__new(NULL);
cpus               75 tools/perf/arch/x86/tests/perf-time-to-tsc.c 	CHECK_NOT_NULL__(cpus);
cpus               80 tools/perf/arch/x86/tests/perf-time-to-tsc.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus              113 tools/perf/arch/x86/util/intel-bts.c 	const struct perf_cpu_map *cpus = evlist->core.cpus;
cpus              140 tools/perf/arch/x86/util/intel-bts.c 	if (opts->full_auxtrace && !perf_cpu_map__empty(cpus)) {
cpus              221 tools/perf/arch/x86/util/intel-bts.c 		if (!perf_cpu_map__empty(cpus))
cpus              371 tools/perf/arch/x86/util/intel-pt.c 	per_cpu_mmaps = !perf_cpu_map__empty(session->evlist->core.cpus);
cpus              583 tools/perf/arch/x86/util/intel-pt.c 	const struct perf_cpu_map *cpus = evlist->core.cpus;
cpus              708 tools/perf/arch/x86/util/intel-pt.c 	if (have_timing_info && !perf_cpu_map__empty(cpus)) {
cpus              766 tools/perf/arch/x86/util/intel-pt.c 		if (!perf_cpu_map__empty(cpus))
cpus              790 tools/perf/arch/x86/util/intel-pt.c 		if (!perf_cpu_map__empty(cpus)) {
cpus              802 tools/perf/arch/x86/util/intel-pt.c 	if (!ptr->have_sched_switch && !perf_cpu_map__empty(cpus) &&
cpus              198 tools/perf/builtin-ftrace.c 	struct perf_cpu_map *cpumap = ftrace->evlist->core.cpus;
cpus              770 tools/perf/builtin-record.c 		if (evsel__open(pos, pos->core.cpus, pos->core.threads) < 0) {
cpus             1316 tools/perf/builtin-record.c 	err = perf_event__synthesize_cpu_map(&rec->tool, rec->evlist->core.cpus,
cpus              172 tools/perf/builtin-sched.c 	struct perf_cpu_map	*cpus;
cpus             1614 tools/perf/builtin-sched.c 		if (sched->map.cpus && !cpu_map__has(sched->map.cpus, cpu))
cpus             1636 tools/perf/builtin-sched.c 	if (sched->map.cpus && !cpu_map__has(sched->map.cpus, this_cpu))
cpus             3198 tools/perf/builtin-sched.c 	sched->map.cpus = map;
cpus             1645 tools/perf/builtin-script.c 	struct perf_cpu_map	*cpus;
cpus             1942 tools/perf/builtin-script.c 				counter->core.cpus->map[cpu],
cpus             3280 tools/perf/builtin-script.c 	if (!script->cpus || !script->threads)
cpus             3286 tools/perf/builtin-script.c 	perf_evlist__set_maps(&evlist->core, script->cpus, script->threads);
cpus             3321 tools/perf/builtin-script.c 	if (script->cpus) {
cpus             3326 tools/perf/builtin-script.c 	script->cpus = cpu_map__new_data(&event->cpu_map.data);
cpus             3327 tools/perf/builtin-script.c 	if (!script->cpus)
cpus              171 tools/perf/builtin-stat.c 	struct perf_cpu_map	*cpus;
cpus              879 tools/perf/builtin-stat.c 		if (cpu_map__build_socket_map(evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus              886 tools/perf/builtin-stat.c 		if (cpu_map__build_die_map(evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus              893 tools/perf/builtin-stat.c 		if (cpu_map__build_core_map(evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus              901 tools/perf/builtin-stat.c 			if (cpu_map__build_core_map(evsel_list->core.cpus,
cpus              921 tools/perf/builtin-stat.c 	nr = perf_cpu_map__max(evsel_list->core.cpus);
cpus             1011 tools/perf/builtin-stat.c static int perf_env__build_socket_map(struct perf_env *env, struct perf_cpu_map *cpus,
cpus             1014 tools/perf/builtin-stat.c 	return cpu_map__build_map(cpus, sockp, perf_env__get_socket, env);
cpus             1017 tools/perf/builtin-stat.c static int perf_env__build_die_map(struct perf_env *env, struct perf_cpu_map *cpus,
cpus             1020 tools/perf/builtin-stat.c 	return cpu_map__build_map(cpus, diep, perf_env__get_die, env);
cpus             1023 tools/perf/builtin-stat.c static int perf_env__build_core_map(struct perf_env *env, struct perf_cpu_map *cpus,
cpus             1026 tools/perf/builtin-stat.c 	return cpu_map__build_map(cpus, corep, perf_env__get_core, env);
cpus             1052 tools/perf/builtin-stat.c 		if (perf_env__build_socket_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus             1059 tools/perf/builtin-stat.c 		if (perf_env__build_die_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus             1066 tools/perf/builtin-stat.c 		if (perf_env__build_core_map(env, evsel_list->core.cpus, &stat_config.aggr_map)) {
cpus             1487 tools/perf/builtin-stat.c 	if (perf_cpu_map__empty(st->cpus)) {
cpus             1506 tools/perf/builtin-stat.c 	if (!st->cpus || !st->threads)
cpus             1512 tools/perf/builtin-stat.c 	perf_evlist__set_maps(&evsel_list->core, st->cpus, st->threads);
cpus             1546 tools/perf/builtin-stat.c 	struct perf_cpu_map *cpus;
cpus             1548 tools/perf/builtin-stat.c 	if (st->cpus) {
cpus             1553 tools/perf/builtin-stat.c 	cpus = cpu_map__new_data(&event->cpu_map.data);
cpus             1554 tools/perf/builtin-stat.c 	if (!cpus)
cpus             1557 tools/perf/builtin-stat.c 	st->cpus = cpus;
cpus             1017 tools/perf/builtin-top.c 		if (evsel__open(counter, top->evlist->core.cpus,
cpus               15 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(int));
cpus               17 tools/perf/lib/cpumap.c 	if (cpus != NULL) {
cpus               18 tools/perf/lib/cpumap.c 		cpus->nr = 1;
cpus               19 tools/perf/lib/cpumap.c 		cpus->map[0] = -1;
cpus               20 tools/perf/lib/cpumap.c 		refcount_set(&cpus->refcnt, 1);
cpus               23 tools/perf/lib/cpumap.c 	return cpus;
cpus               50 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus;
cpus               57 tools/perf/lib/cpumap.c 	cpus = malloc(sizeof(*cpus) + nr_cpus * sizeof(int));
cpus               58 tools/perf/lib/cpumap.c 	if (cpus != NULL) {
cpus               62 tools/perf/lib/cpumap.c 			cpus->map[i] = i;
cpus               64 tools/perf/lib/cpumap.c 		cpus->nr = nr_cpus;
cpus               65 tools/perf/lib/cpumap.c 		refcount_set(&cpus->refcnt, 1);
cpus               68 tools/perf/lib/cpumap.c 	return cpus;
cpus               74 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + payload_size);
cpus               76 tools/perf/lib/cpumap.c 	if (cpus != NULL) {
cpus               77 tools/perf/lib/cpumap.c 		cpus->nr = nr_cpus;
cpus               78 tools/perf/lib/cpumap.c 		memcpy(cpus->map, tmp_cpus, payload_size);
cpus               79 tools/perf/lib/cpumap.c 		refcount_set(&cpus->refcnt, 1);
cpus               82 tools/perf/lib/cpumap.c 	return cpus;
cpus               87 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus = NULL;
cpus              135 tools/perf/lib/cpumap.c 		cpus = cpu_map__trim_new(nr_cpus, tmp_cpus);
cpus              137 tools/perf/lib/cpumap.c 		cpus = cpu_map__default_new();
cpus              140 tools/perf/lib/cpumap.c 	return cpus;
cpus              145 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus = NULL;
cpus              152 tools/perf/lib/cpumap.c 	cpus = perf_cpu_map__read(onlnf);
cpus              154 tools/perf/lib/cpumap.c 	return cpus;
cpus              159 tools/perf/lib/cpumap.c 	struct perf_cpu_map *cpus = NULL;
cpus              223 tools/perf/lib/cpumap.c 		cpus = cpu_map__trim_new(nr_cpus, tmp_cpus);
cpus              225 tools/perf/lib/cpumap.c 		cpus = cpu_map__default_new();
cpus              227 tools/perf/lib/cpumap.c 		cpus = perf_cpu_map__dummy_new();
cpus              231 tools/perf/lib/cpumap.c 	return cpus;
cpus              234 tools/perf/lib/cpumap.c int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx)
cpus              236 tools/perf/lib/cpumap.c 	if (idx < cpus->nr)
cpus              237 tools/perf/lib/cpumap.c 		return cpus->map[idx];
cpus              242 tools/perf/lib/cpumap.c int perf_cpu_map__nr(const struct perf_cpu_map *cpus)
cpus              244 tools/perf/lib/cpumap.c 	return cpus ? cpus->nr : 1;
cpus              252 tools/perf/lib/cpumap.c int perf_cpu_map__idx(struct perf_cpu_map *cpus, int cpu)
cpus              256 tools/perf/lib/cpumap.c 	for (i = 0; i < cpus->nr; ++i) {
cpus              257 tools/perf/lib/cpumap.c 		if (cpus->map[i] == cpu)
cpus               40 tools/perf/lib/evlist.c 		perf_cpu_map__put(evsel->cpus);
cpus               41 tools/perf/lib/evlist.c 		evsel->cpus = perf_cpu_map__get(evlist->cpus);
cpus               42 tools/perf/lib/evlist.c 	} else if (evsel->cpus != evsel->own_cpus) {
cpus               43 tools/perf/lib/evlist.c 		perf_cpu_map__put(evsel->cpus);
cpus               44 tools/perf/lib/evlist.c 		evsel->cpus = perf_cpu_map__get(evsel->own_cpus);
cpus              110 tools/perf/lib/evlist.c 			   struct perf_cpu_map *cpus,
cpus              120 tools/perf/lib/evlist.c 	if (cpus != evlist->cpus) {
cpus              121 tools/perf/lib/evlist.c 		perf_cpu_map__put(evlist->cpus);
cpus              122 tools/perf/lib/evlist.c 		evlist->cpus = perf_cpu_map__get(cpus);
cpus              139 tools/perf/lib/evlist.c 		err = perf_evsel__open(evsel, evsel->cpus, evsel->threads);
cpus              248 tools/perf/lib/evlist.c 	int nr_cpus = perf_cpu_map__nr(evlist->cpus);
cpus               66 tools/perf/lib/evsel.c int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
cpus               71 tools/perf/lib/evsel.c 	if (cpus == NULL) {
cpus               80 tools/perf/lib/evsel.c 		cpus = empty_cpu_map;
cpus               96 tools/perf/lib/evsel.c 	    perf_evsel__alloc_fd(evsel, cpus->nr, threads->nr) < 0)
cpus               99 tools/perf/lib/evsel.c 	for (cpu = 0; cpu < cpus->nr; cpu++) {
cpus              105 tools/perf/lib/evsel.c 						 cpus->map[cpu], -1, 0);
cpus              221 tools/perf/lib/evsel.c 	return evsel->cpus;
cpus               17 tools/perf/lib/include/internal/cpumap.h int perf_cpu_map__idx(struct perf_cpu_map *cpus, int cpu);
cpus               19 tools/perf/lib/include/internal/evlist.h 	struct perf_cpu_map	*cpus;
cpus               40 tools/perf/lib/include/internal/evsel.h 	struct perf_cpu_map	*cpus;
cpus               16 tools/perf/lib/include/perf/cpumap.h LIBPERF_API int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx);
cpus               17 tools/perf/lib/include/perf/cpumap.h LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus);
cpus               21 tools/perf/lib/include/perf/cpumap.h #define perf_cpu_map__for_each_cpu(cpu, idx, cpus)		\
cpus               22 tools/perf/lib/include/perf/cpumap.h 	for ((idx) = 0, (cpu) = perf_cpu_map__cpu(cpus, idx);	\
cpus               23 tools/perf/lib/include/perf/cpumap.h 	     (idx) < perf_cpu_map__nr(cpus);			\
cpus               24 tools/perf/lib/include/perf/cpumap.h 	     (idx)++, (cpu) = perf_cpu_map__cpu(cpus, idx))
cpus              159 tools/perf/lib/include/perf/event.h 	struct perf_record_cpu_map_data	 cpus;
cpus               32 tools/perf/lib/include/perf/evlist.h 				       struct perf_cpu_map *cpus,
cpus               28 tools/perf/lib/include/perf/evsel.h LIBPERF_API int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
cpus               15 tools/perf/lib/tests/test-cpumap.c 	struct perf_cpu_map *cpus;
cpus               21 tools/perf/lib/tests/test-cpumap.c 	cpus = perf_cpu_map__dummy_new();
cpus               22 tools/perf/lib/tests/test-cpumap.c 	if (!cpus)
cpus               25 tools/perf/lib/tests/test-cpumap.c 	perf_cpu_map__get(cpus);
cpus               26 tools/perf/lib/tests/test-cpumap.c 	perf_cpu_map__put(cpus);
cpus               27 tools/perf/lib/tests/test-cpumap.c 	perf_cpu_map__put(cpus);
cpus               19 tools/perf/lib/tests/test-evlist.c 	struct perf_cpu_map *cpus;
cpus               32 tools/perf/lib/tests/test-evlist.c 	cpus = perf_cpu_map__new(NULL);
cpus               33 tools/perf/lib/tests/test-evlist.c 	__T("failed to create cpus", cpus);
cpus               48 tools/perf/lib/tests/test-evlist.c 	perf_evlist__set_maps(evlist, cpus, NULL);
cpus               54 tools/perf/lib/tests/test-evlist.c 		cpus = perf_evsel__cpus(evsel);
cpus               56 tools/perf/lib/tests/test-evlist.c 		perf_cpu_map__for_each_cpu(cpu, tmp, cpus) {
cpus               67 tools/perf/lib/tests/test-evlist.c 	perf_cpu_map__put(cpus);
cpus               18 tools/perf/lib/tests/test-evsel.c 	struct perf_cpu_map *cpus;
cpus               26 tools/perf/lib/tests/test-evsel.c 	cpus = perf_cpu_map__new(NULL);
cpus               27 tools/perf/lib/tests/test-evsel.c 	__T("failed to create cpus", cpus);
cpus               32 tools/perf/lib/tests/test-evsel.c 	err = perf_evsel__open(evsel, cpus, NULL);
cpus               35 tools/perf/lib/tests/test-evsel.c 	perf_cpu_map__for_each_cpu(cpu, tmp, cpus) {
cpus               45 tools/perf/lib/tests/test-evsel.c 	perf_cpu_map__put(cpus);
cpus              568 tools/perf/tests/code-reading.c 	struct perf_cpu_map *cpus = NULL;
cpus              628 tools/perf/tests/code-reading.c 	cpus = perf_cpu_map__new(NULL);
cpus              629 tools/perf/tests/code-reading.c 	if (!cpus) {
cpus              643 tools/perf/tests/code-reading.c 		perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus              670 tools/perf/tests/code-reading.c 				perf_cpu_map__get(cpus);
cpus              720 tools/perf/tests/code-reading.c 		perf_cpu_map__put(cpus);
cpus               54 tools/perf/tests/cpumap.c 	struct cpu_map_entries *cpus;
cpus               62 tools/perf/tests/cpumap.c 	cpus = (struct cpu_map_entries *)data->data;
cpus               64 tools/perf/tests/cpumap.c 	TEST_ASSERT_VAL("wrong nr",   cpus->nr == 2);
cpus               65 tools/perf/tests/cpumap.c 	TEST_ASSERT_VAL("wrong cpu",  cpus->cpu[0] == 1);
cpus               66 tools/perf/tests/cpumap.c 	TEST_ASSERT_VAL("wrong cpu",  cpus->cpu[1] == 256);
cpus               80 tools/perf/tests/cpumap.c 	struct perf_cpu_map *cpus;
cpus               83 tools/perf/tests/cpumap.c 	cpus = perf_cpu_map__new("0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19");
cpus               86 tools/perf/tests/cpumap.c 		!perf_event__synthesize_cpu_map(NULL, cpus, process_event_mask, NULL));
cpus               88 tools/perf/tests/cpumap.c 	perf_cpu_map__put(cpus);
cpus               91 tools/perf/tests/cpumap.c 	cpus = perf_cpu_map__new("1,256");
cpus               94 tools/perf/tests/cpumap.c 		!perf_event__synthesize_cpu_map(NULL, cpus, process_event_cpus, NULL));
cpus               96 tools/perf/tests/cpumap.c 	perf_cpu_map__put(cpus);
cpus              115 tools/perf/tests/event-times.c 	struct perf_cpu_map *cpus;
cpus              120 tools/perf/tests/event-times.c 	cpus = perf_cpu_map__new("0");
cpus              121 tools/perf/tests/event-times.c 	if (cpus == NULL) {
cpus              128 tools/perf/tests/event-times.c 	err = perf_evsel__open_per_cpu(evsel, cpus);
cpus              137 tools/perf/tests/event-times.c 	perf_cpu_map__put(cpus);
cpus              144 tools/perf/tests/event-times.c 	struct perf_cpu_map *cpus;
cpus              149 tools/perf/tests/event-times.c 	cpus = perf_cpu_map__new("0");
cpus              150 tools/perf/tests/event-times.c 	if (cpus == NULL) {
cpus              155 tools/perf/tests/event-times.c 	err = perf_evsel__open_per_cpu(evsel, cpus);
cpus              159 tools/perf/tests/event-times.c 	perf_cpu_map__put(cpus);
cpus               74 tools/perf/tests/event_update.c 	map = cpu_map__new_data(&ev_data->cpus);
cpus               74 tools/perf/tests/keep-tracking.c 	struct perf_cpu_map *cpus = NULL;
cpus               83 tools/perf/tests/keep-tracking.c 	cpus = perf_cpu_map__new(NULL);
cpus               84 tools/perf/tests/keep-tracking.c 	CHECK_NOT_NULL__(cpus);
cpus               89 tools/perf/tests/keep-tracking.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus              157 tools/perf/tests/keep-tracking.c 		perf_cpu_map__put(cpus);
cpus               36 tools/perf/tests/mmap-basic.c 	struct perf_cpu_map *cpus;
cpus               54 tools/perf/tests/mmap-basic.c 	cpus = perf_cpu_map__new(NULL);
cpus               55 tools/perf/tests/mmap-basic.c 	if (cpus == NULL) {
cpus               61 tools/perf/tests/mmap-basic.c 	CPU_SET(cpus->map[0], &cpu_set);
cpus               65 tools/perf/tests/mmap-basic.c 			 cpus->map[0], str_error_r(errno, sbuf, sizeof(sbuf)));
cpus               75 tools/perf/tests/mmap-basic.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus               92 tools/perf/tests/mmap-basic.c 		if (evsel__open(evsels[i], cpus, threads) < 0) {
cpus              160 tools/perf/tests/mmap-basic.c 	cpus	= NULL;
cpus              163 tools/perf/tests/mmap-basic.c 	perf_cpu_map__put(cpus);
cpus               26 tools/perf/tests/openat-syscall-all-cpus.c 	struct perf_cpu_map *cpus;
cpus               39 tools/perf/tests/openat-syscall-all-cpus.c 	cpus = perf_cpu_map__new(NULL);
cpus               40 tools/perf/tests/openat-syscall-all-cpus.c 	if (cpus == NULL) {
cpus               54 tools/perf/tests/openat-syscall-all-cpus.c 	if (evsel__open(evsel, cpus, threads) < 0) {
cpus               61 tools/perf/tests/openat-syscall-all-cpus.c 	for (cpu = 0; cpu < cpus->nr; ++cpu) {
cpus               69 tools/perf/tests/openat-syscall-all-cpus.c 		if (cpus->map[cpu] >= CPU_SETSIZE) {
cpus               70 tools/perf/tests/openat-syscall-all-cpus.c 			pr_debug("Ignoring CPU %d\n", cpus->map[cpu]);
cpus               74 tools/perf/tests/openat-syscall-all-cpus.c 		CPU_SET(cpus->map[cpu], &cpu_set);
cpus               77 tools/perf/tests/openat-syscall-all-cpus.c 				 cpus->map[cpu],
cpus               85 tools/perf/tests/openat-syscall-all-cpus.c 		CPU_CLR(cpus->map[cpu], &cpu_set);
cpus               93 tools/perf/tests/openat-syscall-all-cpus.c 	if (perf_evsel__alloc_counts(evsel, cpus->nr, 1) < 0) {
cpus               94 tools/perf/tests/openat-syscall-all-cpus.c 		pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr);
cpus              100 tools/perf/tests/openat-syscall-all-cpus.c 	for (cpu = 0; cpu < cpus->nr; ++cpu) {
cpus              103 tools/perf/tests/openat-syscall-all-cpus.c 		if (cpus->map[cpu] >= CPU_SETSIZE)
cpus              115 tools/perf/tests/openat-syscall-all-cpus.c 				 expected, cpus->map[cpu], perf_counts(evsel->counts, cpu, 0)->val);
cpus              126 tools/perf/tests/openat-syscall-all-cpus.c 	perf_cpu_map__put(cpus);
cpus               44 tools/perf/tests/sw-clock.c 	struct perf_cpu_map *cpus;
cpus               63 tools/perf/tests/sw-clock.c 	cpus = perf_cpu_map__dummy_new();
cpus               65 tools/perf/tests/sw-clock.c 	if (!cpus || !threads) {
cpus               71 tools/perf/tests/sw-clock.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus               73 tools/perf/tests/sw-clock.c 	cpus	= NULL;
cpus              132 tools/perf/tests/sw-clock.c 	perf_cpu_map__put(cpus);
cpus              336 tools/perf/tests/switch-tracking.c 	struct perf_cpu_map *cpus = NULL;
cpus              349 tools/perf/tests/switch-tracking.c 	cpus = perf_cpu_map__new(NULL);
cpus              350 tools/perf/tests/switch-tracking.c 	if (!cpus) {
cpus              361 tools/perf/tests/switch-tracking.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus              577 tools/perf/tests/switch-tracking.c 		perf_cpu_map__put(cpus);
cpus               53 tools/perf/tests/task-exit.c 	struct perf_cpu_map *cpus;
cpus               72 tools/perf/tests/task-exit.c 	cpus = perf_cpu_map__dummy_new();
cpus               74 tools/perf/tests/task-exit.c 	if (!cpus || !threads) {
cpus               80 tools/perf/tests/task-exit.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus               82 tools/perf/tests/task-exit.c 	cpus	= NULL;
cpus              152 tools/perf/tests/task-exit.c 	perf_cpu_map__put(cpus);
cpus              136 tools/perf/util/auxtrace.c 		mp->cpu = evlist->core.cpus->map[idx];
cpus               21 tools/perf/util/cpumap.c static struct perf_cpu_map *cpu_map__from_entries(struct cpu_map_entries *cpus)
cpus               25 tools/perf/util/cpumap.c 	map = perf_cpu_map__empty_new(cpus->nr);
cpus               29 tools/perf/util/cpumap.c 		for (i = 0; i < cpus->nr; i++) {
cpus               35 tools/perf/util/cpumap.c 			if (cpus->cpu[i] == (u16) -1)
cpus               38 tools/perf/util/cpumap.c 				map->map[i] = (int) cpus->cpu[i];
cpus               83 tools/perf/util/cpumap.c 	struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(int) * nr);
cpus               85 tools/perf/util/cpumap.c 	if (cpus != NULL) {
cpus               88 tools/perf/util/cpumap.c 		cpus->nr = nr;
cpus               90 tools/perf/util/cpumap.c 			cpus->map[i] = -1;
cpus               92 tools/perf/util/cpumap.c 		refcount_set(&cpus->refcnt, 1);
cpus               95 tools/perf/util/cpumap.c 	return cpus;
cpus              131 tools/perf/util/cpumap.c int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res,
cpus              136 tools/perf/util/cpumap.c 	int nr = cpus->nr;
cpus              145 tools/perf/util/cpumap.c 		s1 = f(cpus, cpu, data);
cpus              238 tools/perf/util/cpumap.c int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp)
cpus              240 tools/perf/util/cpumap.c 	return cpu_map__build_map(cpus, sockp, cpu_map__get_socket, NULL);
cpus              243 tools/perf/util/cpumap.c int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep)
cpus              245 tools/perf/util/cpumap.c 	return cpu_map__build_map(cpus, diep, cpu_map__get_die, NULL);
cpus              248 tools/perf/util/cpumap.c int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep)
cpus              250 tools/perf/util/cpumap.c 	return cpu_map__build_map(cpus, corep, cpu_map__get_core, NULL);
cpus              460 tools/perf/util/cpumap.c bool cpu_map__has(struct perf_cpu_map *cpus, int cpu)
cpus              462 tools/perf/util/cpumap.c 	return perf_cpu_map__idx(cpus, cpu) != -1;
cpus              465 tools/perf/util/cpumap.c int cpu_map__cpu(struct perf_cpu_map *cpus, int idx)
cpus              467 tools/perf/util/cpumap.c 	return cpus->map[idx];
cpus               23 tools/perf/util/cpumap.h int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp);
cpus               24 tools/perf/util/cpumap.h int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep);
cpus               25 tools/perf/util/cpumap.h int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep);
cpus               57 tools/perf/util/cpumap.h int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res,
cpus               61 tools/perf/util/cpumap.h int cpu_map__cpu(struct perf_cpu_map *cpus, int idx);
cpus               62 tools/perf/util/cpumap.h bool cpu_map__has(struct perf_cpu_map *cpus, int cpu);
cpus              282 tools/perf/util/cputopo.c 	node->cpus = buf;
cpus              350 tools/perf/util/cputopo.c 		zfree(&tp->nodes[i].cpus);
cpus               17 tools/perf/util/cputopo.h 	char		*cpus;
cpus              297 tools/perf/util/event.c 	struct perf_cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data);
cpus              302 tools/perf/util/event.c 	if (cpus)
cpus              303 tools/perf/util/event.c 		ret += cpu_map__fprintf(cpus, fp);
cpus              307 tools/perf/util/event.c 	perf_cpu_map__put(cpus);
cpus               55 tools/perf/util/evlist.c void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus,
cpus               59 tools/perf/util/evlist.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus              151 tools/perf/util/evlist.c 	perf_cpu_map__put(evlist->core.cpus);
cpus              153 tools/perf/util/evlist.c 	evlist->core.cpus = NULL;
cpus              377 tools/perf/util/evlist.c 	int nr_cpus = perf_cpu_map__nr(evlist->core.cpus);
cpus              393 tools/perf/util/evlist.c 	bool per_cpu_mmaps = !perf_cpu_map__empty(evlist->core.cpus);
cpus              432 tools/perf/util/evlist.c 	if (evlist->core.cpus && cpu >= 0)
cpus              433 tools/perf/util/evlist.c 		sid->cpu = evlist->core.cpus->map[cpu];
cpus              600 tools/perf/util/evlist.c 	evlist->core.nr_mmaps = perf_cpu_map__nr(evlist->core.cpus);
cpus              601 tools/perf/util/evlist.c 	if (perf_cpu_map__empty(evlist->core.cpus))
cpus              639 tools/perf/util/evlist.c 	int evlist_cpu = cpu_map__cpu(evlist->core.cpus, cpu_idx);
cpus              666 tools/perf/util/evlist.c 		cpu = perf_cpu_map__idx(evsel->core.cpus, evlist_cpu);
cpus              715 tools/perf/util/evlist.c 	int nr_cpus = perf_cpu_map__nr(evlist->core.cpus);
cpus              894 tools/perf/util/evlist.c 	const struct perf_cpu_map *cpus = evlist->core.cpus;
cpus              922 tools/perf/util/evlist.c 		    perf_evsel__alloc_id(&evsel->core, perf_cpu_map__nr(cpus), threads->nr) < 0)
cpus              926 tools/perf/util/evlist.c 	if (perf_cpu_map__empty(cpus))
cpus              940 tools/perf/util/evlist.c 	struct perf_cpu_map *cpus;
cpus              968 tools/perf/util/evlist.c 		cpus = perf_cpu_map__dummy_new();
cpus              970 tools/perf/util/evlist.c 		cpus = perf_cpu_map__new(target->cpu_list);
cpus              972 tools/perf/util/evlist.c 	if (!cpus)
cpus              977 tools/perf/util/evlist.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus             1211 tools/perf/util/evlist.c 	struct perf_cpu_map *cpus;
cpus             1224 tools/perf/util/evlist.c 	cpus = perf_cpu_map__new(NULL);
cpus             1225 tools/perf/util/evlist.c 	if (!cpus)
cpus             1232 tools/perf/util/evlist.c 	perf_evlist__set_maps(&evlist->core, cpus, threads);
cpus             1236 tools/perf/util/evlist.c 	perf_cpu_map__put(cpus);
cpus             1249 tools/perf/util/evlist.c 	if (evlist->core.threads == NULL && evlist->core.cpus == NULL) {
cpus             1258 tools/perf/util/evlist.c 		err = evsel__open(evsel, evsel->core.cpus, evsel->core.threads);
cpus             1766 tools/perf/util/evlist.c 		if (evsel__open(counter, evlist->core.cpus,
cpus               87 tools/perf/util/evlist.h void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus,
cpus             1252 tools/perf/util/evsel.c 	perf_cpu_map__put(evsel->core.cpus);
cpus             1572 tools/perf/util/evsel.c int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus,
cpus             1584 tools/perf/util/evsel.c 	if (cpus == NULL) {
cpus             1593 tools/perf/util/evsel.c 		cpus = empty_cpu_map;
cpus             1614 tools/perf/util/evsel.c 	    perf_evsel__alloc_fd(&evsel->core, cpus->nr, nthreads) < 0)
cpus             1650 tools/perf/util/evsel.c 	for (cpu = 0; cpu < cpus->nr; cpu++) {
cpus             1662 tools/perf/util/evsel.c 			fd = perf_event_open(evsel, pid, cpus->map[cpu],
cpus             1670 tools/perf/util/evsel.c 				if (ignore_missing_thread(evsel, cpus->nr, cpu, threads, thread, err)) {
cpus             1832 tools/perf/util/evsel.c 			     struct perf_cpu_map *cpus)
cpus             1834 tools/perf/util/evsel.c 	return evsel__open(evsel, cpus, NULL);
cpus             2539 tools/perf/util/evsel.c 	struct perf_cpu_map *cpus = evsel->core.cpus;
cpus             2542 tools/perf/util/evsel.c 	if (perf_evsel__alloc_id(&evsel->core, cpus->nr, threads->nr))
cpus              225 tools/perf/util/evsel.h 			     struct perf_cpu_map *cpus);
cpus              228 tools/perf/util/evsel.h int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus,
cpus              701 tools/perf/util/header.c 		ret = do_write_string(ff, n->cpus);
cpus             3705 tools/perf/util/header.c 		map = cpu_map__new_data(&ev_cpus->cpus);
cpus             3791 tools/perf/util/header.c 		map = cpu_map__new_data(&ev_cpus->cpus);
cpus              327 tools/perf/util/parse-events.c 	struct perf_cpu_map *cpus = pmu ? pmu->cpus :
cpus              337 tools/perf/util/parse-events.c 	evsel->core.cpus   = perf_cpu_map__get(cpus);
cpus              338 tools/perf/util/parse-events.c 	evsel->core.own_cpus = perf_cpu_map__get(cpus);
cpus              580 tools/perf/util/pmu.c 	struct perf_cpu_map *cpus;
cpus              586 tools/perf/util/pmu.c 	cpus = perf_cpu_map__read(file);
cpus              588 tools/perf/util/pmu.c 	return cpus;
cpus              601 tools/perf/util/pmu.c 	struct perf_cpu_map *cpus;
cpus              615 tools/perf/util/pmu.c 		cpus = __pmu_cpumask(path);
cpus              616 tools/perf/util/pmu.c 		if (cpus)
cpus              617 tools/perf/util/pmu.c 			return cpus;
cpus              626 tools/perf/util/pmu.c 	struct perf_cpu_map *cpus;
cpus              630 tools/perf/util/pmu.c 	cpus = __pmu_cpumask(path);
cpus              631 tools/perf/util/pmu.c 	perf_cpu_map__put(cpus);
cpus              633 tools/perf/util/pmu.c 	return !!cpus;
cpus              841 tools/perf/util/pmu.c 	pmu->cpus = pmu_cpumask(name);
cpus               31 tools/perf/util/pmu.h 	struct perf_cpu_map *cpus;
cpus              550 tools/perf/util/python.c 	struct perf_cpu_map *cpus;
cpus              563 tools/perf/util/python.c 	pcpus->cpus = perf_cpu_map__new(cpustr);
cpus              564 tools/perf/util/python.c 	if (pcpus->cpus == NULL)
cpus              571 tools/perf/util/python.c 	perf_cpu_map__put(pcpus->cpus);
cpus              579 tools/perf/util/python.c 	return pcpus->cpus->nr;
cpus              586 tools/perf/util/python.c 	if (i >= pcpus->cpus->nr)
cpus              589 tools/perf/util/python.c 	return Py_BuildValue("i", pcpus->cpus->map[i]);
cpus              810 tools/perf/util/python.c 	struct perf_cpu_map *cpus = NULL;
cpus              824 tools/perf/util/python.c 		cpus = ((struct pyrf_cpu_map *)pcpus)->cpus;
cpus              831 tools/perf/util/python.c 	if (evsel__open(evsel, cpus, threads) < 0) {
cpus              879 tools/perf/util/python.c 	struct perf_cpu_map *cpus;
cpus              886 tools/perf/util/python.c 	cpus = ((struct pyrf_cpu_map *)pcpus)->cpus;
cpus              887 tools/perf/util/python.c 	evlist__init(&pevlist->evlist, cpus, threads);
cpus               67 tools/perf/util/record.c 	struct perf_cpu_map *cpus;
cpus               70 tools/perf/util/record.c 	cpus = perf_cpu_map__new(NULL);
cpus               71 tools/perf/util/record.c 	if (!cpus)
cpus               73 tools/perf/util/record.c 	cpu = cpus->map[0];
cpus               74 tools/perf/util/record.c 	perf_cpu_map__put(cpus);
cpus              122 tools/perf/util/record.c 	struct perf_cpu_map *cpus;
cpus              125 tools/perf/util/record.c 	cpus = perf_cpu_map__new(NULL);
cpus              126 tools/perf/util/record.c 	if (!cpus)
cpus              128 tools/perf/util/record.c 	cpu = cpus->map[0];
cpus              129 tools/perf/util/record.c 	perf_cpu_map__put(cpus);
cpus              154 tools/perf/util/record.c 	if (evlist->core.cpus->map[0] < 0)
cpus              281 tools/perf/util/record.c 	if (!evlist || perf_cpu_map__empty(evlist->core.cpus)) {
cpus              282 tools/perf/util/record.c 		struct perf_cpu_map *cpus = perf_cpu_map__new(NULL);
cpus              284 tools/perf/util/record.c 		cpu =  cpus ? cpus->map[0] : 0;
cpus              285 tools/perf/util/record.c 		perf_cpu_map__put(cpus);
cpus              287 tools/perf/util/record.c 		cpu = evlist->core.cpus->map[0];
cpus             1396 tools/perf/util/scripting-engines/trace-event-python.c 	struct perf_cpu_map *cpus = counter->core.cpus;
cpus             1406 tools/perf/util/scripting-engines/trace-event-python.c 		for (cpu = 0; cpu < cpus->nr; cpu++) {
cpus             1407 tools/perf/util/scripting-engines/trace-event-python.c 			process_stat(counter, cpus->map[cpu],
cpus              849 tools/perf/util/session.c 	struct cpu_map_entries *cpus;
cpus              857 tools/perf/util/session.c 		cpus = (struct cpu_map_entries *)data->data;
cpus              859 tools/perf/util/session.c 		cpus->nr = bswap_16(cpus->nr);
cpus              861 tools/perf/util/session.c 		for (i = 0; i < cpus->nr; i++)
cpus              862 tools/perf/util/session.c 			cpus->cpu[i] = bswap_16(cpus->cpu[i]);
cpus              331 tools/perf/util/stat-display.c 		if (config->aggr_get_id(config, evlist->core.cpus, cpu2) == id)
cpus              504 tools/perf/util/stat-display.c 				s2 = config->aggr_get_id(config, evlist->core.cpus, cpu);
cpus              749 tools/perf/util/stat-display.c 	int ncpus = perf_cpu_map__nr(counter->core.cpus);
cpus              872 tools/perf/util/stat-display.c 	nrcpus = evlist->core.cpus->nr;
cpus              242 tools/perf/util/stat.c 	struct perf_cpu_map *cpus = evsel__cpus(counter);
cpus              250 tools/perf/util/stat.c 	if (perf_cpu_map__empty(cpus))
cpus              272 tools/perf/util/stat.c 	s = cpu_map__get_socket(cpus, cpu, NULL);
cpus               85 tools/perf/util/svghelper.c void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end)
cpus              107 tools/perf/util/svghelper.c 	total_height = (1 + rows + cpu2slot(cpus)) * SLOT_MULT;
cpus                9 tools/perf/util/svghelper.h void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
cpus              918 tools/perf/util/synthetic-events.c static void synthesize_cpus(struct cpu_map_entries *cpus,
cpus              923 tools/perf/util/synthetic-events.c 	cpus->nr = map->nr;
cpus              926 tools/perf/util/synthetic-events.c 		cpus->cpu[i] = map->map[i];
cpus             1798 tools/perf/util/synthetic-events.c 	err = perf_event__synthesize_cpu_map(tool, evlist->core.cpus, process, NULL);
cpus               35 tools/perf/util/synthetic-events.h int perf_event__synthesize_cpu_map(struct perf_tool *tool, struct perf_cpu_map *cpus, perf_event__handler_t process, struct machine *machine);
cpus               98 tools/perf/util/top.c 				top->evlist->core.cpus->nr > 1 ? "s" : "",
cpus              105 tools/perf/util/top.c 					top->evlist->core.cpus->nr,
cpus              106 tools/perf/util/top.c 					top->evlist->core.cpus->nr > 1 ? "s" : "");
cpus              133 tools/power/cpupower/lib/cpupower.c 	int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF);
cpus              135 tools/power/cpupower/lib/cpupower.c 	cpu_top->core_info = malloc(sizeof(struct cpuid_core_info) * cpus);
cpus              139 tools/power/cpupower/lib/cpupower.c 	for (cpu = 0; cpu < cpus; cpu++) {
cpus              160 tools/power/cpupower/lib/cpupower.c 	qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info),
cpus              167 tools/power/cpupower/lib/cpupower.c 	for(cpu = 1; cpu < cpus; cpu++) {
cpus              185 tools/power/cpupower/lib/cpupower.c 	return cpus;
cpus              393 tools/power/cpupower/utils/cpufreq-info.c 	struct cpufreq_affected_cpus *cpus = cpufreq_get_affected_cpus(cpu);
cpus              396 tools/power/cpupower/utils/cpufreq-info.c 	if (!cpus) {
cpus              401 tools/power/cpupower/utils/cpufreq-info.c 	while (cpus->next) {
cpus              402 tools/power/cpupower/utils/cpufreq-info.c 		printf("%d ", cpus->cpu);
cpus              403 tools/power/cpupower/utils/cpufreq-info.c 		cpus = cpus->next;
cpus              405 tools/power/cpupower/utils/cpufreq-info.c 	printf("%d\n", cpus->cpu);
cpus              406 tools/power/cpupower/utils/cpufreq-info.c 	cpufreq_put_affected_cpus(cpus);
cpus              414 tools/power/cpupower/utils/cpufreq-info.c 	struct cpufreq_affected_cpus *cpus = cpufreq_get_related_cpus(cpu);
cpus              417 tools/power/cpupower/utils/cpufreq-info.c 	if (!cpus) {
cpus              422 tools/power/cpupower/utils/cpufreq-info.c 	while (cpus->next) {
cpus              423 tools/power/cpupower/utils/cpufreq-info.c 		printf("%d ", cpus->cpu);
cpus              424 tools/power/cpupower/utils/cpufreq-info.c 		cpus = cpus->next;
cpus              426 tools/power/cpupower/utils/cpufreq-info.c 	printf("%d\n", cpus->cpu);
cpus              427 tools/power/cpupower/utils/cpufreq-info.c 	cpufreq_put_related_cpus(cpus);
cpus              295 tools/power/cpupower/utils/cpufreq-set.c 			struct cpufreq_affected_cpus *cpus;
cpus              301 tools/power/cpupower/utils/cpufreq-set.c 			cpus = cpufreq_get_related_cpus(cpu);
cpus              302 tools/power/cpupower/utils/cpufreq-set.c 			if (!cpus)
cpus              304 tools/power/cpupower/utils/cpufreq-set.c 			while (cpus->next) {
cpus              305 tools/power/cpupower/utils/cpufreq-set.c 				bitmask_setbit(cpus_chosen, cpus->cpu);
cpus              306 tools/power/cpupower/utils/cpufreq-set.c 				cpus = cpus->next;
cpus              309 tools/power/cpupower/utils/cpufreq-set.c 			bitmask_setbit(cpus_chosen, cpus->cpu);
cpus              310 tools/power/cpupower/utils/cpufreq-set.c 			cpufreq_put_related_cpus(cpus);
cpus              286 tools/power/x86/turbostat/turbostat.c } *cpus;
cpus              949 tools/power/x86/turbostat/turbostat.c 				outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].die_id);
cpus              957 tools/power/x86/turbostat/turbostat.c 					      cpus[t->cpu_id].physical_node_id);
cpus             2516 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].put_ids)
cpus             2517 tools/power/x86/turbostat/turbostat.c 			CPU_FREE(cpus[i].put_ids);
cpus             2519 tools/power/x86/turbostat/turbostat.c 	free(cpus);
cpus             2578 tools/power/x86/turbostat/turbostat.c 		cpus[cpu].logical_node_id = -1;
cpus             2584 tools/power/x86/turbostat/turbostat.c 			if (cpus[cpu].physical_package_id != pkg)
cpus             2587 tools/power/x86/turbostat/turbostat.c 			if (cpus[cpu].logical_node_id != -1)
cpus             2589 tools/power/x86/turbostat/turbostat.c 			cpus[cpu].logical_node_id = lnode;
cpus             2590 tools/power/x86/turbostat/turbostat.c 			node = cpus[cpu].physical_node_id;
cpus             2597 tools/power/x86/turbostat/turbostat.c 				if ((cpus[cpux].physical_package_id == pkg) &&
cpus             2598 tools/power/x86/turbostat/turbostat.c 				   (cpus[cpux].physical_node_id == node)) {
cpus             2599 tools/power/x86/turbostat/turbostat.c 					cpus[cpux].logical_node_id = lnode;
cpus             2665 tools/power/x86/turbostat/turbostat.c 					    (cpus[so].thread_id < 0))
cpus             2666 tools/power/x86/turbostat/turbostat.c 						cpus[so].thread_id =
cpus             2801 tools/power/x86/turbostat/turbostat.c 	cpus[cpu].thread_id = -1;
cpus             4975 tools/power/x86/turbostat/turbostat.c 	cpus = calloc(1, (topo.max_cpu_num  + 1) * sizeof(struct cpu_topology));
cpus             4976 tools/power/x86/turbostat/turbostat.c 	if (cpus == NULL)
cpus             5022 tools/power/x86/turbostat/turbostat.c 		cpus[i].logical_cpu_id = i;
cpus             5025 tools/power/x86/turbostat/turbostat.c 		cpus[i].physical_package_id = get_physical_package_id(i);
cpus             5026 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].physical_package_id > max_package_id)
cpus             5027 tools/power/x86/turbostat/turbostat.c 			max_package_id = cpus[i].physical_package_id;
cpus             5030 tools/power/x86/turbostat/turbostat.c 		cpus[i].die_id = get_die_id(i);
cpus             5031 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].die_id > max_die_id)
cpus             5032 tools/power/x86/turbostat/turbostat.c 			max_die_id = cpus[i].die_id;
cpus             5035 tools/power/x86/turbostat/turbostat.c 		cpus[i].physical_node_id = get_physical_node_id(&cpus[i]);
cpus             5036 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].physical_node_id > topo.max_node_num)
cpus             5037 tools/power/x86/turbostat/turbostat.c 			topo.max_node_num = cpus[i].physical_node_id;
cpus             5040 tools/power/x86/turbostat/turbostat.c 		cpus[i].physical_core_id = get_core_id(i);
cpus             5041 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].physical_core_id > max_core_id)
cpus             5042 tools/power/x86/turbostat/turbostat.c 			max_core_id = cpus[i].physical_core_id;
cpus             5045 tools/power/x86/turbostat/turbostat.c 		siblings = get_thread_siblings(&cpus[i]);
cpus             5048 tools/power/x86/turbostat/turbostat.c 		if (cpus[i].thread_id == 0)
cpus             5091 tools/power/x86/turbostat/turbostat.c 			i, cpus[i].physical_package_id, cpus[i].die_id,
cpus             5092 tools/power/x86/turbostat/turbostat.c 			cpus[i].physical_node_id,
cpus             5093 tools/power/x86/turbostat/turbostat.c 			cpus[i].logical_node_id,
cpus             5094 tools/power/x86/turbostat/turbostat.c 			cpus[i].physical_core_id,
cpus             5095 tools/power/x86/turbostat/turbostat.c 			cpus[i].thread_id);
cpus             5142 tools/power/x86/turbostat/turbostat.c 	int pkg_id = cpus[cpu_id].physical_package_id;
cpus             5143 tools/power/x86/turbostat/turbostat.c 	int node_id = cpus[cpu_id].logical_node_id;
cpus             5144 tools/power/x86/turbostat/turbostat.c 	int core_id = cpus[cpu_id].physical_core_id;
cpus             5145 tools/power/x86/turbostat/turbostat.c 	int thread_id = cpus[cpu_id].thread_id;
cpus              233 tools/testing/selftests/net/reuseport_bpf_cpu.c 	int *rcv_fd, cpus;
cpus              235 tools/testing/selftests/net/reuseport_bpf_cpu.c 	cpus = sysconf(_SC_NPROCESSORS_ONLN);
cpus              236 tools/testing/selftests/net/reuseport_bpf_cpu.c 	if (cpus <= 0)
cpus              239 tools/testing/selftests/net/reuseport_bpf_cpu.c 	rcv_fd = calloc(cpus, sizeof(int));
cpus              244 tools/testing/selftests/net/reuseport_bpf_cpu.c 	test(rcv_fd, cpus, AF_INET, SOCK_DGRAM);
cpus              247 tools/testing/selftests/net/reuseport_bpf_cpu.c 	test(rcv_fd, cpus, AF_INET6, SOCK_DGRAM);
cpus              250 tools/testing/selftests/net/reuseport_bpf_cpu.c 	test(rcv_fd, cpus, AF_INET, SOCK_STREAM);
cpus              253 tools/testing/selftests/net/reuseport_bpf_cpu.c 	test(rcv_fd, cpus, AF_INET6, SOCK_STREAM);
cpus               60 tools/virtio/virtio-trace/trace-agent.c 	s->cpus = get_total_cpus();
cpus               64 tools/virtio/virtio-trace/trace-agent.c 	for (i = 0; i < s->cpus; i++)
cpus              157 tools/virtio/virtio-trace/trace-agent.c 	for (cpu = 0; cpu < s->cpus; cpu++) {
cpus              224 tools/virtio/virtio-trace/trace-agent.c 	for (cpu = 0; cpu < s->cpus; cpu++)
cpus              230 tools/virtio/virtio-trace/trace-agent.c 	for (cpu = 0; cpu < s->cpus; cpu++) {
cpus              246 tools/virtio/virtio-trace/trace-agent.c 	for (i = 0; i < s->cpus; i++) {
cpus               21 tools/virtio/virtio-trace/trace-agent.h 	int cpus;
cpus              236 virt/kvm/kvm_main.c static inline bool kvm_kick_many_cpus(const struct cpumask *cpus, bool wait)
cpus              238 virt/kvm/kvm_main.c 	if (unlikely(!cpus))
cpus              239 virt/kvm/kvm_main.c 		cpus = cpu_online_mask;
cpus              241 virt/kvm/kvm_main.c 	if (cpumask_empty(cpus))
cpus              244 virt/kvm/kvm_main.c 	smp_call_function_many(cpus, ack_flush, NULL, wait);
cpus              280 virt/kvm/kvm_main.c 	cpumask_var_t cpus;
cpus              283 virt/kvm/kvm_main.c 	zalloc_cpumask_var(&cpus, GFP_ATOMIC);
cpus              285 virt/kvm/kvm_main.c 	called = kvm_make_vcpus_request_mask(kvm, req, NULL, cpus);
cpus              287 virt/kvm/kvm_main.c 	free_cpumask_var(cpus);