nr_cpus 60 arch/mips/include/asm/mach-loongson64/boot_param.h u32 nr_cpus; nr_cpus 194 arch/mips/include/asm/mach-loongson64/boot_param.h u32 nr_cpus; nr_cpus 142 arch/mips/loongson64/common/env.c loongson_sysconf.nr_cpus = ecpu->nr_cpus; nr_cpus 145 arch/mips/loongson64/common/env.c if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0) nr_cpus 146 arch/mips/loongson64/common/env.c loongson_sysconf.nr_cpus = NR_CPUS; nr_cpus 147 arch/mips/loongson64/common/env.c loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus + nr_cpus 64 arch/mips/loongson64/loongson-3/numa.c loongson_sysconf.nr_cpus, num_online_nodes()); nr_cpus 223 arch/mips/loongson64/loongson-3/numa.c for (cpu = 0; cpu < loongson_sysconf.nr_cpus; cpu++) { nr_cpus 356 arch/mips/loongson64/loongson-3/smp.c while (i < loongson_sysconf.nr_cpus) { nr_cpus 370 arch/mips/loongson64/loongson-3/smp.c while (num < loongson_sysconf.nr_cpus) { nr_cpus 111 arch/powerpc/platforms/pseries/setup.c int nr_cpus = num_possible_cpus(); nr_cpus 135 arch/powerpc/platforms/pseries/setup.c mce_data_buf = memblock_alloc_try_nid_raw(RTAS_ERROR_LOG_MAX * nr_cpus, nr_cpus 140 arch/powerpc/platforms/pseries/setup.c RTAS_ERROR_LOG_MAX * nr_cpus, &ppc64_rma_size); nr_cpus 150 arch/powerpc/platforms/pseries/setup.c size = sizeof(struct slb_entry) * mmu_slb_size * nr_cpus; nr_cpus 1888 arch/powerpc/sysdev/mpic.c int nr_cpus; nr_cpus 1892 arch/powerpc/sysdev/mpic.c nr_cpus = num_possible_cpus(); nr_cpus 1894 arch/powerpc/sysdev/mpic.c DBG("nr_cpus: %d\n", nr_cpus); nr_cpus 1896 arch/powerpc/sysdev/mpic.c if (nr_cpus > 1) nr_cpus 64 arch/s390/appldata/appldata_os.c u32 nr_cpus; /* number of (virtual) CPUs */ nr_cpus 134 arch/s390/appldata/appldata_os.c os_data->nr_cpus = j; nr_cpus 137 arch/s390/appldata/appldata_os.c (os_data->nr_cpus * sizeof(struct appldata_os_per_cpu)); nr_cpus 107 arch/s390/include/asm/pci_insn.h u16 nr_cpus; nr_cpus 387 arch/s390/pci/pci_irq.c iib.diib.nr_cpus = num_possible_cpus(); nr_cpus 2717 drivers/block/xen-blkfront.c int nr_cpus = num_online_cpus(); nr_cpus 2740 drivers/block/xen-blkfront.c if (xen_blkif_max_queues > nr_cpus) { nr_cpus 2742 drivers/block/xen-blkfront.c xen_blkif_max_queues, nr_cpus); nr_cpus 2743 drivers/block/xen-blkfront.c xen_blkif_max_queues = nr_cpus; nr_cpus 98 drivers/pci/controller/pcie-iproc-msi.c int nr_cpus; nr_cpus 197 drivers/pci/controller/pcie-iproc-msi.c return (hwirq % msi->nr_cpus); nr_cpus 252 drivers/pci/controller/pcie-iproc-msi.c msi->nr_cpus, 0); nr_cpus 254 drivers/pci/controller/pcie-iproc-msi.c bitmap_set(msi->bitmap, hwirq, msi->nr_cpus); nr_cpus 282 drivers/pci/controller/pcie-iproc-msi.c bitmap_clear(msi->bitmap, hwirq, msi->nr_cpus); nr_cpus 470 drivers/pci/controller/pcie-iproc-msi.c for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { nr_cpus 482 drivers/pci/controller/pcie-iproc-msi.c for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { nr_cpus 534 drivers/pci/controller/pcie-iproc-msi.c msi->nr_cpus = num_possible_cpus(); nr_cpus 548 drivers/pci/controller/pcie-iproc-msi.c if (msi->nr_irqs < msi->nr_cpus) { nr_cpus 554 drivers/pci/controller/pcie-iproc-msi.c if (msi->nr_irqs % msi->nr_cpus != 0) { nr_cpus 555 drivers/pci/controller/pcie-iproc-msi.c msi->nr_irqs -= msi->nr_irqs % msi->nr_cpus; nr_cpus 169 drivers/platform/mips/cpu_hwmon.c nr_packages = loongson_sysconf.nr_cpus / nr_cpus 2888 drivers/scsi/mpt3sas/mpt3sas_base.c unsigned int cpu, nr_cpus, nr_msix, index = 0; nr_cpus 2900 drivers/scsi/mpt3sas/mpt3sas_base.c nr_cpus = num_online_cpus(); nr_cpus 2950 drivers/scsi/mpt3sas/mpt3sas_base.c unsigned int i, group = nr_cpus / nr_msix; nr_cpus 2955 drivers/scsi/mpt3sas/mpt3sas_base.c if (cpu >= nr_cpus) nr_cpus 2958 drivers/scsi/mpt3sas/mpt3sas_base.c if (index < nr_cpus % nr_msix) nr_cpus 102 fs/fscache/main.c unsigned int nr_cpus = num_possible_cpus(); nr_cpus 107 fs/fscache/main.c clamp_val(nr_cpus, nr_cpus 1245 kernel/kexec_file.c unsigned long nr_cpus = num_possible_cpus(), nr_phdr, elf_sz; nr_cpus 1252 kernel/kexec_file.c nr_phdr = nr_cpus + 1; nr_cpus 5381 kernel/sched/fair.c atomic_t nr_cpus; nr_cpus 9421 kernel/sched/fair.c if (likely(!atomic_read(&nohz.nr_cpus))) nr_cpus 9536 kernel/sched/fair.c atomic_dec(&nohz.nr_cpus); nr_cpus 9598 kernel/sched/fair.c atomic_inc(&nohz.nr_cpus); nr_cpus 344 kernel/sched/topology.c int i, nr_pd = 0, nr_cs = 0, nr_cpus = cpumask_weight(cpu_map); nr_cpus 397 kernel/sched/topology.c if (nr_pd * (nr_cs + nr_cpus) > EM_MAX_COMPLEXITY) { nr_cpus 545 kernel/smp.c int nr_cpus; nr_cpus 547 kernel/smp.c get_option(&str, &nr_cpus); nr_cpus 548 kernel/smp.c if (nr_cpus > 0 && nr_cpus < nr_cpu_ids) nr_cpus 549 kernel/smp.c nr_cpu_ids = nr_cpus; nr_cpus 31 samples/bpf/lwt_len_hist_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 33 samples/bpf/lwt_len_hist_user.c uint64_t values[nr_cpus], sum, max_value = 0, data[MAX_INDEX] = {}; nr_cpus 54 samples/bpf/lwt_len_hist_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 28 samples/bpf/sampleip_user.c static int nr_cpus; nr_cpus 49 samples/bpf/sampleip_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 68 samples/bpf/sampleip_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 166 samples/bpf/sampleip_user.c nr_cpus = sysconf(_SC_NPROCESSORS_CONF); nr_cpus 167 samples/bpf/sampleip_user.c pmu_fd = malloc(nr_cpus * sizeof(int)); nr_cpus 32 samples/bpf/test_lru_dist.c static int nr_cpus; nr_cpus 225 samples/bpf/test_lru_dist.c if (next_to_try == nr_cpus) nr_cpus 228 samples/bpf/test_lru_dist.c while (next_to_try < nr_cpus) { nr_cpus 321 samples/bpf/test_lru_dist.c nr_cpus * lru_size); nr_cpus 337 samples/bpf/test_lru_dist.c unsigned long long key, value[nr_cpus]; nr_cpus 349 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 900 * nr_cpus); nr_cpus 391 samples/bpf/test_lru_dist.c unsigned long long key, value[nr_cpus]; nr_cpus 401 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 1000 * nr_cpus); nr_cpus 429 samples/bpf/test_lru_dist.c unsigned long long key, value[nr_cpus]; nr_cpus 478 samples/bpf/test_lru_dist.c nr_cpus * (1000 + 200)); nr_cpus 515 samples/bpf/test_lru_dist.c nr_cpus = bpf_num_possible_cpus(); nr_cpus 516 samples/bpf/test_lru_dist.c assert(nr_cpus != -1); nr_cpus 517 samples/bpf/test_lru_dist.c printf("nr_cpus:%d\n\n", nr_cpus); nr_cpus 519 samples/bpf/test_lru_dist.c nr_tasks = min(nr_tasks, nr_cpus); nr_cpus 139 samples/bpf/trace_event_user.c int nr_cpus = sysconf(_SC_NPROCESSORS_CONF); nr_cpus 140 samples/bpf/trace_event_user.c int *pmu_fd = malloc(nr_cpus * sizeof(int)); nr_cpus 147 samples/bpf/trace_event_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 43 samples/bpf/tracex2_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 45 samples/bpf/tracex2_user.c long values[nr_cpus]; nr_cpus 60 samples/bpf/tracex2_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 21 samples/bpf/tracex3_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 22 samples/bpf/tracex3_user.c __u64 values[nr_cpus]; nr_cpus 78 samples/bpf/tracex3_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 80 samples/bpf/tracex3_user.c long values[nr_cpus]; nr_cpus 90 samples/bpf/tracex3_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 75 samples/bpf/tracex6_user.c int i, status, nr_cpus = sysconf(_SC_NPROCESSORS_CONF); nr_cpus 76 samples/bpf/tracex6_user.c pid_t pid[nr_cpus]; nr_cpus 81 samples/bpf/tracex6_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 90 samples/bpf/tracex6_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 46 samples/bpf/xdp1_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 47 samples/bpf/xdp1_user.c __u64 values[nr_cpus], prev[UINT8_MAX] = { 0 }; nr_cpus 59 samples/bpf/xdp1_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 151 samples/bpf/xdp_monitor_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 152 samples/bpf/xdp_monitor_user.c struct datarec values[nr_cpus]; nr_cpus 168 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 188 samples/bpf/xdp_monitor_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 189 samples/bpf/xdp_monitor_user.c struct u64rec values[nr_cpus]; nr_cpus 202 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 298 samples/bpf/xdp_monitor_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 319 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 344 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 370 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 409 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 442 samples/bpf/xdp_monitor_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 510 samples/bpf/xdp_monitor_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 514 samples/bpf/xdp_monitor_user.c size = record_size * nr_cpus; nr_cpus 518 samples/bpf/xdp_monitor_user.c fprintf(stderr, "Mem alloc error (nr_cpus:%u)\n", nr_cpus); nr_cpus 179 samples/bpf/xdp_redirect_cpu_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 180 samples/bpf/xdp_redirect_cpu_user.c struct datarec values[nr_cpus]; nr_cpus 195 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 211 samples/bpf/xdp_redirect_cpu_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 215 samples/bpf/xdp_redirect_cpu_user.c size = sizeof(struct datarec) * nr_cpus; nr_cpus 219 samples/bpf/xdp_redirect_cpu_user.c fprintf(stderr, "Mem alloc error (nr_cpus:%u)\n", nr_cpus); nr_cpus 312 samples/bpf/xdp_redirect_cpu_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 333 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 361 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 398 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 427 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 449 samples/bpf/xdp_redirect_cpu_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 65 samples/bpf/xdp_redirect_map_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 66 samples/bpf/xdp_redirect_map_user.c __u64 values[nr_cpus], prev[nr_cpus]; nr_cpus 77 samples/bpf/xdp_redirect_map_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 65 samples/bpf/xdp_redirect_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 66 samples/bpf/xdp_redirect_user.c __u64 values[nr_cpus], prev[nr_cpus]; nr_cpus 77 samples/bpf/xdp_redirect_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 517 samples/bpf/xdp_router_ipv4_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 520 samples/bpf/xdp_router_ipv4_user.c __u64 prev[nr_keys][nr_cpus]; nr_cpus 522 samples/bpf/xdp_router_ipv4_user.c __u64 values[nr_cpus]; nr_cpus 576 samples/bpf/xdp_router_ipv4_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 199 samples/bpf/xdp_rxq_info_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 203 samples/bpf/xdp_rxq_info_user.c size = sizeof(struct datarec) * nr_cpus; nr_cpus 207 samples/bpf/xdp_rxq_info_user.c fprintf(stderr, "Mem alloc error (nr_cpus:%u)\n", nr_cpus); nr_cpus 265 samples/bpf/xdp_rxq_info_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 266 samples/bpf/xdp_rxq_info_user.c struct datarec values[nr_cpus]; nr_cpus 280 samples/bpf/xdp_rxq_info_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 346 samples/bpf/xdp_rxq_info_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 369 samples/bpf/xdp_rxq_info_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 403 samples/bpf/xdp_rxq_info_user.c for (i = 0; i < nr_cpus; i++) { nr_cpus 54 samples/bpf/xdp_tx_iptunnel_user.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 56 samples/bpf/xdp_tx_iptunnel_user.c __u64 values[nr_cpus], prev[nr_protos][nr_cpus]; nr_cpus 70 samples/bpf/xdp_tx_iptunnel_user.c for (i = 0; i < nr_cpus; i++) nr_cpus 2142 tools/lib/bpf/libbpf.c int nr_cpus = 0; nr_cpus 2167 tools/lib/bpf/libbpf.c if (!nr_cpus) nr_cpus 2168 tools/lib/bpf/libbpf.c nr_cpus = libbpf_num_possible_cpus(); nr_cpus 2169 tools/lib/bpf/libbpf.c if (nr_cpus < 0) { nr_cpus 2171 tools/lib/bpf/libbpf.c nr_cpus); nr_cpus 2172 tools/lib/bpf/libbpf.c err = nr_cpus; nr_cpus 2176 tools/lib/bpf/libbpf.c map->name, nr_cpus); nr_cpus 2177 tools/lib/bpf/libbpf.c create_attr.max_entries = nr_cpus; nr_cpus 21 tools/perf/arch/arm/util/auxtrace.c int ret, i, nr_cpus = sysconf(_SC_NPROCESSORS_CONF); nr_cpus 25 tools/perf/arch/arm/util/auxtrace.c arm_spe_pmus = zalloc(sizeof(struct perf_pmu *) * nr_cpus); nr_cpus 32 tools/perf/arch/arm/util/auxtrace.c for (i = 0; i < nr_cpus; i++) { nr_cpus 125 tools/perf/bench/numa.c int nr_cpus; nr_cpus 276 tools/perf/bench/numa.c for (cpu = 0; cpu < g->p.nr_cpus; cpu++) nr_cpus 279 tools/perf/bench/numa.c BUG_ON(target_cpu < 0 || target_cpu >= g->p.nr_cpus); nr_cpus 291 tools/perf/bench/numa.c int cpus_per_node = g->p.nr_cpus / nr_numa_nodes(); nr_cpus 296 tools/perf/bench/numa.c BUG_ON(cpus_per_node * nr_numa_nodes() != g->p.nr_cpus); nr_cpus 305 tools/perf/bench/numa.c for (cpu = 0; cpu < g->p.nr_cpus; cpu++) nr_cpus 311 tools/perf/bench/numa.c BUG_ON(cpu_stop > g->p.nr_cpus); nr_cpus 537 tools/perf/bench/numa.c BUG_ON(step <= 0 || step >= g->p.nr_cpus); nr_cpus 549 tools/perf/bench/numa.c BUG_ON(bind_len <= 0 || bind_len > g->p.nr_cpus); nr_cpus 562 tools/perf/bench/numa.c if (bind_cpu_0 >= g->p.nr_cpus || bind_cpu_1 >= g->p.nr_cpus) { nr_cpus 563 tools/perf/bench/numa.c printf("\nTest not applicable, system has only %d CPUs.\n", g->p.nr_cpus); nr_cpus 592 tools/perf/bench/numa.c BUG_ON(cpu < 0 || cpu >= g->p.nr_cpus); nr_cpus 1223 tools/perf/bench/numa.c if (this_cpu < g->p.nr_cpus/2) nr_cpus 1224 tools/perf/bench/numa.c target_cpu = g->p.nr_cpus-1; nr_cpus 1346 tools/perf/bench/numa.c g->p.nr_tasks, g->p.nr_tasks == 1 ? "task" : "tasks", nr_numa_nodes(), g->p.nr_cpus); nr_cpus 1375 tools/perf/bench/numa.c for (cpu = 0; cpu < g->p.nr_cpus; cpu++) nr_cpus 1394 tools/perf/bench/numa.c g->p.nr_cpus = numa_num_configured_cpus(); nr_cpus 51 tools/perf/lib/cpumap.c int nr_cpus; nr_cpus 53 tools/perf/lib/cpumap.c nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); nr_cpus 54 tools/perf/lib/cpumap.c if (nr_cpus < 0) nr_cpus 57 tools/perf/lib/cpumap.c cpus = malloc(sizeof(*cpus) + nr_cpus * sizeof(int)); nr_cpus 61 tools/perf/lib/cpumap.c for (i = 0; i < nr_cpus; ++i) nr_cpus 64 tools/perf/lib/cpumap.c cpus->nr = nr_cpus; nr_cpus 71 tools/perf/lib/cpumap.c static struct perf_cpu_map *cpu_map__trim_new(int nr_cpus, int *tmp_cpus) nr_cpus 73 tools/perf/lib/cpumap.c size_t payload_size = nr_cpus * sizeof(int); nr_cpus 77 tools/perf/lib/cpumap.c cpus->nr = nr_cpus; nr_cpus 88 tools/perf/lib/cpumap.c int nr_cpus = 0; nr_cpus 101 tools/perf/lib/cpumap.c int new_max = nr_cpus + cpu - prev - 1; nr_cpus 115 tools/perf/lib/cpumap.c tmp_cpus[nr_cpus++] = prev; nr_cpus 117 tools/perf/lib/cpumap.c if (nr_cpus == max_entries) { nr_cpus 125 tools/perf/lib/cpumap.c tmp_cpus[nr_cpus++] = cpu; nr_cpus 134 tools/perf/lib/cpumap.c if (nr_cpus > 0) nr_cpus 135 tools/perf/lib/cpumap.c cpus = cpu_map__trim_new(nr_cpus, tmp_cpus); nr_cpus 162 tools/perf/lib/cpumap.c int i, nr_cpus = 0; nr_cpus 203 tools/perf/lib/cpumap.c for (i = 0; i < nr_cpus; i++) nr_cpus 207 tools/perf/lib/cpumap.c if (nr_cpus == max_entries) { nr_cpus 214 tools/perf/lib/cpumap.c tmp_cpus[nr_cpus++] = (int)start_cpu; nr_cpus 222 tools/perf/lib/cpumap.c if (nr_cpus > 0) nr_cpus 223 tools/perf/lib/cpumap.c cpus = cpu_map__trim_new(nr_cpus, tmp_cpus); nr_cpus 248 tools/perf/lib/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->cpus); nr_cpus 255 tools/perf/lib/evlist.c nfds += nr_cpus; nr_cpus 257 tools/perf/lib/evlist.c nfds += nr_cpus * nr_threads; nr_cpus 233 tools/perf/util/env.c int cpu, nr_cpus; nr_cpus 241 tools/perf/util/env.c nr_cpus = env->nr_cpus_avail; nr_cpus 242 tools/perf/util/env.c if (nr_cpus == -1) nr_cpus 245 tools/perf/util/env.c env->cpu = calloc(nr_cpus, sizeof(env->cpu[0])); nr_cpus 249 tools/perf/util/env.c for (cpu = 0; cpu < nr_cpus; ++cpu) { nr_cpus 255 tools/perf/util/env.c env->nr_cpus_avail = nr_cpus; nr_cpus 377 tools/perf/util/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->core.cpus); nr_cpus 382 tools/perf/util/evlist.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 715 tools/perf/util/evlist.c int nr_cpus = perf_cpu_map__nr(evlist->core.cpus); nr_cpus 719 tools/perf/util/evlist.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 1450 tools/perf/util/evsel.c int nr_cpus, int nr_threads, nr_cpus 1453 tools/perf/util/evsel.c for (int cpu = 0; cpu < nr_cpus; cpu++) nr_cpus 1459 tools/perf/util/evsel.c int nr_cpus, int cpu_idx, nr_cpus 1464 tools/perf/util/evsel.c if (cpu_idx >= nr_cpus || thread_idx >= nr_threads) nr_cpus 1468 tools/perf/util/evsel.c nr_cpus = pos != evsel ? nr_cpus : cpu_idx; nr_cpus 1470 tools/perf/util/evsel.c perf_evsel__remove_fd(pos, nr_cpus, nr_threads, thread_idx); nr_cpus 1483 tools/perf/util/evsel.c int nr_cpus, int cpu, nr_cpus 1508 tools/perf/util/evsel.c if (update_fds(evsel, nr_cpus, cpu, threads->nr, thread)) nr_cpus 2590 tools/perf/util/machine.c int nr_cpus = min(machine->env->nr_cpus_online, MAX_NR_CPUS); nr_cpus 2592 tools/perf/util/machine.c if (cpu < 0 || cpu >= nr_cpus || !machine->current_tid) nr_cpus 2602 tools/perf/util/machine.c int nr_cpus = min(machine->env->nr_cpus_online, MAX_NR_CPUS); nr_cpus 2610 tools/perf/util/machine.c machine->current_tid = calloc(nr_cpus, sizeof(pid_t)); nr_cpus 2613 tools/perf/util/machine.c for (i = 0; i < nr_cpus; i++) nr_cpus 2617 tools/perf/util/machine.c if (cpu >= nr_cpus) { nr_cpus 332 tools/perf/util/mmap.c int c, cpu, nr_cpus; nr_cpus 339 tools/perf/util/mmap.c nr_cpus = perf_cpu_map__nr(cpu_map); nr_cpus 340 tools/perf/util/mmap.c for (c = 0; c < nr_cpus; c++) { nr_cpus 2316 tools/perf/util/session.c int nr_cpus = min(session->header.env.nr_cpus_online, MAX_NR_CPUS); nr_cpus 2341 tools/perf/util/session.c if (cpu >= nr_cpus) { nr_cpus 700 tools/perf/util/svghelper.c int *pos, int nr_cpus) nr_cpus 709 tools/perf/util/svghelper.c for_each_set_bit(thr, cpumask_bits(&t->sib_thr[i]), nr_cpus) nr_cpus 715 tools/perf/util/svghelper.c static void scan_core_topology(int *map, struct topology *t, int nr_cpus) nr_cpus 722 tools/perf/util/svghelper.c for_each_set_bit(cpu, cpumask_bits(&t->sib_core[i]), nr_cpus) nr_cpus 723 tools/perf/util/svghelper.c scan_thread_topology(map, t, cpu, &pos, nr_cpus); nr_cpus 726 tools/perf/util/svghelper.c static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus) nr_cpus 739 tools/perf/util/svghelper.c if (c >= nr_cpus) { nr_cpus 754 tools/perf/util/svghelper.c int i, nr_cpus; nr_cpus 758 tools/perf/util/svghelper.c nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS); nr_cpus 774 tools/perf/util/svghelper.c if (str_to_bitmap(sib_core, &t.sib_core[i], nr_cpus)) { nr_cpus 783 tools/perf/util/svghelper.c if (str_to_bitmap(sib_thr, &t.sib_thr[i], nr_cpus)) { nr_cpus 791 tools/perf/util/svghelper.c topology_map = malloc(sizeof(int) * nr_cpus); nr_cpus 797 tools/perf/util/svghelper.c for (i = 0; i < nr_cpus; i++) nr_cpus 800 tools/perf/util/svghelper.c scan_core_topology(topology_map, &t, nr_cpus); nr_cpus 58 tools/power/cpupower/utils/cpufreq-info.c unsigned int cpu, nr_cpus; nr_cpus 66 tools/power/cpupower/utils/cpufreq-info.c nr_cpus = count_cpus(); nr_cpus 67 tools/power/cpupower/utils/cpufreq-info.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 6 tools/testing/selftests/bpf/prog_tests/l4lb_all.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 16 tools/testing/selftests/bpf/prog_tests/l4lb_all.c } stats[nr_cpus]; nr_cpus 69 tools/testing/selftests/bpf/prog_tests/l4lb_all.c for (i = 0; i < nr_cpus; i++) { nr_cpus 23 tools/testing/selftests/bpf/prog_tests/perf_buffer.c int err, prog_fd, on_len, nr_on_cpus = 0, nr_cpus, i, duration = 0; nr_cpus 35 tools/testing/selftests/bpf/prog_tests/perf_buffer.c nr_cpus = libbpf_num_possible_cpus(); nr_cpus 36 tools/testing/selftests/bpf/prog_tests/perf_buffer.c if (CHECK(nr_cpus < 0, "nr_cpus", "err %d\n", nr_cpus)) nr_cpus 79 tools/testing/selftests/bpf/prog_tests/perf_buffer.c for (i = 0; i < nr_cpus; i++) { nr_cpus 7 tools/testing/selftests/bpf/prog_tests/xdp_noinline.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 17 tools/testing/selftests/bpf/prog_tests/xdp_noinline.c } stats[nr_cpus]; nr_cpus 70 tools/testing/selftests/bpf/prog_tests/xdp_noinline.c for (i = 0; i < nr_cpus; i++) { nr_cpus 27 tools/testing/selftests/bpf/test_lru_map.c static int nr_cpus; nr_cpus 107 tools/testing/selftests/bpf/test_lru_map.c unsigned long long value0[nr_cpus], value1[nr_cpus]; nr_cpus 138 tools/testing/selftests/bpf/test_lru_map.c while (next < nr_cpus) { nr_cpus 161 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 171 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); nr_cpus 250 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, end_key, value[nr_cpus]; nr_cpus 326 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 434 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, end_key, value[nr_cpus]; nr_cpus 499 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 510 tools/testing/selftests/bpf/test_lru_map.c 3 * tgt_free * nr_cpus); nr_cpus 557 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 575 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 623 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 638 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_size * nr_cpus); nr_cpus 689 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 699 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); nr_cpus 785 tools/testing/selftests/bpf/test_lru_map.c unsigned long long key, value[nr_cpus]; nr_cpus 795 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); nr_cpus 879 tools/testing/selftests/bpf/test_lru_map.c nr_cpus = bpf_num_possible_cpus(); nr_cpus 880 tools/testing/selftests/bpf/test_lru_map.c assert(nr_cpus != -1); nr_cpus 881 tools/testing/selftests/bpf/test_lru_map.c printf("nr_cpus:%d\n\n", nr_cpus); nr_cpus 149 tools/testing/selftests/bpf/test_maps.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 162 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 224 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 402 tools/testing/selftests/bpf/test_maps.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 413 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 432 tools/testing/selftests/bpf/test_maps.c bpf_percpu(values, nr_cpus - 1) == 0); nr_cpus 461 tools/testing/selftests/bpf/test_maps.c unsigned int nr_cpus = bpf_num_possible_cpus(); nr_cpus 477 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 484 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 489 tools/testing/selftests/bpf/test_maps.c for (i = 0; i < nr_cpus; i++) nr_cpus 62 tools/testing/selftests/vm/userfaultfd.c static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; nr_cpus 586 tools/testing/selftests/vm/userfaultfd.c pthread_t locking_threads[nr_cpus]; nr_cpus 587 tools/testing/selftests/vm/userfaultfd.c pthread_t uffd_threads[nr_cpus]; nr_cpus 588 tools/testing/selftests/vm/userfaultfd.c pthread_t background_threads[nr_cpus]; nr_cpus 592 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 611 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) nr_cpus 629 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) nr_cpus 633 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 1041 tools/testing/selftests/vm/userfaultfd.c unsigned long userfaults[nr_cpus]; nr_cpus 1073 tools/testing/selftests/vm/userfaultfd.c pipefd = malloc(sizeof(int) * nr_cpus * 2); nr_cpus 1078 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) { nr_cpus 1213 tools/testing/selftests/vm/userfaultfd.c for (cpu = 0; cpu < nr_cpus; cpu++) nr_cpus 1303 tools/testing/selftests/vm/userfaultfd.c nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); nr_cpus 1305 tools/testing/selftests/vm/userfaultfd.c nr_cpus; nr_cpus 1316 tools/testing/selftests/vm/userfaultfd.c nr_pages = nr_pages_per_cpu * nr_cpus; nr_cpus 31 tools/virtio/virtio-trace/trace-agent.c int nr_cpus = (int)sysconf(_SC_NPROCESSORS_CONF); nr_cpus 33 tools/virtio/virtio-trace/trace-agent.c if (nr_cpus <= 0) { nr_cpus 36 tools/virtio/virtio-trace/trace-agent.c } else if (nr_cpus > MAX_CPUS) { nr_cpus 41 tools/virtio/virtio-trace/trace-agent.c return nr_cpus; nr_cpus 26 virt/kvm/arm/vgic/vgic-debug.c int nr_cpus; nr_cpus 45 virt/kvm/arm/vgic/vgic-debug.c ++iter->vcpu_id < iter->nr_cpus) nr_cpus 58 virt/kvm/arm/vgic/vgic-debug.c int nr_cpus = atomic_read(&kvm->online_vcpus); nr_cpus 62 virt/kvm/arm/vgic/vgic-debug.c iter->nr_cpus = nr_cpus; nr_cpus 78 virt/kvm/arm/vgic/vgic-debug.c iter->vcpu_id == iter->nr_cpus && nr_cpus 234 virt/kvm/arm/vgic/vgic-debug.c if (iter->vcpu_id < iter->nr_cpus)