all_cpus 531 arch/x86/kernel/cpu/microcode/core.c int all_cpus = num_online_cpus(); all_cpus 535 arch/x86/kernel/cpu/microcode/core.c while (atomic_read(t) < all_cpus) { all_cpus 538 arch/x86/kernel/cpu/microcode/core.c all_cpus - atomic_read(t)); all_cpus 1365 arch/x86/kvm/hyperv.c bool all_cpus; all_cpus 1384 arch/x86/kvm/hyperv.c all_cpus = (flush.flags & HV_FLUSH_ALL_PROCESSORS) || all_cpus 1397 arch/x86/kvm/hyperv.c all_cpus = flush_ex.hv_vp_set.format != all_cpus 1404 arch/x86/kvm/hyperv.c if (!sparse_banks_len && !all_cpus) all_cpus 1407 arch/x86/kvm/hyperv.c if (!all_cpus && all_cpus 1418 arch/x86/kvm/hyperv.c vcpu_mask = all_cpus ? NULL : all_cpus 1468 arch/x86/kvm/hyperv.c bool all_cpus; all_cpus 1484 arch/x86/kvm/hyperv.c all_cpus = false; all_cpus 1502 arch/x86/kvm/hyperv.c all_cpus = send_ipi_ex.vp_set.format == HV_GENERIC_SET_ALL; all_cpus 1507 arch/x86/kvm/hyperv.c if (!all_cpus && all_cpus 1519 arch/x86/kvm/hyperv.c vcpu_mask = all_cpus ? NULL : all_cpus 60 tools/bpf/bpftool/map_perf_ring.c bool all_cpus; all_cpus 75 tools/bpf/bpftool/map_perf_ring.c int idx = ctx->all_cpus ? cpu : ctx->idx; all_cpus 130 tools/bpf/bpftool/map_perf_ring.c .all_cpus = true, all_cpus 181 tools/bpf/bpftool/map_perf_ring.c ctx.all_cpus = false; all_cpus 184 tools/bpf/bpftool/map_perf_ring.c if (!ctx.all_cpus) { all_cpus 197 tools/bpf/bpftool/map_perf_ring.c opts.cpu_cnt = ctx.all_cpus ? 0 : 1;