all_cpus          531 arch/x86/kernel/cpu/microcode/core.c 	int all_cpus = num_online_cpus();
all_cpus          535 arch/x86/kernel/cpu/microcode/core.c 	while (atomic_read(t) < all_cpus) {
all_cpus          538 arch/x86/kernel/cpu/microcode/core.c 				all_cpus - atomic_read(t));
all_cpus         1365 arch/x86/kvm/hyperv.c 	bool all_cpus;
all_cpus         1384 arch/x86/kvm/hyperv.c 		all_cpus = (flush.flags & HV_FLUSH_ALL_PROCESSORS) ||
all_cpus         1397 arch/x86/kvm/hyperv.c 		all_cpus = flush_ex.hv_vp_set.format !=
all_cpus         1404 arch/x86/kvm/hyperv.c 		if (!sparse_banks_len && !all_cpus)
all_cpus         1407 arch/x86/kvm/hyperv.c 		if (!all_cpus &&
all_cpus         1418 arch/x86/kvm/hyperv.c 	vcpu_mask = all_cpus ? NULL :
all_cpus         1468 arch/x86/kvm/hyperv.c 	bool all_cpus;
all_cpus         1484 arch/x86/kvm/hyperv.c 		all_cpus = false;
all_cpus         1502 arch/x86/kvm/hyperv.c 		all_cpus = send_ipi_ex.vp_set.format == HV_GENERIC_SET_ALL;
all_cpus         1507 arch/x86/kvm/hyperv.c 		if (!all_cpus &&
all_cpus         1519 arch/x86/kvm/hyperv.c 	vcpu_mask = all_cpus ? NULL :
all_cpus           60 tools/bpf/bpftool/map_perf_ring.c 	bool all_cpus;
all_cpus           75 tools/bpf/bpftool/map_perf_ring.c 	int idx = ctx->all_cpus ? cpu : ctx->idx;
all_cpus          130 tools/bpf/bpftool/map_perf_ring.c 		.all_cpus = true,
all_cpus          181 tools/bpf/bpftool/map_perf_ring.c 		ctx.all_cpus = false;
all_cpus          184 tools/bpf/bpftool/map_perf_ring.c 	if (!ctx.all_cpus) {
all_cpus          197 tools/bpf/bpftool/map_perf_ring.c 	opts.cpu_cnt = ctx.all_cpus ? 0 : 1;