excl_cntrs        957 arch/x86/events/core.c 		    READ_ONCE(cpuc->excl_cntrs->exclusive_present))
excl_cntrs       2806 arch/x86/events/intel/core.c 	struct intel_excl_cntrs *excl_cntrs = cpuc->excl_cntrs;
excl_cntrs       2819 arch/x86/events/intel/core.c 	if (WARN_ON_ONCE(!excl_cntrs))
excl_cntrs       2822 arch/x86/events/intel/core.c 	xl = &excl_cntrs->states[tid];
excl_cntrs       2830 arch/x86/events/intel/core.c 	raw_spin_lock(&excl_cntrs->lock);
excl_cntrs       2835 arch/x86/events/intel/core.c 	struct intel_excl_cntrs *excl_cntrs = cpuc->excl_cntrs;
excl_cntrs       2843 arch/x86/events/intel/core.c 	if (WARN_ON_ONCE(!excl_cntrs))
excl_cntrs       2849 arch/x86/events/intel/core.c 	xl = &excl_cntrs->states[tid];
excl_cntrs       2851 arch/x86/events/intel/core.c 	lockdep_assert_held(&excl_cntrs->lock);
excl_cntrs       2862 arch/x86/events/intel/core.c 	struct intel_excl_cntrs *excl_cntrs = cpuc->excl_cntrs;
excl_cntrs       2874 arch/x86/events/intel/core.c 	if (WARN_ON_ONCE(!excl_cntrs))
excl_cntrs       2877 arch/x86/events/intel/core.c 	xl = &excl_cntrs->states[tid];
excl_cntrs       2883 arch/x86/events/intel/core.c 	raw_spin_unlock(&excl_cntrs->lock);
excl_cntrs       2919 arch/x86/events/intel/core.c 	struct intel_excl_cntrs *excl_cntrs = cpuc->excl_cntrs;
excl_cntrs       2934 arch/x86/events/intel/core.c 	if (WARN_ON_ONCE(!excl_cntrs))
excl_cntrs       2957 arch/x86/events/intel/core.c 	xlo = &excl_cntrs->states[tid ^ 1];
excl_cntrs       2967 arch/x86/events/intel/core.c 			WRITE_ONCE(excl_cntrs->has_exclusive[tid], 1);
excl_cntrs       3036 arch/x86/events/intel/core.c 	if (cpuc->excl_cntrs)
excl_cntrs       3046 arch/x86/events/intel/core.c 	struct intel_excl_cntrs *excl_cntrs = cpuc->excl_cntrs;
excl_cntrs       3056 arch/x86/events/intel/core.c 	if (WARN_ON_ONCE(!excl_cntrs))
excl_cntrs       3062 arch/x86/events/intel/core.c 			WRITE_ONCE(excl_cntrs->has_exclusive[tid], 0);
excl_cntrs       3070 arch/x86/events/intel/core.c 		xl = &excl_cntrs->states[tid];
excl_cntrs       3078 arch/x86/events/intel/core.c 			raw_spin_lock(&excl_cntrs->lock);
excl_cntrs       3083 arch/x86/events/intel/core.c 			raw_spin_unlock(&excl_cntrs->lock);
excl_cntrs       3112 arch/x86/events/intel/core.c 	if (cpuc->excl_cntrs)
excl_cntrs       3667 arch/x86/events/intel/core.c 		cpuc->excl_cntrs = allocate_excl_cntrs(cpu);
excl_cntrs       3668 arch/x86/events/intel/core.c 		if (!cpuc->excl_cntrs)
excl_cntrs       3759 arch/x86/events/intel/core.c 			c = sibling->excl_cntrs;
excl_cntrs       3761 arch/x86/events/intel/core.c 				cpuc->kfree_on_online[1] = cpuc->excl_cntrs;
excl_cntrs       3762 arch/x86/events/intel/core.c 				cpuc->excl_cntrs = c;
excl_cntrs       3768 arch/x86/events/intel/core.c 		cpuc->excl_cntrs->core_id = core_id;
excl_cntrs       3769 arch/x86/events/intel/core.c 		cpuc->excl_cntrs->refcnt++;
excl_cntrs       3777 arch/x86/events/intel/core.c 	c = cpuc->excl_cntrs;
excl_cntrs       3781 arch/x86/events/intel/core.c 		cpuc->excl_cntrs = NULL;
excl_cntrs        262 arch/x86/events/perf_event.h 	struct intel_excl_cntrs		*excl_cntrs;