active_events 101 arch/arm/mach-imx/mmdc.c unsigned int active_events; active_events 372 arch/arm/mach-imx/mmdc.c pmu_mmdc->active_events++; active_events 401 arch/arm/mach-imx/mmdc.c pmu_mmdc->active_events--; active_events 403 arch/arm/mach-imx/mmdc.c if (pmu_mmdc->active_events == 0) active_events 451 arch/arm/mach-imx/mmdc.c .active_events = 0, active_events 545 arch/mips/kernel/perf_event_mipsxx.c static atomic_t active_events = ATOMIC_INIT(0); active_events 596 arch/mips/kernel/perf_event_mipsxx.c if (atomic_dec_and_mutex_lock(&active_events, active_events 630 arch/mips/kernel/perf_event_mipsxx.c if (!atomic_inc_not_zero(&active_events)) { active_events 632 arch/mips/kernel/perf_event_mipsxx.c if (atomic_read(&active_events) == 0) active_events 636 arch/mips/kernel/perf_event_mipsxx.c atomic_inc(&active_events); active_events 83 arch/nds32/include/asm/pmu.h atomic_t active_events; active_events 861 arch/nds32/kernel/perf_event_cpu.c atomic_t *active_events = &nds32_pmu->active_events; active_events 870 arch/nds32/kernel/perf_event_cpu.c if (!atomic_inc_not_zero(active_events)) { active_events 871 arch/nds32/kernel/perf_event_cpu.c if (atomic_read(active_events) == 0) { active_events 877 arch/nds32/kernel/perf_event_cpu.c atomic_inc(active_events); active_events 1042 arch/nds32/kernel/perf_event_cpu.c atomic_set(&nds32_pmu->active_events, 0); active_events 1158 arch/sparc/kernel/perf_event.c static atomic_t active_events = ATOMIC_INIT(0); active_events 1173 arch/sparc/kernel/perf_event.c if (atomic_inc_not_zero(&active_events)) active_events 1177 arch/sparc/kernel/perf_event.c if (atomic_read(&active_events) == 0) { active_events 1182 arch/sparc/kernel/perf_event.c atomic_inc(&active_events); active_events 1189 arch/sparc/kernel/perf_event.c if (atomic_dec_and_mutex_lock(&active_events, &pmc_grab_mutex)) { active_events 1624 arch/sparc/kernel/perf_event.c if (!atomic_read(&active_events)) active_events 141 arch/x86/events/core.c static atomic_t active_events; active_events 283 arch/x86/events/core.c atomic_dec(&active_events); active_events 391 arch/x86/events/core.c atomic_inc(&active_events); active_events 401 arch/x86/events/core.c atomic_dec(&active_events); active_events 600 arch/x86/events/core.c atomic_inc(&active_events); active_events 1562 arch/x86/events/core.c if (!atomic_read(&active_events)) active_events 111 drivers/perf/arm-cci.c atomic_t active_events; active_events 1087 drivers/perf/arm-cci.c atomic_t *active_events = &cci_pmu->active_events; active_events 1090 drivers/perf/arm-cci.c if (atomic_dec_and_mutex_lock(active_events, reserve_mutex)) { active_events 1320 drivers/perf/arm-cci.c atomic_t *active_events = &cci_pmu->active_events; active_events 1344 drivers/perf/arm-cci.c if (!atomic_inc_not_zero(active_events)) { active_events 1346 drivers/perf/arm-cci.c if (atomic_read(active_events) == 0) active_events 1349 drivers/perf/arm-cci.c atomic_inc(active_events); active_events 1686 drivers/perf/arm-cci.c atomic_set(&cci_pmu->active_events, 0); active_events 75 drivers/perf/fsl_imx8_ddr_perf.c int active_events; active_events 390 drivers/perf/fsl_imx8_ddr_perf.c pmu->active_events++; active_events 422 drivers/perf/fsl_imx8_ddr_perf.c pmu->active_events--;