pmcr 213 arch/arm/kernel/perf_event_v6.c armv6_pmcr_has_overflowed(unsigned long pmcr) pmcr 215 arch/arm/kernel/perf_event_v6.c return pmcr & ARMV6_PMCR_OVERFLOWED_MASK; pmcr 219 arch/arm/kernel/perf_event_v6.c armv6_pmcr_counter_has_overflowed(unsigned long pmcr, pmcr 225 arch/arm/kernel/perf_event_v6.c ret = pmcr & ARMV6_PMCR_CCOUNT_OVERFLOW; pmcr 227 arch/arm/kernel/perf_event_v6.c ret = pmcr & ARMV6_PMCR_COUNT0_OVERFLOW; pmcr 229 arch/arm/kernel/perf_event_v6.c ret = pmcr & ARMV6_PMCR_COUNT1_OVERFLOW; pmcr 308 arch/arm/kernel/perf_event_v6.c unsigned long pmcr = armv6_pmcr_read(); pmcr 314 arch/arm/kernel/perf_event_v6.c if (!armv6_pmcr_has_overflowed(pmcr)) pmcr 324 arch/arm/kernel/perf_event_v6.c armv6_pmcr_write(pmcr); pmcr 338 arch/arm/kernel/perf_event_v6.c if (!armv6_pmcr_counter_has_overflowed(pmcr, idx)) pmcr 626 arch/arm64/kvm/sys_regs.c u64 pmcr, val; pmcr 628 arch/arm64/kvm/sys_regs.c pmcr = read_sysreg(pmcr_el0); pmcr 633 arch/arm64/kvm/sys_regs.c val = ((pmcr & ~ARMV8_PMU_PMCR_MASK) pmcr 746 arch/arm64/kvm/sys_regs.c u64 pmcr, val; pmcr 748 arch/arm64/kvm/sys_regs.c pmcr = __vcpu_sys_reg(vcpu, PMCR_EL0); pmcr 749 arch/arm64/kvm/sys_regs.c val = (pmcr >> ARMV8_PMU_PMCR_N_SHIFT) & ARMV8_PMU_PMCR_N_MASK; pmcr 104 arch/powerpc/include/asm/cell-regs.h u64 pmcr; /* 0x0880 */ pmcr 118 arch/powerpc/platforms/cell/pervasive.c out_be64(®s->pmcr, in_be64(®s->pmcr) | pmcr 65 drivers/cpufreq/ppc_cbe_cpufreq_pervasive.c value = in_be64(&pmd_regs->pmcr); pmcr 71 drivers/cpufreq/ppc_cbe_cpufreq_pervasive.c out_be64(&pmd_regs->pmcr, value); pmcr 481 drivers/perf/arm_dsu_pmu.c u32 pmcr; pmcr 490 drivers/perf/arm_dsu_pmu.c pmcr = __dsu_pmu_read_pmcr(); pmcr 491 drivers/perf/arm_dsu_pmu.c pmcr |= CLUSTERPMCR_E; pmcr 492 drivers/perf/arm_dsu_pmu.c __dsu_pmu_write_pmcr(pmcr); pmcr 498 drivers/perf/arm_dsu_pmu.c u32 pmcr; pmcr 503 drivers/perf/arm_dsu_pmu.c pmcr = __dsu_pmu_read_pmcr(); pmcr 504 drivers/perf/arm_dsu_pmu.c pmcr &= ~CLUSTERPMCR_E; pmcr 505 drivers/perf/arm_dsu_pmu.c __dsu_pmu_write_pmcr(pmcr);