prev_raw_count 310 arch/alpha/kernel/perf_event.c long prev_raw_count, new_raw_count; prev_raw_count 314 arch/alpha/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 317 arch/alpha/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 318 arch/alpha/kernel/perf_event.c new_raw_count) != prev_raw_count) prev_raw_count 321 arch/alpha/kernel/perf_event.c delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; prev_raw_count 121 arch/arc/kernel/perf_event.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 123 arch/arc/kernel/perf_event.c s64 delta = new_raw_count - prev_raw_count; prev_raw_count 308 arch/arm/mach-imx/mmdc.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 311 arch/arm/mach-imx/mmdc.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 314 arch/arm/mach-imx/mmdc.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 315 arch/arm/mach-imx/mmdc.c new_raw_count) != prev_raw_count); prev_raw_count 317 arch/arm/mach-imx/mmdc.c delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; prev_raw_count 922 arch/csky/kernel/perf_event.c uint64_t prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 929 arch/csky/kernel/perf_event.c int64_t delta = new_raw_count - prev_raw_count; prev_raw_count 410 arch/mips/kernel/perf_event_mipsxx.c u64 prev_raw_count, new_raw_count; prev_raw_count 414 arch/mips/kernel/perf_event_mipsxx.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 417 arch/mips/kernel/perf_event_mipsxx.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 418 arch/mips/kernel/perf_event_mipsxx.c new_raw_count) != prev_raw_count) prev_raw_count 421 arch/mips/kernel/perf_event_mipsxx.c delta = new_raw_count - prev_raw_count; prev_raw_count 947 arch/nds32/kernel/perf_event_cpu.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 950 arch/nds32/kernel/perf_event_cpu.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 953 arch/nds32/kernel/perf_event_cpu.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 954 arch/nds32/kernel/perf_event_cpu.c new_raw_count) != prev_raw_count) { prev_raw_count 961 arch/nds32/kernel/perf_event_cpu.c delta = (new_raw_count - prev_raw_count) & nds32_pmu->max_period; prev_raw_count 217 arch/riscv/kernel/perf_event.c u64 prev_raw_count, new_raw_count; prev_raw_count 223 arch/riscv/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 226 arch/riscv/kernel/perf_event.c oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 228 arch/riscv/kernel/perf_event.c } while (oldval != prev_raw_count); prev_raw_count 233 arch/riscv/kernel/perf_event.c delta = (new_raw_count - prev_raw_count) & prev_raw_count 179 arch/sh/kernel/perf_event.c u64 prev_raw_count, new_raw_count; prev_raw_count 196 arch/sh/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 199 arch/sh/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 200 arch/sh/kernel/perf_event.c new_raw_count) != prev_raw_count) prev_raw_count 211 arch/sh/kernel/perf_event.c delta = (new_raw_count << shift) - (prev_raw_count << shift); prev_raw_count 867 arch/sparc/kernel/perf_event.c u64 prev_raw_count, new_raw_count; prev_raw_count 871 arch/sparc/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 874 arch/sparc/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 875 arch/sparc/kernel/perf_event.c new_raw_count) != prev_raw_count) prev_raw_count 878 arch/sparc/kernel/perf_event.c delta = (new_raw_count << shift) - (prev_raw_count << shift); prev_raw_count 158 arch/x86/events/amd/ibs.c u64 prev_raw_count; prev_raw_count 168 arch/x86/events/amd/ibs.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 169 arch/x86/events/amd/ibs.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 170 arch/x86/events/amd/ibs.c new_raw_count) != prev_raw_count) prev_raw_count 181 arch/x86/events/amd/ibs.c delta = (new_raw_count << shift) - (prev_raw_count << shift); prev_raw_count 289 arch/x86/events/amd/iommu.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 293 arch/x86/events/amd/iommu.c IOMMU_PC_COUNTER_REG, &prev_raw_count); prev_raw_count 72 arch/x86/events/core.c u64 prev_raw_count, new_raw_count; prev_raw_count 87 arch/x86/events/core.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 90 arch/x86/events/core.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 91 arch/x86/events/core.c new_raw_count) != prev_raw_count) prev_raw_count 102 arch/x86/events/core.c delta = (new_raw_count << shift) - (prev_raw_count << shift); prev_raw_count 361 arch/x86/events/intel/cstate.c u64 prev_raw_count, new_raw_count; prev_raw_count 364 arch/x86/events/intel/cstate.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 367 arch/x86/events/intel/cstate.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 368 arch/x86/events/intel/cstate.c new_raw_count) != prev_raw_count) prev_raw_count 371 arch/x86/events/intel/cstate.c local64_add(new_raw_count - prev_raw_count, &event->count); prev_raw_count 1671 arch/x86/events/intel/ds.c u64 prev_raw_count, new_raw_count; prev_raw_count 1681 arch/x86/events/intel/ds.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 1713 arch/x86/events/intel/ds.c old = ((s64)(prev_raw_count << shift) >> shift); prev_raw_count 179 arch/x86/events/intel/rapl.c u64 prev_raw_count, new_raw_count; prev_raw_count 184 arch/x86/events/intel/rapl.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 187 arch/x86/events/intel/rapl.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 188 arch/x86/events/intel/rapl.c new_raw_count) != prev_raw_count) { prev_raw_count 201 arch/x86/events/intel/rapl.c delta = (new_raw_count << shift) - (prev_raw_count << shift); prev_raw_count 139 arch/xtensa/kernel/perf_event.c uint64_t prev_raw_count, new_raw_count; prev_raw_count 143 arch/xtensa/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 145 arch/xtensa/kernel/perf_event.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 146 arch/xtensa/kernel/perf_event.c new_raw_count) != prev_raw_count); prev_raw_count 148 arch/xtensa/kernel/perf_event.c delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; prev_raw_count 986 drivers/perf/arm-cci.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 989 drivers/perf/arm-cci.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 991 drivers/perf/arm-cci.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 992 drivers/perf/arm-cci.c new_raw_count) != prev_raw_count); prev_raw_count 994 drivers/perf/arm-cci.c delta = (new_raw_count - prev_raw_count) & CCI_PMU_CNTR_MASK; prev_raw_count 166 drivers/perf/arm_pmu.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 170 drivers/perf/arm_pmu.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 173 drivers/perf/arm_pmu.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 174 drivers/perf/arm_pmu.c new_raw_count) != prev_raw_count) prev_raw_count 177 drivers/perf/arm_pmu.c delta = (new_raw_count - prev_raw_count) & max_period; prev_raw_count 308 drivers/perf/fsl_imx8_ddr_perf.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 312 drivers/perf/fsl_imx8_ddr_perf.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 314 drivers/perf/fsl_imx8_ddr_perf.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 315 drivers/perf/fsl_imx8_ddr_perf.c new_raw_count) != prev_raw_count); prev_raw_count 317 drivers/perf/fsl_imx8_ddr_perf.c delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; prev_raw_count 227 drivers/perf/hisilicon/hisi_uncore_pmu.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 232 drivers/perf/hisilicon/hisi_uncore_pmu.c prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 233 drivers/perf/hisilicon/hisi_uncore_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_raw_count 234 drivers/perf/hisilicon/hisi_uncore_pmu.c new_raw_count) != prev_raw_count); prev_raw_count 238 drivers/perf/hisilicon/hisi_uncore_pmu.c delta = (new_raw_count - prev_raw_count) & prev_raw_count 256 drivers/perf/hisilicon/hisi_uncore_pmu.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_raw_count 258 drivers/perf/hisilicon/hisi_uncore_pmu.c hisi_pmu->ops->write_counter(hisi_pmu, hwc, prev_raw_count); prev_raw_count 995 drivers/perf/xgene_pmu.c u64 delta, prev_raw_count, new_raw_count; prev_raw_count 998 drivers/perf/xgene_pmu.c prev_raw_count = local64_read(&hw->prev_count); prev_raw_count 1001 drivers/perf/xgene_pmu.c if (local64_cmpxchg(&hw->prev_count, prev_raw_count, prev_raw_count 1002 drivers/perf/xgene_pmu.c new_raw_count) != prev_raw_count) prev_raw_count 1005 drivers/perf/xgene_pmu.c delta = (new_raw_count - prev_raw_count) & pmu_dev->max_period; prev_raw_count 1030 drivers/perf/xgene_pmu.c u64 prev_raw_count = local64_read(&hw->prev_count); prev_raw_count 1033 drivers/perf/xgene_pmu.c prev_raw_count);