prev_raw_count    310 arch/alpha/kernel/perf_event.c 	long prev_raw_count, new_raw_count;
prev_raw_count    314 arch/alpha/kernel/perf_event.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    317 arch/alpha/kernel/perf_event.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    318 arch/alpha/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
prev_raw_count    321 arch/alpha/kernel/perf_event.c 	delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf;
prev_raw_count    121 arch/arc/kernel/perf_event.c 	u64 prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    123 arch/arc/kernel/perf_event.c 	s64 delta = new_raw_count - prev_raw_count;
prev_raw_count    308 arch/arm/mach-imx/mmdc.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    311 arch/arm/mach-imx/mmdc.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    314 arch/arm/mach-imx/mmdc.c 	} while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    315 arch/arm/mach-imx/mmdc.c 		new_raw_count) != prev_raw_count);
prev_raw_count    317 arch/arm/mach-imx/mmdc.c 	delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF;
prev_raw_count    922 arch/csky/kernel/perf_event.c 	uint64_t prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    929 arch/csky/kernel/perf_event.c 	int64_t delta = new_raw_count - prev_raw_count;
prev_raw_count    410 arch/mips/kernel/perf_event_mipsxx.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    414 arch/mips/kernel/perf_event_mipsxx.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    417 arch/mips/kernel/perf_event_mipsxx.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    418 arch/mips/kernel/perf_event_mipsxx.c 				new_raw_count) != prev_raw_count)
prev_raw_count    421 arch/mips/kernel/perf_event_mipsxx.c 	delta = new_raw_count - prev_raw_count;
prev_raw_count    947 arch/nds32/kernel/perf_event_cpu.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    950 arch/nds32/kernel/perf_event_cpu.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    953 arch/nds32/kernel/perf_event_cpu.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    954 arch/nds32/kernel/perf_event_cpu.c 			    new_raw_count) != prev_raw_count) {
prev_raw_count    961 arch/nds32/kernel/perf_event_cpu.c 	delta = (new_raw_count - prev_raw_count) & nds32_pmu->max_period;
prev_raw_count    217 arch/riscv/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    223 arch/riscv/kernel/perf_event.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    226 arch/riscv/kernel/perf_event.c 		oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    228 arch/riscv/kernel/perf_event.c 	} while (oldval != prev_raw_count);
prev_raw_count    233 arch/riscv/kernel/perf_event.c 	delta = (new_raw_count - prev_raw_count) &
prev_raw_count    179 arch/sh/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    196 arch/sh/kernel/perf_event.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    199 arch/sh/kernel/perf_event.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    200 arch/sh/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
prev_raw_count    211 arch/sh/kernel/perf_event.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
prev_raw_count    867 arch/sparc/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    871 arch/sparc/kernel/perf_event.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    874 arch/sparc/kernel/perf_event.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    875 arch/sparc/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
prev_raw_count    878 arch/sparc/kernel/perf_event.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
prev_raw_count    158 arch/x86/events/amd/ibs.c 	u64 prev_raw_count;
prev_raw_count    168 arch/x86/events/amd/ibs.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    169 arch/x86/events/amd/ibs.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    170 arch/x86/events/amd/ibs.c 					new_raw_count) != prev_raw_count)
prev_raw_count    181 arch/x86/events/amd/ibs.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
prev_raw_count    289 arch/x86/events/amd/iommu.c 		u64 prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    293 arch/x86/events/amd/iommu.c 				     IOMMU_PC_COUNTER_REG, &prev_raw_count);
prev_raw_count     72 arch/x86/events/core.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count     87 arch/x86/events/core.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count     90 arch/x86/events/core.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count     91 arch/x86/events/core.c 					new_raw_count) != prev_raw_count)
prev_raw_count    102 arch/x86/events/core.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
prev_raw_count    361 arch/x86/events/intel/cstate.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    364 arch/x86/events/intel/cstate.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    367 arch/x86/events/intel/cstate.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    368 arch/x86/events/intel/cstate.c 			    new_raw_count) != prev_raw_count)
prev_raw_count    371 arch/x86/events/intel/cstate.c 	local64_add(new_raw_count - prev_raw_count, &event->count);
prev_raw_count   1671 arch/x86/events/intel/ds.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count   1681 arch/x86/events/intel/ds.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count   1713 arch/x86/events/intel/ds.c 	old = ((s64)(prev_raw_count << shift) >> shift);
prev_raw_count    179 arch/x86/events/intel/rapl.c 	u64 prev_raw_count, new_raw_count;
prev_raw_count    184 arch/x86/events/intel/rapl.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    187 arch/x86/events/intel/rapl.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    188 arch/x86/events/intel/rapl.c 			    new_raw_count) != prev_raw_count) {
prev_raw_count    201 arch/x86/events/intel/rapl.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
prev_raw_count    139 arch/xtensa/kernel/perf_event.c 	uint64_t prev_raw_count, new_raw_count;
prev_raw_count    143 arch/xtensa/kernel/perf_event.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    145 arch/xtensa/kernel/perf_event.c 	} while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    146 arch/xtensa/kernel/perf_event.c 				 new_raw_count) != prev_raw_count);
prev_raw_count    148 arch/xtensa/kernel/perf_event.c 	delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK;
prev_raw_count    986 drivers/perf/arm-cci.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    989 drivers/perf/arm-cci.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    991 drivers/perf/arm-cci.c 	} while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    992 drivers/perf/arm-cci.c 		 new_raw_count) != prev_raw_count);
prev_raw_count    994 drivers/perf/arm-cci.c 	delta = (new_raw_count - prev_raw_count) & CCI_PMU_CNTR_MASK;
prev_raw_count    166 drivers/perf/arm_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    170 drivers/perf/arm_pmu.c 	prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    173 drivers/perf/arm_pmu.c 	if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    174 drivers/perf/arm_pmu.c 			     new_raw_count) != prev_raw_count)
prev_raw_count    177 drivers/perf/arm_pmu.c 	delta = (new_raw_count - prev_raw_count) & max_period;
prev_raw_count    308 drivers/perf/fsl_imx8_ddr_perf.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    312 drivers/perf/fsl_imx8_ddr_perf.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    314 drivers/perf/fsl_imx8_ddr_perf.c 	} while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    315 drivers/perf/fsl_imx8_ddr_perf.c 			new_raw_count) != prev_raw_count);
prev_raw_count    317 drivers/perf/fsl_imx8_ddr_perf.c 	delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF;
prev_raw_count    227 drivers/perf/hisilicon/hisi_uncore_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    232 drivers/perf/hisilicon/hisi_uncore_pmu.c 		prev_raw_count = local64_read(&hwc->prev_count);
prev_raw_count    233 drivers/perf/hisilicon/hisi_uncore_pmu.c 	} while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
prev_raw_count    234 drivers/perf/hisilicon/hisi_uncore_pmu.c 				 new_raw_count) != prev_raw_count);
prev_raw_count    238 drivers/perf/hisilicon/hisi_uncore_pmu.c 	delta = (new_raw_count - prev_raw_count) &
prev_raw_count    256 drivers/perf/hisilicon/hisi_uncore_pmu.c 		u64 prev_raw_count =  local64_read(&hwc->prev_count);
prev_raw_count    258 drivers/perf/hisilicon/hisi_uncore_pmu.c 		hisi_pmu->ops->write_counter(hisi_pmu, hwc, prev_raw_count);
prev_raw_count    995 drivers/perf/xgene_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
prev_raw_count    998 drivers/perf/xgene_pmu.c 	prev_raw_count = local64_read(&hw->prev_count);
prev_raw_count   1001 drivers/perf/xgene_pmu.c 	if (local64_cmpxchg(&hw->prev_count, prev_raw_count,
prev_raw_count   1002 drivers/perf/xgene_pmu.c 			    new_raw_count) != prev_raw_count)
prev_raw_count   1005 drivers/perf/xgene_pmu.c 	delta = (new_raw_count - prev_raw_count) & pmu_dev->max_period;
prev_raw_count   1030 drivers/perf/xgene_pmu.c 		u64 prev_raw_count =  local64_read(&hw->prev_count);
prev_raw_count   1033 drivers/perf/xgene_pmu.c 					      prev_raw_count);