new_raw_count     310 arch/alpha/kernel/perf_event.c 	long prev_raw_count, new_raw_count;
new_raw_count     315 arch/alpha/kernel/perf_event.c 	new_raw_count = alpha_read_pmc(idx);
new_raw_count     318 arch/alpha/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
new_raw_count     321 arch/alpha/kernel/perf_event.c 	delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf;
new_raw_count     333 arch/alpha/kernel/perf_event.c 	return new_raw_count;
new_raw_count     122 arch/arc/kernel/perf_event.c 	u64 new_raw_count = arc_pmu_read_counter(idx);
new_raw_count     123 arch/arc/kernel/perf_event.c 	s64 delta = new_raw_count - prev_raw_count;
new_raw_count     129 arch/arc/kernel/perf_event.c 	local64_set(&hwc->prev_count, new_raw_count);
new_raw_count     308 arch/arm/mach-imx/mmdc.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     312 arch/arm/mach-imx/mmdc.c 		new_raw_count = mmdc_pmu_read_counter(pmu_mmdc,
new_raw_count     315 arch/arm/mach-imx/mmdc.c 		new_raw_count) != prev_raw_count);
new_raw_count     317 arch/arm/mach-imx/mmdc.c 	delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF;
new_raw_count     927 arch/csky/kernel/perf_event.c 	uint64_t new_raw_count = sign_extend64(
new_raw_count     929 arch/csky/kernel/perf_event.c 	int64_t delta = new_raw_count - prev_raw_count;
new_raw_count     935 arch/csky/kernel/perf_event.c 	local64_set(&hwc->prev_count, new_raw_count);
new_raw_count     410 arch/mips/kernel/perf_event_mipsxx.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     415 arch/mips/kernel/perf_event_mipsxx.c 	new_raw_count = mipspmu.read_counter(idx);
new_raw_count     418 arch/mips/kernel/perf_event_mipsxx.c 				new_raw_count) != prev_raw_count)
new_raw_count     421 arch/mips/kernel/perf_event_mipsxx.c 	delta = new_raw_count - prev_raw_count;
new_raw_count     947 arch/nds32/kernel/perf_event_cpu.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     951 arch/nds32/kernel/perf_event_cpu.c 	new_raw_count = nds32_pmu->read_counter(event);
new_raw_count     954 arch/nds32/kernel/perf_event_cpu.c 			    new_raw_count) != prev_raw_count) {
new_raw_count     961 arch/nds32/kernel/perf_event_cpu.c 	delta = (new_raw_count - prev_raw_count) & nds32_pmu->max_period;
new_raw_count     966 arch/nds32/kernel/perf_event_cpu.c 	return new_raw_count;
new_raw_count     217 arch/riscv/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     224 arch/riscv/kernel/perf_event.c 		new_raw_count = read_counter(idx);
new_raw_count     227 arch/riscv/kernel/perf_event.c 					 new_raw_count);
new_raw_count     233 arch/riscv/kernel/perf_event.c 	delta = (new_raw_count - prev_raw_count) &
new_raw_count     179 arch/sh/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     197 arch/sh/kernel/perf_event.c 	new_raw_count = sh_pmu->read(idx);
new_raw_count     200 arch/sh/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
new_raw_count     211 arch/sh/kernel/perf_event.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
new_raw_count     867 arch/sparc/kernel/perf_event.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     872 arch/sparc/kernel/perf_event.c 	new_raw_count = sparc_pmu->read_pmc(idx);
new_raw_count     875 arch/sparc/kernel/perf_event.c 			     new_raw_count) != prev_raw_count)
new_raw_count     878 arch/sparc/kernel/perf_event.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
new_raw_count     884 arch/sparc/kernel/perf_event.c 	return new_raw_count;
new_raw_count     154 arch/x86/events/amd/ibs.c perf_event_try_update(struct perf_event *event, u64 new_raw_count, int width)
new_raw_count     170 arch/x86/events/amd/ibs.c 					new_raw_count) != prev_raw_count)
new_raw_count     181 arch/x86/events/amd/ibs.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
new_raw_count      72 arch/x86/events/core.c 	u64 prev_raw_count, new_raw_count;
new_raw_count      88 arch/x86/events/core.c 	rdpmcl(hwc->event_base_rdpmc, new_raw_count);
new_raw_count      91 arch/x86/events/core.c 					new_raw_count) != prev_raw_count)
new_raw_count     102 arch/x86/events/core.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
new_raw_count     108 arch/x86/events/core.c 	return new_raw_count;
new_raw_count     361 arch/x86/events/intel/cstate.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     365 arch/x86/events/intel/cstate.c 	new_raw_count = cstate_pmu_read_counter(event);
new_raw_count     368 arch/x86/events/intel/cstate.c 			    new_raw_count) != prev_raw_count)
new_raw_count     371 arch/x86/events/intel/cstate.c 	local64_add(new_raw_count - prev_raw_count, &event->count);
new_raw_count    1671 arch/x86/events/intel/ds.c 	u64 prev_raw_count, new_raw_count;
new_raw_count    1682 arch/x86/events/intel/ds.c 	rdpmcl(hwc->event_base_rdpmc, new_raw_count);
new_raw_count    1683 arch/x86/events/intel/ds.c 	local64_set(&hwc->prev_count, new_raw_count);
new_raw_count    1712 arch/x86/events/intel/ds.c 	new = ((s64)(new_raw_count << shift) >> shift);
new_raw_count     179 arch/x86/events/intel/rapl.c 	u64 prev_raw_count, new_raw_count;
new_raw_count     185 arch/x86/events/intel/rapl.c 	rdmsrl(event->hw.event_base, new_raw_count);
new_raw_count     188 arch/x86/events/intel/rapl.c 			    new_raw_count) != prev_raw_count) {
new_raw_count     201 arch/x86/events/intel/rapl.c 	delta = (new_raw_count << shift) - (prev_raw_count << shift);
new_raw_count     208 arch/x86/events/intel/rapl.c 	return new_raw_count;
new_raw_count     139 arch/xtensa/kernel/perf_event.c 	uint64_t prev_raw_count, new_raw_count;
new_raw_count     144 arch/xtensa/kernel/perf_event.c 		new_raw_count = xtensa_pmu_read_counter(event->hw.idx);
new_raw_count     146 arch/xtensa/kernel/perf_event.c 				 new_raw_count) != prev_raw_count);
new_raw_count     148 arch/xtensa/kernel/perf_event.c 	delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK;
new_raw_count     986 drivers/perf/arm-cci.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     990 drivers/perf/arm-cci.c 		new_raw_count = pmu_read_counter(event);
new_raw_count     992 drivers/perf/arm-cci.c 		 new_raw_count) != prev_raw_count);
new_raw_count     994 drivers/perf/arm-cci.c 	delta = (new_raw_count - prev_raw_count) & CCI_PMU_CNTR_MASK;
new_raw_count     998 drivers/perf/arm-cci.c 	return new_raw_count;
new_raw_count     166 drivers/perf/arm_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     171 drivers/perf/arm_pmu.c 	new_raw_count = armpmu->read_counter(event);
new_raw_count     174 drivers/perf/arm_pmu.c 			     new_raw_count) != prev_raw_count)
new_raw_count     177 drivers/perf/arm_pmu.c 	delta = (new_raw_count - prev_raw_count) & max_period;
new_raw_count     182 drivers/perf/arm_pmu.c 	return new_raw_count;
new_raw_count     308 drivers/perf/fsl_imx8_ddr_perf.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     313 drivers/perf/fsl_imx8_ddr_perf.c 		new_raw_count = ddr_perf_read_counter(pmu, counter);
new_raw_count     315 drivers/perf/fsl_imx8_ddr_perf.c 			new_raw_count) != prev_raw_count);
new_raw_count     317 drivers/perf/fsl_imx8_ddr_perf.c 	delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF;
new_raw_count     227 drivers/perf/hisilicon/hisi_uncore_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     231 drivers/perf/hisilicon/hisi_uncore_pmu.c 		new_raw_count = hisi_pmu->ops->read_counter(hisi_pmu, hwc);
new_raw_count     234 drivers/perf/hisilicon/hisi_uncore_pmu.c 				 new_raw_count) != prev_raw_count);
new_raw_count     238 drivers/perf/hisilicon/hisi_uncore_pmu.c 	delta = (new_raw_count - prev_raw_count) &
new_raw_count     995 drivers/perf/xgene_pmu.c 	u64 delta, prev_raw_count, new_raw_count;
new_raw_count     999 drivers/perf/xgene_pmu.c 	new_raw_count = xgene_pmu->ops->read_counter(pmu_dev, GET_CNTR(event));
new_raw_count    1002 drivers/perf/xgene_pmu.c 			    new_raw_count) != prev_raw_count)
new_raw_count    1005 drivers/perf/xgene_pmu.c 	delta = (new_raw_count - prev_raw_count) & pmu_dev->max_period;