new_raw_count 310 arch/alpha/kernel/perf_event.c long prev_raw_count, new_raw_count; new_raw_count 315 arch/alpha/kernel/perf_event.c new_raw_count = alpha_read_pmc(idx); new_raw_count 318 arch/alpha/kernel/perf_event.c new_raw_count) != prev_raw_count) new_raw_count 321 arch/alpha/kernel/perf_event.c delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; new_raw_count 333 arch/alpha/kernel/perf_event.c return new_raw_count; new_raw_count 122 arch/arc/kernel/perf_event.c u64 new_raw_count = arc_pmu_read_counter(idx); new_raw_count 123 arch/arc/kernel/perf_event.c s64 delta = new_raw_count - prev_raw_count; new_raw_count 129 arch/arc/kernel/perf_event.c local64_set(&hwc->prev_count, new_raw_count); new_raw_count 308 arch/arm/mach-imx/mmdc.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 312 arch/arm/mach-imx/mmdc.c new_raw_count = mmdc_pmu_read_counter(pmu_mmdc, new_raw_count 315 arch/arm/mach-imx/mmdc.c new_raw_count) != prev_raw_count); new_raw_count 317 arch/arm/mach-imx/mmdc.c delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; new_raw_count 927 arch/csky/kernel/perf_event.c uint64_t new_raw_count = sign_extend64( new_raw_count 929 arch/csky/kernel/perf_event.c int64_t delta = new_raw_count - prev_raw_count; new_raw_count 935 arch/csky/kernel/perf_event.c local64_set(&hwc->prev_count, new_raw_count); new_raw_count 410 arch/mips/kernel/perf_event_mipsxx.c u64 prev_raw_count, new_raw_count; new_raw_count 415 arch/mips/kernel/perf_event_mipsxx.c new_raw_count = mipspmu.read_counter(idx); new_raw_count 418 arch/mips/kernel/perf_event_mipsxx.c new_raw_count) != prev_raw_count) new_raw_count 421 arch/mips/kernel/perf_event_mipsxx.c delta = new_raw_count - prev_raw_count; new_raw_count 947 arch/nds32/kernel/perf_event_cpu.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 951 arch/nds32/kernel/perf_event_cpu.c new_raw_count = nds32_pmu->read_counter(event); new_raw_count 954 arch/nds32/kernel/perf_event_cpu.c new_raw_count) != prev_raw_count) { new_raw_count 961 arch/nds32/kernel/perf_event_cpu.c delta = (new_raw_count - prev_raw_count) & nds32_pmu->max_period; new_raw_count 966 arch/nds32/kernel/perf_event_cpu.c return new_raw_count; new_raw_count 217 arch/riscv/kernel/perf_event.c u64 prev_raw_count, new_raw_count; new_raw_count 224 arch/riscv/kernel/perf_event.c new_raw_count = read_counter(idx); new_raw_count 227 arch/riscv/kernel/perf_event.c new_raw_count); new_raw_count 233 arch/riscv/kernel/perf_event.c delta = (new_raw_count - prev_raw_count) & new_raw_count 179 arch/sh/kernel/perf_event.c u64 prev_raw_count, new_raw_count; new_raw_count 197 arch/sh/kernel/perf_event.c new_raw_count = sh_pmu->read(idx); new_raw_count 200 arch/sh/kernel/perf_event.c new_raw_count) != prev_raw_count) new_raw_count 211 arch/sh/kernel/perf_event.c delta = (new_raw_count << shift) - (prev_raw_count << shift); new_raw_count 867 arch/sparc/kernel/perf_event.c u64 prev_raw_count, new_raw_count; new_raw_count 872 arch/sparc/kernel/perf_event.c new_raw_count = sparc_pmu->read_pmc(idx); new_raw_count 875 arch/sparc/kernel/perf_event.c new_raw_count) != prev_raw_count) new_raw_count 878 arch/sparc/kernel/perf_event.c delta = (new_raw_count << shift) - (prev_raw_count << shift); new_raw_count 884 arch/sparc/kernel/perf_event.c return new_raw_count; new_raw_count 154 arch/x86/events/amd/ibs.c perf_event_try_update(struct perf_event *event, u64 new_raw_count, int width) new_raw_count 170 arch/x86/events/amd/ibs.c new_raw_count) != prev_raw_count) new_raw_count 181 arch/x86/events/amd/ibs.c delta = (new_raw_count << shift) - (prev_raw_count << shift); new_raw_count 72 arch/x86/events/core.c u64 prev_raw_count, new_raw_count; new_raw_count 88 arch/x86/events/core.c rdpmcl(hwc->event_base_rdpmc, new_raw_count); new_raw_count 91 arch/x86/events/core.c new_raw_count) != prev_raw_count) new_raw_count 102 arch/x86/events/core.c delta = (new_raw_count << shift) - (prev_raw_count << shift); new_raw_count 108 arch/x86/events/core.c return new_raw_count; new_raw_count 361 arch/x86/events/intel/cstate.c u64 prev_raw_count, new_raw_count; new_raw_count 365 arch/x86/events/intel/cstate.c new_raw_count = cstate_pmu_read_counter(event); new_raw_count 368 arch/x86/events/intel/cstate.c new_raw_count) != prev_raw_count) new_raw_count 371 arch/x86/events/intel/cstate.c local64_add(new_raw_count - prev_raw_count, &event->count); new_raw_count 1671 arch/x86/events/intel/ds.c u64 prev_raw_count, new_raw_count; new_raw_count 1682 arch/x86/events/intel/ds.c rdpmcl(hwc->event_base_rdpmc, new_raw_count); new_raw_count 1683 arch/x86/events/intel/ds.c local64_set(&hwc->prev_count, new_raw_count); new_raw_count 1712 arch/x86/events/intel/ds.c new = ((s64)(new_raw_count << shift) >> shift); new_raw_count 179 arch/x86/events/intel/rapl.c u64 prev_raw_count, new_raw_count; new_raw_count 185 arch/x86/events/intel/rapl.c rdmsrl(event->hw.event_base, new_raw_count); new_raw_count 188 arch/x86/events/intel/rapl.c new_raw_count) != prev_raw_count) { new_raw_count 201 arch/x86/events/intel/rapl.c delta = (new_raw_count << shift) - (prev_raw_count << shift); new_raw_count 208 arch/x86/events/intel/rapl.c return new_raw_count; new_raw_count 139 arch/xtensa/kernel/perf_event.c uint64_t prev_raw_count, new_raw_count; new_raw_count 144 arch/xtensa/kernel/perf_event.c new_raw_count = xtensa_pmu_read_counter(event->hw.idx); new_raw_count 146 arch/xtensa/kernel/perf_event.c new_raw_count) != prev_raw_count); new_raw_count 148 arch/xtensa/kernel/perf_event.c delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; new_raw_count 986 drivers/perf/arm-cci.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 990 drivers/perf/arm-cci.c new_raw_count = pmu_read_counter(event); new_raw_count 992 drivers/perf/arm-cci.c new_raw_count) != prev_raw_count); new_raw_count 994 drivers/perf/arm-cci.c delta = (new_raw_count - prev_raw_count) & CCI_PMU_CNTR_MASK; new_raw_count 998 drivers/perf/arm-cci.c return new_raw_count; new_raw_count 166 drivers/perf/arm_pmu.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 171 drivers/perf/arm_pmu.c new_raw_count = armpmu->read_counter(event); new_raw_count 174 drivers/perf/arm_pmu.c new_raw_count) != prev_raw_count) new_raw_count 177 drivers/perf/arm_pmu.c delta = (new_raw_count - prev_raw_count) & max_period; new_raw_count 182 drivers/perf/arm_pmu.c return new_raw_count; new_raw_count 308 drivers/perf/fsl_imx8_ddr_perf.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 313 drivers/perf/fsl_imx8_ddr_perf.c new_raw_count = ddr_perf_read_counter(pmu, counter); new_raw_count 315 drivers/perf/fsl_imx8_ddr_perf.c new_raw_count) != prev_raw_count); new_raw_count 317 drivers/perf/fsl_imx8_ddr_perf.c delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; new_raw_count 227 drivers/perf/hisilicon/hisi_uncore_pmu.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 231 drivers/perf/hisilicon/hisi_uncore_pmu.c new_raw_count = hisi_pmu->ops->read_counter(hisi_pmu, hwc); new_raw_count 234 drivers/perf/hisilicon/hisi_uncore_pmu.c new_raw_count) != prev_raw_count); new_raw_count 238 drivers/perf/hisilicon/hisi_uncore_pmu.c delta = (new_raw_count - prev_raw_count) & new_raw_count 995 drivers/perf/xgene_pmu.c u64 delta, prev_raw_count, new_raw_count; new_raw_count 999 drivers/perf/xgene_pmu.c new_raw_count = xgene_pmu->ops->read_counter(pmu_dev, GET_CNTR(event)); new_raw_count 1002 drivers/perf/xgene_pmu.c new_raw_count) != prev_raw_count) new_raw_count 1005 drivers/perf/xgene_pmu.c delta = (new_raw_count - prev_raw_count) & pmu_dev->max_period;