prev_count 283 arch/alpha/kernel/perf_event.c local64_set(&hwc->prev_count, (unsigned long)(-left)); prev_count 314 arch/alpha/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 317 arch/alpha/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 121 arch/arc/kernel/perf_event.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_count 129 arch/arc/kernel/perf_event.c local64_set(&hwc->prev_count, new_raw_count); prev_count 270 arch/arc/kernel/perf_event.c local64_set(&hwc->prev_count, value); prev_count 388 arch/arc/kernel/perf_event.c local64_set(&hwc->prev_count, 0); prev_count 311 arch/arm/mach-imx/mmdc.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 314 arch/arm/mach-imx/mmdc.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 339 arch/arm/mach-imx/mmdc.c local64_set(&hwc->prev_count, 0); prev_count 374 arch/arm/mach-imx/mmdc.c local64_set(&hwc->prev_count, mmdc_pmu_read_counter(pmu_mmdc, cfg)); prev_count 120 arch/arm/mm/cache-l2x0-pmu.c u64 prev_count, new_count, mask; prev_count 123 arch/arm/mm/cache-l2x0-pmu.c prev_count = local64_read(&hw->prev_count); prev_count 125 arch/arm/mm/cache-l2x0-pmu.c } while (local64_xchg(&hw->prev_count, new_count) != prev_count); prev_count 128 arch/arm/mm/cache-l2x0-pmu.c local64_add((new_count - prev_count) & mask, &event->count); prev_count 147 arch/arm/mm/cache-l2x0-pmu.c local64_set(&hw->prev_count, 0); prev_count 906 arch/csky/kernel/perf_event.c local64_set(&hwc->prev_count, (u64)(-left)); prev_count 922 arch/csky/kernel/perf_event.c uint64_t prev_raw_count = local64_read(&hwc->prev_count); prev_count 935 arch/csky/kernel/perf_event.c local64_set(&hwc->prev_count, new_raw_count); prev_count 397 arch/mips/kernel/perf_event_mipsxx.c local64_set(&hwc->prev_count, mipspmu.overflow - left); prev_count 414 arch/mips/kernel/perf_event_mipsxx.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 417 arch/mips/kernel/perf_event_mipsxx.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 216 arch/nds32/kernel/perf_event_cpu.c local64_set(&hwc->prev_count, (u64)(-left)); prev_count 950 arch/nds32/kernel/perf_event_cpu.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 953 arch/nds32/kernel/perf_event_cpu.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 119 arch/powerpc/perf/8xx-pmu.c local64_set(&event->hw.prev_count, val); prev_count 132 arch/powerpc/perf/8xx-pmu.c prev = local64_read(&event->hw.prev_count); prev_count 153 arch/powerpc/perf/8xx-pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev_count 1065 arch/powerpc/perf/core-book3s.c local64_set(&event->hw.prev_count, val); prev_count 1075 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev_count 1081 arch/powerpc/perf/core-book3s.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev_count 1125 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev_count 1144 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev_count 1146 arch/powerpc/perf/core-book3s.c local64_set(&event->hw.prev_count, val); prev_count 1391 arch/powerpc/perf/core-book3s.c val = local64_read(&event->hw.prev_count); prev_count 1399 arch/powerpc/perf/core-book3s.c local64_set(&event->hw.prev_count, val); prev_count 1979 arch/powerpc/perf/core-book3s.c local64_set(&event->hw.prev_count, 0); prev_count 2051 arch/powerpc/perf/core-book3s.c prev = local64_read(&event->hw.prev_count); prev_count 2076 arch/powerpc/perf/core-book3s.c local64_set(&event->hw.prev_count, val); prev_count 188 arch/powerpc/perf/core-fsl-emb.c prev = local64_read(&event->hw.prev_count); prev_count 191 arch/powerpc/perf/core-fsl-emb.c } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); prev_count 327 arch/powerpc/perf/core-fsl-emb.c local64_set(&event->hw.prev_count, val); prev_count 615 arch/powerpc/perf/core-fsl-emb.c prev = local64_read(&event->hw.prev_count); prev_count 638 arch/powerpc/perf/core-fsl-emb.c local64_set(&event->hw.prev_count, val); prev_count 1341 arch/powerpc/perf/hv-24x7.c (void)local64_xchg(&event->hw.prev_count, ct); prev_count 1361 arch/powerpc/perf/hv-24x7.c prev = local64_xchg(&event->hw.prev_count, now); prev_count 1425 arch/powerpc/perf/hv-24x7.c local64_set(&event->hw.prev_count, h_24x7_get_value(event)); prev_count 194 arch/powerpc/perf/hv-gpci.c prev = local64_xchg(&event->hw.prev_count, now); prev_count 200 arch/powerpc/perf/hv-gpci.c local64_set(&event->hw.prev_count, h_gpci_get_value(event)); prev_count 974 arch/powerpc/perf/imc-pmu.c local64_set(&event->hw.prev_count, data); prev_count 983 arch/powerpc/perf/imc-pmu.c counter_prev = local64_read(&event->hw.prev_count); prev_count 223 arch/riscv/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 226 arch/riscv/kernel/perf_event.c oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 291 arch/riscv/kernel/perf_event.c local64_set(&hwc->prev_count, read_counter(hwc->idx)); prev_count 320 arch/s390/kernel/perf_cpum_cf.c prev = local64_read(&event->hw.prev_count); prev_count 332 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev_count 343 arch/s390/kernel/perf_cpum_cf.c prev = local64_read(&event->hw.prev_count); prev_count 347 arch/s390/kernel/perf_cpum_cf.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev_count 196 arch/sh/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 199 arch/sh/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 871 arch/sparc/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 874 arch/sparc/kernel/perf_event.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 914 arch/sparc/kernel/perf_event.c local64_set(&hwc->prev_count, (u64)-left); prev_count 168 arch/x86/events/amd/ibs.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 169 arch/x86/events/amd/ibs.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 323 arch/x86/events/amd/ibs.c local64_set(&hwc->prev_count, 0); prev_count 289 arch/x86/events/amd/iommu.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_count 314 arch/x86/events/amd/iommu.c prev = local64_read(&hwc->prev_count); prev_count 315 arch/x86/events/amd/iommu.c if (local64_cmpxchg(&hwc->prev_count, prev, count) != prev) prev_count 93 arch/x86/events/amd/uncore.c prev = local64_read(&hwc->prev_count); prev_count 95 arch/x86/events/amd/uncore.c local64_set(&hwc->prev_count, new); prev_count 106 arch/x86/events/amd/uncore.c wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count)); prev_count 87 arch/x86/events/core.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 90 arch/x86/events/core.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 1235 arch/x86/events/core.c local64_set(&hwc->prev_count, (u64)-left); prev_count 2287 arch/x86/events/intel/core.c local64_set(&event->hw.prev_count, 0); prev_count 364 arch/x86/events/intel/cstate.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 367 arch/x86/events/intel/cstate.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 376 arch/x86/events/intel/cstate.c local64_set(&event->hw.prev_count, cstate_pmu_read_counter(event)); prev_count 1681 arch/x86/events/intel/ds.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 1683 arch/x86/events/intel/ds.c local64_set(&hwc->prev_count, new_raw_count); prev_count 184 arch/x86/events/intel/rapl.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 187 arch/x86/events/intel/rapl.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 256 arch/x86/events/intel/rapl.c local64_set(&event->hw.prev_count, rapl_read_counter(event)); prev_count 235 arch/x86/events/intel/uncore.c u64 prev_count, new_count, delta; prev_count 247 arch/x86/events/intel/uncore.c prev_count = local64_read(&event->hw.prev_count); prev_count 249 arch/x86/events/intel/uncore.c if (local64_xchg(&event->hw.prev_count, new_count) != prev_count) prev_count 252 arch/x86/events/intel/uncore.c delta = (new_count << shift) - (prev_count << shift); prev_count 487 arch/x86/events/intel/uncore.c local64_set(&event->hw.prev_count, prev_count 502 arch/x86/events/intel/uncore.c local64_set(&event->hw.prev_count, uncore_read_counter(box, event)); prev_count 234 arch/x86/events/msr.c prev = local64_read(&event->hw.prev_count); prev_count 237 arch/x86/events/msr.c if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) prev_count 257 arch/x86/events/msr.c local64_set(&event->hw.prev_count, now); prev_count 143 arch/xtensa/kernel/perf_event.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 145 arch/xtensa/kernel/perf_event.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 181 arch/xtensa/kernel/perf_event.c local64_set(&hwc->prev_count, -left); prev_count 100 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c prev = local64_read(&hwc->prev_count); prev_count 111 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, count) != prev); prev_count 580 drivers/gpu/drm/i915/i915_pmu.c prev = local64_read(&hwc->prev_count); prev_count 583 drivers/gpu/drm/i915/i915_pmu.c if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev) prev_count 645 drivers/gpu/drm/i915/i915_pmu.c local64_set(&event->hw.prev_count, __i915_pmu_event_read(event)); prev_count 906 drivers/perf/arm-cci.c pmu_write_counter(cci_pmu, local64_read(&event->hw.prev_count), i); prev_count 971 drivers/perf/arm-cci.c pmu_write_counter(cci_pmu, local64_read(&event->hw.prev_count), i); prev_count 989 drivers/perf/arm-cci.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 991 drivers/perf/arm-cci.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 1016 drivers/perf/arm-cci.c local64_set(&hwc->prev_count, val); prev_count 877 drivers/perf/arm-ccn.c u64 prev_count, new_count, mask; prev_count 880 drivers/perf/arm-ccn.c prev_count = local64_read(&hw->prev_count); prev_count 882 drivers/perf/arm-ccn.c } while (local64_xchg(&hw->prev_count, new_count) != prev_count); prev_count 886 drivers/perf/arm-ccn.c local64_add((new_count - prev_count) & mask, &event->count); prev_count 927 drivers/perf/arm-ccn.c local64_set(&event->hw.prev_count, prev_count 352 drivers/perf/arm_dsu_pmu.c u64 delta, prev_count, new_count; prev_count 356 drivers/perf/arm_dsu_pmu.c prev_count = local64_read(&hwc->prev_count); prev_count 358 drivers/perf/arm_dsu_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != prev_count 359 drivers/perf/arm_dsu_pmu.c prev_count); prev_count 360 drivers/perf/arm_dsu_pmu.c delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); prev_count 386 drivers/perf/arm_dsu_pmu.c local64_set(&event->hw.prev_count, val); prev_count 153 drivers/perf/arm_pmu.c local64_set(&hwc->prev_count, (u64)-left); prev_count 170 drivers/perf/arm_pmu.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 173 drivers/perf/arm_pmu.c if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 210 drivers/perf/arm_smmuv3_pmu.c prev = local64_read(&hwc->prev_count); prev_count 212 drivers/perf/arm_smmuv3_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); prev_count 247 drivers/perf/arm_smmuv3_pmu.c local64_set(&hwc->prev_count, new); prev_count 443 drivers/perf/arm_smmuv3_pmu.c local64_set(&hwc->prev_count, 0); prev_count 312 drivers/perf/fsl_imx8_ddr_perf.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 314 drivers/perf/fsl_imx8_ddr_perf.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 352 drivers/perf/fsl_imx8_ddr_perf.c local64_set(&hwc->prev_count, 0); prev_count 218 drivers/perf/hisilicon/hisi_uncore_pmu.c local64_set(&hwc->prev_count, val); prev_count 232 drivers/perf/hisilicon/hisi_uncore_pmu.c prev_raw_count = local64_read(&hwc->prev_count); prev_count 233 drivers/perf/hisilicon/hisi_uncore_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, prev_count 256 drivers/perf/hisilicon/hisi_uncore_pmu.c u64 prev_raw_count = local64_read(&hwc->prev_count); prev_count 346 drivers/perf/qcom_l2_pmu.c prev = local64_read(&hwc->prev_count); prev_count 348 drivers/perf/qcom_l2_pmu.c } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); prev_count 377 drivers/perf/qcom_l2_pmu.c local64_set(&hwc->prev_count, new); prev_count 646 drivers/perf/qcom_l2_pmu.c local64_set(&hwc->prev_count, 0); prev_count 206 drivers/perf/qcom_l3_pmu.c local64_set(&event->hw.prev_count, 0); prev_count 247 drivers/perf/qcom_l3_pmu.c prev = local64_read(&event->hw.prev_count); prev_count 253 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev_count 285 drivers/perf/qcom_l3_pmu.c local64_set(&event->hw.prev_count, 0); prev_count 323 drivers/perf/qcom_l3_pmu.c prev = local64_read(&event->hw.prev_count); prev_count 325 drivers/perf/qcom_l3_pmu.c } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); prev_count 276 drivers/perf/thunderx2_pmu.c local64_set(&hwc->prev_count, 0); prev_count 299 drivers/perf/thunderx2_pmu.c local64_set(&hwc->prev_count, 0); prev_count 328 drivers/perf/thunderx2_pmu.c prev = local64_xchg(&hwc->prev_count, new); prev_count 986 drivers/perf/xgene_pmu.c local64_set(&hw->prev_count, val); prev_count 998 drivers/perf/xgene_pmu.c prev_raw_count = local64_read(&hw->prev_count); prev_count 1001 drivers/perf/xgene_pmu.c if (local64_cmpxchg(&hw->prev_count, prev_raw_count, prev_count 1030 drivers/perf/xgene_pmu.c u64 prev_raw_count = local64_read(&hw->prev_count); prev_count 195 include/linux/perf_event.h local64_t prev_count; prev_count 5413 kernel/events/core.c userpg->offset -= local64_read(&event->hw.prev_count); prev_count 9675 kernel/events/core.c prev = local64_xchg(&event->hw.prev_count, now); prev_count 9681 kernel/events/core.c local64_set(&event->hw.prev_count, local_clock()); prev_count 9751 kernel/events/core.c prev = local64_xchg(&event->hw.prev_count, now); prev_count 9758 kernel/events/core.c local64_set(&event->hw.prev_count, event->ctx->time); prev_count 282 kernel/softirq.c int prev_count; prev_count 287 kernel/softirq.c prev_count = preempt_count(); prev_count 294 kernel/softirq.c if (unlikely(prev_count != preempt_count())) { prev_count 297 kernel/softirq.c prev_count, preempt_count()); prev_count 298 kernel/softirq.c preempt_count_set(prev_count); prev_count 192 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/locks.h unsigned int prev_count = __sync_fetch_and_sub(&c->count, 1); prev_count 194 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/locks.h assume(prev_count); prev_count 199 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/locks.h unsigned int prev_count = __sync_fetch_and_add(&c->count, 1); prev_count 201 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/locks.h BUG_ON(prev_count == UINT_MAX);