/linux-4.4.14/arch/x86/kernel/cpu/ |
D | perf_event_msr.c | 159 prev = local64_read(&event->hw.prev_count); in msr_event_update() 162 if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) in msr_event_update() 177 local64_set(&event->hw.prev_count, now); in msr_event_start()
|
D | perf_event_intel_cstate.c | 439 prev_raw_count = local64_read(&hwc->prev_count); in cstate_pmu_event_update() 442 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in cstate_pmu_event_update() 451 local64_set(&event->hw.prev_count, cstate_pmu_read_counter(event)); in cstate_pmu_event_start()
|
D | perf_event_amd_iommu.c | 304 u64 prev_raw_count = local64_read(&hwc->prev_count); in perf_iommu_start() 330 prev_raw_count = local64_read(&hwc->prev_count); in perf_iommu_read() 331 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in perf_iommu_read()
|
D | perf_event_intel_uncore.c | 229 u64 prev_count, new_count, delta; in uncore_perf_event_update() local 239 prev_count = local64_read(&event->hw.prev_count); in uncore_perf_event_update() 241 if (local64_xchg(&event->hw.prev_count, new_count) != prev_count) in uncore_perf_event_update() 244 delta = (new_count << shift) - (prev_count << shift); in uncore_perf_event_update() 476 local64_set(&event->hw.prev_count, uncore_read_counter(box, event)); in uncore_pmu_event_start()
|
D | perf_event_amd_uncore.c | 85 prev = local64_read(&hwc->prev_count); in amd_uncore_read() 87 local64_set(&hwc->prev_count, new); in amd_uncore_read() 98 wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_start()
|
D | perf_event_intel_rapl.c | 176 prev_raw_count = local64_read(&hwc->prev_count); in rapl_event_update() 179 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in rapl_event_update() 254 local64_set(&event->hw.prev_count, rapl_read_counter(event)); in __rapl_pmu_event_start()
|
D | perf_event_amd_ibs.c | 130 prev_raw_count = local64_read(&hwc->prev_count); in perf_event_try_update() 131 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in perf_event_try_update() 297 local64_set(&hwc->prev_count, 0); in perf_ibs_set_period()
|
D | perf_event.c | 83 prev_raw_count = local64_read(&hwc->prev_count); in x86_perf_event_update() 86 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in x86_perf_event_update() 1130 local64_read(&hwc->prev_count) != (u64)-left) { in x86_perf_event_set_period() 1135 local64_set(&hwc->prev_count, (u64)-left); in x86_perf_event_set_period()
|
D | perf_event_intel_uncore_snb.c | 353 local64_set(&event->hw.prev_count, count); in snb_uncore_imc_event_start()
|
D | perf_event_intel.c | 1747 local64_set(&event->hw.prev_count, 0); in intel_pmu_save_and_restart()
|
/linux-4.4.14/kernel/ |
D | softirq.c | 263 int prev_count; in __do_softirq() local 268 prev_count = preempt_count(); in __do_softirq() 275 if (unlikely(prev_count != preempt_count())) { in __do_softirq() 278 prev_count, preempt_count()); in __do_softirq() 279 preempt_count_set(prev_count); in __do_softirq()
|
/linux-4.4.14/arch/powerpc/perf/ |
D | core-fsl-emb.c | 192 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read() 195 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read() 331 local64_set(&event->hw.prev_count, val); in fsl_emb_pmu_add() 619 prev = local64_read(&event->hw.prev_count); in record_and_restart() 642 local64_set(&event->hw.prev_count, val); in record_and_restart()
|
D | core-book3s.c | 1017 local64_set(&event->hw.prev_count, val); in power_pmu_read() 1027 prev = local64_read(&event->hw.prev_count); in power_pmu_read() 1033 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in power_pmu_read() 1077 prev = local64_read(&event->hw.prev_count); in freeze_limited_counters() 1096 prev = local64_read(&event->hw.prev_count); in thaw_limited_counters() 1098 local64_set(&event->hw.prev_count, val); in thaw_limited_counters() 1328 val = local64_read(&event->hw.prev_count); in power_pmu_enable() 1336 local64_set(&event->hw.prev_count, val); in power_pmu_enable() 1893 local64_set(&event->hw.prev_count, 0); in power_pmu_event_init() 1965 prev = local64_read(&event->hw.prev_count); in record_and_restart() [all …]
|
D | hv-gpci.c | 191 prev = local64_xchg(&event->hw.prev_count, now); in h_gpci_event_update() 197 local64_set(&event->hw.prev_count, h_gpci_get_value(event)); in h_gpci_event_start()
|
D | hv-24x7.c | 1236 prev = local64_xchg(&event->hw.prev_count, now); in update_event_count() 1290 local64_set(&event->hw.prev_count, h_24x7_get_value(event)); in h_24x7_event_start()
|
/linux-4.4.14/arch/arc/kernel/ |
D | perf_event.c | 106 uint64_t prev_raw_count = local64_read(&hwc->prev_count); in arc_perf_event_update() 114 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update() 241 local64_set(&hwc->prev_count, value); in arc_pmu_event_set_period() 361 local64_set(&hwc->prev_count, 0); in arc_pmu_add()
|
/linux-4.4.14/arch/s390/kernel/ |
D | perf_cpum_cf.c | 438 prev = local64_read(&event->hw.prev_count); in hw_perf_event_reset() 450 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in hw_perf_event_reset() 461 prev = local64_read(&event->hw.prev_count); in hw_perf_event_update() 465 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in hw_perf_event_update()
|
/linux-4.4.14/arch/metag/kernel/perf/ |
D | perf_event.c | 205 prev_raw_count = local64_read(&hwc->prev_count); in metag_pmu_event_update() 208 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in metag_pmu_event_update() 250 local64_set(&hwc->prev_count, -(s32)left); in metag_pmu_event_set_period() 615 local64_set(&event->prev_count, __core_reg_get(TXTACTCYC)); in metag_pmu_enable_counter() 664 local64_set(&event->prev_count, 0); in metag_pmu_enable_counter()
|
/linux-4.4.14/arch/xtensa/kernel/ |
D | perf_event.c | 146 prev_raw_count = local64_read(&hwc->prev_count); in xtensa_perf_event_update() 148 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in xtensa_perf_event_update() 184 local64_set(&hwc->prev_count, -left); in xtensa_perf_event_set_period()
|
/linux-4.4.14/arch/sh/kernel/ |
D | perf_event.c | 199 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update() 202 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update()
|
/linux-4.4.14/arch/alpha/kernel/ |
D | perf_event.c | 282 local64_set(&hwc->prev_count, (unsigned long)(-left)); in alpha_perf_event_set_period() 313 prev_raw_count = local64_read(&hwc->prev_count); in alpha_perf_event_update() 316 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in alpha_perf_event_update()
|
/linux-4.4.14/arch/blackfin/kernel/ |
D | perf_event.c | 280 prev_raw_count = local64_read(&hwc->prev_count); in bfin_perf_event_update() 283 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in bfin_perf_event_update()
|
/linux-4.4.14/drivers/bus/ |
D | arm-ccn.c | 868 u64 prev_count, new_count, mask; in arm_ccn_pmu_event_update() local 871 prev_count = local64_read(&hw->prev_count); in arm_ccn_pmu_event_update() 873 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in arm_ccn_pmu_event_update() 877 local64_add((new_count - prev_count) & mask, &event->count); in arm_ccn_pmu_event_update() 914 local64_set(&event->hw.prev_count, in arm_ccn_pmu_event_start()
|
D | arm-cci.c | 775 prev_raw_count = local64_read(&hwc->prev_count); in pmu_event_update() 777 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in pmu_event_update() 802 local64_set(&hwc->prev_count, val); in pmu_event_set_period()
|
/linux-4.4.14/drivers/perf/ |
D | arm_pmu.c | 134 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 150 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() 153 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in armpmu_event_update()
|
/linux-4.4.14/arch/tile/kernel/ |
D | perf_event.c | 519 prev_raw_count = local64_read(&hwc->prev_count); in tile_perf_event_update() 522 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in tile_perf_event_update() 579 local64_set(&hwc->prev_count, (u64)-left); in tile_event_set_period()
|
/linux-4.4.14/drivers/net/irda/ |
D | stir4200.c | 595 unsigned long prev_count = 0x1fff; in fifo_txwait() local 598 for (;; prev_count = count) { in fifo_txwait() 632 if (prev_count < count) in fifo_txwait()
|
/linux-4.4.14/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 402 local64_set(&hwc->prev_count, mipspmu.overflow - left); in mipspmu_event_set_period() 419 prev_raw_count = local64_read(&hwc->prev_count); in mipspmu_event_update() 422 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mipspmu_event_update()
|
/linux-4.4.14/arch/sparc/kernel/ |
D | perf_event.c | 869 prev_raw_count = local64_read(&hwc->prev_count); in sparc_perf_event_update() 872 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sparc_perf_event_update() 908 local64_set(&hwc->prev_count, (u64)-left); in sparc_perf_event_set_period()
|
/linux-4.4.14/include/linux/ |
D | perf_event.h | 162 local64_t prev_count; member
|
/linux-4.4.14/kernel/events/ |
D | core.c | 4469 userpg->offset -= local64_read(&event->hw.prev_count); in perf_event_update_userpage() 7248 prev = local64_xchg(&event->hw.prev_count, now); in cpu_clock_event_update() 7254 local64_set(&event->hw.prev_count, local_clock()); in cpu_clock_event_start() 7324 prev = local64_xchg(&event->hw.prev_count, now); in task_clock_event_update() 7331 local64_set(&event->hw.prev_count, event->ctx->time); in task_clock_event_start()
|