/linux-4.1.27/kernel/ |
D | softirq.c | 263 int prev_count; in __do_softirq() local 268 prev_count = preempt_count(); in __do_softirq() 275 if (unlikely(prev_count != preempt_count())) { in __do_softirq() 278 prev_count, preempt_count()); in __do_softirq() 279 preempt_count_set(prev_count); in __do_softirq()
|
/linux-4.1.27/arch/powerpc/perf/ |
D | core-fsl-emb.c | 192 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read() 195 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read() 331 local64_set(&event->hw.prev_count, val); in fsl_emb_pmu_add() 619 prev = local64_read(&event->hw.prev_count); in record_and_restart() 642 local64_set(&event->hw.prev_count, val); in record_and_restart()
|
D | core-book3s.c | 1017 local64_set(&event->hw.prev_count, val); in power_pmu_read() 1027 prev = local64_read(&event->hw.prev_count); in power_pmu_read() 1033 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in power_pmu_read() 1077 prev = local64_read(&event->hw.prev_count); in freeze_limited_counters() 1096 prev = local64_read(&event->hw.prev_count); in thaw_limited_counters() 1098 local64_set(&event->hw.prev_count, val); in thaw_limited_counters() 1328 val = local64_read(&event->hw.prev_count); in power_pmu_enable() 1336 local64_set(&event->hw.prev_count, val); in power_pmu_enable() 1869 local64_set(&event->hw.prev_count, 0); in power_pmu_event_init() 1941 prev = local64_read(&event->hw.prev_count); in record_and_restart() [all …]
|
D | hv-gpci.c | 191 prev = local64_xchg(&event->hw.prev_count, now); in h_gpci_event_update() 197 local64_set(&event->hw.prev_count, h_gpci_get_value(event)); in h_gpci_event_start()
|
D | hv-24x7.c | 1227 prev = local64_xchg(&event->hw.prev_count, now); in update_event_count() 1242 local64_set(&event->hw.prev_count, h_24x7_get_value(event)); in h_24x7_event_start()
|
/linux-4.1.27/arch/arc/kernel/ |
D | perf_event.c | 95 prev_raw_count = local64_read(&hwc->prev_count); in arc_perf_event_update() 97 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in arc_perf_event_update() 256 local64_set(&hwc->prev_count, 0); in arc_pmu_add()
|
/linux-4.1.27/arch/s390/kernel/ |
D | perf_cpum_cf.c | 432 prev = local64_read(&event->hw.prev_count); in hw_perf_event_reset() 444 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in hw_perf_event_reset() 455 prev = local64_read(&event->hw.prev_count); in hw_perf_event_update() 459 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in hw_perf_event_update()
|
/linux-4.1.27/arch/metag/kernel/perf/ |
D | perf_event.c | 205 prev_raw_count = local64_read(&hwc->prev_count); in metag_pmu_event_update() 208 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in metag_pmu_event_update() 250 local64_set(&hwc->prev_count, -(s32)left); in metag_pmu_event_set_period() 615 local64_set(&event->prev_count, __core_reg_get(TXTACTCYC)); in metag_pmu_enable_counter() 664 local64_set(&event->prev_count, 0); in metag_pmu_enable_counter()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | perf_event_intel_uncore.c | 175 u64 prev_count, new_count, delta; in uncore_perf_event_update() local 185 prev_count = local64_read(&event->hw.prev_count); in uncore_perf_event_update() 187 if (local64_xchg(&event->hw.prev_count, new_count) != prev_count) in uncore_perf_event_update() 190 delta = (new_count << shift) - (prev_count << shift); in uncore_perf_event_update() 423 local64_set(&event->hw.prev_count, uncore_read_counter(box, event)); in uncore_pmu_event_start()
|
D | perf_event_amd_iommu.c | 304 u64 prev_raw_count = local64_read(&hwc->prev_count); in perf_iommu_start() 330 prev_raw_count = local64_read(&hwc->prev_count); in perf_iommu_read() 331 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in perf_iommu_read()
|
D | perf_event_amd_uncore.c | 85 prev = local64_read(&hwc->prev_count); in amd_uncore_read() 87 local64_set(&hwc->prev_count, new); in amd_uncore_read() 98 wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_start()
|
D | perf_event_intel_rapl.c | 178 prev_raw_count = local64_read(&hwc->prev_count); in rapl_event_update() 181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in rapl_event_update() 257 local64_set(&event->hw.prev_count, rapl_read_counter(event)); in __rapl_pmu_event_start()
|
D | perf_event_amd_ibs.c | 130 prev_raw_count = local64_read(&hwc->prev_count); in perf_event_try_update() 131 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in perf_event_try_update() 297 local64_set(&hwc->prev_count, 0); in perf_ibs_set_period()
|
D | perf_event_intel_uncore_snb.c | 333 local64_set(&event->hw.prev_count, count); in snb_uncore_imc_event_start()
|
D | perf_event.c | 83 prev_raw_count = local64_read(&hwc->prev_count); in x86_perf_event_update() 86 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in x86_perf_event_update() 1118 local64_set(&hwc->prev_count, (u64)-left); in x86_perf_event_set_period()
|
D | perf_event_intel.c | 1525 local64_set(&event->hw.prev_count, 0); in intel_pmu_save_and_restart()
|
/linux-4.1.27/arch/arm/kernel/ |
D | perf_event.c | 128 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 144 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() 147 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in armpmu_event_update()
|
/linux-4.1.27/arch/sh/kernel/ |
D | perf_event.c | 199 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update() 202 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | perf_event.c | 282 local64_set(&hwc->prev_count, (unsigned long)(-left)); in alpha_perf_event_set_period() 313 prev_raw_count = local64_read(&hwc->prev_count); in alpha_perf_event_update() 316 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in alpha_perf_event_update()
|
/linux-4.1.27/drivers/bus/ |
D | arm-ccn.c | 788 u64 prev_count, new_count, mask; in arm_ccn_pmu_event_update() local 791 prev_count = local64_read(&hw->prev_count); in arm_ccn_pmu_event_update() 793 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in arm_ccn_pmu_event_update() 797 local64_add((new_count - prev_count) & mask, &event->count); in arm_ccn_pmu_event_update() 834 local64_set(&event->hw.prev_count, in arm_ccn_pmu_event_start()
|
D | arm-cci.c | 395 prev_raw_count = local64_read(&hwc->prev_count); in pmu_event_update() 397 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in pmu_event_update() 422 local64_set(&hwc->prev_count, val); in pmu_event_set_period()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | perf_event.c | 280 prev_raw_count = local64_read(&hwc->prev_count); in bfin_perf_event_update() 283 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in bfin_perf_event_update()
|
/linux-4.1.27/arch/tile/kernel/ |
D | perf_event.c | 519 prev_raw_count = local64_read(&hwc->prev_count); in tile_perf_event_update() 522 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in tile_perf_event_update() 579 local64_set(&hwc->prev_count, (u64)-left); in tile_event_set_period()
|
/linux-4.1.27/drivers/net/irda/ |
D | stir4200.c | 595 unsigned long prev_count = 0x1fff; in fifo_txwait() local 598 for (;; prev_count = count) { in fifo_txwait() 632 if (prev_count < count) in fifo_txwait()
|
/linux-4.1.27/arch/arm64/kernel/ |
D | perf_event.c | 183 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 201 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() 204 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in armpmu_event_update()
|
/linux-4.1.27/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 402 local64_set(&hwc->prev_count, mipspmu.overflow - left); in mipspmu_event_set_period() 419 prev_raw_count = local64_read(&hwc->prev_count); in mipspmu_event_update() 422 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mipspmu_event_update()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | perf_event.c | 869 prev_raw_count = local64_read(&hwc->prev_count); in sparc_perf_event_update() 872 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sparc_perf_event_update() 908 local64_set(&hwc->prev_count, (u64)-left); in sparc_perf_event_set_period()
|
/linux-4.1.27/include/linux/ |
D | perf_event.h | 145 local64_t prev_count; member
|
/linux-4.1.27/kernel/events/ |
D | core.c | 4303 userpg->offset -= local64_read(&event->hw.prev_count); in perf_event_update_userpage() 6954 prev = local64_xchg(&event->hw.prev_count, now); in cpu_clock_event_update() 6960 local64_set(&event->hw.prev_count, local_clock()); in cpu_clock_event_start() 7030 prev = local64_xchg(&event->hw.prev_count, now); in task_clock_event_update() 7037 local64_set(&event->hw.prev_count, event->ctx->time); in task_clock_event_start()
|