Lines Matching refs:left
1092 s64 left = local64_read(&hwc->period_left); in x86_perf_event_set_period() local
1102 if (unlikely(left <= -period)) { in x86_perf_event_set_period()
1103 left = period; in x86_perf_event_set_period()
1104 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1109 if (unlikely(left <= 0)) { in x86_perf_event_set_period()
1110 left += period; in x86_perf_event_set_period()
1111 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1118 if (unlikely(left < 2)) in x86_perf_event_set_period()
1119 left = 2; in x86_perf_event_set_period()
1121 if (left > x86_pmu.max_period) in x86_perf_event_set_period()
1122 left = x86_pmu.max_period; in x86_perf_event_set_period()
1125 left = x86_pmu.limit_period(event, left); in x86_perf_event_set_period()
1127 per_cpu(pmc_prev_left[idx], smp_processor_id()) = left; in x86_perf_event_set_period()
1130 local64_read(&hwc->prev_count) != (u64)-left) { in x86_perf_event_set_period()
1135 local64_set(&hwc->prev_count, (u64)-left); in x86_perf_event_set_period()
1137 wrmsrl(hwc->event_base, (u64)(-left) & x86_pmu.cntval_mask); in x86_perf_event_set_period()
1147 (u64)(-left) & x86_pmu.cntval_mask); in x86_perf_event_set_period()