alpha_pmu          95 arch/alpha/kernel/perf_event.c static const struct alpha_pmu_t *alpha_pmu;
alpha_pmu         235 arch/alpha/kernel/perf_event.c 	val &= alpha_pmu->pmc_count_mask[idx];
alpha_pmu         236 arch/alpha/kernel/perf_event.c 	val <<= alpha_pmu->pmc_count_shift[idx];
alpha_pmu         246 arch/alpha/kernel/perf_event.c 	val >>= alpha_pmu->pmc_count_shift[idx];
alpha_pmu         247 arch/alpha/kernel/perf_event.c 	val &= alpha_pmu->pmc_count_mask[idx];
alpha_pmu         277 arch/alpha/kernel/perf_event.c 	if (unlikely(left < alpha_pmu->pmc_left[idx]))
alpha_pmu         278 arch/alpha/kernel/perf_event.c 		left = alpha_pmu->pmc_left[idx];
alpha_pmu         280 arch/alpha/kernel/perf_event.c 	if (left > (long)alpha_pmu->pmc_max_period[idx])
alpha_pmu         281 arch/alpha/kernel/perf_event.c 		left = alpha_pmu->pmc_max_period[idx];
alpha_pmu         321 arch/alpha/kernel/perf_event.c 	delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf;
alpha_pmu         327 arch/alpha/kernel/perf_event.c 		delta += alpha_pmu->pmc_max_period[idx] + 1;
alpha_pmu         379 arch/alpha/kernel/perf_event.c 	if (n_ev > alpha_pmu->num_pmcs)
alpha_pmu         382 arch/alpha/kernel/perf_event.c 	return alpha_pmu->check_constraints(events, evtypes, n_ev);
alpha_pmu         457 arch/alpha/kernel/perf_event.c 	if (n0 < alpha_pmu->num_pmcs) {
alpha_pmu         616 arch/alpha/kernel/perf_event.c 		if (attr->config >= alpha_pmu->max_events)
alpha_pmu         618 arch/alpha/kernel/perf_event.c 		ev = alpha_pmu->event_map[attr->config];
alpha_pmu         622 arch/alpha/kernel/perf_event.c 		if (!alpha_pmu->raw_event_valid(attr->config))
alpha_pmu         651 arch/alpha/kernel/perf_event.c 				alpha_pmu->num_pmcs - 1,
alpha_pmu         679 arch/alpha/kernel/perf_event.c 		hwc->sample_period = alpha_pmu->pmc_max_period[0];
alpha_pmu         708 arch/alpha/kernel/perf_event.c 	if (!alpha_pmu)
alpha_pmu         791 arch/alpha/kernel/perf_event.c 	pcr0 = (pcr >> alpha_pmu->pmc_count_shift[0]) & alpha_pmu->pmc_count_mask[0];
alpha_pmu         792 arch/alpha/kernel/perf_event.c 	pcr1 = (pcr >> alpha_pmu->pmc_count_shift[1]) & alpha_pmu->pmc_count_mask[1];
alpha_pmu         824 arch/alpha/kernel/perf_event.c 	if (unlikely(la_ptr >= alpha_pmu->num_pmcs)) {
alpha_pmu         856 arch/alpha/kernel/perf_event.c 	alpha_perf_event_update(event, hwc, idx, alpha_pmu->pmc_max_period[idx]+1);
alpha_pmu         893 arch/alpha/kernel/perf_event.c 	alpha_pmu = &ev67_pmu;