cpuhw             122 arch/powerpc/perf/core-book3s.c static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw)
cpuhw             124 arch/powerpc/perf/core-book3s.c 	return cpuhw->mmcr[0];
cpuhw             130 arch/powerpc/perf/core-book3s.c static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {}
cpuhw             369 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw             375 arch/powerpc/perf/core-book3s.c 	if (event->ctx->task && cpuhw->bhrb_context != event->ctx) {
cpuhw             377 arch/powerpc/perf/core-book3s.c 		cpuhw->bhrb_context = event->ctx;
cpuhw             379 arch/powerpc/perf/core-book3s.c 	cpuhw->bhrb_users++;
cpuhw             385 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw             390 arch/powerpc/perf/core-book3s.c 	WARN_ON_ONCE(!cpuhw->bhrb_users);
cpuhw             391 arch/powerpc/perf/core-book3s.c 	cpuhw->bhrb_users--;
cpuhw             394 arch/powerpc/perf/core-book3s.c 	if (!cpuhw->disabled && !cpuhw->bhrb_users) {
cpuhw             400 arch/powerpc/perf/core-book3s.c 		cpuhw->bhrb_context = NULL;
cpuhw             447 arch/powerpc/perf/core-book3s.c static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw)
cpuhw             501 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].to = addr;
cpuhw             502 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].mispred = pred;
cpuhw             503 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].predicted = ~pred;
cpuhw             514 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].from = addr;
cpuhw             518 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].from = addr;
cpuhw             519 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].to =
cpuhw             521 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].mispred = pred;
cpuhw             522 arch/powerpc/perf/core-book3s.c 				cpuhw->bhrb_entries[u_index].predicted = ~pred;
cpuhw             528 arch/powerpc/perf/core-book3s.c 	cpuhw->bhrb_stack.nr = u_index;
cpuhw             596 arch/powerpc/perf/core-book3s.c static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw)
cpuhw             598 arch/powerpc/perf/core-book3s.c 	unsigned long mmcr0 = cpuhw->mmcr[0];
cpuhw             632 arch/powerpc/perf/core-book3s.c 	mtspr(SPRN_MMCR2, cpuhw->mmcr[3] | current->thread.mmcr2);
cpuhw             866 arch/powerpc/perf/core-book3s.c static int power_check_constraints(struct cpu_hw_events *cpuhw,
cpuhw             887 arch/powerpc/perf/core-book3s.c 					       cpuhw->alternatives[i]);
cpuhw             888 arch/powerpc/perf/core-book3s.c 			event_id[i] = cpuhw->alternatives[i][0];
cpuhw             890 arch/powerpc/perf/core-book3s.c 		if (ppmu->get_constraint(event_id[i], &cpuhw->amasks[i][0],
cpuhw             891 arch/powerpc/perf/core-book3s.c 					 &cpuhw->avalues[i][0]))
cpuhw             896 arch/powerpc/perf/core-book3s.c 		nv = (value | cpuhw->avalues[i][0]) +
cpuhw             897 arch/powerpc/perf/core-book3s.c 			(value & cpuhw->avalues[i][0] & addf);
cpuhw             902 arch/powerpc/perf/core-book3s.c 		if (((((nv + tadd) ^ cpuhw->avalues[i][0]) & cpuhw->amasks[i][0])
cpuhw             907 arch/powerpc/perf/core-book3s.c 		mask |= cpuhw->amasks[i][0];
cpuhw             922 arch/powerpc/perf/core-book3s.c 						  cpuhw->alternatives[i]);
cpuhw             924 arch/powerpc/perf/core-book3s.c 			ppmu->get_constraint(cpuhw->alternatives[i][j],
cpuhw             925 arch/powerpc/perf/core-book3s.c 					     &cpuhw->amasks[i][j],
cpuhw             926 arch/powerpc/perf/core-book3s.c 					     &cpuhw->avalues[i][j]);
cpuhw             945 arch/powerpc/perf/core-book3s.c 			nv = (value | cpuhw->avalues[i][j]) +
cpuhw             946 arch/powerpc/perf/core-book3s.c 				(value & cpuhw->avalues[i][j] & addf);
cpuhw             948 arch/powerpc/perf/core-book3s.c 			    (((nv + tadd) ^ cpuhw->avalues[i][j])
cpuhw             949 arch/powerpc/perf/core-book3s.c 			     & cpuhw->amasks[i][j]) == 0)
cpuhw             971 arch/powerpc/perf/core-book3s.c 			mask |= cpuhw->amasks[i][j];
cpuhw             979 arch/powerpc/perf/core-book3s.c 		event_id[i] = cpuhw->alternatives[i][choice[i]];
cpuhw            1113 arch/powerpc/perf/core-book3s.c static void freeze_limited_counters(struct cpu_hw_events *cpuhw,
cpuhw            1120 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_limited; ++i) {
cpuhw            1121 arch/powerpc/perf/core-book3s.c 		event = cpuhw->limited_counter[i];
cpuhw            1133 arch/powerpc/perf/core-book3s.c static void thaw_limited_counters(struct cpu_hw_events *cpuhw,
cpuhw            1140 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_limited; ++i) {
cpuhw            1141 arch/powerpc/perf/core-book3s.c 		event = cpuhw->limited_counter[i];
cpuhw            1142 arch/powerpc/perf/core-book3s.c 		event->hw.idx = cpuhw->limited_hwidx[i];
cpuhw            1162 arch/powerpc/perf/core-book3s.c static void write_mmcr0(struct cpu_hw_events *cpuhw, unsigned long mmcr0)
cpuhw            1166 arch/powerpc/perf/core-book3s.c 	if (!cpuhw->n_limited) {
cpuhw            1185 arch/powerpc/perf/core-book3s.c 		freeze_limited_counters(cpuhw, pmc5, pmc6);
cpuhw            1187 arch/powerpc/perf/core-book3s.c 		thaw_limited_counters(cpuhw, pmc5, pmc6);
cpuhw            1203 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1209 arch/powerpc/perf/core-book3s.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1211 arch/powerpc/perf/core-book3s.c 	if (!cpuhw->disabled) {
cpuhw            1215 arch/powerpc/perf/core-book3s.c 		if (!cpuhw->pmcs_enabled) {
cpuhw            1217 arch/powerpc/perf/core-book3s.c 			cpuhw->pmcs_enabled = 1;
cpuhw            1233 arch/powerpc/perf/core-book3s.c 		write_mmcr0(cpuhw, val);
cpuhw            1240 arch/powerpc/perf/core-book3s.c 		if (cpuhw->mmcr[2] & MMCRA_SAMPLE_ENABLE) {
cpuhw            1242 arch/powerpc/perf/core-book3s.c 			      cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE);
cpuhw            1247 arch/powerpc/perf/core-book3s.c 		cpuhw->disabled = 1;
cpuhw            1248 arch/powerpc/perf/core-book3s.c 		cpuhw->n_added = 0;
cpuhw            1277 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1291 arch/powerpc/perf/core-book3s.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1292 arch/powerpc/perf/core-book3s.c 	if (!cpuhw->disabled)
cpuhw            1295 arch/powerpc/perf/core-book3s.c 	if (cpuhw->n_events == 0) {
cpuhw            1300 arch/powerpc/perf/core-book3s.c 	cpuhw->disabled = 0;
cpuhw            1307 arch/powerpc/perf/core-book3s.c 	ebb = is_ebb_event(cpuhw->event[0]);
cpuhw            1315 arch/powerpc/perf/core-book3s.c 	if (!cpuhw->n_added) {
cpuhw            1316 arch/powerpc/perf/core-book3s.c 		mtspr(SPRN_MMCRA, cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE);
cpuhw            1317 arch/powerpc/perf/core-book3s.c 		mtspr(SPRN_MMCR1, cpuhw->mmcr[1]);
cpuhw            1324 arch/powerpc/perf/core-book3s.c 	memset(cpuhw->mmcr, 0, sizeof(cpuhw->mmcr));
cpuhw            1326 arch/powerpc/perf/core-book3s.c 	if (ppmu->compute_mmcr(cpuhw->events, cpuhw->n_events, hwc_index,
cpuhw            1327 arch/powerpc/perf/core-book3s.c 			       cpuhw->mmcr, cpuhw->event)) {
cpuhw            1339 arch/powerpc/perf/core-book3s.c 		event = cpuhw->event[0];
cpuhw            1341 arch/powerpc/perf/core-book3s.c 			cpuhw->mmcr[0] |= MMCR0_FCP;
cpuhw            1343 arch/powerpc/perf/core-book3s.c 			cpuhw->mmcr[0] |= freeze_events_kernel;
cpuhw            1345 arch/powerpc/perf/core-book3s.c 			cpuhw->mmcr[0] |= MMCR0_FCHV;
cpuhw            1354 arch/powerpc/perf/core-book3s.c 	mtspr(SPRN_MMCRA, cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE);
cpuhw            1355 arch/powerpc/perf/core-book3s.c 	mtspr(SPRN_MMCR1, cpuhw->mmcr[1]);
cpuhw            1356 arch/powerpc/perf/core-book3s.c 	mtspr(SPRN_MMCR0, (cpuhw->mmcr[0] & ~(MMCR0_PMC1CE | MMCR0_PMCjCE))
cpuhw            1359 arch/powerpc/perf/core-book3s.c 		mtspr(SPRN_MMCR2, cpuhw->mmcr[3]);
cpuhw            1365 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_events; ++i) {
cpuhw            1366 arch/powerpc/perf/core-book3s.c 		event = cpuhw->event[i];
cpuhw            1377 arch/powerpc/perf/core-book3s.c 	cpuhw->n_limited = n_lim = 0;
cpuhw            1378 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_events; ++i) {
cpuhw            1379 arch/powerpc/perf/core-book3s.c 		event = cpuhw->event[i];
cpuhw            1384 arch/powerpc/perf/core-book3s.c 			cpuhw->limited_counter[n_lim] = event;
cpuhw            1385 arch/powerpc/perf/core-book3s.c 			cpuhw->limited_hwidx[n_lim] = idx;
cpuhw            1409 arch/powerpc/perf/core-book3s.c 	cpuhw->n_limited = n_lim;
cpuhw            1410 arch/powerpc/perf/core-book3s.c 	cpuhw->mmcr[0] |= MMCR0_PMXE | MMCR0_FCECE;
cpuhw            1415 arch/powerpc/perf/core-book3s.c 	mmcr0 = ebb_switch_in(ebb, cpuhw);
cpuhw            1418 arch/powerpc/perf/core-book3s.c 	if (cpuhw->bhrb_users)
cpuhw            1419 arch/powerpc/perf/core-book3s.c 		ppmu->config_bhrb(cpuhw->bhrb_filter);
cpuhw            1421 arch/powerpc/perf/core-book3s.c 	write_mmcr0(cpuhw, mmcr0);
cpuhw            1426 arch/powerpc/perf/core-book3s.c 	if (cpuhw->mmcr[2] & MMCRA_SAMPLE_ENABLE) {
cpuhw            1428 arch/powerpc/perf/core-book3s.c 		mtspr(SPRN_MMCRA, cpuhw->mmcr[2]);
cpuhw            1471 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1483 arch/powerpc/perf/core-book3s.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1484 arch/powerpc/perf/core-book3s.c 	n0 = cpuhw->n_events;
cpuhw            1487 arch/powerpc/perf/core-book3s.c 	cpuhw->event[n0] = event;
cpuhw            1488 arch/powerpc/perf/core-book3s.c 	cpuhw->events[n0] = event->hw.config;
cpuhw            1489 arch/powerpc/perf/core-book3s.c 	cpuhw->flags[n0] = event->hw.event_base;
cpuhw            1507 arch/powerpc/perf/core-book3s.c 	if (cpuhw->txn_flags & PERF_PMU_TXN_ADD)
cpuhw            1510 arch/powerpc/perf/core-book3s.c 	if (check_excludes(cpuhw->event, cpuhw->flags, n0, 1))
cpuhw            1512 arch/powerpc/perf/core-book3s.c 	if (power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n0 + 1))
cpuhw            1514 arch/powerpc/perf/core-book3s.c 	event->hw.config = cpuhw->events[n0];
cpuhw            1519 arch/powerpc/perf/core-book3s.c 	++cpuhw->n_events;
cpuhw            1520 arch/powerpc/perf/core-book3s.c 	++cpuhw->n_added;
cpuhw            1526 arch/powerpc/perf/core-book3s.c 		cpuhw->bhrb_filter = ppmu->bhrb_filter_map(
cpuhw            1540 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1549 arch/powerpc/perf/core-book3s.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1550 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_events; ++i) {
cpuhw            1551 arch/powerpc/perf/core-book3s.c 		if (event == cpuhw->event[i]) {
cpuhw            1552 arch/powerpc/perf/core-book3s.c 			while (++i < cpuhw->n_events) {
cpuhw            1553 arch/powerpc/perf/core-book3s.c 				cpuhw->event[i-1] = cpuhw->event[i];
cpuhw            1554 arch/powerpc/perf/core-book3s.c 				cpuhw->events[i-1] = cpuhw->events[i];
cpuhw            1555 arch/powerpc/perf/core-book3s.c 				cpuhw->flags[i-1] = cpuhw->flags[i];
cpuhw            1557 arch/powerpc/perf/core-book3s.c 			--cpuhw->n_events;
cpuhw            1558 arch/powerpc/perf/core-book3s.c 			ppmu->disable_pmc(event->hw.idx - 1, cpuhw->mmcr);
cpuhw            1567 arch/powerpc/perf/core-book3s.c 	for (i = 0; i < cpuhw->n_limited; ++i)
cpuhw            1568 arch/powerpc/perf/core-book3s.c 		if (event == cpuhw->limited_counter[i])
cpuhw            1570 arch/powerpc/perf/core-book3s.c 	if (i < cpuhw->n_limited) {
cpuhw            1571 arch/powerpc/perf/core-book3s.c 		while (++i < cpuhw->n_limited) {
cpuhw            1572 arch/powerpc/perf/core-book3s.c 			cpuhw->limited_counter[i-1] = cpuhw->limited_counter[i];
cpuhw            1573 arch/powerpc/perf/core-book3s.c 			cpuhw->limited_hwidx[i-1] = cpuhw->limited_hwidx[i];
cpuhw            1575 arch/powerpc/perf/core-book3s.c 		--cpuhw->n_limited;
cpuhw            1577 arch/powerpc/perf/core-book3s.c 	if (cpuhw->n_events == 0) {
cpuhw            1579 arch/powerpc/perf/core-book3s.c 		cpuhw->mmcr[0] &= ~(MMCR0_PMXE | MMCR0_FCECE);
cpuhw            1659 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1661 arch/powerpc/perf/core-book3s.c 	WARN_ON_ONCE(cpuhw->txn_flags);		/* txn already in flight */
cpuhw            1663 arch/powerpc/perf/core-book3s.c 	cpuhw->txn_flags = txn_flags;
cpuhw            1668 arch/powerpc/perf/core-book3s.c 	cpuhw->n_txn_start = cpuhw->n_events;
cpuhw            1678 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1681 arch/powerpc/perf/core-book3s.c 	WARN_ON_ONCE(!cpuhw->txn_flags);	/* no txn in flight */
cpuhw            1683 arch/powerpc/perf/core-book3s.c 	txn_flags = cpuhw->txn_flags;
cpuhw            1684 arch/powerpc/perf/core-book3s.c 	cpuhw->txn_flags = 0;
cpuhw            1698 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1704 arch/powerpc/perf/core-book3s.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1705 arch/powerpc/perf/core-book3s.c 	WARN_ON_ONCE(!cpuhw->txn_flags);	/* no txn in flight */
cpuhw            1707 arch/powerpc/perf/core-book3s.c 	if (cpuhw->txn_flags & ~PERF_PMU_TXN_ADD) {
cpuhw            1708 arch/powerpc/perf/core-book3s.c 		cpuhw->txn_flags = 0;
cpuhw            1712 arch/powerpc/perf/core-book3s.c 	n = cpuhw->n_events;
cpuhw            1713 arch/powerpc/perf/core-book3s.c 	if (check_excludes(cpuhw->event, cpuhw->flags, 0, n))
cpuhw            1715 arch/powerpc/perf/core-book3s.c 	i = power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n);
cpuhw            1719 arch/powerpc/perf/core-book3s.c 	for (i = cpuhw->n_txn_start; i < n; ++i)
cpuhw            1720 arch/powerpc/perf/core-book3s.c 		cpuhw->event[i]->hw.config = cpuhw->events[i];
cpuhw            1722 arch/powerpc/perf/core-book3s.c 	cpuhw->txn_flags = 0;
cpuhw            1848 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw;
cpuhw            1951 arch/powerpc/perf/core-book3s.c 	cpuhw = &get_cpu_var(cpu_hw_events);
cpuhw            1952 arch/powerpc/perf/core-book3s.c 	err = power_check_constraints(cpuhw, events, cflags, n + 1);
cpuhw            1962 arch/powerpc/perf/core-book3s.c 		cpuhw->bhrb_filter = bhrb_filter;
cpuhw            2093 arch/powerpc/perf/core-book3s.c 			struct cpu_hw_events *cpuhw;
cpuhw            2094 arch/powerpc/perf/core-book3s.c 			cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            2095 arch/powerpc/perf/core-book3s.c 			power_pmu_bhrb_read(cpuhw);
cpuhw            2096 arch/powerpc/perf/core-book3s.c 			data.br_stack = &cpuhw->bhrb_stack;
cpuhw            2175 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            2181 arch/powerpc/perf/core-book3s.c 	if (cpuhw->n_limited)
cpuhw            2182 arch/powerpc/perf/core-book3s.c 		freeze_limited_counters(cpuhw, mfspr(SPRN_PMC5),
cpuhw            2211 arch/powerpc/perf/core-book3s.c 		for (j = 0; j < cpuhw->n_events; ++j) {
cpuhw            2212 arch/powerpc/perf/core-book3s.c 			event = cpuhw->event[j];
cpuhw            2225 arch/powerpc/perf/core-book3s.c 		for (i = 0; i < cpuhw->n_events; ++i) {
cpuhw            2226 arch/powerpc/perf/core-book3s.c 			event = cpuhw->event[i];
cpuhw            2248 arch/powerpc/perf/core-book3s.c 	write_mmcr0(cpuhw, cpuhw->mmcr[0]);
cpuhw            2266 arch/powerpc/perf/core-book3s.c 	struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu);
cpuhw            2269 arch/powerpc/perf/core-book3s.c 		memset(cpuhw, 0, sizeof(*cpuhw));
cpuhw            2270 arch/powerpc/perf/core-book3s.c 		cpuhw->mmcr[0] = MMCR0_FC;
cpuhw             205 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw;
cpuhw             209 arch/powerpc/perf/core-fsl-emb.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw             211 arch/powerpc/perf/core-fsl-emb.c 	if (!cpuhw->disabled) {
cpuhw             212 arch/powerpc/perf/core-fsl-emb.c 		cpuhw->disabled = 1;
cpuhw             217 arch/powerpc/perf/core-fsl-emb.c 		if (!cpuhw->pmcs_enabled) {
cpuhw             219 arch/powerpc/perf/core-fsl-emb.c 			cpuhw->pmcs_enabled = 1;
cpuhw             244 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw;
cpuhw             248 arch/powerpc/perf/core-fsl-emb.c 	cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw             249 arch/powerpc/perf/core-fsl-emb.c 	if (!cpuhw->disabled)
cpuhw             252 arch/powerpc/perf/core-fsl-emb.c 	cpuhw->disabled = 0;
cpuhw             253 arch/powerpc/perf/core-fsl-emb.c 	ppc_set_pmu_inuse(cpuhw->n_events != 0);
cpuhw             255 arch/powerpc/perf/core-fsl-emb.c 	if (cpuhw->n_events > 0) {
cpuhw             291 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw;
cpuhw             298 arch/powerpc/perf/core-fsl-emb.c 	cpuhw = &get_cpu_var(cpu_hw_events);
cpuhw             308 arch/powerpc/perf/core-fsl-emb.c 		if (cpuhw->event[i])
cpuhw             318 arch/powerpc/perf/core-fsl-emb.c 	cpuhw->event[i] = event;
cpuhw             319 arch/powerpc/perf/core-fsl-emb.c 	++cpuhw->n_events;
cpuhw             352 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw;
cpuhw             361 arch/powerpc/perf/core-fsl-emb.c 	cpuhw = &get_cpu_var(cpu_hw_events);
cpuhw             363 arch/powerpc/perf/core-fsl-emb.c 	WARN_ON(event != cpuhw->event[event->hw.idx]);
cpuhw             369 arch/powerpc/perf/core-fsl-emb.c 	cpuhw->event[i] = NULL;
cpuhw             380 arch/powerpc/perf/core-fsl-emb.c 	cpuhw->n_events--;
cpuhw             658 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw             671 arch/powerpc/perf/core-fsl-emb.c 		event = cpuhw->event[i];
cpuhw             702 arch/powerpc/perf/core-fsl-emb.c 	struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu);
cpuhw             704 arch/powerpc/perf/core-fsl-emb.c 	memset(cpuhw, 0, sizeof(*cpuhw));
cpuhw              39 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw;
cpuhw              43 arch/s390/kernel/perf_cpum_cf.c 	cpuhw = &get_cpu_var(cpu_cf_events);
cpuhw              49 arch/s390/kernel/perf_cpum_cf.c 		if (cpuhw->info.cfvn < 1)
cpuhw              53 arch/s390/kernel/perf_cpum_cf.c 		if ((cpuhw->info.csvn >= 1 && cpuhw->info.csvn <= 5 &&
cpuhw              55 arch/s390/kernel/perf_cpum_cf.c 		    (cpuhw->info.csvn >= 6 && hwc->config > 83))
cpuhw              59 arch/s390/kernel/perf_cpum_cf.c 		if (cpuhw->info.csvn < 1)
cpuhw              61 arch/s390/kernel/perf_cpum_cf.c 		if ((cpuhw->info.csvn == 1 && hwc->config > 159) ||
cpuhw              62 arch/s390/kernel/perf_cpum_cf.c 		    (cpuhw->info.csvn == 2 && hwc->config > 175) ||
cpuhw              63 arch/s390/kernel/perf_cpum_cf.c 		    (cpuhw->info.csvn >= 3 && cpuhw->info.csvn <= 5
cpuhw              65 arch/s390/kernel/perf_cpum_cf.c 		    (cpuhw->info.csvn >= 6 && hwc->config > 287))
cpuhw              69 arch/s390/kernel/perf_cpum_cf.c 		if (cpuhw->info.csvn <= 3)
cpuhw              84 arch/s390/kernel/perf_cpum_cf.c 		if (!((cpuhw->info.auth_ctl & mtdiag_ctl) &&
cpuhw              85 arch/s390/kernel/perf_cpum_cf.c 		      (cpuhw->info.enable_ctl & mtdiag_ctl) &&
cpuhw              86 arch/s390/kernel/perf_cpum_cf.c 		      (cpuhw->info.act_ctl & mtdiag_ctl)))
cpuhw              97 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw;
cpuhw             101 arch/s390/kernel/perf_cpum_cf.c 	cpuhw = &get_cpu_var(cpu_cf_events);
cpuhw             109 arch/s390/kernel/perf_cpum_cf.c 	if (!(ctrs_state & cpuhw->info.auth_ctl))
cpuhw             123 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             126 arch/s390/kernel/perf_cpum_cf.c 	if (cpuhw->flags & PMU_F_ENABLED)
cpuhw             129 arch/s390/kernel/perf_cpum_cf.c 	err = lcctl(cpuhw->state);
cpuhw             136 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->flags |= PMU_F_ENABLED;
cpuhw             146 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             150 arch/s390/kernel/perf_cpum_cf.c 	if (!(cpuhw->flags & PMU_F_ENABLED))
cpuhw             153 arch/s390/kernel/perf_cpum_cf.c 	inactive = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1);
cpuhw             161 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->flags &= ~PMU_F_ENABLED;
cpuhw             364 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             379 arch/s390/kernel/perf_cpum_cf.c 	ctr_set_enable(&cpuhw->state, hwc->config_base);
cpuhw             380 arch/s390/kernel/perf_cpum_cf.c 	ctr_set_start(&cpuhw->state, hwc->config_base);
cpuhw             390 arch/s390/kernel/perf_cpum_cf.c 	atomic_inc(&cpuhw->ctr_set[hwc->config_base]);
cpuhw             395 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             403 arch/s390/kernel/perf_cpum_cf.c 		if (!atomic_dec_return(&cpuhw->ctr_set[hwc->config_base]))
cpuhw             404 arch/s390/kernel/perf_cpum_cf.c 			ctr_set_stop(&cpuhw->state, hwc->config_base);
cpuhw             416 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             423 arch/s390/kernel/perf_cpum_cf.c 	if (!(cpuhw->txn_flags & PERF_PMU_TXN_ADD))
cpuhw             427 arch/s390/kernel/perf_cpum_cf.c 	ctr_set_enable(&cpuhw->state, event->hw.config_base);
cpuhw             440 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             452 arch/s390/kernel/perf_cpum_cf.c 	if (!atomic_read(&cpuhw->ctr_set[event->hw.config_base]))
cpuhw             453 arch/s390/kernel/perf_cpum_cf.c 		ctr_set_disable(&cpuhw->state, event->hw.config_base);
cpuhw             468 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             470 arch/s390/kernel/perf_cpum_cf.c 	WARN_ON_ONCE(cpuhw->txn_flags);		/* txn already in flight */
cpuhw             472 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->txn_flags = txn_flags;
cpuhw             477 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->tx_state = cpuhw->state;
cpuhw             488 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             490 arch/s390/kernel/perf_cpum_cf.c 	WARN_ON_ONCE(!cpuhw->txn_flags);	/* no txn in flight */
cpuhw             492 arch/s390/kernel/perf_cpum_cf.c 	txn_flags = cpuhw->txn_flags;
cpuhw             493 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->txn_flags = 0;
cpuhw             497 arch/s390/kernel/perf_cpum_cf.c 	WARN_ON(cpuhw->tx_state != cpuhw->state);
cpuhw             509 arch/s390/kernel/perf_cpum_cf.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             512 arch/s390/kernel/perf_cpum_cf.c 	WARN_ON_ONCE(!cpuhw->txn_flags);	/* no txn in flight */
cpuhw             514 arch/s390/kernel/perf_cpum_cf.c 	if (cpuhw->txn_flags & ~PERF_PMU_TXN_ADD) {
cpuhw             515 arch/s390/kernel/perf_cpum_cf.c 		cpuhw->txn_flags = 0;
cpuhw             520 arch/s390/kernel/perf_cpum_cf.c 	state = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1);
cpuhw             522 arch/s390/kernel/perf_cpum_cf.c 	if ((state & cpuhw->info.auth_ctl) != state)
cpuhw             525 arch/s390/kernel/perf_cpum_cf.c 	cpuhw->txn_flags = 0;
cpuhw              42 arch/s390/kernel/perf_cpum_cf_common.c 	struct cpu_cf_events *cpuhw;
cpuhw              48 arch/s390/kernel/perf_cpum_cf_common.c 	cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw              52 arch/s390/kernel/perf_cpum_cf_common.c 	if (!(cpuhw->flags & PMU_F_RESERVED))
cpuhw              57 arch/s390/kernel/perf_cpum_cf_common.c 		qctri(&cpuhw->info);
cpuhw              69 arch/s390/kernel/perf_cpum_cf_common.c 	atomic64_or(alert, &cpuhw->alert);
cpuhw              76 arch/s390/kernel/perf_cpum_cf_common.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw              80 arch/s390/kernel/perf_cpum_cf_common.c 		memset(&cpuhw->info, 0, sizeof(cpuhw->info));
cpuhw              81 arch/s390/kernel/perf_cpum_cf_common.c 		qctri(&cpuhw->info);
cpuhw              82 arch/s390/kernel/perf_cpum_cf_common.c 		cpuhw->flags |= PMU_F_RESERVED;
cpuhw              86 arch/s390/kernel/perf_cpum_cf_common.c 		cpuhw->flags &= ~PMU_F_RESERVED;
cpuhw             130 arch/s390/kernel/perf_cpum_cf_common.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             133 arch/s390/kernel/perf_cpum_cf_common.c 	alert = atomic64_read(&cpuhw->alert);
cpuhw             135 arch/s390/kernel/perf_cpum_cf_common.c 		atomic64_set(&cpuhw->alert, 0);
cpuhw             109 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             112 arch/s390/kernel/perf_cpum_cf_diag.c 	te->cfvn = cpuhw->info.cfvn;		/* Counter version numbers */
cpuhw             113 arch/s390/kernel/perf_cpum_cf_diag.c 	te->csvn = cpuhw->info.csvn;
cpuhw             132 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             137 arch/s390/kernel/perf_cpum_cf_diag.c 			    __func__, pmu, smp_processor_id(), cpuhw->flags,
cpuhw             138 arch/s390/kernel/perf_cpum_cf_diag.c 			    cpuhw->state);
cpuhw             139 arch/s390/kernel/perf_cpum_cf_diag.c 	if (cpuhw->flags & PMU_F_ENABLED)
cpuhw             142 arch/s390/kernel/perf_cpum_cf_diag.c 	err = lcctl(cpuhw->state);
cpuhw             148 arch/s390/kernel/perf_cpum_cf_diag.c 	cpuhw->flags |= PMU_F_ENABLED;
cpuhw             158 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             164 arch/s390/kernel/perf_cpum_cf_diag.c 			    __func__, pmu, smp_processor_id(), cpuhw->flags,
cpuhw             165 arch/s390/kernel/perf_cpum_cf_diag.c 			    cpuhw->state);
cpuhw             166 arch/s390/kernel/perf_cpum_cf_diag.c 	if (!(cpuhw->flags & PMU_F_ENABLED))
cpuhw             169 arch/s390/kernel/perf_cpum_cf_diag.c 	inactive = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1);
cpuhw             176 arch/s390/kernel/perf_cpum_cf_diag.c 	cpuhw->flags &= ~PMU_F_ENABLED;
cpuhw             200 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw;
cpuhw             218 arch/s390/kernel/perf_cpum_cf_diag.c 	cpuhw = &get_cpu_var(cpu_cf_events);
cpuhw             220 arch/s390/kernel/perf_cpum_cf_diag.c 		if (cpuhw->info.auth_ctl & cpumf_ctr_ctl[i])
cpuhw             371 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             378 arch/s390/kernel/perf_cpum_cf_diag.c 	ctrset_size = cf_diag_ctrset_size(ctrset, &cpuhw->info);
cpuhw             394 arch/s390/kernel/perf_cpum_cf_diag.c 			    __func__, ctrset, ctrset_size, cpuhw->info.cfvn,
cpuhw             395 arch/s390/kernel/perf_cpum_cf_diag.c 			    cpuhw->info.csvn, need, rc);
cpuhw             523 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             536 arch/s390/kernel/perf_cpum_cf_diag.c 	ctr_set_multiple_enable(&cpuhw->state, hwc->config_base);
cpuhw             537 arch/s390/kernel/perf_cpum_cf_diag.c 	lcctl(cpuhw->state);	/* Enable counter sets */
cpuhw             540 arch/s390/kernel/perf_cpum_cf_diag.c 	ctr_set_multiple_start(&cpuhw->state, hwc->config_base);
cpuhw             546 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             555 arch/s390/kernel/perf_cpum_cf_diag.c 	ctr_set_multiple_stop(&cpuhw->state, hwc->config_base);
cpuhw             566 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             571 arch/s390/kernel/perf_cpum_cf_diag.c 			    __func__, event, event->cpu, flags, cpuhw);
cpuhw             573 arch/s390/kernel/perf_cpum_cf_diag.c 	if (cpuhw->flags & PMU_F_IN_USE) {
cpuhw             580 arch/s390/kernel/perf_cpum_cf_diag.c 	cpuhw->flags |= PMU_F_IN_USE;
cpuhw             590 arch/s390/kernel/perf_cpum_cf_diag.c 	struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events);
cpuhw             597 arch/s390/kernel/perf_cpum_cf_diag.c 	ctr_set_multiple_stop(&cpuhw->state, event->hw.config_base);
cpuhw             598 arch/s390/kernel/perf_cpum_cf_diag.c 	ctr_set_multiple_disable(&cpuhw->state, event->hw.config_base);
cpuhw             599 arch/s390/kernel/perf_cpum_cf_diag.c 	cpuhw->flags &= ~PMU_F_IN_USE;
cpuhw             116 arch/s390/kernel/perf_cpum_sf.c static int sf_buffer_available(struct cpu_hw_sf *cpuhw)
cpuhw             118 arch/s390/kernel/perf_cpum_sf.c 	return !!cpuhw->sfb.sdbt;
cpuhw             365 arch/s390/kernel/perf_cpum_sf.c static void deallocate_buffers(struct cpu_hw_sf *cpuhw)
cpuhw             367 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->sfb.sdbt)
cpuhw             368 arch/s390/kernel/perf_cpum_sf.c 		free_sampling_buffer(&cpuhw->sfb);
cpuhw             371 arch/s390/kernel/perf_cpum_sf.c static int allocate_buffers(struct cpu_hw_sf *cpuhw, struct hw_perf_event *hwc)
cpuhw             401 arch/s390/kernel/perf_cpum_sf.c 	freq = sample_rate_to_freq(&cpuhw->qsi, SAMPL_RATE(hwc));
cpuhw             416 arch/s390/kernel/perf_cpum_sf.c 	if (sf_buffer_available(cpuhw))
cpuhw             423 arch/s390/kernel/perf_cpum_sf.c 			    sample_size, cpuhw);
cpuhw             425 arch/s390/kernel/perf_cpum_sf.c 	return alloc_sampling_buffer(&cpuhw->sfb,
cpuhw             426 arch/s390/kernel/perf_cpum_sf.c 				     sfb_pending_allocs(&cpuhw->sfb, hwc));
cpuhw             458 arch/s390/kernel/perf_cpum_sf.c static void sfb_account_overflows(struct cpu_hw_sf *cpuhw,
cpuhw             473 arch/s390/kernel/perf_cpum_sf.c 	ratio = DIV_ROUND_UP(100 * OVERFLOW_REG(hwc) * cpuhw->sfb.num_sdb,
cpuhw             474 arch/s390/kernel/perf_cpum_sf.c 			     sample_rate_to_freq(&cpuhw->qsi, SAMPL_RATE(hwc)));
cpuhw             477 arch/s390/kernel/perf_cpum_sf.c 	num = compute_sfb_extent(ratio, cpuhw->sfb.num_sdb);
cpuhw             768 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw;
cpuhw             799 arch/s390/kernel/perf_cpum_sf.c 	cpuhw = NULL;
cpuhw             806 arch/s390/kernel/perf_cpum_sf.c 		cpuhw = &per_cpu(cpu_hw_sf, event->cpu);
cpuhw             807 arch/s390/kernel/perf_cpum_sf.c 		si = cpuhw->qsi;
cpuhw             860 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw)
cpuhw             862 arch/s390/kernel/perf_cpum_sf.c 		err = allocate_buffers(cpuhw, hwc);
cpuhw             868 arch/s390/kernel/perf_cpum_sf.c 			cpuhw = &per_cpu(cpu_hw_sf, cpu);
cpuhw             869 arch/s390/kernel/perf_cpum_sf.c 			err = allocate_buffers(cpuhw, hwc);
cpuhw             936 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw             940 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->flags & PMU_F_ENABLED)
cpuhw             943 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->flags & PMU_F_ERR_MASK)
cpuhw             957 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->event) {
cpuhw             958 arch/s390/kernel/perf_cpum_sf.c 		hwc = &cpuhw->event->hw;
cpuhw             964 arch/s390/kernel/perf_cpum_sf.c 			sfb_account_overflows(cpuhw, hwc);
cpuhw             965 arch/s390/kernel/perf_cpum_sf.c 			if (sfb_has_pending_allocs(&cpuhw->sfb, hwc))
cpuhw             966 arch/s390/kernel/perf_cpum_sf.c 				extend_sampling_buffer(&cpuhw->sfb, hwc);
cpuhw             969 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->lsctl.interval = SAMPL_RATE(&cpuhw->event->hw);
cpuhw             973 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->flags |= PMU_F_ENABLED;
cpuhw             976 arch/s390/kernel/perf_cpum_sf.c 	err = lsctl(&cpuhw->lsctl);
cpuhw             978 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->flags &= ~PMU_F_ENABLED;
cpuhw             989 arch/s390/kernel/perf_cpum_sf.c 			    cpuhw->lsctl.es, cpuhw->lsctl.cs, cpuhw->lsctl.ed,
cpuhw             990 arch/s390/kernel/perf_cpum_sf.c 			    cpuhw->lsctl.cd, cpuhw->lsctl.interval,
cpuhw             991 arch/s390/kernel/perf_cpum_sf.c 			    (void *) cpuhw->lsctl.tear,
cpuhw             992 arch/s390/kernel/perf_cpum_sf.c 			    (void *) cpuhw->lsctl.dear);
cpuhw             997 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            1002 arch/s390/kernel/perf_cpum_sf.c 	if (!(cpuhw->flags & PMU_F_ENABLED))
cpuhw            1005 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->flags & PMU_F_ERR_MASK)
cpuhw            1009 arch/s390/kernel/perf_cpum_sf.c 	inactive = cpuhw->lsctl;
cpuhw            1028 arch/s390/kernel/perf_cpum_sf.c 			cpuhw->lsctl.tear = si.tear;
cpuhw            1029 arch/s390/kernel/perf_cpum_sf.c 			cpuhw->lsctl.dear = si.dear;
cpuhw            1035 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->flags &= ~PMU_F_ENABLED;
cpuhw            1408 arch/s390/kernel/perf_cpum_sf.c 			    struct cpu_hw_sf *cpuhw)
cpuhw            1449 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.tear = base + offset * sizeof(unsigned long);
cpuhw            1450 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.dear = aux->sdb_index[head];
cpuhw            1459 arch/s390/kernel/perf_cpum_sf.c 			    cpuhw->lsctl.tear,
cpuhw            1460 arch/s390/kernel/perf_cpum_sf.c 			    cpuhw->lsctl.dear);
cpuhw            1574 arch/s390/kernel/perf_cpum_sf.c static void hw_collect_aux(struct cpu_hw_sf *cpuhw)
cpuhw            1580 arch/s390/kernel/perf_cpum_sf.c 	struct perf_output_handle *handle = &cpuhw->handle;
cpuhw            1595 arch/s390/kernel/perf_cpum_sf.c 		aux = perf_aux_output_begin(handle, cpuhw->event);
cpuhw            1620 arch/s390/kernel/perf_cpum_sf.c 			perf_aux_output_end(&cpuhw->handle, size);
cpuhw            1628 arch/s390/kernel/perf_cpum_sf.c 			perf_aux_output_end(&cpuhw->handle, size);
cpuhw            1813 arch/s390/kernel/perf_cpum_sf.c 		struct cpu_hw_sf *cpuhw = &per_cpu(cpu_hw_sf, event->cpu);
cpuhw            1815 arch/s390/kernel/perf_cpum_sf.c 		si = cpuhw->qsi;
cpuhw            1838 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            1848 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.cs = 1;
cpuhw            1850 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->lsctl.cd = 1;
cpuhw            1859 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            1865 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.cs = 0;
cpuhw            1866 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.cd = 0;
cpuhw            1878 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            1882 arch/s390/kernel/perf_cpum_sf.c 	if (cpuhw->flags & PMU_F_IN_USE)
cpuhw            1885 arch/s390/kernel/perf_cpum_sf.c 	if (!SAMPL_DIAG_MODE(&event->hw) && !cpuhw->sfb.sdbt)
cpuhw            1898 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.s = 0;
cpuhw            1899 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.h = 1;
cpuhw            1900 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.interval = SAMPL_RATE(&event->hw);
cpuhw            1902 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->lsctl.tear = (unsigned long) cpuhw->sfb.sdbt;
cpuhw            1903 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->lsctl.dear = *(unsigned long *) cpuhw->sfb.sdbt;
cpuhw            1904 arch/s390/kernel/perf_cpum_sf.c 		hw_reset_registers(&event->hw, cpuhw->sfb.sdbt);
cpuhw            1909 arch/s390/kernel/perf_cpum_sf.c 	if (WARN_ON_ONCE(cpuhw->lsctl.es == 1 || cpuhw->lsctl.ed == 1)) {
cpuhw            1914 arch/s390/kernel/perf_cpum_sf.c 		aux = perf_aux_output_begin(&cpuhw->handle, event);
cpuhw            1919 arch/s390/kernel/perf_cpum_sf.c 		err = aux_output_begin(&cpuhw->handle, aux, cpuhw);
cpuhw            1922 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->lsctl.ed = 1;
cpuhw            1924 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.es = 1;
cpuhw            1927 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->event = event;
cpuhw            1928 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->flags |= PMU_F_IN_USE;
cpuhw            1940 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            1945 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.es = 0;
cpuhw            1946 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->lsctl.ed = 0;
cpuhw            1947 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->flags &= ~PMU_F_IN_USE;
cpuhw            1948 arch/s390/kernel/perf_cpum_sf.c 	cpuhw->event = NULL;
cpuhw            1951 arch/s390/kernel/perf_cpum_sf.c 		aux_output_end(&cpuhw->handle);
cpuhw            2031 arch/s390/kernel/perf_cpum_sf.c 	struct cpu_hw_sf *cpuhw;
cpuhw            2036 arch/s390/kernel/perf_cpum_sf.c 	cpuhw = this_cpu_ptr(&cpu_hw_sf);
cpuhw            2040 arch/s390/kernel/perf_cpum_sf.c 	if (!(cpuhw->flags & PMU_F_RESERVED))
cpuhw            2048 arch/s390/kernel/perf_cpum_sf.c 		if (cpuhw->flags & PMU_F_IN_USE)
cpuhw            2049 arch/s390/kernel/perf_cpum_sf.c 			if (SAMPL_DIAG_MODE(&cpuhw->event->hw))
cpuhw            2050 arch/s390/kernel/perf_cpum_sf.c 				hw_collect_aux(cpuhw);
cpuhw            2052 arch/s390/kernel/perf_cpum_sf.c 				hw_perf_event_update(cpuhw->event, 0);
cpuhw            2054 arch/s390/kernel/perf_cpum_sf.c 			WARN_ON_ONCE(!(cpuhw->flags & PMU_F_IN_USE));
cpuhw            2064 arch/s390/kernel/perf_cpum_sf.c 		qsi(&cpuhw->qsi);
cpuhw            2069 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->flags |= PMU_F_ERR_LSDA;
cpuhw            2077 arch/s390/kernel/perf_cpum_sf.c 		cpuhw->flags |= PMU_F_ERR_IBE;
cpuhw             354 arch/sh/kernel/perf_event.c 	struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu);
cpuhw             356 arch/sh/kernel/perf_event.c 	memset(cpuhw, 0, sizeof(struct cpu_hw_events));
cpuhw            1514 arch/sparc/kernel/perf_event.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1516 arch/sparc/kernel/perf_event.c 	WARN_ON_ONCE(cpuhw->txn_flags);		/* txn already in flight */
cpuhw            1518 arch/sparc/kernel/perf_event.c 	cpuhw->txn_flags = txn_flags;
cpuhw            1532 arch/sparc/kernel/perf_event.c 	struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events);
cpuhw            1535 arch/sparc/kernel/perf_event.c 	WARN_ON_ONCE(!cpuhw->txn_flags);	/* no txn in flight */
cpuhw            1537 arch/sparc/kernel/perf_event.c 	txn_flags = cpuhw->txn_flags;
cpuhw            1538 arch/sparc/kernel/perf_event.c 	cpuhw->txn_flags = 0;
cpuhw             555 arch/x86/events/amd/core.c 	struct cpu_hw_events *cpuhw;
cpuhw             560 arch/x86/events/amd/core.c 	cpuhw = &per_cpu(cpu_hw_events, cpu);
cpuhw             562 arch/x86/events/amd/core.c 	if (cpuhw->amd_nb) {
cpuhw             563 arch/x86/events/amd/core.c 		struct amd_nb *nb = cpuhw->amd_nb;
cpuhw             568 arch/x86/events/amd/core.c 		cpuhw->amd_nb = NULL;