cpuhw 122 arch/powerpc/perf/core-book3s.c static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) cpuhw 124 arch/powerpc/perf/core-book3s.c return cpuhw->mmcr[0]; cpuhw 130 arch/powerpc/perf/core-book3s.c static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} cpuhw 369 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 375 arch/powerpc/perf/core-book3s.c if (event->ctx->task && cpuhw->bhrb_context != event->ctx) { cpuhw 377 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_context = event->ctx; cpuhw 379 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_users++; cpuhw 385 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 390 arch/powerpc/perf/core-book3s.c WARN_ON_ONCE(!cpuhw->bhrb_users); cpuhw 391 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_users--; cpuhw 394 arch/powerpc/perf/core-book3s.c if (!cpuhw->disabled && !cpuhw->bhrb_users) { cpuhw 400 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_context = NULL; cpuhw 447 arch/powerpc/perf/core-book3s.c static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) cpuhw 501 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].to = addr; cpuhw 502 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].mispred = pred; cpuhw 503 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].predicted = ~pred; cpuhw 514 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].from = addr; cpuhw 518 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].from = addr; cpuhw 519 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].to = cpuhw 521 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].mispred = pred; cpuhw 522 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_entries[u_index].predicted = ~pred; cpuhw 528 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_stack.nr = u_index; cpuhw 596 arch/powerpc/perf/core-book3s.c static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) cpuhw 598 arch/powerpc/perf/core-book3s.c unsigned long mmcr0 = cpuhw->mmcr[0]; cpuhw 632 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCR2, cpuhw->mmcr[3] | current->thread.mmcr2); cpuhw 866 arch/powerpc/perf/core-book3s.c static int power_check_constraints(struct cpu_hw_events *cpuhw, cpuhw 887 arch/powerpc/perf/core-book3s.c cpuhw->alternatives[i]); cpuhw 888 arch/powerpc/perf/core-book3s.c event_id[i] = cpuhw->alternatives[i][0]; cpuhw 890 arch/powerpc/perf/core-book3s.c if (ppmu->get_constraint(event_id[i], &cpuhw->amasks[i][0], cpuhw 891 arch/powerpc/perf/core-book3s.c &cpuhw->avalues[i][0])) cpuhw 896 arch/powerpc/perf/core-book3s.c nv = (value | cpuhw->avalues[i][0]) + cpuhw 897 arch/powerpc/perf/core-book3s.c (value & cpuhw->avalues[i][0] & addf); cpuhw 902 arch/powerpc/perf/core-book3s.c if (((((nv + tadd) ^ cpuhw->avalues[i][0]) & cpuhw->amasks[i][0]) cpuhw 907 arch/powerpc/perf/core-book3s.c mask |= cpuhw->amasks[i][0]; cpuhw 922 arch/powerpc/perf/core-book3s.c cpuhw->alternatives[i]); cpuhw 924 arch/powerpc/perf/core-book3s.c ppmu->get_constraint(cpuhw->alternatives[i][j], cpuhw 925 arch/powerpc/perf/core-book3s.c &cpuhw->amasks[i][j], cpuhw 926 arch/powerpc/perf/core-book3s.c &cpuhw->avalues[i][j]); cpuhw 945 arch/powerpc/perf/core-book3s.c nv = (value | cpuhw->avalues[i][j]) + cpuhw 946 arch/powerpc/perf/core-book3s.c (value & cpuhw->avalues[i][j] & addf); cpuhw 948 arch/powerpc/perf/core-book3s.c (((nv + tadd) ^ cpuhw->avalues[i][j]) cpuhw 949 arch/powerpc/perf/core-book3s.c & cpuhw->amasks[i][j]) == 0) cpuhw 971 arch/powerpc/perf/core-book3s.c mask |= cpuhw->amasks[i][j]; cpuhw 979 arch/powerpc/perf/core-book3s.c event_id[i] = cpuhw->alternatives[i][choice[i]]; cpuhw 1113 arch/powerpc/perf/core-book3s.c static void freeze_limited_counters(struct cpu_hw_events *cpuhw, cpuhw 1120 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_limited; ++i) { cpuhw 1121 arch/powerpc/perf/core-book3s.c event = cpuhw->limited_counter[i]; cpuhw 1133 arch/powerpc/perf/core-book3s.c static void thaw_limited_counters(struct cpu_hw_events *cpuhw, cpuhw 1140 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_limited; ++i) { cpuhw 1141 arch/powerpc/perf/core-book3s.c event = cpuhw->limited_counter[i]; cpuhw 1142 arch/powerpc/perf/core-book3s.c event->hw.idx = cpuhw->limited_hwidx[i]; cpuhw 1162 arch/powerpc/perf/core-book3s.c static void write_mmcr0(struct cpu_hw_events *cpuhw, unsigned long mmcr0) cpuhw 1166 arch/powerpc/perf/core-book3s.c if (!cpuhw->n_limited) { cpuhw 1185 arch/powerpc/perf/core-book3s.c freeze_limited_counters(cpuhw, pmc5, pmc6); cpuhw 1187 arch/powerpc/perf/core-book3s.c thaw_limited_counters(cpuhw, pmc5, pmc6); cpuhw 1203 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1209 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1211 arch/powerpc/perf/core-book3s.c if (!cpuhw->disabled) { cpuhw 1215 arch/powerpc/perf/core-book3s.c if (!cpuhw->pmcs_enabled) { cpuhw 1217 arch/powerpc/perf/core-book3s.c cpuhw->pmcs_enabled = 1; cpuhw 1233 arch/powerpc/perf/core-book3s.c write_mmcr0(cpuhw, val); cpuhw 1240 arch/powerpc/perf/core-book3s.c if (cpuhw->mmcr[2] & MMCRA_SAMPLE_ENABLE) { cpuhw 1242 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE); cpuhw 1247 arch/powerpc/perf/core-book3s.c cpuhw->disabled = 1; cpuhw 1248 arch/powerpc/perf/core-book3s.c cpuhw->n_added = 0; cpuhw 1277 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1291 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1292 arch/powerpc/perf/core-book3s.c if (!cpuhw->disabled) cpuhw 1295 arch/powerpc/perf/core-book3s.c if (cpuhw->n_events == 0) { cpuhw 1300 arch/powerpc/perf/core-book3s.c cpuhw->disabled = 0; cpuhw 1307 arch/powerpc/perf/core-book3s.c ebb = is_ebb_event(cpuhw->event[0]); cpuhw 1315 arch/powerpc/perf/core-book3s.c if (!cpuhw->n_added) { cpuhw 1316 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCRA, cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE); cpuhw 1317 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCR1, cpuhw->mmcr[1]); cpuhw 1324 arch/powerpc/perf/core-book3s.c memset(cpuhw->mmcr, 0, sizeof(cpuhw->mmcr)); cpuhw 1326 arch/powerpc/perf/core-book3s.c if (ppmu->compute_mmcr(cpuhw->events, cpuhw->n_events, hwc_index, cpuhw 1327 arch/powerpc/perf/core-book3s.c cpuhw->mmcr, cpuhw->event)) { cpuhw 1339 arch/powerpc/perf/core-book3s.c event = cpuhw->event[0]; cpuhw 1341 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] |= MMCR0_FCP; cpuhw 1343 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] |= freeze_events_kernel; cpuhw 1345 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] |= MMCR0_FCHV; cpuhw 1354 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCRA, cpuhw->mmcr[2] & ~MMCRA_SAMPLE_ENABLE); cpuhw 1355 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCR1, cpuhw->mmcr[1]); cpuhw 1356 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCR0, (cpuhw->mmcr[0] & ~(MMCR0_PMC1CE | MMCR0_PMCjCE)) cpuhw 1359 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCR2, cpuhw->mmcr[3]); cpuhw 1365 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_events; ++i) { cpuhw 1366 arch/powerpc/perf/core-book3s.c event = cpuhw->event[i]; cpuhw 1377 arch/powerpc/perf/core-book3s.c cpuhw->n_limited = n_lim = 0; cpuhw 1378 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_events; ++i) { cpuhw 1379 arch/powerpc/perf/core-book3s.c event = cpuhw->event[i]; cpuhw 1384 arch/powerpc/perf/core-book3s.c cpuhw->limited_counter[n_lim] = event; cpuhw 1385 arch/powerpc/perf/core-book3s.c cpuhw->limited_hwidx[n_lim] = idx; cpuhw 1409 arch/powerpc/perf/core-book3s.c cpuhw->n_limited = n_lim; cpuhw 1410 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] |= MMCR0_PMXE | MMCR0_FCECE; cpuhw 1415 arch/powerpc/perf/core-book3s.c mmcr0 = ebb_switch_in(ebb, cpuhw); cpuhw 1418 arch/powerpc/perf/core-book3s.c if (cpuhw->bhrb_users) cpuhw 1419 arch/powerpc/perf/core-book3s.c ppmu->config_bhrb(cpuhw->bhrb_filter); cpuhw 1421 arch/powerpc/perf/core-book3s.c write_mmcr0(cpuhw, mmcr0); cpuhw 1426 arch/powerpc/perf/core-book3s.c if (cpuhw->mmcr[2] & MMCRA_SAMPLE_ENABLE) { cpuhw 1428 arch/powerpc/perf/core-book3s.c mtspr(SPRN_MMCRA, cpuhw->mmcr[2]); cpuhw 1471 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1483 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1484 arch/powerpc/perf/core-book3s.c n0 = cpuhw->n_events; cpuhw 1487 arch/powerpc/perf/core-book3s.c cpuhw->event[n0] = event; cpuhw 1488 arch/powerpc/perf/core-book3s.c cpuhw->events[n0] = event->hw.config; cpuhw 1489 arch/powerpc/perf/core-book3s.c cpuhw->flags[n0] = event->hw.event_base; cpuhw 1507 arch/powerpc/perf/core-book3s.c if (cpuhw->txn_flags & PERF_PMU_TXN_ADD) cpuhw 1510 arch/powerpc/perf/core-book3s.c if (check_excludes(cpuhw->event, cpuhw->flags, n0, 1)) cpuhw 1512 arch/powerpc/perf/core-book3s.c if (power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n0 + 1)) cpuhw 1514 arch/powerpc/perf/core-book3s.c event->hw.config = cpuhw->events[n0]; cpuhw 1519 arch/powerpc/perf/core-book3s.c ++cpuhw->n_events; cpuhw 1520 arch/powerpc/perf/core-book3s.c ++cpuhw->n_added; cpuhw 1526 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_filter = ppmu->bhrb_filter_map( cpuhw 1540 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1549 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1550 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_events; ++i) { cpuhw 1551 arch/powerpc/perf/core-book3s.c if (event == cpuhw->event[i]) { cpuhw 1552 arch/powerpc/perf/core-book3s.c while (++i < cpuhw->n_events) { cpuhw 1553 arch/powerpc/perf/core-book3s.c cpuhw->event[i-1] = cpuhw->event[i]; cpuhw 1554 arch/powerpc/perf/core-book3s.c cpuhw->events[i-1] = cpuhw->events[i]; cpuhw 1555 arch/powerpc/perf/core-book3s.c cpuhw->flags[i-1] = cpuhw->flags[i]; cpuhw 1557 arch/powerpc/perf/core-book3s.c --cpuhw->n_events; cpuhw 1558 arch/powerpc/perf/core-book3s.c ppmu->disable_pmc(event->hw.idx - 1, cpuhw->mmcr); cpuhw 1567 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_limited; ++i) cpuhw 1568 arch/powerpc/perf/core-book3s.c if (event == cpuhw->limited_counter[i]) cpuhw 1570 arch/powerpc/perf/core-book3s.c if (i < cpuhw->n_limited) { cpuhw 1571 arch/powerpc/perf/core-book3s.c while (++i < cpuhw->n_limited) { cpuhw 1572 arch/powerpc/perf/core-book3s.c cpuhw->limited_counter[i-1] = cpuhw->limited_counter[i]; cpuhw 1573 arch/powerpc/perf/core-book3s.c cpuhw->limited_hwidx[i-1] = cpuhw->limited_hwidx[i]; cpuhw 1575 arch/powerpc/perf/core-book3s.c --cpuhw->n_limited; cpuhw 1577 arch/powerpc/perf/core-book3s.c if (cpuhw->n_events == 0) { cpuhw 1579 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] &= ~(MMCR0_PMXE | MMCR0_FCECE); cpuhw 1659 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1661 arch/powerpc/perf/core-book3s.c WARN_ON_ONCE(cpuhw->txn_flags); /* txn already in flight */ cpuhw 1663 arch/powerpc/perf/core-book3s.c cpuhw->txn_flags = txn_flags; cpuhw 1668 arch/powerpc/perf/core-book3s.c cpuhw->n_txn_start = cpuhw->n_events; cpuhw 1678 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1681 arch/powerpc/perf/core-book3s.c WARN_ON_ONCE(!cpuhw->txn_flags); /* no txn in flight */ cpuhw 1683 arch/powerpc/perf/core-book3s.c txn_flags = cpuhw->txn_flags; cpuhw 1684 arch/powerpc/perf/core-book3s.c cpuhw->txn_flags = 0; cpuhw 1698 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1704 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1705 arch/powerpc/perf/core-book3s.c WARN_ON_ONCE(!cpuhw->txn_flags); /* no txn in flight */ cpuhw 1707 arch/powerpc/perf/core-book3s.c if (cpuhw->txn_flags & ~PERF_PMU_TXN_ADD) { cpuhw 1708 arch/powerpc/perf/core-book3s.c cpuhw->txn_flags = 0; cpuhw 1712 arch/powerpc/perf/core-book3s.c n = cpuhw->n_events; cpuhw 1713 arch/powerpc/perf/core-book3s.c if (check_excludes(cpuhw->event, cpuhw->flags, 0, n)) cpuhw 1715 arch/powerpc/perf/core-book3s.c i = power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n); cpuhw 1719 arch/powerpc/perf/core-book3s.c for (i = cpuhw->n_txn_start; i < n; ++i) cpuhw 1720 arch/powerpc/perf/core-book3s.c cpuhw->event[i]->hw.config = cpuhw->events[i]; cpuhw 1722 arch/powerpc/perf/core-book3s.c cpuhw->txn_flags = 0; cpuhw 1848 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 1951 arch/powerpc/perf/core-book3s.c cpuhw = &get_cpu_var(cpu_hw_events); cpuhw 1952 arch/powerpc/perf/core-book3s.c err = power_check_constraints(cpuhw, events, cflags, n + 1); cpuhw 1962 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_filter = bhrb_filter; cpuhw 2093 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw; cpuhw 2094 arch/powerpc/perf/core-book3s.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 2095 arch/powerpc/perf/core-book3s.c power_pmu_bhrb_read(cpuhw); cpuhw 2096 arch/powerpc/perf/core-book3s.c data.br_stack = &cpuhw->bhrb_stack; cpuhw 2175 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 2181 arch/powerpc/perf/core-book3s.c if (cpuhw->n_limited) cpuhw 2182 arch/powerpc/perf/core-book3s.c freeze_limited_counters(cpuhw, mfspr(SPRN_PMC5), cpuhw 2211 arch/powerpc/perf/core-book3s.c for (j = 0; j < cpuhw->n_events; ++j) { cpuhw 2212 arch/powerpc/perf/core-book3s.c event = cpuhw->event[j]; cpuhw 2225 arch/powerpc/perf/core-book3s.c for (i = 0; i < cpuhw->n_events; ++i) { cpuhw 2226 arch/powerpc/perf/core-book3s.c event = cpuhw->event[i]; cpuhw 2248 arch/powerpc/perf/core-book3s.c write_mmcr0(cpuhw, cpuhw->mmcr[0]); cpuhw 2266 arch/powerpc/perf/core-book3s.c struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); cpuhw 2269 arch/powerpc/perf/core-book3s.c memset(cpuhw, 0, sizeof(*cpuhw)); cpuhw 2270 arch/powerpc/perf/core-book3s.c cpuhw->mmcr[0] = MMCR0_FC; cpuhw 205 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw; cpuhw 209 arch/powerpc/perf/core-fsl-emb.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 211 arch/powerpc/perf/core-fsl-emb.c if (!cpuhw->disabled) { cpuhw 212 arch/powerpc/perf/core-fsl-emb.c cpuhw->disabled = 1; cpuhw 217 arch/powerpc/perf/core-fsl-emb.c if (!cpuhw->pmcs_enabled) { cpuhw 219 arch/powerpc/perf/core-fsl-emb.c cpuhw->pmcs_enabled = 1; cpuhw 244 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw; cpuhw 248 arch/powerpc/perf/core-fsl-emb.c cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 249 arch/powerpc/perf/core-fsl-emb.c if (!cpuhw->disabled) cpuhw 252 arch/powerpc/perf/core-fsl-emb.c cpuhw->disabled = 0; cpuhw 253 arch/powerpc/perf/core-fsl-emb.c ppc_set_pmu_inuse(cpuhw->n_events != 0); cpuhw 255 arch/powerpc/perf/core-fsl-emb.c if (cpuhw->n_events > 0) { cpuhw 291 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw; cpuhw 298 arch/powerpc/perf/core-fsl-emb.c cpuhw = &get_cpu_var(cpu_hw_events); cpuhw 308 arch/powerpc/perf/core-fsl-emb.c if (cpuhw->event[i]) cpuhw 318 arch/powerpc/perf/core-fsl-emb.c cpuhw->event[i] = event; cpuhw 319 arch/powerpc/perf/core-fsl-emb.c ++cpuhw->n_events; cpuhw 352 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw; cpuhw 361 arch/powerpc/perf/core-fsl-emb.c cpuhw = &get_cpu_var(cpu_hw_events); cpuhw 363 arch/powerpc/perf/core-fsl-emb.c WARN_ON(event != cpuhw->event[event->hw.idx]); cpuhw 369 arch/powerpc/perf/core-fsl-emb.c cpuhw->event[i] = NULL; cpuhw 380 arch/powerpc/perf/core-fsl-emb.c cpuhw->n_events--; cpuhw 658 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 671 arch/powerpc/perf/core-fsl-emb.c event = cpuhw->event[i]; cpuhw 702 arch/powerpc/perf/core-fsl-emb.c struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); cpuhw 704 arch/powerpc/perf/core-fsl-emb.c memset(cpuhw, 0, sizeof(*cpuhw)); cpuhw 39 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw; cpuhw 43 arch/s390/kernel/perf_cpum_cf.c cpuhw = &get_cpu_var(cpu_cf_events); cpuhw 49 arch/s390/kernel/perf_cpum_cf.c if (cpuhw->info.cfvn < 1) cpuhw 53 arch/s390/kernel/perf_cpum_cf.c if ((cpuhw->info.csvn >= 1 && cpuhw->info.csvn <= 5 && cpuhw 55 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.csvn >= 6 && hwc->config > 83)) cpuhw 59 arch/s390/kernel/perf_cpum_cf.c if (cpuhw->info.csvn < 1) cpuhw 61 arch/s390/kernel/perf_cpum_cf.c if ((cpuhw->info.csvn == 1 && hwc->config > 159) || cpuhw 62 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.csvn == 2 && hwc->config > 175) || cpuhw 63 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.csvn >= 3 && cpuhw->info.csvn <= 5 cpuhw 65 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.csvn >= 6 && hwc->config > 287)) cpuhw 69 arch/s390/kernel/perf_cpum_cf.c if (cpuhw->info.csvn <= 3) cpuhw 84 arch/s390/kernel/perf_cpum_cf.c if (!((cpuhw->info.auth_ctl & mtdiag_ctl) && cpuhw 85 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.enable_ctl & mtdiag_ctl) && cpuhw 86 arch/s390/kernel/perf_cpum_cf.c (cpuhw->info.act_ctl & mtdiag_ctl))) cpuhw 97 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw; cpuhw 101 arch/s390/kernel/perf_cpum_cf.c cpuhw = &get_cpu_var(cpu_cf_events); cpuhw 109 arch/s390/kernel/perf_cpum_cf.c if (!(ctrs_state & cpuhw->info.auth_ctl)) cpuhw 123 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 126 arch/s390/kernel/perf_cpum_cf.c if (cpuhw->flags & PMU_F_ENABLED) cpuhw 129 arch/s390/kernel/perf_cpum_cf.c err = lcctl(cpuhw->state); cpuhw 136 arch/s390/kernel/perf_cpum_cf.c cpuhw->flags |= PMU_F_ENABLED; cpuhw 146 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 150 arch/s390/kernel/perf_cpum_cf.c if (!(cpuhw->flags & PMU_F_ENABLED)) cpuhw 153 arch/s390/kernel/perf_cpum_cf.c inactive = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1); cpuhw 161 arch/s390/kernel/perf_cpum_cf.c cpuhw->flags &= ~PMU_F_ENABLED; cpuhw 364 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 379 arch/s390/kernel/perf_cpum_cf.c ctr_set_enable(&cpuhw->state, hwc->config_base); cpuhw 380 arch/s390/kernel/perf_cpum_cf.c ctr_set_start(&cpuhw->state, hwc->config_base); cpuhw 390 arch/s390/kernel/perf_cpum_cf.c atomic_inc(&cpuhw->ctr_set[hwc->config_base]); cpuhw 395 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 403 arch/s390/kernel/perf_cpum_cf.c if (!atomic_dec_return(&cpuhw->ctr_set[hwc->config_base])) cpuhw 404 arch/s390/kernel/perf_cpum_cf.c ctr_set_stop(&cpuhw->state, hwc->config_base); cpuhw 416 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 423 arch/s390/kernel/perf_cpum_cf.c if (!(cpuhw->txn_flags & PERF_PMU_TXN_ADD)) cpuhw 427 arch/s390/kernel/perf_cpum_cf.c ctr_set_enable(&cpuhw->state, event->hw.config_base); cpuhw 440 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 452 arch/s390/kernel/perf_cpum_cf.c if (!atomic_read(&cpuhw->ctr_set[event->hw.config_base])) cpuhw 453 arch/s390/kernel/perf_cpum_cf.c ctr_set_disable(&cpuhw->state, event->hw.config_base); cpuhw 468 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 470 arch/s390/kernel/perf_cpum_cf.c WARN_ON_ONCE(cpuhw->txn_flags); /* txn already in flight */ cpuhw 472 arch/s390/kernel/perf_cpum_cf.c cpuhw->txn_flags = txn_flags; cpuhw 477 arch/s390/kernel/perf_cpum_cf.c cpuhw->tx_state = cpuhw->state; cpuhw 488 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 490 arch/s390/kernel/perf_cpum_cf.c WARN_ON_ONCE(!cpuhw->txn_flags); /* no txn in flight */ cpuhw 492 arch/s390/kernel/perf_cpum_cf.c txn_flags = cpuhw->txn_flags; cpuhw 493 arch/s390/kernel/perf_cpum_cf.c cpuhw->txn_flags = 0; cpuhw 497 arch/s390/kernel/perf_cpum_cf.c WARN_ON(cpuhw->tx_state != cpuhw->state); cpuhw 509 arch/s390/kernel/perf_cpum_cf.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 512 arch/s390/kernel/perf_cpum_cf.c WARN_ON_ONCE(!cpuhw->txn_flags); /* no txn in flight */ cpuhw 514 arch/s390/kernel/perf_cpum_cf.c if (cpuhw->txn_flags & ~PERF_PMU_TXN_ADD) { cpuhw 515 arch/s390/kernel/perf_cpum_cf.c cpuhw->txn_flags = 0; cpuhw 520 arch/s390/kernel/perf_cpum_cf.c state = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1); cpuhw 522 arch/s390/kernel/perf_cpum_cf.c if ((state & cpuhw->info.auth_ctl) != state) cpuhw 525 arch/s390/kernel/perf_cpum_cf.c cpuhw->txn_flags = 0; cpuhw 42 arch/s390/kernel/perf_cpum_cf_common.c struct cpu_cf_events *cpuhw; cpuhw 48 arch/s390/kernel/perf_cpum_cf_common.c cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 52 arch/s390/kernel/perf_cpum_cf_common.c if (!(cpuhw->flags & PMU_F_RESERVED)) cpuhw 57 arch/s390/kernel/perf_cpum_cf_common.c qctri(&cpuhw->info); cpuhw 69 arch/s390/kernel/perf_cpum_cf_common.c atomic64_or(alert, &cpuhw->alert); cpuhw 76 arch/s390/kernel/perf_cpum_cf_common.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 80 arch/s390/kernel/perf_cpum_cf_common.c memset(&cpuhw->info, 0, sizeof(cpuhw->info)); cpuhw 81 arch/s390/kernel/perf_cpum_cf_common.c qctri(&cpuhw->info); cpuhw 82 arch/s390/kernel/perf_cpum_cf_common.c cpuhw->flags |= PMU_F_RESERVED; cpuhw 86 arch/s390/kernel/perf_cpum_cf_common.c cpuhw->flags &= ~PMU_F_RESERVED; cpuhw 130 arch/s390/kernel/perf_cpum_cf_common.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 133 arch/s390/kernel/perf_cpum_cf_common.c alert = atomic64_read(&cpuhw->alert); cpuhw 135 arch/s390/kernel/perf_cpum_cf_common.c atomic64_set(&cpuhw->alert, 0); cpuhw 109 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 112 arch/s390/kernel/perf_cpum_cf_diag.c te->cfvn = cpuhw->info.cfvn; /* Counter version numbers */ cpuhw 113 arch/s390/kernel/perf_cpum_cf_diag.c te->csvn = cpuhw->info.csvn; cpuhw 132 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 137 arch/s390/kernel/perf_cpum_cf_diag.c __func__, pmu, smp_processor_id(), cpuhw->flags, cpuhw 138 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->state); cpuhw 139 arch/s390/kernel/perf_cpum_cf_diag.c if (cpuhw->flags & PMU_F_ENABLED) cpuhw 142 arch/s390/kernel/perf_cpum_cf_diag.c err = lcctl(cpuhw->state); cpuhw 148 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->flags |= PMU_F_ENABLED; cpuhw 158 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 164 arch/s390/kernel/perf_cpum_cf_diag.c __func__, pmu, smp_processor_id(), cpuhw->flags, cpuhw 165 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->state); cpuhw 166 arch/s390/kernel/perf_cpum_cf_diag.c if (!(cpuhw->flags & PMU_F_ENABLED)) cpuhw 169 arch/s390/kernel/perf_cpum_cf_diag.c inactive = cpuhw->state & ~((1 << CPUMF_LCCTL_ENABLE_SHIFT) - 1); cpuhw 176 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->flags &= ~PMU_F_ENABLED; cpuhw 200 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw; cpuhw 218 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw = &get_cpu_var(cpu_cf_events); cpuhw 220 arch/s390/kernel/perf_cpum_cf_diag.c if (cpuhw->info.auth_ctl & cpumf_ctr_ctl[i]) cpuhw 371 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 378 arch/s390/kernel/perf_cpum_cf_diag.c ctrset_size = cf_diag_ctrset_size(ctrset, &cpuhw->info); cpuhw 394 arch/s390/kernel/perf_cpum_cf_diag.c __func__, ctrset, ctrset_size, cpuhw->info.cfvn, cpuhw 395 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->info.csvn, need, rc); cpuhw 523 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 536 arch/s390/kernel/perf_cpum_cf_diag.c ctr_set_multiple_enable(&cpuhw->state, hwc->config_base); cpuhw 537 arch/s390/kernel/perf_cpum_cf_diag.c lcctl(cpuhw->state); /* Enable counter sets */ cpuhw 540 arch/s390/kernel/perf_cpum_cf_diag.c ctr_set_multiple_start(&cpuhw->state, hwc->config_base); cpuhw 546 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 555 arch/s390/kernel/perf_cpum_cf_diag.c ctr_set_multiple_stop(&cpuhw->state, hwc->config_base); cpuhw 566 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 571 arch/s390/kernel/perf_cpum_cf_diag.c __func__, event, event->cpu, flags, cpuhw); cpuhw 573 arch/s390/kernel/perf_cpum_cf_diag.c if (cpuhw->flags & PMU_F_IN_USE) { cpuhw 580 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->flags |= PMU_F_IN_USE; cpuhw 590 arch/s390/kernel/perf_cpum_cf_diag.c struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); cpuhw 597 arch/s390/kernel/perf_cpum_cf_diag.c ctr_set_multiple_stop(&cpuhw->state, event->hw.config_base); cpuhw 598 arch/s390/kernel/perf_cpum_cf_diag.c ctr_set_multiple_disable(&cpuhw->state, event->hw.config_base); cpuhw 599 arch/s390/kernel/perf_cpum_cf_diag.c cpuhw->flags &= ~PMU_F_IN_USE; cpuhw 116 arch/s390/kernel/perf_cpum_sf.c static int sf_buffer_available(struct cpu_hw_sf *cpuhw) cpuhw 118 arch/s390/kernel/perf_cpum_sf.c return !!cpuhw->sfb.sdbt; cpuhw 365 arch/s390/kernel/perf_cpum_sf.c static void deallocate_buffers(struct cpu_hw_sf *cpuhw) cpuhw 367 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->sfb.sdbt) cpuhw 368 arch/s390/kernel/perf_cpum_sf.c free_sampling_buffer(&cpuhw->sfb); cpuhw 371 arch/s390/kernel/perf_cpum_sf.c static int allocate_buffers(struct cpu_hw_sf *cpuhw, struct hw_perf_event *hwc) cpuhw 401 arch/s390/kernel/perf_cpum_sf.c freq = sample_rate_to_freq(&cpuhw->qsi, SAMPL_RATE(hwc)); cpuhw 416 arch/s390/kernel/perf_cpum_sf.c if (sf_buffer_available(cpuhw)) cpuhw 423 arch/s390/kernel/perf_cpum_sf.c sample_size, cpuhw); cpuhw 425 arch/s390/kernel/perf_cpum_sf.c return alloc_sampling_buffer(&cpuhw->sfb, cpuhw 426 arch/s390/kernel/perf_cpum_sf.c sfb_pending_allocs(&cpuhw->sfb, hwc)); cpuhw 458 arch/s390/kernel/perf_cpum_sf.c static void sfb_account_overflows(struct cpu_hw_sf *cpuhw, cpuhw 473 arch/s390/kernel/perf_cpum_sf.c ratio = DIV_ROUND_UP(100 * OVERFLOW_REG(hwc) * cpuhw->sfb.num_sdb, cpuhw 474 arch/s390/kernel/perf_cpum_sf.c sample_rate_to_freq(&cpuhw->qsi, SAMPL_RATE(hwc))); cpuhw 477 arch/s390/kernel/perf_cpum_sf.c num = compute_sfb_extent(ratio, cpuhw->sfb.num_sdb); cpuhw 768 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw; cpuhw 799 arch/s390/kernel/perf_cpum_sf.c cpuhw = NULL; cpuhw 806 arch/s390/kernel/perf_cpum_sf.c cpuhw = &per_cpu(cpu_hw_sf, event->cpu); cpuhw 807 arch/s390/kernel/perf_cpum_sf.c si = cpuhw->qsi; cpuhw 860 arch/s390/kernel/perf_cpum_sf.c if (cpuhw) cpuhw 862 arch/s390/kernel/perf_cpum_sf.c err = allocate_buffers(cpuhw, hwc); cpuhw 868 arch/s390/kernel/perf_cpum_sf.c cpuhw = &per_cpu(cpu_hw_sf, cpu); cpuhw 869 arch/s390/kernel/perf_cpum_sf.c err = allocate_buffers(cpuhw, hwc); cpuhw 936 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 940 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->flags & PMU_F_ENABLED) cpuhw 943 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->flags & PMU_F_ERR_MASK) cpuhw 957 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->event) { cpuhw 958 arch/s390/kernel/perf_cpum_sf.c hwc = &cpuhw->event->hw; cpuhw 964 arch/s390/kernel/perf_cpum_sf.c sfb_account_overflows(cpuhw, hwc); cpuhw 965 arch/s390/kernel/perf_cpum_sf.c if (sfb_has_pending_allocs(&cpuhw->sfb, hwc)) cpuhw 966 arch/s390/kernel/perf_cpum_sf.c extend_sampling_buffer(&cpuhw->sfb, hwc); cpuhw 969 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.interval = SAMPL_RATE(&cpuhw->event->hw); cpuhw 973 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags |= PMU_F_ENABLED; cpuhw 976 arch/s390/kernel/perf_cpum_sf.c err = lsctl(&cpuhw->lsctl); cpuhw 978 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags &= ~PMU_F_ENABLED; cpuhw 989 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.es, cpuhw->lsctl.cs, cpuhw->lsctl.ed, cpuhw 990 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.cd, cpuhw->lsctl.interval, cpuhw 991 arch/s390/kernel/perf_cpum_sf.c (void *) cpuhw->lsctl.tear, cpuhw 992 arch/s390/kernel/perf_cpum_sf.c (void *) cpuhw->lsctl.dear); cpuhw 997 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 1002 arch/s390/kernel/perf_cpum_sf.c if (!(cpuhw->flags & PMU_F_ENABLED)) cpuhw 1005 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->flags & PMU_F_ERR_MASK) cpuhw 1009 arch/s390/kernel/perf_cpum_sf.c inactive = cpuhw->lsctl; cpuhw 1028 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.tear = si.tear; cpuhw 1029 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.dear = si.dear; cpuhw 1035 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags &= ~PMU_F_ENABLED; cpuhw 1408 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw) cpuhw 1449 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.tear = base + offset * sizeof(unsigned long); cpuhw 1450 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.dear = aux->sdb_index[head]; cpuhw 1459 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.tear, cpuhw 1460 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.dear); cpuhw 1574 arch/s390/kernel/perf_cpum_sf.c static void hw_collect_aux(struct cpu_hw_sf *cpuhw) cpuhw 1580 arch/s390/kernel/perf_cpum_sf.c struct perf_output_handle *handle = &cpuhw->handle; cpuhw 1595 arch/s390/kernel/perf_cpum_sf.c aux = perf_aux_output_begin(handle, cpuhw->event); cpuhw 1620 arch/s390/kernel/perf_cpum_sf.c perf_aux_output_end(&cpuhw->handle, size); cpuhw 1628 arch/s390/kernel/perf_cpum_sf.c perf_aux_output_end(&cpuhw->handle, size); cpuhw 1813 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = &per_cpu(cpu_hw_sf, event->cpu); cpuhw 1815 arch/s390/kernel/perf_cpum_sf.c si = cpuhw->qsi; cpuhw 1838 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 1848 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.cs = 1; cpuhw 1850 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.cd = 1; cpuhw 1859 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 1865 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.cs = 0; cpuhw 1866 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.cd = 0; cpuhw 1878 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 1882 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->flags & PMU_F_IN_USE) cpuhw 1885 arch/s390/kernel/perf_cpum_sf.c if (!SAMPL_DIAG_MODE(&event->hw) && !cpuhw->sfb.sdbt) cpuhw 1898 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.s = 0; cpuhw 1899 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.h = 1; cpuhw 1900 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.interval = SAMPL_RATE(&event->hw); cpuhw 1902 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.tear = (unsigned long) cpuhw->sfb.sdbt; cpuhw 1903 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.dear = *(unsigned long *) cpuhw->sfb.sdbt; cpuhw 1904 arch/s390/kernel/perf_cpum_sf.c hw_reset_registers(&event->hw, cpuhw->sfb.sdbt); cpuhw 1909 arch/s390/kernel/perf_cpum_sf.c if (WARN_ON_ONCE(cpuhw->lsctl.es == 1 || cpuhw->lsctl.ed == 1)) { cpuhw 1914 arch/s390/kernel/perf_cpum_sf.c aux = perf_aux_output_begin(&cpuhw->handle, event); cpuhw 1919 arch/s390/kernel/perf_cpum_sf.c err = aux_output_begin(&cpuhw->handle, aux, cpuhw); cpuhw 1922 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.ed = 1; cpuhw 1924 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.es = 1; cpuhw 1927 arch/s390/kernel/perf_cpum_sf.c cpuhw->event = event; cpuhw 1928 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags |= PMU_F_IN_USE; cpuhw 1940 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 1945 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.es = 0; cpuhw 1946 arch/s390/kernel/perf_cpum_sf.c cpuhw->lsctl.ed = 0; cpuhw 1947 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags &= ~PMU_F_IN_USE; cpuhw 1948 arch/s390/kernel/perf_cpum_sf.c cpuhw->event = NULL; cpuhw 1951 arch/s390/kernel/perf_cpum_sf.c aux_output_end(&cpuhw->handle); cpuhw 2031 arch/s390/kernel/perf_cpum_sf.c struct cpu_hw_sf *cpuhw; cpuhw 2036 arch/s390/kernel/perf_cpum_sf.c cpuhw = this_cpu_ptr(&cpu_hw_sf); cpuhw 2040 arch/s390/kernel/perf_cpum_sf.c if (!(cpuhw->flags & PMU_F_RESERVED)) cpuhw 2048 arch/s390/kernel/perf_cpum_sf.c if (cpuhw->flags & PMU_F_IN_USE) cpuhw 2049 arch/s390/kernel/perf_cpum_sf.c if (SAMPL_DIAG_MODE(&cpuhw->event->hw)) cpuhw 2050 arch/s390/kernel/perf_cpum_sf.c hw_collect_aux(cpuhw); cpuhw 2052 arch/s390/kernel/perf_cpum_sf.c hw_perf_event_update(cpuhw->event, 0); cpuhw 2054 arch/s390/kernel/perf_cpum_sf.c WARN_ON_ONCE(!(cpuhw->flags & PMU_F_IN_USE)); cpuhw 2064 arch/s390/kernel/perf_cpum_sf.c qsi(&cpuhw->qsi); cpuhw 2069 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags |= PMU_F_ERR_LSDA; cpuhw 2077 arch/s390/kernel/perf_cpum_sf.c cpuhw->flags |= PMU_F_ERR_IBE; cpuhw 354 arch/sh/kernel/perf_event.c struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); cpuhw 356 arch/sh/kernel/perf_event.c memset(cpuhw, 0, sizeof(struct cpu_hw_events)); cpuhw 1514 arch/sparc/kernel/perf_event.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1516 arch/sparc/kernel/perf_event.c WARN_ON_ONCE(cpuhw->txn_flags); /* txn already in flight */ cpuhw 1518 arch/sparc/kernel/perf_event.c cpuhw->txn_flags = txn_flags; cpuhw 1532 arch/sparc/kernel/perf_event.c struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); cpuhw 1535 arch/sparc/kernel/perf_event.c WARN_ON_ONCE(!cpuhw->txn_flags); /* no txn in flight */ cpuhw 1537 arch/sparc/kernel/perf_event.c txn_flags = cpuhw->txn_flags; cpuhw 1538 arch/sparc/kernel/perf_event.c cpuhw->txn_flags = 0; cpuhw 555 arch/x86/events/amd/core.c struct cpu_hw_events *cpuhw; cpuhw 560 arch/x86/events/amd/core.c cpuhw = &per_cpu(cpu_hw_events, cpu); cpuhw 562 arch/x86/events/amd/core.c if (cpuhw->amd_nb) { cpuhw 563 arch/x86/events/amd/core.c struct amd_nb *nb = cpuhw->amd_nb; cpuhw 568 arch/x86/events/amd/core.c cpuhw->amd_nb = NULL;