ppmu 66 arch/powerpc/perf/core-book3s.c static struct power_pmu *ppmu; ppmu 136 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 165 arch/powerpc/perf/core-book3s.c if ((ppmu->flags & PPMU_HAS_SSLOT) && (mmcra & MMCRA_SAMPLE_ENABLE)) { ppmu 187 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 192 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_SIAR_VALID) ppmu 194 arch/powerpc/perf/core-book3s.c else if (ppmu->flags & PPMU_ALT_SIPR) ppmu 196 arch/powerpc/perf/core-book3s.c else if (ppmu->flags & PPMU_NO_SIAR) ppmu 216 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 219 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ALT_SIPR) ppmu 229 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 232 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ALT_SIPR) ppmu 260 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_NO_SIPR) { ppmu 292 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 314 arch/powerpc/perf/core-book3s.c else if ((ppmu->flags & PPMU_NO_SIAR)) ppmu 318 arch/powerpc/perf/core-book3s.c else if ((ppmu->flags & PPMU_NO_CONT_SAMPLING)) ppmu 320 arch/powerpc/perf/core-book3s.c else if (!(ppmu->flags & PPMU_NO_SIPR) && regs_sipr(regs)) ppmu 350 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 353 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_SIAR_VALID) ppmu 371 arch/powerpc/perf/core-book3s.c if (!ppmu->bhrb_nr) ppmu 387 arch/powerpc/perf/core-book3s.c if (!ppmu->bhrb_nr) ppmu 409 arch/powerpc/perf/core-book3s.c if (!ppmu->bhrb_nr) ppmu 455 arch/powerpc/perf/core-book3s.c while (r_index < ppmu->bhrb_nr) { ppmu 539 arch/powerpc/perf/core-book3s.c return (ppmu->flags & PPMU_ARCH_207S) && ppmu 811 arch/powerpc/perf/core-book3s.c if (!ppmu) { ppmu 816 arch/powerpc/perf/core-book3s.c if (!ppmu->n_counter) ppmu 822 arch/powerpc/perf/core-book3s.c smp_processor_id(), ppmu->name, ppmu->n_counter); ppmu 824 arch/powerpc/perf/core-book3s.c for (i = 0; i < ppmu->n_counter; i++) ppmu 833 arch/powerpc/perf/core-book3s.c if (ppmu->n_counter > 4) ppmu 844 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_HAS_SIER) ppmu 847 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ARCH_207S) { ppmu 874 arch/powerpc/perf/core-book3s.c unsigned long addf = ppmu->add_fields; ppmu 875 arch/powerpc/perf/core-book3s.c unsigned long tadd = ppmu->test_adder; ppmu 876 arch/powerpc/perf/core-book3s.c unsigned long grp_mask = ppmu->group_constraint_mask; ppmu 877 arch/powerpc/perf/core-book3s.c unsigned long grp_val = ppmu->group_constraint_val; ppmu 879 arch/powerpc/perf/core-book3s.c if (n_ev > ppmu->n_counter) ppmu 885 arch/powerpc/perf/core-book3s.c && !ppmu->limited_pmc_event(event_id[i])) { ppmu 886 arch/powerpc/perf/core-book3s.c ppmu->get_alternatives(event_id[i], cflags[i], ppmu 890 arch/powerpc/perf/core-book3s.c if (ppmu->get_constraint(event_id[i], &cpuhw->amasks[i][0], ppmu 917 arch/powerpc/perf/core-book3s.c if (!ppmu->get_alternatives) ppmu 921 arch/powerpc/perf/core-book3s.c n_alt[i] = ppmu->get_alternatives(event_id[i], cflags[i], ppmu 924 arch/powerpc/perf/core-book3s.c ppmu->get_constraint(cpuhw->alternatives[i][j], ppmu 1000 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ARCH_207S) ppmu 1109 arch/powerpc/perf/core-book3s.c return (ppmu->flags & PPMU_LIMITED_PMC5_6) ppmu 1206 arch/powerpc/perf/core-book3s.c if (!ppmu) ppmu 1259 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ARCH_207S) { ppmu 1287 arch/powerpc/perf/core-book3s.c if (!ppmu) ppmu 1326 arch/powerpc/perf/core-book3s.c if (ppmu->compute_mmcr(cpuhw->events, cpuhw->n_events, hwc_index, ppmu 1333 arch/powerpc/perf/core-book3s.c if (!(ppmu->flags & PPMU_ARCH_207S)) { ppmu 1358 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_ARCH_207S) ppmu 1419 arch/powerpc/perf/core-book3s.c ppmu->config_bhrb(cpuhw->bhrb_filter); ppmu 1485 arch/powerpc/perf/core-book3s.c if (n0 >= ppmu->n_counter) ppmu 1526 arch/powerpc/perf/core-book3s.c cpuhw->bhrb_filter = ppmu->bhrb_filter_map( ppmu 1558 arch/powerpc/perf/core-book3s.c ppmu->disable_pmc(event->hw.idx - 1, cpuhw->mmcr); ppmu 1701 arch/powerpc/perf/core-book3s.c if (!ppmu) ppmu 1746 arch/powerpc/perf/core-book3s.c if (ppmu->limited_pmc_event(ev)) ppmu 1753 arch/powerpc/perf/core-book3s.c if (!ppmu->get_alternatives) ppmu 1757 arch/powerpc/perf/core-book3s.c n = ppmu->get_alternatives(ev, flags, alt); ppmu 1773 arch/powerpc/perf/core-book3s.c n = ppmu->get_alternatives(ev, flags, alt); ppmu 1805 arch/powerpc/perf/core-book3s.c if (!ppmu->cache_events) ppmu 1818 arch/powerpc/perf/core-book3s.c ev = (*ppmu->cache_events)[type][op][result]; ppmu 1831 arch/powerpc/perf/core-book3s.c for (i=0; i < ppmu->n_blacklist_ev; i++) { ppmu 1832 arch/powerpc/perf/core-book3s.c if (ppmu->blacklist_ev[i] == ev) ppmu 1851 arch/powerpc/perf/core-book3s.c if (!ppmu) ppmu 1856 arch/powerpc/perf/core-book3s.c if (!(ppmu->flags & PPMU_ARCH_207S)) ppmu 1863 arch/powerpc/perf/core-book3s.c if (ev >= ppmu->n_generic || ppmu->generic_events[ev] == 0) ppmu 1866 arch/powerpc/perf/core-book3s.c if (ppmu->blacklist_ev && is_event_blacklisted(ev)) ppmu 1868 arch/powerpc/perf/core-book3s.c ev = ppmu->generic_events[ev]; ppmu 1875 arch/powerpc/perf/core-book3s.c if (ppmu->blacklist_ev && is_event_blacklisted(ev)) ppmu 1881 arch/powerpc/perf/core-book3s.c if (ppmu->blacklist_ev && is_event_blacklisted(ev)) ppmu 1913 arch/powerpc/perf/core-book3s.c if (ppmu->flags & PPMU_LIMITED_PMC5_6) { ppmu 1916 arch/powerpc/perf/core-book3s.c } else if (ppmu->limited_pmc_event(ev)) { ppmu 1940 arch/powerpc/perf/core-book3s.c n = collect_events(event->group_leader, ppmu->n_counter - 1, ppmu 1955 arch/powerpc/perf/core-book3s.c bhrb_filter = ppmu->bhrb_filter_map( ppmu 2100 arch/powerpc/perf/core-book3s.c ppmu->get_mem_data_src) ppmu 2101 arch/powerpc/perf/core-book3s.c ppmu->get_mem_data_src(&data.data_src, ppmu->flags, regs); ppmu 2104 arch/powerpc/perf/core-book3s.c ppmu->get_mem_weight) ppmu 2105 arch/powerpc/perf/core-book3s.c ppmu->get_mem_weight(&data.weight); ppmu 2194 arch/powerpc/perf/core-book3s.c for (i = 0; i < ppmu->n_counter; ++i) ppmu 2199 arch/powerpc/perf/core-book3s.c for (i = 0; i < ppmu->n_counter; ++i) { ppmu 2268 arch/powerpc/perf/core-book3s.c if (ppmu) { ppmu 2277 arch/powerpc/perf/core-book3s.c if (ppmu) ppmu 2280 arch/powerpc/perf/core-book3s.c ppmu = pmu; ppmu 2284 arch/powerpc/perf/core-book3s.c power_pmu.attr_groups = ppmu->attr_groups; ppmu 27 arch/powerpc/perf/core-fsl-emb.c static struct fsl_emb_pmu *ppmu; ppmu 293 arch/powerpc/perf/core-fsl-emb.c int num_counters = ppmu->n_counter; ppmu 301 arch/powerpc/perf/core-fsl-emb.c num_counters = ppmu->n_restricted; ppmu 460 arch/powerpc/perf/core-fsl-emb.c if (!ppmu->cache_events) ppmu 473 arch/powerpc/perf/core-fsl-emb.c ev = (*ppmu->cache_events)[type][op][result]; ppmu 491 arch/powerpc/perf/core-fsl-emb.c if (ppmu->n_counter > MAX_HWEVENTS) { ppmu 493 arch/powerpc/perf/core-fsl-emb.c ppmu->n_counter, MAX_HWEVENTS); ppmu 494 arch/powerpc/perf/core-fsl-emb.c ppmu->n_counter = MAX_HWEVENTS; ppmu 500 arch/powerpc/perf/core-fsl-emb.c if (ev >= ppmu->n_generic || ppmu->generic_events[ev] == 0) ppmu 502 arch/powerpc/perf/core-fsl-emb.c ev = ppmu->generic_events[ev]; ppmu 519 arch/powerpc/perf/core-fsl-emb.c event->hw.config = ppmu->xlate_event(ev); ppmu 531 arch/powerpc/perf/core-fsl-emb.c ppmu->n_counter - 1, events); ppmu 543 arch/powerpc/perf/core-fsl-emb.c if (num_restricted >= ppmu->n_restricted) ppmu 670 arch/powerpc/perf/core-fsl-emb.c for (i = 0; i < ppmu->n_counter; ++i) { ppmu 709 arch/powerpc/perf/core-fsl-emb.c if (ppmu) ppmu 712 arch/powerpc/perf/core-fsl-emb.c ppmu = pmu; ppmu 41 drivers/devfreq/event/exynos-ppmu.c struct exynos_ppmu_data ppmu; ppmu 631 drivers/devfreq/event/exynos-ppmu.c info->ppmu.clk = devm_clk_get(dev, "ppmu"); ppmu 632 drivers/devfreq/event/exynos-ppmu.c if (IS_ERR(info->ppmu.clk)) { ppmu 633 drivers/devfreq/event/exynos-ppmu.c info->ppmu.clk = NULL; ppmu 689 drivers/devfreq/event/exynos-ppmu.c ret = clk_prepare_enable(info->ppmu.clk); ppmu 702 drivers/devfreq/event/exynos-ppmu.c clk_disable_unprepare(info->ppmu.clk); ppmu 175 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c int index, struct nvkm_pmu **ppmu) ppmu 178 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c if (!(pmu = *ppmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) ppmu 180 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c return nvkm_pmu_ctor(func, device, index, *ppmu); ppmu 59 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gf100.c gf100_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 61 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gf100.c return nvkm_pmu_new_(&gf100_pmu, device, index, ppmu); ppmu 43 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gf119.c gf119_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 45 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gf119.c return nvkm_pmu_new_(&gf119_pmu, device, index, ppmu); ppmu 123 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk104.c gk104_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 125 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk104.c return nvkm_pmu_new_(&gk104_pmu, device, index, ppmu); ppmu 102 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk110.c gk110_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 104 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk110.c return nvkm_pmu_new_(&gk110_pmu, device, index, ppmu); ppmu 44 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk208.c gk208_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 46 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk208.c return nvkm_pmu_new_(&gk208_pmu, device, index, ppmu); ppmu 206 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk20a.c gk20a_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 212 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gk20a.c *ppmu = &pmu->base; ppmu 45 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gm107.c gm107_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 47 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gm107.c return nvkm_pmu_new_(&gm107_pmu, device, index, ppmu); ppmu 47 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gm20b.c gm20b_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 51 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gm20b.c ret = nvkm_pmu_new_(&gm20b_pmu, device, index, ppmu); ppmu 33 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gp100.c gp100_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 35 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gp100.c return nvkm_pmu_new_(&gp100_pmu, device, index, ppmu); ppmu 47 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gp102.c gp102_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 49 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gp102.c return nvkm_pmu_new_(&gp102_pmu, device, index, ppmu); ppmu 260 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gt215.c gt215_pmu_new(struct nvkm_device *device, int index, struct nvkm_pmu **ppmu) ppmu 262 drivers/gpu/drm/nouveau/nvkm/subdev/pmu/gt215.c return nvkm_pmu_new_(>215_pmu, device, index, ppmu); ppmu 596 tools/perf/pmu-events/jevents.c const char *ppmu; ppmu 598 tools/perf/pmu-events/jevents.c ppmu = field_to_perf(unit_to_pmu, map, val); ppmu 599 tools/perf/pmu-events/jevents.c if (ppmu) { ppmu 600 tools/perf/pmu-events/jevents.c pmu = strdup(ppmu);