kvm_pmc           472 arch/x86/include/asm/kvm_host.h 	struct kvm_pmc gp_counters[INTEL_PMC_MAX_GENERIC];
kvm_pmc           473 arch/x86/include/asm/kvm_host.h 	struct kvm_pmc fixed_counters[INTEL_PMC_MAX_FIXED];
kvm_pmc            62 arch/x86/kvm/pmu.c 	struct kvm_pmc *pmc = perf_event->overflow_handler_context;
kvm_pmc            76 arch/x86/kvm/pmu.c 	struct kvm_pmc *pmc = perf_event->overflow_handler_context;
kvm_pmc            99 arch/x86/kvm/pmu.c static void pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type,
kvm_pmc           143 arch/x86/kvm/pmu.c void reprogram_gp_counter(struct kvm_pmc *pmc, u64 eventsel)
kvm_pmc           205 arch/x86/kvm/pmu.c void reprogram_fixed_counter(struct kvm_pmc *pmc, u8 ctrl, int idx)
kvm_pmc           237 arch/x86/kvm/pmu.c 	struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx);
kvm_pmc           262 arch/x86/kvm/pmu.c 		struct kvm_pmc *pmc = kvm_x86_ops->pmu_ops->pmc_idx_to_pmc(pmu, bit);
kvm_pmc           317 arch/x86/kvm/pmu.c 	struct kvm_pmc *pmc;
kvm_pmc            28 arch/x86/kvm/pmu.h 	bool (*pmc_is_enabled)(struct kvm_pmc *pmc);
kvm_pmc            29 arch/x86/kvm/pmu.h 	struct kvm_pmc *(*pmc_idx_to_pmc)(struct kvm_pmu *pmu, int pmc_idx);
kvm_pmc            30 arch/x86/kvm/pmu.h 	struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, unsigned idx,
kvm_pmc            41 arch/x86/kvm/pmu.h static inline u64 pmc_bitmask(struct kvm_pmc *pmc)
kvm_pmc            48 arch/x86/kvm/pmu.h static inline u64 pmc_read_counter(struct kvm_pmc *pmc)
kvm_pmc            60 arch/x86/kvm/pmu.h static inline void pmc_stop_counter(struct kvm_pmc *pmc)
kvm_pmc            69 arch/x86/kvm/pmu.h static inline bool pmc_is_gp(struct kvm_pmc *pmc)
kvm_pmc            74 arch/x86/kvm/pmu.h static inline bool pmc_is_fixed(struct kvm_pmc *pmc)
kvm_pmc            79 arch/x86/kvm/pmu.h static inline bool pmc_is_enabled(struct kvm_pmc *pmc)
kvm_pmc            88 arch/x86/kvm/pmu.h static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr,
kvm_pmc           102 arch/x86/kvm/pmu.h static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr)
kvm_pmc           116 arch/x86/kvm/pmu.h void reprogram_gp_counter(struct kvm_pmc *pmc, u64 eventsel);
kvm_pmc           117 arch/x86/kvm/pmu.h void reprogram_fixed_counter(struct kvm_pmc *pmc, u8 ctrl, int fixed_idx);
kvm_pmc            98 arch/x86/kvm/pmu_amd.c static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr,
kvm_pmc           155 arch/x86/kvm/pmu_amd.c static bool amd_pmc_is_enabled(struct kvm_pmc *pmc)
kvm_pmc           160 arch/x86/kvm/pmu_amd.c static struct kvm_pmc *amd_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx)
kvm_pmc           187 arch/x86/kvm/pmu_amd.c static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *mask)
kvm_pmc           190 arch/x86/kvm/pmu_amd.c 	struct kvm_pmc *counters;
kvm_pmc           214 arch/x86/kvm/pmu_amd.c 	struct kvm_pmc *pmc;
kvm_pmc           235 arch/x86/kvm/pmu_amd.c 	struct kvm_pmc *pmc;
kvm_pmc           297 arch/x86/kvm/pmu_amd.c 		struct kvm_pmc *pmc = &pmu->gp_counters[i];
kvm_pmc            42 arch/x86/kvm/vmx/pmu_intel.c 		struct kvm_pmc *pmc;
kvm_pmc            98 arch/x86/kvm/vmx/pmu_intel.c static bool intel_pmc_is_enabled(struct kvm_pmc *pmc)
kvm_pmc           105 arch/x86/kvm/vmx/pmu_intel.c static struct kvm_pmc *intel_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx)
kvm_pmc           129 arch/x86/kvm/vmx/pmu_intel.c static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu,
kvm_pmc           134 arch/x86/kvm/vmx/pmu_intel.c 	struct kvm_pmc *counters;
kvm_pmc           176 arch/x86/kvm/vmx/pmu_intel.c 	struct kvm_pmc *pmc;
kvm_pmc           212 arch/x86/kvm/vmx/pmu_intel.c 	struct kvm_pmc *pmc;
kvm_pmc           351 arch/x86/kvm/vmx/pmu_intel.c 	struct kvm_pmc *pmc = NULL;
kvm_pmc            25 include/kvm/arm_pmu.h 	struct kvm_pmc pmc[ARMV8_PMU_MAX_COUNTERS];
kvm_pmc            32 virt/kvm/arm/pmu.c static struct kvm_vcpu *kvm_pmc_to_vcpu(struct kvm_pmc *pmc)
kvm_pmc            47 virt/kvm/arm/pmu.c static bool kvm_pmu_pmc_is_chained(struct kvm_pmc *pmc)
kvm_pmc            70 virt/kvm/arm/pmu.c static struct kvm_pmc *kvm_pmu_get_canonical_pmc(struct kvm_pmc *pmc)
kvm_pmc           105 virt/kvm/arm/pmu.c 					  struct kvm_pmc *pmc)
kvm_pmc           143 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc = &pmu->pmc[select_idx];
kvm_pmc           178 virt/kvm/arm/pmu.c static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
kvm_pmc           194 virt/kvm/arm/pmu.c static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc)
kvm_pmc           286 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc;
kvm_pmc           327 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc;
kvm_pmc           445 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc = perf_event->overflow_handler_context;
kvm_pmc           567 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc;
kvm_pmc           649 virt/kvm/arm/pmu.c 	struct kvm_pmc *pmc = &pmu->pmc[select_idx];