/linux-4.1.27/arch/s390/kernel/ |
D | perf_cpum_cf.c | 70 struct cpu_hw_events { struct 76 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { argument 125 struct cpu_hw_events *cpuhw; in validate_ctr_version() 128 cpuhw = &get_cpu_var(cpu_hw_events); in validate_ctr_version() 148 put_cpu_var(cpu_hw_events); in validate_ctr_version() 154 struct cpu_hw_events *cpuhw; in validate_ctr_auth() 158 cpuhw = &get_cpu_var(cpu_hw_events); in validate_ctr_auth() 165 put_cpu_var(cpu_hw_events); in validate_ctr_auth() 176 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_enable() 199 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_disable() [all …]
|
/linux-4.1.27/arch/sh/kernel/ |
D | perf_event.c | 31 struct cpu_hw_events { struct 37 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 222 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop() 240 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start() 257 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del() 267 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add() 357 struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); in sh_pmu_setup() 359 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); in sh_pmu_setup()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | perf_event.c | 230 struct cpu_hw_events { struct 234 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 303 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_stop() 321 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_start() 338 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_del() 348 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_add() 424 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_enable() 458 struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); in bfin_pmu_setup() 460 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); in bfin_pmu_setup()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | perf_event_amd.c | 207 static inline int amd_has_nb(struct cpu_hw_events *cpuc) in amd_has_nb() 235 static void __amd_put_nb_event_constraints(struct cpu_hw_events *cpuc, in __amd_put_nb_event_constraints() 292 __amd_get_nb_event_constraints(struct cpu_hw_events *cpuc, struct perf_event *event, in __amd_get_nb_event_constraints() 368 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_prepare() 384 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_starting() 398 nb = per_cpu(cpu_hw_events, i).amd_nb; in amd_pmu_cpu_starting() 415 struct cpu_hw_events *cpuhw; in amd_pmu_cpu_dead() 420 cpuhw = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_dead() 433 amd_get_event_constraints(struct cpu_hw_events *cpuc, int idx, in amd_get_event_constraints() 445 static void amd_put_event_constraints(struct cpu_hw_events *cpuc, in amd_put_event_constraints() [all …]
|
D | perf_event_intel_ds.c | 229 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in init_debug_store_on_cpu() 241 if (!per_cpu(cpu_hw_events, cpu).ds) in fini_debug_store_on_cpu() 251 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in alloc_pebs_buffer() 291 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_pebs_buffer() 305 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in alloc_bts_buffer() 334 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_bts_buffer() 352 per_cpu(cpu_hw_events, cpu).ds = ds; in alloc_ds_buffer() 359 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_ds_buffer() 364 per_cpu(cpu_hw_events, cpu).ds = NULL; in release_ds_buffer() 478 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_bts() [all …]
|
D | perf_event_intel_lbr.c | 128 static void intel_pmu_lbr_filter(struct cpu_hw_events *cpuc); 137 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_enable() 262 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_sched_task() 309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable() 337 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable() 362 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable_all() 370 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable_all() 376 static void intel_pmu_lbr_read_32(struct cpu_hw_events *cpuc) in intel_pmu_lbr_read_32() 408 static void intel_pmu_lbr_read_64(struct cpu_hw_events *cpuc) in intel_pmu_lbr_read_64() 462 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_read() [all …]
|
D | perf_event.c | 44 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { 575 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable_all() 593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable() 610 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable_all() 813 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in x86_schedule_events() 929 static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader, bool dogrp) in collect_events() 963 struct cpu_hw_events *cpuc, int i) in x86_assign_hw_event() 986 struct cpu_hw_events *cpuc, in match_prev_assignment() 998 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable() 1139 if (__this_cpu_read(cpu_hw_events.enabled)) in x86_pmu_enable_event() [all …]
|
D | perf_event_intel.c | 1251 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_disable_all() 1271 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_enable_all() 1311 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_nhm_workaround() 1410 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_event() 1474 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_enable_event() 1477 if (!__this_cpu_read(cpu_hw_events.enabled)) in intel_pmu_enable_event() 1532 struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds); in intel_pmu_reset() 1575 struct cpu_hw_events *cpuc; in intel_pmu_handle_irq() 1580 cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_handle_irq() 1739 __intel_shared_reg_get_constraints(struct cpu_hw_events *cpuc, in __intel_shared_reg_get_constraints() [all …]
|
D | perf_event.h | 163 struct cpu_hw_events { struct 504 int (*schedule_events)(struct cpu_hw_events *cpuc, int n, int *assign); 523 (*get_event_constraints)(struct cpu_hw_events *cpuc, 527 void (*put_event_constraints)(struct cpu_hw_events *cpuc, 530 void (*commit_scheduling)(struct cpu_hw_events *cpuc, int idx, int cntr); 532 void (*start_scheduling)(struct cpu_hw_events *cpuc); 534 void (*stop_scheduling)(struct cpu_hw_events *cpuc); 660 DECLARE_PER_CPU(struct cpu_hw_events, cpu_hw_events); 721 u64 disable_mask = __this_cpu_read(cpu_hw_events.perf_ctr_virt_mask); in __x86_pmu_enable_event() 732 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign); [all …]
|
D | perf_event_intel_bts.c | 141 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_config_buffer() 192 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_update() 422 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_del() 447 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_add()
|
D | perf_event_knc.c | 215 struct cpu_hw_events *cpuc; in knc_pmu_handle_irq() 220 cpuc = this_cpu_ptr(&cpu_hw_events); in knc_pmu_handle_irq()
|
D | perf_event_p4.c | 918 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_disable_all() 987 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_enable_all() 1001 struct cpu_hw_events *cpuc; in p4_pmu_handle_irq() 1007 cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_handle_irq() 1206 static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in p4_pmu_schedule_events()
|
/linux-4.1.27/arch/metag/kernel/perf/ |
D | perf_event.c | 38 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); 261 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_start() 309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_add() 351 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_del() 600 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_enable_counter() 673 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_disable_counter() 721 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_write_counter() 754 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in metag_pmu_counter_overflow() 810 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in metag_pmu_cpu_notify() 815 memset(cpuc, 0, sizeof(struct cpu_hw_events)); in metag_pmu_cpu_notify() [all …]
|
D | perf_event.h | 50 struct cpu_hw_events { struct
|
/linux-4.1.27/arch/powerpc/perf/ |
D | core-fsl-emb.c | 23 struct cpu_hw_events { struct 29 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 209 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_disable() 213 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_disable() 248 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_enable() 252 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_enable() 295 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_add() 302 cpuhw = &get_cpu_var(cpu_hw_events); in fsl_emb_pmu_add() 348 put_cpu_var(cpu_hw_events); in fsl_emb_pmu_add() 356 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_del() [all …]
|
D | core-book3s.c | 29 struct cpu_hw_events { struct 62 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 120 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in() 128 static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} in power_pmu_bhrb_read() 351 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_enable() 367 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_disable() 425 static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) in power_pmu_bhrb_read() 564 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in() 828 static int power_check_constraints(struct cpu_hw_events *cpuhw, in power_check_constraints() 1065 static void freeze_limited_counters(struct cpu_hw_events *cpuhw, in freeze_limited_counters() [all …]
|
/linux-4.1.27/arch/alpha/kernel/ |
D | perf_event.c | 33 struct cpu_hw_events { struct 52 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument 390 static void maybe_change_configuration(struct cpu_hw_events *cpuc) in maybe_change_configuration() 434 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_add() 486 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_del() 534 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_stop() 554 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_start() 727 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_enable() 753 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_disable() 811 struct cpu_hw_events *cpuc; in alpha_perf_event_irq_handler() [all …]
|
/linux-4.1.27/arch/sparc/kernel/ |
D | perf_event.c | 76 struct cpu_hw_events { struct 113 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { .enabled = 1, }; argument 825 static inline void sparc_pmu_enable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, in… in sparc_pmu_enable_event() 843 static inline void sparc_pmu_disable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, i… in sparc_pmu_disable_event() 917 static void read_in_all_counters(struct cpu_hw_events *cpuc) in read_in_all_counters() 939 static void calculate_single_pcr(struct cpu_hw_events *cpuc) in calculate_single_pcr() 973 static void calculate_multiple_pcrs(struct cpu_hw_events *cpuc) in calculate_multiple_pcrs() 1004 static void update_pcrs_for_enable(struct cpu_hw_events *cpuc) in update_pcrs_for_enable() 1018 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_enable() 1036 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_disable() [all …]
|
/linux-4.1.27/arch/tile/kernel/ |
D | perf_event.c | 50 struct cpu_hw_events { struct 84 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); 593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_stop() 619 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_start() 653 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_add() 709 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_del() 883 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_handle_irq()
|
/linux-4.1.27/arch/arm/kernel/ |
D | perf_event_cpu.c | 199 struct pmu_hw_events __percpu *cpu_hw_events; in cpu_pmu_init() local 201 cpu_hw_events = alloc_percpu(struct pmu_hw_events); in cpu_pmu_init() 202 if (!cpu_hw_events) in cpu_pmu_init() 211 struct pmu_hw_events *events = per_cpu_ptr(cpu_hw_events, cpu); in cpu_pmu_init() 216 cpu_pmu->hw_events = cpu_hw_events; in cpu_pmu_init() 231 free_percpu(cpu_hw_events); in cpu_pmu_init()
|
/linux-4.1.27/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 34 struct cpu_hw_events { struct 51 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { argument 311 static int mipsxx_pmu_alloc_counter(struct cpu_hw_events *cpuc, in mipsxx_pmu_alloc_counter() 343 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_enable_event() 363 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_disable_event() 463 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_add() 499 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_del() 719 struct cpu_hw_events fake_cpuc; in validate_group() 738 static void handle_associated_event(struct cpu_hw_events *cpuc, in handle_associated_event() 1342 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in pause_local_counters() [all …]
|
/linux-4.1.27/arch/arm64/kernel/ |
D | perf_event.c | 49 static DEFINE_PER_CPU(struct pmu_hw_events, cpu_hw_events); 407 free_percpu_irq(irq, &cpu_hw_events); in armpmu_release_hardware() 457 "arm-pmu", &cpu_hw_events); in armpmu_reserve_hardware() 1122 cpuc = this_cpu_ptr(&cpu_hw_events); in armv8pmu_handle_irq() 1376 return this_cpu_ptr(&cpu_hw_events); in armpmu_get_cpu_events() 1383 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); in cpu_pmu_init()
|