Home
last modified time | relevance | path

Searched refs:cpu_hw_events (Results 1 – 22 of 22) sorted by relevance

/linux-4.1.27/arch/s390/kernel/
Dperf_cpum_cf.c70 struct cpu_hw_events { struct
76 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { argument
125 struct cpu_hw_events *cpuhw; in validate_ctr_version()
128 cpuhw = &get_cpu_var(cpu_hw_events); in validate_ctr_version()
148 put_cpu_var(cpu_hw_events); in validate_ctr_version()
154 struct cpu_hw_events *cpuhw; in validate_ctr_auth()
158 cpuhw = &get_cpu_var(cpu_hw_events); in validate_ctr_auth()
165 put_cpu_var(cpu_hw_events); in validate_ctr_auth()
176 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_enable()
199 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_disable()
[all …]
/linux-4.1.27/arch/sh/kernel/
Dperf_event.c31 struct cpu_hw_events { struct
37 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument
222 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop()
240 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start()
257 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del()
267 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add()
357 struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); in sh_pmu_setup()
359 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); in sh_pmu_setup()
/linux-4.1.27/arch/blackfin/kernel/
Dperf_event.c230 struct cpu_hw_events { struct
234 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument
303 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_stop()
321 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_start()
338 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_del()
348 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_add()
424 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_enable()
458 struct cpu_hw_events *cpuhw = &per_cpu(cpu_hw_events, cpu); in bfin_pmu_setup()
460 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); in bfin_pmu_setup()
/linux-4.1.27/arch/x86/kernel/cpu/
Dperf_event_amd.c207 static inline int amd_has_nb(struct cpu_hw_events *cpuc) in amd_has_nb()
235 static void __amd_put_nb_event_constraints(struct cpu_hw_events *cpuc, in __amd_put_nb_event_constraints()
292 __amd_get_nb_event_constraints(struct cpu_hw_events *cpuc, struct perf_event *event, in __amd_get_nb_event_constraints()
368 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_prepare()
384 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_starting()
398 nb = per_cpu(cpu_hw_events, i).amd_nb; in amd_pmu_cpu_starting()
415 struct cpu_hw_events *cpuhw; in amd_pmu_cpu_dead()
420 cpuhw = &per_cpu(cpu_hw_events, cpu); in amd_pmu_cpu_dead()
433 amd_get_event_constraints(struct cpu_hw_events *cpuc, int idx, in amd_get_event_constraints()
445 static void amd_put_event_constraints(struct cpu_hw_events *cpuc, in amd_put_event_constraints()
[all …]
Dperf_event_intel_ds.c229 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in init_debug_store_on_cpu()
241 if (!per_cpu(cpu_hw_events, cpu).ds) in fini_debug_store_on_cpu()
251 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in alloc_pebs_buffer()
291 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_pebs_buffer()
305 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in alloc_bts_buffer()
334 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_bts_buffer()
352 per_cpu(cpu_hw_events, cpu).ds = ds; in alloc_ds_buffer()
359 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in release_ds_buffer()
364 per_cpu(cpu_hw_events, cpu).ds = NULL; in release_ds_buffer()
478 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_bts()
[all …]
Dperf_event_intel_lbr.c128 static void intel_pmu_lbr_filter(struct cpu_hw_events *cpuc);
137 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_enable()
262 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_sched_task()
309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable()
337 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable()
362 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable_all()
370 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable_all()
376 static void intel_pmu_lbr_read_32(struct cpu_hw_events *cpuc) in intel_pmu_lbr_read_32()
408 static void intel_pmu_lbr_read_64(struct cpu_hw_events *cpuc) in intel_pmu_lbr_read_64()
462 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_read()
[all …]
Dperf_event.c44 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = {
575 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable_all()
593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable()
610 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable_all()
813 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in x86_schedule_events()
929 static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader, bool dogrp) in collect_events()
963 struct cpu_hw_events *cpuc, int i) in x86_assign_hw_event()
986 struct cpu_hw_events *cpuc, in match_prev_assignment()
998 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable()
1139 if (__this_cpu_read(cpu_hw_events.enabled)) in x86_pmu_enable_event()
[all …]
Dperf_event_intel.c1251 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_disable_all()
1271 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_enable_all()
1311 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_nhm_workaround()
1410 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_event()
1474 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_enable_event()
1477 if (!__this_cpu_read(cpu_hw_events.enabled)) in intel_pmu_enable_event()
1532 struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds); in intel_pmu_reset()
1575 struct cpu_hw_events *cpuc; in intel_pmu_handle_irq()
1580 cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_handle_irq()
1739 __intel_shared_reg_get_constraints(struct cpu_hw_events *cpuc, in __intel_shared_reg_get_constraints()
[all …]
Dperf_event.h163 struct cpu_hw_events { struct
504 int (*schedule_events)(struct cpu_hw_events *cpuc, int n, int *assign);
523 (*get_event_constraints)(struct cpu_hw_events *cpuc,
527 void (*put_event_constraints)(struct cpu_hw_events *cpuc,
530 void (*commit_scheduling)(struct cpu_hw_events *cpuc, int idx, int cntr);
532 void (*start_scheduling)(struct cpu_hw_events *cpuc);
534 void (*stop_scheduling)(struct cpu_hw_events *cpuc);
660 DECLARE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
721 u64 disable_mask = __this_cpu_read(cpu_hw_events.perf_ctr_virt_mask); in __x86_pmu_enable_event()
732 int x86_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign);
[all …]
Dperf_event_intel_bts.c141 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_config_buffer()
192 struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; in bts_update()
422 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_del()
447 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_add()
Dperf_event_knc.c215 struct cpu_hw_events *cpuc; in knc_pmu_handle_irq()
220 cpuc = this_cpu_ptr(&cpu_hw_events); in knc_pmu_handle_irq()
Dperf_event_p4.c918 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_disable_all()
987 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_enable_all()
1001 struct cpu_hw_events *cpuc; in p4_pmu_handle_irq()
1007 cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_handle_irq()
1206 static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign) in p4_pmu_schedule_events()
/linux-4.1.27/arch/metag/kernel/perf/
Dperf_event.c38 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
261 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_start()
309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_add()
351 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_del()
600 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_enable_counter()
673 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_disable_counter()
721 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_write_counter()
754 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in metag_pmu_counter_overflow()
810 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); in metag_pmu_cpu_notify()
815 memset(cpuc, 0, sizeof(struct cpu_hw_events)); in metag_pmu_cpu_notify()
[all …]
Dperf_event.h50 struct cpu_hw_events { struct
/linux-4.1.27/arch/powerpc/perf/
Dcore-fsl-emb.c23 struct cpu_hw_events { struct
29 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument
209 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_disable()
213 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_disable()
248 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_enable()
252 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_enable()
295 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_add()
302 cpuhw = &get_cpu_var(cpu_hw_events); in fsl_emb_pmu_add()
348 put_cpu_var(cpu_hw_events); in fsl_emb_pmu_add()
356 struct cpu_hw_events *cpuhw; in fsl_emb_pmu_del()
[all …]
Dcore-book3s.c29 struct cpu_hw_events { struct
62 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument
120 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in()
128 static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} in power_pmu_bhrb_read()
351 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_enable()
367 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_disable()
425 static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) in power_pmu_bhrb_read()
564 static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) in ebb_switch_in()
828 static int power_check_constraints(struct cpu_hw_events *cpuhw, in power_check_constraints()
1065 static void freeze_limited_counters(struct cpu_hw_events *cpuhw, in freeze_limited_counters()
[all …]
/linux-4.1.27/arch/alpha/kernel/
Dperf_event.c33 struct cpu_hw_events { struct
52 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); argument
390 static void maybe_change_configuration(struct cpu_hw_events *cpuc) in maybe_change_configuration()
434 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_add()
486 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_del()
534 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_stop()
554 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_start()
727 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_enable()
753 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_disable()
811 struct cpu_hw_events *cpuc; in alpha_perf_event_irq_handler()
[all …]
/linux-4.1.27/arch/sparc/kernel/
Dperf_event.c76 struct cpu_hw_events { struct
113 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { .enabled = 1, }; argument
825 static inline void sparc_pmu_enable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, in… in sparc_pmu_enable_event()
843 static inline void sparc_pmu_disable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, i… in sparc_pmu_disable_event()
917 static void read_in_all_counters(struct cpu_hw_events *cpuc) in read_in_all_counters()
939 static void calculate_single_pcr(struct cpu_hw_events *cpuc) in calculate_single_pcr()
973 static void calculate_multiple_pcrs(struct cpu_hw_events *cpuc) in calculate_multiple_pcrs()
1004 static void update_pcrs_for_enable(struct cpu_hw_events *cpuc) in update_pcrs_for_enable()
1018 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_enable()
1036 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_disable()
[all …]
/linux-4.1.27/arch/tile/kernel/
Dperf_event.c50 struct cpu_hw_events { struct
84 static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_stop()
619 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_start()
653 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_add()
709 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_del()
883 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_handle_irq()
/linux-4.1.27/arch/arm/kernel/
Dperf_event_cpu.c199 struct pmu_hw_events __percpu *cpu_hw_events; in cpu_pmu_init() local
201 cpu_hw_events = alloc_percpu(struct pmu_hw_events); in cpu_pmu_init()
202 if (!cpu_hw_events) in cpu_pmu_init()
211 struct pmu_hw_events *events = per_cpu_ptr(cpu_hw_events, cpu); in cpu_pmu_init()
216 cpu_pmu->hw_events = cpu_hw_events; in cpu_pmu_init()
231 free_percpu(cpu_hw_events); in cpu_pmu_init()
/linux-4.1.27/arch/mips/kernel/
Dperf_event_mipsxx.c34 struct cpu_hw_events { struct
51 DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { argument
311 static int mipsxx_pmu_alloc_counter(struct cpu_hw_events *cpuc, in mipsxx_pmu_alloc_counter()
343 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_enable_event()
363 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_disable_event()
463 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_add()
499 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_del()
719 struct cpu_hw_events fake_cpuc; in validate_group()
738 static void handle_associated_event(struct cpu_hw_events *cpuc, in handle_associated_event()
1342 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in pause_local_counters()
[all …]
/linux-4.1.27/arch/arm64/kernel/
Dperf_event.c49 static DEFINE_PER_CPU(struct pmu_hw_events, cpu_hw_events);
407 free_percpu_irq(irq, &cpu_hw_events); in armpmu_release_hardware()
457 "arm-pmu", &cpu_hw_events); in armpmu_reserve_hardware()
1122 cpuc = this_cpu_ptr(&cpu_hw_events); in armv8pmu_handle_irq()
1376 return this_cpu_ptr(&cpu_hw_events); in armpmu_get_cpu_events()
1383 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); in cpu_pmu_init()