/linux-4.4.14/tools/perf/ |
D | builtin-sched.c | 963 struct thread *sched_out, *sched_in; in latency_switch_event() local 983 sched_in = machine__findnew_thread(machine, -1, next_pid); in latency_switch_event() 984 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1000 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1002 if (thread_atoms_insert(sched, sched_in)) in latency_switch_event() 1004 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1020 thread__put(sched_in); in latency_switch_event() 1346 struct thread *sched_in; in map_switch_event() local 1369 sched_in = machine__findnew_thread(machine, -1, next_pid); in map_switch_event() 1370 if (sched_in == NULL) in map_switch_event() [all …]
|
/linux-4.4.14/include/linux/ |
D | preempt.h | 270 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member
|
D | perf_event.h | 366 bool sched_in);
|
/linux-4.4.14/arch/x86/kernel/cpu/ |
D | perf_event.h | 577 bool sched_in); 881 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in); 885 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in);
|
D | perf_event_intel_lbr.c | 278 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_lbr_sched_task() argument 290 if (sched_in) { in intel_pmu_lbr_sched_task() 311 if (sched_in) { in intel_pmu_lbr_sched_task()
|
D | perf_event_intel_ds.c | 606 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 608 if (!sched_in) in intel_pmu_pebs_sched_task()
|
D | perf_event_intel.c | 2908 bool sched_in) in intel_pmu_sched_task() argument 2911 intel_pmu_pebs_sched_task(ctx, sched_in); in intel_pmu_sched_task() 2913 intel_pmu_lbr_sched_task(ctx, sched_in); in intel_pmu_sched_task()
|
D | perf_event.c | 2099 static void x86_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in x86_pmu_sched_task() argument 2102 x86_pmu.sched_task(ctx, sched_in); in x86_pmu_sched_task()
|
/linux-4.4.14/arch/powerpc/perf/ |
D | core-book3s.c | 127 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) {} in power_pmu_sched_task() argument 389 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in power_pmu_sched_task() argument 394 if (sched_in) in power_pmu_sched_task()
|
/linux-4.4.14/arch/x86/include/asm/ |
D | kvm_host.h | 872 void (*sched_in)(struct kvm_vcpu *kvm, int cpu); member
|
/linux-4.4.14/kernel/events/ |
D | core.c | 2625 bool sched_in) in perf_pmu_sched_task() argument 2646 pmu->sched_task(cpuctx->task_ctx, sched_in); in perf_pmu_sched_task() 2660 struct task_struct *next_prev, bool sched_in); 6290 struct task_struct *next_prev, bool sched_in) in perf_event_switch() argument 6302 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch()
|
/linux-4.4.14/virt/kvm/ |
D | kvm_main.c | 3597 kvm_preempt_ops.sched_in = kvm_sched_in; in kvm_init()
|
/linux-4.4.14/arch/x86/kvm/ |
D | svm.c | 4364 .sched_in = svm_sched_in,
|
D | x86.c | 7624 kvm_x86_ops->sched_in(vcpu, cpu); in kvm_arch_sched_in()
|
D | vmx.c | 10882 .sched_in = vmx_sched_in,
|
/linux-4.4.14/kernel/sched/ |
D | core.c | 2454 notifier->ops->sched_in(notifier, raw_smp_processor_id()); in __fire_sched_in_preempt_notifiers()
|