/linux-6.1.9/arch/x86/events/amd/ |
D | brs.c | 387 void amd_pmu_brs_sched_task(struct perf_event_context *ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 400 if (sched_in) in amd_pmu_brs_sched_task()
|
D | lbr.c | 376 void amd_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in) in amd_pmu_lbr_sched_task() argument 385 if (cpuc->lbr_users && sched_in) in amd_pmu_lbr_sched_task()
|
/linux-6.1.9/tools/perf/ |
D | builtin-sched.c | 1137 struct thread *sched_out, *sched_in; in latency_switch_event() local 1157 sched_in = machine__findnew_thread(machine, -1, next_pid); in latency_switch_event() 1158 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1174 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1176 if (thread_atoms_insert(sched, sched_in)) in latency_switch_event() 1178 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1194 thread__put(sched_in); in latency_switch_event() 1555 struct thread *sched_in; in map_switch_event() local 1595 sched_in = map__findnew_thread(sched, machine, -1, next_pid); in map_switch_event() 1596 if (sched_in == NULL) in map_switch_event() [all …]
|
/linux-6.1.9/arch/x86/events/ |
D | perf_event.h | 815 bool sched_in); 1309 void amd_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in); 1351 void amd_pmu_brs_sched_task(struct perf_event_context *ctx, bool sched_in); 1376 static inline void amd_pmu_brs_sched_task(struct perf_event_context *ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 1536 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in); 1547 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in);
|
D | core.c | 2649 static void x86_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in x86_pmu_sched_task() argument 2651 static_call_cond(x86_pmu_sched_task)(ctx, sched_in); in x86_pmu_sched_task()
|
/linux-6.1.9/arch/x86/include/asm/ |
D | kvm-x86-ops.h | 102 KVM_X86_OP(sched_in)
|
D | kvm_host.h | 1589 void (*sched_in)(struct kvm_vcpu *kvm, int cpu); member
|
/linux-6.1.9/include/linux/ |
D | preempt.h | 326 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member
|
D | perf_event.h | 447 bool sched_in);
|
/linux-6.1.9/arch/s390/kernel/ |
D | perf_pai_crypto.c | 382 static void paicrypt_sched_task(struct perf_event_context *ctx, bool sched_in) in paicrypt_sched_task() argument 387 if (!sched_in) in paicrypt_sched_task()
|
D | perf_pai_ext.c | 475 static void paiext_sched_task(struct perf_event_context *ctx, bool sched_in) in paiext_sched_task() argument 480 if (!sched_in) in paiext_sched_task()
|
/linux-6.1.9/arch/x86/events/intel/ |
D | lbr.c | 541 void intel_pmu_lbr_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_lbr_sched_task() argument 556 if (sched_in) in intel_pmu_lbr_sched_task() 569 if (sched_in) in intel_pmu_lbr_sched_task()
|
D | ds.c | 1072 void intel_pmu_pebs_sched_task(struct perf_event_context *ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 1076 if (!sched_in && pebs_needs_sched_cb(cpuc)) in intel_pmu_pebs_sched_task()
|
D | core.c | 4675 bool sched_in) in intel_pmu_sched_task() argument 4677 intel_pmu_pebs_sched_task(ctx, sched_in); in intel_pmu_sched_task() 4678 intel_pmu_lbr_sched_task(ctx, sched_in); in intel_pmu_sched_task()
|
/linux-6.1.9/arch/powerpc/perf/ |
D | core-book3s.c | 135 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) {} in power_pmu_sched_task() argument 454 static void power_pmu_sched_task(struct perf_event_context *ctx, bool sched_in) in power_pmu_sched_task() argument 459 if (sched_in) in power_pmu_sched_task()
|
/linux-6.1.9/kernel/events/ |
D | core.c | 3554 static void __perf_pmu_sched_task(struct perf_cpu_context *cpuctx, bool sched_in) in __perf_pmu_sched_task() argument 3566 pmu->sched_task(cpuctx->task_ctx, sched_in); in __perf_pmu_sched_task() 3574 bool sched_in) in perf_pmu_sched_task() argument 3586 __perf_pmu_sched_task(cpuctx, sched_in); in perf_pmu_sched_task() 3591 struct task_struct *next_prev, bool sched_in); 8820 struct task_struct *next_prev, bool sched_in) in perf_event_switch() argument 8832 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch() 8840 if (!sched_in && task->on_rq) { in perf_event_switch()
|
/linux-6.1.9/arch/x86/kvm/svm/ |
D | svm.c | 4777 .sched_in = svm_sched_in,
|
/linux-6.1.9/virt/kvm/ |
D | kvm_main.c | 5946 kvm_preempt_ops.sched_in = kvm_sched_in; in kvm_init()
|
/linux-6.1.9/kernel/sched/ |
D | core.c | 4798 notifier->ops->sched_in(notifier, raw_smp_processor_id()); in __fire_sched_in_preempt_notifiers()
|
/linux-6.1.9/arch/x86/kvm/vmx/ |
D | vmx.c | 8110 .sched_in = vmx_sched_in,
|