/linux-master/arch/x86/events/amd/ |
H A D | brs.c | 381 * On ctxswin, sched_in = true, called after the PMU has started 382 * On ctxswout, sched_in = false, called before the PMU is stopped 384 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 397 if (sched_in)
|
H A D | lbr.c | 378 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 387 if (cpuc->lbr_users && sched_in)
|
/linux-master/include/linux/ |
H A D | preempt.h | 325 * @sched_in: we're about to be rescheduled: 332 * Please note that sched_in and out are called under different 334 * while sched_in is called without rq lock and irq enabled. This 338 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member in struct:preempt_ops
|
H A D | perf_event.h | 456 bool sched_in);
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm-x86-ops.h | 106 KVM_X86_OP(sched_in)
|
H A D | kvm_host.h | 1747 void (*sched_in)(struct kvm_vcpu *vcpu, int cpu); member in struct:kvm_x86_ops
|
/linux-master/arch/x86/events/ |
H A D | perf_event.h | 832 bool sched_in); 1327 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1368 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1393 static inline void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 1555 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1570 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in);
|
H A D | core.c | 2623 static void x86_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 2625 static_call_cond(x86_pmu_sched_task)(pmu_ctx, sched_in);
|
/linux-master/tools/perf/ |
H A D | builtin-sched.c | 1112 struct thread *sched_out, *sched_in; local 1132 sched_in = machine__findnew_thread(machine, -1, next_pid); 1133 if (sched_out == NULL || sched_in == NULL) 1149 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); 1151 if (thread_atoms_insert(sched, sched_in)) 1153 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); 1169 thread__put(sched_in); 1545 struct thread *sched_in; local 1585 sched_in = map__findnew_thread(sched, machine, -1, next_pid); 1586 if (sched_in [all...] |
/linux-master/arch/s390/kernel/ |
H A D | perf_pai_crypto.c | 476 static void paicrypt_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 481 if (!sched_in)
|
H A D | perf_pai_ext.c | 489 static void paiext_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 494 if (!sched_in)
|
/linux-master/arch/x86/events/intel/ |
H A D | lbr.c | 541 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 556 if (sched_in) 569 if (sched_in)
|
H A D | ds.c | 1129 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 1133 if (!sched_in && pebs_needs_sched_cb(cpuc))
|
H A D | core.c | 4960 bool sched_in) 4962 intel_pmu_pebs_sched_task(pmu_ctx, sched_in); 4963 intel_pmu_lbr_sched_task(pmu_ctx, sched_in); 4959 intel_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument
|
/linux-master/arch/powerpc/perf/ |
H A D | core-book3s.c | 135 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) {} argument 454 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) argument 459 if (sched_in)
|
/linux-master/kernel/events/ |
H A D | core.c | 3479 static void perf_ctx_sched_task_cb(struct perf_event_context *ctx, bool sched_in) argument 3488 pmu_ctx->pmu->sched_task(pmu_ctx, sched_in); 3618 static void __perf_pmu_sched_task(struct perf_cpu_pmu_context *cpc, bool sched_in) argument 3632 pmu->sched_task(cpc->task_epc, sched_in); 3640 bool sched_in) 3650 __perf_pmu_sched_task(cpc, sched_in); 3654 struct task_struct *next_prev, bool sched_in); 9072 struct task_struct *next_prev, bool sched_in) 9084 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, 9092 if (!sched_in 3638 perf_pmu_sched_task(struct task_struct *prev, struct task_struct *next, bool sched_in) argument 9071 perf_event_switch(struct task_struct *task, struct task_struct *next_prev, bool sched_in) argument [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 5011 .sched_in = svm_sched_in,
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 6469 kvm_preempt_ops.sched_in = kvm_sched_in;
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.c | 8393 .sched_in = vmx_sched_in,
|
/linux-master/kernel/sched/ |
H A D | core.c | 4964 notifier->ops->sched_in(notifier, raw_smp_processor_id());
|