/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_raw_tp_test_run.c | 9 __u32 on_cpu = 0xffffffff; variable 17 on_cpu = bpf_get_smp_processor_id();
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | raw_tp_test_run.c | 43 ASSERT_EQ(skel->data->on_cpu, 0xffffffff, "check_on_cpu"); 65 ASSERT_EQ(skel->data->on_cpu, i, "check_on_cpu");
|
/linux-master/drivers/perf/hisilicon/ |
H A D | hisi_uncore_pmu.c | 61 return sysfs_emit(buf, "%d\n", hisi_pmu->on_cpu); 218 if (hisi_pmu->on_cpu == -1) 232 event->cpu = hisi_pmu->on_cpu; 490 if (hisi_pmu->on_cpu != -1) 494 hisi_pmu->on_cpu = cpu; 514 if (hisi_pmu->on_cpu != cpu) 518 hisi_pmu->on_cpu = -1; 529 hisi_pmu->on_cpu = target;
|
H A D | hisi_pcie_pmu.c | 76 int on_cpu; member in struct:hisi_pcie_pmu 135 return cpumap_print_to_pagebuf(true, buf, cpumask_of(pcie_pmu->on_cpu)); 388 event->cpu = pcie_pmu->on_cpu; 664 if (pcie_pmu->on_cpu == -1) { 665 pcie_pmu->on_cpu = cpumask_local_spread(0, dev_to_node(&pcie_pmu->pdev->dev)); 666 WARN_ON(irq_set_affinity(pcie_pmu->irq, cpumask_of(pcie_pmu->on_cpu))); 680 if (pcie_pmu->on_cpu != cpu) 683 pcie_pmu->on_cpu = -1; 700 pcie_pmu->on_cpu = target; 805 pcie_pmu->on_cpu [all...] |
H A D | hisi_uncore_pmu.h | 93 int on_cpu; member in struct:hisi_pmu
|
H A D | hns3_pmu.c | 308 int on_cpu; member in struct:hns3_pmu 482 return sysfs_emit(buf, "%d\n", hns3_pmu->on_cpu); 1246 event->cpu = hns3_pmu->on_cpu; 1417 hns3_pmu->on_cpu = -1; 1467 if (hns3_pmu->on_cpu == -1) { 1468 hns3_pmu->on_cpu = cpu; 1485 if (hns3_pmu->on_cpu != cpu) 1494 hns3_pmu->on_cpu = target;
|
H A D | hisi_uncore_cpa_pmu.c | 295 cpa_pmu->on_cpu = -1;
|
H A D | hisi_uncore_ddrc_pmu.c | 481 ddrc_pmu->on_cpu = -1;
|
H A D | hisi_uncore_pa_pmu.c | 472 pa_pmu->on_cpu = -1;
|
H A D | hisi_uncore_sllc_pmu.c | 417 sllc_pmu->on_cpu = -1;
|
H A D | hisi_uncore_hha_pmu.c | 492 hha_pmu->on_cpu = -1;
|
H A D | hisi_uncore_l3c_pmu.c | 526 l3c_pmu->on_cpu = -1;
|
H A D | hisi_uncore_uc_pmu.c | 510 uc_pmu->on_cpu = -1;
|
/linux-master/drivers/perf/ |
H A D | dwc_pcie_pmu.c | 88 int on_cpu; member in struct:dwc_pcie_pmu 119 return cpumap_print_to_pagebuf(true, buf, cpumask_of(pcie_pmu->on_cpu)); 398 event->cpu = pcie_pmu->on_cpu; 633 pcie_pmu->on_cpu = -1; 682 if (pcie_pmu->on_cpu == -1) 683 pcie_pmu->on_cpu = cpumask_local_spread( 699 if (cpu != pcie_pmu->on_cpu) 702 pcie_pmu->on_cpu = -1; 718 pcie_pmu->on_cpu = target;
|
H A D | cxl_pmu.c | 106 int on_cpu; member in struct:cxl_pmu_info 504 return cpumap_print_to_pagebuf(true, buf, cpumask_of(info->on_cpu)); 591 event->cpu = info->on_cpu; 840 info->on_cpu = -1; 923 if (info->on_cpu != -1) 926 info->on_cpu = cpu; 941 if (info->on_cpu != cpu) 944 info->on_cpu = -1; 952 info->on_cpu = target;
|
H A D | qcom_l2_pmu.c | 144 int on_cpu; member in struct:cluster_pmu 498 (cluster->on_cpu != event->group_leader->cpu)) { 536 event->cpu = cluster->on_cpu; 785 if (cluster->on_cpu != -1) 792 cluster->on_cpu = cpu; 815 if (cluster->on_cpu != cpu) 820 cluster->on_cpu = -1; 832 cluster->on_cpu = target; 868 cluster->on_cpu = -1;
|
H A D | arm_smmuv3_pmu.c | 128 unsigned int on_cpu; member in struct:smmu_pmu 451 event->cpu = smmu_pmu->on_cpu; 537 return cpumap_print_to_pagebuf(true, buf, cpumask_of(smmu_pmu->on_cpu)); 671 if (cpu != smmu_pmu->on_cpu) 679 smmu_pmu->on_cpu = target; 938 smmu_pmu->on_cpu = raw_smp_processor_id(); 939 WARN_ON(irq_set_affinity(smmu_pmu->irq, cpumask_of(smmu_pmu->on_cpu)));
|
/linux-master/arch/x86/include/asm/ |
H A D | unwind.h | 148 return task != current && task->on_cpu;
|
/linux-master/drivers/hwtracing/ptt/ |
H A D | hisi_ptt.h | 123 * @on_cpu: current tracing cpu 135 int on_cpu; member in struct:hisi_ptt_trace_ctrl
|
H A D | hisi_ptt.c | 773 hisi_ptt->trace_ctrl.on_cpu = -1; 1111 hisi_ptt->trace_ctrl.on_cpu = cpu; 1389 src = hisi_ptt->trace_ctrl.on_cpu; 1410 hisi_ptt->trace_ctrl.on_cpu = target;
|
/linux-master/include/linux/ |
H A D | sched.h | 774 int on_cpu; member in struct:task_struct 909 * queueing no longer being serialized by p->on_cpu. However: 913 * smp_mb__after_spinlock(); if (smp_load_acquire(&p->on_cpu) && //true 2169 return READ_ONCE(owner->on_cpu) && !vcpu_is_preempted(task_cpu(owner));
|
/linux-master/kernel/sched/ |
H A D | core.c | 517 * p->on_cpu <- { 0, 1 }: 524 * CPU to have ->on_cpu = 1 at the same time. ] 2766 * smp_cond_load_acquire(&p->on_cpu, !VAL), and we're in __schedule() 2772 SCHED_WARN_ON(!p->on_cpu); 3860 * When on_rq && !on_cpu the task is preempted, see if 3889 if (WARN_ON_ONCE(p->on_cpu)) 3890 smp_cond_load_acquire(&p->on_cpu, !VAL); 4018 * p->on_cpu can be whatever, we've done the dequeue, so 4159 * 1) smp_store_release(X->on_cpu, 0) -- finish_task() 4160 * 2) smp_cond_load_acquire(!X->on_cpu) [all...] |
H A D | sched.h | 2161 return p->on_cpu;
|
H A D | fair.c | 1404 (lockdep_is_held(__rq_lockp(task_rq(p))) && !READ_ONCE(p->on_cpu))); 12319 * This is OK, because current is on_cpu, which avoids it being picked
|
/linux-master/kernel/rcu/ |
H A D | tree_nocb.h | 1621 // The ->on_cpu field is available only in CONFIG_SMP=y, so... 1625 return tsp && task_is_running(tsp) && !tsp->on_cpu ? "!" : "";
|