/linux-master/arch/xtensa/include/asm/ |
H A D | mmu_context.h | 130 int migrated = next->context.cpu != cpu; local 131 /* Flush the icache if we migrated to a new core. */ 132 if (migrated) { 136 if (migrated || prev != next)
|
/linux-master/tools/perf/scripts/python/ |
H A D | compaction-times.py | 72 self.migrated = pair(0, 0, "moved", "failed") 78 self.migrated += rhs.migrated 87 s += "migration: %s" % self.migrated 99 def increment(self, migrated, fscan, mscan): 100 if (migrated != None): 101 self.migrated += migrated 132 def increment_pending(cls, pid, migrated, fscan, mscan): 136 head.do_increment(migrated, fsca [all...] |
/linux-master/fs/nfsd/ |
H A D | export.h | 32 * list of replicas that do serve it) then we set "migrated": */ 33 int migrated; member in struct:nfsd4_fs_locations
|
H A D | export.c | 454 int migrated, i, err; local 494 /* migrated */ 495 err = get_int(mesg, &migrated); 499 if (migrated < 0 || migrated > 1) 501 fsloc->migrated = migrated; 824 new->ex_fslocs.migrated = 0; 849 new->ex_fslocs.migrated = item->ex_fslocs.migrated; [all...] |
H A D | nfs4xdr.c | 3041 if (unlikely(args->exp->ex_fslocs.migrated)) { 3527 if (exp->ex_fslocs.migrated) {
|
H A D | nfs4proc.c | 2712 current_fh->fh_export->ex_fslocs.migrated &&
|
/linux-master/arch/loongarch/kvm/ |
H A D | main.c | 220 bool migrated; local 230 migrated = (vcpu->cpu != cpu); 241 if (migrated || (ver != old)) {
|
H A D | vcpu.c | 960 bool migrated; local 965 * Have we migrated to a different CPU? 968 migrated = (vcpu->arch.last_sched_cpu != cpu); 975 if (migrated || (context->last_vcpu != vcpu))
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | ruc.c | 36 bool migrated = packet->migrated; local 39 if (qp->s_mig_state == IB_MIG_ARMED && migrated) { 109 if (qp->s_mig_state == IB_MIG_REARM && !migrated)
|
H A D | driver.c | 1514 packet->migrated = ib_bth_is_migration(packet->ohdr); 1551 packet->migrated = opa_bth_is_migration(packet->ohdr); 1563 packet->migrated = opa_bth_is_migration(packet->ohdr); 1577 packet->migrated = false;
|
H A D | hfi.h | 361 bool migrated; member in struct:hfi1_packet
|
/linux-master/fs/xfs/ |
H A D | xfs_mru_cache.c | 65 * this happens, all the elements on the LRU list need to be migrated to the end 118 * Case a) above is detected by counting how many groups are migrated and 120 * time_zero field, which is updated as each group is migrated. 131 unsigned int migrated = 0; local 158 * lists have been migrated to the reap list, it's now empty. 160 if (++migrated == mru->grp_count) { 185 * MRU list. The lists must be migrated first to ensure that they're 219 * When destroying or reaping, all the elements that were migrated to the reap
|
/linux-master/arch/mips/kvm/ |
H A D | vz.c | 2506 bool migrated; local 2512 migrated = (vcpu->arch.last_exec_cpu != cpu); 2529 if (migrated || 2548 if (migrated || last_exec_vcpu[cpu] != vcpu) 2566 bool migrated, all; local 2569 * Have we migrated to a different CPU? 2572 migrated = (vcpu->arch.last_sched_cpu != cpu); 2578 all = migrated || (last_vcpu[cpu] != vcpu);
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 1863 * @event: mmu notifier event when range is evicted or migrated 3378 * @migrated: output, true if migration is triggered 3401 bool *migrated) 3406 *migrated = false; 3421 *migrated = !r; 3427 *migrated = !r; 3585 bool migrated; local 3589 r = svm_range_trigger_migration(mm, prange, &migrated); 3593 if (migrated && (!p->xnack_enabled || 3601 if (!migrated 3400 svm_range_trigger_migration(struct mm_struct *mm, struct svm_range *prange, bool *migrated) argument [all...] |
/linux-master/fs/f2fs/ |
H A D | gc.c | 1676 int seg_freed = 0, migrated = 0; local 1731 migrated >= sbi->migration_granularity) 1763 migrated++; 1782 if (migrated)
|
/linux-master/tools/perf/ |
H A D | builtin-sched.c | 2466 struct thread *migrated) 2485 timehist_skip_sample(sched, migrated, evsel, sample)) { 2509 printf("migrated: %s", timehist_get_commstr(migrated)); 2524 /* want pid of migrated task not pid in sample */ 2462 timehist_print_migration_event(struct perf_sched *sched, struct evsel *evsel, struct perf_sample *sample, struct machine *machine, struct thread *migrated) argument
|
/linux-master/mm/ |
H A D | huge_memory.c | 1722 bool migrated = false, writable = false; local 1766 migrated = migrate_misplaced_folio(folio, vma, target_nid); 1767 if (migrated) {
|
/linux-master/kernel/sched/ |
H A D | fair.c | 1728 * memory pages will be migrated to fast memory node without 1840 * The pages in slow memory node should be migrated according 2159 * be improved if the source tasks was migrated to the target dst_cpu taking 2185 * Because we have preemption enabled we can get migrated around and 2449 * Tasks that are "trapped" in such domains cannot be migrated 2507 * A task that migrated to a second choice node will be better off 3074 bool migrated = flags & TNF_MIGRATED; local 3142 if (migrated) 3292 * migrated as it is expected they are cache replicated. Avoid 4741 * Track task load average for carrying it to new CPU after migrated, an [all...] |