Lines Matching defs:util

1958 	unsigned long util;
1992 (((ns->compute_capacity * 100) < (ns->util * imbalance_pct)) ||
1997 (((ns->compute_capacity * 100) > (ns->util * imbalance_pct)) &&
2050 ns->util += cpu_util_cfs(cpu);
4005 * As is, the util number is not freq-invariant (we'd have to
4221 * And since, like util, the runnable part should be directly transferable,
4563 * update_cfs_rq_load_avg - update the cfs_rq's load/util averages
4977 static inline int util_fits_cpu(unsigned long util,
4987 * Check if the real util fits without any uclamp boost/cap applied.
4989 fits = fits_capacity(util, capacity);
4997 * capacity_of()) for comparing against the real util.
5027 * | | | | | | | (util somewhere in this region)
5036 * * util = 80% of x then it does not fit on CPU0 and should migrate
5038 * * util = 80% of y then it is forced to fit on CPU1 to honour
5060 * | ___ (region a, capped, util >= uclamp_max)
5064 * | ___ | | | | (region b, uclamp_min <= util <= uclamp_max)
5067 * | | | | | | | (region c, boosted, util < uclamp_min)
5071 * a) If util > uclamp_max, then we're capped, we don't care about
5076 * b) If uclamp_min <= util <= uclamp_max, then the normal
5080 * c) If util < uclamp_min, then we are boosted. Same as (b) but we
5089 if (fits && (util < uclamp_min) &&
5100 unsigned long util = task_util_est(p);
5105 return (util_fits_cpu(util, uclamp_min, uclamp_max, cpu) > 0);
6961 /* Discount task's util from CPU's util */
7216 * We need task's util for cpu_util_without, sync it up to
7523 static inline bool asym_fits_cpu(unsigned long util,
7533 return (util_fits_cpu(util, util_min, util_max, cpu) > 0);
7701 * range. Otherwise a group of CPUs (CPU0 util = 121% + CPU1 util = 80%)
7713 unsigned long util = READ_ONCE(cfs_rq->avg.util_avg);
7718 util = max(util, runnable);
7728 lsub_positive(&util, task_util(p));
7730 util += task_util(p);
7768 util = max(util, util_est);
7771 return min(util, arch_scale_cpu_capacity(cpu));
7871 unsigned long util = cpu_util(cpu, p, -1, 0);
7873 busy_time += effective_cpu_util(cpu, util, NULL, NULL);
7895 unsigned long util = cpu_util(cpu, p, dst_cpu, 1);
7905 eff_util = effective_cpu_util(cpu, util, &min, &max);
8030 unsigned long cpu_cap, cpu_actual_cap, util;
8060 util = cpu_util(cpu, p, cpu, 0);
8075 * operate on non clamped util but must use the
8085 fits = util_fits_cpu(util, util_min, util_max, cpu);
8089 lsub_positive(&cpu_cap, util);
9117 * detach_tasks() -- tries to detach up to imbalance load/util/tasks from
9125 unsigned long util, load;
9200 util = task_util_est(p);
9202 if (shr_bound(util, env->sd->nr_balance_failed) > env->imbalance)
9205 env->imbalance -= util;
9238 * load/util/tasks.
11032 unsigned long capacity, load, util;
11121 util = cpu_util_cfs_boost(i);
11131 if (busiest_util < util) {
11132 busiest_util = util;
11477 * load/util migration, don't pollute nr_balance_failed.