Searched refs:shares (Results 1 - 4 of 4) sorted by path
/linux-master/kernel/sched/ |
H A D | autogroup.c | 235 unsigned long shares; local 256 shares = scale_load(sched_prio_to_weight[idx]); 259 err = sched_group_set_shares(ag->tg, shares);
|
H A D | core.c | 9952 root_task_group.shares = ROOT_TASK_GROUP_LOAD; 10796 return (u64) scale_load_down(tg->shares); 11188 .name = "shares", 11301 u64 weight = scale_load_down(tg->shares); 11312 * a bit of range on both ends, it maps pretty well onto the shares 11327 unsigned long weight = scale_load_down(css_tg(css)->shares);
|
H A D | fair.c | 3928 long tg_weight, tg_shares, load, shares; local 3931 tg_shares = READ_ONCE(tg->shares); 3941 shares = (tg_shares * load); 3943 shares /= tg_weight; 3947 * of a group with small tg->shares value. It is a floor value which is 3951 * E.g. on 64-bit for a group with tg->shares of scale_load(15)=15*1024 3952 * on an 8-core system with 8 tasks each runnable on one CPU shares has 3957 return clamp_t(long, shares, MIN_SHARES, tg_shares); 3968 long shares; local 3977 shares 3982 reweight_entity(cfs_rq_of(se), se, shares); local 13044 __sched_group_set_shares(struct task_group *tg, unsigned long shares) argument 13080 sched_group_set_shares(struct task_group *tg, unsigned long shares) argument [all...] |
H A D | sched.h | 135 * The extra resolution improves shares distribution and load balancing of 138 * and does not change the user-interface for setting shares/weights. 401 unsigned long shares; member in struct:task_group 454 * too large, so as the shares value of a task group. 523 extern int sched_group_set_shares(struct task_group *tg, unsigned long shares);
|
Completed in 259 milliseconds