Searched refs:shares (Results 1 - 4 of 4) sorted by relevance
/linux-master/kernel/sched/ |
H A D | autogroup.c | 235 unsigned long shares; local 256 shares = scale_load(sched_prio_to_weight[idx]); 259 err = sched_group_set_shares(ag->tg, shares);
|
H A D | fair.c | 3934 long tg_weight, tg_shares, load, shares; local 3937 tg_shares = READ_ONCE(tg->shares); 3947 shares = (tg_shares * load); 3949 shares /= tg_weight; 3953 * of a group with small tg->shares value. It is a floor value which is 3957 * E.g. on 64-bit for a group with tg->shares of scale_load(15)=15*1024 3958 * on an 8-core system with 8 tasks each runnable on one CPU shares has 3963 return clamp_t(long, shares, MIN_SHARES, tg_shares); 3974 long shares; local 3983 shares 3988 reweight_entity(cfs_rq_of(se), se, shares); local 12971 __sched_group_set_shares(struct task_group *tg, unsigned long shares) argument 13007 sched_group_set_shares(struct task_group *tg, unsigned long shares) argument [all...] |
H A D | sched.h | 121 * The extra resolution improves shares distribution and load balancing of 124 * and does not change the user-interface for setting shares/weights. 387 unsigned long shares; member in struct:task_group 440 * too large, so as the shares value of a task group. 509 extern int sched_group_set_shares(struct task_group *tg, unsigned long shares);
|
H A D | core.c | 9952 root_task_group.shares = ROOT_TASK_GROUP_LOAD; 10796 return (u64) scale_load_down(tg->shares); 11188 .name = "shares", 11301 u64 weight = scale_load_down(tg->shares); 11312 * a bit of range on both ends, it maps pretty well onto the shares 11327 unsigned long weight = scale_load_down(css_tg(css)->shares);
|
Completed in 240 milliseconds