Searched refs:shares (Results 1 - 4 of 4) sorted by relevance

/linux-master/kernel/sched/
H A Dautogroup.c235 unsigned long shares; local
256 shares = scale_load(sched_prio_to_weight[idx]);
259 err = sched_group_set_shares(ag->tg, shares);
H A Dfair.c3934 long tg_weight, tg_shares, load, shares; local
3937 tg_shares = READ_ONCE(tg->shares);
3947 shares = (tg_shares * load);
3949 shares /= tg_weight;
3953 * of a group with small tg->shares value. It is a floor value which is
3957 * E.g. on 64-bit for a group with tg->shares of scale_load(15)=15*1024
3958 * on an 8-core system with 8 tasks each runnable on one CPU shares has
3963 return clamp_t(long, shares, MIN_SHARES, tg_shares);
3974 long shares; local
3983 shares
3988 reweight_entity(cfs_rq_of(se), se, shares); local
12971 __sched_group_set_shares(struct task_group *tg, unsigned long shares) argument
13007 sched_group_set_shares(struct task_group *tg, unsigned long shares) argument
[all...]
H A Dsched.h121 * The extra resolution improves shares distribution and load balancing of
124 * and does not change the user-interface for setting shares/weights.
387 unsigned long shares; member in struct:task_group
440 * too large, so as the shares value of a task group.
509 extern int sched_group_set_shares(struct task_group *tg, unsigned long shares);
H A Dcore.c9952 root_task_group.shares = ROOT_TASK_GROUP_LOAD;
10796 return (u64) scale_load_down(tg->shares);
11188 .name = "shares",
11301 u64 weight = scale_load_down(tg->shares);
11312 * a bit of range on both ends, it maps pretty well onto the shares
11327 unsigned long weight = scale_load_down(css_tg(css)->shares);

Completed in 240 milliseconds