mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 03:36:19 +00:00
sched: cleanup: rename task_grp to task_group
cleanup: rename task_grp to task_group. No need to save two characters and 'grp' is annoying to read. Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
06877c33fe
commit
4cf86d77f5
4 changed files with 28 additions and 28 deletions
|
@ -136,7 +136,7 @@ extern unsigned long weighted_cpuload(const int cpu);
|
||||||
|
|
||||||
struct seq_file;
|
struct seq_file;
|
||||||
struct cfs_rq;
|
struct cfs_rq;
|
||||||
struct task_grp;
|
struct task_group;
|
||||||
#ifdef CONFIG_SCHED_DEBUG
|
#ifdef CONFIG_SCHED_DEBUG
|
||||||
extern void proc_sched_show_task(struct task_struct *p, struct seq_file *m);
|
extern void proc_sched_show_task(struct task_struct *p, struct seq_file *m);
|
||||||
extern void proc_sched_set_task(struct task_struct *p);
|
extern void proc_sched_set_task(struct task_struct *p);
|
||||||
|
@ -598,7 +598,7 @@ struct user_struct {
|
||||||
uid_t uid;
|
uid_t uid;
|
||||||
|
|
||||||
#ifdef CONFIG_FAIR_USER_SCHED
|
#ifdef CONFIG_FAIR_USER_SCHED
|
||||||
struct task_grp *tg;
|
struct task_group *tg;
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -1842,12 +1842,12 @@ extern void normalize_rt_tasks(void);
|
||||||
|
|
||||||
#ifdef CONFIG_FAIR_GROUP_SCHED
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
||||||
|
|
||||||
extern struct task_grp init_task_grp;
|
extern struct task_group init_task_group;
|
||||||
|
|
||||||
extern struct task_grp *sched_create_group(void);
|
extern struct task_group *sched_create_group(void);
|
||||||
extern void sched_destroy_group(struct task_grp *tg);
|
extern void sched_destroy_group(struct task_group *tg);
|
||||||
extern void sched_move_task(struct task_struct *tsk);
|
extern void sched_move_task(struct task_struct *tsk);
|
||||||
extern int sched_group_set_shares(struct task_grp *tg, unsigned long shares);
|
extern int sched_group_set_shares(struct task_group *tg, unsigned long shares);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -156,7 +156,7 @@ struct rt_prio_array {
|
||||||
struct cfs_rq;
|
struct cfs_rq;
|
||||||
|
|
||||||
/* task group related information */
|
/* task group related information */
|
||||||
struct task_grp {
|
struct task_group {
|
||||||
/* schedulable entities of this group on each cpu */
|
/* schedulable entities of this group on each cpu */
|
||||||
struct sched_entity **se;
|
struct sched_entity **se;
|
||||||
/* runqueue "owned" by this group on each cpu */
|
/* runqueue "owned" by this group on each cpu */
|
||||||
|
@ -175,7 +175,7 @@ static struct cfs_rq *init_cfs_rq_p[NR_CPUS];
|
||||||
/* Default task group.
|
/* Default task group.
|
||||||
* Every task in system belong to this group at bootup.
|
* Every task in system belong to this group at bootup.
|
||||||
*/
|
*/
|
||||||
struct task_grp init_task_grp = {
|
struct task_group init_task_group = {
|
||||||
.se = init_sched_entity_p,
|
.se = init_sched_entity_p,
|
||||||
.cfs_rq = init_cfs_rq_p,
|
.cfs_rq = init_cfs_rq_p,
|
||||||
};
|
};
|
||||||
|
@ -186,17 +186,17 @@ struct task_grp init_task_grp = {
|
||||||
# define INIT_TASK_GRP_LOAD NICE_0_LOAD
|
# define INIT_TASK_GRP_LOAD NICE_0_LOAD
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static int init_task_grp_load = INIT_TASK_GRP_LOAD;
|
static int init_task_group_load = INIT_TASK_GRP_LOAD;
|
||||||
|
|
||||||
/* return group to which a task belongs */
|
/* return group to which a task belongs */
|
||||||
static inline struct task_grp *task_grp(struct task_struct *p)
|
static inline struct task_group *task_group(struct task_struct *p)
|
||||||
{
|
{
|
||||||
struct task_grp *tg;
|
struct task_group *tg;
|
||||||
|
|
||||||
#ifdef CONFIG_FAIR_USER_SCHED
|
#ifdef CONFIG_FAIR_USER_SCHED
|
||||||
tg = p->user->tg;
|
tg = p->user->tg;
|
||||||
#else
|
#else
|
||||||
tg = &init_task_grp;
|
tg = &init_task_group;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
return tg;
|
return tg;
|
||||||
|
@ -205,8 +205,8 @@ static inline struct task_grp *task_grp(struct task_struct *p)
|
||||||
/* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */
|
/* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */
|
||||||
static inline void set_task_cfs_rq(struct task_struct *p)
|
static inline void set_task_cfs_rq(struct task_struct *p)
|
||||||
{
|
{
|
||||||
p->se.cfs_rq = task_grp(p)->cfs_rq[task_cpu(p)];
|
p->se.cfs_rq = task_group(p)->cfs_rq[task_cpu(p)];
|
||||||
p->se.parent = task_grp(p)->se[task_cpu(p)];
|
p->se.parent = task_group(p)->se[task_cpu(p)];
|
||||||
}
|
}
|
||||||
|
|
||||||
#else
|
#else
|
||||||
|
@ -244,7 +244,7 @@ struct cfs_rq {
|
||||||
* list is used during load balance.
|
* list is used during load balance.
|
||||||
*/
|
*/
|
||||||
struct list_head leaf_cfs_rq_list; /* Better name : task_cfs_rq_list? */
|
struct list_head leaf_cfs_rq_list; /* Better name : task_cfs_rq_list? */
|
||||||
struct task_grp *tg; /* group that "owns" this runqueue */
|
struct task_group *tg; /* group that "owns" this runqueue */
|
||||||
struct rcu_head rcu;
|
struct rcu_head rcu;
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
@ -6522,19 +6522,19 @@ void __init sched_init(void)
|
||||||
|
|
||||||
init_cfs_rq_p[i] = cfs_rq;
|
init_cfs_rq_p[i] = cfs_rq;
|
||||||
init_cfs_rq(cfs_rq, rq);
|
init_cfs_rq(cfs_rq, rq);
|
||||||
cfs_rq->tg = &init_task_grp;
|
cfs_rq->tg = &init_task_group;
|
||||||
list_add(&cfs_rq->leaf_cfs_rq_list,
|
list_add(&cfs_rq->leaf_cfs_rq_list,
|
||||||
&rq->leaf_cfs_rq_list);
|
&rq->leaf_cfs_rq_list);
|
||||||
|
|
||||||
init_sched_entity_p[i] = se;
|
init_sched_entity_p[i] = se;
|
||||||
se->cfs_rq = &rq->cfs;
|
se->cfs_rq = &rq->cfs;
|
||||||
se->my_q = cfs_rq;
|
se->my_q = cfs_rq;
|
||||||
se->load.weight = init_task_grp_load;
|
se->load.weight = init_task_group_load;
|
||||||
se->load.inv_weight =
|
se->load.inv_weight =
|
||||||
div64_64(1ULL<<32, init_task_grp_load);
|
div64_64(1ULL<<32, init_task_group_load);
|
||||||
se->parent = NULL;
|
se->parent = NULL;
|
||||||
}
|
}
|
||||||
init_task_grp.shares = init_task_grp_load;
|
init_task_group.shares = init_task_group_load;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
for (j = 0; j < CPU_LOAD_IDX_MAX; j++)
|
for (j = 0; j < CPU_LOAD_IDX_MAX; j++)
|
||||||
|
@ -6725,9 +6725,9 @@ void set_curr_task(int cpu, struct task_struct *p)
|
||||||
#ifdef CONFIG_FAIR_GROUP_SCHED
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
||||||
|
|
||||||
/* allocate runqueue etc for a new task group */
|
/* allocate runqueue etc for a new task group */
|
||||||
struct task_grp *sched_create_group(void)
|
struct task_group *sched_create_group(void)
|
||||||
{
|
{
|
||||||
struct task_grp *tg;
|
struct task_group *tg;
|
||||||
struct cfs_rq *cfs_rq;
|
struct cfs_rq *cfs_rq;
|
||||||
struct sched_entity *se;
|
struct sched_entity *se;
|
||||||
struct rq *rq;
|
struct rq *rq;
|
||||||
|
@ -6800,7 +6800,7 @@ err:
|
||||||
static void free_sched_group(struct rcu_head *rhp)
|
static void free_sched_group(struct rcu_head *rhp)
|
||||||
{
|
{
|
||||||
struct cfs_rq *cfs_rq = container_of(rhp, struct cfs_rq, rcu);
|
struct cfs_rq *cfs_rq = container_of(rhp, struct cfs_rq, rcu);
|
||||||
struct task_grp *tg = cfs_rq->tg;
|
struct task_group *tg = cfs_rq->tg;
|
||||||
struct sched_entity *se;
|
struct sched_entity *se;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
|
@ -6819,7 +6819,7 @@ static void free_sched_group(struct rcu_head *rhp)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Destroy runqueue etc associated with a task group */
|
/* Destroy runqueue etc associated with a task group */
|
||||||
void sched_destroy_group(struct task_grp *tg)
|
void sched_destroy_group(struct task_group *tg)
|
||||||
{
|
{
|
||||||
struct cfs_rq *cfs_rq;
|
struct cfs_rq *cfs_rq;
|
||||||
int i;
|
int i;
|
||||||
|
@ -6895,7 +6895,7 @@ static void set_se_shares(struct sched_entity *se, unsigned long shares)
|
||||||
spin_unlock_irq(&rq->lock);
|
spin_unlock_irq(&rq->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
int sched_group_set_shares(struct task_grp *tg, unsigned long shares)
|
int sched_group_set_shares(struct task_group *tg, unsigned long shares)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
|
|
|
@ -239,7 +239,7 @@ static int
|
||||||
root_user_share_read_proc(char *page, char **start, off_t off, int count,
|
root_user_share_read_proc(char *page, char **start, off_t off, int count,
|
||||||
int *eof, void *data)
|
int *eof, void *data)
|
||||||
{
|
{
|
||||||
return sprintf(page, "%d\n", init_task_grp_load);
|
return sprintf(page, "%d\n", init_task_group_load);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
|
@ -260,8 +260,8 @@ root_user_share_write_proc(struct file *file, const char __user *buffer,
|
||||||
|
|
||||||
mutex_lock(&root_user_share_mutex);
|
mutex_lock(&root_user_share_mutex);
|
||||||
|
|
||||||
init_task_grp_load = shares;
|
init_task_group_load = shares;
|
||||||
rc = sched_group_set_shares(&init_task_grp, shares);
|
rc = sched_group_set_shares(&init_task_group, shares);
|
||||||
|
|
||||||
mutex_unlock(&root_user_share_mutex);
|
mutex_unlock(&root_user_share_mutex);
|
||||||
|
|
||||||
|
|
|
@ -51,7 +51,7 @@ struct user_struct root_user = {
|
||||||
.session_keyring = &root_session_keyring,
|
.session_keyring = &root_session_keyring,
|
||||||
#endif
|
#endif
|
||||||
#ifdef CONFIG_FAIR_USER_SCHED
|
#ifdef CONFIG_FAIR_USER_SCHED
|
||||||
.tg = &init_task_grp,
|
.tg = &init_task_group,
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue