summaryrefslogtreecommitdiff
path: root/kernel
diff options
context:
space:
mode:
authorIngo Molnar <mingo@elte.hu>2007-10-15 17:00:12 +0200
committerIngo Molnar <mingo@elte.hu>2007-10-15 17:00:12 +0200
commit3a2520157234d58abce89526756a32c272824f3f (patch)
tree5a6ffe45173a6f6c5df03a66b6d82a53f52f1460 /kernel
parent5522d5d5f70005faeffff3ffc0cfa8eec0155de4 (diff)
sched: whitespace cleanups
more whitespace cleanups. No code changed: text data bss dec hex filename 26553 2790 288 29631 73bf sched.o.before 26553 2790 288 29631 73bf sched.o.after Signed-off-by: Ingo Molnar <mingo@elte.hu> Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched.c52
1 files changed, 26 insertions, 26 deletions
diff --git a/kernel/sched.c b/kernel/sched.c
index f582e2cedb09..e717047be5cf 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -193,17 +193,17 @@ static struct sched_entity *init_sched_entity_p[NR_CPUS];
static struct cfs_rq *init_cfs_rq_p[NR_CPUS];
/* Default task group.
- * Every task in system belong to this group at bootup.
+ * Every task in system belong to this group at bootup.
*/
-struct task_grp init_task_grp = {
- .se = init_sched_entity_p,
- .cfs_rq = init_cfs_rq_p,
- };
+struct task_grp init_task_grp = {
+ .se = init_sched_entity_p,
+ .cfs_rq = init_cfs_rq_p,
+};
#ifdef CONFIG_FAIR_USER_SCHED
-#define INIT_TASK_GRP_LOAD 2*NICE_0_LOAD
+# define INIT_TASK_GRP_LOAD 2*NICE_0_LOAD
#else
-#define INIT_TASK_GRP_LOAD NICE_0_LOAD
+# define INIT_TASK_GRP_LOAD NICE_0_LOAD
#endif
static int init_task_grp_load = INIT_TASK_GRP_LOAD;
@@ -6516,25 +6516,25 @@ void __init sched_init(void)
init_cfs_rq(&rq->cfs, rq);
#ifdef CONFIG_FAIR_GROUP_SCHED
INIT_LIST_HEAD(&rq->leaf_cfs_rq_list);
- {
- struct cfs_rq *cfs_rq = &per_cpu(init_cfs_rq, i);
- struct sched_entity *se =
- &per_cpu(init_sched_entity, i);
-
- init_cfs_rq_p[i] = cfs_rq;
- init_cfs_rq(cfs_rq, rq);
- cfs_rq->tg = &init_task_grp;
- list_add(&cfs_rq->leaf_cfs_rq_list,
+ {
+ struct cfs_rq *cfs_rq = &per_cpu(init_cfs_rq, i);
+ struct sched_entity *se =
+ &per_cpu(init_sched_entity, i);
+
+ init_cfs_rq_p[i] = cfs_rq;
+ init_cfs_rq(cfs_rq, rq);
+ cfs_rq->tg = &init_task_grp;
+ list_add(&cfs_rq->leaf_cfs_rq_list,
&rq->leaf_cfs_rq_list);
- init_sched_entity_p[i] = se;
- se->cfs_rq = &rq->cfs;
- se->my_q = cfs_rq;
- se->load.weight = init_task_grp_load;
+ init_sched_entity_p[i] = se;
+ se->cfs_rq = &rq->cfs;
+ se->my_q = cfs_rq;
+ se->load.weight = init_task_grp_load;
se->load.inv_weight =
div64_64(1ULL<<32, init_task_grp_load);
- se->parent = NULL;
- }
+ se->parent = NULL;
+ }
init_task_grp.shares = init_task_grp_load;
#endif
@@ -6840,9 +6840,9 @@ void sched_destroy_group(struct task_grp *tg)
}
/* change task's runqueue when it moves between groups.
- * The caller of this function should have put the task in its new group
- * by now. This function just updates tsk->se.cfs_rq and tsk->se.parent to
- * reflect its new group.
+ * The caller of this function should have put the task in its new group
+ * by now. This function just updates tsk->se.cfs_rq and tsk->se.parent to
+ * reflect its new group.
*/
void sched_move_task(struct task_struct *tsk)
{
@@ -6915,4 +6915,4 @@ int sched_group_set_shares(struct task_grp *tg, unsigned long shares)
return 0;
}
-#endif /* CONFIG_FAIR_GROUP_SCHED */
+#endif /* CONFIG_FAIR_GROUP_SCHED */