2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* kernel/sched.c
|
|
|
|
*
|
|
|
|
* Kernel scheduler and related syscalls
|
|
|
|
*
|
|
|
|
* Copyright (C) 1991-2002 Linus Torvalds
|
|
|
|
*
|
|
|
|
* 1996-12-23 Modified by Dave Grothe to fix bugs in semaphores and
|
|
|
|
* make semaphores SMP safe
|
|
|
|
* 1998-11-19 Implemented schedule_timeout() and related stuff
|
|
|
|
* by Andrea Arcangeli
|
|
|
|
* 2002-01-04 New ultra-scalable O(1) scheduler by Ingo Molnar:
|
|
|
|
* hybrid priority-list and round-robin design with
|
|
|
|
* an array-switch method of distributing timeslices
|
|
|
|
* and per-CPU runqueues. Cleanups and useful suggestions
|
|
|
|
* by Davide Libenzi, preemptible kernel bits by Robert Love.
|
|
|
|
* 2003-09-03 Interactivity tuning by Con Kolivas.
|
|
|
|
* 2004-04-02 Scheduler domains code by Nick Piggin
|
2007-07-09 16:52:01 +00:00
|
|
|
* 2007-04-15 Work begun on replacing all interactivity tuning with a
|
|
|
|
* fair scheduling design by Con Kolivas.
|
|
|
|
* 2007-05-05 Load balancing (smp-nice) and other improvements
|
|
|
|
* by Peter Williams
|
|
|
|
* 2007-05-06 Interactivity improvements to CFS by Mike Galbraith
|
|
|
|
* 2007-07-01 Group scheduling enhancements by Srivatsa Vaddagiri
|
2008-01-25 20:08:19 +00:00
|
|
|
* 2007-11-29 RT balancing improvements by Steven Rostedt, Gregory Haskins,
|
|
|
|
* Thomas Gleixner, Mike Kravetz
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/nmi.h>
|
|
|
|
#include <linux/init.h>
|
2007-07-09 16:52:00 +00:00
|
|
|
#include <linux/uaccess.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/highmem.h>
|
|
|
|
#include <linux/smp_lock.h>
|
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
#include <linux/interrupt.h>
|
2006-01-11 20:17:46 +00:00
|
|
|
#include <linux/capability.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/completion.h>
|
|
|
|
#include <linux/kernel_stat.h>
|
2006-07-03 07:24:33 +00:00
|
|
|
#include <linux/debug_locks.h>
|
2009-05-04 17:13:30 +00:00
|
|
|
#include <linux/perf_counter.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/security.h>
|
|
|
|
#include <linux/notifier.h>
|
|
|
|
#include <linux/profile.h>
|
2006-12-07 04:34:23 +00:00
|
|
|
#include <linux/freezer.h>
|
[PATCH] scheduler cache-hot-autodetect
)
From: Ingo Molnar <mingo@elte.hu>
This is the latest version of the scheduler cache-hot-auto-tune patch.
The first problem was that detection time scaled with O(N^2), which is
unacceptable on larger SMP and NUMA systems. To solve this:
- I've added a 'domain distance' function, which is used to cache
measurement results. Each distance is only measured once. This means
that e.g. on NUMA distances of 0, 1 and 2 might be measured, on HT
distances 0 and 1, and on SMP distance 0 is measured. The code walks
the domain tree to determine the distance, so it automatically follows
whatever hierarchy an architecture sets up. This cuts down on the boot
time significantly and removes the O(N^2) limit. The only assumption
is that migration costs can be expressed as a function of domain
distance - this covers the overwhelming majority of existing systems,
and is a good guess even for more assymetric systems.
[ People hacking systems that have assymetries that break this
assumption (e.g. different CPU speeds) should experiment a bit with
the cpu_distance() function. Adding a ->migration_distance factor to
the domain structure would be one possible solution - but lets first
see the problem systems, if they exist at all. Lets not overdesign. ]
Another problem was that only a single cache-size was used for measuring
the cost of migration, and most architectures didnt set that variable
up. Furthermore, a single cache-size does not fit NUMA hierarchies with
L3 caches and does not fit HT setups, where different CPUs will often
have different 'effective cache sizes'. To solve this problem:
- Instead of relying on a single cache-size provided by the platform and
sticking to it, the code now auto-detects the 'effective migration
cost' between two measured CPUs, via iterating through a wide range of
cachesizes. The code searches for the maximum migration cost, which
occurs when the working set of the test-workload falls just below the
'effective cache size'. I.e. real-life optimized search is done for
the maximum migration cost, between two real CPUs.
This, amongst other things, has the positive effect hat if e.g. two
CPUs share a L2/L3 cache, a different (and accurate) migration cost
will be found than between two CPUs on the same system that dont share
any caches.
(The reliable measurement of migration costs is tricky - see the source
for details.)
Furthermore i've added various boot-time options to override/tune
migration behavior.
Firstly, there's a blanket override for autodetection:
migration_cost=1000,2000,3000
will override the depth 0/1/2 values with 1msec/2msec/3msec values.
Secondly, there's a global factor that can be used to increase (or
decrease) the autodetected values:
migration_factor=120
will increase the autodetected values by 20%. This option is useful to
tune things in a workload-dependent way - e.g. if a workload is
cache-insensitive then CPU utilization can be maximized by specifying
migration_factor=0.
I've tested the autodetection code quite extensively on x86, on 3
P3/Xeon/2MB, and the autodetected values look pretty good:
Dual Celeron (128K L2 cache):
---------------------
migration cost matrix (max_cache_size: 131072, cpu: 467 MHz):
---------------------
[00] [01]
[00]: - 1.7(1)
[01]: 1.7(1) -
---------------------
cacheflush times [2]: 0.0 (0) 1.7 (1784008)
---------------------
Here the slow memory subsystem dominates system performance, and even
though caches are small, the migration cost is 1.7 msecs.
Dual HT P4 (512K L2 cache):
---------------------
migration cost matrix (max_cache_size: 524288, cpu: 2379 MHz):
---------------------
[00] [01] [02] [03]
[00]: - 0.4(1) 0.0(0) 0.4(1)
[01]: 0.4(1) - 0.4(1) 0.0(0)
[02]: 0.0(0) 0.4(1) - 0.4(1)
[03]: 0.4(1) 0.0(0) 0.4(1) -
---------------------
cacheflush times [2]: 0.0 (33900) 0.4 (448514)
---------------------
Here it can be seen that there is no migration cost between two HT
siblings (CPU#0/2 and CPU#1/3 are separate physical CPUs). A fast memory
system makes inter-physical-CPU migration pretty cheap: 0.4 msecs.
8-way P3/Xeon [2MB L2 cache]:
---------------------
migration cost matrix (max_cache_size: 2097152, cpu: 700 MHz):
---------------------
[00] [01] [02] [03] [04] [05] [06] [07]
[00]: - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[01]: 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[02]: 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[03]: 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[04]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1)
[05]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1)
[06]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1)
[07]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) -
---------------------
cacheflush times [2]: 0.0 (0) 19.2 (19281756)
---------------------
This one has huge caches and a relatively slow memory subsystem - so the
migration cost is 19 msecs.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Ashok Raj <ashok.raj@intel.com>
Signed-off-by: Ken Chen <kenneth.w.chen@intel.com>
Cc: <wilder@us.ibm.com>
Signed-off-by: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-01-12 09:05:30 +00:00
|
|
|
#include <linux/vmalloc.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/blkdev.h>
|
|
|
|
#include <linux/delay.h>
|
2007-10-19 06:40:14 +00:00
|
|
|
#include <linux/pid_namespace.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/smp.h>
|
|
|
|
#include <linux/threads.h>
|
|
|
|
#include <linux/timer.h>
|
|
|
|
#include <linux/rcupdate.h>
|
|
|
|
#include <linux/cpu.h>
|
|
|
|
#include <linux/cpuset.h>
|
|
|
|
#include <linux/percpu.h>
|
|
|
|
#include <linux/kthread.h>
|
2008-10-06 09:23:43 +00:00
|
|
|
#include <linux/proc_fs.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/seq_file.h>
|
2007-07-26 11:40:43 +00:00
|
|
|
#include <linux/sysctl.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/syscalls.h>
|
|
|
|
#include <linux/times.h>
|
2006-10-01 06:28:59 +00:00
|
|
|
#include <linux/tsacct_kern.h>
|
2006-03-26 09:38:20 +00:00
|
|
|
#include <linux/kprobes.h>
|
2006-07-14 07:24:37 +00:00
|
|
|
#include <linux/delayacct.h>
|
2007-05-08 07:32:57 +00:00
|
|
|
#include <linux/reciprocal_div.h>
|
2007-07-09 16:52:00 +00:00
|
|
|
#include <linux/unistd.h>
|
2007-09-21 07:19:54 +00:00
|
|
|
#include <linux/pagemap.h>
|
2008-01-25 20:08:29 +00:00
|
|
|
#include <linux/hrtimer.h>
|
2008-03-17 23:19:05 +00:00
|
|
|
#include <linux/tick.h>
|
2008-04-19 17:45:00 +00:00
|
|
|
#include <linux/debugfs.h>
|
|
|
|
#include <linux/ctype.h>
|
2008-05-12 19:20:42 +00:00
|
|
|
#include <linux/ftrace.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-05-08 07:32:57 +00:00
|
|
|
#include <asm/tlb.h>
|
2007-10-24 16:23:50 +00:00
|
|
|
#include <asm/irq_regs.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-05-12 19:21:01 +00:00
|
|
|
#include "sched_cpupri.h"
|
|
|
|
|
2009-04-10 13:36:00 +00:00
|
|
|
#define CREATE_TRACE_POINTS
|
2009-04-14 23:39:12 +00:00
|
|
|
#include <trace/events/sched.h>
|
2009-04-10 13:36:00 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Convert user-nice values [ -20 ... 0 ... 19 ]
|
|
|
|
* to static priority [ MAX_RT_PRIO..MAX_PRIO-1 ],
|
|
|
|
* and back.
|
|
|
|
*/
|
|
|
|
#define NICE_TO_PRIO(nice) (MAX_RT_PRIO + (nice) + 20)
|
|
|
|
#define PRIO_TO_NICE(prio) ((prio) - MAX_RT_PRIO - 20)
|
|
|
|
#define TASK_NICE(p) PRIO_TO_NICE((p)->static_prio)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 'User priority' is the nice value converted to something we
|
|
|
|
* can work with better when scaling various scheduler parameters,
|
|
|
|
* it's a [ 0 ... 39 ] range.
|
|
|
|
*/
|
|
|
|
#define USER_PRIO(p) ((p)-MAX_RT_PRIO)
|
|
|
|
#define TASK_USER_PRIO(p) USER_PRIO((p)->static_prio)
|
|
|
|
#define MAX_USER_PRIO (USER_PRIO(MAX_PRIO))
|
|
|
|
|
|
|
|
/*
|
2008-01-25 20:08:19 +00:00
|
|
|
* Helpers for converting nanosecond timing to jiffy resolution
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-11-09 21:39:38 +00:00
|
|
|
#define NS_TO_JIFFIES(TIME) ((unsigned long)(TIME) / (NSEC_PER_SEC / HZ))
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
#define NICE_0_LOAD SCHED_LOAD_SCALE
|
|
|
|
#define NICE_0_SHIFT SCHED_LOAD_SHIFT
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* These are the 'tuning knobs' of the scheduler:
|
|
|
|
*
|
2007-10-15 15:00:13 +00:00
|
|
|
* default timeslice is 100 msecs (used only for SCHED_RR tasks).
|
2005-04-16 22:20:36 +00:00
|
|
|
* Timeslices get refilled after they expire.
|
|
|
|
*/
|
|
|
|
#define DEF_TIMESLICE (100 * HZ / 1000)
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
/*
|
|
|
|
* single value that denotes runtime == period, ie unlimited time.
|
|
|
|
*/
|
|
|
|
#define RUNTIME_INF ((u64)~0ULL)
|
|
|
|
|
2007-05-08 07:32:57 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2009-01-11 09:04:22 +00:00
|
|
|
|
|
|
|
static void double_rq_lock(struct rq *rq1, struct rq *rq2);
|
|
|
|
|
2007-05-08 07:32:57 +00:00
|
|
|
/*
|
|
|
|
* Divide a load by a sched group cpu_power : (load / sg->__cpu_power)
|
|
|
|
* Since cpu_power is a 'constant', we can use a reciprocal divide.
|
|
|
|
*/
|
|
|
|
static inline u32 sg_div_cpu_power(const struct sched_group *sg, u32 load)
|
|
|
|
{
|
|
|
|
return reciprocal_divide(load, sg->reciprocal_cpu_power);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Each time a sched group cpu_power is changed,
|
|
|
|
* we must compute its reciprocal value
|
|
|
|
*/
|
|
|
|
static inline void sg_inc_cpu_power(struct sched_group *sg, u32 val)
|
|
|
|
{
|
|
|
|
sg->__cpu_power += val;
|
|
|
|
sg->reciprocal_cpu_power = reciprocal_value(sg->__cpu_power);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
static inline int rt_policy(int policy)
|
|
|
|
{
|
2008-05-13 21:44:11 +00:00
|
|
|
if (unlikely(policy == SCHED_FIFO || policy == SCHED_RR))
|
2007-07-09 16:51:59 +00:00
|
|
|
return 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int task_has_rt_policy(struct task_struct *p)
|
|
|
|
{
|
|
|
|
return rt_policy(p->policy);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-07-09 16:51:58 +00:00
|
|
|
* This is the priority-queue data structure of the RT scheduling class:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:58 +00:00
|
|
|
struct rt_prio_array {
|
|
|
|
DECLARE_BITMAP(bitmap, MAX_RT_PRIO+1); /* include 1 bit for delimiter */
|
|
|
|
struct list_head queue[MAX_RT_PRIO];
|
|
|
|
};
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
struct rt_bandwidth {
|
2008-03-25 12:51:45 +00:00
|
|
|
/* nests inside the rq lock: */
|
|
|
|
spinlock_t rt_runtime_lock;
|
|
|
|
ktime_t rt_period;
|
|
|
|
u64 rt_runtime;
|
|
|
|
struct hrtimer rt_period_timer;
|
2008-04-19 17:44:57 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct rt_bandwidth def_rt_bandwidth;
|
|
|
|
|
|
|
|
static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun);
|
|
|
|
|
|
|
|
static enum hrtimer_restart sched_rt_period_timer(struct hrtimer *timer)
|
|
|
|
{
|
|
|
|
struct rt_bandwidth *rt_b =
|
|
|
|
container_of(timer, struct rt_bandwidth, rt_period_timer);
|
|
|
|
ktime_t now;
|
|
|
|
int overrun;
|
|
|
|
int idle = 0;
|
|
|
|
|
|
|
|
for (;;) {
|
|
|
|
now = hrtimer_cb_get_time(timer);
|
|
|
|
overrun = hrtimer_forward(timer, now, rt_b->rt_period);
|
|
|
|
|
|
|
|
if (!overrun)
|
|
|
|
break;
|
|
|
|
|
|
|
|
idle = do_sched_rt_period_timer(rt_b, overrun);
|
|
|
|
}
|
|
|
|
|
|
|
|
return idle ? HRTIMER_NORESTART : HRTIMER_RESTART;
|
|
|
|
}
|
|
|
|
|
|
|
|
static
|
|
|
|
void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime)
|
|
|
|
{
|
|
|
|
rt_b->rt_period = ns_to_ktime(period);
|
|
|
|
rt_b->rt_runtime = runtime;
|
|
|
|
|
2008-04-19 17:44:58 +00:00
|
|
|
spin_lock_init(&rt_b->rt_runtime_lock);
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
hrtimer_init(&rt_b->rt_period_timer,
|
|
|
|
CLOCK_MONOTONIC, HRTIMER_MODE_REL);
|
|
|
|
rt_b->rt_period_timer.function = sched_rt_period_timer;
|
|
|
|
}
|
|
|
|
|
2008-09-05 21:46:19 +00:00
|
|
|
static inline int rt_bandwidth_enabled(void)
|
|
|
|
{
|
|
|
|
return sysctl_sched_rt_runtime >= 0;
|
2008-04-19 17:44:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void start_rt_bandwidth(struct rt_bandwidth *rt_b)
|
|
|
|
{
|
|
|
|
ktime_t now;
|
|
|
|
|
2009-02-25 17:59:26 +00:00
|
|
|
if (!rt_bandwidth_enabled() || rt_b->rt_runtime == RUNTIME_INF)
|
2008-04-19 17:44:57 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
if (hrtimer_active(&rt_b->rt_period_timer))
|
|
|
|
return;
|
|
|
|
|
|
|
|
spin_lock(&rt_b->rt_runtime_lock);
|
|
|
|
for (;;) {
|
2009-03-13 11:21:27 +00:00
|
|
|
unsigned long delta;
|
|
|
|
ktime_t soft, hard;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
if (hrtimer_active(&rt_b->rt_period_timer))
|
|
|
|
break;
|
|
|
|
|
|
|
|
now = hrtimer_cb_get_time(&rt_b->rt_period_timer);
|
|
|
|
hrtimer_forward(&rt_b->rt_period_timer, now, rt_b->rt_period);
|
2009-03-13 11:21:27 +00:00
|
|
|
|
|
|
|
soft = hrtimer_get_softexpires(&rt_b->rt_period_timer);
|
|
|
|
hard = hrtimer_get_expires(&rt_b->rt_period_timer);
|
|
|
|
delta = ktime_to_ns(ktime_sub(hard, soft));
|
|
|
|
__hrtimer_start_range_ns(&rt_b->rt_period_timer, soft, delta,
|
2009-04-16 06:44:37 +00:00
|
|
|
HRTIMER_MODE_ABS_PINNED, 0);
|
2008-04-19 17:44:57 +00:00
|
|
|
}
|
|
|
|
spin_unlock(&rt_b->rt_runtime_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
static void destroy_rt_bandwidth(struct rt_bandwidth *rt_b)
|
|
|
|
{
|
|
|
|
hrtimer_cancel(&rt_b->rt_period_timer);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-04-28 09:33:07 +00:00
|
|
|
/*
|
|
|
|
* sched_domains_mutex serializes calls to arch_init_sched_domains,
|
|
|
|
* detach_destroy_domains and partition_sched_domains.
|
|
|
|
*/
|
|
|
|
static DEFINE_MUTEX(sched_domains_mutex);
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_GROUP_SCHED
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2007-10-19 06:41:03 +00:00
|
|
|
#include <linux/cgroup.h>
|
|
|
|
|
2007-10-15 15:00:07 +00:00
|
|
|
struct cfs_rq;
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
static LIST_HEAD(task_groups);
|
|
|
|
|
2007-10-15 15:00:07 +00:00
|
|
|
/* task group related information */
|
2007-10-15 15:00:14 +00:00
|
|
|
struct task_group {
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_CGROUP_SCHED
|
2007-10-19 06:41:03 +00:00
|
|
|
struct cgroup_subsys_state css;
|
|
|
|
#endif
|
2008-02-13 14:45:40 +00:00
|
|
|
|
2008-12-01 15:19:05 +00:00
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
uid_t uid;
|
|
|
|
#endif
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-10-15 15:00:07 +00:00
|
|
|
/* schedulable entities of this group on each cpu */
|
|
|
|
struct sched_entity **se;
|
|
|
|
/* runqueue "owned" by this group on each cpu */
|
|
|
|
struct cfs_rq **cfs_rq;
|
|
|
|
unsigned long shares;
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
struct sched_rt_entity **rt_se;
|
|
|
|
struct rt_rq **rt_rq;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
struct rt_bandwidth rt_bandwidth;
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
sched: group scheduler, fix fairness of cpu bandwidth allocation for task groups
The current load balancing scheme isn't good enough for precise
group fairness.
For example: on a 8-cpu system, I created 3 groups as under:
a = 8 tasks (cpu.shares = 1024)
b = 4 tasks (cpu.shares = 1024)
c = 3 tasks (cpu.shares = 1024)
a, b and c are task groups that have equal weight. We would expect each
of the groups to receive 33.33% of cpu bandwidth under a fair scheduler.
This is what I get with the latest scheduler git tree:
Signed-off-by: Ingo Molnar <mingo@elte.hu>
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 277.676 | 57.8% | 54.1% 54.1% 54.1% 54.2% 56.7% 62.2% 62.8% 64.5%
b | 116.108 | 24.2% | 47.4% 48.1% 48.7% 49.3%
c | 86.326 | 18.0% | 47.5% 47.9% 48.5%
--------------------------------------------------------------------------------
Explanation of o/p:
Col1 -> Group name
Col2 -> Cumulative execution time (in seconds) received by all tasks of that
group in a 60sec window across 8 cpus
Col3 -> CPU bandwidth received by the group in the 60sec window, expressed in
percentage. Col3 data is derived as:
Col3 = 100 * Col2 / (NR_CPUS * 60)
Col4 -> CPU bandwidth received by each individual task of the group.
Col4 = 100 * cpu_time_recd_by_task / 60
[I can share the test case that produces a similar o/p if reqd]
The deviation from desired group fairness is as below:
a = +24.47%
b = -9.13%
c = -15.33%
which is quite high.
After the patch below is applied, here are the results:
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 163.112 | 34.0% | 33.2% 33.4% 33.5% 33.5% 33.7% 34.4% 34.8% 35.3%
b | 156.220 | 32.5% | 63.3% 64.5% 66.1% 66.5%
c | 160.653 | 33.5% | 85.8% 90.6% 91.4%
--------------------------------------------------------------------------------
Deviation from desired group fairness is as below:
a = +0.67%
b = -0.83%
c = +0.17%
which is far better IMO. Most of other runs have yielded a deviation within
+-2% at the most, which is good.
Why do we see bad (group) fairness with current scheuler?
=========================================================
Currently cpu's weight is just the summation of individual task weights.
This can yield incorrect results. For ex: consider three groups as below
on a 2-cpu system:
CPU0 CPU1
---------------------------
A (10) B(5)
C(5)
---------------------------
Group A has 10 tasks, all on CPU0, Group B and C have 5 tasks each all
of which are on CPU1. Each task has the same weight (NICE_0_LOAD =
1024).
The current scheme would yield a cpu weight of 10240 (10*1024) for each cpu and
the load balancer will think both CPUs are perfectly balanced and won't
move around any tasks. This, however, would yield this bandwidth:
A = 50%
B = 25%
C = 25%
which is not the desired result.
What's changing in the patch?
=============================
- How cpu weights are calculated when CONFIF_FAIR_GROUP_SCHED is
defined (see below)
- API Change
- Two tunables introduced in sysfs (under SCHED_DEBUG) to
control the frequency at which the load balance monitor
thread runs.
The basic change made in this patch is how cpu weight (rq->load.weight) is
calculated. Its now calculated as the summation of group weights on a cpu,
rather than summation of task weights. Weight exerted by a group on a
cpu is dependent on the shares allocated to it and also the number of
tasks the group has on that cpu compared to the total number of
(runnable) tasks the group has in the system.
Let,
W(K,i) = Weight of group K on cpu i
T(K,i) = Task load present in group K's cfs_rq on cpu i
T(K) = Total task load of group K across various cpus
S(K) = Shares allocated to group K
NRCPUS = Number of online cpus in the scheduler domain to
which group K is assigned.
Then,
W(K,i) = S(K) * NRCPUS * T(K,i) / T(K)
A load balance monitor thread is created at bootup, which periodically
runs and adjusts group's weight on each cpu. To avoid its overhead, two
min/max tunables are introduced (under SCHED_DEBUG) to control the rate
at which it runs.
Fixes from: Peter Zijlstra <a.p.zijlstra@chello.nl>
- don't start the load_balance_monitor when there is only a single cpu.
- rename the kthread because its currently longer than TASK_COMM_LEN
Signed-off-by: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-01-25 20:08:00 +00:00
|
|
|
|
2007-10-29 20:18:11 +00:00
|
|
|
struct rcu_head rcu;
|
2008-01-25 20:08:30 +00:00
|
|
|
struct list_head list;
|
2008-04-19 17:45:00 +00:00
|
|
|
|
|
|
|
struct task_group *parent;
|
|
|
|
struct list_head siblings;
|
|
|
|
struct list_head children;
|
2007-10-15 15:00:07 +00:00
|
|
|
};
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
#ifdef CONFIG_USER_SCHED
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-12-01 15:19:05 +00:00
|
|
|
/* Helper function to pass uid information to create_sched_user() */
|
|
|
|
void set_tg_uid(struct user_struct *user)
|
|
|
|
{
|
|
|
|
user->tg->uid = user->uid;
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
/*
|
|
|
|
* Root task group.
|
|
|
|
* Every UID task group (including init_task_group aka UID-0) will
|
|
|
|
* be a child to this group.
|
|
|
|
*/
|
|
|
|
struct task_group root_task_group;
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-10-15 15:00:07 +00:00
|
|
|
/* Default task group's sched entity on each cpu */
|
|
|
|
static DEFINE_PER_CPU(struct sched_entity, init_sched_entity);
|
|
|
|
/* Default task group's cfs_rq on each cpu */
|
|
|
|
static DEFINE_PER_CPU(struct cfs_rq, init_cfs_rq) ____cacheline_aligned_in_smp;
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
static DEFINE_PER_CPU(struct sched_rt_entity, init_sched_rt_entity);
|
|
|
|
static DEFINE_PER_CPU(struct rt_rq, init_rt_rq) ____cacheline_aligned_in_smp;
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2008-08-19 10:33:06 +00:00
|
|
|
#else /* !CONFIG_USER_SCHED */
|
2008-04-19 17:45:00 +00:00
|
|
|
#define root_task_group init_task_group
|
2008-08-19 10:33:06 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
/* task_group_lock serializes add/remove of task groups and also changes to
|
2008-01-25 20:07:59 +00:00
|
|
|
* a task group's cpu shares.
|
|
|
|
*/
|
2008-02-13 14:45:39 +00:00
|
|
|
static DEFINE_SPINLOCK(task_group_lock);
|
2008-01-25 20:07:59 +00:00
|
|
|
|
2009-03-09 12:56:21 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static int root_task_group_empty(void)
|
|
|
|
{
|
|
|
|
return list_empty(&root_task_group.children);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
# define INIT_TASK_GROUP_LOAD (2*NICE_0_LOAD)
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_USER_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
# define INIT_TASK_GROUP_LOAD NICE_0_LOAD
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
|
sched: fair-group: fix a Div0 error of the fair group scheduler
When I echoed 0 into the "cpu.shares" file, a Div0 error occured.
We found it is caused by the following calling.
sched_group_set_shares(tg, shares)
set_se_shares(tg->se[i], shares/nr_cpu_ids)
__set_se_shares(se, shares)
div64_64((1ULL<<32), shares)
When the echoed value was less than the number of processores, the result of the
sentence "shares/nr_cpu_ids" was 0, and then the system called div64() to divide
the result, the Div0 error occured.
It is unnecessary that the shares value is divided by nr_cpu_ids, I think.
Because in the function __update_group_shares_cpu() and init_tg_cfs_entry(),
the shares value isn't divided by nr_cpu_ids when setting shares of the sched
entity.
This patch fixes this bug. And echoing ULONG_MAX value into cpu.shares also
causes Div0 error, so we set a macro MAX_SHARES to limit the max value of
shares.
Signed-off-by: Miao Xie <miaox@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-04-28 04:54:56 +00:00
|
|
|
/*
|
sched: fair group: fix overflow(was: fix divide by zero)
I found a bug which can be reproduced by this way:(linux-2.6.26-rc5, x86-64)
(use 2^32, 2^33, ...., 2^63 as shares value)
# mkdir /dev/cpuctl
# mount -t cgroup -o cpu cpuctl /dev/cpuctl
# cd /dev/cpuctl
# mkdir sub
# echo 0x8000000000000000 > sub/cpu.shares
# echo $$ > sub/tasks
oops here! divide by zero.
This is because do_div() expects the 2th parameter to be 32 bits,
but unsigned long is 64 bits in x86_64.
Peter Zijstra pointed it out that the sane thing to do is limit the
shares value to something smaller instead of using an even more
expensive divide.
Also, I found another bug about "the shares value is too large":
pid1 and pid2 are set affinity to cpu#0
pid1 is attached to cg1 and pid2 is attached to cg2
if cg1/cpu.shares = 1024 cg2/cpu.shares = 2000000000
then pid2 got 100% usage of cpu, and pid1 0%
if cg1/cpu.shares = 1024 cg2/cpu.shares = 20000000000
then pid2 got 0% usage of cpu, and pid1 100%
And a weight of a cfs_rq is the sum of weights of which entities
are queued on this cfs_rq, so the shares value should be limited
to a smaller value.
I think that (1UL << 18) is a good limited value:
1) it's not too large, we can create a lot of group before overflow
2) it's several times the weight value for nice=-19 (not too small)
Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-06-12 08:42:58 +00:00
|
|
|
* A weight of 0 or 1 can cause arithmetics problems.
|
|
|
|
* A weight of a cfs_rq is the sum of weights of which entities
|
|
|
|
* are queued on this cfs_rq, so a weight of a entity should not be
|
|
|
|
* too large, so as the shares value of a task group.
|
sched: fair-group: fix a Div0 error of the fair group scheduler
When I echoed 0 into the "cpu.shares" file, a Div0 error occured.
We found it is caused by the following calling.
sched_group_set_shares(tg, shares)
set_se_shares(tg->se[i], shares/nr_cpu_ids)
__set_se_shares(se, shares)
div64_64((1ULL<<32), shares)
When the echoed value was less than the number of processores, the result of the
sentence "shares/nr_cpu_ids" was 0, and then the system called div64() to divide
the result, the Div0 error occured.
It is unnecessary that the shares value is divided by nr_cpu_ids, I think.
Because in the function __update_group_shares_cpu() and init_tg_cfs_entry(),
the shares value isn't divided by nr_cpu_ids when setting shares of the sched
entity.
This patch fixes this bug. And echoing ULONG_MAX value into cpu.shares also
causes Div0 error, so we set a macro MAX_SHARES to limit the max value of
shares.
Signed-off-by: Miao Xie <miaox@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-04-28 04:54:56 +00:00
|
|
|
* (The default weight is 1024 - so there's no practical
|
|
|
|
* limitation from this.)
|
|
|
|
*/
|
2008-04-19 17:45:00 +00:00
|
|
|
#define MIN_SHARES 2
|
sched: fair group: fix overflow(was: fix divide by zero)
I found a bug which can be reproduced by this way:(linux-2.6.26-rc5, x86-64)
(use 2^32, 2^33, ...., 2^63 as shares value)
# mkdir /dev/cpuctl
# mount -t cgroup -o cpu cpuctl /dev/cpuctl
# cd /dev/cpuctl
# mkdir sub
# echo 0x8000000000000000 > sub/cpu.shares
# echo $$ > sub/tasks
oops here! divide by zero.
This is because do_div() expects the 2th parameter to be 32 bits,
but unsigned long is 64 bits in x86_64.
Peter Zijstra pointed it out that the sane thing to do is limit the
shares value to something smaller instead of using an even more
expensive divide.
Also, I found another bug about "the shares value is too large":
pid1 and pid2 are set affinity to cpu#0
pid1 is attached to cg1 and pid2 is attached to cg2
if cg1/cpu.shares = 1024 cg2/cpu.shares = 2000000000
then pid2 got 100% usage of cpu, and pid1 0%
if cg1/cpu.shares = 1024 cg2/cpu.shares = 20000000000
then pid2 got 0% usage of cpu, and pid1 100%
And a weight of a cfs_rq is the sum of weights of which entities
are queued on this cfs_rq, so the shares value should be limited
to a smaller value.
I think that (1UL << 18) is a good limited value:
1) it's not too large, we can create a lot of group before overflow
2) it's several times the weight value for nice=-19 (not too small)
Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-06-12 08:42:58 +00:00
|
|
|
#define MAX_SHARES (1UL << 18)
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
static int init_task_group_load = INIT_TASK_GROUP_LOAD;
|
|
|
|
#endif
|
|
|
|
|
2007-10-15 15:00:07 +00:00
|
|
|
/* Default task group.
|
2007-10-15 15:00:12 +00:00
|
|
|
* Every task in system belong to this group at bootup.
|
2007-10-15 15:00:07 +00:00
|
|
|
*/
|
2008-04-05 01:11:04 +00:00
|
|
|
struct task_group init_task_group;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
|
|
|
/* return group to which a task belongs */
|
2007-10-15 15:00:14 +00:00
|
|
|
static inline struct task_group *task_group(struct task_struct *p)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
struct task_group *tg;
|
2007-10-15 15:00:09 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_USER_SCHED
|
2008-11-13 23:39:19 +00:00
|
|
|
rcu_read_lock();
|
|
|
|
tg = __task_cred(p)->user->tg;
|
|
|
|
rcu_read_unlock();
|
2008-02-13 14:45:40 +00:00
|
|
|
#elif defined(CONFIG_CGROUP_SCHED)
|
2007-10-19 06:41:03 +00:00
|
|
|
tg = container_of(task_subsys_state(p, cpu_cgroup_subsys_id),
|
|
|
|
struct task_group, css);
|
2007-10-15 15:00:09 +00:00
|
|
|
#else
|
2007-12-05 14:46:09 +00:00
|
|
|
tg = &init_task_group;
|
2007-10-15 15:00:09 +00:00
|
|
|
#endif
|
2007-10-15 15:00:09 +00:00
|
|
|
return tg;
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */
|
2008-01-25 20:08:30 +00:00
|
|
|
static inline void set_task_rq(struct task_struct *p, unsigned int cpu)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-11-15 19:57:40 +00:00
|
|
|
p->se.cfs_rq = task_group(p)->cfs_rq[cpu];
|
|
|
|
p->se.parent = task_group(p)->se[cpu];
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
p->rt.rt_rq = task_group(p)->rt_rq[cpu];
|
|
|
|
p->rt.parent = task_group(p)->rt_se[cpu];
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2009-03-09 12:56:21 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static int root_task_group_empty(void)
|
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
static inline void set_task_rq(struct task_struct *p, unsigned int cpu) { }
|
sched: correct wakeup weight calculations
rw_i = {2, 4, 1, 0}
s_i = {2/7, 4/7, 1/7, 0}
wakeup on cpu0, weight=1
rw'_i = {3, 4, 1, 0}
s'_i = {3/8, 4/8, 1/8, 0}
s_0 = S * rw_0 / \Sum rw_j ->
\Sum rw_j = S*rw_0/s_0 = 1*2*7/2 = 7 (correct)
s'_0 = S * (rw_0 + 1) / (\Sum rw_j + 1) =
1 * (2+1) / (7+1) = 3/8 (correct
so we find that adding 1 to cpu0 gains 5/56 in weight
if say the other cpu were, cpu1, we'd also have to calculate its 4/56 loss
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Cc: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-06-27 11:41:37 +00:00
|
|
|
static inline struct task_group *task_group(struct task_struct *p)
|
|
|
|
{
|
|
|
|
return NULL;
|
|
|
|
}
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif /* CONFIG_GROUP_SCHED */
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
/* CFS-related fields in a runqueue */
|
|
|
|
struct cfs_rq {
|
|
|
|
struct load_weight load;
|
|
|
|
unsigned long nr_running;
|
|
|
|
|
|
|
|
u64 exec_clock;
|
2007-10-15 15:00:04 +00:00
|
|
|
u64 min_vruntime;
|
2007-07-09 16:51:58 +00:00
|
|
|
|
|
|
|
struct rb_root tasks_timeline;
|
|
|
|
struct rb_node *rb_leftmost;
|
2008-04-19 17:45:00 +00:00
|
|
|
|
|
|
|
struct list_head tasks;
|
|
|
|
struct list_head *balance_iterator;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* 'curr' points to currently running entity on this cfs_rq.
|
2007-07-09 16:51:58 +00:00
|
|
|
* It is set to NULL otherwise (i.e when none are currently running).
|
|
|
|
*/
|
sched: backward looking buddy
Impact: improve/change/fix wakeup-buddy scheduling
Currently we only have a forward looking buddy, that is, we prefer to
schedule to the task we last woke up, under the presumption that its
going to consume the data we just produced, and therefore will have
cache hot benefits.
This allows co-waking producer/consumer task pairs to run ahead of the
pack for a little while, keeping their cache warm. Without this, we
would interleave all pairs, utterly trashing the cache.
This patch introduces a backward looking buddy, that is, suppose that
in the above scenario, the consumer preempts the producer before it
can go to sleep, we will therefore miss the wakeup from consumer to
producer (its already running, after all), breaking the cycle and
reverting to the cache-trashing interleaved schedule pattern.
The backward buddy will try to schedule back to the task that woke us
up in case the forward buddy is not available, under the assumption
that the last task will be the one with the most cache hot task around
barring current.
This will basically allow a task to continue after it got preempted.
In order to avoid starvation, we allow either buddy to get wakeup_gran
ahead of the pack.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Acked-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 20:25:09 +00:00
|
|
|
struct sched_entity *curr, *next, *last;
|
2007-10-15 15:00:10 +00:00
|
|
|
|
2008-11-10 09:46:32 +00:00
|
|
|
unsigned int nr_spread_over;
|
2007-10-15 15:00:10 +00:00
|
|
|
|
2007-10-15 15:00:03 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-07-09 16:51:58 +00:00
|
|
|
struct rq *rq; /* cpu runqueue to which this cfs_rq is attached */
|
|
|
|
|
2007-12-05 14:46:09 +00:00
|
|
|
/*
|
|
|
|
* leaf cfs_rqs are those that hold tasks (lowest schedulable entity in
|
2007-07-09 16:51:58 +00:00
|
|
|
* a hierarchy). Non-leaf lrqs hold other higher schedulable entities
|
|
|
|
* (like users, containers etc.)
|
|
|
|
*
|
|
|
|
* leaf_cfs_rq_list ties together list of leaf cfs_rq's in a cpu. This
|
|
|
|
* list is used during load balance.
|
|
|
|
*/
|
2007-12-05 14:46:09 +00:00
|
|
|
struct list_head leaf_cfs_rq_list;
|
|
|
|
struct task_group *tg; /* group that "owns" this runqueue */
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/*
|
2008-06-27 11:41:23 +00:00
|
|
|
* the part of load.weight contributed by tasks
|
2008-06-27 11:41:14 +00:00
|
|
|
*/
|
2008-06-27 11:41:23 +00:00
|
|
|
unsigned long task_weight;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
/*
|
|
|
|
* h_load = weight * f(tg)
|
|
|
|
*
|
|
|
|
* Where f(tg) is the recursive weight fraction assigned to
|
|
|
|
* this group.
|
|
|
|
*/
|
|
|
|
unsigned long h_load;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
/*
|
|
|
|
* this cpu's part of tg->shares
|
|
|
|
*/
|
|
|
|
unsigned long shares;
|
2008-06-27 11:41:38 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* load.weight at the time we set shares
|
|
|
|
*/
|
|
|
|
unsigned long rq_weight;
|
2008-06-27 11:41:14 +00:00
|
|
|
#endif
|
2007-07-09 16:51:58 +00:00
|
|
|
#endif
|
|
|
|
};
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
/* Real-Time classes' related field in a runqueue: */
|
|
|
|
struct rt_rq {
|
|
|
|
struct rt_prio_array active;
|
2008-01-25 20:08:03 +00:00
|
|
|
unsigned long rt_nr_running;
|
2008-02-13 14:45:40 +00:00
|
|
|
#if defined CONFIG_SMP || defined CONFIG_RT_GROUP_SCHED
|
2008-12-29 14:39:49 +00:00
|
|
|
struct {
|
|
|
|
int curr; /* highest queued rt task prio */
|
2009-01-14 14:10:04 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-12-29 14:39:49 +00:00
|
|
|
int next; /* next highest */
|
2009-01-14 14:10:04 +00:00
|
|
|
#endif
|
2008-12-29 14:39:49 +00:00
|
|
|
} highest_prio;
|
2008-01-25 20:08:30 +00:00
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-01-25 20:08:07 +00:00
|
|
|
unsigned long rt_nr_migratory;
|
2009-04-01 16:40:15 +00:00
|
|
|
unsigned long rt_nr_total;
|
2008-01-25 20:08:12 +00:00
|
|
|
int overloaded;
|
sched: create "pushable_tasks" list to limit pushing to one attempt
The RT scheduler employs a "push/pull" design to actively balance tasks
within the system (on a per disjoint cpuset basis). When a task is
awoken, it is immediately determined if there are any lower priority
cpus which should be preempted. This is opposed to the way normal
SCHED_OTHER tasks behave, which will wait for a periodic rebalancing
operation to occur before spreading out load.
When a particular RQ has more than 1 active RT task, it is said to
be in an "overloaded" state. Once this occurs, the system enters
the active balancing mode, where it will try to push the task away,
or persuade a different cpu to pull it over. The system will stay
in this state until the system falls back below the <= 1 queued RT
task per RQ.
However, the current implementation suffers from a limitation in the
push logic. Once overloaded, all tasks (other than current) on the
RQ are analyzed on every push operation, even if it was previously
unpushable (due to affinity, etc). Whats more, the operation stops
at the first task that is unpushable and will not look at items
lower in the queue. This causes two problems:
1) We can have the same tasks analyzed over and over again during each
push, which extends out the fast path in the scheduler for no
gain. Consider a RQ that has dozens of tasks that are bound to a
core. Each one of those tasks will be encountered and skipped
for each push operation while they are queued.
2) There may be lower-priority tasks under the unpushable task that
could have been successfully pushed, but will never be considered
until either the unpushable task is cleared, or a pull operation
succeeds. The net result is a potential latency source for mid
priority tasks.
This patch aims to rectify these two conditions by introducing a new
priority sorted list: "pushable_tasks". A task is added to the list
each time a task is activated or preempted. It is removed from the
list any time it is deactivated, made current, or fails to push.
This works because a task only needs to be attempted to push once.
After an initial failure to push, the other cpus will eventually try to
pull the task when the conditions are proper. This also solves the
problem that we don't completely analyze all tasks due to encountering
an unpushable tasks. Now every task will have a push attempted (when
appropriate).
This reduces latency both by shorting the critical section of the
rq->lock for certain workloads, and by making sure the algorithm
considers all eligible tasks in the system.
[ rostedt: added a couple more BUG_ONs ]
Signed-off-by: Gregory Haskins <ghaskins@novell.com>
Acked-by: Steven Rostedt <srostedt@redhat.com>
2008-12-29 14:39:53 +00:00
|
|
|
struct plist_head pushable_tasks;
|
2008-01-25 20:08:29 +00:00
|
|
|
#endif
|
2008-01-25 20:08:30 +00:00
|
|
|
int rt_throttled;
|
2008-01-25 20:08:29 +00:00
|
|
|
u64 rt_time;
|
2008-04-19 17:44:58 +00:00
|
|
|
u64 rt_runtime;
|
2008-03-25 12:51:45 +00:00
|
|
|
/* Nests inside the rq lock: */
|
2008-04-19 17:44:58 +00:00
|
|
|
spinlock_t rt_runtime_lock;
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-02-13 14:45:39 +00:00
|
|
|
unsigned long rt_nr_boosted;
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
struct rq *rq;
|
|
|
|
struct list_head leaf_rt_rq_list;
|
|
|
|
struct task_group *tg;
|
|
|
|
struct sched_rt_entity *rt_se;
|
|
|
|
#endif
|
2007-07-09 16:51:58 +00:00
|
|
|
};
|
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We add the notion of a root-domain which will be used to define per-domain
|
2008-01-25 20:08:19 +00:00
|
|
|
* variables. Each exclusive cpuset essentially defines an island domain by
|
|
|
|
* fully partitioning the member cpus from any other cpuset. Whenever a new
|
2008-01-25 20:08:18 +00:00
|
|
|
* exclusive cpuset is created, we also create and attach a new root-domain
|
|
|
|
* object.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
struct root_domain {
|
|
|
|
atomic_t refcount;
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_var_t span;
|
|
|
|
cpumask_var_t online;
|
2008-01-25 20:08:18 +00:00
|
|
|
|
2008-01-25 20:08:19 +00:00
|
|
|
/*
|
2008-01-25 20:08:18 +00:00
|
|
|
* The "RT overload" flag: it gets set if a CPU has more than
|
|
|
|
* one runnable RT task.
|
|
|
|
*/
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_var_t rto_mask;
|
2008-01-25 20:08:19 +00:00
|
|
|
atomic_t rto_count;
|
2008-05-12 19:21:01 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
struct cpupri cpupri;
|
|
|
|
#endif
|
2008-12-18 17:56:22 +00:00
|
|
|
#if defined(CONFIG_SCHED_MC) || defined(CONFIG_SCHED_SMT)
|
|
|
|
/*
|
|
|
|
* Preferred wake up cpu nominated by sched_mc balance that will be
|
|
|
|
* used when most cpus are idle in the system indicating overall very
|
|
|
|
* low system utilisation. Triggered at POWERSAVINGS_BALANCE_WAKEUP(2)
|
|
|
|
*/
|
|
|
|
unsigned int sched_mc_preferred_wakeup_cpu;
|
|
|
|
#endif
|
2008-01-25 20:08:18 +00:00
|
|
|
};
|
|
|
|
|
2008-01-25 20:08:26 +00:00
|
|
|
/*
|
|
|
|
* By default the system creates a single root-domain with all cpus as
|
|
|
|
* members (mimicking the global state we have today).
|
|
|
|
*/
|
2008-01-25 20:08:18 +00:00
|
|
|
static struct root_domain def_root_domain;
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* This is the main, per-CPU runqueue data structure.
|
|
|
|
*
|
|
|
|
* Locking rule: those places that want to lock multiple runqueues
|
|
|
|
* (such as the load balancing or the thread migration code), lock
|
|
|
|
* acquire operations must be ordered by ascending &runqueue.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq {
|
2007-10-18 19:32:55 +00:00
|
|
|
/* runqueue lock: */
|
|
|
|
spinlock_t lock;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* nr_running and cpu_load should be in the same cacheline because
|
|
|
|
* remote CPUs use both these fields when doing load calculation.
|
|
|
|
*/
|
|
|
|
unsigned long nr_running;
|
2007-07-09 16:51:58 +00:00
|
|
|
#define CPU_LOAD_IDX_MAX 5
|
|
|
|
unsigned long cpu_load[CPU_LOAD_IDX_MAX];
|
2007-05-08 07:32:51 +00:00
|
|
|
#ifdef CONFIG_NO_HZ
|
2008-04-19 17:44:57 +00:00
|
|
|
unsigned long last_tick_seen;
|
2007-05-08 07:32:51 +00:00
|
|
|
unsigned char in_nohz_recently;
|
|
|
|
#endif
|
2007-10-18 19:32:55 +00:00
|
|
|
/* capture load from *all* tasks on this cpu: */
|
|
|
|
struct load_weight load;
|
2007-07-09 16:51:58 +00:00
|
|
|
unsigned long nr_load_updates;
|
|
|
|
u64 nr_switches;
|
perf_counters: make software counters work as per-cpu counters
Impact: kernel crash fix
Yanmin Zhang reported that using a PERF_COUNT_TASK_CLOCK software
counter as a per-cpu counter would reliably crash the system, because
it calls __task_delta_exec with a null pointer. The page fault,
context switch and cpu migration counters also won't function
correctly as per-cpu counters since they reference the current task.
This fixes the problem by redirecting the task_clock counter to the
cpu_clock counter when used as a per-cpu counter, and by implementing
per-cpu page fault, context switch and cpu migration counters.
Along the way, this:
- Initializes counter->ctx earlier, in perf_counter_alloc, so that
sw_perf_counter_init can use it
- Adds code to kernel/sched.c to count task migrations into each
cpu, in rq->nr_migrations_in
- Exports the per-cpu context switch and task migration counts
via new functions added to kernel/sched.c
- Makes sure that if sw_perf_counter_init fails, we don't try to
initialize the counter as a hardware counter. Since the user has
passed a negative, non-raw event type, they clearly don't intend
for it to be interpreted as a hardware event.
Reported-by: "Zhang Yanmin" <yanmin_zhang@linux.intel.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-09 11:42:47 +00:00
|
|
|
u64 nr_migrations_in;
|
2007-07-09 16:51:58 +00:00
|
|
|
|
|
|
|
struct cfs_rq cfs;
|
2008-01-25 20:08:30 +00:00
|
|
|
struct rt_rq rt;
|
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-10-18 19:32:55 +00:00
|
|
|
/* list of leaf cfs_rq on this cpu: */
|
|
|
|
struct list_head leaf_cfs_rq_list;
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
struct list_head leaf_rt_rq_list;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This is part of a global counter where only the total sum
|
|
|
|
* over all CPUs matters. A task can increase this counter on
|
|
|
|
* one CPU and if it got migrated afterwards it may decrease
|
|
|
|
* it on another CPU. Always updated under the runqueue lock:
|
|
|
|
*/
|
|
|
|
unsigned long nr_uninterruptible;
|
|
|
|
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *curr, *idle;
|
2006-12-10 10:20:25 +00:00
|
|
|
unsigned long next_balance;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct mm_struct *prev_mm;
|
2007-07-09 16:51:58 +00:00
|
|
|
|
2008-05-03 16:29:28 +00:00
|
|
|
u64 clock;
|
2007-07-09 16:51:58 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_t nr_iowait;
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
2008-01-25 20:08:19 +00:00
|
|
|
struct root_domain *rd;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sched_domain *sd;
|
|
|
|
|
2009-02-13 19:35:45 +00:00
|
|
|
unsigned char idle_at_tick;
|
2005-04-16 22:20:36 +00:00
|
|
|
/* For active balancing */
|
|
|
|
int active_balance;
|
|
|
|
int push_cpu;
|
2007-10-18 19:32:55 +00:00
|
|
|
/* cpu of this runqueue: */
|
|
|
|
int cpu;
|
2008-06-04 19:04:05 +00:00
|
|
|
int online;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-06-27 11:41:26 +00:00
|
|
|
unsigned long avg_load_per_task;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *migration_thread;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct list_head migration_queue;
|
|
|
|
#endif
|
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
/* calc_load related fields */
|
|
|
|
unsigned long calc_load_update;
|
|
|
|
long calc_load_active;
|
|
|
|
|
2008-01-25 20:08:29 +00:00
|
|
|
#ifdef CONFIG_SCHED_HRTICK
|
2008-07-18 16:01:23 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
int hrtick_csd_pending;
|
|
|
|
struct call_single_data hrtick_csd;
|
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
struct hrtimer hrtick_timer;
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
|
|
|
/* latency stats */
|
|
|
|
struct sched_info rq_sched_info;
|
2008-12-17 07:41:22 +00:00
|
|
|
unsigned long long rq_cpu_time;
|
|
|
|
/* could above be rq->cfs_rq.exec_clock + rq->rt_rq.rt_runtime ? */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* sys_sched_yield() stats */
|
2007-10-18 19:32:56 +00:00
|
|
|
unsigned int yld_count;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* schedule() stats */
|
2007-10-18 19:32:56 +00:00
|
|
|
unsigned int sched_switch;
|
|
|
|
unsigned int sched_count;
|
|
|
|
unsigned int sched_goidle;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* try_to_wake_up() stats */
|
2007-10-18 19:32:56 +00:00
|
|
|
unsigned int ttwu_count;
|
|
|
|
unsigned int ttwu_local;
|
2007-10-15 15:00:10 +00:00
|
|
|
|
|
|
|
/* BKL stats */
|
2007-10-18 19:32:56 +00:00
|
|
|
unsigned int bkl_count;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
};
|
|
|
|
|
2007-07-19 08:48:13 +00:00
|
|
|
static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-09-20 21:38:02 +00:00
|
|
|
static inline void check_preempt_curr(struct rq *rq, struct task_struct *p, int sync)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
2008-09-20 21:38:02 +00:00
|
|
|
rq->curr->sched_class->check_preempt_curr(rq, p, sync);
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2006-09-26 06:30:51 +00:00
|
|
|
static inline int cpu_of(struct rq *rq)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
return rq->cpu;
|
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:27 +00:00
|
|
|
/*
|
|
|
|
* The domain tree (rq->sd) is protected by RCU's quiescent state transition.
|
2005-06-25 21:57:33 +00:00
|
|
|
* See detach_destroy_domains: synchronize_sched for details.
|
2005-06-25 21:57:27 +00:00
|
|
|
*
|
|
|
|
* The domain tree of any CPU may only be accessed from within
|
|
|
|
* preempt-disabled sections.
|
|
|
|
*/
|
2006-07-03 07:25:40 +00:00
|
|
|
#define for_each_domain(cpu, __sd) \
|
|
|
|
for (__sd = rcu_dereference(cpu_rq(cpu)->sd); __sd; __sd = __sd->parent)
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
#define cpu_rq(cpu) (&per_cpu(runqueues, (cpu)))
|
|
|
|
#define this_rq() (&__get_cpu_var(runqueues))
|
|
|
|
#define task_rq(p) cpu_rq(task_cpu(p))
|
|
|
|
#define cpu_curr(cpu) (cpu_rq(cpu)->curr)
|
|
|
|
|
2008-12-17 13:10:57 +00:00
|
|
|
inline void update_rq_clock(struct rq *rq)
|
2008-05-03 16:29:28 +00:00
|
|
|
{
|
|
|
|
rq->clock = sched_clock_cpu(cpu_of(rq));
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
/*
|
|
|
|
* Tunables that become constants when CONFIG_SCHED_DEBUG is off:
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
|
# define const_debug __read_mostly
|
|
|
|
#else
|
|
|
|
# define const_debug static const
|
|
|
|
#endif
|
|
|
|
|
2008-05-12 19:20:52 +00:00
|
|
|
/**
|
|
|
|
* runqueue_is_locked
|
|
|
|
*
|
|
|
|
* Returns true if the current cpu runqueue is locked.
|
|
|
|
* This interface allows printk to be called with the runqueue lock
|
|
|
|
* held and know whether or not it is OK to wake up the klogd.
|
|
|
|
*/
|
|
|
|
int runqueue_is_locked(void)
|
|
|
|
{
|
|
|
|
int cpu = get_cpu();
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = spin_is_locked(&rq->lock);
|
|
|
|
put_cpu();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
/*
|
|
|
|
* Debugging: various feature bits
|
|
|
|
*/
|
2008-04-19 17:45:00 +00:00
|
|
|
|
|
|
|
#define SCHED_FEAT(name, enabled) \
|
|
|
|
__SCHED_FEAT_##name ,
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
enum {
|
2008-04-19 17:45:00 +00:00
|
|
|
#include "sched_features.h"
|
2007-10-15 15:00:04 +00:00
|
|
|
};
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
#undef SCHED_FEAT
|
|
|
|
|
|
|
|
#define SCHED_FEAT(name, enabled) \
|
|
|
|
(1UL << __SCHED_FEAT_##name) * enabled |
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
const_debug unsigned int sysctl_sched_features =
|
2008-04-19 17:45:00 +00:00
|
|
|
#include "sched_features.h"
|
|
|
|
0;
|
|
|
|
|
|
|
|
#undef SCHED_FEAT
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
|
#define SCHED_FEAT(name, enabled) \
|
|
|
|
#name ,
|
|
|
|
|
2008-04-25 01:17:55 +00:00
|
|
|
static __read_mostly char *sched_feat_names[] = {
|
2008-04-19 17:45:00 +00:00
|
|
|
#include "sched_features.h"
|
|
|
|
NULL
|
|
|
|
};
|
|
|
|
|
|
|
|
#undef SCHED_FEAT
|
|
|
|
|
2008-10-30 07:23:32 +00:00
|
|
|
static int sched_feat_show(struct seq_file *m, void *v)
|
2008-04-19 17:45:00 +00:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; sched_feat_names[i]; i++) {
|
2008-10-30 07:23:32 +00:00
|
|
|
if (!(sysctl_sched_features & (1UL << i)))
|
|
|
|
seq_puts(m, "NO_");
|
|
|
|
seq_printf(m, "%s ", sched_feat_names[i]);
|
2008-04-19 17:45:00 +00:00
|
|
|
}
|
2008-10-30 07:23:32 +00:00
|
|
|
seq_puts(m, "\n");
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-10-30 07:23:32 +00:00
|
|
|
return 0;
|
2008-04-19 17:45:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t
|
|
|
|
sched_feat_write(struct file *filp, const char __user *ubuf,
|
|
|
|
size_t cnt, loff_t *ppos)
|
|
|
|
{
|
|
|
|
char buf[64];
|
|
|
|
char *cmp = buf;
|
|
|
|
int neg = 0;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (cnt > 63)
|
|
|
|
cnt = 63;
|
|
|
|
|
|
|
|
if (copy_from_user(&buf, ubuf, cnt))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
buf[cnt] = 0;
|
|
|
|
|
2008-04-18 08:55:34 +00:00
|
|
|
if (strncmp(buf, "NO_", 3) == 0) {
|
2008-04-19 17:45:00 +00:00
|
|
|
neg = 1;
|
|
|
|
cmp += 3;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; sched_feat_names[i]; i++) {
|
|
|
|
int len = strlen(sched_feat_names[i]);
|
|
|
|
|
|
|
|
if (strncmp(cmp, sched_feat_names[i], len) == 0) {
|
|
|
|
if (neg)
|
|
|
|
sysctl_sched_features &= ~(1UL << i);
|
|
|
|
else
|
|
|
|
sysctl_sched_features |= (1UL << i);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!sched_feat_names[i])
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
filp->f_pos += cnt;
|
|
|
|
|
|
|
|
return cnt;
|
|
|
|
}
|
|
|
|
|
2008-10-30 07:23:32 +00:00
|
|
|
static int sched_feat_open(struct inode *inode, struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, sched_feat_show, NULL);
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
static struct file_operations sched_feat_fops = {
|
2008-10-30 07:23:32 +00:00
|
|
|
.open = sched_feat_open,
|
|
|
|
.write = sched_feat_write,
|
|
|
|
.read = seq_read,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
2008-04-19 17:45:00 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static __init int sched_init_debug(void)
|
|
|
|
{
|
|
|
|
debugfs_create_file("sched_features", 0644, NULL, NULL,
|
|
|
|
&sched_feat_fops);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
late_initcall(sched_init_debug);
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define sched_feat(x) (sysctl_sched_features & (1UL << __SCHED_FEAT_##x))
|
2007-10-15 15:00:04 +00:00
|
|
|
|
2007-11-09 21:39:39 +00:00
|
|
|
/*
|
|
|
|
* Number of tasks to iterate in a single balance run.
|
|
|
|
* Limited because this is done with IRQs disabled.
|
|
|
|
*/
|
|
|
|
const_debug unsigned int sysctl_sched_nr_migrate = 32;
|
|
|
|
|
2008-06-27 11:41:35 +00:00
|
|
|
/*
|
|
|
|
* ratelimit for updating the group shares.
|
2008-08-04 06:54:26 +00:00
|
|
|
* default: 0.25ms
|
2008-06-27 11:41:35 +00:00
|
|
|
*/
|
2008-08-04 06:54:26 +00:00
|
|
|
unsigned int sysctl_sched_shares_ratelimit = 250000;
|
2008-06-27 11:41:35 +00:00
|
|
|
|
2008-10-17 17:27:02 +00:00
|
|
|
/*
|
|
|
|
* Inject some fuzzyness into changing the per-cpu group shares
|
|
|
|
* this avoids remote rq-locks at the expense of fairness.
|
|
|
|
* default: 4
|
|
|
|
*/
|
|
|
|
unsigned int sysctl_sched_shares_thresh = 4;
|
|
|
|
|
2008-01-25 20:08:29 +00:00
|
|
|
/*
|
2008-02-13 14:45:39 +00:00
|
|
|
* period over which we measure -rt task cpu usage in us.
|
2008-01-25 20:08:29 +00:00
|
|
|
* default: 1s
|
|
|
|
*/
|
2008-02-13 14:45:39 +00:00
|
|
|
unsigned int sysctl_sched_rt_period = 1000000;
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-02-13 13:02:36 +00:00
|
|
|
static __read_mostly int scheduler_running;
|
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
/*
|
|
|
|
* part of the period that we allow rt tasks to run in us.
|
|
|
|
* default: 0.95s
|
|
|
|
*/
|
|
|
|
int sysctl_sched_rt_runtime = 950000;
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
static inline u64 global_rt_period(void)
|
|
|
|
{
|
|
|
|
return (u64)sysctl_sched_rt_period * NSEC_PER_USEC;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 global_rt_runtime(void)
|
|
|
|
{
|
2008-07-22 20:51:15 +00:00
|
|
|
if (sysctl_sched_rt_runtime < 0)
|
2008-04-19 17:44:57 +00:00
|
|
|
return RUNTIME_INF;
|
|
|
|
|
|
|
|
return (u64)sysctl_sched_rt_runtime * NSEC_PER_USEC;
|
|
|
|
}
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifndef prepare_arch_switch
|
2005-06-25 21:57:23 +00:00
|
|
|
# define prepare_arch_switch(next) do { } while (0)
|
|
|
|
#endif
|
|
|
|
#ifndef finish_arch_switch
|
|
|
|
# define finish_arch_switch(prev) do { } while (0)
|
|
|
|
#endif
|
|
|
|
|
2007-12-18 14:21:13 +00:00
|
|
|
static inline int task_current(struct rq *rq, struct task_struct *p)
|
|
|
|
{
|
|
|
|
return rq->curr == p;
|
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:23 +00:00
|
|
|
#ifndef __ARCH_WANT_UNLOCKED_CTXSW
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline int task_running(struct rq *rq, struct task_struct *p)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
2007-12-18 14:21:13 +00:00
|
|
|
return task_current(rq, p);
|
2005-06-25 21:57:23 +00:00
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void prepare_lock_switch(struct rq *rq, struct task_struct *next)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void finish_lock_switch(struct rq *rq, struct task_struct *prev)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
2005-09-13 09:17:59 +00:00
|
|
|
#ifdef CONFIG_DEBUG_SPINLOCK
|
|
|
|
/* this is a valid case when another task releases the spinlock */
|
|
|
|
rq->lock.owner = current;
|
|
|
|
#endif
|
2006-07-03 07:24:54 +00:00
|
|
|
/*
|
|
|
|
* If we are tracking spinlock dependencies then we have to
|
|
|
|
* fix up the runqueue lock - which gets 'carried over' from
|
|
|
|
* prev into current:
|
|
|
|
*/
|
|
|
|
spin_acquire(&rq->lock.dep_map, 0, 0, _THIS_IP_);
|
|
|
|
|
2005-06-25 21:57:23 +00:00
|
|
|
spin_unlock_irq(&rq->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* __ARCH_WANT_UNLOCKED_CTXSW */
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline int task_running(struct rq *rq, struct task_struct *p)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
return p->oncpu;
|
|
|
|
#else
|
2007-12-18 14:21:13 +00:00
|
|
|
return task_current(rq, p);
|
2005-06-25 21:57:23 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void prepare_lock_switch(struct rq *rq, struct task_struct *next)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/*
|
|
|
|
* We can optimise this out completely for !SMP, because the
|
|
|
|
* SMP rebalancing from interrupt is the only thing that cares
|
|
|
|
* here.
|
|
|
|
*/
|
|
|
|
next->oncpu = 1;
|
|
|
|
#endif
|
|
|
|
#ifdef __ARCH_WANT_INTERRUPTS_ON_CTXSW
|
|
|
|
spin_unlock_irq(&rq->lock);
|
|
|
|
#else
|
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void finish_lock_switch(struct rq *rq, struct task_struct *prev)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/*
|
|
|
|
* After ->oncpu is cleared, the task can be moved to a different CPU.
|
|
|
|
* We must ensure this doesn't happen until the switch is completely
|
|
|
|
* finished.
|
|
|
|
*/
|
|
|
|
smp_wmb();
|
|
|
|
prev->oncpu = 0;
|
|
|
|
#endif
|
|
|
|
#ifndef __ARCH_WANT_INTERRUPTS_ON_CTXSW
|
|
|
|
local_irq_enable();
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2005-06-25 21:57:23 +00:00
|
|
|
}
|
|
|
|
#endif /* __ARCH_WANT_UNLOCKED_CTXSW */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
/*
|
|
|
|
* __task_rq_lock - lock the runqueue a given task resides on.
|
|
|
|
* Must be called interrupts disabled.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline struct rq *__task_rq_lock(struct task_struct *p)
|
2006-06-27 09:54:51 +00:00
|
|
|
__acquires(rq->lock)
|
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
for (;;) {
|
|
|
|
struct rq *rq = task_rq(p);
|
|
|
|
spin_lock(&rq->lock);
|
|
|
|
if (likely(rq == task_rq(p)))
|
|
|
|
return rq;
|
2006-06-27 09:54:51 +00:00
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* task_rq_lock - lock the runqueue a given task resides on and disable
|
2007-12-05 14:46:09 +00:00
|
|
|
* interrupts. Note the ordering: we can safely lookup the task_rq without
|
2005-04-16 22:20:36 +00:00
|
|
|
* explicitly disabling preemption.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static struct rq *task_rq_lock(struct task_struct *p, unsigned long *flags)
|
2005-04-16 22:20:36 +00:00
|
|
|
__acquires(rq->lock)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
for (;;) {
|
|
|
|
local_irq_save(*flags);
|
|
|
|
rq = task_rq(p);
|
|
|
|
spin_lock(&rq->lock);
|
|
|
|
if (likely(rq == task_rq(p)))
|
|
|
|
return rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irqrestore(&rq->lock, *flags);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-11-10 14:39:30 +00:00
|
|
|
void task_rq_unlock_wait(struct task_struct *p)
|
|
|
|
{
|
|
|
|
struct rq *rq = task_rq(p);
|
|
|
|
|
|
|
|
smp_mb(); /* spin-unlock-wait is not a full memory barrier */
|
|
|
|
spin_unlock_wait(&rq->lock);
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:13 +00:00
|
|
|
static void __task_rq_unlock(struct rq *rq)
|
2006-06-27 09:54:51 +00:00
|
|
|
__releases(rq->lock)
|
|
|
|
{
|
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void task_rq_unlock(struct rq *rq, unsigned long *flags)
|
2005-04-16 22:20:36 +00:00
|
|
|
__releases(rq->lock)
|
|
|
|
{
|
|
|
|
spin_unlock_irqrestore(&rq->lock, *flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2006-12-10 10:20:00 +00:00
|
|
|
* this_rq_lock - lock this runqueue and disable interrupts.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static struct rq *this_rq_lock(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
__acquires(rq->lock)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
local_irq_disable();
|
|
|
|
rq = this_rq();
|
|
|
|
spin_lock(&rq->lock);
|
|
|
|
|
|
|
|
return rq;
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:29 +00:00
|
|
|
#ifdef CONFIG_SCHED_HRTICK
|
|
|
|
/*
|
|
|
|
* Use HR-timers to deliver accurate preemption points.
|
|
|
|
*
|
|
|
|
* Its all a bit involved since we cannot program an hrt while holding the
|
|
|
|
* rq->lock. So what we do is store a state in in rq->hrtick_* and ask for a
|
|
|
|
* reschedule event.
|
|
|
|
*
|
|
|
|
* When we get rescheduled we reprogram the hrtick_timer outside of the
|
|
|
|
* rq->lock.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Use hrtick when:
|
|
|
|
* - enabled by features
|
|
|
|
* - hrtimer is actually high res
|
|
|
|
*/
|
|
|
|
static inline int hrtick_enabled(struct rq *rq)
|
|
|
|
{
|
|
|
|
if (!sched_feat(HRTICK))
|
|
|
|
return 0;
|
2008-07-20 09:02:06 +00:00
|
|
|
if (!cpu_active(cpu_of(rq)))
|
2008-04-29 08:02:46 +00:00
|
|
|
return 0;
|
2008-01-25 20:08:29 +00:00
|
|
|
return hrtimer_is_hres_active(&rq->hrtick_timer);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hrtick_clear(struct rq *rq)
|
|
|
|
{
|
|
|
|
if (hrtimer_active(&rq->hrtick_timer))
|
|
|
|
hrtimer_cancel(&rq->hrtick_timer);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* High-resolution timer tick.
|
|
|
|
* Runs from hardirq context with interrupts disabled.
|
|
|
|
*/
|
|
|
|
static enum hrtimer_restart hrtick(struct hrtimer *timer)
|
|
|
|
{
|
|
|
|
struct rq *rq = container_of(timer, struct rq, hrtick_timer);
|
|
|
|
|
|
|
|
WARN_ON_ONCE(cpu_of(rq) != smp_processor_id());
|
|
|
|
|
|
|
|
spin_lock(&rq->lock);
|
2008-05-03 16:29:28 +00:00
|
|
|
update_rq_clock(rq);
|
2008-01-25 20:08:29 +00:00
|
|
|
rq->curr->sched_class->task_tick(rq, rq->curr, 1);
|
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
|
|
|
|
return HRTIMER_NORESTART;
|
|
|
|
}
|
|
|
|
|
2008-05-11 00:25:33 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-07-18 16:01:23 +00:00
|
|
|
/*
|
|
|
|
* called from hardirq (IPI) context
|
|
|
|
*/
|
|
|
|
static void __hrtick_start(void *arg)
|
2008-04-29 08:02:46 +00:00
|
|
|
{
|
2008-07-18 16:01:23 +00:00
|
|
|
struct rq *rq = arg;
|
2008-04-29 08:02:46 +00:00
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
spin_lock(&rq->lock);
|
|
|
|
hrtimer_restart(&rq->hrtick_timer);
|
|
|
|
rq->hrtick_csd_pending = 0;
|
|
|
|
spin_unlock(&rq->lock);
|
2008-04-29 08:02:46 +00:00
|
|
|
}
|
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
/*
|
|
|
|
* Called to set the hrtick timer state.
|
|
|
|
*
|
|
|
|
* called with rq->lock held and irqs disabled
|
|
|
|
*/
|
|
|
|
static void hrtick_start(struct rq *rq, u64 delay)
|
2008-04-29 08:02:46 +00:00
|
|
|
{
|
2008-07-18 16:01:23 +00:00
|
|
|
struct hrtimer *timer = &rq->hrtick_timer;
|
|
|
|
ktime_t time = ktime_add_ns(timer->base->get_time(), delay);
|
2008-04-29 08:02:46 +00:00
|
|
|
|
2008-09-01 22:02:30 +00:00
|
|
|
hrtimer_set_expires(timer, time);
|
2008-07-18 16:01:23 +00:00
|
|
|
|
|
|
|
if (rq == this_rq()) {
|
|
|
|
hrtimer_restart(timer);
|
|
|
|
} else if (!rq->hrtick_csd_pending) {
|
2009-02-25 12:59:48 +00:00
|
|
|
__smp_call_function_single(cpu_of(rq), &rq->hrtick_csd, 0);
|
2008-07-18 16:01:23 +00:00
|
|
|
rq->hrtick_csd_pending = 1;
|
|
|
|
}
|
2008-04-29 08:02:46 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
hotplug_hrtick(struct notifier_block *nfb, unsigned long action, void *hcpu)
|
|
|
|
{
|
|
|
|
int cpu = (int)(long)hcpu;
|
|
|
|
|
|
|
|
switch (action) {
|
|
|
|
case CPU_UP_CANCELED:
|
|
|
|
case CPU_UP_CANCELED_FROZEN:
|
|
|
|
case CPU_DOWN_PREPARE:
|
|
|
|
case CPU_DOWN_PREPARE_FROZEN:
|
|
|
|
case CPU_DEAD:
|
|
|
|
case CPU_DEAD_FROZEN:
|
2008-07-18 16:01:23 +00:00
|
|
|
hrtick_clear(cpu_rq(cpu));
|
2008-04-29 08:02:46 +00:00
|
|
|
return NOTIFY_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
|
2008-09-22 21:55:45 +00:00
|
|
|
static __init void init_hrtick(void)
|
2008-04-29 08:02:46 +00:00
|
|
|
{
|
|
|
|
hotcpu_notifier(hotplug_hrtick, 0);
|
|
|
|
}
|
2008-07-18 16:01:23 +00:00
|
|
|
#else
|
|
|
|
/*
|
|
|
|
* Called to set the hrtick timer state.
|
|
|
|
*
|
|
|
|
* called with rq->lock held and irqs disabled
|
|
|
|
*/
|
|
|
|
static void hrtick_start(struct rq *rq, u64 delay)
|
|
|
|
{
|
2009-03-13 11:21:27 +00:00
|
|
|
__hrtimer_start_range_ns(&rq->hrtick_timer, ns_to_ktime(delay), 0,
|
2009-04-16 06:44:37 +00:00
|
|
|
HRTIMER_MODE_REL_PINNED, 0);
|
2008-07-18 16:01:23 +00:00
|
|
|
}
|
2008-04-29 08:02:46 +00:00
|
|
|
|
2008-09-22 21:55:46 +00:00
|
|
|
static inline void init_hrtick(void)
|
2008-01-25 20:08:29 +00:00
|
|
|
{
|
|
|
|
}
|
2008-07-18 16:01:23 +00:00
|
|
|
#endif /* CONFIG_SMP */
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
static void init_rq_hrtick(struct rq *rq)
|
2008-01-25 20:08:29 +00:00
|
|
|
{
|
2008-07-18 16:01:23 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
rq->hrtick_csd_pending = 0;
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
rq->hrtick_csd.flags = 0;
|
|
|
|
rq->hrtick_csd.func = __hrtick_start;
|
|
|
|
rq->hrtick_csd.info = rq;
|
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
hrtimer_init(&rq->hrtick_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
|
|
|
|
rq->hrtick_timer.function = hrtick;
|
2008-01-25 20:08:29 +00:00
|
|
|
}
|
2008-09-22 21:55:46 +00:00
|
|
|
#else /* CONFIG_SCHED_HRTICK */
|
2008-01-25 20:08:29 +00:00
|
|
|
static inline void hrtick_clear(struct rq *rq)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void init_rq_hrtick(struct rq *rq)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-04-29 08:02:46 +00:00
|
|
|
static inline void init_hrtick(void)
|
|
|
|
{
|
|
|
|
}
|
2008-09-22 21:55:46 +00:00
|
|
|
#endif /* CONFIG_SCHED_HRTICK */
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* resched_task - mark a task 'to be rescheduled now'.
|
|
|
|
*
|
|
|
|
* On UP this means the setting of the need_resched flag, on SMP it
|
|
|
|
* might also involve a cross-CPU call to trigger the scheduler on
|
|
|
|
* the target CPU.
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
|
|
|
|
#ifndef tsk_is_polling
|
|
|
|
#define tsk_is_polling(t) test_tsk_thread_flag(t, TIF_POLLING_NRFLAG)
|
|
|
|
#endif
|
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
static void resched_task(struct task_struct *p)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
assert_spin_locked(&task_rq(p)->lock);
|
|
|
|
|
2009-03-06 11:40:20 +00:00
|
|
|
if (test_tsk_need_resched(p))
|
2007-07-09 16:51:59 +00:00
|
|
|
return;
|
|
|
|
|
2009-03-06 11:40:20 +00:00
|
|
|
set_tsk_need_resched(p);
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
cpu = task_cpu(p);
|
|
|
|
if (cpu == smp_processor_id())
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* NEED_RESCHED must be visible before we test polling */
|
|
|
|
smp_mb();
|
|
|
|
if (!tsk_is_polling(p))
|
|
|
|
smp_send_reschedule(cpu);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void resched_cpu(int cpu)
|
|
|
|
{
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (!spin_trylock_irqsave(&rq->lock, flags))
|
|
|
|
return;
|
|
|
|
resched_task(cpu_curr(cpu));
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
}
|
2008-03-22 08:20:24 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_NO_HZ
|
|
|
|
/*
|
|
|
|
* When add_timer_on() enqueues a timer into the timer wheel of an
|
|
|
|
* idle CPU then this timer might expire before the next timer event
|
|
|
|
* which is scheduled to wake up that CPU. In case of a completely
|
|
|
|
* idle system the next event might even be infinite time into the
|
|
|
|
* future. wake_up_idle_cpu() ensures that the CPU is woken up and
|
|
|
|
* leaves the inner idle loop so the newly added timer is taken into
|
|
|
|
* account when the CPU goes back to idle and evaluates the timer
|
|
|
|
* wheel for the next timer event.
|
|
|
|
*/
|
|
|
|
void wake_up_idle_cpu(int cpu)
|
|
|
|
{
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
|
|
|
|
|
if (cpu == smp_processor_id())
|
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This is safe, as this function is called with the timer
|
|
|
|
* wheel base lock of (cpu) held. When the CPU is on the way
|
|
|
|
* to idle and has not yet set rq->curr to idle then it will
|
|
|
|
* be serialized on the timer wheel base lock and take the new
|
|
|
|
* timer into account automatically.
|
|
|
|
*/
|
|
|
|
if (rq->curr != rq->idle)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We can set TIF_RESCHED on the idle task of the other CPU
|
|
|
|
* lockless. The worst case is that the other CPU runs the
|
|
|
|
* idle task through an additional NOOP schedule()
|
|
|
|
*/
|
2009-03-06 11:40:20 +00:00
|
|
|
set_tsk_need_resched(rq->idle);
|
2008-03-22 08:20:24 +00:00
|
|
|
|
|
|
|
/* NEED_RESCHED must be visible before we test polling */
|
|
|
|
smp_mb();
|
|
|
|
if (!tsk_is_polling(rq->idle))
|
|
|
|
smp_send_reschedule(cpu);
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_NO_HZ */
|
2008-03-22 08:20:24 +00:00
|
|
|
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_SMP */
|
2008-07-18 16:01:23 +00:00
|
|
|
static void resched_task(struct task_struct *p)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
assert_spin_locked(&task_rq(p)->lock);
|
2008-07-18 16:01:23 +00:00
|
|
|
set_tsk_need_resched(p);
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SMP */
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
#if BITS_PER_LONG == 32
|
|
|
|
# define WMULT_CONST (~0UL)
|
|
|
|
#else
|
|
|
|
# define WMULT_CONST (1UL << 32)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define WMULT_SHIFT 32
|
|
|
|
|
2007-08-09 09:16:51 +00:00
|
|
|
/*
|
|
|
|
* Shift right and round:
|
|
|
|
*/
|
2007-09-05 12:32:49 +00:00
|
|
|
#define SRR(x, y) (((x) + (1UL << ((y) - 1))) >> (y))
|
2007-08-09 09:16:51 +00:00
|
|
|
|
2008-06-27 11:41:11 +00:00
|
|
|
/*
|
|
|
|
* delta *= weight / lw
|
|
|
|
*/
|
2007-08-02 15:41:40 +00:00
|
|
|
static unsigned long
|
2007-07-09 16:51:59 +00:00
|
|
|
calc_delta_mine(unsigned long delta_exec, unsigned long weight,
|
|
|
|
struct load_weight *lw)
|
|
|
|
{
|
|
|
|
u64 tmp;
|
|
|
|
|
2008-06-12 08:43:07 +00:00
|
|
|
if (!lw->inv_weight) {
|
|
|
|
if (BITS_PER_LONG > 32 && unlikely(lw->weight >= WMULT_CONST))
|
|
|
|
lw->inv_weight = 1;
|
|
|
|
else
|
|
|
|
lw->inv_weight = 1 + (WMULT_CONST-lw->weight/2)
|
|
|
|
/ (lw->weight+1);
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
tmp = (u64)delta_exec * weight;
|
|
|
|
/*
|
|
|
|
* Check whether we'd overflow the 64-bit multiplication:
|
|
|
|
*/
|
2007-08-09 09:16:51 +00:00
|
|
|
if (unlikely(tmp > WMULT_CONST))
|
2007-09-05 12:32:49 +00:00
|
|
|
tmp = SRR(SRR(tmp, WMULT_SHIFT/2) * lw->inv_weight,
|
2007-08-09 09:16:51 +00:00
|
|
|
WMULT_SHIFT/2);
|
|
|
|
else
|
2007-09-05 12:32:49 +00:00
|
|
|
tmp = SRR(tmp * lw->inv_weight, WMULT_SHIFT);
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2007-08-02 15:41:40 +00:00
|
|
|
return (unsigned long)min(tmp, (u64)(unsigned long)LONG_MAX);
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
static inline void update_load_add(struct load_weight *lw, unsigned long inc)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
lw->weight += inc;
|
2008-03-14 22:48:28 +00:00
|
|
|
lw->inv_weight = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:04 +00:00
|
|
|
static inline void update_load_sub(struct load_weight *lw, unsigned long dec)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
lw->weight -= dec;
|
2008-03-14 22:48:28 +00:00
|
|
|
lw->inv_weight = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
/*
|
|
|
|
* To aid in avoiding the subversion of "niceness" due to uneven distribution
|
|
|
|
* of tasks with abnormal "nice" values across CPUs the contribution that
|
|
|
|
* each task makes to its run queue's load is weighted according to its
|
2007-12-05 14:46:09 +00:00
|
|
|
* scheduling class and "nice" value. For SCHED_NORMAL tasks this is just a
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
* scaled version of the new time slice allocation that they receive on time
|
|
|
|
* slice expiry etc.
|
|
|
|
*/
|
|
|
|
|
2009-01-15 13:53:37 +00:00
|
|
|
#define WEIGHT_IDLEPRIO 3
|
|
|
|
#define WMULT_IDLEPRIO 1431655765
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Nice levels are multiplicative, with a gentle 10% change for every
|
|
|
|
* nice level changed. I.e. when a CPU-bound task goes from nice 0 to
|
|
|
|
* nice 1, it will get ~10% less CPU time than another CPU-bound task
|
|
|
|
* that remained on nice 0.
|
|
|
|
*
|
|
|
|
* The "10% effect" is relative and cumulative: from _any_ nice level,
|
|
|
|
* if you go up 1 level, it's -10% CPU usage, if you go down 1 level
|
2007-07-16 07:46:30 +00:00
|
|
|
* it's +10% CPU usage. (to achieve that we use a multiplier of 1.25.
|
|
|
|
* If a task goes up by ~10% and another task goes down by ~10% then
|
|
|
|
* the relative distance between them is ~25%.)
|
2007-07-09 16:51:59 +00:00
|
|
|
*/
|
|
|
|
static const int prio_to_weight[40] = {
|
2007-08-09 09:16:51 +00:00
|
|
|
/* -20 */ 88761, 71755, 56483, 46273, 36291,
|
|
|
|
/* -15 */ 29154, 23254, 18705, 14949, 11916,
|
|
|
|
/* -10 */ 9548, 7620, 6100, 4904, 3906,
|
|
|
|
/* -5 */ 3121, 2501, 1991, 1586, 1277,
|
|
|
|
/* 0 */ 1024, 820, 655, 526, 423,
|
|
|
|
/* 5 */ 335, 272, 215, 172, 137,
|
|
|
|
/* 10 */ 110, 87, 70, 56, 45,
|
|
|
|
/* 15 */ 36, 29, 23, 18, 15,
|
2007-07-09 16:51:59 +00:00
|
|
|
};
|
|
|
|
|
2007-07-16 07:46:31 +00:00
|
|
|
/*
|
|
|
|
* Inverse (2^32/x) values of the prio_to_weight[] array, precalculated.
|
|
|
|
*
|
|
|
|
* In cases where the weight does not change often, we can use the
|
|
|
|
* precalculated inverse to speed up arithmetics by turning divisions
|
|
|
|
* into multiplications:
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static const u32 prio_to_wmult[40] = {
|
2007-08-09 09:16:51 +00:00
|
|
|
/* -20 */ 48388, 59856, 76040, 92818, 118348,
|
|
|
|
/* -15 */ 147320, 184698, 229616, 287308, 360437,
|
|
|
|
/* -10 */ 449829, 563644, 704093, 875809, 1099582,
|
|
|
|
/* -5 */ 1376151, 1717300, 2157191, 2708050, 3363326,
|
|
|
|
/* 0 */ 4194304, 5237765, 6557202, 8165337, 10153587,
|
|
|
|
/* 5 */ 12820798, 15790321, 19976592, 24970740, 31350126,
|
|
|
|
/* 10 */ 39045157, 49367440, 61356676, 76695844, 95443717,
|
|
|
|
/* 15 */ 119304647, 148102320, 186737708, 238609294, 286331153,
|
2007-07-09 16:51:59 +00:00
|
|
|
};
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
static void activate_task(struct rq *rq, struct task_struct *p, int wakeup);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* runqueue iterator, to support SMP load-balancing between different
|
|
|
|
* scheduling classes, without having to expose their internal data
|
|
|
|
* structures to the load-balancing proper:
|
|
|
|
*/
|
|
|
|
struct rq_iterator {
|
|
|
|
void *arg;
|
|
|
|
struct task_struct *(*start)(void *);
|
|
|
|
struct task_struct *(*next)(void *);
|
|
|
|
};
|
|
|
|
|
2007-10-24 16:23:51 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static unsigned long
|
|
|
|
balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
|
|
|
unsigned long max_load_move, struct sched_domain *sd,
|
|
|
|
enum cpu_idle_type idle, int *all_pinned,
|
|
|
|
int *this_best_prio, struct rq_iterator *iterator);
|
|
|
|
|
|
|
|
static int
|
|
|
|
iter_move_one_task(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle,
|
|
|
|
struct rq_iterator *iterator);
|
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2009-03-31 04:32:22 +00:00
|
|
|
/* Time spent by the tasks of the cpu accounting group executing in ... */
|
|
|
|
enum cpuacct_stat_index {
|
|
|
|
CPUACCT_STAT_USER, /* ... user mode */
|
|
|
|
CPUACCT_STAT_SYSTEM, /* ... kernel mode */
|
|
|
|
|
|
|
|
CPUACCT_STAT_NSTATS,
|
|
|
|
};
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
#ifdef CONFIG_CGROUP_CPUACCT
|
|
|
|
static void cpuacct_charge(struct task_struct *tsk, u64 cputime);
|
2009-03-31 04:32:22 +00:00
|
|
|
static void cpuacct_update_stats(struct task_struct *tsk,
|
|
|
|
enum cpuacct_stat_index idx, cputime_t val);
|
2007-12-02 19:04:49 +00:00
|
|
|
#else
|
|
|
|
static inline void cpuacct_charge(struct task_struct *tsk, u64 cputime) {}
|
2009-03-31 04:32:22 +00:00
|
|
|
static inline void cpuacct_update_stats(struct task_struct *tsk,
|
|
|
|
enum cpuacct_stat_index idx, cputime_t val) {}
|
2007-12-02 19:04:49 +00:00
|
|
|
#endif
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
static inline void inc_cpu_load(struct rq *rq, unsigned long load)
|
|
|
|
{
|
|
|
|
update_load_add(&rq->load, load);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void dec_cpu_load(struct rq *rq, unsigned long load)
|
|
|
|
{
|
|
|
|
update_load_sub(&rq->load, load);
|
|
|
|
}
|
|
|
|
|
2008-08-19 11:40:47 +00:00
|
|
|
#if (defined(CONFIG_SMP) && defined(CONFIG_FAIR_GROUP_SCHED)) || defined(CONFIG_RT_GROUP_SCHED)
|
2008-08-19 10:33:05 +00:00
|
|
|
typedef int (*tg_visitor)(struct task_group *, void *);
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Iterate the full tree, calling @down when first entering a node and @up when
|
|
|
|
* leaving it for the final time.
|
|
|
|
*/
|
2008-08-19 10:33:05 +00:00
|
|
|
static int walk_tg_tree(tg_visitor down, tg_visitor up, void *data)
|
2008-06-27 11:41:14 +00:00
|
|
|
{
|
|
|
|
struct task_group *parent, *child;
|
2008-08-19 10:33:05 +00:00
|
|
|
int ret;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
parent = &root_task_group;
|
|
|
|
down:
|
2008-08-19 10:33:05 +00:00
|
|
|
ret = (*down)(parent, data);
|
|
|
|
if (ret)
|
|
|
|
goto out_unlock;
|
2008-06-27 11:41:14 +00:00
|
|
|
list_for_each_entry_rcu(child, &parent->children, siblings) {
|
|
|
|
parent = child;
|
|
|
|
goto down;
|
|
|
|
|
|
|
|
up:
|
|
|
|
continue;
|
|
|
|
}
|
2008-08-19 10:33:05 +00:00
|
|
|
ret = (*up)(parent, data);
|
|
|
|
if (ret)
|
|
|
|
goto out_unlock;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
child = parent;
|
|
|
|
parent = parent->parent;
|
|
|
|
if (parent)
|
|
|
|
goto up;
|
2008-08-19 10:33:05 +00:00
|
|
|
out_unlock:
|
2008-06-27 11:41:14 +00:00
|
|
|
rcu_read_unlock();
|
2008-08-19 10:33:05 +00:00
|
|
|
|
|
|
|
return ret;
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
|
|
|
|
2008-08-19 10:33:05 +00:00
|
|
|
static int tg_nop(struct task_group *tg, void *data)
|
|
|
|
{
|
|
|
|
return 0;
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
2008-08-19 10:33:05 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static unsigned long source_load(int cpu, int type);
|
|
|
|
static unsigned long target_load(int cpu, int type);
|
|
|
|
static int task_hot(struct task_struct *p, u64 now, struct sched_domain *sd);
|
|
|
|
|
|
|
|
static unsigned long cpu_avg_load_per_task(int cpu)
|
|
|
|
{
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2008-11-29 19:45:15 +00:00
|
|
|
unsigned long nr_running = ACCESS_ONCE(rq->nr_running);
|
2008-08-19 10:33:05 +00:00
|
|
|
|
2008-11-27 02:04:24 +00:00
|
|
|
if (nr_running)
|
|
|
|
rq->avg_load_per_task = rq->load.weight / nr_running;
|
2008-11-12 10:49:00 +00:00
|
|
|
else
|
|
|
|
rq->avg_load_per_task = 0;
|
2008-08-19 10:33:05 +00:00
|
|
|
|
|
|
|
return rq->avg_load_per_task;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
static void __set_se_shares(struct sched_entity *se, unsigned long shares);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Calculate and set the cpu's group shares.
|
|
|
|
*/
|
|
|
|
static void
|
2008-10-17 17:27:02 +00:00
|
|
|
update_group_shares_cpu(struct task_group *tg, int cpu,
|
|
|
|
unsigned long sd_shares, unsigned long sd_rq_weight)
|
2008-04-19 17:45:00 +00:00
|
|
|
{
|
2008-06-27 11:41:14 +00:00
|
|
|
unsigned long shares;
|
|
|
|
unsigned long rq_weight;
|
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
if (!tg->se[cpu])
|
2008-06-27 11:41:14 +00:00
|
|
|
return;
|
|
|
|
|
2008-11-19 06:41:57 +00:00
|
|
|
rq_weight = tg->cfs_rq[cpu]->rq_weight;
|
2008-06-27 11:41:23 +00:00
|
|
|
|
2008-06-27 11:41:14 +00:00
|
|
|
/*
|
|
|
|
* \Sum shares * rq_weight
|
|
|
|
* shares = -----------------------
|
|
|
|
* \Sum rq_weight
|
|
|
|
*
|
|
|
|
*/
|
2008-11-19 06:41:57 +00:00
|
|
|
shares = (sd_shares * rq_weight) / sd_rq_weight;
|
2008-10-17 17:27:02 +00:00
|
|
|
shares = clamp_t(unsigned long, shares, MIN_SHARES, MAX_SHARES);
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-10-17 17:27:02 +00:00
|
|
|
if (abs(shares - tg->se[cpu]->load.weight) >
|
|
|
|
sysctl_sched_shares_thresh) {
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
|
unsigned long flags;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-10-17 17:27:02 +00:00
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
2008-11-19 06:41:57 +00:00
|
|
|
tg->cfs_rq[cpu]->shares = shares;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-10-17 17:27:02 +00:00
|
|
|
__set_se_shares(tg->se[cpu], shares);
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
}
|
2008-04-19 17:45:00 +00:00
|
|
|
}
|
2008-06-27 11:41:14 +00:00
|
|
|
|
|
|
|
/*
|
2008-06-27 11:41:23 +00:00
|
|
|
* Re-compute the task group their per cpu shares over the given domain.
|
|
|
|
* This needs to be done in a bottom-up fashion because the rq weight of a
|
|
|
|
* parent group depends on the shares of its child groups.
|
2008-06-27 11:41:14 +00:00
|
|
|
*/
|
2008-08-19 10:33:05 +00:00
|
|
|
static int tg_shares_up(struct task_group *tg, void *data)
|
2008-06-27 11:41:14 +00:00
|
|
|
{
|
2008-11-19 06:41:57 +00:00
|
|
|
unsigned long weight, rq_weight = 0;
|
2008-06-27 11:41:23 +00:00
|
|
|
unsigned long shares = 0;
|
2008-08-19 10:33:05 +00:00
|
|
|
struct sched_domain *sd = data;
|
2008-06-27 11:41:23 +00:00
|
|
|
int i;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu(i, sched_domain_span(sd)) {
|
2008-11-19 06:41:57 +00:00
|
|
|
/*
|
|
|
|
* If there are currently no tasks on the cpu pretend there
|
|
|
|
* is one of average load so that when a new task gets to
|
|
|
|
* run here it will not get delayed by group starvation.
|
|
|
|
*/
|
|
|
|
weight = tg->cfs_rq[i]->load.weight;
|
|
|
|
if (!weight)
|
|
|
|
weight = NICE_0_LOAD;
|
|
|
|
|
|
|
|
tg->cfs_rq[i]->rq_weight = weight;
|
|
|
|
rq_weight += weight;
|
2008-06-27 11:41:23 +00:00
|
|
|
shares += tg->cfs_rq[i]->shares;
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
if ((!shares && rq_weight) || shares > tg->shares)
|
|
|
|
shares = tg->shares;
|
|
|
|
|
|
|
|
if (!sd->parent || !(sd->parent->flags & SD_LOAD_BALANCE))
|
|
|
|
shares = tg->shares;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu(i, sched_domain_span(sd))
|
2008-10-17 17:27:02 +00:00
|
|
|
update_group_shares_cpu(tg, i, shares, rq_weight);
|
2008-08-19 10:33:05 +00:00
|
|
|
|
|
|
|
return 0;
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2008-06-27 11:41:23 +00:00
|
|
|
* Compute the cpu's hierarchical load factor for each task group.
|
|
|
|
* This needs to be done in a top-down fashion because the load of a child
|
|
|
|
* group is a fraction of its parents load.
|
2008-06-27 11:41:14 +00:00
|
|
|
*/
|
2008-08-19 10:33:05 +00:00
|
|
|
static int tg_load_down(struct task_group *tg, void *data)
|
2008-06-27 11:41:14 +00:00
|
|
|
{
|
2008-06-27 11:41:23 +00:00
|
|
|
unsigned long load;
|
2008-08-19 10:33:05 +00:00
|
|
|
long cpu = (long)data;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
if (!tg->parent) {
|
|
|
|
load = cpu_rq(cpu)->load.weight;
|
|
|
|
} else {
|
|
|
|
load = tg->parent->cfs_rq[cpu]->h_load;
|
|
|
|
load *= tg->cfs_rq[cpu]->shares;
|
|
|
|
load /= tg->parent->cfs_rq[cpu]->load.weight + 1;
|
|
|
|
}
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
tg->cfs_rq[cpu]->h_load = load;
|
2008-06-27 11:41:14 +00:00
|
|
|
|
2008-08-19 10:33:05 +00:00
|
|
|
return 0;
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
static void update_shares(struct sched_domain *sd)
|
2008-06-27 11:41:19 +00:00
|
|
|
{
|
2008-06-27 11:41:35 +00:00
|
|
|
u64 now = cpu_clock(raw_smp_processor_id());
|
|
|
|
s64 elapsed = now - sd->last_update;
|
|
|
|
|
|
|
|
if (elapsed >= (s64)(u64)sysctl_sched_shares_ratelimit) {
|
|
|
|
sd->last_update = now;
|
2008-08-19 10:33:05 +00:00
|
|
|
walk_tg_tree(tg_nop, tg_shares_up, sd);
|
2008-06-27 11:41:35 +00:00
|
|
|
}
|
2008-06-27 11:41:19 +00:00
|
|
|
}
|
|
|
|
|
2008-06-27 11:41:24 +00:00
|
|
|
static void update_shares_locked(struct rq *rq, struct sched_domain *sd)
|
|
|
|
{
|
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
update_shares(sd);
|
|
|
|
spin_lock(&rq->lock);
|
|
|
|
}
|
|
|
|
|
2008-08-19 10:33:05 +00:00
|
|
|
static void update_h_load(long cpu)
|
2008-06-27 11:41:14 +00:00
|
|
|
{
|
2008-08-19 10:33:05 +00:00
|
|
|
walk_tg_tree(tg_load_down, tg_nop, (void *)cpu);
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2008-06-27 11:41:23 +00:00
|
|
|
static inline void update_shares(struct sched_domain *sd)
|
2008-06-27 11:41:19 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-06-27 11:41:24 +00:00
|
|
|
static inline void update_shares_locked(struct rq *rq, struct sched_domain *sd)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
#endif
|
|
|
|
|
2008-12-29 14:39:51 +00:00
|
|
|
#ifdef CONFIG_PREEMPT
|
|
|
|
|
2008-11-28 19:08:00 +00:00
|
|
|
/*
|
2008-12-29 14:39:51 +00:00
|
|
|
* fair double_lock_balance: Safely acquires both rq->locks in a fair
|
|
|
|
* way at the expense of forcing extra atomic operations in all
|
|
|
|
* invocations. This assures that the double_lock is acquired using the
|
|
|
|
* same underlying policy as the spinlock_t on this architecture, which
|
|
|
|
* reduces latency compared to the unfair variant below. However, it
|
|
|
|
* also adds more overhead and therefore may reduce throughput.
|
2008-11-28 19:08:00 +00:00
|
|
|
*/
|
2008-12-29 14:39:51 +00:00
|
|
|
static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest)
|
|
|
|
__releases(this_rq->lock)
|
|
|
|
__acquires(busiest->lock)
|
|
|
|
__acquires(this_rq->lock)
|
|
|
|
{
|
|
|
|
spin_unlock(&this_rq->lock);
|
|
|
|
double_rq_lock(this_rq, busiest);
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
/*
|
|
|
|
* Unfair double_lock_balance: Optimizes throughput at the expense of
|
|
|
|
* latency by eliminating extra atomic operations when the locks are
|
|
|
|
* already in proper order on entry. This favors lower cpu-ids and will
|
|
|
|
* grant the double lock to lower cpus over higher ids under contention,
|
|
|
|
* regardless of entry order into the function.
|
|
|
|
*/
|
|
|
|
static int _double_lock_balance(struct rq *this_rq, struct rq *busiest)
|
2008-11-28 19:08:00 +00:00
|
|
|
__releases(this_rq->lock)
|
|
|
|
__acquires(busiest->lock)
|
|
|
|
__acquires(this_rq->lock)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
if (unlikely(!spin_trylock(&busiest->lock))) {
|
|
|
|
if (busiest < this_rq) {
|
|
|
|
spin_unlock(&this_rq->lock);
|
|
|
|
spin_lock(&busiest->lock);
|
|
|
|
spin_lock_nested(&this_rq->lock, SINGLE_DEPTH_NESTING);
|
|
|
|
ret = 1;
|
|
|
|
} else
|
|
|
|
spin_lock_nested(&busiest->lock, SINGLE_DEPTH_NESTING);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-12-29 14:39:51 +00:00
|
|
|
#endif /* CONFIG_PREEMPT */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* double_lock_balance - lock the busiest runqueue, this_rq is locked already.
|
|
|
|
*/
|
|
|
|
static int double_lock_balance(struct rq *this_rq, struct rq *busiest)
|
|
|
|
{
|
|
|
|
if (unlikely(!irqs_disabled())) {
|
|
|
|
/* printk() doesn't work good under rq->lock */
|
|
|
|
spin_unlock(&this_rq->lock);
|
|
|
|
BUG_ON(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
return _double_lock_balance(this_rq, busiest);
|
|
|
|
}
|
|
|
|
|
2008-11-28 19:08:00 +00:00
|
|
|
static inline void double_unlock_balance(struct rq *this_rq, struct rq *busiest)
|
|
|
|
__releases(busiest->lock)
|
|
|
|
{
|
|
|
|
spin_unlock(&busiest->lock);
|
|
|
|
lock_set_subclass(&this_rq->lock.dep_map, 0, _RET_IP_);
|
|
|
|
}
|
2008-04-19 17:45:00 +00:00
|
|
|
#endif
|
|
|
|
|
2008-06-27 19:35:50 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2008-06-27 13:42:36 +00:00
|
|
|
static void cfs_rq_set_shares(struct cfs_rq *cfs_rq, unsigned long shares)
|
|
|
|
{
|
2008-06-27 19:35:50 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-06-27 13:42:36 +00:00
|
|
|
cfs_rq->shares = shares;
|
|
|
|
#endif
|
|
|
|
}
|
2008-06-27 19:35:50 +00:00
|
|
|
#endif
|
2008-01-25 20:08:09 +00:00
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
static void calc_load_account_active(struct rq *this_rq);
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
#include "sched_stats.h"
|
|
|
|
#include "sched_idletask.c"
|
2007-10-15 15:00:12 +00:00
|
|
|
#include "sched_fair.c"
|
|
|
|
#include "sched_rt.c"
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
|
# include "sched_debug.c"
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define sched_class_highest (&rt_sched_class)
|
2008-06-04 19:04:05 +00:00
|
|
|
#define for_each_class(class) \
|
|
|
|
for (class = sched_class_highest; class; class = class->next)
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2008-06-27 11:41:14 +00:00
|
|
|
static void inc_nr_running(struct rq *rq)
|
2007-08-02 15:41:40 +00:00
|
|
|
{
|
|
|
|
rq->nr_running++;
|
|
|
|
}
|
|
|
|
|
2008-06-27 11:41:14 +00:00
|
|
|
static void dec_nr_running(struct rq *rq)
|
2007-08-02 15:41:40 +00:00
|
|
|
{
|
|
|
|
rq->nr_running--;
|
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
static void set_load_weight(struct task_struct *p)
|
|
|
|
{
|
|
|
|
if (task_has_rt_policy(p)) {
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.load.weight = prio_to_weight[0] * 2;
|
|
|
|
p->se.load.inv_weight = prio_to_wmult[0] >> 1;
|
|
|
|
return;
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* SCHED_IDLE tasks get minimal weight:
|
|
|
|
*/
|
|
|
|
if (p->policy == SCHED_IDLE) {
|
|
|
|
p->se.load.weight = WEIGHT_IDLEPRIO;
|
|
|
|
p->se.load.inv_weight = WMULT_IDLEPRIO;
|
|
|
|
return;
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.load.weight = prio_to_weight[p->static_prio - MAX_RT_PRIO];
|
|
|
|
p->se.load.inv_weight = prio_to_wmult[p->static_prio - MAX_RT_PRIO];
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2008-06-27 20:30:00 +00:00
|
|
|
static void update_avg(u64 *avg, u64 sample)
|
|
|
|
{
|
|
|
|
s64 diff = sample - *avg;
|
|
|
|
*avg += diff >> 3;
|
|
|
|
}
|
|
|
|
|
2007-08-09 09:16:49 +00:00
|
|
|
static void enqueue_task(struct rq *rq, struct task_struct *p, int wakeup)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
2009-01-14 11:39:18 +00:00
|
|
|
if (wakeup)
|
|
|
|
p->se.start_runtime = p->se.sum_exec_runtime;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
sched_info_queued(p);
|
2007-08-09 09:16:48 +00:00
|
|
|
p->sched_class->enqueue_task(rq, p, wakeup);
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.on_rq = 1;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2007-08-09 09:16:49 +00:00
|
|
|
static void dequeue_task(struct rq *rq, struct task_struct *p, int sleep)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
2009-01-14 11:39:18 +00:00
|
|
|
if (sleep) {
|
|
|
|
if (p->se.last_wakeup) {
|
|
|
|
update_avg(&p->se.avg_overlap,
|
|
|
|
p->se.sum_exec_runtime - p->se.last_wakeup);
|
|
|
|
p->se.last_wakeup = 0;
|
|
|
|
} else {
|
|
|
|
update_avg(&p->se.avg_wakeup,
|
|
|
|
sysctl_sched_wakeup_granularity);
|
|
|
|
}
|
2008-06-27 20:30:00 +00:00
|
|
|
}
|
|
|
|
|
sched: fix accounting in task delay accounting & migration
On Thu, Jun 19, 2008 at 12:27:14PM +0200, Peter Zijlstra wrote:
> On Thu, 2008-06-05 at 10:50 +0530, Ankita Garg wrote:
>
> > Thanks Peter for the explanation...
> >
> > I agree with the above and that is the reason why I did not see weird
> > values with cpu_time. But, run_delay still would suffer skews as the end
> > points for delta could be taken on different cpus due to migration (more
> > so on RT kernel due to the push-pull operations). With the below patch,
> > I could not reproduce the issue I had seen earlier. After every dequeue,
> > we take the delta and start wait measurements from zero when moved to a
> > different rq.
>
> OK, so task delay delay accounting is broken because it doesn't take
> migration into account.
>
> What you've done is make it symmetric wrt enqueue, and account it like
>
> cpu0 cpu1
>
> enqueue
> <wait-d1>
> dequeue
> enqueue
> <wait-d2>
> run
>
> Where you add both d1 and d2 to the run_delay,.. right?
>
Thanks for reviewing the patch. The above is exactly what I have done.
> This seems like a good fix, however it looks like the patch will break
> compilation in !CONFIG_SCHEDSTATS && !CONFIG_TASK_DELAY_ACCT, of it
> failing to provide a stub for sched_info_dequeue() in that case.
Fixed. Pl. find the new patch below.
Signed-off-by: Ankita Garg <ankita@in.ibm.com>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Cc: Gregory Haskins <ghaskins@novell.com>
Cc: rostedt@goodmis.org
Cc: suresh.b.siddha@intel.com
Cc: aneesh.kumar@linux.vnet.ibm.com
Cc: dhaval@linux.vnet.ibm.com
Cc: vatsa@linux.vnet.ibm.com
Cc: David Bahi <DBahi@novell.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-07-01 09:00:06 +00:00
|
|
|
sched_info_dequeued(p);
|
2007-08-09 09:16:48 +00:00
|
|
|
p->sched_class->dequeue_task(rq, p, sleep);
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.on_rq = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* __normal_prio - return the priority that is based on the static prio
|
2007-07-09 16:51:59 +00:00
|
|
|
*/
|
|
|
|
static inline int __normal_prio(struct task_struct *p)
|
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
return p->static_prio;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
/*
|
|
|
|
* Calculate the expected normal priority: i.e. priority
|
|
|
|
* without taking RT-inheritance into account. Might be
|
|
|
|
* boosted by interactivity modifiers. Changes upon fork,
|
|
|
|
* setprio syscalls, and whenever the interactivity
|
|
|
|
* estimator recalculates.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
static inline int normal_prio(struct task_struct *p)
|
2006-06-27 09:54:51 +00:00
|
|
|
{
|
|
|
|
int prio;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (task_has_rt_policy(p))
|
2006-06-27 09:54:51 +00:00
|
|
|
prio = MAX_RT_PRIO-1 - p->rt_priority;
|
|
|
|
else
|
|
|
|
prio = __normal_prio(p);
|
|
|
|
return prio;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Calculate the current priority, i.e. the priority
|
|
|
|
* taken into account by the scheduler. This value might
|
|
|
|
* be boosted by RT tasks, or might be boosted by
|
|
|
|
* interactivity modifiers. Will be RT if the task got
|
|
|
|
* RT-boosted. If not then it returns p->normal_prio.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
static int effective_prio(struct task_struct *p)
|
2006-06-27 09:54:51 +00:00
|
|
|
{
|
|
|
|
p->normal_prio = normal_prio(p);
|
|
|
|
/*
|
|
|
|
* If we are RT tasks or we were boosted to RT priority,
|
|
|
|
* keep the priority unchanged. Otherwise, update priority
|
|
|
|
* to the normal priority:
|
|
|
|
*/
|
|
|
|
if (!rt_prio(p->prio))
|
|
|
|
return p->normal_prio;
|
|
|
|
return p->prio;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* activate_task - move a task to the runqueue.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static void activate_task(struct rq *rq, struct task_struct *p, int wakeup)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-12-06 16:07:07 +00:00
|
|
|
if (task_contributes_to_load(p))
|
2007-07-09 16:51:59 +00:00
|
|
|
rq->nr_uninterruptible--;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-08-09 09:16:49 +00:00
|
|
|
enqueue_task(rq, p, wakeup);
|
2008-06-27 11:41:14 +00:00
|
|
|
inc_nr_running(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* deactivate_task - remove a task from the runqueue.
|
|
|
|
*/
|
2007-08-09 09:16:49 +00:00
|
|
|
static void deactivate_task(struct rq *rq, struct task_struct *p, int sleep)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-12-06 16:07:07 +00:00
|
|
|
if (task_contributes_to_load(p))
|
2007-07-09 16:51:59 +00:00
|
|
|
rq->nr_uninterruptible++;
|
|
|
|
|
2007-08-09 09:16:49 +00:00
|
|
|
dequeue_task(rq, p, sleep);
|
2008-06-27 11:41:14 +00:00
|
|
|
dec_nr_running(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* task_curr - is this task currently executing on a CPU?
|
|
|
|
* @p: the task in question.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
inline int task_curr(const struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return cpu_curr(task_cpu(p)) == p;
|
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu)
|
|
|
|
{
|
2008-01-25 20:08:30 +00:00
|
|
|
set_task_rq(p, cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2007-11-15 19:57:40 +00:00
|
|
|
/*
|
|
|
|
* After ->cpu is set up to a new value, task_rq_lock(p, ...) can be
|
|
|
|
* successfuly executed on another CPU. We must ensure that updates of
|
|
|
|
* per-task data have been completed by this moment.
|
|
|
|
*/
|
|
|
|
smp_wmb();
|
2007-07-09 16:51:59 +00:00
|
|
|
task_thread_info(p)->cpu = cpu;
|
|
|
|
#endif
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:22 +00:00
|
|
|
static inline void check_class_changed(struct rq *rq, struct task_struct *p,
|
|
|
|
const struct sched_class *prev_class,
|
|
|
|
int oldprio, int running)
|
|
|
|
{
|
|
|
|
if (prev_class != p->sched_class) {
|
|
|
|
if (prev_class->switched_from)
|
|
|
|
prev_class->switched_from(rq, p, running);
|
|
|
|
p->sched_class->switched_to(rq, p, running);
|
|
|
|
} else
|
|
|
|
p->sched_class->prio_changed(rq, p, oldprio, running);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2007-07-09 16:51:58 +00:00
|
|
|
|
2008-06-04 21:22:32 +00:00
|
|
|
/* Used instead of source_load when we know the type == 0 */
|
|
|
|
static unsigned long weighted_cpuload(const int cpu)
|
|
|
|
{
|
|
|
|
return cpu_rq(cpu)->load.weight;
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
/*
|
|
|
|
* Is this task likely cache-hot:
|
|
|
|
*/
|
2008-01-25 20:08:09 +00:00
|
|
|
static int
|
2007-10-15 15:00:18 +00:00
|
|
|
task_hot(struct task_struct *p, u64 now, struct sched_domain *sd)
|
|
|
|
{
|
|
|
|
s64 delta;
|
|
|
|
|
2008-03-15 16:10:34 +00:00
|
|
|
/*
|
|
|
|
* Buddy candidates are cache hot:
|
|
|
|
*/
|
sched: backward looking buddy
Impact: improve/change/fix wakeup-buddy scheduling
Currently we only have a forward looking buddy, that is, we prefer to
schedule to the task we last woke up, under the presumption that its
going to consume the data we just produced, and therefore will have
cache hot benefits.
This allows co-waking producer/consumer task pairs to run ahead of the
pack for a little while, keeping their cache warm. Without this, we
would interleave all pairs, utterly trashing the cache.
This patch introduces a backward looking buddy, that is, suppose that
in the above scenario, the consumer preempts the producer before it
can go to sleep, we will therefore miss the wakeup from consumer to
producer (its already running, after all), breaking the cycle and
reverting to the cache-trashing interleaved schedule pattern.
The backward buddy will try to schedule back to the task that woke us
up in case the forward buddy is not available, under the assumption
that the last task will be the one with the most cache hot task around
barring current.
This will basically allow a task to continue after it got preempted.
In order to avoid starvation, we allow either buddy to get wakeup_gran
ahead of the pack.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Acked-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-11-04 20:25:09 +00:00
|
|
|
if (sched_feat(CACHE_HOT_BUDDY) &&
|
|
|
|
(&p->se == cfs_rq_of(&p->se)->next ||
|
|
|
|
&p->se == cfs_rq_of(&p->se)->last))
|
2008-03-15 16:10:34 +00:00
|
|
|
return 1;
|
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
if (p->sched_class != &fair_sched_class)
|
|
|
|
return 0;
|
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
if (sysctl_sched_migration_cost == -1)
|
|
|
|
return 1;
|
|
|
|
if (sysctl_sched_migration_cost == 0)
|
|
|
|
return 0;
|
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
delta = now - p->se.exec_start;
|
|
|
|
|
|
|
|
return delta < (s64)sysctl_sched_migration_cost;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
void set_task_cpu(struct task_struct *p, unsigned int new_cpu)
|
2007-07-09 16:51:58 +00:00
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
int old_cpu = task_cpu(p);
|
|
|
|
struct rq *old_rq = cpu_rq(old_cpu), *new_rq = cpu_rq(new_cpu);
|
2007-10-15 15:00:12 +00:00
|
|
|
struct cfs_rq *old_cfsrq = task_cfs_rq(p),
|
|
|
|
*new_cfsrq = cpu_cfs_rq(old_cfsrq, new_cpu);
|
2007-10-15 15:00:06 +00:00
|
|
|
u64 clock_offset;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
clock_offset = old_rq->clock - new_rq->clock;
|
2007-08-02 15:41:40 +00:00
|
|
|
|
2009-05-05 08:49:59 +00:00
|
|
|
trace_sched_migrate_task(p, new_cpu);
|
2008-12-10 07:08:22 +00:00
|
|
|
|
2007-08-02 15:41:40 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
|
|
|
if (p->se.wait_start)
|
|
|
|
p->se.wait_start -= clock_offset;
|
2007-07-09 16:51:59 +00:00
|
|
|
if (p->se.sleep_start)
|
|
|
|
p->se.sleep_start -= clock_offset;
|
|
|
|
if (p->se.block_start)
|
|
|
|
p->se.block_start -= clock_offset;
|
2008-12-14 11:34:15 +00:00
|
|
|
#endif
|
2007-10-15 15:00:18 +00:00
|
|
|
if (old_cpu != new_cpu) {
|
2008-12-14 11:34:15 +00:00
|
|
|
p->se.nr_migrations++;
|
perf_counters: make software counters work as per-cpu counters
Impact: kernel crash fix
Yanmin Zhang reported that using a PERF_COUNT_TASK_CLOCK software
counter as a per-cpu counter would reliably crash the system, because
it calls __task_delta_exec with a null pointer. The page fault,
context switch and cpu migration counters also won't function
correctly as per-cpu counters since they reference the current task.
This fixes the problem by redirecting the task_clock counter to the
cpu_clock counter when used as a per-cpu counter, and by implementing
per-cpu page fault, context switch and cpu migration counters.
Along the way, this:
- Initializes counter->ctx earlier, in perf_counter_alloc, so that
sw_perf_counter_init can use it
- Adds code to kernel/sched.c to count task migrations into each
cpu, in rq->nr_migrations_in
- Exports the per-cpu context switch and task migration counts
via new functions added to kernel/sched.c
- Makes sure that if sw_perf_counter_init fails, we don't try to
initialize the counter as a hardware counter. Since the user has
passed a negative, non-raw event type, they clearly don't intend
for it to be interpreted as a hardware event.
Reported-by: "Zhang Yanmin" <yanmin_zhang@linux.intel.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-09 11:42:47 +00:00
|
|
|
new_rq->nr_migrations_in++;
|
2008-12-14 11:34:15 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
2007-10-15 15:00:18 +00:00
|
|
|
if (task_hot(p, old_rq->clock, NULL))
|
|
|
|
schedstat_inc(p, se.nr_forced2_migrations);
|
2007-08-02 15:41:40 +00:00
|
|
|
#endif
|
2009-06-19 11:22:51 +00:00
|
|
|
perf_swcounter_event(PERF_COUNT_SW_CPU_MIGRATIONS,
|
|
|
|
1, 1, NULL, 0);
|
2008-12-14 11:34:15 +00:00
|
|
|
}
|
2007-10-15 15:00:12 +00:00
|
|
|
p->se.vruntime -= old_cfsrq->min_vruntime -
|
|
|
|
new_cfsrq->min_vruntime;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
__set_task_cpu(p, new_cpu);
|
2007-07-09 16:51:58 +00:00
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req {
|
2005-04-16 22:20:36 +00:00
|
|
|
struct list_head list;
|
|
|
|
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *task;
|
2005-04-16 22:20:36 +00:00
|
|
|
int dest_cpu;
|
|
|
|
|
|
|
|
struct completion done;
|
2006-07-03 07:25:42 +00:00
|
|
|
};
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The task's runqueue lock must be held.
|
|
|
|
* Returns true if you have to wait for migration thread.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
static int
|
2006-07-03 07:25:42 +00:00
|
|
|
migrate_task(struct task_struct *p, int dest_cpu, struct migration_req *req)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = task_rq(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If the task is not on a runqueue (and not running), then
|
|
|
|
* it is sufficient to simply update the task's cpu field.
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
if (!p->se.on_rq && !task_running(rq, p)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
set_task_cpu(p, dest_cpu);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
init_completion(&req->done);
|
|
|
|
req->task = p;
|
|
|
|
req->dest_cpu = dest_cpu;
|
|
|
|
list_add(&req->list, &rq->migration_queue);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2009-04-03 14:43:34 +00:00
|
|
|
/*
|
|
|
|
* wait_task_context_switch - wait for a thread to complete at least one
|
|
|
|
* context switch.
|
|
|
|
*
|
|
|
|
* @p must not be current.
|
|
|
|
*/
|
|
|
|
void wait_task_context_switch(struct task_struct *p)
|
|
|
|
{
|
|
|
|
unsigned long nvcsw, nivcsw, flags;
|
|
|
|
int running;
|
|
|
|
struct rq *rq;
|
|
|
|
|
|
|
|
nvcsw = p->nvcsw;
|
|
|
|
nivcsw = p->nivcsw;
|
|
|
|
for (;;) {
|
|
|
|
/*
|
|
|
|
* The runqueue is assigned before the actual context
|
|
|
|
* switch. We need to take the runqueue lock.
|
|
|
|
*
|
|
|
|
* We could check initially without the lock but it is
|
|
|
|
* very likely that we need to take the lock in every
|
|
|
|
* iteration.
|
|
|
|
*/
|
|
|
|
rq = task_rq_lock(p, &flags);
|
|
|
|
running = task_running(rq, p);
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
|
|
|
|
if (likely(!running))
|
|
|
|
break;
|
|
|
|
/*
|
|
|
|
* The switch count is incremented before the actual
|
|
|
|
* context switch. We thus wait for two switches to be
|
|
|
|
* sure at least one completed.
|
|
|
|
*/
|
|
|
|
if ((p->nvcsw - nvcsw) > 1)
|
|
|
|
break;
|
|
|
|
if ((p->nivcsw - nivcsw) > 1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
cpu_relax();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* wait_task_inactive - wait for a thread to unschedule.
|
|
|
|
*
|
2008-07-26 02:45:58 +00:00
|
|
|
* If @match_state is nonzero, it's the @p->state value just checked and
|
|
|
|
* not expected to change. If it changes, i.e. @p might have woken up,
|
|
|
|
* then return zero. When we succeed in waiting for @p to be off its CPU,
|
|
|
|
* we return a positive number (its total switch count). If a second call
|
|
|
|
* a short while later returns the same number, the caller can be sure that
|
|
|
|
* @p has remained unscheduled the whole time.
|
|
|
|
*
|
2005-04-16 22:20:36 +00:00
|
|
|
* The caller must ensure that the task *will* unschedule sometime soon,
|
|
|
|
* else this function might spin for a *long* time. This function can't
|
|
|
|
* be called with interrupts off, or it may introduce deadlock with
|
|
|
|
* smp_call_function() if an IPI is sent by the same process we are
|
|
|
|
* waiting to become inactive.
|
|
|
|
*/
|
2008-07-26 02:45:58 +00:00
|
|
|
unsigned long wait_task_inactive(struct task_struct *p, long match_state)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
2007-07-09 16:51:59 +00:00
|
|
|
int running, on_rq;
|
2008-07-26 02:45:58 +00:00
|
|
|
unsigned long ncsw;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
for (;;) {
|
|
|
|
/*
|
|
|
|
* We do the initial early heuristics without holding
|
|
|
|
* any task-queue locks at all. We'll only try to get
|
|
|
|
* the runqueue lock when things look like they will
|
|
|
|
* work out!
|
|
|
|
*/
|
|
|
|
rq = task_rq(p);
|
Fix possible runqueue lock starvation in wait_task_inactive()
Miklos Szeredi reported very long pauses (several seconds, sometimes
more) on his T60 (with a Core2Duo) which he managed to track down to
wait_task_inactive()'s open-coded busy-loop.
He observed that an interrupt on one core tries to acquire the
runqueue-lock but does not succeed in doing so for a very long time -
while wait_task_inactive() on the other core loops waiting for the first
core to deschedule a task (which it wont do while spinning in an
interrupt handler).
This rewrites wait_task_inactive() to do all its waiting optimistically
without any locks taken at all, and then just double-check the end
result with the proper runqueue lock held over just a very short
section. If there were races in the optimistic wait, of a preemption
event scheduled the process away, we simply re-synchronize, and start
over.
So the code now looks like this:
repeat:
/* Unlocked, optimistic looping! */
rq = task_rq(p);
while (task_running(rq, p))
cpu_relax();
/* Get the *real* values */
rq = task_rq_lock(p, &flags);
running = task_running(rq, p);
array = p->array;
task_rq_unlock(rq, &flags);
/* Check them.. */
if (unlikely(running)) {
cpu_relax();
goto repeat;
}
/* Preempted away? Yield if so.. */
if (unlikely(array)) {
yield();
goto repeat;
}
Basically, that first "while()" loop is done entirely without any
locking at all (and doesn't check for the case where the target process
might have been preempted away), and so it's possibly "incorrect", but
we don't really care. Both the runqueue used, and the "task_running()"
check might be the wrong tests, but they won't oops - they just mean
that we could possibly get the wrong results due to lack of locking and
exit the loop early in the case of a race condition.
So once we've exited the loop, we then get the proper (and careful) rq
lock, and check the running/runnable state _safely_. And if it turns
out that our quick-and-dirty and unsafe loop was wrong after all, we
just go back and try it all again.
(The patch also adds a lot of comments, which is the actual bulk of it
all, to make it more obvious why we can do these things without holding
the locks).
Thanks to Miklos for all the testing and tracking it down.
Tested-by: Miklos Szeredi <miklos@szeredi.hu>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-06-18 16:34:40 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* If the task is actively running on another CPU
|
|
|
|
* still, just relax and busy-wait without holding
|
|
|
|
* any locks.
|
|
|
|
*
|
|
|
|
* NOTE! Since we don't hold any locks, it's not
|
|
|
|
* even sure that "rq" stays as the right runqueue!
|
|
|
|
* But we don't care, since "task_running()" will
|
|
|
|
* return false if the runqueue has changed and p
|
|
|
|
* is actually now running somewhere else!
|
|
|
|
*/
|
2008-07-26 02:45:58 +00:00
|
|
|
while (task_running(rq, p)) {
|
|
|
|
if (match_state && unlikely(p->state != match_state))
|
|
|
|
return 0;
|
2007-10-15 15:00:14 +00:00
|
|
|
cpu_relax();
|
2008-07-26 02:45:58 +00:00
|
|
|
}
|
Fix possible runqueue lock starvation in wait_task_inactive()
Miklos Szeredi reported very long pauses (several seconds, sometimes
more) on his T60 (with a Core2Duo) which he managed to track down to
wait_task_inactive()'s open-coded busy-loop.
He observed that an interrupt on one core tries to acquire the
runqueue-lock but does not succeed in doing so for a very long time -
while wait_task_inactive() on the other core loops waiting for the first
core to deschedule a task (which it wont do while spinning in an
interrupt handler).
This rewrites wait_task_inactive() to do all its waiting optimistically
without any locks taken at all, and then just double-check the end
result with the proper runqueue lock held over just a very short
section. If there were races in the optimistic wait, of a preemption
event scheduled the process away, we simply re-synchronize, and start
over.
So the code now looks like this:
repeat:
/* Unlocked, optimistic looping! */
rq = task_rq(p);
while (task_running(rq, p))
cpu_relax();
/* Get the *real* values */
rq = task_rq_lock(p, &flags);
running = task_running(rq, p);
array = p->array;
task_rq_unlock(rq, &flags);
/* Check them.. */
if (unlikely(running)) {
cpu_relax();
goto repeat;
}
/* Preempted away? Yield if so.. */
if (unlikely(array)) {
yield();
goto repeat;
}
Basically, that first "while()" loop is done entirely without any
locking at all (and doesn't check for the case where the target process
might have been preempted away), and so it's possibly "incorrect", but
we don't really care. Both the runqueue used, and the "task_running()"
check might be the wrong tests, but they won't oops - they just mean
that we could possibly get the wrong results due to lack of locking and
exit the loop early in the case of a race condition.
So once we've exited the loop, we then get the proper (and careful) rq
lock, and check the running/runnable state _safely_. And if it turns
out that our quick-and-dirty and unsafe loop was wrong after all, we
just go back and try it all again.
(The patch also adds a lot of comments, which is the actual bulk of it
all, to make it more obvious why we can do these things without holding
the locks).
Thanks to Miklos for all the testing and tracking it down.
Tested-by: Miklos Szeredi <miklos@szeredi.hu>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-06-18 16:34:40 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Ok, time to look more closely! We need the rq
|
|
|
|
* lock now, to be *sure*. If we're wrong, we'll
|
|
|
|
* just go back and repeat.
|
|
|
|
*/
|
|
|
|
rq = task_rq_lock(p, &flags);
|
tracing, sched: LTTng instrumentation - scheduler
Instrument the scheduler activity (sched_switch, migration, wakeups,
wait for a task, signal delivery) and process/thread
creation/destruction (fork, exit, kthread stop). Actually, kthread
creation is not instrumented in this patch because it is architecture
dependent. It allows to connect tracers such as ftrace which detects
scheduling latencies, good/bad scheduler decisions. Tools like LTTng can
export this scheduler information along with instrumentation of the rest
of the kernel activity to perform post-mortem analysis on the scheduler
activity.
About the performance impact of tracepoints (which is comparable to
markers), even without immediate values optimizations, tests done by
Hideo Aoki on ia64 show no regression. His test case was using hackbench
on a kernel where scheduler instrumentation (about 5 events in code
scheduler code) was added. See the "Tracepoints" patch header for
performance result detail.
Changelog :
- Change instrumentation location and parameter to match ftrace
instrumentation, previously done with kernel markers.
[ mingo@elte.hu: conflict resolutions ]
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
Acked-by: 'Peter Zijlstra' <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-07-18 16:16:17 +00:00
|
|
|
trace_sched_wait_task(rq, p);
|
2007-10-15 15:00:14 +00:00
|
|
|
running = task_running(rq, p);
|
|
|
|
on_rq = p->se.on_rq;
|
2008-07-26 02:45:58 +00:00
|
|
|
ncsw = 0;
|
2008-08-20 23:54:44 +00:00
|
|
|
if (!match_state || p->state == match_state)
|
2008-08-20 23:54:44 +00:00
|
|
|
ncsw = p->nvcsw | LONG_MIN; /* sets MSB */
|
2007-10-15 15:00:14 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
Fix possible runqueue lock starvation in wait_task_inactive()
Miklos Szeredi reported very long pauses (several seconds, sometimes
more) on his T60 (with a Core2Duo) which he managed to track down to
wait_task_inactive()'s open-coded busy-loop.
He observed that an interrupt on one core tries to acquire the
runqueue-lock but does not succeed in doing so for a very long time -
while wait_task_inactive() on the other core loops waiting for the first
core to deschedule a task (which it wont do while spinning in an
interrupt handler).
This rewrites wait_task_inactive() to do all its waiting optimistically
without any locks taken at all, and then just double-check the end
result with the proper runqueue lock held over just a very short
section. If there were races in the optimistic wait, of a preemption
event scheduled the process away, we simply re-synchronize, and start
over.
So the code now looks like this:
repeat:
/* Unlocked, optimistic looping! */
rq = task_rq(p);
while (task_running(rq, p))
cpu_relax();
/* Get the *real* values */
rq = task_rq_lock(p, &flags);
running = task_running(rq, p);
array = p->array;
task_rq_unlock(rq, &flags);
/* Check them.. */
if (unlikely(running)) {
cpu_relax();
goto repeat;
}
/* Preempted away? Yield if so.. */
if (unlikely(array)) {
yield();
goto repeat;
}
Basically, that first "while()" loop is done entirely without any
locking at all (and doesn't check for the case where the target process
might have been preempted away), and so it's possibly "incorrect", but
we don't really care. Both the runqueue used, and the "task_running()"
check might be the wrong tests, but they won't oops - they just mean
that we could possibly get the wrong results due to lack of locking and
exit the loop early in the case of a race condition.
So once we've exited the loop, we then get the proper (and careful) rq
lock, and check the running/runnable state _safely_. And if it turns
out that our quick-and-dirty and unsafe loop was wrong after all, we
just go back and try it all again.
(The patch also adds a lot of comments, which is the actual bulk of it
all, to make it more obvious why we can do these things without holding
the locks).
Thanks to Miklos for all the testing and tracking it down.
Tested-by: Miklos Szeredi <miklos@szeredi.hu>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-06-18 16:34:40 +00:00
|
|
|
|
2008-07-26 02:45:58 +00:00
|
|
|
/*
|
|
|
|
* If it changed from the expected state, bail out now.
|
|
|
|
*/
|
|
|
|
if (unlikely(!ncsw))
|
|
|
|
break;
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Was it really running after all now that we
|
|
|
|
* checked with the proper locks actually held?
|
|
|
|
*
|
|
|
|
* Oops. Go back and try again..
|
|
|
|
*/
|
|
|
|
if (unlikely(running)) {
|
|
|
|
cpu_relax();
|
|
|
|
continue;
|
|
|
|
}
|
Fix possible runqueue lock starvation in wait_task_inactive()
Miklos Szeredi reported very long pauses (several seconds, sometimes
more) on his T60 (with a Core2Duo) which he managed to track down to
wait_task_inactive()'s open-coded busy-loop.
He observed that an interrupt on one core tries to acquire the
runqueue-lock but does not succeed in doing so for a very long time -
while wait_task_inactive() on the other core loops waiting for the first
core to deschedule a task (which it wont do while spinning in an
interrupt handler).
This rewrites wait_task_inactive() to do all its waiting optimistically
without any locks taken at all, and then just double-check the end
result with the proper runqueue lock held over just a very short
section. If there were races in the optimistic wait, of a preemption
event scheduled the process away, we simply re-synchronize, and start
over.
So the code now looks like this:
repeat:
/* Unlocked, optimistic looping! */
rq = task_rq(p);
while (task_running(rq, p))
cpu_relax();
/* Get the *real* values */
rq = task_rq_lock(p, &flags);
running = task_running(rq, p);
array = p->array;
task_rq_unlock(rq, &flags);
/* Check them.. */
if (unlikely(running)) {
cpu_relax();
goto repeat;
}
/* Preempted away? Yield if so.. */
if (unlikely(array)) {
yield();
goto repeat;
}
Basically, that first "while()" loop is done entirely without any
locking at all (and doesn't check for the case where the target process
might have been preempted away), and so it's possibly "incorrect", but
we don't really care. Both the runqueue used, and the "task_running()"
check might be the wrong tests, but they won't oops - they just mean
that we could possibly get the wrong results due to lack of locking and
exit the loop early in the case of a race condition.
So once we've exited the loop, we then get the proper (and careful) rq
lock, and check the running/runnable state _safely_. And if it turns
out that our quick-and-dirty and unsafe loop was wrong after all, we
just go back and try it all again.
(The patch also adds a lot of comments, which is the actual bulk of it
all, to make it more obvious why we can do these things without holding
the locks).
Thanks to Miklos for all the testing and tracking it down.
Tested-by: Miklos Szeredi <miklos@szeredi.hu>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-06-18 16:34:40 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* It's not enough that it's not actively running,
|
|
|
|
* it must be off the runqueue _entirely_, and not
|
|
|
|
* preempted!
|
|
|
|
*
|
2009-03-16 19:58:09 +00:00
|
|
|
* So if it was still runnable (but just not actively
|
2007-10-15 15:00:14 +00:00
|
|
|
* running right now), it's preempted, and we should
|
|
|
|
* yield - it could be a while.
|
|
|
|
*/
|
|
|
|
if (unlikely(on_rq)) {
|
|
|
|
schedule_timeout_uninterruptible(1);
|
|
|
|
continue;
|
|
|
|
}
|
Fix possible runqueue lock starvation in wait_task_inactive()
Miklos Szeredi reported very long pauses (several seconds, sometimes
more) on his T60 (with a Core2Duo) which he managed to track down to
wait_task_inactive()'s open-coded busy-loop.
He observed that an interrupt on one core tries to acquire the
runqueue-lock but does not succeed in doing so for a very long time -
while wait_task_inactive() on the other core loops waiting for the first
core to deschedule a task (which it wont do while spinning in an
interrupt handler).
This rewrites wait_task_inactive() to do all its waiting optimistically
without any locks taken at all, and then just double-check the end
result with the proper runqueue lock held over just a very short
section. If there were races in the optimistic wait, of a preemption
event scheduled the process away, we simply re-synchronize, and start
over.
So the code now looks like this:
repeat:
/* Unlocked, optimistic looping! */
rq = task_rq(p);
while (task_running(rq, p))
cpu_relax();
/* Get the *real* values */
rq = task_rq_lock(p, &flags);
running = task_running(rq, p);
array = p->array;
task_rq_unlock(rq, &flags);
/* Check them.. */
if (unlikely(running)) {
cpu_relax();
goto repeat;
}
/* Preempted away? Yield if so.. */
if (unlikely(array)) {
yield();
goto repeat;
}
Basically, that first "while()" loop is done entirely without any
locking at all (and doesn't check for the case where the target process
might have been preempted away), and so it's possibly "incorrect", but
we don't really care. Both the runqueue used, and the "task_running()"
check might be the wrong tests, but they won't oops - they just mean
that we could possibly get the wrong results due to lack of locking and
exit the loop early in the case of a race condition.
So once we've exited the loop, we then get the proper (and careful) rq
lock, and check the running/runnable state _safely_. And if it turns
out that our quick-and-dirty and unsafe loop was wrong after all, we
just go back and try it all again.
(The patch also adds a lot of comments, which is the actual bulk of it
all, to make it more obvious why we can do these things without holding
the locks).
Thanks to Miklos for all the testing and tracking it down.
Tested-by: Miklos Szeredi <miklos@szeredi.hu>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-06-18 16:34:40 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Ahh, all good. It wasn't running, and it wasn't
|
|
|
|
* runnable, which means that it will never become
|
|
|
|
* running in the future either. We're all done!
|
|
|
|
*/
|
|
|
|
break;
|
|
|
|
}
|
2008-07-26 02:45:58 +00:00
|
|
|
|
|
|
|
return ncsw;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/***
|
|
|
|
* kick_process - kick a running thread to enter/exit the kernel
|
|
|
|
* @p: the to-be-kicked thread
|
|
|
|
*
|
|
|
|
* Cause a process which is running on another CPU to enter
|
|
|
|
* kernel-mode, without any delay. (to get signals handled.)
|
|
|
|
*
|
|
|
|
* NOTE: this function doesnt have to take the runqueue lock,
|
|
|
|
* because all it wants to ensure is that the remote task enters
|
|
|
|
* the kernel. If the IPI races and the task has been migrated
|
|
|
|
* to another CPU then no harm is done and the purpose has been
|
|
|
|
* achieved as well.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
void kick_process(struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
preempt_disable();
|
|
|
|
cpu = task_cpu(p);
|
|
|
|
if ((cpu != smp_processor_id()) && task_curr(p))
|
|
|
|
smp_send_reschedule(cpu);
|
|
|
|
preempt_enable();
|
|
|
|
}
|
2009-06-13 04:27:00 +00:00
|
|
|
EXPORT_SYMBOL_GPL(kick_process);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
* Return a low guess at the load of a migration-source cpu weighted
|
|
|
|
* according to the scheduling class and "nice" value.
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* We want to under-estimate the load of migration sources, to
|
|
|
|
* balance conservatively.
|
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static unsigned long source_load(int cpu, int type)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
unsigned long total = weighted_cpuload(cpu);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2008-06-27 11:41:33 +00:00
|
|
|
if (type == 0 || !sched_feat(LB_BIAS))
|
2007-07-09 16:51:59 +00:00
|
|
|
return total;
|
2005-11-09 05:38:55 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
return min(rq->cpu_load[type-1], total);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
* Return a high guess at the load of a migration-target cpu weighted
|
|
|
|
* according to the scheduling class and "nice" value.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static unsigned long target_load(int cpu, int type)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
unsigned long total = weighted_cpuload(cpu);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2008-06-27 11:41:33 +00:00
|
|
|
if (type == 0 || !sched_feat(LB_BIAS))
|
2007-07-09 16:51:59 +00:00
|
|
|
return total;
|
2005-11-09 05:38:58 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
return max(rq->cpu_load[type-1], total);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:19 +00:00
|
|
|
/*
|
|
|
|
* find_idlest_group finds and returns the least busy CPU group within the
|
|
|
|
* domain.
|
|
|
|
*/
|
|
|
|
static struct sched_group *
|
|
|
|
find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
|
|
|
|
{
|
|
|
|
struct sched_group *idlest = NULL, *this = NULL, *group = sd->groups;
|
|
|
|
unsigned long min_load = ULONG_MAX, this_load = 0;
|
|
|
|
int load_idx = sd->forkexec_idx;
|
|
|
|
int imbalance = 100 + (sd->imbalance_pct-100)/2;
|
|
|
|
|
|
|
|
do {
|
|
|
|
unsigned long load, avg_load;
|
|
|
|
int local_group;
|
|
|
|
int i;
|
|
|
|
|
2005-09-10 07:26:09 +00:00
|
|
|
/* Skip over this group if it has no CPUs allowed */
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_intersects(sched_group_cpus(group),
|
|
|
|
&p->cpus_allowed))
|
2007-10-15 15:00:14 +00:00
|
|
|
continue;
|
2005-09-10 07:26:09 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
local_group = cpumask_test_cpu(this_cpu,
|
|
|
|
sched_group_cpus(group));
|
2005-06-25 21:57:19 +00:00
|
|
|
|
|
|
|
/* Tally up the load of all CPUs in the group */
|
|
|
|
avg_load = 0;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu(i, sched_group_cpus(group)) {
|
2005-06-25 21:57:19 +00:00
|
|
|
/* Bias balancing toward cpus of our domain */
|
|
|
|
if (local_group)
|
|
|
|
load = source_load(i, load_idx);
|
|
|
|
else
|
|
|
|
load = target_load(i, load_idx);
|
|
|
|
|
|
|
|
avg_load += load;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Adjust by relative CPU power of the group */
|
2007-05-08 07:32:57 +00:00
|
|
|
avg_load = sg_div_cpu_power(group,
|
|
|
|
avg_load * SCHED_LOAD_SCALE);
|
2005-06-25 21:57:19 +00:00
|
|
|
|
|
|
|
if (local_group) {
|
|
|
|
this_load = avg_load;
|
|
|
|
this = group;
|
|
|
|
} else if (avg_load < min_load) {
|
|
|
|
min_load = avg_load;
|
|
|
|
idlest = group;
|
|
|
|
}
|
2007-10-15 15:00:14 +00:00
|
|
|
} while (group = group->next, group != sd->groups);
|
2005-06-25 21:57:19 +00:00
|
|
|
|
|
|
|
if (!idlest || 100*this_load < imbalance*min_load)
|
|
|
|
return NULL;
|
|
|
|
return idlest;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2006-10-03 08:14:10 +00:00
|
|
|
* find_idlest_cpu - find the idlest cpu among the cpus in group.
|
2005-06-25 21:57:19 +00:00
|
|
|
*/
|
2005-09-10 07:26:11 +00:00
|
|
|
static int
|
2008-11-24 16:05:04 +00:00
|
|
|
find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
|
2005-06-25 21:57:19 +00:00
|
|
|
{
|
|
|
|
unsigned long load, min_load = ULONG_MAX;
|
|
|
|
int idlest = -1;
|
|
|
|
int i;
|
|
|
|
|
2005-09-10 07:26:09 +00:00
|
|
|
/* Traverse only the allowed CPUs */
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu_and(i, sched_group_cpus(group), &p->cpus_allowed) {
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
load = weighted_cpuload(i);
|
2005-06-25 21:57:19 +00:00
|
|
|
|
|
|
|
if (load < min_load || (load == min_load && i == this_cpu)) {
|
|
|
|
min_load = load;
|
|
|
|
idlest = i;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return idlest;
|
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:29 +00:00
|
|
|
/*
|
|
|
|
* sched_balance_self: balance the current task (running on cpu) in domains
|
|
|
|
* that have the 'flag' flag set. In practice, this is SD_BALANCE_FORK and
|
|
|
|
* SD_BALANCE_EXEC.
|
|
|
|
*
|
|
|
|
* Balance, ie. select the least loaded group.
|
|
|
|
*
|
|
|
|
* Returns the target CPU number, or the same CPU if no balancing is needed.
|
|
|
|
*
|
|
|
|
* preempt must be disabled.
|
|
|
|
*/
|
|
|
|
static int sched_balance_self(int cpu, int flag)
|
|
|
|
{
|
|
|
|
struct task_struct *t = current;
|
|
|
|
struct sched_domain *tmp, *sd = NULL;
|
2005-06-25 21:57:19 +00:00
|
|
|
|
2006-06-27 09:54:28 +00:00
|
|
|
for_each_domain(cpu, tmp) {
|
2007-07-09 16:52:00 +00:00
|
|
|
/*
|
|
|
|
* If power savings logic is enabled for a domain, stop there.
|
|
|
|
*/
|
2006-06-27 09:54:42 +00:00
|
|
|
if (tmp->flags & SD_POWERSAVINGS_BALANCE)
|
|
|
|
break;
|
2005-06-25 21:57:29 +00:00
|
|
|
if (tmp->flags & flag)
|
|
|
|
sd = tmp;
|
2006-06-27 09:54:28 +00:00
|
|
|
}
|
2005-06-25 21:57:29 +00:00
|
|
|
|
2008-06-27 11:41:25 +00:00
|
|
|
if (sd)
|
|
|
|
update_shares(sd);
|
|
|
|
|
2005-06-25 21:57:29 +00:00
|
|
|
while (sd) {
|
|
|
|
struct sched_group *group;
|
2006-10-03 08:14:08 +00:00
|
|
|
int new_cpu, weight;
|
|
|
|
|
|
|
|
if (!(sd->flags & flag)) {
|
|
|
|
sd = sd->child;
|
|
|
|
continue;
|
|
|
|
}
|
2005-06-25 21:57:29 +00:00
|
|
|
|
|
|
|
group = find_idlest_group(sd, t, cpu);
|
2006-10-03 08:14:08 +00:00
|
|
|
if (!group) {
|
|
|
|
sd = sd->child;
|
|
|
|
continue;
|
|
|
|
}
|
2005-06-25 21:57:29 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
new_cpu = find_idlest_cpu(group, t, cpu);
|
2006-10-03 08:14:08 +00:00
|
|
|
if (new_cpu == -1 || new_cpu == cpu) {
|
|
|
|
/* Now try balancing at a lower domain level of cpu */
|
|
|
|
sd = sd->child;
|
|
|
|
continue;
|
|
|
|
}
|
2005-06-25 21:57:29 +00:00
|
|
|
|
2006-10-03 08:14:08 +00:00
|
|
|
/* Now try balancing at a lower domain level of new_cpu */
|
2005-06-25 21:57:29 +00:00
|
|
|
cpu = new_cpu;
|
2008-11-24 16:05:04 +00:00
|
|
|
weight = cpumask_weight(sched_domain_span(sd));
|
2005-06-25 21:57:29 +00:00
|
|
|
sd = NULL;
|
|
|
|
for_each_domain(cpu, tmp) {
|
2008-11-24 16:05:04 +00:00
|
|
|
if (weight <= cpumask_weight(sched_domain_span(tmp)))
|
2005-06-25 21:57:29 +00:00
|
|
|
break;
|
|
|
|
if (tmp->flags & flag)
|
|
|
|
sd = tmp;
|
|
|
|
}
|
|
|
|
/* while loop will break here if sd == NULL */
|
|
|
|
}
|
|
|
|
|
|
|
|
return cpu;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_SMP */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-12-04 19:12:29 +00:00
|
|
|
/**
|
|
|
|
* task_oncpu_function_call - call a function on the cpu on which a task runs
|
|
|
|
* @p: the task to evaluate
|
|
|
|
* @func: the function to be called
|
|
|
|
* @info: the function call argument
|
|
|
|
*
|
|
|
|
* Calls the function @func when the task is currently running. This might
|
|
|
|
* be on the current CPU, which just calls the function directly
|
|
|
|
*/
|
|
|
|
void task_oncpu_function_call(struct task_struct *p,
|
|
|
|
void (*func) (void *info), void *info)
|
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
preempt_disable();
|
|
|
|
cpu = task_cpu(p);
|
|
|
|
if (task_curr(p))
|
|
|
|
smp_call_function_single(cpu, func, info, 1);
|
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/***
|
|
|
|
* try_to_wake_up - wake up a thread
|
|
|
|
* @p: the to-be-woken-up thread
|
|
|
|
* @state: the mask of task states that can be woken
|
|
|
|
* @sync: do a synchronous wakeup?
|
|
|
|
*
|
|
|
|
* Put it on the run-queue if it's not already there. The "current"
|
|
|
|
* thread is always on the run-queue (except when the actual
|
|
|
|
* re-schedule is in progress), and as such you're allowed to do
|
|
|
|
* the simpler "current->state = TASK_RUNNING" to mark yourself
|
|
|
|
* runnable without the overhead of this.
|
|
|
|
*
|
|
|
|
* returns failure only if the task is already active.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
static int try_to_wake_up(struct task_struct *p, unsigned int state, int sync)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:18 +00:00
|
|
|
int cpu, orig_cpu, this_cpu, success = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
|
|
|
long old_state;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-03-16 19:03:22 +00:00
|
|
|
if (!sched_feat(SYNC_WAKEUPS))
|
|
|
|
sync = 0;
|
|
|
|
|
2008-06-27 11:41:35 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2009-03-09 12:56:21 +00:00
|
|
|
if (sched_feat(LB_WAKEUP_UPDATE) && !root_task_group_empty()) {
|
2008-06-27 11:41:35 +00:00
|
|
|
struct sched_domain *sd;
|
|
|
|
|
|
|
|
this_cpu = raw_smp_processor_id();
|
|
|
|
cpu = task_cpu(p);
|
|
|
|
|
|
|
|
for_each_domain(this_cpu, sd) {
|
2008-11-24 16:05:04 +00:00
|
|
|
if (cpumask_test_cpu(cpu, sched_domain_span(sd))) {
|
2008-06-27 11:41:35 +00:00
|
|
|
update_shares(sd);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-02-24 02:05:03 +00:00
|
|
|
smp_wmb();
|
2005-04-16 22:20:36 +00:00
|
|
|
rq = task_rq_lock(p, &flags);
|
2008-12-16 07:45:30 +00:00
|
|
|
update_rq_clock(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
old_state = p->state;
|
|
|
|
if (!(old_state & state))
|
|
|
|
goto out;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (p->se.on_rq)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out_running;
|
|
|
|
|
|
|
|
cpu = task_cpu(p);
|
2007-10-15 15:00:18 +00:00
|
|
|
orig_cpu = cpu;
|
2005-04-16 22:20:36 +00:00
|
|
|
this_cpu = smp_processor_id();
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
if (unlikely(task_running(rq, p)))
|
|
|
|
goto out_activate;
|
|
|
|
|
2008-01-25 20:08:21 +00:00
|
|
|
cpu = p->sched_class->select_task_rq(p, sync);
|
|
|
|
if (cpu != orig_cpu) {
|
|
|
|
set_task_cpu(p, cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
/* might preempt at this point */
|
|
|
|
rq = task_rq_lock(p, &flags);
|
|
|
|
old_state = p->state;
|
|
|
|
if (!(old_state & state))
|
|
|
|
goto out;
|
2007-07-09 16:51:59 +00:00
|
|
|
if (p->se.on_rq)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out_running;
|
|
|
|
|
|
|
|
this_cpu = smp_processor_id();
|
|
|
|
cpu = task_cpu(p);
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:09 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
|
|
|
schedstat_inc(rq, ttwu_count);
|
|
|
|
if (cpu == this_cpu)
|
|
|
|
schedstat_inc(rq, ttwu_local);
|
|
|
|
else {
|
|
|
|
struct sched_domain *sd;
|
|
|
|
for_each_domain(this_cpu, sd) {
|
2008-11-24 16:05:04 +00:00
|
|
|
if (cpumask_test_cpu(cpu, sched_domain_span(sd))) {
|
2008-01-25 20:08:09 +00:00
|
|
|
schedstat_inc(sd, ttwu_wake_remote);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SCHEDSTATS */
|
2008-01-25 20:08:09 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
out_activate:
|
|
|
|
#endif /* CONFIG_SMP */
|
2007-10-15 15:00:18 +00:00
|
|
|
schedstat_inc(p, se.nr_wakeups);
|
|
|
|
if (sync)
|
|
|
|
schedstat_inc(p, se.nr_wakeups_sync);
|
|
|
|
if (orig_cpu != cpu)
|
|
|
|
schedstat_inc(p, se.nr_wakeups_migrate);
|
|
|
|
if (cpu == this_cpu)
|
|
|
|
schedstat_inc(p, se.nr_wakeups_local);
|
|
|
|
else
|
|
|
|
schedstat_inc(p, se.nr_wakeups_remote);
|
2007-07-09 16:51:59 +00:00
|
|
|
activate_task(rq, p, 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
success = 1;
|
|
|
|
|
2009-01-14 11:39:18 +00:00
|
|
|
/*
|
|
|
|
* Only attribute actual wakeups done by this task.
|
|
|
|
*/
|
|
|
|
if (!in_interrupt()) {
|
|
|
|
struct sched_entity *se = ¤t->se;
|
|
|
|
u64 sample = se->sum_exec_runtime;
|
|
|
|
|
|
|
|
if (se->last_wakeup)
|
|
|
|
sample -= se->last_wakeup;
|
|
|
|
else
|
|
|
|
sample -= se->start_runtime;
|
|
|
|
update_avg(&se->avg_wakeup, sample);
|
|
|
|
|
|
|
|
se->last_wakeup = se->sum_exec_runtime;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
out_running:
|
2008-12-16 07:07:03 +00:00
|
|
|
trace_sched_wakeup(rq, p, success);
|
2008-09-20 21:38:02 +00:00
|
|
|
check_preempt_curr(rq, p, sync);
|
2008-03-19 00:42:00 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
p->state = TASK_RUNNING;
|
2008-01-25 20:08:22 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
if (p->sched_class->task_wake_up)
|
|
|
|
p->sched_class->task_wake_up(rq, p);
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
out:
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
|
|
|
|
return success;
|
|
|
|
}
|
|
|
|
|
2009-04-28 14:01:38 +00:00
|
|
|
/**
|
|
|
|
* wake_up_process - Wake up a specific process
|
|
|
|
* @p: The process to be woken up.
|
|
|
|
*
|
|
|
|
* Attempt to wake up the nominated process and move it to the set of runnable
|
|
|
|
* processes. Returns 1 if the process was woken up, 0 if it was already
|
|
|
|
* running.
|
|
|
|
*
|
|
|
|
* It may be assumed that this function implies a write memory barrier before
|
|
|
|
* changing the task state if and only if any tasks are woken up.
|
|
|
|
*/
|
2008-02-08 12:19:53 +00:00
|
|
|
int wake_up_process(struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-12-06 16:07:07 +00:00
|
|
|
return try_to_wake_up(p, TASK_ALL, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(wake_up_process);
|
|
|
|
|
2008-02-08 12:19:53 +00:00
|
|
|
int wake_up_state(struct task_struct *p, unsigned int state)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return try_to_wake_up(p, state, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Perform scheduler related setup for a newly forked process p.
|
|
|
|
* p is forked by current.
|
2007-07-09 16:51:59 +00:00
|
|
|
*
|
|
|
|
* __sched_fork() is basic setup used by init_idle() too:
|
|
|
|
*/
|
|
|
|
static void __sched_fork(struct task_struct *p)
|
|
|
|
{
|
|
|
|
p->se.exec_start = 0;
|
|
|
|
p->se.sum_exec_runtime = 0;
|
2007-08-28 10:53:24 +00:00
|
|
|
p->se.prev_sum_exec_runtime = 0;
|
2008-12-14 11:34:15 +00:00
|
|
|
p->se.nr_migrations = 0;
|
2008-03-19 00:42:00 +00:00
|
|
|
p->se.last_wakeup = 0;
|
|
|
|
p->se.avg_overlap = 0;
|
2009-01-14 11:39:18 +00:00
|
|
|
p->se.start_runtime = 0;
|
|
|
|
p->se.avg_wakeup = sysctl_sched_wakeup_granularity;
|
2007-08-02 15:41:40 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
|
|
|
p->se.wait_start = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.sum_sleep_runtime = 0;
|
|
|
|
p->se.sleep_start = 0;
|
|
|
|
p->se.block_start = 0;
|
|
|
|
p->se.sleep_max = 0;
|
|
|
|
p->se.block_max = 0;
|
|
|
|
p->se.exec_max = 0;
|
2007-10-15 15:00:02 +00:00
|
|
|
p->se.slice_max = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.wait_max = 0;
|
2007-08-02 15:41:40 +00:00
|
|
|
#endif
|
2005-06-25 21:57:29 +00:00
|
|
|
|
2008-01-25 20:08:27 +00:00
|
|
|
INIT_LIST_HEAD(&p->rt.run_list);
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.on_rq = 0;
|
2008-04-19 17:45:00 +00:00
|
|
|
INIT_LIST_HEAD(&p->se.group_node);
|
2005-06-25 21:57:29 +00:00
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_NOTIFIERS
|
|
|
|
INIT_HLIST_HEAD(&p->preempt_notifiers);
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* We mark the process as running here, but have not actually
|
|
|
|
* inserted it onto the runqueue yet. This guarantees that
|
|
|
|
* nobody will actually run it, and a signal or other external
|
|
|
|
* event cannot wake it up and insert it on the runqueue either.
|
|
|
|
*/
|
|
|
|
p->state = TASK_RUNNING;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* fork()/clone()-time setup:
|
|
|
|
*/
|
|
|
|
void sched_fork(struct task_struct *p, int clone_flags)
|
|
|
|
{
|
|
|
|
int cpu = get_cpu();
|
|
|
|
|
|
|
|
__sched_fork(p);
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
cpu = sched_balance_self(cpu, SD_BALANCE_FORK);
|
|
|
|
#endif
|
2007-10-15 15:00:11 +00:00
|
|
|
set_task_cpu(p, cpu);
|
2006-06-27 09:54:51 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Make sure we do not leak PI boosting priority to the child:
|
|
|
|
*/
|
|
|
|
p->prio = current->normal_prio;
|
2007-10-15 15:00:11 +00:00
|
|
|
if (!rt_prio(p->prio))
|
|
|
|
p->sched_class = &fair_sched_class;
|
2006-06-27 09:54:51 +00:00
|
|
|
|
2006-07-14 07:24:38 +00:00
|
|
|
#if defined(CONFIG_SCHEDSTATS) || defined(CONFIG_TASK_DELAY_ACCT)
|
2007-07-09 16:51:59 +00:00
|
|
|
if (likely(sched_info_on()))
|
2006-07-14 07:24:38 +00:00
|
|
|
memset(&p->sched_info, 0, sizeof(p->sched_info));
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
[PATCH] sched: revert "filter affine wakeups"
Revert commit d7102e95b7b9c00277562c29aad421d2d521c5f6:
[PATCH] sched: filter affine wakeups
Apparently caused more than 10% performance regression for aim7 benchmark.
The setup in use is 16-cpu HP rx8620, 64Gb of memory and 12 MSA1000s with 144
disks. Each disk is 72Gb with a single ext3 filesystem (courtesy of HP, who
supplied benchmark results).
The problem is, for aim7, the wake-up pattern is random, but it still needs
load balancing action in the wake-up path to achieve best performance. With
the above commit, lack of load balancing hurts that workload.
However, for workloads like database transaction processing, the requirement
is exactly opposite. In the wake up path, best performance is achieved with
absolutely zero load balancing. We simply wake up the process on the CPU that
it was previously run. Worst performance is obtained when we do load
balancing at wake up.
There isn't an easy way to auto detect the workload characteristics. Ingo's
earlier patch that detects idle CPU and decide whether to load balance or not
doesn't perform with aim7 either since all CPUs are busy (it causes even
bigger perf. regression).
Revert commit d7102e95b7b9c00277562c29aad421d2d521c5f6, which causes more
than 10% performance regression with aim7.
Signed-off-by: Ken Chen <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-02-14 21:53:10 +00:00
|
|
|
#if defined(CONFIG_SMP) && defined(__ARCH_WANT_UNLOCKED_CTXSW)
|
2005-06-25 21:57:23 +00:00
|
|
|
p->oncpu = 0;
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_PREEMPT
|
2005-06-25 21:57:23 +00:00
|
|
|
/* Want to start with kernel preemption disabled. */
|
2005-11-14 00:06:55 +00:00
|
|
|
task_thread_info(p)->preempt_count = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
sched: create "pushable_tasks" list to limit pushing to one attempt
The RT scheduler employs a "push/pull" design to actively balance tasks
within the system (on a per disjoint cpuset basis). When a task is
awoken, it is immediately determined if there are any lower priority
cpus which should be preempted. This is opposed to the way normal
SCHED_OTHER tasks behave, which will wait for a periodic rebalancing
operation to occur before spreading out load.
When a particular RQ has more than 1 active RT task, it is said to
be in an "overloaded" state. Once this occurs, the system enters
the active balancing mode, where it will try to push the task away,
or persuade a different cpu to pull it over. The system will stay
in this state until the system falls back below the <= 1 queued RT
task per RQ.
However, the current implementation suffers from a limitation in the
push logic. Once overloaded, all tasks (other than current) on the
RQ are analyzed on every push operation, even if it was previously
unpushable (due to affinity, etc). Whats more, the operation stops
at the first task that is unpushable and will not look at items
lower in the queue. This causes two problems:
1) We can have the same tasks analyzed over and over again during each
push, which extends out the fast path in the scheduler for no
gain. Consider a RQ that has dozens of tasks that are bound to a
core. Each one of those tasks will be encountered and skipped
for each push operation while they are queued.
2) There may be lower-priority tasks under the unpushable task that
could have been successfully pushed, but will never be considered
until either the unpushable task is cleared, or a pull operation
succeeds. The net result is a potential latency source for mid
priority tasks.
This patch aims to rectify these two conditions by introducing a new
priority sorted list: "pushable_tasks". A task is added to the list
each time a task is activated or preempted. It is removed from the
list any time it is deactivated, made current, or fails to push.
This works because a task only needs to be attempted to push once.
After an initial failure to push, the other cpus will eventually try to
pull the task when the conditions are proper. This also solves the
problem that we don't completely analyze all tasks due to encountering
an unpushable tasks. Now every task will have a push attempted (when
appropriate).
This reduces latency both by shorting the critical section of the
rq->lock for certain workloads, and by making sure the algorithm
considers all eligible tasks in the system.
[ rostedt: added a couple more BUG_ONs ]
Signed-off-by: Gregory Haskins <ghaskins@novell.com>
Acked-by: Steven Rostedt <srostedt@redhat.com>
2008-12-29 14:39:53 +00:00
|
|
|
plist_node_init(&p->pushable_tasks, MAX_PRIO);
|
|
|
|
|
2005-06-25 21:57:29 +00:00
|
|
|
put_cpu();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* wake_up_new_task - wake up a newly created task for the first time.
|
|
|
|
*
|
|
|
|
* This function will do some initial scheduler statistics housekeeping
|
|
|
|
* that must be done for every newly created context, then puts the task
|
|
|
|
* on the runqueue and wakes it.
|
|
|
|
*/
|
2008-02-08 12:19:53 +00:00
|
|
|
void wake_up_new_task(struct task_struct *p, unsigned long clone_flags)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
2007-07-09 16:51:59 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2005-06-25 21:57:19 +00:00
|
|
|
BUG_ON(p->state != TASK_RUNNING);
|
2007-08-09 09:16:47 +00:00
|
|
|
update_rq_clock(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
p->prio = effective_prio(p);
|
|
|
|
|
2007-10-17 14:55:11 +00:00
|
|
|
if (!p->sched_class->task_new || !current->se.on_rq) {
|
2007-07-09 16:51:59 +00:00
|
|
|
activate_task(rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else {
|
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* Let the scheduling class do new task startup
|
|
|
|
* management (if any):
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-08-09 09:16:49 +00:00
|
|
|
p->sched_class->task_new(rq, p);
|
2008-06-27 11:41:14 +00:00
|
|
|
inc_nr_running(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-12-19 00:09:51 +00:00
|
|
|
trace_sched_wakeup_new(rq, p, 1);
|
2008-09-20 21:38:02 +00:00
|
|
|
check_preempt_curr(rq, p, 0);
|
2008-01-25 20:08:22 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
if (p->sched_class->task_wake_up)
|
|
|
|
p->sched_class->task_wake_up(rq, p);
|
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_NOTIFIERS
|
|
|
|
|
|
|
|
/**
|
2009-03-16 19:58:09 +00:00
|
|
|
* preempt_notifier_register - tell me when current is being preempted & rescheduled
|
2007-07-31 07:37:50 +00:00
|
|
|
* @notifier: notifier struct to register
|
2007-07-26 11:40:43 +00:00
|
|
|
*/
|
|
|
|
void preempt_notifier_register(struct preempt_notifier *notifier)
|
|
|
|
{
|
|
|
|
hlist_add_head(¬ifier->link, ¤t->preempt_notifiers);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(preempt_notifier_register);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* preempt_notifier_unregister - no longer interested in preemption notifications
|
2007-07-31 07:37:50 +00:00
|
|
|
* @notifier: notifier struct to unregister
|
2007-07-26 11:40:43 +00:00
|
|
|
*
|
|
|
|
* This is safe to call from within a preemption notifier.
|
|
|
|
*/
|
|
|
|
void preempt_notifier_unregister(struct preempt_notifier *notifier)
|
|
|
|
{
|
|
|
|
hlist_del(¬ifier->link);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(preempt_notifier_unregister);
|
|
|
|
|
|
|
|
static void fire_sched_in_preempt_notifiers(struct task_struct *curr)
|
|
|
|
{
|
|
|
|
struct preempt_notifier *notifier;
|
|
|
|
struct hlist_node *node;
|
|
|
|
|
|
|
|
hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
|
|
|
|
notifier->ops->sched_in(notifier, raw_smp_processor_id());
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
fire_sched_out_preempt_notifiers(struct task_struct *curr,
|
|
|
|
struct task_struct *next)
|
|
|
|
{
|
|
|
|
struct preempt_notifier *notifier;
|
|
|
|
struct hlist_node *node;
|
|
|
|
|
|
|
|
hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
|
|
|
|
notifier->ops->sched_out(notifier, next);
|
|
|
|
}
|
|
|
|
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_PREEMPT_NOTIFIERS */
|
2007-07-26 11:40:43 +00:00
|
|
|
|
|
|
|
static void fire_sched_in_preempt_notifiers(struct task_struct *curr)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
fire_sched_out_preempt_notifiers(struct task_struct *curr,
|
|
|
|
struct task_struct *next)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_PREEMPT_NOTIFIERS */
|
2007-07-26 11:40:43 +00:00
|
|
|
|
2005-06-25 21:57:23 +00:00
|
|
|
/**
|
|
|
|
* prepare_task_switch - prepare to switch tasks
|
|
|
|
* @rq: the runqueue preparing to switch
|
2007-07-31 07:37:50 +00:00
|
|
|
* @prev: the current task that is being switched out
|
2005-06-25 21:57:23 +00:00
|
|
|
* @next: the task we are going to switch to.
|
|
|
|
*
|
|
|
|
* This is called with the rq lock held and interrupts off. It must
|
|
|
|
* be paired with a subsequent finish_task_switch after the context
|
|
|
|
* switch.
|
|
|
|
*
|
|
|
|
* prepare_task_switch sets up locking and calls architecture specific
|
|
|
|
* hooks.
|
|
|
|
*/
|
2007-07-26 11:40:43 +00:00
|
|
|
static inline void
|
|
|
|
prepare_task_switch(struct rq *rq, struct task_struct *prev,
|
|
|
|
struct task_struct *next)
|
2005-06-25 21:57:23 +00:00
|
|
|
{
|
2007-07-26 11:40:43 +00:00
|
|
|
fire_sched_out_preempt_notifiers(prev, next);
|
2005-06-25 21:57:23 +00:00
|
|
|
prepare_lock_switch(rq, next);
|
|
|
|
prepare_arch_switch(next);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/**
|
|
|
|
* finish_task_switch - clean up after a task-switch
|
2005-09-07 05:15:17 +00:00
|
|
|
* @rq: runqueue associated with task-switch
|
2005-04-16 22:20:36 +00:00
|
|
|
* @prev: the thread we just switched away from.
|
|
|
|
*
|
2005-06-25 21:57:23 +00:00
|
|
|
* finish_task_switch must be called after the context switch, paired
|
|
|
|
* with a prepare_task_switch call before the context switch.
|
|
|
|
* finish_task_switch will reconcile locking set up by prepare_task_switch,
|
|
|
|
* and do any other architecture-specific cleanup actions.
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* Note that we may have delayed dropping an mm in context_switch(). If
|
2007-12-05 14:46:09 +00:00
|
|
|
* so, we finish that here outside of the runqueue lock. (Doing it
|
2005-04-16 22:20:36 +00:00
|
|
|
* with the lock held can cause deadlocks; see schedule() for
|
|
|
|
* details.)
|
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static void finish_task_switch(struct rq *rq, struct task_struct *prev)
|
2005-04-16 22:20:36 +00:00
|
|
|
__releases(rq->lock)
|
|
|
|
{
|
|
|
|
struct mm_struct *mm = rq->prev_mm;
|
2006-09-29 09:01:10 +00:00
|
|
|
long prev_state;
|
2008-12-29 14:39:52 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
int post_schedule = 0;
|
|
|
|
|
|
|
|
if (current->sched_class->needs_post_schedule)
|
|
|
|
post_schedule = current->sched_class->needs_post_schedule(rq);
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq->prev_mm = NULL;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A task struct has one reference for the use as "current".
|
2006-09-29 09:01:11 +00:00
|
|
|
* If a task dies, then it sets TASK_DEAD in tsk->state and calls
|
2006-09-29 09:01:10 +00:00
|
|
|
* schedule one last time. The schedule call will never return, and
|
|
|
|
* the scheduled task must drop that reference.
|
2006-09-29 09:01:11 +00:00
|
|
|
* The test for TASK_DEAD must occur while the runqueue locks are
|
2005-04-16 22:20:36 +00:00
|
|
|
* still held, otherwise prev could be scheduled on another cpu, die
|
|
|
|
* there before we look at prev->state, and then the reference would
|
|
|
|
* be dropped twice.
|
|
|
|
* Manfred Spraul <manfred@colorfullife.com>
|
|
|
|
*/
|
2006-09-29 09:01:10 +00:00
|
|
|
prev_state = prev->state;
|
2005-06-25 21:57:23 +00:00
|
|
|
finish_arch_switch(prev);
|
2008-12-04 19:12:29 +00:00
|
|
|
perf_counter_task_sched_in(current, cpu_of(rq));
|
2005-06-25 21:57:23 +00:00
|
|
|
finish_lock_switch(rq, prev);
|
2008-01-25 20:08:22 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-12-29 14:39:52 +00:00
|
|
|
if (post_schedule)
|
2008-01-25 20:08:22 +00:00
|
|
|
current->sched_class->post_schedule(rq);
|
|
|
|
#endif
|
2008-01-25 20:08:05 +00:00
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
fire_sched_in_preempt_notifiers(current);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (mm)
|
|
|
|
mmdrop(mm);
|
2006-09-29 09:01:11 +00:00
|
|
|
if (unlikely(prev_state == TASK_DEAD)) {
|
2006-03-26 09:38:20 +00:00
|
|
|
/*
|
|
|
|
* Remove function-return probe instances associated with this
|
|
|
|
* task and put them back on the free list.
|
2007-07-09 16:52:00 +00:00
|
|
|
*/
|
2006-03-26 09:38:20 +00:00
|
|
|
kprobe_flush_task(prev);
|
2005-04-16 22:20:36 +00:00
|
|
|
put_task_struct(prev);
|
2006-03-26 09:38:20 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* schedule_tail - first thing a freshly forked thread must call.
|
|
|
|
* @prev: the thread we just switched away from.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
asmlinkage void schedule_tail(struct task_struct *prev)
|
2005-04-16 22:20:36 +00:00
|
|
|
__releases(rq->lock)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = this_rq();
|
|
|
|
|
2005-06-25 21:57:23 +00:00
|
|
|
finish_task_switch(rq, prev);
|
|
|
|
#ifdef __ARCH_WANT_UNLOCKED_CTXSW
|
|
|
|
/* In this case, finish_task_switch does not reenable preemption */
|
|
|
|
preempt_enable();
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
if (current->set_child_tid)
|
2007-10-19 06:40:14 +00:00
|
|
|
put_user(task_pid_vnr(current), current->set_child_tid);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* context_switch - switch to the new MM and the new
|
|
|
|
* thread's register state.
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static inline void
|
2006-07-03 07:25:42 +00:00
|
|
|
context_switch(struct rq *rq, struct task_struct *prev,
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *next)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
struct mm_struct *mm, *oldmm;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
prepare_task_switch(rq, prev, next);
|
tracing, sched: LTTng instrumentation - scheduler
Instrument the scheduler activity (sched_switch, migration, wakeups,
wait for a task, signal delivery) and process/thread
creation/destruction (fork, exit, kthread stop). Actually, kthread
creation is not instrumented in this patch because it is architecture
dependent. It allows to connect tracers such as ftrace which detects
scheduling latencies, good/bad scheduler decisions. Tools like LTTng can
export this scheduler information along with instrumentation of the rest
of the kernel activity to perform post-mortem analysis on the scheduler
activity.
About the performance impact of tracepoints (which is comparable to
markers), even without immediate values optimizations, tests done by
Hideo Aoki on ia64 show no regression. His test case was using hackbench
on a kernel where scheduler instrumentation (about 5 events in code
scheduler code) was added. See the "Tracepoints" patch header for
performance result detail.
Changelog :
- Change instrumentation location and parameter to match ftrace
instrumentation, previously done with kernel markers.
[ mingo@elte.hu: conflict resolutions ]
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
Acked-by: 'Peter Zijlstra' <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-07-18 16:16:17 +00:00
|
|
|
trace_sched_switch(rq, prev, next);
|
2007-07-09 16:51:59 +00:00
|
|
|
mm = next->mm;
|
|
|
|
oldmm = prev->active_mm;
|
2007-02-13 12:26:21 +00:00
|
|
|
/*
|
|
|
|
* For paravirt, this is coupled with an exit in switch_to to
|
|
|
|
* combine the page table reload and the switch backend into
|
|
|
|
* one hypercall.
|
|
|
|
*/
|
2009-02-18 19:18:57 +00:00
|
|
|
arch_start_context_switch(prev);
|
2007-02-13 12:26:21 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (unlikely(!mm)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
next->active_mm = oldmm;
|
|
|
|
atomic_inc(&oldmm->mm_count);
|
|
|
|
enter_lazy_tlb(oldmm, next);
|
|
|
|
} else
|
|
|
|
switch_mm(oldmm, mm, next);
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (unlikely(!prev->mm)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
prev->active_mm = NULL;
|
|
|
|
rq->prev_mm = oldmm;
|
|
|
|
}
|
2006-07-14 07:24:27 +00:00
|
|
|
/*
|
|
|
|
* Since the runqueue lock will be released by the next
|
|
|
|
* task (which is an invalid locking op but in the case
|
|
|
|
* of the scheduler it's an obvious special-case), so we
|
|
|
|
* do an early lockdep release here:
|
|
|
|
*/
|
|
|
|
#ifndef __ARCH_WANT_UNLOCKED_CTXSW
|
2006-07-03 07:24:54 +00:00
|
|
|
spin_release(&rq->lock.dep_map, 1, _THIS_IP_);
|
2006-07-14 07:24:27 +00:00
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Here we just switch the register state and the stack. */
|
|
|
|
switch_to(prev, next, prev);
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
barrier();
|
|
|
|
/*
|
|
|
|
* this_rq must be evaluated again because prev may have moved
|
|
|
|
* CPUs since it called schedule(), thus the 'rq' on its stack
|
|
|
|
* frame will be invalid.
|
|
|
|
*/
|
|
|
|
finish_task_switch(this_rq(), prev);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* nr_running, nr_uninterruptible and nr_context_switches:
|
|
|
|
*
|
|
|
|
* externally visible scheduler statistics: current number of runnable
|
|
|
|
* threads, current number of uninterruptible-sleeping threads, total
|
|
|
|
* number of context switches performed since bootup.
|
|
|
|
*/
|
|
|
|
unsigned long nr_running(void)
|
|
|
|
{
|
|
|
|
unsigned long i, sum = 0;
|
|
|
|
|
|
|
|
for_each_online_cpu(i)
|
|
|
|
sum += cpu_rq(i)->nr_running;
|
|
|
|
|
|
|
|
return sum;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned long nr_uninterruptible(void)
|
|
|
|
{
|
|
|
|
unsigned long i, sum = 0;
|
|
|
|
|
2006-03-28 09:56:37 +00:00
|
|
|
for_each_possible_cpu(i)
|
2005-04-16 22:20:36 +00:00
|
|
|
sum += cpu_rq(i)->nr_uninterruptible;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Since we read the counters lockless, it might be slightly
|
|
|
|
* inaccurate. Do not allow it to go below zero though:
|
|
|
|
*/
|
|
|
|
if (unlikely((long)sum < 0))
|
|
|
|
sum = 0;
|
|
|
|
|
|
|
|
return sum;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned long long nr_context_switches(void)
|
|
|
|
{
|
2006-06-27 09:54:31 +00:00
|
|
|
int i;
|
|
|
|
unsigned long long sum = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-03-28 09:56:37 +00:00
|
|
|
for_each_possible_cpu(i)
|
2005-04-16 22:20:36 +00:00
|
|
|
sum += cpu_rq(i)->nr_switches;
|
|
|
|
|
|
|
|
return sum;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned long nr_iowait(void)
|
|
|
|
{
|
|
|
|
unsigned long i, sum = 0;
|
|
|
|
|
2006-03-28 09:56:37 +00:00
|
|
|
for_each_possible_cpu(i)
|
2005-04-16 22:20:36 +00:00
|
|
|
sum += atomic_read(&cpu_rq(i)->nr_iowait);
|
|
|
|
|
|
|
|
return sum;
|
|
|
|
}
|
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
/* Variables and functions for calc_load */
|
|
|
|
static atomic_long_t calc_load_tasks;
|
|
|
|
static unsigned long calc_load_update;
|
|
|
|
unsigned long avenrun[3];
|
|
|
|
EXPORT_SYMBOL(avenrun);
|
|
|
|
|
2009-05-02 18:08:52 +00:00
|
|
|
/**
|
|
|
|
* get_avenrun - get the load average array
|
|
|
|
* @loads: pointer to dest load array
|
|
|
|
* @offset: offset to add
|
|
|
|
* @shift: shift count to shift the result left
|
|
|
|
*
|
|
|
|
* These values are estimates at best, so no need for locking.
|
|
|
|
*/
|
|
|
|
void get_avenrun(unsigned long *loads, unsigned long offset, int shift)
|
|
|
|
{
|
|
|
|
loads[0] = (avenrun[0] + offset) << shift;
|
|
|
|
loads[1] = (avenrun[1] + offset) << shift;
|
|
|
|
loads[2] = (avenrun[2] + offset) << shift;
|
|
|
|
}
|
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
static unsigned long
|
|
|
|
calc_load(unsigned long load, unsigned long exp, unsigned long active)
|
2006-03-31 10:31:21 +00:00
|
|
|
{
|
2009-04-11 08:43:41 +00:00
|
|
|
load *= exp;
|
|
|
|
load += active * (FIXED_1 - exp);
|
|
|
|
return load >> FSHIFT;
|
|
|
|
}
|
2006-03-31 10:31:21 +00:00
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
/*
|
|
|
|
* calc_load - update the avenrun load estimates 10 ticks after the
|
|
|
|
* CPUs have updated calc_load_tasks.
|
|
|
|
*/
|
|
|
|
void calc_global_load(void)
|
|
|
|
{
|
|
|
|
unsigned long upd = calc_load_update + 10;
|
|
|
|
long active;
|
|
|
|
|
|
|
|
if (time_before(jiffies, upd))
|
|
|
|
return;
|
2006-03-31 10:31:21 +00:00
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
active = atomic_long_read(&calc_load_tasks);
|
|
|
|
active = active > 0 ? active * FIXED_1 : 0;
|
2006-03-31 10:31:21 +00:00
|
|
|
|
2009-04-11 08:43:41 +00:00
|
|
|
avenrun[0] = calc_load(avenrun[0], EXP_1, active);
|
|
|
|
avenrun[1] = calc_load(avenrun[1], EXP_5, active);
|
|
|
|
avenrun[2] = calc_load(avenrun[2], EXP_15, active);
|
|
|
|
|
|
|
|
calc_load_update += LOAD_FREQ;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Either called from update_cpu_load() or from a cpu going idle
|
|
|
|
*/
|
|
|
|
static void calc_load_account_active(struct rq *this_rq)
|
|
|
|
{
|
|
|
|
long nr_active, delta;
|
|
|
|
|
|
|
|
nr_active = this_rq->nr_running;
|
|
|
|
nr_active += (long) this_rq->nr_uninterruptible;
|
|
|
|
|
|
|
|
if (nr_active != this_rq->calc_load_active) {
|
|
|
|
delta = nr_active - this_rq->calc_load_active;
|
|
|
|
this_rq->calc_load_active = nr_active;
|
|
|
|
atomic_long_add(delta, &calc_load_tasks);
|
|
|
|
}
|
2006-03-31 10:31:21 +00:00
|
|
|
}
|
|
|
|
|
perf_counters: make software counters work as per-cpu counters
Impact: kernel crash fix
Yanmin Zhang reported that using a PERF_COUNT_TASK_CLOCK software
counter as a per-cpu counter would reliably crash the system, because
it calls __task_delta_exec with a null pointer. The page fault,
context switch and cpu migration counters also won't function
correctly as per-cpu counters since they reference the current task.
This fixes the problem by redirecting the task_clock counter to the
cpu_clock counter when used as a per-cpu counter, and by implementing
per-cpu page fault, context switch and cpu migration counters.
Along the way, this:
- Initializes counter->ctx earlier, in perf_counter_alloc, so that
sw_perf_counter_init can use it
- Adds code to kernel/sched.c to count task migrations into each
cpu, in rq->nr_migrations_in
- Exports the per-cpu context switch and task migration counts
via new functions added to kernel/sched.c
- Makes sure that if sw_perf_counter_init fails, we don't try to
initialize the counter as a hardware counter. Since the user has
passed a negative, non-raw event type, they clearly don't intend
for it to be interpreted as a hardware event.
Reported-by: "Zhang Yanmin" <yanmin_zhang@linux.intel.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-09 11:42:47 +00:00
|
|
|
/*
|
|
|
|
* Externally visible per-cpu scheduler statistics:
|
|
|
|
* cpu_nr_migrations(cpu) - number of migrations into that cpu
|
|
|
|
*/
|
|
|
|
u64 cpu_nr_migrations(int cpu)
|
|
|
|
{
|
|
|
|
return cpu_rq(cpu)->nr_migrations_in;
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* Update rq->cpu_load[] statistics. This function is usually called every
|
|
|
|
* scheduler tick (TICK_NSEC).
|
2006-07-03 07:25:40 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static void update_cpu_load(struct rq *this_rq)
|
2006-07-03 07:25:40 +00:00
|
|
|
{
|
2007-10-15 15:00:06 +00:00
|
|
|
unsigned long this_load = this_rq->load.weight;
|
2007-07-09 16:51:59 +00:00
|
|
|
int i, scale;
|
|
|
|
|
|
|
|
this_rq->nr_load_updates++;
|
|
|
|
|
|
|
|
/* Update our load: */
|
|
|
|
for (i = 0, scale = 1; i < CPU_LOAD_IDX_MAX; i++, scale += scale) {
|
|
|
|
unsigned long old_load, new_load;
|
|
|
|
|
|
|
|
/* scale is effectively 1 << i now, and >> i divides by scale */
|
|
|
|
|
|
|
|
old_load = this_rq->cpu_load[i];
|
|
|
|
new_load = this_load;
|
2007-10-15 15:00:03 +00:00
|
|
|
/*
|
|
|
|
* Round up the averaging division if load is increasing. This
|
|
|
|
* prevents us from getting stuck on 9 if the load is 10, for
|
|
|
|
* example.
|
|
|
|
*/
|
|
|
|
if (new_load > old_load)
|
|
|
|
new_load += scale-1;
|
2007-07-09 16:51:59 +00:00
|
|
|
this_rq->cpu_load[i] = (old_load*(scale-1) + new_load) >> i;
|
|
|
|
}
|
2009-04-11 08:43:41 +00:00
|
|
|
|
|
|
|
if (time_after_eq(jiffies, this_rq->calc_load_update)) {
|
|
|
|
this_rq->calc_load_update += LOAD_FREQ;
|
|
|
|
calc_load_account_active(this_rq);
|
|
|
|
}
|
2006-07-03 07:25:40 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* double_rq_lock - safely lock two runqueues
|
|
|
|
*
|
|
|
|
* Note this does not disable interrupts like task_rq_lock,
|
|
|
|
* you need to do so manually before calling.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static void double_rq_lock(struct rq *rq1, struct rq *rq2)
|
2005-04-16 22:20:36 +00:00
|
|
|
__acquires(rq1->lock)
|
|
|
|
__acquires(rq2->lock)
|
|
|
|
{
|
2006-12-10 10:20:11 +00:00
|
|
|
BUG_ON(!irqs_disabled());
|
2005-04-16 22:20:36 +00:00
|
|
|
if (rq1 == rq2) {
|
|
|
|
spin_lock(&rq1->lock);
|
|
|
|
__acquire(rq2->lock); /* Fake it out ;) */
|
|
|
|
} else {
|
2006-06-27 09:54:28 +00:00
|
|
|
if (rq1 < rq2) {
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_lock(&rq1->lock);
|
2008-07-30 11:26:57 +00:00
|
|
|
spin_lock_nested(&rq2->lock, SINGLE_DEPTH_NESTING);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else {
|
|
|
|
spin_lock(&rq2->lock);
|
2008-07-30 11:26:57 +00:00
|
|
|
spin_lock_nested(&rq1->lock, SINGLE_DEPTH_NESTING);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
2007-08-09 09:16:51 +00:00
|
|
|
update_rq_clock(rq1);
|
|
|
|
update_rq_clock(rq2);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* double_rq_unlock - safely unlock two runqueues
|
|
|
|
*
|
|
|
|
* Note this does not restore interrupts like task_rq_unlock,
|
|
|
|
* you need to do so manually after calling.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static void double_rq_unlock(struct rq *rq1, struct rq *rq2)
|
2005-04-16 22:20:36 +00:00
|
|
|
__releases(rq1->lock)
|
|
|
|
__releases(rq2->lock)
|
|
|
|
{
|
|
|
|
spin_unlock(&rq1->lock);
|
|
|
|
if (rq1 != rq2)
|
|
|
|
spin_unlock(&rq2->lock);
|
|
|
|
else
|
|
|
|
__release(rq2->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If dest_cpu is allowed for this process, migrate the task to it.
|
|
|
|
* This is accomplished by forcing the cpu_allowed mask to only
|
2007-12-05 14:46:09 +00:00
|
|
|
* allow dest_cpu, which will force the cpu onto dest_cpu. Then
|
2005-04-16 22:20:36 +00:00
|
|
|
* the cpu_allowed mask is restored.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
static void sched_migrate_task(struct task_struct *p, int dest_cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req req;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_test_cpu(dest_cpu, &p->cpus_allowed)
|
2008-07-15 11:43:49 +00:00
|
|
|
|| unlikely(!cpu_active(dest_cpu)))
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
/* force the process onto the specified CPU */
|
|
|
|
if (migrate_task(p, dest_cpu, &req)) {
|
|
|
|
/* Need to wait for migration thread (might exit: take ref). */
|
|
|
|
struct task_struct *mt = rq->migration_thread;
|
2006-07-03 07:25:41 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
get_task_struct(mt);
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
wake_up_process(mt);
|
|
|
|
put_task_struct(mt);
|
|
|
|
wait_for_completion(&req.done);
|
2006-07-03 07:25:41 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2005-06-25 21:57:29 +00:00
|
|
|
* sched_exec - execve() is a valuable balancing opportunity, because at
|
|
|
|
* this point the task has the smallest effective memory and cache footprint.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
void sched_exec(void)
|
|
|
|
{
|
|
|
|
int new_cpu, this_cpu = get_cpu();
|
2005-06-25 21:57:29 +00:00
|
|
|
new_cpu = sched_balance_self(this_cpu, SD_BALANCE_EXEC);
|
2005-04-16 22:20:36 +00:00
|
|
|
put_cpu();
|
2005-06-25 21:57:29 +00:00
|
|
|
if (new_cpu != this_cpu)
|
|
|
|
sched_migrate_task(current, new_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* pull_task - move a task from a remote runqueue to the local runqueue.
|
|
|
|
* Both runqueues must be locked.
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static void pull_task(struct rq *src_rq, struct task_struct *p,
|
|
|
|
struct rq *this_rq, int this_cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-08-09 09:16:49 +00:00
|
|
|
deactivate_task(src_rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
set_task_cpu(p, this_cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
activate_task(this_rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Note that idle threads have a prio of MAX_PRIO, for this test
|
|
|
|
* to be always true for them.
|
|
|
|
*/
|
2008-09-20 21:38:02 +00:00
|
|
|
check_preempt_curr(this_rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* can_migrate_task - may task p from runqueue rq be migrated to this_cpu?
|
|
|
|
*/
|
2006-01-14 21:20:43 +00:00
|
|
|
static
|
2006-07-03 07:25:42 +00:00
|
|
|
int can_migrate_task(struct task_struct *p, struct rq *rq, int this_cpu,
|
2007-07-09 16:51:57 +00:00
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle,
|
2005-09-10 07:26:11 +00:00
|
|
|
int *all_pinned)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2009-03-16 19:59:02 +00:00
|
|
|
int tsk_cache_hot = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* We do not migrate tasks that are:
|
|
|
|
* 1) running (obviously), or
|
|
|
|
* 2) cannot be migrated to this CPU due to cpus_allowed, or
|
|
|
|
* 3) are cache-hot on their current CPU.
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_test_cpu(this_cpu, &p->cpus_allowed)) {
|
2007-10-15 15:00:18 +00:00
|
|
|
schedstat_inc(p, se.nr_failed_migrations_affine);
|
2005-04-16 22:20:36 +00:00
|
|
|
return 0;
|
2007-10-15 15:00:18 +00:00
|
|
|
}
|
2005-06-25 21:57:07 +00:00
|
|
|
*all_pinned = 0;
|
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
if (task_running(rq, p)) {
|
|
|
|
schedstat_inc(p, se.nr_failed_migrations_running);
|
2005-06-25 21:57:07 +00:00
|
|
|
return 0;
|
2007-10-15 15:00:18 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
/*
|
|
|
|
* Aggressive migration if:
|
|
|
|
* 1) task is cache cold, or
|
|
|
|
* 2) too many balance attempts have failed.
|
|
|
|
*/
|
|
|
|
|
2009-03-16 19:59:02 +00:00
|
|
|
tsk_cache_hot = task_hot(p, rq->clock, sd);
|
|
|
|
if (!tsk_cache_hot ||
|
|
|
|
sd->nr_balance_failed > sd->cache_nice_tries) {
|
2007-10-15 15:00:18 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
2009-03-16 19:59:02 +00:00
|
|
|
if (tsk_cache_hot) {
|
2007-10-15 15:00:18 +00:00
|
|
|
schedstat_inc(sd, lb_hot_gained[idle]);
|
2007-10-15 15:00:18 +00:00
|
|
|
schedstat_inc(p, se.nr_forced_migrations);
|
|
|
|
}
|
2007-10-15 15:00:18 +00:00
|
|
|
#endif
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2009-03-16 19:59:02 +00:00
|
|
|
if (tsk_cache_hot) {
|
2007-10-15 15:00:18 +00:00
|
|
|
schedstat_inc(p, se.nr_failed_migrations_hot);
|
2007-10-15 15:00:18 +00:00
|
|
|
return 0;
|
2007-10-15 15:00:18 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2007-10-24 16:23:51 +00:00
|
|
|
static unsigned long
|
|
|
|
balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
|
|
|
unsigned long max_load_move, struct sched_domain *sd,
|
|
|
|
enum cpu_idle_type idle, int *all_pinned,
|
|
|
|
int *this_best_prio, struct rq_iterator *iterator)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-06-27 11:41:31 +00:00
|
|
|
int loops = 0, pulled = 0, pinned = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
struct task_struct *p;
|
|
|
|
long rem_load_move = max_load_move;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:51 +00:00
|
|
|
if (max_load_move == 0)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out;
|
|
|
|
|
2005-06-25 21:57:07 +00:00
|
|
|
pinned = 1;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* Start the load-balancing iterator:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
p = iterator->start(iterator->arg);
|
|
|
|
next:
|
2007-11-09 21:39:39 +00:00
|
|
|
if (!p || loops++ > sysctl_sched_nr_migrate)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out;
|
2008-06-27 11:41:31 +00:00
|
|
|
|
|
|
|
if ((p->se.load.weight >> 1) > rem_load_move ||
|
2007-07-09 16:51:59 +00:00
|
|
|
!can_migrate_task(p, busiest, this_cpu, sd, idle, &pinned)) {
|
|
|
|
p = iterator->next(iterator->arg);
|
|
|
|
goto next;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
pull_task(busiest, p, this_rq, this_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
pulled++;
|
2007-07-09 16:51:59 +00:00
|
|
|
rem_load_move -= p->se.load.weight;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-12-29 14:39:50 +00:00
|
|
|
#ifdef CONFIG_PREEMPT
|
|
|
|
/*
|
|
|
|
* NEWIDLE balancing is a source of latency, so preemptible kernels
|
|
|
|
* will stop after the first task is pulled to minimize the critical
|
|
|
|
* section.
|
|
|
|
*/
|
|
|
|
if (idle == CPU_NEWLY_IDLE)
|
|
|
|
goto out;
|
|
|
|
#endif
|
|
|
|
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
/*
|
2007-11-09 21:39:39 +00:00
|
|
|
* We only want to steal up to the prescribed amount of weighted load.
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
*/
|
2007-10-24 16:23:51 +00:00
|
|
|
if (rem_load_move > 0) {
|
2007-08-09 09:16:46 +00:00
|
|
|
if (p->prio < *this_best_prio)
|
|
|
|
*this_best_prio = p->prio;
|
2007-07-09 16:51:59 +00:00
|
|
|
p = iterator->next(iterator->arg);
|
|
|
|
goto next;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
/*
|
2007-10-24 16:23:51 +00:00
|
|
|
* Right now, this is one of only two places pull_task() is called,
|
2005-04-16 22:20:36 +00:00
|
|
|
* so we can safely collect pull_task() stats here rather than
|
|
|
|
* inside pull_task().
|
|
|
|
*/
|
|
|
|
schedstat_add(sd, lb_gained[idle], pulled);
|
2005-06-25 21:57:07 +00:00
|
|
|
|
|
|
|
if (all_pinned)
|
|
|
|
*all_pinned = pinned;
|
2007-10-24 16:23:51 +00:00
|
|
|
|
|
|
|
return max_load_move - rem_load_move;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
* move_tasks tries to move up to max_load_move weighted load from busiest to
|
|
|
|
* this_rq, as part of a balancing operation within domain "sd".
|
|
|
|
* Returns 1 if successful and 0 otherwise.
|
2007-07-09 16:51:59 +00:00
|
|
|
*
|
|
|
|
* Called with both runqueues locked.
|
|
|
|
*/
|
|
|
|
static int move_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
unsigned long max_load_move,
|
2007-07-09 16:51:59 +00:00
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle,
|
|
|
|
int *all_pinned)
|
|
|
|
{
|
2007-10-15 15:00:12 +00:00
|
|
|
const struct sched_class *class = sched_class_highest;
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
unsigned long total_load_moved = 0;
|
2007-08-09 09:16:46 +00:00
|
|
|
int this_best_prio = this_rq->curr->prio;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
do {
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
total_load_moved +=
|
|
|
|
class->load_balance(this_rq, this_cpu, busiest,
|
2007-10-24 16:23:51 +00:00
|
|
|
max_load_move - total_load_moved,
|
2007-08-09 09:16:46 +00:00
|
|
|
sd, idle, all_pinned, &this_best_prio);
|
2007-07-09 16:51:59 +00:00
|
|
|
class = class->next;
|
2008-06-27 20:29:55 +00:00
|
|
|
|
2008-12-29 14:39:50 +00:00
|
|
|
#ifdef CONFIG_PREEMPT
|
|
|
|
/*
|
|
|
|
* NEWIDLE balancing is a source of latency, so preemptible
|
|
|
|
* kernels will stop after the first task is pulled to minimize
|
|
|
|
* the critical section.
|
|
|
|
*/
|
2008-06-27 20:29:55 +00:00
|
|
|
if (idle == CPU_NEWLY_IDLE && this_rq->nr_running)
|
|
|
|
break;
|
2008-12-29 14:39:50 +00:00
|
|
|
#endif
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
} while (class && max_load_move > total_load_moved);
|
2007-07-09 16:51:59 +00:00
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
return total_load_moved > 0;
|
|
|
|
}
|
|
|
|
|
2007-10-24 16:23:51 +00:00
|
|
|
static int
|
|
|
|
iter_move_one_task(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle,
|
|
|
|
struct rq_iterator *iterator)
|
|
|
|
{
|
|
|
|
struct task_struct *p = iterator->start(iterator->arg);
|
|
|
|
int pinned = 0;
|
|
|
|
|
|
|
|
while (p) {
|
|
|
|
if (can_migrate_task(p, busiest, this_cpu, sd, idle, &pinned)) {
|
|
|
|
pull_task(busiest, p, this_rq, this_cpu);
|
|
|
|
/*
|
|
|
|
* Right now, this is only the second place pull_task()
|
|
|
|
* is called, so we can safely collect pull_task()
|
|
|
|
* stats here rather than inside pull_task().
|
|
|
|
*/
|
|
|
|
schedstat_inc(sd, lb_gained[idle]);
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
p = iterator->next(iterator->arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
/*
|
|
|
|
* move_one_task tries to move exactly one task from busiest to this_rq, as
|
|
|
|
* part of active balancing operations within "domain".
|
|
|
|
* Returns 1 if successful and 0 otherwise.
|
|
|
|
*
|
|
|
|
* Called with both runqueues locked.
|
|
|
|
*/
|
|
|
|
static int move_one_task(struct rq *this_rq, int this_cpu, struct rq *busiest,
|
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle)
|
|
|
|
{
|
2007-10-15 15:00:12 +00:00
|
|
|
const struct sched_class *class;
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
|
|
|
|
for (class = sched_class_highest; class; class = class->next)
|
2007-10-24 16:23:51 +00:00
|
|
|
if (class->move_one_task(this_rq, this_cpu, busiest, sd, idle))
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
return 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
2009-03-25 09:13:35 +00:00
|
|
|
/********** Helpers for find_busiest_group ************************/
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2009-03-25 09:13:56 +00:00
|
|
|
* sd_lb_stats - Structure to store the statistics of a sched_domain
|
|
|
|
* during load balancing.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-03-25 09:13:56 +00:00
|
|
|
struct sd_lb_stats {
|
|
|
|
struct sched_group *busiest; /* Busiest group in this sd */
|
|
|
|
struct sched_group *this; /* Local group in this sd */
|
|
|
|
unsigned long total_load; /* Total load of all groups in sd */
|
|
|
|
unsigned long total_pwr; /* Total power of all groups in sd */
|
|
|
|
unsigned long avg_load; /* Average load across all groups in sd */
|
|
|
|
|
|
|
|
/** Statistics of this group */
|
|
|
|
unsigned long this_load;
|
|
|
|
unsigned long this_load_per_task;
|
|
|
|
unsigned long this_nr_running;
|
|
|
|
|
|
|
|
/* Statistics of the busiest group */
|
|
|
|
unsigned long max_load;
|
|
|
|
unsigned long busiest_load_per_task;
|
|
|
|
unsigned long busiest_nr_running;
|
|
|
|
|
|
|
|
int group_imb; /* Is there imbalance in this sd */
|
2006-06-27 09:54:42 +00:00
|
|
|
#if defined(CONFIG_SCHED_MC) || defined(CONFIG_SCHED_SMT)
|
2009-03-25 09:13:56 +00:00
|
|
|
int power_savings_balance; /* Is powersave balance needed for this sd */
|
|
|
|
struct sched_group *group_min; /* Least loaded group in sd */
|
|
|
|
struct sched_group *group_leader; /* Group which relieves group_min */
|
|
|
|
unsigned long min_load_per_task; /* load_per_task in group_min */
|
|
|
|
unsigned long leader_nr_running; /* Nr running of group_leader */
|
|
|
|
unsigned long min_nr_running; /* Nr running of group_min */
|
2006-06-27 09:54:42 +00:00
|
|
|
#endif
|
2009-03-25 09:13:56 +00:00
|
|
|
};
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-29 04:52:47 +00:00
|
|
|
/*
|
2009-03-25 09:13:46 +00:00
|
|
|
* sg_lb_stats - stats of a sched_group required for load_balancing
|
|
|
|
*/
|
|
|
|
struct sg_lb_stats {
|
|
|
|
unsigned long avg_load; /*Avg load across the CPUs of the group */
|
|
|
|
unsigned long group_load; /* Total load over the CPUs of the group */
|
|
|
|
unsigned long sum_nr_running; /* Nr tasks running in the group */
|
|
|
|
unsigned long sum_weighted_load; /* Weighted load of group's tasks */
|
|
|
|
unsigned long group_capacity;
|
|
|
|
int group_imb; /* Is there an imbalance in the group ? */
|
|
|
|
};
|
2008-06-27 11:41:28 +00:00
|
|
|
|
2009-03-25 09:13:35 +00:00
|
|
|
/**
|
|
|
|
* group_first_cpu - Returns the first cpu in the cpumask of a sched_group.
|
|
|
|
* @group: The group whose first cpu is to be returned.
|
|
|
|
*/
|
|
|
|
static inline unsigned int group_first_cpu(struct sched_group *group)
|
|
|
|
{
|
|
|
|
return cpumask_first(sched_group_cpus(group));
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* get_sd_load_idx - Obtain the load index for a given sched domain.
|
|
|
|
* @sd: The sched_domain whose load_idx is to be obtained.
|
|
|
|
* @idle: The Idle status of the CPU for whose sd load_icx is obtained.
|
|
|
|
*/
|
|
|
|
static inline int get_sd_load_idx(struct sched_domain *sd,
|
|
|
|
enum cpu_idle_type idle)
|
|
|
|
{
|
|
|
|
int load_idx;
|
|
|
|
|
|
|
|
switch (idle) {
|
|
|
|
case CPU_NOT_IDLE:
|
2005-06-25 21:57:13 +00:00
|
|
|
load_idx = sd->busy_idx;
|
2009-03-25 09:13:35 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case CPU_NEWLY_IDLE:
|
2005-06-25 21:57:13 +00:00
|
|
|
load_idx = sd->newidle_idx;
|
2009-03-25 09:13:35 +00:00
|
|
|
break;
|
|
|
|
default:
|
2005-06-25 21:57:13 +00:00
|
|
|
load_idx = sd->idle_idx;
|
2009-03-25 09:13:35 +00:00
|
|
|
break;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:13:35 +00:00
|
|
|
return load_idx;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
#if defined(CONFIG_SCHED_MC) || defined(CONFIG_SCHED_SMT)
|
|
|
|
/**
|
|
|
|
* init_sd_power_savings_stats - Initialize power savings statistics for
|
|
|
|
* the given sched_domain, during load balancing.
|
|
|
|
*
|
|
|
|
* @sd: Sched domain whose power-savings statistics are to be initialized.
|
|
|
|
* @sds: Variable containing the statistics for sd.
|
|
|
|
* @idle: Idle status of the CPU at which we're performing load-balancing.
|
|
|
|
*/
|
|
|
|
static inline void init_sd_power_savings_stats(struct sched_domain *sd,
|
|
|
|
struct sd_lb_stats *sds, enum cpu_idle_type idle)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Busy processors will not participate in power savings
|
|
|
|
* balance.
|
|
|
|
*/
|
|
|
|
if (idle == CPU_NOT_IDLE || !(sd->flags & SD_POWERSAVINGS_BALANCE))
|
|
|
|
sds->power_savings_balance = 0;
|
|
|
|
else {
|
|
|
|
sds->power_savings_balance = 1;
|
|
|
|
sds->min_nr_running = ULONG_MAX;
|
|
|
|
sds->leader_nr_running = 0;
|
|
|
|
}
|
|
|
|
}
|
2006-12-10 10:20:33 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/**
|
|
|
|
* update_sd_power_savings_stats - Update the power saving stats for a
|
|
|
|
* sched_domain while performing load balancing.
|
|
|
|
*
|
|
|
|
* @group: sched_group belonging to the sched_domain under consideration.
|
|
|
|
* @sds: Variable containing the statistics of the sched_domain
|
|
|
|
* @local_group: Does group contain the CPU for which we're performing
|
|
|
|
* load balancing ?
|
|
|
|
* @sgs: Variable containing the statistics of the group.
|
|
|
|
*/
|
|
|
|
static inline void update_sd_power_savings_stats(struct sched_group *group,
|
|
|
|
struct sd_lb_stats *sds, int local_group, struct sg_lb_stats *sgs)
|
|
|
|
{
|
2008-06-27 11:41:28 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
if (!sds->power_savings_balance)
|
|
|
|
return;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/*
|
|
|
|
* If the local group is idle or completely loaded
|
|
|
|
* no need to do power savings balance at this domain
|
|
|
|
*/
|
|
|
|
if (local_group && (sds->this_nr_running >= sgs->group_capacity ||
|
|
|
|
!sds->this_nr_running))
|
|
|
|
sds->power_savings_balance = 0;
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/*
|
|
|
|
* If a group is already running at full capacity or idle,
|
|
|
|
* don't include that group in power savings calculations
|
|
|
|
*/
|
|
|
|
if (!sds->power_savings_balance ||
|
|
|
|
sgs->sum_nr_running >= sgs->group_capacity ||
|
|
|
|
!sgs->sum_nr_running)
|
|
|
|
return;
|
2005-09-10 07:26:19 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/*
|
|
|
|
* Calculate the group which has the least non-idle load.
|
|
|
|
* This is the group from where we need to pick up the load
|
|
|
|
* for saving power
|
|
|
|
*/
|
|
|
|
if ((sgs->sum_nr_running < sds->min_nr_running) ||
|
|
|
|
(sgs->sum_nr_running == sds->min_nr_running &&
|
|
|
|
group_first_cpu(group) > group_first_cpu(sds->group_min))) {
|
|
|
|
sds->group_min = group;
|
|
|
|
sds->min_nr_running = sgs->sum_nr_running;
|
|
|
|
sds->min_load_per_task = sgs->sum_weighted_load /
|
|
|
|
sgs->sum_nr_running;
|
|
|
|
}
|
2006-12-10 10:20:33 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/*
|
|
|
|
* Calculate the group which is almost near its
|
|
|
|
* capacity but still has some space to pick up some load
|
|
|
|
* from other group and save more power
|
|
|
|
*/
|
|
|
|
if (sgs->sum_nr_running > sgs->group_capacity - 1)
|
|
|
|
return;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
if (sgs->sum_nr_running > sds->leader_nr_running ||
|
|
|
|
(sgs->sum_nr_running == sds->leader_nr_running &&
|
|
|
|
group_first_cpu(group) < group_first_cpu(sds->group_leader))) {
|
|
|
|
sds->group_leader = group;
|
|
|
|
sds->leader_nr_running = sgs->sum_nr_running;
|
|
|
|
}
|
|
|
|
}
|
2008-06-27 11:41:28 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
/**
|
2009-03-29 04:52:47 +00:00
|
|
|
* check_power_save_busiest_group - see if there is potential for some power-savings balance
|
2009-03-25 09:14:22 +00:00
|
|
|
* @sds: Variable containing the statistics of the sched_domain
|
|
|
|
* under consideration.
|
|
|
|
* @this_cpu: Cpu at which we're currently performing load-balancing.
|
|
|
|
* @imbalance: Variable to store the imbalance.
|
|
|
|
*
|
2009-03-29 04:52:47 +00:00
|
|
|
* Description:
|
|
|
|
* Check if we have potential to perform some power-savings balance.
|
|
|
|
* If yes, set the busiest group to be the least loaded group in the
|
|
|
|
* sched_domain, so that it's CPUs can be put to idle.
|
|
|
|
*
|
2009-03-25 09:14:22 +00:00
|
|
|
* Returns 1 if there is potential to perform power-savings balance.
|
|
|
|
* Else returns 0.
|
|
|
|
*/
|
|
|
|
static inline int check_power_save_busiest_group(struct sd_lb_stats *sds,
|
|
|
|
int this_cpu, unsigned long *imbalance)
|
|
|
|
{
|
|
|
|
if (!sds->power_savings_balance)
|
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
if (sds->this != sds->group_leader ||
|
|
|
|
sds->group_leader == sds->group_min)
|
|
|
|
return 0;
|
2006-12-10 10:20:33 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
*imbalance = sds->min_load_per_task;
|
|
|
|
sds->busiest = sds->group_min;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
if (sched_mc_power_savings >= POWERSAVINGS_BALANCE_WAKEUP) {
|
|
|
|
cpu_rq(this_cpu)->rd->sched_mc_preferred_wakeup_cpu =
|
|
|
|
group_first_cpu(sds->group_leader);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
}
|
|
|
|
#else /* CONFIG_SCHED_MC || CONFIG_SCHED_SMT */
|
|
|
|
static inline void init_sd_power_savings_stats(struct sched_domain *sd,
|
|
|
|
struct sd_lb_stats *sds, enum cpu_idle_type idle)
|
|
|
|
{
|
|
|
|
return;
|
|
|
|
}
|
2008-06-27 11:41:28 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
static inline void update_sd_power_savings_stats(struct sched_group *group,
|
|
|
|
struct sd_lb_stats *sds, int local_group, struct sg_lb_stats *sgs)
|
|
|
|
{
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int check_power_save_busiest_group(struct sd_lb_stats *sds,
|
|
|
|
int this_cpu, unsigned long *imbalance)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif /* CONFIG_SCHED_MC || CONFIG_SCHED_SMT */
|
|
|
|
|
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
/**
|
|
|
|
* update_sg_lb_stats - Update sched_group's statistics for load balancing.
|
|
|
|
* @group: sched_group whose statistics are to be updated.
|
|
|
|
* @this_cpu: Cpu for which load balance is currently performed.
|
|
|
|
* @idle: Idle status of this_cpu
|
|
|
|
* @load_idx: Load index of sched_domain of this_cpu for load calc.
|
|
|
|
* @sd_idle: Idle status of the sched_domain containing group.
|
|
|
|
* @local_group: Does group contain this_cpu.
|
|
|
|
* @cpus: Set of cpus considered for load balancing.
|
|
|
|
* @balance: Should we balance.
|
|
|
|
* @sgs: variable to hold the statistics for this group.
|
|
|
|
*/
|
|
|
|
static inline void update_sg_lb_stats(struct sched_group *group, int this_cpu,
|
|
|
|
enum cpu_idle_type idle, int load_idx, int *sd_idle,
|
|
|
|
int local_group, const struct cpumask *cpus,
|
|
|
|
int *balance, struct sg_lb_stats *sgs)
|
|
|
|
{
|
|
|
|
unsigned long load, max_cpu_load, min_cpu_load;
|
|
|
|
int i;
|
|
|
|
unsigned int balance_cpu = -1, first_idle_cpu = 0;
|
|
|
|
unsigned long sum_avg_load_per_task;
|
|
|
|
unsigned long avg_load_per_task;
|
|
|
|
|
|
|
|
if (local_group)
|
|
|
|
balance_cpu = group_first_cpu(group);
|
|
|
|
|
|
|
|
/* Tally up the load of all CPUs in the group */
|
|
|
|
sum_avg_load_per_task = avg_load_per_task = 0;
|
|
|
|
max_cpu_load = 0;
|
|
|
|
min_cpu_load = ~0UL;
|
2008-06-27 11:41:28 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
for_each_cpu_and(i, sched_group_cpus(group), cpus) {
|
|
|
|
struct rq *rq = cpu_rq(i);
|
sched: fix improper load balance across sched domain
We recently discovered a nasty performance bug in the kernel CPU load
balancer where we were hit by 50% performance regression.
When tasks are assigned to a subset of CPUs that span across
sched_domains (either ccNUMA node or the new multi-core domain) via
cpu affinity, kernel fails to perform proper load balance at
these domains, due to several logic in find_busiest_group() miss
identified busiest sched group within a given domain. This leads to
inadequate load balance and causes 50% performance hit.
To give you a concrete example, on a dual-core, 2 socket numa system,
there are 4 logical cpu, organized as:
CPU0 attaching sched-domain:
domain 0: span 0003 groups: 0001 0002
domain 1: span 000f groups: 0003 000c
CPU1 attaching sched-domain:
domain 0: span 0003 groups: 0002 0001
domain 1: span 000f groups: 0003 000c
CPU2 attaching sched-domain:
domain 0: span 000c groups: 0004 0008
domain 1: span 000f groups: 000c 0003
CPU3 attaching sched-domain:
domain 0: span 000c groups: 0008 0004
domain 1: span 000f groups: 000c 0003
If I run 2 tasks with CPU affinity set to 0x5. There are situation
where cpu0 has run queue length of 2, and cpu2 will be idle. The
kernel load balancer is unable to balance out these two tasks over
cpu0 and cpu2 due to at least three logics in find_busiest_group()
that heavily bias load balance towards power saving mode. e.g. while
determining "busiest" variable, kernel only set it when
"sum_nr_running > group_capacity". This test is flawed that
"sum_nr_running" is not necessary same as
sum-tasks-allowed-to-run-within-the sched-group. The end result is
that kernel "think" everything is balanced, but in reality we have an
imbalance and thus causing one CPU to be over-subscribed and leaving
other idle. There are two other logic in the same function will also
causing similar effect. The nastiness of this bug is that kernel not
be able to get unstuck in this unfortunate broken state. From what
we've seen in our environment, kernel will stuck in imbalanced state
for extended period of time and it is also very easy for the kernel to
stuck into that state (it's pretty much 100% reproducible for us).
So proposing the following fix: add addition logic in
find_busiest_group to detect intrinsic imbalance within the busiest
group. When such condition is detected, load balance goes into spread
mode instead of default grouping mode.
Signed-off-by: Ken Chen <kenchen@google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-10-17 14:55:11 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
if (*sd_idle && rq->nr_running)
|
|
|
|
*sd_idle = 0;
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
/* Bias balancing toward cpus of our domain */
|
2005-04-16 22:20:36 +00:00
|
|
|
if (local_group) {
|
2009-03-25 09:13:51 +00:00
|
|
|
if (idle_cpu(i) && !first_idle_cpu) {
|
|
|
|
first_idle_cpu = 1;
|
|
|
|
balance_cpu = i;
|
|
|
|
}
|
|
|
|
|
|
|
|
load = target_load(i, load_idx);
|
|
|
|
} else {
|
|
|
|
load = source_load(i, load_idx);
|
|
|
|
if (load > max_cpu_load)
|
|
|
|
max_cpu_load = load;
|
|
|
|
if (min_cpu_load > load)
|
|
|
|
min_cpu_load = load;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
sgs->group_load += load;
|
|
|
|
sgs->sum_nr_running += rq->nr_running;
|
|
|
|
sgs->sum_weighted_load += weighted_cpuload(i);
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
sum_avg_load_per_task += cpu_avg_load_per_task(i);
|
|
|
|
}
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
/*
|
|
|
|
* First idle cpu or the first cpu(busiest) in this sched group
|
|
|
|
* is eligible for doing load balancing at this and above
|
|
|
|
* domains. In the newly idle case, we will allow all the cpu's
|
|
|
|
* to do the newly idle load balance.
|
|
|
|
*/
|
|
|
|
if (idle != CPU_NEWLY_IDLE && local_group &&
|
|
|
|
balance_cpu != this_cpu && balance) {
|
|
|
|
*balance = 0;
|
|
|
|
return;
|
|
|
|
}
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
/* Adjust by relative CPU power of the group */
|
|
|
|
sgs->avg_load = sg_div_cpu_power(group,
|
|
|
|
sgs->group_load * SCHED_LOAD_SCALE);
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:13:51 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Consider the group unbalanced when the imbalance is larger
|
|
|
|
* than the average weight of two tasks.
|
|
|
|
*
|
|
|
|
* APZ: with cgroup the avg task weight can vary wildly and
|
|
|
|
* might not be a suitable number - should we keep a
|
|
|
|
* normalized nr_running number somewhere that negates
|
|
|
|
* the hierarchy?
|
|
|
|
*/
|
|
|
|
avg_load_per_task = sg_div_cpu_power(group,
|
|
|
|
sum_avg_load_per_task * SCHED_LOAD_SCALE);
|
|
|
|
|
|
|
|
if ((max_cpu_load - min_cpu_load) > 2*avg_load_per_task)
|
|
|
|
sgs->group_imb = 1;
|
|
|
|
|
|
|
|
sgs->group_capacity = group->__cpu_power / SCHED_LOAD_SCALE;
|
|
|
|
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
/**
|
|
|
|
* update_sd_lb_stats - Update sched_group's statistics for load balancing.
|
|
|
|
* @sd: sched_domain whose statistics are to be updated.
|
|
|
|
* @this_cpu: Cpu for which load balance is currently performed.
|
|
|
|
* @idle: Idle status of this_cpu
|
|
|
|
* @sd_idle: Idle status of the sched_domain containing group.
|
|
|
|
* @cpus: Set of cpus considered for load balancing.
|
|
|
|
* @balance: Should we balance.
|
|
|
|
* @sds: variable to hold the statistics for this sched_domain.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-03-25 09:14:01 +00:00
|
|
|
static inline void update_sd_lb_stats(struct sched_domain *sd, int this_cpu,
|
|
|
|
enum cpu_idle_type idle, int *sd_idle,
|
|
|
|
const struct cpumask *cpus, int *balance,
|
|
|
|
struct sd_lb_stats *sds)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2009-03-25 09:13:56 +00:00
|
|
|
struct sched_group *group = sd->groups;
|
2009-03-25 09:14:01 +00:00
|
|
|
struct sg_lb_stats sgs;
|
2009-03-25 09:13:56 +00:00
|
|
|
int load_idx;
|
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
init_sd_power_savings_stats(sd, sds, idle);
|
2009-03-25 09:13:35 +00:00
|
|
|
load_idx = get_sd_load_idx(sd, idle);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
do {
|
|
|
|
int local_group;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
local_group = cpumask_test_cpu(this_cpu,
|
|
|
|
sched_group_cpus(group));
|
2009-03-25 09:13:46 +00:00
|
|
|
memset(&sgs, 0, sizeof(sgs));
|
2009-03-25 09:13:51 +00:00
|
|
|
update_sg_lb_stats(group, this_cpu, idle, load_idx, sd_idle,
|
|
|
|
local_group, cpus, balance, &sgs);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
if (local_group && balance && !(*balance))
|
|
|
|
return;
|
2006-12-10 10:20:33 +00:00
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
sds->total_load += sgs.group_load;
|
|
|
|
sds->total_pwr += group->__cpu_power;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (local_group) {
|
2009-03-25 09:14:01 +00:00
|
|
|
sds->this_load = sgs.avg_load;
|
|
|
|
sds->this = group;
|
|
|
|
sds->this_nr_running = sgs.sum_nr_running;
|
|
|
|
sds->this_load_per_task = sgs.sum_weighted_load;
|
|
|
|
} else if (sgs.avg_load > sds->max_load &&
|
2009-03-25 09:13:46 +00:00
|
|
|
(sgs.sum_nr_running > sgs.group_capacity ||
|
|
|
|
sgs.group_imb)) {
|
2009-03-25 09:14:01 +00:00
|
|
|
sds->max_load = sgs.avg_load;
|
|
|
|
sds->busiest = group;
|
|
|
|
sds->busiest_nr_running = sgs.sum_nr_running;
|
|
|
|
sds->busiest_load_per_task = sgs.sum_weighted_load;
|
|
|
|
sds->group_imb = sgs.group_imb;
|
2006-07-03 07:25:40 +00:00
|
|
|
}
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-03-25 09:14:22 +00:00
|
|
|
update_sd_power_savings_stats(group, sds, local_group, &sgs);
|
2005-04-16 22:20:36 +00:00
|
|
|
group = group->next;
|
|
|
|
} while (group != sd->groups);
|
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:06 +00:00
|
|
|
/**
|
|
|
|
* fix_small_imbalance - Calculate the minor imbalance that exists
|
2009-03-25 09:14:12 +00:00
|
|
|
* amongst the groups of a sched_domain, during
|
|
|
|
* load balancing.
|
2009-03-25 09:14:06 +00:00
|
|
|
* @sds: Statistics of the sched_domain whose imbalance is to be calculated.
|
|
|
|
* @this_cpu: The cpu at whose sched_domain we're performing load-balance.
|
|
|
|
* @imbalance: Variable to store the imbalance.
|
|
|
|
*/
|
|
|
|
static inline void fix_small_imbalance(struct sd_lb_stats *sds,
|
|
|
|
int this_cpu, unsigned long *imbalance)
|
|
|
|
{
|
|
|
|
unsigned long tmp, pwr_now = 0, pwr_move = 0;
|
|
|
|
unsigned int imbn = 2;
|
|
|
|
|
|
|
|
if (sds->this_nr_running) {
|
|
|
|
sds->this_load_per_task /= sds->this_nr_running;
|
|
|
|
if (sds->busiest_load_per_task >
|
|
|
|
sds->this_load_per_task)
|
|
|
|
imbn = 1;
|
|
|
|
} else
|
|
|
|
sds->this_load_per_task =
|
|
|
|
cpu_avg_load_per_task(this_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:06 +00:00
|
|
|
if (sds->max_load - sds->this_load + sds->busiest_load_per_task >=
|
|
|
|
sds->busiest_load_per_task * imbn) {
|
|
|
|
*imbalance = sds->busiest_load_per_task;
|
|
|
|
return;
|
|
|
|
}
|
sched: fix improper load balance across sched domain
We recently discovered a nasty performance bug in the kernel CPU load
balancer where we were hit by 50% performance regression.
When tasks are assigned to a subset of CPUs that span across
sched_domains (either ccNUMA node or the new multi-core domain) via
cpu affinity, kernel fails to perform proper load balance at
these domains, due to several logic in find_busiest_group() miss
identified busiest sched group within a given domain. This leads to
inadequate load balance and causes 50% performance hit.
To give you a concrete example, on a dual-core, 2 socket numa system,
there are 4 logical cpu, organized as:
CPU0 attaching sched-domain:
domain 0: span 0003 groups: 0001 0002
domain 1: span 000f groups: 0003 000c
CPU1 attaching sched-domain:
domain 0: span 0003 groups: 0002 0001
domain 1: span 000f groups: 0003 000c
CPU2 attaching sched-domain:
domain 0: span 000c groups: 0004 0008
domain 1: span 000f groups: 000c 0003
CPU3 attaching sched-domain:
domain 0: span 000c groups: 0008 0004
domain 1: span 000f groups: 000c 0003
If I run 2 tasks with CPU affinity set to 0x5. There are situation
where cpu0 has run queue length of 2, and cpu2 will be idle. The
kernel load balancer is unable to balance out these two tasks over
cpu0 and cpu2 due to at least three logics in find_busiest_group()
that heavily bias load balance towards power saving mode. e.g. while
determining "busiest" variable, kernel only set it when
"sum_nr_running > group_capacity". This test is flawed that
"sum_nr_running" is not necessary same as
sum-tasks-allowed-to-run-within-the sched-group. The end result is
that kernel "think" everything is balanced, but in reality we have an
imbalance and thus causing one CPU to be over-subscribed and leaving
other idle. There are two other logic in the same function will also
causing similar effect. The nastiness of this bug is that kernel not
be able to get unstuck in this unfortunate broken state. From what
we've seen in our environment, kernel will stuck in imbalanced state
for extended period of time and it is also very easy for the kernel to
stuck into that state (it's pretty much 100% reproducible for us).
So proposing the following fix: add addition logic in
find_busiest_group to detect intrinsic imbalance within the busiest
group. When such condition is detected, load balance goes into spread
mode instead of default grouping mode.
Signed-off-by: Ken Chen <kenchen@google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-10-17 14:55:11 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2009-03-25 09:14:06 +00:00
|
|
|
* OK, we don't have enough imbalance to justify moving tasks,
|
|
|
|
* however we may be able to increase total CPU power used by
|
|
|
|
* moving them.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2009-03-25 09:14:06 +00:00
|
|
|
pwr_now += sds->busiest->__cpu_power *
|
|
|
|
min(sds->busiest_load_per_task, sds->max_load);
|
|
|
|
pwr_now += sds->this->__cpu_power *
|
|
|
|
min(sds->this_load_per_task, sds->this_load);
|
|
|
|
pwr_now /= SCHED_LOAD_SCALE;
|
|
|
|
|
|
|
|
/* Amount of load we'd subtract */
|
|
|
|
tmp = sg_div_cpu_power(sds->busiest,
|
|
|
|
sds->busiest_load_per_task * SCHED_LOAD_SCALE);
|
|
|
|
if (sds->max_load > tmp)
|
|
|
|
pwr_move += sds->busiest->__cpu_power *
|
|
|
|
min(sds->busiest_load_per_task, sds->max_load - tmp);
|
|
|
|
|
|
|
|
/* Amount of load we'd add */
|
|
|
|
if (sds->max_load * sds->busiest->__cpu_power <
|
|
|
|
sds->busiest_load_per_task * SCHED_LOAD_SCALE)
|
|
|
|
tmp = sg_div_cpu_power(sds->this,
|
|
|
|
sds->max_load * sds->busiest->__cpu_power);
|
|
|
|
else
|
|
|
|
tmp = sg_div_cpu_power(sds->this,
|
|
|
|
sds->busiest_load_per_task * SCHED_LOAD_SCALE);
|
|
|
|
pwr_move += sds->this->__cpu_power *
|
|
|
|
min(sds->this_load_per_task, sds->this_load + tmp);
|
|
|
|
pwr_move /= SCHED_LOAD_SCALE;
|
|
|
|
|
|
|
|
/* Move if we gain throughput */
|
|
|
|
if (pwr_move > pwr_now)
|
|
|
|
*imbalance = sds->busiest_load_per_task;
|
|
|
|
}
|
2009-03-25 09:14:12 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* calculate_imbalance - Calculate the amount of imbalance present within the
|
|
|
|
* groups of a given sched_domain during load balance.
|
|
|
|
* @sds: statistics of the sched_domain whose imbalance is to be calculated.
|
|
|
|
* @this_cpu: Cpu for which currently load balance is being performed.
|
|
|
|
* @imbalance: The variable to store the imbalance.
|
|
|
|
*/
|
|
|
|
static inline void calculate_imbalance(struct sd_lb_stats *sds, int this_cpu,
|
|
|
|
unsigned long *imbalance)
|
|
|
|
{
|
|
|
|
unsigned long max_pull;
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
/*
|
|
|
|
* In the presence of smp nice balancing, certain scenarios can have
|
|
|
|
* max load less than avg load(as we skip the groups at or below
|
|
|
|
* its cpu_power, while calculating max_load..)
|
|
|
|
*/
|
2009-03-25 09:14:12 +00:00
|
|
|
if (sds->max_load < sds->avg_load) {
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
*imbalance = 0;
|
2009-03-25 09:14:12 +00:00
|
|
|
return fix_small_imbalance(sds, this_cpu, imbalance);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
}
|
2005-09-10 07:26:21 +00:00
|
|
|
|
|
|
|
/* Don't want to pull so many tasks that a group would go idle */
|
2009-03-25 09:14:12 +00:00
|
|
|
max_pull = min(sds->max_load - sds->avg_load,
|
|
|
|
sds->max_load - sds->busiest_load_per_task);
|
2005-09-10 07:26:21 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* How much load to actually move to equalise the imbalance */
|
2009-03-25 09:14:12 +00:00
|
|
|
*imbalance = min(max_pull * sds->busiest->__cpu_power,
|
|
|
|
(sds->avg_load - sds->this_load) * sds->this->__cpu_power)
|
2005-04-16 22:20:36 +00:00
|
|
|
/ SCHED_LOAD_SCALE;
|
|
|
|
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
/*
|
|
|
|
* if *imbalance is less than the average load per runnable task
|
|
|
|
* there is no gaurantee that any tasks will be moved so we'll have
|
|
|
|
* a think about bumping its value to force at least one task to be
|
|
|
|
* moved
|
|
|
|
*/
|
2009-03-25 09:14:12 +00:00
|
|
|
if (*imbalance < sds->busiest_load_per_task)
|
|
|
|
return fix_small_imbalance(sds, this_cpu, imbalance);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:12 +00:00
|
|
|
}
|
2009-03-25 09:14:01 +00:00
|
|
|
/******* find_busiest_group() helpers end here *********************/
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:27 +00:00
|
|
|
/**
|
|
|
|
* find_busiest_group - Returns the busiest group within the sched_domain
|
|
|
|
* if there is an imbalance. If there isn't an imbalance, and
|
|
|
|
* the user has opted for power-savings, it returns a group whose
|
|
|
|
* CPUs can be put to idle by rebalancing those tasks elsewhere, if
|
|
|
|
* such a group exists.
|
|
|
|
*
|
|
|
|
* Also calculates the amount of weighted load which should be moved
|
|
|
|
* to restore balance.
|
|
|
|
*
|
|
|
|
* @sd: The sched_domain whose busiest group is to be returned.
|
|
|
|
* @this_cpu: The cpu for which load balancing is currently being performed.
|
|
|
|
* @imbalance: Variable which stores amount of weighted load which should
|
|
|
|
* be moved to restore balance/put a group to idle.
|
|
|
|
* @idle: The idle status of this_cpu.
|
|
|
|
* @sd_idle: The idleness of sd
|
|
|
|
* @cpus: The set of CPUs under consideration for load-balancing.
|
|
|
|
* @balance: Pointer to a variable indicating if this_cpu
|
|
|
|
* is the appropriate cpu to perform load balancing at this_level.
|
|
|
|
*
|
|
|
|
* Returns: - the busiest group if imbalance exists.
|
|
|
|
* - If no imbalance and user has opted for power-savings balance,
|
|
|
|
* return the least loaded group whose CPUs can be
|
|
|
|
* put to idle by rebalancing its tasks onto our group.
|
2009-03-25 09:14:01 +00:00
|
|
|
*/
|
|
|
|
static struct sched_group *
|
|
|
|
find_busiest_group(struct sched_domain *sd, int this_cpu,
|
|
|
|
unsigned long *imbalance, enum cpu_idle_type idle,
|
|
|
|
int *sd_idle, const struct cpumask *cpus, int *balance)
|
|
|
|
{
|
|
|
|
struct sd_lb_stats sds;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
memset(&sds, 0, sizeof(sds));
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:01 +00:00
|
|
|
/*
|
|
|
|
* Compute the various statistics relavent for load balancing at
|
|
|
|
* this level.
|
|
|
|
*/
|
|
|
|
update_sd_lb_stats(sd, this_cpu, idle, sd_idle, cpus,
|
|
|
|
balance, &sds);
|
|
|
|
|
2009-03-25 09:14:27 +00:00
|
|
|
/* Cases where imbalance does not exist from POV of this_cpu */
|
|
|
|
/* 1) this_cpu is not the appropriate cpu to perform load balancing
|
|
|
|
* at this level.
|
|
|
|
* 2) There is no busy sibling group to pull from.
|
|
|
|
* 3) This group is the busiest group.
|
|
|
|
* 4) This group is more busy than the avg busieness at this
|
|
|
|
* sched_domain.
|
|
|
|
* 5) The imbalance is within the specified limit.
|
|
|
|
* 6) Any rebalance would lead to ping-pong
|
|
|
|
*/
|
2009-03-25 09:14:01 +00:00
|
|
|
if (balance && !(*balance))
|
|
|
|
goto ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:27 +00:00
|
|
|
if (!sds.busiest || sds.busiest_nr_running == 0)
|
|
|
|
goto out_balanced;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:27 +00:00
|
|
|
if (sds.this_load >= sds.max_load)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out_balanced;
|
|
|
|
|
2009-03-25 09:13:56 +00:00
|
|
|
sds.avg_load = (SCHED_LOAD_SCALE * sds.total_load) / sds.total_pwr;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-03-25 09:14:27 +00:00
|
|
|
if (sds.this_load >= sds.avg_load)
|
|
|
|
goto out_balanced;
|
|
|
|
|
|
|
|
if (100 * sds.max_load <= sd->imbalance_pct * sds.this_load)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out_balanced;
|
|
|
|
|
2009-03-25 09:13:56 +00:00
|
|
|
sds.busiest_load_per_task /= sds.busiest_nr_running;
|
|
|
|
if (sds.group_imb)
|
|
|
|
sds.busiest_load_per_task =
|
|
|
|
min(sds.busiest_load_per_task, sds.avg_load);
|
sched: fix improper load balance across sched domain
We recently discovered a nasty performance bug in the kernel CPU load
balancer where we were hit by 50% performance regression.
When tasks are assigned to a subset of CPUs that span across
sched_domains (either ccNUMA node or the new multi-core domain) via
cpu affinity, kernel fails to perform proper load balance at
these domains, due to several logic in find_busiest_group() miss
identified busiest sched group within a given domain. This leads to
inadequate load balance and causes 50% performance hit.
To give you a concrete example, on a dual-core, 2 socket numa system,
there are 4 logical cpu, organized as:
CPU0 attaching sched-domain:
domain 0: span 0003 groups: 0001 0002
domain 1: span 000f groups: 0003 000c
CPU1 attaching sched-domain:
domain 0: span 0003 groups: 0002 0001
domain 1: span 000f groups: 0003 000c
CPU2 attaching sched-domain:
domain 0: span 000c groups: 0004 0008
domain 1: span 000f groups: 000c 0003
CPU3 attaching sched-domain:
domain 0: span 000c groups: 0008 0004
domain 1: span 000f groups: 000c 0003
If I run 2 tasks with CPU affinity set to 0x5. There are situation
where cpu0 has run queue length of 2, and cpu2 will be idle. The
kernel load balancer is unable to balance out these two tasks over
cpu0 and cpu2 due to at least three logics in find_busiest_group()
that heavily bias load balance towards power saving mode. e.g. while
determining "busiest" variable, kernel only set it when
"sum_nr_running > group_capacity". This test is flawed that
"sum_nr_running" is not necessary same as
sum-tasks-allowed-to-run-within-the sched-group. The end result is
that kernel "think" everything is balanced, but in reality we have an
imbalance and thus causing one CPU to be over-subscribed and leaving
other idle. There are two other logic in the same function will also
causing similar effect. The nastiness of this bug is that kernel not
be able to get unstuck in this unfortunate broken state. From what
we've seen in our environment, kernel will stuck in imbalanced state
for extended period of time and it is also very easy for the kernel to
stuck into that state (it's pretty much 100% reproducible for us).
So proposing the following fix: add addition logic in
find_busiest_group to detect intrinsic imbalance within the busiest
group. When such condition is detected, load balance goes into spread
mode instead of default grouping mode.
Signed-off-by: Ken Chen <kenchen@google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-10-17 14:55:11 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* We're trying to get all the cpus to the average_load, so we don't
|
|
|
|
* want to push ourselves above the average load, nor do we wish to
|
|
|
|
* reduce the max loaded cpu below the average load, as either of these
|
|
|
|
* actions would just result in more rebalancing later, and ping-pong
|
|
|
|
* tasks around. Thus we look for the minimum possible imbalance.
|
|
|
|
* Negative imbalances (*we* are more loaded than anyone else) will
|
|
|
|
* be counted as no imbalance for these purposes -- we can't fix that
|
2007-12-05 14:46:09 +00:00
|
|
|
* by pulling tasks to us. Be careful of negative numbers as they'll
|
2005-04-16 22:20:36 +00:00
|
|
|
* appear as very large values with unsigned longs.
|
|
|
|
*/
|
2009-03-25 09:13:56 +00:00
|
|
|
if (sds.max_load <= sds.busiest_load_per_task)
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
goto out_balanced;
|
|
|
|
|
2009-03-25 09:14:12 +00:00
|
|
|
/* Looks like there is an imbalance. Compute it */
|
|
|
|
calculate_imbalance(&sds, this_cpu, imbalance);
|
2009-03-25 09:13:56 +00:00
|
|
|
return sds.busiest;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
out_balanced:
|
2009-03-25 09:14:22 +00:00
|
|
|
/*
|
|
|
|
* There is no obvious imbalance. But check if we can do some balancing
|
|
|
|
* to save power.
|
|
|
|
*/
|
|
|
|
if (check_power_save_busiest_group(&sds, this_cpu, imbalance))
|
|
|
|
return sds.busiest;
|
2006-12-10 10:20:33 +00:00
|
|
|
ret:
|
2005-04-16 22:20:36 +00:00
|
|
|
*imbalance = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* find_busiest_queue - find the busiest runqueue among the cpus in group.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static struct rq *
|
2007-07-09 16:51:57 +00:00
|
|
|
find_busiest_queue(struct sched_group *group, enum cpu_idle_type idle,
|
2008-11-24 16:05:14 +00:00
|
|
|
unsigned long imbalance, const struct cpumask *cpus)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *busiest = NULL, *rq;
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
unsigned long max_load = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
int i;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu(i, sched_group_cpus(group)) {
|
2007-07-09 16:51:59 +00:00
|
|
|
unsigned long wl;
|
2006-09-26 06:30:51 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_test_cpu(i, cpus))
|
2006-09-26 06:30:51 +00:00
|
|
|
continue;
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
rq = cpu_rq(i);
|
2007-07-09 16:51:59 +00:00
|
|
|
wl = weighted_cpuload(i);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (rq->nr_running == 1 && wl > imbalance)
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
continue;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (wl > max_load) {
|
|
|
|
max_load = wl;
|
2006-07-03 07:25:40 +00:00
|
|
|
busiest = rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return busiest;
|
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:30 +00:00
|
|
|
/*
|
|
|
|
* Max backoff if we encounter pinned tasks. Pretty arbitrary value, but
|
|
|
|
* so long as it is large enough.
|
|
|
|
*/
|
|
|
|
#define MAX_PINNED_INTERVAL 512
|
|
|
|
|
2009-03-19 04:52:20 +00:00
|
|
|
/* Working cpumask for load_balance and load_balance_newidle. */
|
|
|
|
static DEFINE_PER_CPU(cpumask_var_t, load_balance_tmpmask);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Check this_cpu to ensure it is balanced within domain. Attempt to move
|
|
|
|
* tasks if there is an imbalance.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static int load_balance(int this_cpu, struct rq *this_rq,
|
2007-07-09 16:51:57 +00:00
|
|
|
struct sched_domain *sd, enum cpu_idle_type idle,
|
2009-03-19 04:52:20 +00:00
|
|
|
int *balance)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
int ld_moved, all_pinned = 0, active_balance = 0, sd_idle = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sched_group *group;
|
|
|
|
unsigned long imbalance;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *busiest;
|
2006-12-10 10:20:21 +00:00
|
|
|
unsigned long flags;
|
2009-03-19 04:52:20 +00:00
|
|
|
struct cpumask *cpus = __get_cpu_var(load_balance_tmpmask);
|
2005-09-10 07:26:19 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_setall(cpus);
|
2008-04-05 01:11:11 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
/*
|
|
|
|
* When power savings policy is enabled for the parent domain, idle
|
|
|
|
* sibling can pick up load irrespective of busy siblings. In this case,
|
2007-07-09 16:51:59 +00:00
|
|
|
* let the state of idle sibling percolate up as CPU_IDLE, instead of
|
2007-07-09 16:51:57 +00:00
|
|
|
* portraying it as CPU_NOT_IDLE.
|
2006-10-03 08:14:09 +00:00
|
|
|
*/
|
2007-07-09 16:51:57 +00:00
|
|
|
if (idle != CPU_NOT_IDLE && sd->flags & SD_SHARE_CPUPOWER &&
|
2006-10-03 08:14:09 +00:00
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2005-09-10 07:26:19 +00:00
|
|
|
sd_idle = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(sd, lb_count[idle]);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-09-26 06:30:51 +00:00
|
|
|
redo:
|
2008-06-27 11:41:23 +00:00
|
|
|
update_shares(sd);
|
2006-09-26 06:30:51 +00:00
|
|
|
group = find_busiest_group(sd, this_cpu, &imbalance, idle, &sd_idle,
|
2008-04-05 01:11:11 +00:00
|
|
|
cpus, balance);
|
2006-12-10 10:20:33 +00:00
|
|
|
|
2006-12-10 10:20:35 +00:00
|
|
|
if (*balance == 0)
|
2006-12-10 10:20:33 +00:00
|
|
|
goto out_balanced;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!group) {
|
|
|
|
schedstat_inc(sd, lb_nobusyg[idle]);
|
|
|
|
goto out_balanced;
|
|
|
|
}
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
busiest = find_busiest_queue(group, idle, imbalance, cpus);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!busiest) {
|
|
|
|
schedstat_inc(sd, lb_nobusyq[idle]);
|
|
|
|
goto out_balanced;
|
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:11 +00:00
|
|
|
BUG_ON(busiest == this_rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
schedstat_add(sd, lb_imbalance[idle], imbalance);
|
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
ld_moved = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
if (busiest->nr_running > 1) {
|
|
|
|
/*
|
|
|
|
* Attempt to move tasks. If find_busiest_group has found
|
|
|
|
* an imbalance but busiest->nr_running <= 1, the group is
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
* still unbalanced. ld_moved simply stays zero, so it is
|
2005-04-16 22:20:36 +00:00
|
|
|
* correctly treated as an imbalance.
|
|
|
|
*/
|
2006-12-10 10:20:21 +00:00
|
|
|
local_irq_save(flags);
|
2005-09-10 07:26:18 +00:00
|
|
|
double_rq_lock(this_rq, busiest);
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
ld_moved = move_tasks(this_rq, this_cpu, busiest,
|
2006-07-03 07:25:40 +00:00
|
|
|
imbalance, sd, idle, &all_pinned);
|
2005-09-10 07:26:18 +00:00
|
|
|
double_rq_unlock(this_rq, busiest);
|
2006-12-10 10:20:21 +00:00
|
|
|
local_irq_restore(flags);
|
2005-06-25 21:57:07 +00:00
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
/*
|
|
|
|
* some other cpu did the load balance for us.
|
|
|
|
*/
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
if (ld_moved && this_cpu != smp_processor_id())
|
2007-05-08 07:32:51 +00:00
|
|
|
resched_cpu(this_cpu);
|
|
|
|
|
2005-06-25 21:57:07 +00:00
|
|
|
/* All tasks on this runqueue were pinned by CPU affinity */
|
2006-09-26 06:30:51 +00:00
|
|
|
if (unlikely(all_pinned)) {
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_clear_cpu(cpu_of(busiest), cpus);
|
|
|
|
if (!cpumask_empty(cpus))
|
2006-09-26 06:30:51 +00:00
|
|
|
goto redo;
|
2005-06-25 21:57:07 +00:00
|
|
|
goto out_balanced;
|
2006-09-26 06:30:51 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2005-06-25 21:57:07 +00:00
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
if (!ld_moved) {
|
2005-04-16 22:20:36 +00:00
|
|
|
schedstat_inc(sd, lb_failed[idle]);
|
|
|
|
sd->nr_balance_failed++;
|
|
|
|
|
|
|
|
if (unlikely(sd->nr_balance_failed > sd->cache_nice_tries+2)) {
|
|
|
|
|
2006-12-10 10:20:21 +00:00
|
|
|
spin_lock_irqsave(&busiest->lock, flags);
|
2005-09-10 07:26:21 +00:00
|
|
|
|
|
|
|
/* don't kick the migration_thread, if the curr
|
|
|
|
* task on busiest cpu can't be moved to this_cpu
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_test_cpu(this_cpu,
|
|
|
|
&busiest->curr->cpus_allowed)) {
|
2006-12-10 10:20:21 +00:00
|
|
|
spin_unlock_irqrestore(&busiest->lock, flags);
|
2005-09-10 07:26:21 +00:00
|
|
|
all_pinned = 1;
|
|
|
|
goto out_one_pinned;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!busiest->active_balance) {
|
|
|
|
busiest->active_balance = 1;
|
|
|
|
busiest->push_cpu = this_cpu;
|
2005-06-25 21:57:07 +00:00
|
|
|
active_balance = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-12-10 10:20:21 +00:00
|
|
|
spin_unlock_irqrestore(&busiest->lock, flags);
|
2005-06-25 21:57:07 +00:00
|
|
|
if (active_balance)
|
2005-04-16 22:20:36 +00:00
|
|
|
wake_up_process(busiest->migration_thread);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We've kicked active balancing, reset the failure
|
|
|
|
* counter.
|
|
|
|
*/
|
2005-06-25 21:57:09 +00:00
|
|
|
sd->nr_balance_failed = sd->cache_nice_tries+1;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2005-06-25 21:57:07 +00:00
|
|
|
} else
|
2005-04-16 22:20:36 +00:00
|
|
|
sd->nr_balance_failed = 0;
|
|
|
|
|
2005-06-25 21:57:07 +00:00
|
|
|
if (likely(!active_balance)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/* We were unbalanced, so reset the balancing interval */
|
|
|
|
sd->balance_interval = sd->min_interval;
|
2005-06-25 21:57:07 +00:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* If we've begun active balancing, start to back off. This
|
|
|
|
* case may not be covered by the all_pinned logic if there
|
|
|
|
* is only 1 task on the busy runqueue (because we don't call
|
|
|
|
* move_tasks).
|
|
|
|
*/
|
|
|
|
if (sd->balance_interval < sd->max_interval)
|
|
|
|
sd->balance_interval *= 2;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
if (!ld_moved && !sd_idle && sd->flags & SD_SHARE_CPUPOWER &&
|
2006-10-03 08:14:09 +00:00
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2008-06-27 11:41:14 +00:00
|
|
|
ld_moved = -1;
|
|
|
|
|
|
|
|
goto out;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
out_balanced:
|
|
|
|
schedstat_inc(sd, lb_balanced[idle]);
|
|
|
|
|
2005-06-25 21:57:08 +00:00
|
|
|
sd->nr_balance_failed = 0;
|
2005-09-10 07:26:21 +00:00
|
|
|
|
|
|
|
out_one_pinned:
|
2005-04-16 22:20:36 +00:00
|
|
|
/* tune up the balancing interval */
|
2005-06-25 21:57:30 +00:00
|
|
|
if ((all_pinned && sd->balance_interval < MAX_PINNED_INTERVAL) ||
|
|
|
|
(sd->balance_interval < sd->max_interval))
|
2005-04-16 22:20:36 +00:00
|
|
|
sd->balance_interval *= 2;
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER &&
|
2006-10-03 08:14:09 +00:00
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2008-06-27 11:41:14 +00:00
|
|
|
ld_moved = -1;
|
|
|
|
else
|
|
|
|
ld_moved = 0;
|
|
|
|
out:
|
2008-06-27 11:41:23 +00:00
|
|
|
if (ld_moved)
|
|
|
|
update_shares(sd);
|
2008-06-27 11:41:14 +00:00
|
|
|
return ld_moved;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check this_cpu to ensure it is balanced within domain. Attempt to move
|
|
|
|
* tasks if there is an imbalance.
|
|
|
|
*
|
2007-07-09 16:51:57 +00:00
|
|
|
* Called from schedule when this_rq is about to become idle (CPU_NEWLY_IDLE).
|
2005-04-16 22:20:36 +00:00
|
|
|
* this_rq is locked.
|
|
|
|
*/
|
2006-07-03 07:25:40 +00:00
|
|
|
static int
|
2009-03-19 04:52:20 +00:00
|
|
|
load_balance_newidle(int this_cpu, struct rq *this_rq, struct sched_domain *sd)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sched_group *group;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *busiest = NULL;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long imbalance;
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
int ld_moved = 0;
|
2005-09-10 07:26:19 +00:00
|
|
|
int sd_idle = 0;
|
2007-07-19 19:28:35 +00:00
|
|
|
int all_pinned = 0;
|
2009-03-19 04:52:20 +00:00
|
|
|
struct cpumask *cpus = __get_cpu_var(load_balance_tmpmask);
|
2008-04-05 01:11:11 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_setall(cpus);
|
2005-09-10 07:26:19 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
/*
|
|
|
|
* When power savings policy is enabled for the parent domain, idle
|
|
|
|
* sibling can pick up load irrespective of busy siblings. In this case,
|
|
|
|
* let the state of idle sibling percolate up as IDLE, instead of
|
2007-07-09 16:51:57 +00:00
|
|
|
* portraying it as CPU_NOT_IDLE.
|
2006-10-03 08:14:09 +00:00
|
|
|
*/
|
|
|
|
if (sd->flags & SD_SHARE_CPUPOWER &&
|
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2005-09-10 07:26:19 +00:00
|
|
|
sd_idle = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(sd, lb_count[CPU_NEWLY_IDLE]);
|
2006-09-26 06:30:51 +00:00
|
|
|
redo:
|
2008-06-27 11:41:24 +00:00
|
|
|
update_shares_locked(this_rq, sd);
|
2007-07-09 16:51:57 +00:00
|
|
|
group = find_busiest_group(sd, this_cpu, &imbalance, CPU_NEWLY_IDLE,
|
2008-04-05 01:11:11 +00:00
|
|
|
&sd_idle, cpus, NULL);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!group) {
|
2007-07-09 16:51:57 +00:00
|
|
|
schedstat_inc(sd, lb_nobusyg[CPU_NEWLY_IDLE]);
|
2005-06-25 21:57:08 +00:00
|
|
|
goto out_balanced;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
busiest = find_busiest_queue(group, CPU_NEWLY_IDLE, imbalance, cpus);
|
2005-06-25 21:57:11 +00:00
|
|
|
if (!busiest) {
|
2007-07-09 16:51:57 +00:00
|
|
|
schedstat_inc(sd, lb_nobusyq[CPU_NEWLY_IDLE]);
|
2005-06-25 21:57:08 +00:00
|
|
|
goto out_balanced;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:11 +00:00
|
|
|
BUG_ON(busiest == this_rq);
|
|
|
|
|
2007-07-09 16:51:57 +00:00
|
|
|
schedstat_add(sd, lb_imbalance[CPU_NEWLY_IDLE], imbalance);
|
2005-09-10 07:26:16 +00:00
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
ld_moved = 0;
|
2005-09-10 07:26:16 +00:00
|
|
|
if (busiest->nr_running > 1) {
|
|
|
|
/* Attempt to move tasks */
|
|
|
|
double_lock_balance(this_rq, busiest);
|
2007-08-09 09:16:51 +00:00
|
|
|
/* this_rq->clock is already updated */
|
|
|
|
update_rq_clock(busiest);
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
ld_moved = move_tasks(this_rq, this_cpu, busiest,
|
2007-07-19 19:28:35 +00:00
|
|
|
imbalance, sd, CPU_NEWLY_IDLE,
|
|
|
|
&all_pinned);
|
2008-08-11 07:30:22 +00:00
|
|
|
double_unlock_balance(this_rq, busiest);
|
2006-09-26 06:30:51 +00:00
|
|
|
|
2007-07-19 19:28:35 +00:00
|
|
|
if (unlikely(all_pinned)) {
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_clear_cpu(cpu_of(busiest), cpus);
|
|
|
|
if (!cpumask_empty(cpus))
|
2006-09-26 06:30:51 +00:00
|
|
|
goto redo;
|
|
|
|
}
|
2005-09-10 07:26:16 +00:00
|
|
|
}
|
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
if (!ld_moved) {
|
2008-12-20 04:36:38 +00:00
|
|
|
int active_balance = 0;
|
2008-12-18 17:56:36 +00:00
|
|
|
|
2007-07-09 16:51:57 +00:00
|
|
|
schedstat_inc(sd, lb_failed[CPU_NEWLY_IDLE]);
|
2006-10-03 08:14:09 +00:00
|
|
|
if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER &&
|
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2005-09-10 07:26:19 +00:00
|
|
|
return -1;
|
2008-12-18 17:56:36 +00:00
|
|
|
|
|
|
|
if (sched_mc_power_savings < POWERSAVINGS_BALANCE_WAKEUP)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
if (sd->nr_balance_failed++ < 2)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The only task running in a non-idle cpu can be moved to this
|
|
|
|
* cpu in an attempt to completely freeup the other CPU
|
|
|
|
* package. The same method used to move task in load_balance()
|
|
|
|
* have been extended for load_balance_newidle() to speedup
|
|
|
|
* consolidation at sched_mc=POWERSAVINGS_BALANCE_WAKEUP (2)
|
|
|
|
*
|
|
|
|
* The package power saving logic comes from
|
|
|
|
* find_busiest_group(). If there are no imbalance, then
|
|
|
|
* f_b_g() will return NULL. However when sched_mc={1,2} then
|
|
|
|
* f_b_g() will select a group from which a running task may be
|
|
|
|
* pulled to this cpu in order to make the other package idle.
|
|
|
|
* If there is no opportunity to make a package idle and if
|
|
|
|
* there are no imbalance, then f_b_g() will return NULL and no
|
|
|
|
* action will be taken in load_balance_newidle().
|
|
|
|
*
|
|
|
|
* Under normal task pull operation due to imbalance, there
|
|
|
|
* will be more than one task in the source run queue and
|
|
|
|
* move_tasks() will succeed. ld_moved will be true and this
|
|
|
|
* active balance code will not be triggered.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Lock busiest in correct order while this_rq is held */
|
|
|
|
double_lock_balance(this_rq, busiest);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* don't kick the migration_thread, if the curr
|
|
|
|
* task on busiest cpu can't be moved to this_cpu
|
|
|
|
*/
|
2009-01-01 02:08:45 +00:00
|
|
|
if (!cpumask_test_cpu(this_cpu, &busiest->curr->cpus_allowed)) {
|
2008-12-18 17:56:36 +00:00
|
|
|
double_unlock_balance(this_rq, busiest);
|
|
|
|
all_pinned = 1;
|
|
|
|
return ld_moved;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!busiest->active_balance) {
|
|
|
|
busiest->active_balance = 1;
|
|
|
|
busiest->push_cpu = this_cpu;
|
|
|
|
active_balance = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
double_unlock_balance(this_rq, busiest);
|
2009-01-07 14:28:57 +00:00
|
|
|
/*
|
|
|
|
* Should not call ttwu while holding a rq->lock
|
|
|
|
*/
|
|
|
|
spin_unlock(&this_rq->lock);
|
2008-12-18 17:56:36 +00:00
|
|
|
if (active_balance)
|
|
|
|
wake_up_process(busiest->migration_thread);
|
2009-01-07 14:28:57 +00:00
|
|
|
spin_lock(&this_rq->lock);
|
2008-12-18 17:56:36 +00:00
|
|
|
|
2005-09-10 07:26:19 +00:00
|
|
|
} else
|
2005-06-25 21:57:08 +00:00
|
|
|
sd->nr_balance_failed = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-06-27 11:41:24 +00:00
|
|
|
update_shares_locked(this_rq, sd);
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
return ld_moved;
|
2005-06-25 21:57:08 +00:00
|
|
|
|
|
|
|
out_balanced:
|
2007-07-09 16:51:57 +00:00
|
|
|
schedstat_inc(sd, lb_balanced[CPU_NEWLY_IDLE]);
|
2006-07-03 07:25:40 +00:00
|
|
|
if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER &&
|
2006-10-03 08:14:09 +00:00
|
|
|
!test_sd_parent(sd, SD_POWERSAVINGS_BALANCE))
|
2005-09-10 07:26:19 +00:00
|
|
|
return -1;
|
2005-06-25 21:57:08 +00:00
|
|
|
sd->nr_balance_failed = 0;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-06-25 21:57:08 +00:00
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* idle_balance is called by schedule() if this_cpu is about to become
|
|
|
|
* idle. Attempts to pull tasks from other CPUs.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static void idle_balance(int this_cpu, struct rq *this_rq)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sched_domain *sd;
|
2008-12-08 15:22:49 +00:00
|
|
|
int pulled_task = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
unsigned long next_balance = jiffies + HZ;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
for_each_domain(this_cpu, sd) {
|
2007-06-24 00:16:33 +00:00
|
|
|
unsigned long interval;
|
|
|
|
|
|
|
|
if (!(sd->flags & SD_LOAD_BALANCE))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (sd->flags & SD_BALANCE_NEWIDLE)
|
2006-07-03 07:25:40 +00:00
|
|
|
/* If we've pulled tasks over stop searching: */
|
2008-04-05 01:11:11 +00:00
|
|
|
pulled_task = load_balance_newidle(this_cpu, this_rq,
|
2009-03-19 04:52:20 +00:00
|
|
|
sd);
|
2007-06-24 00:16:33 +00:00
|
|
|
|
|
|
|
interval = msecs_to_jiffies(sd->balance_interval);
|
|
|
|
if (time_after(next_balance, sd->last_balance + interval))
|
|
|
|
next_balance = sd->last_balance + interval;
|
|
|
|
if (pulled_task)
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
if (pulled_task || time_after(jiffies, this_rq->next_balance)) {
|
2006-12-10 10:20:27 +00:00
|
|
|
/*
|
|
|
|
* We are going idle. next_balance may be set based on
|
|
|
|
* a busy processor. So reset next_balance.
|
|
|
|
*/
|
|
|
|
this_rq->next_balance = next_balance;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* active_load_balance is run by migration threads. It pushes running tasks
|
|
|
|
* off the busiest CPU onto idle CPUs. It requires at least 1 task to be
|
|
|
|
* running on each physical CPU where possible, and avoids physical /
|
|
|
|
* logical imbalances.
|
|
|
|
*
|
|
|
|
* Called with busiest_rq locked.
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static void active_load_balance(struct rq *busiest_rq, int busiest_cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2005-06-25 21:57:09 +00:00
|
|
|
int target_cpu = busiest_rq->push_cpu;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct sched_domain *sd;
|
|
|
|
struct rq *target_rq;
|
2005-06-25 21:57:09 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
/* Is there any task to move? */
|
2005-06-25 21:57:09 +00:00
|
|
|
if (busiest_rq->nr_running <= 1)
|
|
|
|
return;
|
|
|
|
|
|
|
|
target_rq = cpu_rq(target_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2005-06-25 21:57:09 +00:00
|
|
|
* This condition is "impossible", if it occurs
|
2007-12-05 14:46:09 +00:00
|
|
|
* we need to fix it. Originally reported by
|
2005-06-25 21:57:09 +00:00
|
|
|
* Bjorn Helgaas on a 128-cpu setup.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2005-06-25 21:57:09 +00:00
|
|
|
BUG_ON(busiest_rq == target_rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-06-25 21:57:09 +00:00
|
|
|
/* move a task from busiest_rq to target_rq */
|
|
|
|
double_lock_balance(busiest_rq, target_rq);
|
2007-08-09 09:16:51 +00:00
|
|
|
update_rq_clock(busiest_rq);
|
|
|
|
update_rq_clock(target_rq);
|
2005-06-25 21:57:09 +00:00
|
|
|
|
|
|
|
/* Search for an sd spanning us and the target CPU. */
|
2006-06-27 09:54:28 +00:00
|
|
|
for_each_domain(target_cpu, sd) {
|
2005-06-25 21:57:09 +00:00
|
|
|
if ((sd->flags & SD_LOAD_BALANCE) &&
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_test_cpu(busiest_cpu, sched_domain_span(sd)))
|
2005-06-25 21:57:09 +00:00
|
|
|
break;
|
2006-06-27 09:54:28 +00:00
|
|
|
}
|
2005-06-25 21:57:09 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
if (likely(sd)) {
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(sd, alb_count);
|
2005-06-25 21:57:09 +00:00
|
|
|
|
sched: simplify move_tasks()
The move_tasks() function is currently multiplexed with two distinct
capabilities:
1. attempt to move a specified amount of weighted load from one run
queue to another; and
2. attempt to move a specified number of tasks from one run queue to
another.
The first of these capabilities is used in two places, load_balance()
and load_balance_idle(), and in both of these cases the return value of
move_tasks() is used purely to decide if tasks/load were moved and no
notice of the actual number of tasks moved is taken.
The second capability is used in exactly one place,
active_load_balance(), to attempt to move exactly one task and, as
before, the return value is only used as an indicator of success or failure.
This multiplexing of sched_task() was introduced, by me, as part of the
smpnice patches and was motivated by the fact that the alternative, one
function to move specified load and one to move a single task, would
have led to two functions of roughly the same complexity as the old
move_tasks() (or the new balance_tasks()). However, the new modular
design of the new CFS scheduler allows a simpler solution to be adopted
and this patch addresses that solution by:
1. adding a new function, move_one_task(), to be used by
active_load_balance(); and
2. making move_tasks() a single purpose function that tries to move a
specified weighted load and returns 1 for success and 0 for failure.
One of the consequences of these changes is that neither move_one_task()
or the new move_tasks() care how many tasks sched_class.load_balance()
moves and this enables its interface to be simplified by returning the
amount of load moved as its result and removing the load_moved pointer
from the argument list. This helps simplify the new move_tasks() and
slightly reduces the amount of work done in each of
sched_class.load_balance()'s implementations.
Further simplification, e.g. changes to balance_tasks(), are possible
but (slightly) complicated by the special needs of load_balance_fair()
so I've left them to a later patch (if this one gets accepted).
NB Since move_tasks() gets called with two run queue locks held even
small reductions in overhead are worthwhile.
[ mingo@elte.hu ]
this change also reduces code size nicely:
text data bss dec hex filename
39216 3618 24 42858 a76a sched.o.before
39173 3618 24 42815 a73f sched.o.after
Signed-off-by: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2007-08-09 09:16:46 +00:00
|
|
|
if (move_one_task(target_rq, target_cpu, busiest_rq,
|
|
|
|
sd, CPU_IDLE))
|
2006-07-03 07:25:40 +00:00
|
|
|
schedstat_inc(sd, alb_pushed);
|
|
|
|
else
|
|
|
|
schedstat_inc(sd, alb_failed);
|
|
|
|
}
|
2008-08-11 07:30:22 +00:00
|
|
|
double_unlock_balance(busiest_rq, target_rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
#ifdef CONFIG_NO_HZ
|
|
|
|
static struct {
|
|
|
|
atomic_t load_balancer;
|
2008-11-24 16:05:09 +00:00
|
|
|
cpumask_var_t cpu_mask;
|
2009-04-14 04:55:30 +00:00
|
|
|
cpumask_var_t ilb_grp_nohz_mask;
|
2007-05-08 07:32:51 +00:00
|
|
|
} nohz ____cacheline_aligned = {
|
|
|
|
.load_balancer = ATOMIC_INIT(-1),
|
|
|
|
};
|
|
|
|
|
2009-04-16 06:46:41 +00:00
|
|
|
int get_nohz_load_balancer(void)
|
|
|
|
{
|
|
|
|
return atomic_read(&nohz.load_balancer);
|
|
|
|
}
|
|
|
|
|
2009-04-14 04:55:30 +00:00
|
|
|
#if defined(CONFIG_SCHED_MC) || defined(CONFIG_SCHED_SMT)
|
|
|
|
/**
|
|
|
|
* lowest_flag_domain - Return lowest sched_domain containing flag.
|
|
|
|
* @cpu: The cpu whose lowest level of sched domain is to
|
|
|
|
* be returned.
|
|
|
|
* @flag: The flag to check for the lowest sched_domain
|
|
|
|
* for the given cpu.
|
|
|
|
*
|
|
|
|
* Returns the lowest sched_domain of a cpu which contains the given flag.
|
|
|
|
*/
|
|
|
|
static inline struct sched_domain *lowest_flag_domain(int cpu, int flag)
|
|
|
|
{
|
|
|
|
struct sched_domain *sd;
|
|
|
|
|
|
|
|
for_each_domain(cpu, sd)
|
|
|
|
if (sd && (sd->flags & flag))
|
|
|
|
break;
|
|
|
|
|
|
|
|
return sd;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* for_each_flag_domain - Iterates over sched_domains containing the flag.
|
|
|
|
* @cpu: The cpu whose domains we're iterating over.
|
|
|
|
* @sd: variable holding the value of the power_savings_sd
|
|
|
|
* for cpu.
|
|
|
|
* @flag: The flag to filter the sched_domains to be iterated.
|
|
|
|
*
|
|
|
|
* Iterates over all the scheduler domains for a given cpu that has the 'flag'
|
|
|
|
* set, starting from the lowest sched_domain to the highest.
|
|
|
|
*/
|
|
|
|
#define for_each_flag_domain(cpu, sd, flag) \
|
|
|
|
for (sd = lowest_flag_domain(cpu, flag); \
|
|
|
|
(sd && (sd->flags & flag)); sd = sd->parent)
|
|
|
|
|
|
|
|
/**
|
|
|
|
* is_semi_idle_group - Checks if the given sched_group is semi-idle.
|
|
|
|
* @ilb_group: group to be checked for semi-idleness
|
|
|
|
*
|
|
|
|
* Returns: 1 if the group is semi-idle. 0 otherwise.
|
|
|
|
*
|
|
|
|
* We define a sched_group to be semi idle if it has atleast one idle-CPU
|
|
|
|
* and atleast one non-idle CPU. This helper function checks if the given
|
|
|
|
* sched_group is semi-idle or not.
|
|
|
|
*/
|
|
|
|
static inline int is_semi_idle_group(struct sched_group *ilb_group)
|
|
|
|
{
|
|
|
|
cpumask_and(nohz.ilb_grp_nohz_mask, nohz.cpu_mask,
|
|
|
|
sched_group_cpus(ilb_group));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A sched_group is semi-idle when it has atleast one busy cpu
|
|
|
|
* and atleast one idle cpu.
|
|
|
|
*/
|
|
|
|
if (cpumask_empty(nohz.ilb_grp_nohz_mask))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (cpumask_equal(nohz.ilb_grp_nohz_mask, sched_group_cpus(ilb_group)))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
/**
|
|
|
|
* find_new_ilb - Finds the optimum idle load balancer for nomination.
|
|
|
|
* @cpu: The cpu which is nominating a new idle_load_balancer.
|
|
|
|
*
|
|
|
|
* Returns: Returns the id of the idle load balancer if it exists,
|
|
|
|
* Else, returns >= nr_cpu_ids.
|
|
|
|
*
|
|
|
|
* This algorithm picks the idle load balancer such that it belongs to a
|
|
|
|
* semi-idle powersavings sched_domain. The idea is to try and avoid
|
|
|
|
* completely idle packages/cores just for the purpose of idle load balancing
|
|
|
|
* when there are other idle cpu's which are better suited for that job.
|
|
|
|
*/
|
|
|
|
static int find_new_ilb(int cpu)
|
|
|
|
{
|
|
|
|
struct sched_domain *sd;
|
|
|
|
struct sched_group *ilb_group;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Have idle load balancer selection from semi-idle packages only
|
|
|
|
* when power-aware load balancing is enabled
|
|
|
|
*/
|
|
|
|
if (!(sched_smt_power_savings || sched_mc_power_savings))
|
|
|
|
goto out_done;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Optimize for the case when we have no idle CPUs or only one
|
|
|
|
* idle CPU. Don't walk the sched_domain hierarchy in such cases
|
|
|
|
*/
|
|
|
|
if (cpumask_weight(nohz.cpu_mask) < 2)
|
|
|
|
goto out_done;
|
|
|
|
|
|
|
|
for_each_flag_domain(cpu, sd, SD_POWERSAVINGS_BALANCE) {
|
|
|
|
ilb_group = sd->groups;
|
|
|
|
|
|
|
|
do {
|
|
|
|
if (is_semi_idle_group(ilb_group))
|
|
|
|
return cpumask_first(nohz.ilb_grp_nohz_mask);
|
|
|
|
|
|
|
|
ilb_group = ilb_group->next;
|
|
|
|
|
|
|
|
} while (ilb_group != sd->groups);
|
|
|
|
}
|
|
|
|
|
|
|
|
out_done:
|
|
|
|
return cpumask_first(nohz.cpu_mask);
|
|
|
|
}
|
|
|
|
#else /* (CONFIG_SCHED_MC || CONFIG_SCHED_SMT) */
|
|
|
|
static inline int find_new_ilb(int call_cpu)
|
|
|
|
{
|
2009-04-21 03:10:49 +00:00
|
|
|
return cpumask_first(nohz.cpu_mask);
|
2009-04-14 04:55:30 +00:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2006-12-10 10:20:22 +00:00
|
|
|
/*
|
2007-05-08 07:32:51 +00:00
|
|
|
* This routine will try to nominate the ilb (idle load balancing)
|
|
|
|
* owner among the cpus whose ticks are stopped. ilb owner will do the idle
|
|
|
|
* load balancing on behalf of all those cpus. If all the cpus in the system
|
|
|
|
* go into this tickless mode, then there will be no ilb owner (as there is
|
|
|
|
* no need for one) and all the cpus will sleep till the next wakeup event
|
|
|
|
* arrives...
|
|
|
|
*
|
|
|
|
* For the ilb owner, tick is not stopped. And this tick will be used
|
|
|
|
* for idle load balancing. ilb owner will still be part of
|
|
|
|
* nohz.cpu_mask..
|
2006-12-10 10:20:22 +00:00
|
|
|
*
|
2007-05-08 07:32:51 +00:00
|
|
|
* While stopping the tick, this cpu will become the ilb owner if there
|
|
|
|
* is no other owner. And will be the owner till that cpu becomes busy
|
|
|
|
* or if all cpus in the system stop their ticks at which point
|
|
|
|
* there is no need for ilb owner.
|
|
|
|
*
|
|
|
|
* When the ilb owner becomes busy, it nominates another owner, during the
|
|
|
|
* next busy scheduler_tick()
|
|
|
|
*/
|
|
|
|
int select_nohz_load_balancer(int stop_tick)
|
|
|
|
{
|
|
|
|
int cpu = smp_processor_id();
|
|
|
|
|
|
|
|
if (stop_tick) {
|
|
|
|
cpu_rq(cpu)->in_nohz_recently = 1;
|
|
|
|
|
2009-02-04 19:59:44 +00:00
|
|
|
if (!cpu_active(cpu)) {
|
|
|
|
if (atomic_read(&nohz.load_balancer) != cpu)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we are going offline and still the leader,
|
|
|
|
* give up!
|
|
|
|
*/
|
2007-05-08 07:32:51 +00:00
|
|
|
if (atomic_cmpxchg(&nohz.load_balancer, cpu, -1) != cpu)
|
|
|
|
BUG();
|
2009-02-04 19:59:44 +00:00
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-02-04 19:59:44 +00:00
|
|
|
cpumask_set_cpu(cpu, nohz.cpu_mask);
|
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
/* time for ilb owner also to sleep */
|
2008-11-24 16:05:09 +00:00
|
|
|
if (cpumask_weight(nohz.cpu_mask) == num_online_cpus()) {
|
2007-05-08 07:32:51 +00:00
|
|
|
if (atomic_read(&nohz.load_balancer) == cpu)
|
|
|
|
atomic_set(&nohz.load_balancer, -1);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (atomic_read(&nohz.load_balancer) == -1) {
|
|
|
|
/* make me the ilb owner */
|
|
|
|
if (atomic_cmpxchg(&nohz.load_balancer, -1, cpu) == -1)
|
|
|
|
return 1;
|
2009-04-14 04:55:35 +00:00
|
|
|
} else if (atomic_read(&nohz.load_balancer) == cpu) {
|
|
|
|
int new_ilb;
|
|
|
|
|
|
|
|
if (!(sched_smt_power_savings ||
|
|
|
|
sched_mc_power_savings))
|
|
|
|
return 1;
|
|
|
|
/*
|
|
|
|
* Check to see if there is a more power-efficient
|
|
|
|
* ilb.
|
|
|
|
*/
|
|
|
|
new_ilb = find_new_ilb(cpu);
|
|
|
|
if (new_ilb < nr_cpu_ids && new_ilb != cpu) {
|
|
|
|
atomic_set(&nohz.load_balancer, -1);
|
|
|
|
resched_cpu(new_ilb);
|
|
|
|
return 0;
|
|
|
|
}
|
2007-05-08 07:32:51 +00:00
|
|
|
return 1;
|
2009-04-14 04:55:35 +00:00
|
|
|
}
|
2007-05-08 07:32:51 +00:00
|
|
|
} else {
|
2008-11-24 16:05:09 +00:00
|
|
|
if (!cpumask_test_cpu(cpu, nohz.cpu_mask))
|
2007-05-08 07:32:51 +00:00
|
|
|
return 0;
|
|
|
|
|
2008-11-24 16:05:09 +00:00
|
|
|
cpumask_clear_cpu(cpu, nohz.cpu_mask);
|
2007-05-08 07:32:51 +00:00
|
|
|
|
|
|
|
if (atomic_read(&nohz.load_balancer) == cpu)
|
|
|
|
if (atomic_cmpxchg(&nohz.load_balancer, cpu, -1) != cpu)
|
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static DEFINE_SPINLOCK(balancing);
|
|
|
|
|
|
|
|
/*
|
2006-12-10 10:20:22 +00:00
|
|
|
* It checks each scheduling domain to see if it is due to be balanced,
|
|
|
|
* and initiates a balancing operation if so.
|
|
|
|
*
|
|
|
|
* Balancing parameters are set up in arch_init_sched_domains.
|
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static void rebalance_domains(int cpu, enum cpu_idle_type idle)
|
2006-12-10 10:20:22 +00:00
|
|
|
{
|
2007-05-08 07:32:51 +00:00
|
|
|
int balance = 1;
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2006-12-10 10:20:22 +00:00
|
|
|
unsigned long interval;
|
|
|
|
struct sched_domain *sd;
|
2007-05-08 07:32:51 +00:00
|
|
|
/* Earliest time when we have to do rebalance again */
|
2006-12-10 10:20:25 +00:00
|
|
|
unsigned long next_balance = jiffies + 60*HZ;
|
2007-08-23 13:18:02 +00:00
|
|
|
int update_next_balance = 0;
|
2008-05-12 19:21:15 +00:00
|
|
|
int need_serialize;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
for_each_domain(cpu, sd) {
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!(sd->flags & SD_LOAD_BALANCE))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
interval = sd->balance_interval;
|
2007-07-09 16:51:57 +00:00
|
|
|
if (idle != CPU_IDLE)
|
2005-04-16 22:20:36 +00:00
|
|
|
interval *= sd->busy_factor;
|
|
|
|
|
|
|
|
/* scale ms to jiffies */
|
|
|
|
interval = msecs_to_jiffies(interval);
|
|
|
|
if (unlikely(!interval))
|
|
|
|
interval = 1;
|
2007-07-09 16:51:59 +00:00
|
|
|
if (interval > HZ*NR_CPUS/10)
|
|
|
|
interval = HZ*NR_CPUS/10;
|
|
|
|
|
2008-05-12 19:21:15 +00:00
|
|
|
need_serialize = sd->flags & SD_SERIALIZE;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-05-12 19:21:15 +00:00
|
|
|
if (need_serialize) {
|
2006-12-10 10:20:29 +00:00
|
|
|
if (!spin_trylock(&balancing))
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2006-12-10 10:20:25 +00:00
|
|
|
if (time_after_eq(jiffies, sd->last_balance + interval)) {
|
2009-03-19 04:52:20 +00:00
|
|
|
if (load_balance(cpu, rq, sd, idle, &balance)) {
|
2005-09-10 07:26:21 +00:00
|
|
|
/*
|
|
|
|
* We've pulled tasks over so either we're no
|
2005-09-10 07:26:19 +00:00
|
|
|
* longer idle, or one of our SMT siblings is
|
|
|
|
* not idle.
|
|
|
|
*/
|
2007-07-09 16:51:57 +00:00
|
|
|
idle = CPU_NOT_IDLE;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-12-10 10:20:27 +00:00
|
|
|
sd->last_balance = jiffies;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-05-12 19:21:15 +00:00
|
|
|
if (need_serialize)
|
2006-12-10 10:20:29 +00:00
|
|
|
spin_unlock(&balancing);
|
|
|
|
out:
|
2007-08-23 13:18:02 +00:00
|
|
|
if (time_after(next_balance, sd->last_balance + interval)) {
|
2006-12-10 10:20:25 +00:00
|
|
|
next_balance = sd->last_balance + interval;
|
2007-08-23 13:18:02 +00:00
|
|
|
update_next_balance = 1;
|
|
|
|
}
|
2006-12-10 10:20:33 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Stop the load balance at this level. There is another
|
|
|
|
* CPU in our sched group which is doing load balancing more
|
|
|
|
* actively.
|
|
|
|
*/
|
|
|
|
if (!balance)
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-08-23 13:18:02 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* next_balance will be updated only when there is a need.
|
|
|
|
* When the cpu is attached to null domain for ex, it will not be
|
|
|
|
* updated.
|
|
|
|
*/
|
|
|
|
if (likely(update_next_balance))
|
|
|
|
rq->next_balance = next_balance;
|
2007-05-08 07:32:51 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* run_rebalance_domains is triggered when needed from the scheduler tick.
|
|
|
|
* In CONFIG_NO_HZ case, the idle load balance owner will do the
|
|
|
|
* rebalancing for all the cpus for whom scheduler ticks are stopped.
|
|
|
|
*/
|
|
|
|
static void run_rebalance_domains(struct softirq_action *h)
|
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
int this_cpu = smp_processor_id();
|
|
|
|
struct rq *this_rq = cpu_rq(this_cpu);
|
|
|
|
enum cpu_idle_type idle = this_rq->idle_at_tick ?
|
|
|
|
CPU_IDLE : CPU_NOT_IDLE;
|
2007-05-08 07:32:51 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
rebalance_domains(this_cpu, idle);
|
2007-05-08 07:32:51 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_NO_HZ
|
|
|
|
/*
|
|
|
|
* If this cpu is the owner for idle load balancing, then do the
|
|
|
|
* balancing on behalf of the other idle cpus whose ticks are
|
|
|
|
* stopped.
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
if (this_rq->idle_at_tick &&
|
|
|
|
atomic_read(&nohz.load_balancer) == this_cpu) {
|
2007-05-08 07:32:51 +00:00
|
|
|
struct rq *rq;
|
|
|
|
int balance_cpu;
|
|
|
|
|
2008-11-24 16:05:09 +00:00
|
|
|
for_each_cpu(balance_cpu, nohz.cpu_mask) {
|
|
|
|
if (balance_cpu == this_cpu)
|
|
|
|
continue;
|
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
/*
|
|
|
|
* If this cpu gets work to do, stop the load balancing
|
|
|
|
* work being done for other cpus. Next load
|
|
|
|
* balancing owner will pick it up.
|
|
|
|
*/
|
|
|
|
if (need_resched())
|
|
|
|
break;
|
|
|
|
|
2007-08-12 16:08:19 +00:00
|
|
|
rebalance_domains(balance_cpu, CPU_IDLE);
|
2007-05-08 07:32:51 +00:00
|
|
|
|
|
|
|
rq = cpu_rq(balance_cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (time_after(this_rq->next_balance, rq->next_balance))
|
|
|
|
this_rq->next_balance = rq->next_balance;
|
2007-05-08 07:32:51 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
sched: don't rebalance if attached on NULL domain
Impact: fix function graph trace hang / drop pointless softirq on UP
While debugging a function graph trace hang on an old PII, I saw
that it consumed most of its time on the timer interrupt. And
the domain rebalancing softirq was the most concerned.
The timer interrupt calls trigger_load_balance() which will
decide if it is worth to schedule a rebalancing softirq.
In case of builtin UP kernel, no problem arises because there is
no domain question.
In case of builtin SMP kernel running on an SMP box, still no
problem, the softirq will be raised each time we reach the
next_balance time.
In case of builtin SMP kernel running on a UP box (most distros
provide default SMP kernels, whatever the box you have), then
the CPU is attached to the NULL sched domain. So a kind of
unexpected behaviour happen:
trigger_load_balance() -> raises the rebalancing softirq later
on softirq: run_rebalance_domains() -> rebalance_domains() where
the for_each_domain(cpu, sd) is not taken because of the NULL
domain we are attached at. Which means rq->next_balance is never
updated. So on the next timer tick, we will enter
trigger_load_balance() which will always reschedule() the
rebalacing softirq:
if (time_after_eq(jiffies, rq->next_balance))
raise_softirq(SCHED_SOFTIRQ);
So for each tick, we process this pointless softirq.
This patch fixes it by checking if we are attached to the null
domain before raising the softirq, another possible fix would be
to set the maximal possible JIFFIES value to rq->next_balance if
we are attached to the NULL domain.
v2: build fix on UP
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Peter Zijlstra <peterz@infradead.org>
LKML-Reference: <49af242d.1c07d00a.32d5.ffffc019@mx.google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-03-05 00:27:02 +00:00
|
|
|
static inline int on_null_domain(int cpu)
|
|
|
|
{
|
|
|
|
return !rcu_dereference(cpu_rq(cpu)->sd);
|
|
|
|
}
|
|
|
|
|
2007-05-08 07:32:51 +00:00
|
|
|
/*
|
|
|
|
* Trigger the SCHED_SOFTIRQ if it is time to do periodic load balancing.
|
|
|
|
*
|
|
|
|
* In case of CONFIG_NO_HZ, this is the place where we nominate a new
|
|
|
|
* idle load balancing owner or decide to stop the periodic load balancing,
|
|
|
|
* if the whole system is idle.
|
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static inline void trigger_load_balance(struct rq *rq, int cpu)
|
2007-05-08 07:32:51 +00:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NO_HZ
|
|
|
|
/*
|
|
|
|
* If we were in the nohz mode recently and busy at the current
|
|
|
|
* scheduler tick, then check if we need to nominate new idle
|
|
|
|
* load balancer.
|
|
|
|
*/
|
|
|
|
if (rq->in_nohz_recently && !rq->idle_at_tick) {
|
|
|
|
rq->in_nohz_recently = 0;
|
|
|
|
|
|
|
|
if (atomic_read(&nohz.load_balancer) == cpu) {
|
2008-11-24 16:05:09 +00:00
|
|
|
cpumask_clear_cpu(cpu, nohz.cpu_mask);
|
2007-05-08 07:32:51 +00:00
|
|
|
atomic_set(&nohz.load_balancer, -1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (atomic_read(&nohz.load_balancer) == -1) {
|
2009-04-14 04:55:30 +00:00
|
|
|
int ilb = find_new_ilb(cpu);
|
2007-05-08 07:32:51 +00:00
|
|
|
|
2008-04-05 01:11:04 +00:00
|
|
|
if (ilb < nr_cpu_ids)
|
2007-05-08 07:32:51 +00:00
|
|
|
resched_cpu(ilb);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this cpu is idle and doing idle load balancing for all the
|
|
|
|
* cpus with ticks stopped, is it time for that to stop?
|
|
|
|
*/
|
|
|
|
if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) == cpu &&
|
2008-11-24 16:05:09 +00:00
|
|
|
cpumask_weight(nohz.cpu_mask) == num_online_cpus()) {
|
2007-05-08 07:32:51 +00:00
|
|
|
resched_cpu(cpu);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this cpu is idle and the idle load balancing is done by
|
|
|
|
* someone else, then no need raise the SCHED_SOFTIRQ
|
|
|
|
*/
|
|
|
|
if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) != cpu &&
|
2008-11-24 16:05:09 +00:00
|
|
|
cpumask_test_cpu(cpu, nohz.cpu_mask))
|
2007-05-08 07:32:51 +00:00
|
|
|
return;
|
|
|
|
#endif
|
sched: don't rebalance if attached on NULL domain
Impact: fix function graph trace hang / drop pointless softirq on UP
While debugging a function graph trace hang on an old PII, I saw
that it consumed most of its time on the timer interrupt. And
the domain rebalancing softirq was the most concerned.
The timer interrupt calls trigger_load_balance() which will
decide if it is worth to schedule a rebalancing softirq.
In case of builtin UP kernel, no problem arises because there is
no domain question.
In case of builtin SMP kernel running on an SMP box, still no
problem, the softirq will be raised each time we reach the
next_balance time.
In case of builtin SMP kernel running on a UP box (most distros
provide default SMP kernels, whatever the box you have), then
the CPU is attached to the NULL sched domain. So a kind of
unexpected behaviour happen:
trigger_load_balance() -> raises the rebalancing softirq later
on softirq: run_rebalance_domains() -> rebalance_domains() where
the for_each_domain(cpu, sd) is not taken because of the NULL
domain we are attached at. Which means rq->next_balance is never
updated. So on the next timer tick, we will enter
trigger_load_balance() which will always reschedule() the
rebalacing softirq:
if (time_after_eq(jiffies, rq->next_balance))
raise_softirq(SCHED_SOFTIRQ);
So for each tick, we process this pointless softirq.
This patch fixes it by checking if we are attached to the null
domain before raising the softirq, another possible fix would be
to set the maximal possible JIFFIES value to rq->next_balance if
we are attached to the NULL domain.
v2: build fix on UP
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Peter Zijlstra <peterz@infradead.org>
LKML-Reference: <49af242d.1c07d00a.32d5.ffffc019@mx.google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-03-05 00:27:02 +00:00
|
|
|
/* Don't need to rebalance while attached to NULL domain */
|
|
|
|
if (time_after_eq(jiffies, rq->next_balance) &&
|
|
|
|
likely(!on_null_domain(cpu)))
|
2007-05-08 07:32:51 +00:00
|
|
|
raise_softirq(SCHED_SOFTIRQ);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* on UP we do not need to balance between CPUs:
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static inline void idle_balance(int cpu, struct rq *rq)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
DEFINE_PER_CPU(struct kernel_stat, kstat);
|
|
|
|
|
|
|
|
EXPORT_PER_CPU_SYMBOL(kstat);
|
|
|
|
|
|
|
|
/*
|
2009-03-31 07:56:03 +00:00
|
|
|
* Return any ns on the sched_clock that have not yet been accounted in
|
timers: fix itimer/many thread hang
Overview
This patch reworks the handling of POSIX CPU timers, including the
ITIMER_PROF, ITIMER_VIRT timers and rlimit handling. It was put together
with the help of Roland McGrath, the owner and original writer of this code.
The problem we ran into, and the reason for this rework, has to do with using
a profiling timer in a process with a large number of threads. It appears
that the performance of the old implementation of run_posix_cpu_timers() was
at least O(n*3) (where "n" is the number of threads in a process) or worse.
Everything is fine with an increasing number of threads until the time taken
for that routine to run becomes the same as or greater than the tick time, at
which point things degrade rather quickly.
This patch fixes bug 9906, "Weird hang with NPTL and SIGPROF."
Code Changes
This rework corrects the implementation of run_posix_cpu_timers() to make it
run in constant time for a particular machine. (Performance may vary between
one machine and another depending upon whether the kernel is built as single-
or multiprocessor and, in the latter case, depending upon the number of
running processors.) To do this, at each tick we now update fields in
signal_struct as well as task_struct. The run_posix_cpu_timers() function
uses those fields to make its decisions.
We define a new structure, "task_cputime," to contain user, system and
scheduler times and use these in appropriate places:
struct task_cputime {
cputime_t utime;
cputime_t stime;
unsigned long long sum_exec_runtime;
};
This is included in the structure "thread_group_cputime," which is a new
substructure of signal_struct and which varies for uniprocessor versus
multiprocessor kernels. For uniprocessor kernels, it uses "task_cputime" as
a simple substructure, while for multiprocessor kernels it is a pointer:
struct thread_group_cputime {
struct task_cputime totals;
};
struct thread_group_cputime {
struct task_cputime *totals;
};
We also add a new task_cputime substructure directly to signal_struct, to
cache the earliest expiration of process-wide timers, and task_cputime also
replaces the it_*_expires fields of task_struct (used for earliest expiration
of thread timers). The "thread_group_cputime" structure contains process-wide
timers that are updated via account_user_time() and friends. In the non-SMP
case the structure is a simple aggregator; unfortunately in the SMP case that
simplicity was not achievable due to cache-line contention between CPUs (in
one measured case performance was actually _worse_ on a 16-cpu system than
the same test on a 4-cpu system, due to this contention). For SMP, the
thread_group_cputime counters are maintained as a per-cpu structure allocated
using alloc_percpu(). The timer functions update only the timer field in
the structure corresponding to the running CPU, obtained using per_cpu_ptr().
We define a set of inline functions in sched.h that we use to maintain the
thread_group_cputime structure and hide the differences between UP and SMP
implementations from the rest of the kernel. The thread_group_cputime_init()
function initializes the thread_group_cputime structure for the given task.
The thread_group_cputime_alloc() is a no-op for UP; for SMP it calls the
out-of-line function thread_group_cputime_alloc_smp() to allocate and fill
in the per-cpu structures and fields. The thread_group_cputime_free()
function, also a no-op for UP, in SMP frees the per-cpu structures. The
thread_group_cputime_clone_thread() function (also a UP no-op) for SMP calls
thread_group_cputime_alloc() if the per-cpu structures haven't yet been
allocated. The thread_group_cputime() function fills the task_cputime
structure it is passed with the contents of the thread_group_cputime fields;
in UP it's that simple but in SMP it must also safely check that tsk->signal
is non-NULL (if it is it just uses the appropriate fields of task_struct) and,
if so, sums the per-cpu values for each online CPU. Finally, the three
functions account_group_user_time(), account_group_system_time() and
account_group_exec_runtime() are used by timer functions to update the
respective fields of the thread_group_cputime structure.
Non-SMP operation is trivial and will not be mentioned further.
The per-cpu structure is always allocated when a task creates its first new
thread, via a call to thread_group_cputime_clone_thread() from copy_signal().
It is freed at process exit via a call to thread_group_cputime_free() from
cleanup_signal().
All functions that formerly summed utime/stime/sum_sched_runtime values from
from all threads in the thread group now use thread_group_cputime() to
snapshot the values in the thread_group_cputime structure or the values in
the task structure itself if the per-cpu structure hasn't been allocated.
Finally, the code in kernel/posix-cpu-timers.c has changed quite a bit.
The run_posix_cpu_timers() function has been split into a fast path and a
slow path; the former safely checks whether there are any expired thread
timers and, if not, just returns, while the slow path does the heavy lifting.
With the dedicated thread group fields, timers are no longer "rebalanced" and
the process_timer_rebalance() function and related code has gone away. All
summing loops are gone and all code that used them now uses the
thread_group_cputime() inline. When process-wide timers are set, the new
task_cputime structure in signal_struct is used to cache the earliest
expiration; this is checked in the fast path.
Performance
The fix appears not to add significant overhead to existing operations. It
generally performs the same as the current code except in two cases, one in
which it performs slightly worse (Case 5 below) and one in which it performs
very significantly better (Case 2 below). Overall it's a wash except in those
two cases.
I've since done somewhat more involved testing on a dual-core Opteron system.
Case 1: With no itimer running, for a test with 100,000 threads, the fixed
kernel took 1428.5 seconds, 513 seconds more than the unfixed system,
all of which was spent in the system. There were twice as many
voluntary context switches with the fix as without it.
Case 2: With an itimer running at .01 second ticks and 4000 threads (the most
an unmodified kernel can handle), the fixed kernel ran the test in
eight percent of the time (5.8 seconds as opposed to 70 seconds) and
had better tick accuracy (.012 seconds per tick as opposed to .023
seconds per tick).
Case 3: A 4000-thread test with an initial timer tick of .01 second and an
interval of 10,000 seconds (i.e. a timer that ticks only once) had
very nearly the same performance in both cases: 6.3 seconds elapsed
for the fixed kernel versus 5.5 seconds for the unfixed kernel.
With fewer threads (eight in these tests), the Case 1 test ran in essentially
the same time on both the modified and unmodified kernels (5.2 seconds versus
5.8 seconds). The Case 2 test ran in about the same time as well, 5.9 seconds
versus 5.4 seconds but again with much better tick accuracy, .013 seconds per
tick versus .025 seconds per tick for the unmodified kernel.
Since the fix affected the rlimit code, I also tested soft and hard CPU limits.
Case 4: With a hard CPU limit of 20 seconds and eight threads (and an itimer
running), the modified kernel was very slightly favored in that while
it killed the process in 19.997 seconds of CPU time (5.002 seconds of
wall time), only .003 seconds of that was system time, the rest was
user time. The unmodified kernel killed the process in 20.001 seconds
of CPU (5.014 seconds of wall time) of which .016 seconds was system
time. Really, though, the results were too close to call. The results
were essentially the same with no itimer running.
Case 5: With a soft limit of 20 seconds and a hard limit of 2000 seconds
(where the hard limit would never be reached) and an itimer running,
the modified kernel exhibited worse tick accuracy than the unmodified
kernel: .050 seconds/tick versus .028 seconds/tick. Otherwise,
performance was almost indistinguishable. With no itimer running this
test exhibited virtually identical behavior and times in both cases.
In times past I did some limited performance testing. those results are below.
On a four-cpu Opteron system without this fix, a sixteen-thread test executed
in 3569.991 seconds, of which user was 3568.435s and system was 1.556s. On
the same system with the fix, user and elapsed time were about the same, but
system time dropped to 0.007 seconds. Performance with eight, four and one
thread were comparable. Interestingly, the timer ticks with the fix seemed
more accurate: The sixteen-thread test with the fix received 149543 ticks
for 0.024 seconds per tick, while the same test without the fix received 58720
for 0.061 seconds per tick. Both cases were configured for an interval of
0.01 seconds. Again, the other tests were comparable. Each thread in this
test computed the primes up to 25,000,000.
I also did a test with a large number of threads, 100,000 threads, which is
impossible without the fix. In this case each thread computed the primes only
up to 10,000 (to make the runtime manageable). System time dominated, at
1546.968 seconds out of a total 2176.906 seconds (giving a user time of
629.938s). It received 147651 ticks for 0.015 seconds per tick, still quite
accurate. There is obviously no comparable test without the fix.
Signed-off-by: Frank Mayhar <fmayhar@google.com>
Cc: Roland McGrath <roland@redhat.com>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-09-12 16:54:39 +00:00
|
|
|
* @p in case that task is currently running.
|
2009-03-31 07:56:03 +00:00
|
|
|
*
|
|
|
|
* Called with task_rq_lock() held on @rq.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-03-31 07:56:03 +00:00
|
|
|
static u64 do_task_delta_exec(struct task_struct *p, struct rq *rq)
|
|
|
|
{
|
|
|
|
u64 ns = 0;
|
|
|
|
|
|
|
|
if (task_current(rq, p)) {
|
|
|
|
update_rq_clock(rq);
|
|
|
|
ns = rq->clock - p->se.exec_start;
|
|
|
|
if ((s64)ns < 0)
|
|
|
|
ns = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ns;
|
|
|
|
}
|
|
|
|
|
2008-09-12 16:54:39 +00:00
|
|
|
unsigned long long task_delta_exec(struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
2007-07-09 16:51:58 +00:00
|
|
|
struct rq *rq;
|
2008-09-12 16:54:39 +00:00
|
|
|
u64 ns = 0;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
rq = task_rq_lock(p, &flags);
|
2009-03-31 07:56:03 +00:00
|
|
|
ns = do_task_delta_exec(p, rq);
|
|
|
|
task_rq_unlock(rq, &flags);
|
2008-09-30 06:28:17 +00:00
|
|
|
|
2009-03-31 07:56:03 +00:00
|
|
|
return ns;
|
|
|
|
}
|
timers: fix itimer/many thread hang
Overview
This patch reworks the handling of POSIX CPU timers, including the
ITIMER_PROF, ITIMER_VIRT timers and rlimit handling. It was put together
with the help of Roland McGrath, the owner and original writer of this code.
The problem we ran into, and the reason for this rework, has to do with using
a profiling timer in a process with a large number of threads. It appears
that the performance of the old implementation of run_posix_cpu_timers() was
at least O(n*3) (where "n" is the number of threads in a process) or worse.
Everything is fine with an increasing number of threads until the time taken
for that routine to run becomes the same as or greater than the tick time, at
which point things degrade rather quickly.
This patch fixes bug 9906, "Weird hang with NPTL and SIGPROF."
Code Changes
This rework corrects the implementation of run_posix_cpu_timers() to make it
run in constant time for a particular machine. (Performance may vary between
one machine and another depending upon whether the kernel is built as single-
or multiprocessor and, in the latter case, depending upon the number of
running processors.) To do this, at each tick we now update fields in
signal_struct as well as task_struct. The run_posix_cpu_timers() function
uses those fields to make its decisions.
We define a new structure, "task_cputime," to contain user, system and
scheduler times and use these in appropriate places:
struct task_cputime {
cputime_t utime;
cputime_t stime;
unsigned long long sum_exec_runtime;
};
This is included in the structure "thread_group_cputime," which is a new
substructure of signal_struct and which varies for uniprocessor versus
multiprocessor kernels. For uniprocessor kernels, it uses "task_cputime" as
a simple substructure, while for multiprocessor kernels it is a pointer:
struct thread_group_cputime {
struct task_cputime totals;
};
struct thread_group_cputime {
struct task_cputime *totals;
};
We also add a new task_cputime substructure directly to signal_struct, to
cache the earliest expiration of process-wide timers, and task_cputime also
replaces the it_*_expires fields of task_struct (used for earliest expiration
of thread timers). The "thread_group_cputime" structure contains process-wide
timers that are updated via account_user_time() and friends. In the non-SMP
case the structure is a simple aggregator; unfortunately in the SMP case that
simplicity was not achievable due to cache-line contention between CPUs (in
one measured case performance was actually _worse_ on a 16-cpu system than
the same test on a 4-cpu system, due to this contention). For SMP, the
thread_group_cputime counters are maintained as a per-cpu structure allocated
using alloc_percpu(). The timer functions update only the timer field in
the structure corresponding to the running CPU, obtained using per_cpu_ptr().
We define a set of inline functions in sched.h that we use to maintain the
thread_group_cputime structure and hide the differences between UP and SMP
implementations from the rest of the kernel. The thread_group_cputime_init()
function initializes the thread_group_cputime structure for the given task.
The thread_group_cputime_alloc() is a no-op for UP; for SMP it calls the
out-of-line function thread_group_cputime_alloc_smp() to allocate and fill
in the per-cpu structures and fields. The thread_group_cputime_free()
function, also a no-op for UP, in SMP frees the per-cpu structures. The
thread_group_cputime_clone_thread() function (also a UP no-op) for SMP calls
thread_group_cputime_alloc() if the per-cpu structures haven't yet been
allocated. The thread_group_cputime() function fills the task_cputime
structure it is passed with the contents of the thread_group_cputime fields;
in UP it's that simple but in SMP it must also safely check that tsk->signal
is non-NULL (if it is it just uses the appropriate fields of task_struct) and,
if so, sums the per-cpu values for each online CPU. Finally, the three
functions account_group_user_time(), account_group_system_time() and
account_group_exec_runtime() are used by timer functions to update the
respective fields of the thread_group_cputime structure.
Non-SMP operation is trivial and will not be mentioned further.
The per-cpu structure is always allocated when a task creates its first new
thread, via a call to thread_group_cputime_clone_thread() from copy_signal().
It is freed at process exit via a call to thread_group_cputime_free() from
cleanup_signal().
All functions that formerly summed utime/stime/sum_sched_runtime values from
from all threads in the thread group now use thread_group_cputime() to
snapshot the values in the thread_group_cputime structure or the values in
the task structure itself if the per-cpu structure hasn't been allocated.
Finally, the code in kernel/posix-cpu-timers.c has changed quite a bit.
The run_posix_cpu_timers() function has been split into a fast path and a
slow path; the former safely checks whether there are any expired thread
timers and, if not, just returns, while the slow path does the heavy lifting.
With the dedicated thread group fields, timers are no longer "rebalanced" and
the process_timer_rebalance() function and related code has gone away. All
summing loops are gone and all code that used them now uses the
thread_group_cputime() inline. When process-wide timers are set, the new
task_cputime structure in signal_struct is used to cache the earliest
expiration; this is checked in the fast path.
Performance
The fix appears not to add significant overhead to existing operations. It
generally performs the same as the current code except in two cases, one in
which it performs slightly worse (Case 5 below) and one in which it performs
very significantly better (Case 2 below). Overall it's a wash except in those
two cases.
I've since done somewhat more involved testing on a dual-core Opteron system.
Case 1: With no itimer running, for a test with 100,000 threads, the fixed
kernel took 1428.5 seconds, 513 seconds more than the unfixed system,
all of which was spent in the system. There were twice as many
voluntary context switches with the fix as without it.
Case 2: With an itimer running at .01 second ticks and 4000 threads (the most
an unmodified kernel can handle), the fixed kernel ran the test in
eight percent of the time (5.8 seconds as opposed to 70 seconds) and
had better tick accuracy (.012 seconds per tick as opposed to .023
seconds per tick).
Case 3: A 4000-thread test with an initial timer tick of .01 second and an
interval of 10,000 seconds (i.e. a timer that ticks only once) had
very nearly the same performance in both cases: 6.3 seconds elapsed
for the fixed kernel versus 5.5 seconds for the unfixed kernel.
With fewer threads (eight in these tests), the Case 1 test ran in essentially
the same time on both the modified and unmodified kernels (5.2 seconds versus
5.8 seconds). The Case 2 test ran in about the same time as well, 5.9 seconds
versus 5.4 seconds but again with much better tick accuracy, .013 seconds per
tick versus .025 seconds per tick for the unmodified kernel.
Since the fix affected the rlimit code, I also tested soft and hard CPU limits.
Case 4: With a hard CPU limit of 20 seconds and eight threads (and an itimer
running), the modified kernel was very slightly favored in that while
it killed the process in 19.997 seconds of CPU time (5.002 seconds of
wall time), only .003 seconds of that was system time, the rest was
user time. The unmodified kernel killed the process in 20.001 seconds
of CPU (5.014 seconds of wall time) of which .016 seconds was system
time. Really, though, the results were too close to call. The results
were essentially the same with no itimer running.
Case 5: With a soft limit of 20 seconds and a hard limit of 2000 seconds
(where the hard limit would never be reached) and an itimer running,
the modified kernel exhibited worse tick accuracy than the unmodified
kernel: .050 seconds/tick versus .028 seconds/tick. Otherwise,
performance was almost indistinguishable. With no itimer running this
test exhibited virtually identical behavior and times in both cases.
In times past I did some limited performance testing. those results are below.
On a four-cpu Opteron system without this fix, a sixteen-thread test executed
in 3569.991 seconds, of which user was 3568.435s and system was 1.556s. On
the same system with the fix, user and elapsed time were about the same, but
system time dropped to 0.007 seconds. Performance with eight, four and one
thread were comparable. Interestingly, the timer ticks with the fix seemed
more accurate: The sixteen-thread test with the fix received 149543 ticks
for 0.024 seconds per tick, while the same test without the fix received 58720
for 0.061 seconds per tick. Both cases were configured for an interval of
0.01 seconds. Again, the other tests were comparable. Each thread in this
test computed the primes up to 25,000,000.
I also did a test with a large number of threads, 100,000 threads, which is
impossible without the fix. In this case each thread computed the primes only
up to 10,000 (to make the runtime manageable). System time dominated, at
1546.968 seconds out of a total 2176.906 seconds (giving a user time of
629.938s). It received 147651 ticks for 0.015 seconds per tick, still quite
accurate. There is obviously no comparable test without the fix.
Signed-off-by: Frank Mayhar <fmayhar@google.com>
Cc: Roland McGrath <roland@redhat.com>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-09-12 16:54:39 +00:00
|
|
|
|
2009-03-31 07:56:03 +00:00
|
|
|
/*
|
|
|
|
* Return accounted runtime for the task.
|
|
|
|
* In case the task is currently running, return the runtime plus current's
|
|
|
|
* pending runtime that have not been accounted yet.
|
|
|
|
*/
|
|
|
|
unsigned long long task_sched_runtime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
struct rq *rq;
|
|
|
|
u64 ns = 0;
|
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
|
|
|
ns = p->se.sum_exec_runtime + do_task_delta_exec(p, rq);
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
|
|
|
|
return ns;
|
|
|
|
}
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2009-03-31 07:56:03 +00:00
|
|
|
/*
|
|
|
|
* Return sum_exec_runtime for the thread group.
|
|
|
|
* In case the task is currently running, return the sum plus current's
|
|
|
|
* pending runtime that have not been accounted yet.
|
|
|
|
*
|
|
|
|
* Note that the thread group might have other running tasks as well,
|
|
|
|
* so the return value not includes other pending runtime that other
|
|
|
|
* running tasks might have.
|
|
|
|
*/
|
|
|
|
unsigned long long thread_group_sched_runtime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
struct task_cputime totals;
|
|
|
|
unsigned long flags;
|
|
|
|
struct rq *rq;
|
|
|
|
u64 ns;
|
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
|
|
|
thread_group_cputime(p, &totals);
|
|
|
|
ns = totals.sum_exec_runtime + do_task_delta_exec(p, rq);
|
2007-07-09 16:51:58 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return ns;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Account user cpu time to a process.
|
|
|
|
* @p: the process that the cpu time gets accounted to
|
|
|
|
* @cputime: the cpu time spent in user space since the last update
|
2008-12-31 14:11:37 +00:00
|
|
|
* @cputime_scaled: cputime scaled by cpu frequency
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-12-31 14:11:37 +00:00
|
|
|
void account_user_time(struct task_struct *p, cputime_t cputime,
|
|
|
|
cputime_t cputime_scaled)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
|
|
|
|
cputime64_t tmp;
|
|
|
|
|
2008-12-31 14:11:37 +00:00
|
|
|
/* Add user time to process. */
|
2005-04-16 22:20:36 +00:00
|
|
|
p->utime = cputime_add(p->utime, cputime);
|
2008-12-31 14:11:37 +00:00
|
|
|
p->utimescaled = cputime_add(p->utimescaled, cputime_scaled);
|
timers: fix itimer/many thread hang
Overview
This patch reworks the handling of POSIX CPU timers, including the
ITIMER_PROF, ITIMER_VIRT timers and rlimit handling. It was put together
with the help of Roland McGrath, the owner and original writer of this code.
The problem we ran into, and the reason for this rework, has to do with using
a profiling timer in a process with a large number of threads. It appears
that the performance of the old implementation of run_posix_cpu_timers() was
at least O(n*3) (where "n" is the number of threads in a process) or worse.
Everything is fine with an increasing number of threads until the time taken
for that routine to run becomes the same as or greater than the tick time, at
which point things degrade rather quickly.
This patch fixes bug 9906, "Weird hang with NPTL and SIGPROF."
Code Changes
This rework corrects the implementation of run_posix_cpu_timers() to make it
run in constant time for a particular machine. (Performance may vary between
one machine and another depending upon whether the kernel is built as single-
or multiprocessor and, in the latter case, depending upon the number of
running processors.) To do this, at each tick we now update fields in
signal_struct as well as task_struct. The run_posix_cpu_timers() function
uses those fields to make its decisions.
We define a new structure, "task_cputime," to contain user, system and
scheduler times and use these in appropriate places:
struct task_cputime {
cputime_t utime;
cputime_t stime;
unsigned long long sum_exec_runtime;
};
This is included in the structure "thread_group_cputime," which is a new
substructure of signal_struct and which varies for uniprocessor versus
multiprocessor kernels. For uniprocessor kernels, it uses "task_cputime" as
a simple substructure, while for multiprocessor kernels it is a pointer:
struct thread_group_cputime {
struct task_cputime totals;
};
struct thread_group_cputime {
struct task_cputime *totals;
};
We also add a new task_cputime substructure directly to signal_struct, to
cache the earliest expiration of process-wide timers, and task_cputime also
replaces the it_*_expires fields of task_struct (used for earliest expiration
of thread timers). The "thread_group_cputime" structure contains process-wide
timers that are updated via account_user_time() and friends. In the non-SMP
case the structure is a simple aggregator; unfortunately in the SMP case that
simplicity was not achievable due to cache-line contention between CPUs (in
one measured case performance was actually _worse_ on a 16-cpu system than
the same test on a 4-cpu system, due to this contention). For SMP, the
thread_group_cputime counters are maintained as a per-cpu structure allocated
using alloc_percpu(). The timer functions update only the timer field in
the structure corresponding to the running CPU, obtained using per_cpu_ptr().
We define a set of inline functions in sched.h that we use to maintain the
thread_group_cputime structure and hide the differences between UP and SMP
implementations from the rest of the kernel. The thread_group_cputime_init()
function initializes the thread_group_cputime structure for the given task.
The thread_group_cputime_alloc() is a no-op for UP; for SMP it calls the
out-of-line function thread_group_cputime_alloc_smp() to allocate and fill
in the per-cpu structures and fields. The thread_group_cputime_free()
function, also a no-op for UP, in SMP frees the per-cpu structures. The
thread_group_cputime_clone_thread() function (also a UP no-op) for SMP calls
thread_group_cputime_alloc() if the per-cpu structures haven't yet been
allocated. The thread_group_cputime() function fills the task_cputime
structure it is passed with the contents of the thread_group_cputime fields;
in UP it's that simple but in SMP it must also safely check that tsk->signal
is non-NULL (if it is it just uses the appropriate fields of task_struct) and,
if so, sums the per-cpu values for each online CPU. Finally, the three
functions account_group_user_time(), account_group_system_time() and
account_group_exec_runtime() are used by timer functions to update the
respective fields of the thread_group_cputime structure.
Non-SMP operation is trivial and will not be mentioned further.
The per-cpu structure is always allocated when a task creates its first new
thread, via a call to thread_group_cputime_clone_thread() from copy_signal().
It is freed at process exit via a call to thread_group_cputime_free() from
cleanup_signal().
All functions that formerly summed utime/stime/sum_sched_runtime values from
from all threads in the thread group now use thread_group_cputime() to
snapshot the values in the thread_group_cputime structure or the values in
the task structure itself if the per-cpu structure hasn't been allocated.
Finally, the code in kernel/posix-cpu-timers.c has changed quite a bit.
The run_posix_cpu_timers() function has been split into a fast path and a
slow path; the former safely checks whether there are any expired thread
timers and, if not, just returns, while the slow path does the heavy lifting.
With the dedicated thread group fields, timers are no longer "rebalanced" and
the process_timer_rebalance() function and related code has gone away. All
summing loops are gone and all code that used them now uses the
thread_group_cputime() inline. When process-wide timers are set, the new
task_cputime structure in signal_struct is used to cache the earliest
expiration; this is checked in the fast path.
Performance
The fix appears not to add significant overhead to existing operations. It
generally performs the same as the current code except in two cases, one in
which it performs slightly worse (Case 5 below) and one in which it performs
very significantly better (Case 2 below). Overall it's a wash except in those
two cases.
I've since done somewhat more involved testing on a dual-core Opteron system.
Case 1: With no itimer running, for a test with 100,000 threads, the fixed
kernel took 1428.5 seconds, 513 seconds more than the unfixed system,
all of which was spent in the system. There were twice as many
voluntary context switches with the fix as without it.
Case 2: With an itimer running at .01 second ticks and 4000 threads (the most
an unmodified kernel can handle), the fixed kernel ran the test in
eight percent of the time (5.8 seconds as opposed to 70 seconds) and
had better tick accuracy (.012 seconds per tick as opposed to .023
seconds per tick).
Case 3: A 4000-thread test with an initial timer tick of .01 second and an
interval of 10,000 seconds (i.e. a timer that ticks only once) had
very nearly the same performance in both cases: 6.3 seconds elapsed
for the fixed kernel versus 5.5 seconds for the unfixed kernel.
With fewer threads (eight in these tests), the Case 1 test ran in essentially
the same time on both the modified and unmodified kernels (5.2 seconds versus
5.8 seconds). The Case 2 test ran in about the same time as well, 5.9 seconds
versus 5.4 seconds but again with much better tick accuracy, .013 seconds per
tick versus .025 seconds per tick for the unmodified kernel.
Since the fix affected the rlimit code, I also tested soft and hard CPU limits.
Case 4: With a hard CPU limit of 20 seconds and eight threads (and an itimer
running), the modified kernel was very slightly favored in that while
it killed the process in 19.997 seconds of CPU time (5.002 seconds of
wall time), only .003 seconds of that was system time, the rest was
user time. The unmodified kernel killed the process in 20.001 seconds
of CPU (5.014 seconds of wall time) of which .016 seconds was system
time. Really, though, the results were too close to call. The results
were essentially the same with no itimer running.
Case 5: With a soft limit of 20 seconds and a hard limit of 2000 seconds
(where the hard limit would never be reached) and an itimer running,
the modified kernel exhibited worse tick accuracy than the unmodified
kernel: .050 seconds/tick versus .028 seconds/tick. Otherwise,
performance was almost indistinguishable. With no itimer running this
test exhibited virtually identical behavior and times in both cases.
In times past I did some limited performance testing. those results are below.
On a four-cpu Opteron system without this fix, a sixteen-thread test executed
in 3569.991 seconds, of which user was 3568.435s and system was 1.556s. On
the same system with the fix, user and elapsed time were about the same, but
system time dropped to 0.007 seconds. Performance with eight, four and one
thread were comparable. Interestingly, the timer ticks with the fix seemed
more accurate: The sixteen-thread test with the fix received 149543 ticks
for 0.024 seconds per tick, while the same test without the fix received 58720
for 0.061 seconds per tick. Both cases were configured for an interval of
0.01 seconds. Again, the other tests were comparable. Each thread in this
test computed the primes up to 25,000,000.
I also did a test with a large number of threads, 100,000 threads, which is
impossible without the fix. In this case each thread computed the primes only
up to 10,000 (to make the runtime manageable). System time dominated, at
1546.968 seconds out of a total 2176.906 seconds (giving a user time of
629.938s). It received 147651 ticks for 0.015 seconds per tick, still quite
accurate. There is obviously no comparable test without the fix.
Signed-off-by: Frank Mayhar <fmayhar@google.com>
Cc: Roland McGrath <roland@redhat.com>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-09-12 16:54:39 +00:00
|
|
|
account_group_user_time(p, cputime);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Add user time to cpustat. */
|
|
|
|
tmp = cputime_to_cputime64(cputime);
|
|
|
|
if (TASK_NICE(p) > 0)
|
|
|
|
cpustat->nice = cputime64_add(cpustat->nice, tmp);
|
|
|
|
else
|
|
|
|
cpustat->user = cputime64_add(cpustat->user, tmp);
|
2009-03-31 04:32:22 +00:00
|
|
|
|
|
|
|
cpuacct_update_stats(p, CPUACCT_STAT_USER, cputime);
|
2008-07-25 08:48:40 +00:00
|
|
|
/* Account for user time used */
|
|
|
|
acct_update_integrals(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:19 +00:00
|
|
|
/*
|
|
|
|
* Account guest cpu time to a process.
|
|
|
|
* @p: the process that the cpu time gets accounted to
|
|
|
|
* @cputime: the cpu time spent in virtual machine since the last update
|
2008-12-31 14:11:37 +00:00
|
|
|
* @cputime_scaled: cputime scaled by cpu frequency
|
2007-10-15 15:00:19 +00:00
|
|
|
*/
|
2008-12-31 14:11:37 +00:00
|
|
|
static void account_guest_time(struct task_struct *p, cputime_t cputime,
|
|
|
|
cputime_t cputime_scaled)
|
2007-10-15 15:00:19 +00:00
|
|
|
{
|
|
|
|
cputime64_t tmp;
|
|
|
|
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
|
|
|
|
|
|
|
|
tmp = cputime_to_cputime64(cputime);
|
|
|
|
|
2008-12-31 14:11:37 +00:00
|
|
|
/* Add guest time to process. */
|
2007-10-15 15:00:19 +00:00
|
|
|
p->utime = cputime_add(p->utime, cputime);
|
2008-12-31 14:11:37 +00:00
|
|
|
p->utimescaled = cputime_add(p->utimescaled, cputime_scaled);
|
timers: fix itimer/many thread hang
Overview
This patch reworks the handling of POSIX CPU timers, including the
ITIMER_PROF, ITIMER_VIRT timers and rlimit handling. It was put together
with the help of Roland McGrath, the owner and original writer of this code.
The problem we ran into, and the reason for this rework, has to do with using
a profiling timer in a process with a large number of threads. It appears
that the performance of the old implementation of run_posix_cpu_timers() was
at least O(n*3) (where "n" is the number of threads in a process) or worse.
Everything is fine with an increasing number of threads until the time taken
for that routine to run becomes the same as or greater than the tick time, at
which point things degrade rather quickly.
This patch fixes bug 9906, "Weird hang with NPTL and SIGPROF."
Code Changes
This rework corrects the implementation of run_posix_cpu_timers() to make it
run in constant time for a particular machine. (Performance may vary between
one machine and another depending upon whether the kernel is built as single-
or multiprocessor and, in the latter case, depending upon the number of
running processors.) To do this, at each tick we now update fields in
signal_struct as well as task_struct. The run_posix_cpu_timers() function
uses those fields to make its decisions.
We define a new structure, "task_cputime," to contain user, system and
scheduler times and use these in appropriate places:
struct task_cputime {
cputime_t utime;
cputime_t stime;
unsigned long long sum_exec_runtime;
};
This is included in the structure "thread_group_cputime," which is a new
substructure of signal_struct and which varies for uniprocessor versus
multiprocessor kernels. For uniprocessor kernels, it uses "task_cputime" as
a simple substructure, while for multiprocessor kernels it is a pointer:
struct thread_group_cputime {
struct task_cputime totals;
};
struct thread_group_cputime {
struct task_cputime *totals;
};
We also add a new task_cputime substructure directly to signal_struct, to
cache the earliest expiration of process-wide timers, and task_cputime also
replaces the it_*_expires fields of task_struct (used for earliest expiration
of thread timers). The "thread_group_cputime" structure contains process-wide
timers that are updated via account_user_time() and friends. In the non-SMP
case the structure is a simple aggregator; unfortunately in the SMP case that
simplicity was not achievable due to cache-line contention between CPUs (in
one measured case performance was actually _worse_ on a 16-cpu system than
the same test on a 4-cpu system, due to this contention). For SMP, the
thread_group_cputime counters are maintained as a per-cpu structure allocated
using alloc_percpu(). The timer functions update only the timer field in
the structure corresponding to the running CPU, obtained using per_cpu_ptr().
We define a set of inline functions in sched.h that we use to maintain the
thread_group_cputime structure and hide the differences between UP and SMP
implementations from the rest of the kernel. The thread_group_cputime_init()
function initializes the thread_group_cputime structure for the given task.
The thread_group_cputime_alloc() is a no-op for UP; for SMP it calls the
out-of-line function thread_group_cputime_alloc_smp() to allocate and fill
in the per-cpu structures and fields. The thread_group_cputime_free()
function, also a no-op for UP, in SMP frees the per-cpu structures. The
thread_group_cputime_clone_thread() function (also a UP no-op) for SMP calls
thread_group_cputime_alloc() if the per-cpu structures haven't yet been
allocated. The thread_group_cputime() function fills the task_cputime
structure it is passed with the contents of the thread_group_cputime fields;
in UP it's that simple but in SMP it must also safely check that tsk->signal
is non-NULL (if it is it just uses the appropriate fields of task_struct) and,
if so, sums the per-cpu values for each online CPU. Finally, the three
functions account_group_user_time(), account_group_system_time() and
account_group_exec_runtime() are used by timer functions to update the
respective fields of the thread_group_cputime structure.
Non-SMP operation is trivial and will not be mentioned further.
The per-cpu structure is always allocated when a task creates its first new
thread, via a call to thread_group_cputime_clone_thread() from copy_signal().
It is freed at process exit via a call to thread_group_cputime_free() from
cleanup_signal().
All functions that formerly summed utime/stime/sum_sched_runtime values from
from all threads in the thread group now use thread_group_cputime() to
snapshot the values in the thread_group_cputime structure or the values in
the task structure itself if the per-cpu structure hasn't been allocated.
Finally, the code in kernel/posix-cpu-timers.c has changed quite a bit.
The run_posix_cpu_timers() function has been split into a fast path and a
slow path; the former safely checks whether there are any expired thread
timers and, if not, just returns, while the slow path does the heavy lifting.
With the dedicated thread group fields, timers are no longer "rebalanced" and
the process_timer_rebalance() function and related code has gone away. All
summing loops are gone and all code that used them now uses the
thread_group_cputime() inline. When process-wide timers are set, the new
task_cputime structure in signal_struct is used to cache the earliest
expiration; this is checked in the fast path.
Performance
The fix appears not to add significant overhead to existing operations. It
generally performs the same as the current code except in two cases, one in
which it performs slightly worse (Case 5 below) and one in which it performs
very significantly better (Case 2 below). Overall it's a wash except in those
two cases.
I've since done somewhat more involved testing on a dual-core Opteron system.
Case 1: With no itimer running, for a test with 100,000 threads, the fixed
kernel took 1428.5 seconds, 513 seconds more than the unfixed system,
all of which was spent in the system. There were twice as many
voluntary context switches with the fix as without it.
Case 2: With an itimer running at .01 second ticks and 4000 threads (the most
an unmodified kernel can handle), the fixed kernel ran the test in
eight percent of the time (5.8 seconds as opposed to 70 seconds) and
had better tick accuracy (.012 seconds per tick as opposed to .023
seconds per tick).
Case 3: A 4000-thread test with an initial timer tick of .01 second and an
interval of 10,000 seconds (i.e. a timer that ticks only once) had
very nearly the same performance in both cases: 6.3 seconds elapsed
for the fixed kernel versus 5.5 seconds for the unfixed kernel.
With fewer threads (eight in these tests), the Case 1 test ran in essentially
the same time on both the modified and unmodified kernels (5.2 seconds versus
5.8 seconds). The Case 2 test ran in about the same time as well, 5.9 seconds
versus 5.4 seconds but again with much better tick accuracy, .013 seconds per
tick versus .025 seconds per tick for the unmodified kernel.
Since the fix affected the rlimit code, I also tested soft and hard CPU limits.
Case 4: With a hard CPU limit of 20 seconds and eight threads (and an itimer
running), the modified kernel was very slightly favored in that while
it killed the process in 19.997 seconds of CPU time (5.002 seconds of
wall time), only .003 seconds of that was system time, the rest was
user time. The unmodified kernel killed the process in 20.001 seconds
of CPU (5.014 seconds of wall time) of which .016 seconds was system
time. Really, though, the results were too close to call. The results
were essentially the same with no itimer running.
Case 5: With a soft limit of 20 seconds and a hard limit of 2000 seconds
(where the hard limit would never be reached) and an itimer running,
the modified kernel exhibited worse tick accuracy than the unmodified
kernel: .050 seconds/tick versus .028 seconds/tick. Otherwise,
performance was almost indistinguishable. With no itimer running this
test exhibited virtually identical behavior and times in both cases.
In times past I did some limited performance testing. those results are below.
On a four-cpu Opteron system without this fix, a sixteen-thread test executed
in 3569.991 seconds, of which user was 3568.435s and system was 1.556s. On
the same system with the fix, user and elapsed time were about the same, but
system time dropped to 0.007 seconds. Performance with eight, four and one
thread were comparable. Interestingly, the timer ticks with the fix seemed
more accurate: The sixteen-thread test with the fix received 149543 ticks
for 0.024 seconds per tick, while the same test without the fix received 58720
for 0.061 seconds per tick. Both cases were configured for an interval of
0.01 seconds. Again, the other tests were comparable. Each thread in this
test computed the primes up to 25,000,000.
I also did a test with a large number of threads, 100,000 threads, which is
impossible without the fix. In this case each thread computed the primes only
up to 10,000 (to make the runtime manageable). System time dominated, at
1546.968 seconds out of a total 2176.906 seconds (giving a user time of
629.938s). It received 147651 ticks for 0.015 seconds per tick, still quite
accurate. There is obviously no comparable test without the fix.
Signed-off-by: Frank Mayhar <fmayhar@google.com>
Cc: Roland McGrath <roland@redhat.com>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-09-12 16:54:39 +00:00
|
|
|
account_group_user_time(p, cputime);
|
2007-10-15 15:00:19 +00:00
|
|
|
p->gtime = cputime_add(p->gtime, cputime);
|
|
|
|
|
2008-12-31 14:11:37 +00:00
|
|
|
/* Add guest time to cpustat. */
|
2007-10-15 15:00:19 +00:00
|
|
|
cpustat->user = cputime64_add(cpustat->user, tmp);
|
|
|
|
cpustat->guest = cputime64_add(cpustat->guest, tmp);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Account system cpu time to a process.
|
|
|
|
* @p: the process that the cpu time gets accounted to
|
|
|
|
* @hardirq_offset: the offset to subtract from hardirq_count()
|
|
|
|
* @cputime: the cpu time spent in kernel space since the last update
|
2008-12-31 14:11:37 +00:00
|
|
|
* @cputime_scaled: cputime scaled by cpu frequency
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
void account_system_time(struct task_struct *p, int hardirq_offset,
|
2008-12-31 14:11:37 +00:00
|
|
|
cputime_t cputime, cputime_t cputime_scaled)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
|
|
|
|
cputime64_t tmp;
|
|
|
|
|
2008-04-25 01:17:55 +00:00
|
|
|
if ((p->flags & PF_VCPU) && (irq_count() - hardirq_offset == 0)) {
|
2008-12-31 14:11:37 +00:00
|
|
|
account_guest_time(p, cputime, cputime_scaled);
|
2008-04-25 01:17:55 +00:00
|
|
|
return;
|
|
|
|
}
|
2007-10-15 15:00:19 +00:00
|
|
|
|
2008-12-31 14:11:37 +00:00
|
|
|
/* Add system time to process. */
|
2005-04-16 22:20:36 +00:00
|
|
|
p->stime = cputime_add(p->stime, cputime);
|
2008-12-31 14:11:37 +00:00
|
|
|
p->stimescaled = cputime_add(p->stimescaled, cputime_scaled);
|
timers: fix itimer/many thread hang
Overview
This patch reworks the handling of POSIX CPU timers, including the
ITIMER_PROF, ITIMER_VIRT timers and rlimit handling. It was put together
with the help of Roland McGrath, the owner and original writer of this code.
The problem we ran into, and the reason for this rework, has to do with using
a profiling timer in a process with a large number of threads. It appears
that the performance of the old implementation of run_posix_cpu_timers() was
at least O(n*3) (where "n" is the number of threads in a process) or worse.
Everything is fine with an increasing number of threads until the time taken
for that routine to run becomes the same as or greater than the tick time, at
which point things degrade rather quickly.
This patch fixes bug 9906, "Weird hang with NPTL and SIGPROF."
Code Changes
This rework corrects the implementation of run_posix_cpu_timers() to make it
run in constant time for a particular machine. (Performance may vary between
one machine and another depending upon whether the kernel is built as single-
or multiprocessor and, in the latter case, depending upon the number of
running processors.) To do this, at each tick we now update fields in
signal_struct as well as task_struct. The run_posix_cpu_timers() function
uses those fields to make its decisions.
We define a new structure, "task_cputime," to contain user, system and
scheduler times and use these in appropriate places:
struct task_cputime {
cputime_t utime;
cputime_t stime;
unsigned long long sum_exec_runtime;
};
This is included in the structure "thread_group_cputime," which is a new
substructure of signal_struct and which varies for uniprocessor versus
multiprocessor kernels. For uniprocessor kernels, it uses "task_cputime" as
a simple substructure, while for multiprocessor kernels it is a pointer:
struct thread_group_cputime {
struct task_cputime totals;
};
struct thread_group_cputime {
struct task_cputime *totals;
};
We also add a new task_cputime substructure directly to signal_struct, to
cache the earliest expiration of process-wide timers, and task_cputime also
replaces the it_*_expires fields of task_struct (used for earliest expiration
of thread timers). The "thread_group_cputime" structure contains process-wide
timers that are updated via account_user_time() and friends. In the non-SMP
case the structure is a simple aggregator; unfortunately in the SMP case that
simplicity was not achievable due to cache-line contention between CPUs (in
one measured case performance was actually _worse_ on a 16-cpu system than
the same test on a 4-cpu system, due to this contention). For SMP, the
thread_group_cputime counters are maintained as a per-cpu structure allocated
using alloc_percpu(). The timer functions update only the timer field in
the structure corresponding to the running CPU, obtained using per_cpu_ptr().
We define a set of inline functions in sched.h that we use to maintain the
thread_group_cputime structure and hide the differences between UP and SMP
implementations from the rest of the kernel. The thread_group_cputime_init()
function initializes the thread_group_cputime structure for the given task.
The thread_group_cputime_alloc() is a no-op for UP; for SMP it calls the
out-of-line function thread_group_cputime_alloc_smp() to allocate and fill
in the per-cpu structures and fields. The thread_group_cputime_free()
function, also a no-op for UP, in SMP frees the per-cpu structures. The
thread_group_cputime_clone_thread() function (also a UP no-op) for SMP calls
thread_group_cputime_alloc() if the per-cpu structures haven't yet been
allocated. The thread_group_cputime() function fills the task_cputime
structure it is passed with the contents of the thread_group_cputime fields;
in UP it's that simple but in SMP it must also safely check that tsk->signal
is non-NULL (if it is it just uses the appropriate fields of task_struct) and,
if so, sums the per-cpu values for each online CPU. Finally, the three
functions account_group_user_time(), account_group_system_time() and
account_group_exec_runtime() are used by timer functions to update the
respective fields of the thread_group_cputime structure.
Non-SMP operation is trivial and will not be mentioned further.
The per-cpu structure is always allocated when a task creates its first new
thread, via a call to thread_group_cputime_clone_thread() from copy_signal().
It is freed at process exit via a call to thread_group_cputime_free() from
cleanup_signal().
All functions that formerly summed utime/stime/sum_sched_runtime values from
from all threads in the thread group now use thread_group_cputime() to
snapshot the values in the thread_group_cputime structure or the values in
the task structure itself if the per-cpu structure hasn't been allocated.
Finally, the code in kernel/posix-cpu-timers.c has changed quite a bit.
The run_posix_cpu_timers() function has been split into a fast path and a
slow path; the former safely checks whether there are any expired thread
timers and, if not, just returns, while the slow path does the heavy lifting.
With the dedicated thread group fields, timers are no longer "rebalanced" and
the process_timer_rebalance() function and related code has gone away. All
summing loops are gone and all code that used them now uses the
thread_group_cputime() inline. When process-wide timers are set, the new
task_cputime structure in signal_struct is used to cache the earliest
expiration; this is checked in the fast path.
Performance
The fix appears not to add significant overhead to existing operations. It
generally performs the same as the current code except in two cases, one in
which it performs slightly worse (Case 5 below) and one in which it performs
very significantly better (Case 2 below). Overall it's a wash except in those
two cases.
I've since done somewhat more involved testing on a dual-core Opteron system.
Case 1: With no itimer running, for a test with 100,000 threads, the fixed
kernel took 1428.5 seconds, 513 seconds more than the unfixed system,
all of which was spent in the system. There were twice as many
voluntary context switches with the fix as without it.
Case 2: With an itimer running at .01 second ticks and 4000 threads (the most
an unmodified kernel can handle), the fixed kernel ran the test in
eight percent of the time (5.8 seconds as opposed to 70 seconds) and
had better tick accuracy (.012 seconds per tick as opposed to .023
seconds per tick).
Case 3: A 4000-thread test with an initial timer tick of .01 second and an
interval of 10,000 seconds (i.e. a timer that ticks only once) had
very nearly the same performance in both cases: 6.3 seconds elapsed
for the fixed kernel versus 5.5 seconds for the unfixed kernel.
With fewer threads (eight in these tests), the Case 1 test ran in essentially
the same time on both the modified and unmodified kernels (5.2 seconds versus
5.8 seconds). The Case 2 test ran in about the same time as well, 5.9 seconds
versus 5.4 seconds but again with much better tick accuracy, .013 seconds per
tick versus .025 seconds per tick for the unmodified kernel.
Since the fix affected the rlimit code, I also tested soft and hard CPU limits.
Case 4: With a hard CPU limit of 20 seconds and eight threads (and an itimer
running), the modified kernel was very slightly favored in that while
it killed the process in 19.997 seconds of CPU time (5.002 seconds of
wall time), only .003 seconds of that was system time, the rest was
user time. The unmodified kernel killed the process in 20.001 seconds
of CPU (5.014 seconds of wall time) of which .016 seconds was system
time. Really, though, the results were too close to call. The results
were essentially the same with no itimer running.
Case 5: With a soft limit of 20 seconds and a hard limit of 2000 seconds
(where the hard limit would never be reached) and an itimer running,
the modified kernel exhibited worse tick accuracy than the unmodified
kernel: .050 seconds/tick versus .028 seconds/tick. Otherwise,
performance was almost indistinguishable. With no itimer running this
test exhibited virtually identical behavior and times in both cases.
In times past I did some limited performance testing. those results are below.
On a four-cpu Opteron system without this fix, a sixteen-thread test executed
in 3569.991 seconds, of which user was 3568.435s and system was 1.556s. On
the same system with the fix, user and elapsed time were about the same, but
system time dropped to 0.007 seconds. Performance with eight, four and one
thread were comparable. Interestingly, the timer ticks with the fix seemed
more accurate: The sixteen-thread test with the fix received 149543 ticks
for 0.024 seconds per tick, while the same test without the fix received 58720
for 0.061 seconds per tick. Both cases were configured for an interval of
0.01 seconds. Again, the other tests were comparable. Each thread in this
test computed the primes up to 25,000,000.
I also did a test with a large number of threads, 100,000 threads, which is
impossible without the fix. In this case each thread computed the primes only
up to 10,000 (to make the runtime manageable). System time dominated, at
1546.968 seconds out of a total 2176.906 seconds (giving a user time of
629.938s). It received 147651 ticks for 0.015 seconds per tick, still quite
accurate. There is obviously no comparable test without the fix.
Signed-off-by: Frank Mayhar <fmayhar@google.com>
Cc: Roland McGrath <roland@redhat.com>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-09-12 16:54:39 +00:00
|
|
|
account_group_system_time(p, cputime);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Add system time to cpustat. */
|
|
|
|
tmp = cputime_to_cputime64(cputime);
|
|
|
|
if (hardirq_count() - hardirq_offset)
|
|
|
|
cpustat->irq = cputime64_add(cpustat->irq, tmp);
|
|
|
|
else if (softirq_count())
|
|
|
|
cpustat->softirq = cputime64_add(cpustat->softirq, tmp);
|
|
|
|
else
|
2008-12-31 14:11:38 +00:00
|
|
|
cpustat->system = cputime64_add(cpustat->system, tmp);
|
|
|
|
|
2009-03-31 04:32:22 +00:00
|
|
|
cpuacct_update_stats(p, CPUACCT_STAT_SYSTEM, cputime);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Account for system time used */
|
|
|
|
acct_update_integrals(p);
|
|
|
|
}
|
|
|
|
|
2007-10-18 10:06:34 +00:00
|
|
|
/*
|
2005-04-16 22:20:36 +00:00
|
|
|
* Account for involuntary wait time.
|
|
|
|
* @steal: the cpu time spent in involuntary wait
|
2007-10-18 10:06:34 +00:00
|
|
|
*/
|
2008-12-31 14:11:38 +00:00
|
|
|
void account_steal_time(cputime_t cputime)
|
2007-10-18 10:06:34 +00:00
|
|
|
{
|
2008-12-31 14:11:38 +00:00
|
|
|
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
|
|
|
|
cputime64_t cputime64 = cputime_to_cputime64(cputime);
|
|
|
|
|
|
|
|
cpustat->steal = cputime64_add(cpustat->steal, cputime64);
|
2007-10-18 10:06:34 +00:00
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2008-12-31 14:11:38 +00:00
|
|
|
* Account for idle time.
|
|
|
|
* @cputime: the cpu time spent in idle wait
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-12-31 14:11:38 +00:00
|
|
|
void account_idle_time(cputime_t cputime)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
|
2008-12-31 14:11:38 +00:00
|
|
|
cputime64_t cputime64 = cputime_to_cputime64(cputime);
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = this_rq();
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-12-31 14:11:38 +00:00
|
|
|
if (atomic_read(&rq->nr_iowait) > 0)
|
|
|
|
cpustat->iowait = cputime64_add(cpustat->iowait, cputime64);
|
|
|
|
else
|
|
|
|
cpustat->idle = cputime64_add(cpustat->idle, cputime64);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-12-31 14:11:38 +00:00
|
|
|
#ifndef CONFIG_VIRT_CPU_ACCOUNTING
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Account a single tick of cpu time.
|
|
|
|
* @p: the process that the cpu time gets accounted to
|
|
|
|
* @user_tick: indicates if the tick is a user or a system tick
|
|
|
|
*/
|
|
|
|
void account_process_tick(struct task_struct *p, int user_tick)
|
|
|
|
{
|
|
|
|
cputime_t one_jiffy = jiffies_to_cputime(1);
|
|
|
|
cputime_t one_jiffy_scaled = cputime_to_scaled(one_jiffy);
|
|
|
|
struct rq *rq = this_rq();
|
|
|
|
|
|
|
|
if (user_tick)
|
|
|
|
account_user_time(p, one_jiffy, one_jiffy_scaled);
|
2009-04-29 12:44:49 +00:00
|
|
|
else if ((p != rq->idle) || (irq_count() != HARDIRQ_OFFSET))
|
2008-12-31 14:11:38 +00:00
|
|
|
account_system_time(p, HARDIRQ_OFFSET, one_jiffy,
|
|
|
|
one_jiffy_scaled);
|
|
|
|
else
|
|
|
|
account_idle_time(one_jiffy);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Account multiple ticks of steal time.
|
|
|
|
* @p: the process from which the cpu time has been stolen
|
|
|
|
* @ticks: number of stolen ticks
|
|
|
|
*/
|
|
|
|
void account_steal_ticks(unsigned long ticks)
|
|
|
|
{
|
|
|
|
account_steal_time(jiffies_to_cputime(ticks));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Account multiple ticks of idle time.
|
|
|
|
* @ticks: number of stolen ticks
|
|
|
|
*/
|
|
|
|
void account_idle_ticks(unsigned long ticks)
|
|
|
|
{
|
|
|
|
account_idle_time(jiffies_to_cputime(ticks));
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-12-31 14:11:38 +00:00
|
|
|
#endif
|
|
|
|
|
2008-09-05 16:12:23 +00:00
|
|
|
/*
|
|
|
|
* Use precise platform statistics if available:
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
|
|
|
|
cputime_t task_utime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
return p->utime;
|
|
|
|
}
|
|
|
|
|
|
|
|
cputime_t task_stime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
return p->stime;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
cputime_t task_utime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
clock_t utime = cputime_to_clock_t(p->utime),
|
|
|
|
total = utime + cputime_to_clock_t(p->stime);
|
|
|
|
u64 temp;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Use CFS's precise accounting:
|
|
|
|
*/
|
|
|
|
temp = (u64)nsec_to_clock_t(p->se.sum_exec_runtime);
|
|
|
|
|
|
|
|
if (total) {
|
|
|
|
temp *= utime;
|
|
|
|
do_div(temp, total);
|
|
|
|
}
|
|
|
|
utime = (clock_t)temp;
|
|
|
|
|
|
|
|
p->prev_utime = max(p->prev_utime, clock_t_to_cputime(utime));
|
|
|
|
return p->prev_utime;
|
|
|
|
}
|
|
|
|
|
|
|
|
cputime_t task_stime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
clock_t stime;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Use CFS's precise accounting. (we subtract utime from
|
|
|
|
* the total, to make sure the total observed by userspace
|
|
|
|
* grows monotonically - apps rely on that):
|
|
|
|
*/
|
|
|
|
stime = nsec_to_clock_t(p->se.sum_exec_runtime) -
|
|
|
|
cputime_to_clock_t(task_utime(p));
|
|
|
|
|
|
|
|
if (stime >= 0)
|
|
|
|
p->prev_stime = max(p->prev_stime, clock_t_to_cputime(stime));
|
|
|
|
|
|
|
|
return p->prev_stime;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
inline cputime_t task_gtime(struct task_struct *p)
|
|
|
|
{
|
|
|
|
return p->gtime;
|
|
|
|
}
|
|
|
|
|
2006-12-10 10:20:22 +00:00
|
|
|
/*
|
|
|
|
* This function gets called by the timer code, with HZ frequency.
|
|
|
|
* We call it with interrupts disabled.
|
|
|
|
*
|
|
|
|
* It also gets called by the fork code, when changing the parent's
|
|
|
|
* timeslices.
|
|
|
|
*/
|
|
|
|
void scheduler_tick(void)
|
|
|
|
{
|
|
|
|
int cpu = smp_processor_id();
|
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
struct task_struct *curr = rq->curr;
|
2008-05-03 16:29:28 +00:00
|
|
|
|
|
|
|
sched_clock_tick();
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
spin_lock(&rq->lock);
|
2008-05-03 16:29:28 +00:00
|
|
|
update_rq_clock(rq);
|
2007-08-09 09:16:45 +00:00
|
|
|
update_cpu_load(rq);
|
2008-01-25 20:08:29 +00:00
|
|
|
curr->sched_class->task_tick(rq, curr, 0);
|
2007-07-09 16:51:59 +00:00
|
|
|
spin_unlock(&rq->lock);
|
2006-12-10 10:20:22 +00:00
|
|
|
|
2009-05-23 16:28:55 +00:00
|
|
|
perf_counter_task_tick(curr, cpu);
|
|
|
|
|
2006-12-10 10:20:23 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2007-07-09 16:51:59 +00:00
|
|
|
rq->idle_at_tick = idle_cpu(cpu);
|
|
|
|
trigger_load_balance(rq, cpu);
|
2006-12-10 10:20:23 +00:00
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2009-04-02 06:18:25 +00:00
|
|
|
notrace unsigned long get_parent_ip(unsigned long addr)
|
2008-05-12 19:20:42 +00:00
|
|
|
{
|
|
|
|
if (in_lock_functions(addr)) {
|
|
|
|
addr = CALLER_ADDR2;
|
|
|
|
if (in_lock_functions(addr))
|
|
|
|
addr = CALLER_ADDR3;
|
|
|
|
}
|
|
|
|
return addr;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-01-23 00:01:40 +00:00
|
|
|
#if defined(CONFIG_PREEMPT) && (defined(CONFIG_DEBUG_PREEMPT) || \
|
|
|
|
defined(CONFIG_PREEMPT_TRACER))
|
|
|
|
|
2008-02-23 23:24:04 +00:00
|
|
|
void __kprobes add_preempt_count(int val)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-05-12 19:20:42 +00:00
|
|
|
#ifdef CONFIG_DEBUG_PREEMPT
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Underflow?
|
|
|
|
*/
|
2006-07-03 07:24:33 +00:00
|
|
|
if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0)))
|
|
|
|
return;
|
2008-05-12 19:20:42 +00:00
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
preempt_count() += val;
|
2008-05-12 19:20:42 +00:00
|
|
|
#ifdef CONFIG_DEBUG_PREEMPT
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Spinlock count overflowing soon?
|
|
|
|
*/
|
2006-12-10 10:20:38 +00:00
|
|
|
DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >=
|
|
|
|
PREEMPT_MASK - 10);
|
2008-05-12 19:20:42 +00:00
|
|
|
#endif
|
|
|
|
if (preempt_count() == val)
|
|
|
|
trace_preempt_off(CALLER_ADDR0, get_parent_ip(CALLER_ADDR1));
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(add_preempt_count);
|
|
|
|
|
2008-02-23 23:24:04 +00:00
|
|
|
void __kprobes sub_preempt_count(int val)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-05-12 19:20:42 +00:00
|
|
|
#ifdef CONFIG_DEBUG_PREEMPT
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Underflow?
|
|
|
|
*/
|
2009-01-12 12:00:50 +00:00
|
|
|
if (DEBUG_LOCKS_WARN_ON(val > preempt_count()))
|
2006-07-03 07:24:33 +00:00
|
|
|
return;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Is the spinlock portion underflowing?
|
|
|
|
*/
|
2006-07-03 07:24:33 +00:00
|
|
|
if (DEBUG_LOCKS_WARN_ON((val < PREEMPT_MASK) &&
|
|
|
|
!(preempt_count() & PREEMPT_MASK)))
|
|
|
|
return;
|
2008-05-12 19:20:42 +00:00
|
|
|
#endif
|
2006-07-03 07:24:33 +00:00
|
|
|
|
2008-05-12 19:20:42 +00:00
|
|
|
if (preempt_count() == val)
|
|
|
|
trace_preempt_on(CALLER_ADDR0, get_parent_ip(CALLER_ADDR1));
|
2005-04-16 22:20:36 +00:00
|
|
|
preempt_count() -= val;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(sub_preempt_count);
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* Print scheduling while atomic bug:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
static noinline void __schedule_bug(struct task_struct *prev)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-24 16:23:50 +00:00
|
|
|
struct pt_regs *regs = get_irq_regs();
|
|
|
|
|
|
|
|
printk(KERN_ERR "BUG: scheduling while atomic: %s/%d/0x%08x\n",
|
|
|
|
prev->comm, prev->pid, preempt_count());
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
debug_show_held_locks(prev);
|
2008-05-23 16:05:58 +00:00
|
|
|
print_modules();
|
2007-07-09 16:51:59 +00:00
|
|
|
if (irqs_disabled())
|
|
|
|
print_irqtrace_events(prev);
|
2007-10-24 16:23:50 +00:00
|
|
|
|
|
|
|
if (regs)
|
|
|
|
show_regs(regs);
|
|
|
|
else
|
|
|
|
dump_stack();
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* Various schedule()-time debugging checks and statistics:
|
|
|
|
*/
|
|
|
|
static inline void schedule_debug(struct task_struct *prev)
|
|
|
|
{
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-12-05 14:46:09 +00:00
|
|
|
* Test if we are atomic. Since do_exit() needs to call into
|
2005-04-16 22:20:36 +00:00
|
|
|
* schedule() atomically, we ignore that path for now.
|
|
|
|
* Otherwise, whine if we are scheduling when we should not be.
|
|
|
|
*/
|
2008-05-13 21:44:11 +00:00
|
|
|
if (unlikely(in_atomic_preempt_off() && !prev->exit_state))
|
2007-07-09 16:51:59 +00:00
|
|
|
__schedule_bug(prev);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
profile_hit(SCHED_PROFILING, __builtin_return_address(0));
|
|
|
|
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(this_rq(), sched_count);
|
2007-10-15 15:00:10 +00:00
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
|
|
|
if (unlikely(prev->lock_depth >= 0)) {
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(this_rq(), bkl_count);
|
|
|
|
schedstat_inc(prev, sched_info.bkl_count);
|
2007-10-15 15:00:10 +00:00
|
|
|
}
|
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2009-03-10 18:08:11 +00:00
|
|
|
static void put_prev_task(struct rq *rq, struct task_struct *prev)
|
|
|
|
{
|
|
|
|
if (prev->state == TASK_RUNNING) {
|
|
|
|
u64 runtime = prev->se.sum_exec_runtime;
|
|
|
|
|
|
|
|
runtime -= prev->se.prev_sum_exec_runtime;
|
|
|
|
runtime = min_t(u64, runtime, 2*sysctl_sched_migration_cost);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In order to avoid avg_overlap growing stale when we are
|
|
|
|
* indeed overlapping and hence not getting put to sleep, grow
|
|
|
|
* the avg_overlap on preemption.
|
|
|
|
*
|
|
|
|
* We use the average preemption runtime because that
|
|
|
|
* correlates to the amount of cache footprint a task can
|
|
|
|
* build up.
|
|
|
|
*/
|
|
|
|
update_avg(&prev->se.avg_overlap, runtime);
|
|
|
|
}
|
|
|
|
prev->sched_class->put_prev_task(rq, prev);
|
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* Pick up the highest-prio task:
|
|
|
|
*/
|
|
|
|
static inline struct task_struct *
|
2009-03-02 05:55:26 +00:00
|
|
|
pick_next_task(struct rq *rq)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
2007-10-15 15:00:12 +00:00
|
|
|
const struct sched_class *class;
|
2007-07-09 16:51:59 +00:00
|
|
|
struct task_struct *p;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2007-07-09 16:51:59 +00:00
|
|
|
* Optimization: we know that if all tasks are in
|
|
|
|
* the fair class we can call that function directly:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
if (likely(rq->nr_running == rq->cfs.nr_running)) {
|
2007-08-09 09:16:48 +00:00
|
|
|
p = fair_sched_class.pick_next_task(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (likely(p))
|
|
|
|
return p;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
class = sched_class_highest;
|
|
|
|
for ( ; ; ) {
|
2007-08-09 09:16:48 +00:00
|
|
|
p = class->pick_next_task(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (p)
|
|
|
|
return p;
|
|
|
|
/*
|
|
|
|
* Will never be NULL as the idle class always
|
|
|
|
* returns a non-NULL p:
|
|
|
|
*/
|
|
|
|
class = class->next;
|
|
|
|
}
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* schedule() is the main scheduler function.
|
|
|
|
*/
|
2009-03-13 11:21:26 +00:00
|
|
|
asmlinkage void __sched schedule(void)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
struct task_struct *prev, *next;
|
2008-02-15 17:56:36 +00:00
|
|
|
unsigned long *switch_count;
|
2007-07-09 16:51:59 +00:00
|
|
|
struct rq *rq;
|
2008-07-18 16:01:23 +00:00
|
|
|
int cpu;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2009-03-13 11:21:26 +00:00
|
|
|
need_resched:
|
|
|
|
preempt_disable();
|
2007-07-09 16:51:59 +00:00
|
|
|
cpu = smp_processor_id();
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
rcu_qsctr_inc(cpu);
|
|
|
|
prev = rq->curr;
|
|
|
|
switch_count = &prev->nivcsw;
|
|
|
|
|
|
|
|
release_kernel_lock(prev);
|
|
|
|
need_resched_nonpreemptible:
|
|
|
|
|
|
|
|
schedule_debug(prev);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-07-18 16:01:23 +00:00
|
|
|
if (sched_feat(HRTICK))
|
2008-05-12 19:20:55 +00:00
|
|
|
hrtick_clear(rq);
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2008-10-15 18:37:23 +00:00
|
|
|
spin_lock_irq(&rq->lock);
|
2008-05-03 16:29:28 +00:00
|
|
|
update_rq_clock(rq);
|
2007-10-15 15:00:13 +00:00
|
|
|
clear_tsk_need_resched(prev);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) {
|
sched: fix TASK_WAKEKILL vs SIGKILL race
schedule() has the special "TASK_INTERRUPTIBLE && signal_pending()" case,
this allows us to do
current->state = TASK_INTERRUPTIBLE;
schedule();
without fear to sleep with pending signal.
However, the code like
current->state = TASK_KILLABLE;
schedule();
is not right, schedule() doesn't take TASK_WAKEKILL into account. This means
that mutex_lock_killable(), wait_for_completion_killable(), down_killable(),
schedule_timeout_killable() can miss SIGKILL (and btw the second SIGKILL has
no effect).
Introduce the new helper, signal_pending_state(), and change schedule() to
use it. Hopefully it will have more users, that is why the task's state is
passed separately.
Note this "__TASK_STOPPED | __TASK_TRACED" check in signal_pending_state().
This is needed to preserve the current behaviour (ptrace_notify). I hope
this check will be removed soon, but this (afaics good) change needs the
separate discussion.
The fast path is "(state & (INTERRUPTIBLE | WAKEKILL)) + signal_pending(p)",
basically the same that schedule() does now. However, this patch of course
bloats schedule().
Signed-off-by: Oleg Nesterov <oleg@tv-sign.ru>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-06-08 17:20:41 +00:00
|
|
|
if (unlikely(signal_pending_state(prev->state, prev)))
|
2005-04-16 22:20:36 +00:00
|
|
|
prev->state = TASK_RUNNING;
|
sched: fix TASK_WAKEKILL vs SIGKILL race
schedule() has the special "TASK_INTERRUPTIBLE && signal_pending()" case,
this allows us to do
current->state = TASK_INTERRUPTIBLE;
schedule();
without fear to sleep with pending signal.
However, the code like
current->state = TASK_KILLABLE;
schedule();
is not right, schedule() doesn't take TASK_WAKEKILL into account. This means
that mutex_lock_killable(), wait_for_completion_killable(), down_killable(),
schedule_timeout_killable() can miss SIGKILL (and btw the second SIGKILL has
no effect).
Introduce the new helper, signal_pending_state(), and change schedule() to
use it. Hopefully it will have more users, that is why the task's state is
passed separately.
Note this "__TASK_STOPPED | __TASK_TRACED" check in signal_pending_state().
This is needed to preserve the current behaviour (ptrace_notify). I hope
this check will be removed soon, but this (afaics good) change needs the
separate discussion.
The fast path is "(state & (INTERRUPTIBLE | WAKEKILL)) + signal_pending(p)",
basically the same that schedule() does now. However, this patch of course
bloats schedule().
Signed-off-by: Oleg Nesterov <oleg@tv-sign.ru>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-06-08 17:20:41 +00:00
|
|
|
else
|
2007-08-09 09:16:49 +00:00
|
|
|
deactivate_task(rq, prev, 1);
|
2007-07-09 16:51:59 +00:00
|
|
|
switch_count = &prev->nvcsw;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:22 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
if (prev->sched_class->pre_schedule)
|
|
|
|
prev->sched_class->pre_schedule(rq, prev);
|
|
|
|
#endif
|
2008-01-25 20:08:07 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (unlikely(!rq->nr_running))
|
2005-04-16 22:20:36 +00:00
|
|
|
idle_balance(cpu, rq);
|
|
|
|
|
2009-03-10 18:08:11 +00:00
|
|
|
put_prev_task(rq, prev);
|
2009-03-02 05:55:26 +00:00
|
|
|
next = pick_next_task(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (likely(prev != next)) {
|
2008-04-29 09:08:59 +00:00
|
|
|
sched_info_switch(prev, next);
|
perf_counter: Optimize context switch between identical inherited contexts
When monitoring a process and its descendants with a set of inherited
counters, we can often get the situation in a context switch where
both the old (outgoing) and new (incoming) process have the same set
of counters, and their values are ultimately going to be added together.
In that situation it doesn't matter which set of counters are used to
count the activity for the new process, so there is really no need to
go through the process of reading the hardware counters and updating
the old task's counters and then setting up the PMU for the new task.
This optimizes the context switch in this situation. Instead of
scheduling out the perf_counter_context for the old task and
scheduling in the new context, we simply transfer the old context
to the new task and keep using it without interruption. The new
context gets transferred to the old task. This means that both
tasks still have a valid perf_counter_context, so no special case
is introduced when the old task gets scheduled in again, either on
this CPU or another CPU.
The equivalence of contexts is detected by keeping a pointer in
each cloned context pointing to the context it was cloned from.
To cope with the situation where a context is changed by adding
or removing counters after it has been cloned, we also keep a
generation number on each context which is incremented every time
a context is changed. When a context is cloned we take a copy
of the parent's generation number, and two cloned contexts are
equivalent only if they have the same parent and the same
generation number. In order that the parent context pointer
remains valid (and is not reused), we increment the parent
context's reference count for each context cloned from it.
Since we don't have individual fds for the counters in a cloned
context, the only thing that can make two clones of a given parent
different after they have been cloned is enabling or disabling all
counters with prctl. To account for this, we keep a count of the
number of enabled counters in each context. Two contexts must have
the same number of enabled counters to be considered equivalent.
Here are some measurements of the context switch time as measured with
the lat_ctx benchmark from lmbench, comparing the times obtained with
and without this patch series:
-----Unmodified----- With this patch series
Counters: none 2 HW 4H+4S none 2 HW 4H+4S
2 processes:
Average 3.44 6.45 11.24 3.12 3.39 3.60
St dev 0.04 0.04 0.13 0.05 0.17 0.19
8 processes:
Average 6.45 8.79 14.00 5.57 6.23 7.57
St dev 1.27 1.04 0.88 1.42 1.46 1.42
32 processes:
Average 5.56 8.43 13.78 5.28 5.55 7.15
St dev 0.41 0.47 0.53 0.54 0.57 0.81
The numbers are the mean and standard deviation of 20 runs of
lat_ctx. The "none" columns are lat_ctx run directly without any
counters. The "2 HW" columns are with lat_ctx run under perfstat,
counting cycles and instructions. The "4H+4S" columns are lat_ctx run
under perfstat with 4 hardware counters and 4 software counters
(cycles, instructions, cache references, cache misses, task
clock, context switch, cpu migrations, and page faults).
[ Impact: performance optimization of counter context-switches ]
Signed-off-by: Paul Mackerras <paulus@samba.org>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Corey Ashford <cjashfor@linux.vnet.ibm.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
LKML-Reference: <18966.10666.517218.332164@cargo.ozlabs.ibm.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-05-22 04:27:22 +00:00
|
|
|
perf_counter_task_sched_out(prev, next, cpu);
|
2008-04-29 09:08:59 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->nr_switches++;
|
|
|
|
rq->curr = next;
|
|
|
|
++*switch_count;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
context_switch(rq, prev, next); /* unlocks the rq */
|
2008-01-25 20:08:29 +00:00
|
|
|
/*
|
|
|
|
* the context switch might have flipped the stack from under
|
|
|
|
* us, hence refresh the local variables.
|
|
|
|
*/
|
|
|
|
cpu = smp_processor_id();
|
|
|
|
rq = cpu_rq(cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else
|
|
|
|
spin_unlock_irq(&rq->lock);
|
|
|
|
|
2008-01-25 20:08:29 +00:00
|
|
|
if (unlikely(reacquire_kernel_lock(current) < 0))
|
2005-04-16 22:20:36 +00:00
|
|
|
goto need_resched_nonpreemptible;
|
2008-01-25 20:08:29 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
preempt_enable_no_resched();
|
2009-03-13 11:21:26 +00:00
|
|
|
if (need_resched())
|
2005-04-16 22:20:36 +00:00
|
|
|
goto need_resched;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(schedule);
|
|
|
|
|
2009-01-12 13:01:47 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/*
|
|
|
|
* Look out! "owner" is an entirely speculative pointer
|
|
|
|
* access and not reliable.
|
|
|
|
*/
|
|
|
|
int mutex_spin_on_owner(struct mutex *lock, struct thread_info *owner)
|
|
|
|
{
|
|
|
|
unsigned int cpu;
|
|
|
|
struct rq *rq;
|
|
|
|
|
|
|
|
if (!sched_feat(OWNER_SPIN))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
#ifdef CONFIG_DEBUG_PAGEALLOC
|
|
|
|
/*
|
|
|
|
* Need to access the cpu field knowing that
|
|
|
|
* DEBUG_PAGEALLOC could have unmapped it if
|
|
|
|
* the mutex owner just released it and exited.
|
|
|
|
*/
|
|
|
|
if (probe_kernel_address(&owner->cpu, cpu))
|
|
|
|
goto out;
|
|
|
|
#else
|
|
|
|
cpu = owner->cpu;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Even if the access succeeded (likely case),
|
|
|
|
* the cpu field may no longer be valid.
|
|
|
|
*/
|
|
|
|
if (cpu >= nr_cpumask_bits)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We need to validate that we can do a
|
|
|
|
* get_cpu() and that we have the percpu area.
|
|
|
|
*/
|
|
|
|
if (!cpu_online(cpu))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
|
|
|
|
for (;;) {
|
|
|
|
/*
|
|
|
|
* Owner changed, break to re-assess state.
|
|
|
|
*/
|
|
|
|
if (lock->owner != owner)
|
|
|
|
break;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Is that owner really running on that cpu?
|
|
|
|
*/
|
|
|
|
if (task_thread_info(rq->curr) != owner || need_resched())
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
cpu_relax();
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_PREEMPT
|
|
|
|
/*
|
2006-07-10 11:43:52 +00:00
|
|
|
* this is the entry point to schedule() from in-kernel preemption
|
2007-12-05 14:46:09 +00:00
|
|
|
* off of preempt_enable. Kernel preemptions off return from interrupt
|
2005-04-16 22:20:36 +00:00
|
|
|
* occur there and call schedule directly.
|
|
|
|
*/
|
|
|
|
asmlinkage void __sched preempt_schedule(void)
|
|
|
|
{
|
|
|
|
struct thread_info *ti = current_thread_info();
|
2008-01-25 20:08:33 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* If there is a non-zero preempt_count or interrupts are disabled,
|
2007-12-05 14:46:09 +00:00
|
|
|
* we do not want to preempt the current task. Just return..
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2006-10-11 08:21:52 +00:00
|
|
|
if (likely(ti->preempt_count || irqs_disabled()))
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
do {
|
|
|
|
add_preempt_count(PREEMPT_ACTIVE);
|
|
|
|
schedule();
|
|
|
|
sub_preempt_count(PREEMPT_ACTIVE);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Check again in case we missed a preemption opportunity
|
|
|
|
* between schedule and now.
|
|
|
|
*/
|
|
|
|
barrier();
|
2009-03-06 11:40:20 +00:00
|
|
|
} while (need_resched());
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(preempt_schedule);
|
|
|
|
|
|
|
|
/*
|
2006-07-10 11:43:52 +00:00
|
|
|
* this is the entry point to schedule() from kernel preemption
|
2005-04-16 22:20:36 +00:00
|
|
|
* off of irq context.
|
|
|
|
* Note, that this is called and return with irqs disabled. This will
|
|
|
|
* protect us against recursive calling from irq.
|
|
|
|
*/
|
|
|
|
asmlinkage void __sched preempt_schedule_irq(void)
|
|
|
|
{
|
|
|
|
struct thread_info *ti = current_thread_info();
|
2008-01-25 20:08:33 +00:00
|
|
|
|
2006-07-10 11:43:52 +00:00
|
|
|
/* Catch callers which need to be fixed */
|
2005-04-16 22:20:36 +00:00
|
|
|
BUG_ON(ti->preempt_count || !irqs_disabled());
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
do {
|
|
|
|
add_preempt_count(PREEMPT_ACTIVE);
|
|
|
|
local_irq_enable();
|
|
|
|
schedule();
|
|
|
|
local_irq_disable();
|
|
|
|
sub_preempt_count(PREEMPT_ACTIVE);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Check again in case we missed a preemption opportunity
|
|
|
|
* between schedule and now.
|
|
|
|
*/
|
|
|
|
barrier();
|
2009-03-06 11:40:20 +00:00
|
|
|
} while (need_resched());
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_PREEMPT */
|
|
|
|
|
2005-09-10 07:26:11 +00:00
|
|
|
int default_wake_function(wait_queue_t *curr, unsigned mode, int sync,
|
|
|
|
void *key)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:40 +00:00
|
|
|
return try_to_wake_up(curr->private, mode, sync);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(default_wake_function);
|
|
|
|
|
|
|
|
/*
|
2007-12-05 14:46:09 +00:00
|
|
|
* The core wakeup function. Non-exclusive wakeups (nr_exclusive == 0) just
|
|
|
|
* wake everything up. If it's an exclusive wakeup (nr_exclusive == small +ve
|
2005-04-16 22:20:36 +00:00
|
|
|
* number) then we wake all the non-exclusive tasks and one exclusive task.
|
|
|
|
*
|
|
|
|
* There are circumstances in which we can try to wake a task which has already
|
2007-12-05 14:46:09 +00:00
|
|
|
* started to run but is not in state TASK_RUNNING. try_to_wake_up() returns
|
2005-04-16 22:20:36 +00:00
|
|
|
* zero in this (rare) case, and we handle it by continuing to scan the queue.
|
|
|
|
*/
|
2009-04-14 14:53:05 +00:00
|
|
|
static void __wake_up_common(wait_queue_head_t *q, unsigned int mode,
|
2009-02-04 23:12:14 +00:00
|
|
|
int nr_exclusive, int sync, void *key)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:02 +00:00
|
|
|
wait_queue_t *curr, *next;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:02 +00:00
|
|
|
list_for_each_entry_safe(curr, next, &q->task_list, task_list) {
|
2006-07-03 07:25:40 +00:00
|
|
|
unsigned flags = curr->flags;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (curr->func(curr, mode, sync, key) &&
|
2006-07-03 07:25:40 +00:00
|
|
|
(flags & WQ_FLAG_EXCLUSIVE) && !--nr_exclusive)
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* __wake_up - wake up threads blocked on a waitqueue.
|
|
|
|
* @q: the waitqueue
|
|
|
|
* @mode: which threads
|
|
|
|
* @nr_exclusive: how many wake-one or wake-many threads to wake up
|
2005-05-01 15:59:26 +00:00
|
|
|
* @key: is directly passed to the wakeup function
|
2009-04-28 14:01:38 +00:00
|
|
|
*
|
|
|
|
* It may be assumed that this function implies a write memory barrier before
|
|
|
|
* changing the task state if and only if any tasks are woken up.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-02-08 12:19:53 +00:00
|
|
|
void __wake_up(wait_queue_head_t *q, unsigned int mode,
|
2005-09-10 07:26:11 +00:00
|
|
|
int nr_exclusive, void *key)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&q->lock, flags);
|
|
|
|
__wake_up_common(q, mode, nr_exclusive, 0, key);
|
|
|
|
spin_unlock_irqrestore(&q->lock, flags);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__wake_up);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Same as __wake_up but called with the spinlock in wait_queue_head_t held.
|
|
|
|
*/
|
2008-02-08 12:19:53 +00:00
|
|
|
void __wake_up_locked(wait_queue_head_t *q, unsigned int mode)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
__wake_up_common(q, mode, 1, 0, NULL);
|
|
|
|
}
|
|
|
|
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
void __wake_up_locked_key(wait_queue_head_t *q, unsigned int mode, void *key)
|
|
|
|
{
|
|
|
|
__wake_up_common(q, mode, 1, 0, key);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/**
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
* __wake_up_sync_key - wake up threads blocked on a waitqueue.
|
2005-04-16 22:20:36 +00:00
|
|
|
* @q: the waitqueue
|
|
|
|
* @mode: which threads
|
|
|
|
* @nr_exclusive: how many wake-one or wake-many threads to wake up
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
* @key: opaque value to be passed to wakeup targets
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* The sync wakeup differs that the waker knows that it will schedule
|
|
|
|
* away soon, so while the target thread will be woken up, it will not
|
|
|
|
* be migrated to another CPU - ie. the two threads are 'synchronized'
|
|
|
|
* with each other. This can prevent needless bouncing between CPUs.
|
|
|
|
*
|
|
|
|
* On UP it can prevent extra preemption.
|
2009-04-28 14:01:38 +00:00
|
|
|
*
|
|
|
|
* It may be assumed that this function implies a write memory barrier before
|
|
|
|
* changing the task state if and only if any tasks are woken up.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
void __wake_up_sync_key(wait_queue_head_t *q, unsigned int mode,
|
|
|
|
int nr_exclusive, void *key)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
int sync = 1;
|
|
|
|
|
|
|
|
if (unlikely(!q))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (unlikely(!nr_exclusive))
|
|
|
|
sync = 0;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&q->lock, flags);
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
__wake_up_common(q, mode, nr_exclusive, sync, key);
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irqrestore(&q->lock, flags);
|
|
|
|
}
|
epoll keyed wakeups: add __wake_up_locked_key() and __wake_up_sync_key()
This patchset introduces wakeup hints for some of the most popular (from
epoll POV) devices, so that epoll code can avoid spurious wakeups on its
waiters.
The problem with epoll is that the callback-based wakeups do not, ATM,
carry any information about the events the wakeup is related to. So the
only choice epoll has (not being able to call f_op->poll() from inside the
callback), is to add the file* to a ready-list and resolve the real events
later on, at epoll_wait() (or its own f_op->poll()) time. This can cause
spurious wakeups, since the wake_up() itself might be for an event the
caller is not interested into.
The rate of these spurious wakeup can be pretty high in case of many
network sockets being monitored.
By allowing devices to report the events the wakeups refer to (at least
the two major classes - POLLIN/POLLOUT), we are able to spare useless
wakeups by proper handling inside the epoll's poll callback.
Epoll will have in any case to call f_op->poll() on the file* later on,
since the change to be done in order to have the full event set sent via
wakeup, is too invasive for the way our f_op->poll() system works (the
full event set is calculated inside the poll function - there are too many
of them to even start thinking the change - also poll/select would need
change too).
Epoll is changed in a way that both devices which send event hints, and
the ones that don't, are correctly handled. The former will gain some
efficiency though.
As a general rule for devices, would be to add an event mask by using
key-aware wakeup macros, when making up poll wait queues. I tested it
(together with the epoll's poll fix patch Andrew has in -mm) and wakeups
for the supported devices are correctly filtered.
Test program available here:
http://www.xmailserver.org/epoll_test.c
This patch:
Nothing revolutionary here. Just using the available "key" that our
wakeup core already support. The __wake_up_locked_key() was no brainer,
since both __wake_up_locked() and __wake_up_locked_key() are thin wrappers
around __wake_up_common().
The __wake_up_sync() function had a body, so the choice was between
borrowing the body for __wake_up_sync_key() and calling it from
__wake_up_sync(), or make an inline and calling it from both. I chose the
former since in most archs it all resolves to "mov $0, REG; jmp ADDR".
Signed-off-by: Davide Libenzi <davidel@xmailserver.org>
Cc: Alan Cox <alan@lxorguk.ukuu.org.uk>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: David Miller <davem@davemloft.net>
Cc: William Lee Irwin III <wli@movementarian.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-03-31 22:24:20 +00:00
|
|
|
EXPORT_SYMBOL_GPL(__wake_up_sync_key);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* __wake_up_sync - see __wake_up_sync_key()
|
|
|
|
*/
|
|
|
|
void __wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr_exclusive)
|
|
|
|
{
|
|
|
|
__wake_up_sync_key(q, mode, nr_exclusive, NULL);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
EXPORT_SYMBOL_GPL(__wake_up_sync); /* For internal use only */
|
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* complete: - signals a single thread waiting on this completion
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
*
|
|
|
|
* This will wake up a single thread waiting on this completion. Threads will be
|
|
|
|
* awakened in the same order in which they were queued.
|
|
|
|
*
|
|
|
|
* See also complete_all(), wait_for_completion() and related routines.
|
2009-04-28 14:01:38 +00:00
|
|
|
*
|
|
|
|
* It may be assumed that this function implies a write memory barrier before
|
|
|
|
* changing the task state if and only if any tasks are woken up.
|
2008-08-26 08:26:54 +00:00
|
|
|
*/
|
2007-10-24 16:23:48 +00:00
|
|
|
void complete(struct completion *x)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&x->wait.lock, flags);
|
|
|
|
x->done++;
|
2007-12-06 16:07:07 +00:00
|
|
|
__wake_up_common(&x->wait, TASK_NORMAL, 1, 0, NULL);
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irqrestore(&x->wait.lock, flags);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(complete);
|
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* complete_all: - signals all threads waiting on this completion
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
*
|
|
|
|
* This will wake up all threads waiting on this particular completion event.
|
2009-04-28 14:01:38 +00:00
|
|
|
*
|
|
|
|
* It may be assumed that this function implies a write memory barrier before
|
|
|
|
* changing the task state if and only if any tasks are woken up.
|
2008-08-26 08:26:54 +00:00
|
|
|
*/
|
2007-10-24 16:23:48 +00:00
|
|
|
void complete_all(struct completion *x)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&x->wait.lock, flags);
|
|
|
|
x->done += UINT_MAX/2;
|
2007-12-06 16:07:07 +00:00
|
|
|
__wake_up_common(&x->wait, TASK_NORMAL, 0, 0, NULL);
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irqrestore(&x->wait.lock, flags);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(complete_all);
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
static inline long __sched
|
|
|
|
do_wait_for_common(struct completion *x, long timeout, int state)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
if (!x->done) {
|
|
|
|
DECLARE_WAITQUEUE(wait, current);
|
|
|
|
|
|
|
|
wait.flags |= WQ_FLAG_EXCLUSIVE;
|
|
|
|
__add_wait_queue_tail(&x->wait, &wait);
|
|
|
|
do {
|
2008-08-20 23:54:41 +00:00
|
|
|
if (signal_pending_state(state, current)) {
|
2008-06-20 14:32:20 +00:00
|
|
|
timeout = -ERESTARTSYS;
|
|
|
|
break;
|
2007-10-15 15:00:14 +00:00
|
|
|
}
|
|
|
|
__set_current_state(state);
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irq(&x->wait.lock);
|
|
|
|
timeout = schedule_timeout(timeout);
|
|
|
|
spin_lock_irq(&x->wait.lock);
|
2008-06-20 14:32:20 +00:00
|
|
|
} while (!x->done && timeout);
|
2005-04-16 22:20:36 +00:00
|
|
|
__remove_wait_queue(&x->wait, &wait);
|
2008-06-20 14:32:20 +00:00
|
|
|
if (!x->done)
|
|
|
|
return timeout;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
x->done--;
|
2008-06-20 14:32:20 +00:00
|
|
|
return timeout ?: 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
static long __sched
|
|
|
|
wait_for_common(struct completion *x, long timeout, int state)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
might_sleep();
|
|
|
|
|
|
|
|
spin_lock_irq(&x->wait.lock);
|
2007-10-15 15:00:14 +00:00
|
|
|
timeout = do_wait_for_common(x, timeout, state);
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irq(&x->wait.lock);
|
2007-10-15 15:00:14 +00:00
|
|
|
return timeout;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* wait_for_completion: - waits for completion of a task
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
*
|
|
|
|
* This waits to be signaled for completion of a specific task. It is NOT
|
|
|
|
* interruptible and there is no timeout.
|
|
|
|
*
|
|
|
|
* See also similar routines (i.e. wait_for_completion_timeout()) with timeout
|
|
|
|
* and interrupt capability. Also see complete().
|
|
|
|
*/
|
2007-10-24 16:23:48 +00:00
|
|
|
void __sched wait_for_completion(struct completion *x)
|
2007-10-15 15:00:14 +00:00
|
|
|
{
|
|
|
|
wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_UNINTERRUPTIBLE);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-10-15 15:00:14 +00:00
|
|
|
EXPORT_SYMBOL(wait_for_completion);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* wait_for_completion_timeout: - waits for completion of a task (w/timeout)
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
* @timeout: timeout value in jiffies
|
|
|
|
*
|
|
|
|
* This waits for either a completion of a specific task to be signaled or for a
|
|
|
|
* specified timeout to expire. The timeout is in jiffies. It is not
|
|
|
|
* interruptible.
|
|
|
|
*/
|
2007-10-24 16:23:48 +00:00
|
|
|
unsigned long __sched
|
2007-10-15 15:00:14 +00:00
|
|
|
wait_for_completion_timeout(struct completion *x, unsigned long timeout)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
return wait_for_common(x, timeout, TASK_UNINTERRUPTIBLE);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-10-15 15:00:14 +00:00
|
|
|
EXPORT_SYMBOL(wait_for_completion_timeout);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* wait_for_completion_interruptible: - waits for completion of a task (w/intr)
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
*
|
|
|
|
* This waits for completion of a specific task to be signaled. It is
|
|
|
|
* interruptible.
|
|
|
|
*/
|
2007-10-15 15:00:14 +00:00
|
|
|
int __sched wait_for_completion_interruptible(struct completion *x)
|
2007-07-09 16:52:01 +00:00
|
|
|
{
|
2007-10-18 19:32:55 +00:00
|
|
|
long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_INTERRUPTIBLE);
|
|
|
|
if (t == -ERESTARTSYS)
|
|
|
|
return t;
|
|
|
|
return 0;
|
2007-07-09 16:52:01 +00:00
|
|
|
}
|
2007-10-15 15:00:14 +00:00
|
|
|
EXPORT_SYMBOL(wait_for_completion_interruptible);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* wait_for_completion_interruptible_timeout: - waits for completion (w/(to,intr))
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
* @timeout: timeout value in jiffies
|
|
|
|
*
|
|
|
|
* This waits for either a completion of a specific task to be signaled or for a
|
|
|
|
* specified timeout to expire. It is interruptible. The timeout is in jiffies.
|
|
|
|
*/
|
2007-10-24 16:23:48 +00:00
|
|
|
unsigned long __sched
|
2007-10-15 15:00:14 +00:00
|
|
|
wait_for_completion_interruptible_timeout(struct completion *x,
|
|
|
|
unsigned long timeout)
|
2007-07-09 16:52:01 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
return wait_for_common(x, timeout, TASK_INTERRUPTIBLE);
|
2007-07-09 16:52:01 +00:00
|
|
|
}
|
2007-10-15 15:00:14 +00:00
|
|
|
EXPORT_SYMBOL(wait_for_completion_interruptible_timeout);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-26 08:26:54 +00:00
|
|
|
/**
|
|
|
|
* wait_for_completion_killable: - waits for completion of a task (killable)
|
|
|
|
* @x: holds the state of this particular completion
|
|
|
|
*
|
|
|
|
* This waits to be signaled for completion of a specific task. It can be
|
|
|
|
* interrupted by a kill signal.
|
|
|
|
*/
|
2007-12-06 17:29:54 +00:00
|
|
|
int __sched wait_for_completion_killable(struct completion *x)
|
|
|
|
{
|
|
|
|
long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_KILLABLE);
|
|
|
|
if (t == -ERESTARTSYS)
|
|
|
|
return t;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(wait_for_completion_killable);
|
|
|
|
|
2008-08-15 07:40:44 +00:00
|
|
|
/**
|
|
|
|
* try_wait_for_completion - try to decrement a completion without blocking
|
|
|
|
* @x: completion structure
|
|
|
|
*
|
|
|
|
* Returns: 0 if a decrement cannot be done without blocking
|
|
|
|
* 1 if a decrement succeeded.
|
|
|
|
*
|
|
|
|
* If a completion is being used as a counting completion,
|
|
|
|
* attempt to decrement the counter without blocking. This
|
|
|
|
* enables us to avoid waiting if the resource the completion
|
|
|
|
* is protecting is not available.
|
|
|
|
*/
|
|
|
|
bool try_wait_for_completion(struct completion *x)
|
|
|
|
{
|
|
|
|
int ret = 1;
|
|
|
|
|
|
|
|
spin_lock_irq(&x->wait.lock);
|
|
|
|
if (!x->done)
|
|
|
|
ret = 0;
|
|
|
|
else
|
|
|
|
x->done--;
|
|
|
|
spin_unlock_irq(&x->wait.lock);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(try_wait_for_completion);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* completion_done - Test to see if a completion has any waiters
|
|
|
|
* @x: completion structure
|
|
|
|
*
|
|
|
|
* Returns: 0 if there are waiters (wait_for_completion() in progress)
|
|
|
|
* 1 if there are no waiters.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
bool completion_done(struct completion *x)
|
|
|
|
{
|
|
|
|
int ret = 1;
|
|
|
|
|
|
|
|
spin_lock_irq(&x->wait.lock);
|
|
|
|
if (!x->done)
|
|
|
|
ret = 0;
|
|
|
|
spin_unlock_irq(&x->wait.lock);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(completion_done);
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
static long __sched
|
|
|
|
sleep_on_common(wait_queue_head_t *q, int state, long timeout)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-07-09 16:52:01 +00:00
|
|
|
unsigned long flags;
|
|
|
|
wait_queue_t wait;
|
|
|
|
|
|
|
|
init_waitqueue_entry(&wait, current);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
__set_current_state(state);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
spin_lock_irqsave(&q->lock, flags);
|
|
|
|
__add_wait_queue(q, &wait);
|
|
|
|
spin_unlock(&q->lock);
|
|
|
|
timeout = schedule_timeout(timeout);
|
|
|
|
spin_lock_irq(&q->lock);
|
|
|
|
__remove_wait_queue(q, &wait);
|
|
|
|
spin_unlock_irqrestore(&q->lock, flags);
|
|
|
|
|
|
|
|
return timeout;
|
|
|
|
}
|
|
|
|
|
|
|
|
void __sched interruptible_sleep_on(wait_queue_head_t *q)
|
|
|
|
{
|
|
|
|
sleep_on_common(q, TASK_INTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(interruptible_sleep_on);
|
|
|
|
|
2007-07-09 16:52:01 +00:00
|
|
|
long __sched
|
2005-09-10 07:26:11 +00:00
|
|
|
interruptible_sleep_on_timeout(wait_queue_head_t *q, long timeout)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
return sleep_on_common(q, TASK_INTERRUPTIBLE, timeout);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(interruptible_sleep_on_timeout);
|
|
|
|
|
2007-07-09 16:52:01 +00:00
|
|
|
void __sched sleep_on(wait_queue_head_t *q)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
sleep_on_common(q, TASK_UNINTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(sleep_on);
|
|
|
|
|
2007-07-09 16:52:01 +00:00
|
|
|
long __sched sleep_on_timeout(wait_queue_head_t *q, long timeout)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:14 +00:00
|
|
|
return sleep_on_common(q, TASK_UNINTERRUPTIBLE, timeout);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(sleep_on_timeout);
|
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
|
|
|
|
/*
|
|
|
|
* rt_mutex_setprio - set the current priority of a task
|
|
|
|
* @p: task
|
|
|
|
* @prio: prio value (kernel-internal form)
|
|
|
|
*
|
|
|
|
* This function changes the 'effective' priority of a task. It does
|
|
|
|
* not touch ->normal_prio like __setscheduler().
|
|
|
|
*
|
|
|
|
* Used by the rt_mutex code to implement priority inheritance logic.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
void rt_mutex_setprio(struct task_struct *p, int prio)
|
2006-06-27 09:54:51 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
2007-10-15 15:00:08 +00:00
|
|
|
int oldprio, on_rq, running;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2008-01-25 20:08:22 +00:00
|
|
|
const struct sched_class *prev_class = p->sched_class;
|
2006-06-27 09:54:51 +00:00
|
|
|
|
|
|
|
BUG_ON(prio < 0 || prio > MAX_PRIO);
|
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2007-08-09 09:16:47 +00:00
|
|
|
update_rq_clock(rq);
|
2006-06-27 09:54:51 +00:00
|
|
|
|
2007-05-09 03:27:06 +00:00
|
|
|
oldprio = p->prio;
|
2007-07-09 16:51:59 +00:00
|
|
|
on_rq = p->se.on_rq;
|
2007-12-18 14:21:13 +00:00
|
|
|
running = task_current(rq, p);
|
2008-03-10 18:01:20 +00:00
|
|
|
if (on_rq)
|
2007-08-09 09:16:49 +00:00
|
|
|
dequeue_task(rq, p, 0);
|
2008-03-10 18:01:20 +00:00
|
|
|
if (running)
|
|
|
|
p->sched_class->put_prev_task(rq, p);
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
if (rt_prio(prio))
|
|
|
|
p->sched_class = &rt_sched_class;
|
|
|
|
else
|
|
|
|
p->sched_class = &fair_sched_class;
|
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
p->prio = prio;
|
|
|
|
|
2008-03-10 18:01:20 +00:00
|
|
|
if (running)
|
|
|
|
p->sched_class->set_curr_task(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (on_rq) {
|
2007-08-09 09:16:49 +00:00
|
|
|
enqueue_task(rq, p, 0);
|
2008-01-25 20:08:22 +00:00
|
|
|
|
|
|
|
check_class_changed(rq, p, prev_class, oldprio, running);
|
2006-06-27 09:54:51 +00:00
|
|
|
}
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2006-07-03 07:25:41 +00:00
|
|
|
void set_user_nice(struct task_struct *p, long nice)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
int old_prio, delta, on_rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (TASK_NICE(p) == nice || nice < -20 || nice > 19)
|
|
|
|
return;
|
|
|
|
/*
|
|
|
|
* We have to be careful, if called from sys_setpriority(),
|
|
|
|
* the task might be in the middle of scheduling on another CPU.
|
|
|
|
*/
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2007-08-09 09:16:47 +00:00
|
|
|
update_rq_clock(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* The RT priorities are set via sched_setscheduler(), but we still
|
|
|
|
* allow the 'normal' nice value to be set - but as expected
|
|
|
|
* it wont have any effect on scheduling until the task is
|
2007-07-09 16:51:59 +00:00
|
|
|
* SCHED_FIFO/SCHED_RR:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-07-09 16:51:59 +00:00
|
|
|
if (task_has_rt_policy(p)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
p->static_prio = NICE_TO_PRIO(nice);
|
|
|
|
goto out_unlock;
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
on_rq = p->se.on_rq;
|
2008-06-27 11:41:14 +00:00
|
|
|
if (on_rq)
|
2007-08-09 09:16:49 +00:00
|
|
|
dequeue_task(rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
p->static_prio = NICE_TO_PRIO(nice);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
set_load_weight(p);
|
2006-06-27 09:54:51 +00:00
|
|
|
old_prio = p->prio;
|
|
|
|
p->prio = effective_prio(p);
|
|
|
|
delta = p->prio - old_prio;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
if (on_rq) {
|
2007-08-09 09:16:49 +00:00
|
|
|
enqueue_task(rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-05-09 03:27:06 +00:00
|
|
|
* If the task increased its priority or is running and
|
|
|
|
* lowered its priority, then reschedule its CPU:
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-05-09 03:27:06 +00:00
|
|
|
if (delta < 0 || (delta > 0 && task_running(rq, p)))
|
2005-04-16 22:20:36 +00:00
|
|
|
resched_task(rq->curr);
|
|
|
|
}
|
|
|
|
out_unlock:
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(set_user_nice);
|
|
|
|
|
2005-05-01 15:59:00 +00:00
|
|
|
/*
|
|
|
|
* can_nice - check if a task can reduce its nice value
|
|
|
|
* @p: task
|
|
|
|
* @nice: nice value
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
int can_nice(const struct task_struct *p, const int nice)
|
2005-05-01 15:59:00 +00:00
|
|
|
{
|
2005-08-18 18:24:19 +00:00
|
|
|
/* convert nice value [19,-20] to rlimit style value [1,40] */
|
|
|
|
int nice_rlim = 20 - nice;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-05-01 15:59:00 +00:00
|
|
|
return (nice_rlim <= p->signal->rlim[RLIMIT_NICE].rlim_cur ||
|
|
|
|
capable(CAP_SYS_NICE));
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef __ARCH_WANT_SYS_NICE
|
|
|
|
|
|
|
|
/*
|
|
|
|
* sys_nice - change the priority of the current process.
|
|
|
|
* @increment: priority increment
|
|
|
|
*
|
|
|
|
* sys_setpriority is a more generic, but much slower function that
|
|
|
|
* does similar things.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE1(nice, int, increment)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:40 +00:00
|
|
|
long nice, retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Setpriority might change our priority at the same moment.
|
|
|
|
* We don't have to worry. Conceptually one call occurs first
|
|
|
|
* and we have a single winner.
|
|
|
|
*/
|
2005-05-01 15:59:00 +00:00
|
|
|
if (increment < -40)
|
|
|
|
increment = -40;
|
2005-04-16 22:20:36 +00:00
|
|
|
if (increment > 40)
|
|
|
|
increment = 40;
|
|
|
|
|
2009-02-16 10:54:21 +00:00
|
|
|
nice = TASK_NICE(current) + increment;
|
2005-04-16 22:20:36 +00:00
|
|
|
if (nice < -20)
|
|
|
|
nice = -20;
|
|
|
|
if (nice > 19)
|
|
|
|
nice = 19;
|
|
|
|
|
2005-05-01 15:59:00 +00:00
|
|
|
if (increment < 0 && !can_nice(current, nice))
|
|
|
|
return -EPERM;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
retval = security_task_setnice(current, nice);
|
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
set_user_nice(current, nice);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/**
|
|
|
|
* task_prio - return the priority value of a given task.
|
|
|
|
* @p: the task in question.
|
|
|
|
*
|
|
|
|
* This is the priority value as seen by users in /proc.
|
|
|
|
* RT tasks are offset by -200. Normal tasks are centered
|
|
|
|
* around 0, value goes from -16 to +15.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
int task_prio(const struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return p->prio - MAX_RT_PRIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* task_nice - return the nice value of a given task.
|
|
|
|
* @p: the task in question.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
int task_nice(const struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return TASK_NICE(p);
|
|
|
|
}
|
2008-03-05 21:56:37 +00:00
|
|
|
EXPORT_SYMBOL(task_nice);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* idle_cpu - is a given cpu idle currently?
|
|
|
|
* @cpu: the processor in question.
|
|
|
|
*/
|
|
|
|
int idle_cpu(int cpu)
|
|
|
|
{
|
|
|
|
return cpu_curr(cpu) == cpu_rq(cpu)->idle;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* idle_task - return the idle task for a given cpu.
|
|
|
|
* @cpu: the processor in question.
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *idle_task(int cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return cpu_rq(cpu)->idle;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* find_process_by_pid - find a process with a matching PID value.
|
|
|
|
* @pid: the pid in question.
|
|
|
|
*/
|
2007-10-15 15:00:13 +00:00
|
|
|
static struct task_struct *find_process_by_pid(pid_t pid)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-19 06:40:16 +00:00
|
|
|
return pid ? find_task_by_vpid(pid) : current;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Actually do priority change: must hold rq lock. */
|
2007-07-09 16:51:59 +00:00
|
|
|
static void
|
|
|
|
__setscheduler(struct rq *rq, struct task_struct *p, int policy, int prio)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
BUG_ON(p->se.on_rq);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
p->policy = policy;
|
2007-07-09 16:51:59 +00:00
|
|
|
switch (p->policy) {
|
|
|
|
case SCHED_NORMAL:
|
|
|
|
case SCHED_BATCH:
|
|
|
|
case SCHED_IDLE:
|
|
|
|
p->sched_class = &fair_sched_class;
|
|
|
|
break;
|
|
|
|
case SCHED_FIFO:
|
|
|
|
case SCHED_RR:
|
|
|
|
p->sched_class = &rt_sched_class;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
p->rt_priority = prio;
|
2006-06-27 09:54:51 +00:00
|
|
|
p->normal_prio = normal_prio(p);
|
|
|
|
/* we are holding p->pi_lock already */
|
|
|
|
p->prio = rt_mutex_getprio(p);
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
set_load_weight(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-11-13 23:39:19 +00:00
|
|
|
/*
|
|
|
|
* check the target process has a UID that matches the current process's
|
|
|
|
*/
|
|
|
|
static bool check_same_owner(struct task_struct *p)
|
|
|
|
{
|
|
|
|
const struct cred *cred = current_cred(), *pcred;
|
|
|
|
bool match;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
pcred = __task_cred(p);
|
|
|
|
match = (cred->euid == pcred->euid ||
|
|
|
|
cred->euid == pcred->uid);
|
|
|
|
rcu_read_unlock();
|
|
|
|
return match;
|
|
|
|
}
|
|
|
|
|
2008-06-23 03:55:38 +00:00
|
|
|
static int __sched_setscheduler(struct task_struct *p, int policy,
|
|
|
|
struct sched_param *param, bool user)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-15 15:00:08 +00:00
|
|
|
int retval, oldprio, oldpolicy = -1, on_rq, running;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2008-01-25 20:08:22 +00:00
|
|
|
const struct sched_class *prev_class = p->sched_class;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-06-27 09:54:44 +00:00
|
|
|
/* may grab non-irq protected spin_locks */
|
|
|
|
BUG_ON(in_interrupt());
|
2005-04-16 22:20:36 +00:00
|
|
|
recheck:
|
|
|
|
/* double check policy once rq lock held */
|
|
|
|
if (policy < 0)
|
|
|
|
policy = oldpolicy = p->policy;
|
|
|
|
else if (policy != SCHED_FIFO && policy != SCHED_RR &&
|
2007-07-09 16:51:59 +00:00
|
|
|
policy != SCHED_NORMAL && policy != SCHED_BATCH &&
|
|
|
|
policy != SCHED_IDLE)
|
2006-01-14 21:20:41 +00:00
|
|
|
return -EINVAL;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Valid priorities for SCHED_FIFO and SCHED_RR are
|
2007-07-09 16:51:59 +00:00
|
|
|
* 1..MAX_USER_RT_PRIO-1, valid priority for SCHED_NORMAL,
|
|
|
|
* SCHED_BATCH and SCHED_IDLE is 0.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
if (param->sched_priority < 0 ||
|
2005-09-10 07:26:11 +00:00
|
|
|
(p->mm && param->sched_priority > MAX_USER_RT_PRIO-1) ||
|
2005-07-25 20:28:39 +00:00
|
|
|
(!p->mm && param->sched_priority > MAX_RT_PRIO-1))
|
2005-04-16 22:20:36 +00:00
|
|
|
return -EINVAL;
|
2007-07-09 16:51:59 +00:00
|
|
|
if (rt_policy(policy) != (param->sched_priority != 0))
|
2005-04-16 22:20:36 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
[PATCH] Changing RT priority without CAP_SYS_NICE
Presently, a process without the capability CAP_SYS_NICE can not change
its own policy, which is OK.
But it can also not decrease its RT priority (if scheduled with policy
SCHED_RR or SCHED_FIFO), which is what this patch changes.
The rationale is the same as for the nice value: a process should be
able to require less priority for itself. Increasing the priority is
still not allowed.
This is for example useful if you give a multithreaded user process a RT
priority, and the process would like to organize its internal threads
using priorities also. Then you can give the process the highest
priority needed N, and the process starts its threads with lower
priorities: N-1, N-2...
The POSIX norm says that the permissions are implementation specific, so
I think we can do that.
In a sense, it makes the permissions consistent whatever the policy is:
with this patch, process scheduled by SCHED_FIFO, SCHED_RR and
SCHED_OTHER can all decrease their priority.
From: Ingo Molnar <mingo@elte.hu>
cleaned up and merged to -mm.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-06-25 21:57:32 +00:00
|
|
|
/*
|
|
|
|
* Allow unprivileged RT tasks to decrease priority:
|
|
|
|
*/
|
2008-06-23 03:55:38 +00:00
|
|
|
if (user && !capable(CAP_SYS_NICE)) {
|
2007-07-09 16:51:59 +00:00
|
|
|
if (rt_policy(policy)) {
|
2006-09-29 09:00:50 +00:00
|
|
|
unsigned long rlim_rtprio;
|
|
|
|
|
|
|
|
if (!lock_task_sighand(p, &flags))
|
|
|
|
return -ESRCH;
|
|
|
|
rlim_rtprio = p->signal->rlim[RLIMIT_RTPRIO].rlim_cur;
|
|
|
|
unlock_task_sighand(p, &flags);
|
|
|
|
|
|
|
|
/* can't set/change the rt policy */
|
|
|
|
if (policy != p->policy && !rlim_rtprio)
|
|
|
|
return -EPERM;
|
|
|
|
|
|
|
|
/* can't increase priority */
|
|
|
|
if (param->sched_priority > p->rt_priority &&
|
|
|
|
param->sched_priority > rlim_rtprio)
|
|
|
|
return -EPERM;
|
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* Like positive nice levels, dont allow tasks to
|
|
|
|
* move out of SCHED_IDLE either:
|
|
|
|
*/
|
|
|
|
if (p->policy == SCHED_IDLE && policy != SCHED_IDLE)
|
|
|
|
return -EPERM;
|
2006-09-29 09:00:48 +00:00
|
|
|
|
[PATCH] Changing RT priority without CAP_SYS_NICE
Presently, a process without the capability CAP_SYS_NICE can not change
its own policy, which is OK.
But it can also not decrease its RT priority (if scheduled with policy
SCHED_RR or SCHED_FIFO), which is what this patch changes.
The rationale is the same as for the nice value: a process should be
able to require less priority for itself. Increasing the priority is
still not allowed.
This is for example useful if you give a multithreaded user process a RT
priority, and the process would like to organize its internal threads
using priorities also. Then you can give the process the highest
priority needed N, and the process starts its threads with lower
priorities: N-1, N-2...
The POSIX norm says that the permissions are implementation specific, so
I think we can do that.
In a sense, it makes the permissions consistent whatever the policy is:
with this patch, process scheduled by SCHED_FIFO, SCHED_RR and
SCHED_OTHER can all decrease their priority.
From: Ingo Molnar <mingo@elte.hu>
cleaned up and merged to -mm.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-06-25 21:57:32 +00:00
|
|
|
/* can't change other user's priorities */
|
2008-11-13 23:39:19 +00:00
|
|
|
if (!check_same_owner(p))
|
[PATCH] Changing RT priority without CAP_SYS_NICE
Presently, a process without the capability CAP_SYS_NICE can not change
its own policy, which is OK.
But it can also not decrease its RT priority (if scheduled with policy
SCHED_RR or SCHED_FIFO), which is what this patch changes.
The rationale is the same as for the nice value: a process should be
able to require less priority for itself. Increasing the priority is
still not allowed.
This is for example useful if you give a multithreaded user process a RT
priority, and the process would like to organize its internal threads
using priorities also. Then you can give the process the highest
priority needed N, and the process starts its threads with lower
priorities: N-1, N-2...
The POSIX norm says that the permissions are implementation specific, so
I think we can do that.
In a sense, it makes the permissions consistent whatever the policy is:
with this patch, process scheduled by SCHED_FIFO, SCHED_RR and
SCHED_OTHER can all decrease their priority.
From: Ingo Molnar <mingo@elte.hu>
cleaned up and merged to -mm.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-06-25 21:57:32 +00:00
|
|
|
return -EPERM;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-08-03 16:33:03 +00:00
|
|
|
if (user) {
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-08-03 16:33:03 +00:00
|
|
|
/*
|
|
|
|
* Do not allow realtime tasks into groups that have no runtime
|
|
|
|
* assigned.
|
|
|
|
*/
|
2008-08-19 10:33:06 +00:00
|
|
|
if (rt_bandwidth_enabled() && rt_policy(policy) &&
|
|
|
|
task_group(p)->rt_bandwidth.rt_runtime == 0)
|
2008-08-03 16:33:03 +00:00
|
|
|
return -EPERM;
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
|
|
|
|
2008-08-03 16:33:03 +00:00
|
|
|
retval = security_task_setscheduler(p, policy, param);
|
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
/*
|
|
|
|
* make sure no PI-waiters arrive (or leave) while we are
|
|
|
|
* changing the priority of the task:
|
|
|
|
*/
|
|
|
|
spin_lock_irqsave(&p->pi_lock, flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* To be able to change p->policy safely, the apropriate
|
|
|
|
* runqueue lock must be held.
|
|
|
|
*/
|
2006-06-27 09:54:51 +00:00
|
|
|
rq = __task_rq_lock(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* recheck policy now with rq lock held */
|
|
|
|
if (unlikely(oldpolicy != -1 && oldpolicy != p->policy)) {
|
|
|
|
policy = oldpolicy = -1;
|
2006-06-27 09:54:51 +00:00
|
|
|
__task_rq_unlock(rq);
|
|
|
|
spin_unlock_irqrestore(&p->pi_lock, flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
goto recheck;
|
|
|
|
}
|
2007-08-09 09:16:51 +00:00
|
|
|
update_rq_clock(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
on_rq = p->se.on_rq;
|
2007-12-18 14:21:13 +00:00
|
|
|
running = task_current(rq, p);
|
2008-03-10 18:01:20 +00:00
|
|
|
if (on_rq)
|
2007-08-09 09:16:49 +00:00
|
|
|
deactivate_task(rq, p, 0);
|
2008-03-10 18:01:20 +00:00
|
|
|
if (running)
|
|
|
|
p->sched_class->put_prev_task(rq, p);
|
2007-10-15 15:00:08 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
oldprio = p->prio;
|
2007-07-09 16:51:59 +00:00
|
|
|
__setscheduler(rq, p, policy, param->sched_priority);
|
2007-10-15 15:00:08 +00:00
|
|
|
|
2008-03-10 18:01:20 +00:00
|
|
|
if (running)
|
|
|
|
p->sched_class->set_curr_task(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (on_rq) {
|
|
|
|
activate_task(rq, p, 0);
|
2008-01-25 20:08:22 +00:00
|
|
|
|
|
|
|
check_class_changed(rq, p, prev_class, oldprio, running);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-06-27 09:54:51 +00:00
|
|
|
__task_rq_unlock(rq);
|
|
|
|
spin_unlock_irqrestore(&p->pi_lock, flags);
|
|
|
|
|
2006-06-27 09:55:02 +00:00
|
|
|
rt_mutex_adjust_pi(p);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2008-06-23 03:55:38 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* sched_setscheduler - change the scheduling policy and/or RT priority of a thread.
|
|
|
|
* @p: the task in question.
|
|
|
|
* @policy: new policy.
|
|
|
|
* @param: structure containing the new RT priority.
|
|
|
|
*
|
|
|
|
* NOTE that the task may be already dead.
|
|
|
|
*/
|
|
|
|
int sched_setscheduler(struct task_struct *p, int policy,
|
|
|
|
struct sched_param *param)
|
|
|
|
{
|
|
|
|
return __sched_setscheduler(p, policy, param, true);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
EXPORT_SYMBOL_GPL(sched_setscheduler);
|
|
|
|
|
2008-06-23 03:55:38 +00:00
|
|
|
/**
|
|
|
|
* sched_setscheduler_nocheck - change the scheduling policy and/or RT priority of a thread from kernelspace.
|
|
|
|
* @p: the task in question.
|
|
|
|
* @policy: new policy.
|
|
|
|
* @param: structure containing the new RT priority.
|
|
|
|
*
|
|
|
|
* Just like sched_setscheduler, only don't bother checking if the
|
|
|
|
* current context has permission. For example, this is needed in
|
|
|
|
* stop_machine(): we create temporary high priority worker threads,
|
|
|
|
* but our caller might not have that capability.
|
|
|
|
*/
|
|
|
|
int sched_setscheduler_nocheck(struct task_struct *p, int policy,
|
|
|
|
struct sched_param *param)
|
|
|
|
{
|
|
|
|
return __sched_setscheduler(p, policy, param, false);
|
|
|
|
}
|
|
|
|
|
2005-09-10 07:26:11 +00:00
|
|
|
static int
|
|
|
|
do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sched_param lparam;
|
|
|
|
struct task_struct *p;
|
2006-07-03 07:25:41 +00:00
|
|
|
int retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (!param || pid < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
if (copy_from_user(&lparam, param, sizeof(struct sched_param)))
|
|
|
|
return -EFAULT;
|
2006-09-29 09:00:48 +00:00
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
retval = -ESRCH;
|
2005-04-16 22:20:36 +00:00
|
|
|
p = find_process_by_pid(pid);
|
2006-09-29 09:00:48 +00:00
|
|
|
if (p != NULL)
|
|
|
|
retval = sched_setscheduler(p, policy, &lparam);
|
|
|
|
rcu_read_unlock();
|
2006-07-03 07:25:41 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_setscheduler - set/change the scheduler policy and RT priority
|
|
|
|
* @pid: the pid in question.
|
|
|
|
* @policy: new policy.
|
|
|
|
* @param: structure containing the new RT priority.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy,
|
|
|
|
struct sched_param __user *, param)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-01-19 01:43:03 +00:00
|
|
|
/* negative values for policy are not valid */
|
|
|
|
if (policy < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return do_sched_setscheduler(pid, policy, param);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_setparam - set/change the RT priority of a thread
|
|
|
|
* @pid: the pid in question.
|
|
|
|
* @param: structure containing the new RT priority.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
return do_sched_setscheduler(pid, -1, param);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_getscheduler - get the policy (scheduling class) of a thread
|
|
|
|
* @pid: the pid in question.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *p;
|
2007-10-15 15:00:14 +00:00
|
|
|
int retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (pid < 0)
|
2007-10-15 15:00:14 +00:00
|
|
|
return -EINVAL;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
retval = -ESRCH;
|
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
p = find_process_by_pid(pid);
|
|
|
|
if (p) {
|
|
|
|
retval = security_task_getscheduler(p);
|
|
|
|
if (!retval)
|
|
|
|
retval = p->policy;
|
|
|
|
}
|
|
|
|
read_unlock(&tasklist_lock);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_getscheduler - get the RT priority of a thread
|
|
|
|
* @pid: the pid in question.
|
|
|
|
* @param: structure containing the RT priority.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE2(sched_getparam, pid_t, pid, struct sched_param __user *, param)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sched_param lp;
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *p;
|
2007-10-15 15:00:14 +00:00
|
|
|
int retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (!param || pid < 0)
|
2007-10-15 15:00:14 +00:00
|
|
|
return -EINVAL;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
p = find_process_by_pid(pid);
|
|
|
|
retval = -ESRCH;
|
|
|
|
if (!p)
|
|
|
|
goto out_unlock;
|
|
|
|
|
|
|
|
retval = security_task_getscheduler(p);
|
|
|
|
if (retval)
|
|
|
|
goto out_unlock;
|
|
|
|
|
|
|
|
lp.sched_priority = p->rt_priority;
|
|
|
|
read_unlock(&tasklist_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This one might sleep, we cannot do it with a spinlock held ...
|
|
|
|
*/
|
|
|
|
retval = copy_to_user(param, &lp, sizeof(*param)) ? -EFAULT : 0;
|
|
|
|
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
out_unlock:
|
|
|
|
read_unlock(&tasklist_lock);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
long sched_setaffinity(pid_t pid, const struct cpumask *in_mask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-11-24 16:05:11 +00:00
|
|
|
cpumask_var_t cpus_allowed, new_mask;
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *p;
|
|
|
|
int retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-01-25 20:08:02 +00:00
|
|
|
get_online_cpus();
|
2005-04-16 22:20:36 +00:00
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
|
|
|
|
p = find_process_by_pid(pid);
|
|
|
|
if (!p) {
|
|
|
|
read_unlock(&tasklist_lock);
|
2008-01-25 20:08:02 +00:00
|
|
|
put_online_cpus();
|
2005-04-16 22:20:36 +00:00
|
|
|
return -ESRCH;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* It is not safe to call set_cpus_allowed with the
|
2007-12-05 14:46:09 +00:00
|
|
|
* tasklist_lock held. We will bump the task_struct's
|
2005-04-16 22:20:36 +00:00
|
|
|
* usage count and then drop tasklist_lock.
|
|
|
|
*/
|
|
|
|
get_task_struct(p);
|
|
|
|
read_unlock(&tasklist_lock);
|
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) {
|
|
|
|
retval = -ENOMEM;
|
|
|
|
goto out_put_task;
|
|
|
|
}
|
|
|
|
if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) {
|
|
|
|
retval = -ENOMEM;
|
|
|
|
goto out_free_cpus_allowed;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
retval = -EPERM;
|
2008-11-13 23:39:19 +00:00
|
|
|
if (!check_same_owner(p) && !capable(CAP_SYS_NICE))
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out_unlock;
|
|
|
|
|
2006-06-23 09:03:59 +00:00
|
|
|
retval = security_task_setscheduler(p, 0, NULL);
|
|
|
|
if (retval)
|
|
|
|
goto out_unlock;
|
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
cpuset_cpus_allowed(p, cpus_allowed);
|
|
|
|
cpumask_and(new_mask, in_mask, cpus_allowed);
|
2007-10-19 06:40:22 +00:00
|
|
|
again:
|
2008-11-24 16:05:11 +00:00
|
|
|
retval = set_cpus_allowed_ptr(p, new_mask);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-19 06:40:22 +00:00
|
|
|
if (!retval) {
|
2008-11-24 16:05:11 +00:00
|
|
|
cpuset_cpus_allowed(p, cpus_allowed);
|
|
|
|
if (!cpumask_subset(new_mask, cpus_allowed)) {
|
2007-10-19 06:40:22 +00:00
|
|
|
/*
|
|
|
|
* We must have raced with a concurrent cpuset
|
|
|
|
* update. Just reset the cpus_allowed to the
|
|
|
|
* cpuset's cpus_allowed
|
|
|
|
*/
|
2008-11-24 16:05:11 +00:00
|
|
|
cpumask_copy(new_mask, cpus_allowed);
|
2007-10-19 06:40:22 +00:00
|
|
|
goto again;
|
|
|
|
}
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
out_unlock:
|
2008-11-24 16:05:11 +00:00
|
|
|
free_cpumask_var(new_mask);
|
|
|
|
out_free_cpus_allowed:
|
|
|
|
free_cpumask_var(cpus_allowed);
|
|
|
|
out_put_task:
|
2005-04-16 22:20:36 +00:00
|
|
|
put_task_struct(p);
|
2008-01-25 20:08:02 +00:00
|
|
|
put_online_cpus();
|
2005-04-16 22:20:36 +00:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int get_user_cpu_mask(unsigned long __user *user_mask_ptr, unsigned len,
|
2008-11-24 16:05:14 +00:00
|
|
|
struct cpumask *new_mask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-11-24 16:05:14 +00:00
|
|
|
if (len < cpumask_size())
|
|
|
|
cpumask_clear(new_mask);
|
|
|
|
else if (len > cpumask_size())
|
|
|
|
len = cpumask_size();
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return copy_from_user(new_mask, user_mask_ptr, len) ? -EFAULT : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_setaffinity - set the cpu affinity of a process
|
|
|
|
* @pid: pid of the process
|
|
|
|
* @len: length in bytes of the bitmask pointed to by user_mask_ptr
|
|
|
|
* @user_mask_ptr: user-space pointer to the new cpu mask
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE3(sched_setaffinity, pid_t, pid, unsigned int, len,
|
|
|
|
unsigned long __user *, user_mask_ptr)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-11-24 16:05:11 +00:00
|
|
|
cpumask_var_t new_mask;
|
2005-04-16 22:20:36 +00:00
|
|
|
int retval;
|
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
if (!alloc_cpumask_var(&new_mask, GFP_KERNEL))
|
|
|
|
return -ENOMEM;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
retval = get_user_cpu_mask(user_mask_ptr, len, new_mask);
|
|
|
|
if (retval == 0)
|
|
|
|
retval = sched_setaffinity(pid, new_mask);
|
|
|
|
free_cpumask_var(new_mask);
|
|
|
|
return retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
long sched_getaffinity(pid_t pid, struct cpumask *mask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *p;
|
2005-04-16 22:20:36 +00:00
|
|
|
int retval;
|
|
|
|
|
2008-01-25 20:08:02 +00:00
|
|
|
get_online_cpus();
|
2005-04-16 22:20:36 +00:00
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
|
|
|
|
retval = -ESRCH;
|
|
|
|
p = find_process_by_pid(pid);
|
|
|
|
if (!p)
|
|
|
|
goto out_unlock;
|
|
|
|
|
2006-06-23 09:03:59 +00:00
|
|
|
retval = security_task_getscheduler(p);
|
|
|
|
if (retval)
|
|
|
|
goto out_unlock;
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_and(mask, &p->cpus_allowed, cpu_online_mask);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
out_unlock:
|
|
|
|
read_unlock(&tasklist_lock);
|
2008-01-25 20:08:02 +00:00
|
|
|
put_online_cpus();
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-08-09 09:16:46 +00:00
|
|
|
return retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_getaffinity - get the cpu affinity of a process
|
|
|
|
* @pid: pid of the process
|
|
|
|
* @len: length in bytes of the bitmask pointed to by user_mask_ptr
|
|
|
|
* @user_mask_ptr: user-space pointer to hold the current cpu mask
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len,
|
|
|
|
unsigned long __user *, user_mask_ptr)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int ret;
|
2008-11-24 16:05:11 +00:00
|
|
|
cpumask_var_t mask;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
if (len < cpumask_size())
|
2005-04-16 22:20:36 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
if (!alloc_cpumask_var(&mask, GFP_KERNEL))
|
|
|
|
return -ENOMEM;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
ret = sched_getaffinity(pid, mask);
|
|
|
|
if (ret == 0) {
|
|
|
|
if (copy_to_user(user_mask_ptr, mask, cpumask_size()))
|
|
|
|
ret = -EFAULT;
|
|
|
|
else
|
|
|
|
ret = cpumask_size();
|
|
|
|
}
|
|
|
|
free_cpumask_var(mask);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_yield - yield the current processor to other threads.
|
|
|
|
*
|
2007-07-09 16:51:59 +00:00
|
|
|
* This function yields the current CPU to other tasks. If there are no
|
|
|
|
* other threads running on this CPU then this function will return.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE0(sched_yield)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = this_rq_lock();
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:12 +00:00
|
|
|
schedstat_inc(rq, yld_count);
|
2007-10-15 15:00:08 +00:00
|
|
|
current->sched_class->yield_task(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Since we are going to call schedule() anyway, there's
|
|
|
|
* no need to preempt or enable interrupts:
|
|
|
|
*/
|
|
|
|
__release(rq->lock);
|
2006-07-03 07:24:54 +00:00
|
|
|
spin_release(&rq->lock.dep_map, 1, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
_raw_spin_unlock(&rq->lock);
|
|
|
|
preempt_enable_no_resched();
|
|
|
|
|
|
|
|
schedule();
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2006-06-30 08:56:00 +00:00
|
|
|
static void __cond_resched(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-06-23 09:05:23 +00:00
|
|
|
#ifdef CONFIG_DEBUG_SPINLOCK_SLEEP
|
|
|
|
__might_sleep(__FILE__, __LINE__);
|
|
|
|
#endif
|
2005-07-08 00:57:04 +00:00
|
|
|
/*
|
|
|
|
* The BKS might be reacquired before we have dropped
|
|
|
|
* PREEMPT_ACTIVE, which could trigger a second
|
|
|
|
* cond_resched() call.
|
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
do {
|
|
|
|
add_preempt_count(PREEMPT_ACTIVE);
|
|
|
|
schedule();
|
|
|
|
sub_preempt_count(PREEMPT_ACTIVE);
|
|
|
|
} while (need_resched());
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:28 +00:00
|
|
|
int __sched _cond_resched(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-12-30 00:48:13 +00:00
|
|
|
if (need_resched() && !(preempt_count() & PREEMPT_ACTIVE) &&
|
|
|
|
system_state == SYSTEM_RUNNING) {
|
2005-04-16 22:20:36 +00:00
|
|
|
__cond_resched();
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2008-01-25 20:08:28 +00:00
|
|
|
EXPORT_SYMBOL(_cond_resched);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* cond_resched_lock() - if a reschedule is pending, drop the given lock,
|
|
|
|
* call schedule, and on return reacquire the lock.
|
|
|
|
*
|
2007-12-05 14:46:09 +00:00
|
|
|
* This works OK both with and without CONFIG_PREEMPT. We do strange low-level
|
2005-04-16 22:20:36 +00:00
|
|
|
* operations here to prevent schedule() from being called twice (once via
|
|
|
|
* spin_unlock(), once by hand).
|
|
|
|
*/
|
2005-09-10 07:26:11 +00:00
|
|
|
int cond_resched_lock(spinlock_t *lock)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-01-30 12:31:20 +00:00
|
|
|
int resched = need_resched() && system_state == SYSTEM_RUNNING;
|
2005-06-13 22:52:32 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
2008-01-30 12:31:20 +00:00
|
|
|
if (spin_needbreak(lock) || resched) {
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock(lock);
|
2008-01-30 12:31:20 +00:00
|
|
|
if (resched && need_resched())
|
|
|
|
__cond_resched();
|
|
|
|
else
|
|
|
|
cpu_relax();
|
2005-06-13 22:52:32 +00:00
|
|
|
ret = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_lock(lock);
|
|
|
|
}
|
2005-06-13 22:52:32 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(cond_resched_lock);
|
|
|
|
|
|
|
|
int __sched cond_resched_softirq(void)
|
|
|
|
{
|
|
|
|
BUG_ON(!in_softirq());
|
|
|
|
|
2006-12-30 00:48:13 +00:00
|
|
|
if (need_resched() && system_state == SYSTEM_RUNNING) {
|
2007-05-23 20:58:18 +00:00
|
|
|
local_bh_enable();
|
2005-04-16 22:20:36 +00:00
|
|
|
__cond_resched();
|
|
|
|
local_bh_disable();
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(cond_resched_softirq);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* yield - yield the current processor to other threads.
|
|
|
|
*
|
2007-02-10 09:45:59 +00:00
|
|
|
* This is a shortcut for kernel-space yielding - it marks the
|
2005-04-16 22:20:36 +00:00
|
|
|
* thread runnable and calls sys_sched_yield().
|
|
|
|
*/
|
|
|
|
void __sched yield(void)
|
|
|
|
{
|
|
|
|
set_current_state(TASK_RUNNING);
|
|
|
|
sys_sched_yield();
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(yield);
|
|
|
|
|
|
|
|
/*
|
2007-12-05 14:46:09 +00:00
|
|
|
* This task is about to go to sleep on IO. Increment rq->nr_iowait so
|
2005-04-16 22:20:36 +00:00
|
|
|
* that process accounting knows that this is a task in IO wait state.
|
|
|
|
*
|
|
|
|
* But don't do that if it is a deliberate, throttling IO wait (this task
|
|
|
|
* has set its backing_dev_info: the queue against which it should throttle)
|
|
|
|
*/
|
|
|
|
void __sched io_schedule(void)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = &__raw_get_cpu_var(runqueues);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-14 07:24:37 +00:00
|
|
|
delayacct_blkio_start();
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_inc(&rq->nr_iowait);
|
|
|
|
schedule();
|
|
|
|
atomic_dec(&rq->nr_iowait);
|
2006-07-14 07:24:37 +00:00
|
|
|
delayacct_blkio_end();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(io_schedule);
|
|
|
|
|
|
|
|
long __sched io_schedule_timeout(long timeout)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = &__raw_get_cpu_var(runqueues);
|
2005-04-16 22:20:36 +00:00
|
|
|
long ret;
|
|
|
|
|
2006-07-14 07:24:37 +00:00
|
|
|
delayacct_blkio_start();
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_inc(&rq->nr_iowait);
|
|
|
|
ret = schedule_timeout(timeout);
|
|
|
|
atomic_dec(&rq->nr_iowait);
|
2006-07-14 07:24:37 +00:00
|
|
|
delayacct_blkio_end();
|
2005-04-16 22:20:36 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_get_priority_max - return maximum RT priority.
|
|
|
|
* @policy: scheduling class.
|
|
|
|
*
|
|
|
|
* this syscall returns the maximum rt_priority that can be used
|
|
|
|
* by a given scheduling class.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE1(sched_get_priority_max, int, policy)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int ret = -EINVAL;
|
|
|
|
|
|
|
|
switch (policy) {
|
|
|
|
case SCHED_FIFO:
|
|
|
|
case SCHED_RR:
|
|
|
|
ret = MAX_USER_RT_PRIO-1;
|
|
|
|
break;
|
|
|
|
case SCHED_NORMAL:
|
2006-01-14 21:20:41 +00:00
|
|
|
case SCHED_BATCH:
|
2007-07-09 16:51:59 +00:00
|
|
|
case SCHED_IDLE:
|
2005-04-16 22:20:36 +00:00
|
|
|
ret = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_get_priority_min - return minimum RT priority.
|
|
|
|
* @policy: scheduling class.
|
|
|
|
*
|
|
|
|
* this syscall returns the minimum rt_priority that can be used
|
|
|
|
* by a given scheduling class.
|
|
|
|
*/
|
2009-01-14 13:14:08 +00:00
|
|
|
SYSCALL_DEFINE1(sched_get_priority_min, int, policy)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int ret = -EINVAL;
|
|
|
|
|
|
|
|
switch (policy) {
|
|
|
|
case SCHED_FIFO:
|
|
|
|
case SCHED_RR:
|
|
|
|
ret = 1;
|
|
|
|
break;
|
|
|
|
case SCHED_NORMAL:
|
2006-01-14 21:20:41 +00:00
|
|
|
case SCHED_BATCH:
|
2007-07-09 16:51:59 +00:00
|
|
|
case SCHED_IDLE:
|
2005-04-16 22:20:36 +00:00
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sys_sched_rr_get_interval - return the default timeslice of a process.
|
|
|
|
* @pid: pid of the process.
|
|
|
|
* @interval: userspace pointer to the timeslice value.
|
|
|
|
*
|
|
|
|
* this syscall writes the default timeslice value of a given process
|
|
|
|
* into the user-space timespec buffer. A value of '0' means infinity.
|
|
|
|
*/
|
2009-01-14 13:14:10 +00:00
|
|
|
SYSCALL_DEFINE2(sched_rr_get_interval, pid_t, pid,
|
2009-01-14 13:14:09 +00:00
|
|
|
struct timespec __user *, interval)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *p;
|
2007-10-15 15:00:13 +00:00
|
|
|
unsigned int time_slice;
|
2007-10-15 15:00:14 +00:00
|
|
|
int retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct timespec t;
|
|
|
|
|
|
|
|
if (pid < 0)
|
2007-10-15 15:00:14 +00:00
|
|
|
return -EINVAL;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
retval = -ESRCH;
|
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
p = find_process_by_pid(pid);
|
|
|
|
if (!p)
|
|
|
|
goto out_unlock;
|
|
|
|
|
|
|
|
retval = security_task_getscheduler(p);
|
|
|
|
if (retval)
|
|
|
|
goto out_unlock;
|
|
|
|
|
2007-12-04 16:04:39 +00:00
|
|
|
/*
|
|
|
|
* Time slice is 0 for SCHED_FIFO tasks and for SCHED_OTHER
|
|
|
|
* tasks that are on an otherwise idle runqueue:
|
|
|
|
*/
|
|
|
|
time_slice = 0;
|
|
|
|
if (p->policy == SCHED_RR) {
|
2007-10-15 15:00:13 +00:00
|
|
|
time_slice = DEF_TIMESLICE;
|
2008-03-07 01:35:06 +00:00
|
|
|
} else if (p->policy != SCHED_FIFO) {
|
2007-10-15 15:00:13 +00:00
|
|
|
struct sched_entity *se = &p->se;
|
|
|
|
unsigned long flags;
|
|
|
|
struct rq *rq;
|
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2007-12-04 16:04:39 +00:00
|
|
|
if (rq->cfs.load.weight)
|
|
|
|
time_slice = NS_TO_JIFFIES(sched_slice(&rq->cfs, se));
|
2007-10-15 15:00:13 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
read_unlock(&tasklist_lock);
|
2007-10-15 15:00:13 +00:00
|
|
|
jiffies_to_timespec(time_slice, &t);
|
2005-04-16 22:20:36 +00:00
|
|
|
retval = copy_to_user(interval, &t, sizeof(t)) ? -EFAULT : 0;
|
|
|
|
return retval;
|
2007-10-15 15:00:14 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
out_unlock:
|
|
|
|
read_unlock(&tasklist_lock);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2008-05-12 19:20:41 +00:00
|
|
|
static const char stat_nam[] = TASK_STATE_TO_CHAR_STR;
|
2006-07-03 07:25:41 +00:00
|
|
|
|
2008-01-25 20:08:02 +00:00
|
|
|
void sched_show_task(struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long free = 0;
|
2006-07-03 07:25:41 +00:00
|
|
|
unsigned state;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
state = p->state ? __ffs(p->state) + 1 : 0;
|
2007-10-18 19:32:56 +00:00
|
|
|
printk(KERN_INFO "%-13.13s %c", p->comm,
|
2006-07-10 11:43:52 +00:00
|
|
|
state < sizeof(stat_nam) - 1 ? stat_nam[state] : '?');
|
2007-07-11 19:21:47 +00:00
|
|
|
#if BITS_PER_LONG == 32
|
2005-04-16 22:20:36 +00:00
|
|
|
if (state == TASK_RUNNING)
|
2007-10-18 19:32:56 +00:00
|
|
|
printk(KERN_CONT " running ");
|
2005-04-16 22:20:36 +00:00
|
|
|
else
|
2007-10-18 19:32:56 +00:00
|
|
|
printk(KERN_CONT " %08lx ", thread_saved_pc(p));
|
2005-04-16 22:20:36 +00:00
|
|
|
#else
|
|
|
|
if (state == TASK_RUNNING)
|
2007-10-18 19:32:56 +00:00
|
|
|
printk(KERN_CONT " running task ");
|
2005-04-16 22:20:36 +00:00
|
|
|
else
|
2007-10-18 19:32:56 +00:00
|
|
|
printk(KERN_CONT " %016lx ", thread_saved_pc(p));
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_DEBUG_STACK_USAGE
|
2008-04-22 21:38:23 +00:00
|
|
|
free = stack_not_used(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2009-05-04 08:38:05 +00:00
|
|
|
printk(KERN_CONT "%5lu %5d %6d 0x%08lx\n", free,
|
|
|
|
task_pid_nr(p), task_pid_nr(p->real_parent),
|
|
|
|
(unsigned long)task_thread_info(p)->flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-01-25 20:08:34 +00:00
|
|
|
show_stack(p, NULL);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2006-12-07 04:35:59 +00:00
|
|
|
void show_state_filter(unsigned long state_filter)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *g, *p;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-11 19:21:47 +00:00
|
|
|
#if BITS_PER_LONG == 32
|
|
|
|
printk(KERN_INFO
|
|
|
|
" task PC stack pid father\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
#else
|
2007-07-11 19:21:47 +00:00
|
|
|
printk(KERN_INFO
|
|
|
|
" task PC stack pid father\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
read_lock(&tasklist_lock);
|
|
|
|
do_each_thread(g, p) {
|
|
|
|
/*
|
|
|
|
* reset the NMI-timeout, listing all files on a slow
|
|
|
|
* console might take alot of time:
|
|
|
|
*/
|
|
|
|
touch_nmi_watchdog();
|
2007-04-26 03:50:03 +00:00
|
|
|
if (!state_filter || (p->state & state_filter))
|
2008-01-25 20:08:02 +00:00
|
|
|
sched_show_task(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
} while_each_thread(g, p);
|
|
|
|
|
2007-05-08 07:28:05 +00:00
|
|
|
touch_all_softlockup_watchdogs();
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
|
sysrq_sched_debug_show();
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
read_unlock(&tasklist_lock);
|
2006-12-07 04:35:59 +00:00
|
|
|
/*
|
|
|
|
* Only show locks if all tasks are dumped:
|
|
|
|
*/
|
|
|
|
if (state_filter == -1)
|
|
|
|
debug_show_all_locks();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:58 +00:00
|
|
|
void __cpuinit init_idle_bootup_task(struct task_struct *idle)
|
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
idle->sched_class = &idle_sched_class;
|
2007-07-09 16:51:58 +00:00
|
|
|
}
|
|
|
|
|
2005-06-28 14:40:42 +00:00
|
|
|
/**
|
|
|
|
* init_idle - set up an idle thread for a given CPU
|
|
|
|
* @idle: task in question
|
|
|
|
* @cpu: cpu the idle task belongs to
|
|
|
|
*
|
|
|
|
* NOTE: this function does not set the idle thread's NEED_RESCHED
|
|
|
|
* flag, to make booting more robust.
|
|
|
|
*/
|
2006-10-03 08:14:04 +00:00
|
|
|
void __cpuinit init_idle(struct task_struct *idle, int cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
|
|
|
|
2008-11-12 19:05:50 +00:00
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
__sched_fork(idle);
|
|
|
|
idle->se.exec_start = sched_clock();
|
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
idle->prio = idle->normal_prio = MAX_PRIO;
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_copy(&idle->cpus_allowed, cpumask_of(cpu));
|
2007-07-09 16:51:59 +00:00
|
|
|
__set_task_cpu(idle, cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq->curr = rq->idle = idle;
|
2005-06-25 21:57:23 +00:00
|
|
|
#if defined(CONFIG_SMP) && defined(__ARCH_WANT_UNLOCKED_CTXSW)
|
|
|
|
idle->oncpu = 1;
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
|
|
|
|
/* Set the preempt count _outside_ the spinlocks! */
|
2008-05-11 03:58:02 +00:00
|
|
|
#if defined(CONFIG_PREEMPT)
|
|
|
|
task_thread_info(idle)->preempt_count = (idle->lock_depth >= 0);
|
|
|
|
#else
|
2005-11-14 00:06:55 +00:00
|
|
|
task_thread_info(idle)->preempt_count = 0;
|
2008-05-11 03:58:02 +00:00
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* The idle tasks have their own, simple scheduling class:
|
|
|
|
*/
|
|
|
|
idle->sched_class = &idle_sched_class;
|
2008-11-25 20:07:04 +00:00
|
|
|
ftrace_graph_init_task(idle);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In a system that switches off the HZ timer nohz_cpu_mask
|
|
|
|
* indicates which cpus entered this state. This is used
|
|
|
|
* in the rcu update to wait only for active cpus. For system
|
|
|
|
* which do not switch off the HZ timer nohz_cpu_mask should
|
2008-11-24 16:05:04 +00:00
|
|
|
* always be CPU_BITS_NONE.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_var_t nohz_cpu_mask;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-11-09 21:39:38 +00:00
|
|
|
/*
|
|
|
|
* Increase the granularity value when there are more CPUs,
|
|
|
|
* because with more CPUs the 'effective latency' as visible
|
|
|
|
* to users decreases. But the relationship is not linear,
|
|
|
|
* so pick a second-best guess by going with the log2 of the
|
|
|
|
* number of CPUs.
|
|
|
|
*
|
|
|
|
* This idea comes from the SD scheduler of Con Kolivas:
|
|
|
|
*/
|
|
|
|
static inline void sched_init_granularity(void)
|
|
|
|
{
|
|
|
|
unsigned int factor = 1 + ilog2(num_online_cpus());
|
|
|
|
const unsigned long limit = 200000000;
|
|
|
|
|
|
|
|
sysctl_sched_min_granularity *= factor;
|
|
|
|
if (sysctl_sched_min_granularity > limit)
|
|
|
|
sysctl_sched_min_granularity = limit;
|
|
|
|
|
|
|
|
sysctl_sched_latency *= factor;
|
|
|
|
if (sysctl_sched_latency > limit)
|
|
|
|
sysctl_sched_latency = limit;
|
|
|
|
|
|
|
|
sysctl_sched_wakeup_granularity *= factor;
|
2008-08-04 06:54:26 +00:00
|
|
|
|
|
|
|
sysctl_sched_shares_ratelimit *= factor;
|
2007-11-09 21:39:38 +00:00
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/*
|
|
|
|
* This is how migration works:
|
|
|
|
*
|
2006-07-03 07:25:42 +00:00
|
|
|
* 1) we queue a struct migration_req structure in the source CPU's
|
2005-04-16 22:20:36 +00:00
|
|
|
* runqueue and wake up that CPU's migration thread.
|
|
|
|
* 2) we down() the locked semaphore => thread blocks.
|
|
|
|
* 3) migration thread wakes up (implicitly it forces the migrated
|
|
|
|
* thread off the CPU)
|
|
|
|
* 4) it gets the migration request and checks whether the migrated
|
|
|
|
* task is still in the wrong runqueue.
|
|
|
|
* 5) if it's in the wrong runqueue then the migration thread removes
|
|
|
|
* it and puts it into the right queue.
|
|
|
|
* 6) migration thread up()s the semaphore.
|
|
|
|
* 7) we wake up and the migration is done.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Change a given task's CPU affinity. Migrate the thread to a
|
|
|
|
* proper CPU and schedule it away if the CPU it's executing on
|
|
|
|
* is removed from the allowed bitmask.
|
|
|
|
*
|
|
|
|
* NOTE: the caller must have a valid reference to the task, the
|
2007-12-05 14:46:09 +00:00
|
|
|
* task must not exit() & deallocate itself prematurely. The
|
2005-04-16 22:20:36 +00:00
|
|
|
* call is not atomic; no spinlocks may be held.
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req req;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2006-07-03 07:25:40 +00:00
|
|
|
int ret = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_intersects(new_mask, cpu_online_mask)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
ret = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2008-06-05 19:57:11 +00:00
|
|
|
if (unlikely((p->flags & PF_THREAD_BOUND) && p != current &&
|
2008-11-24 16:05:14 +00:00
|
|
|
!cpumask_equal(&p->cpus_allowed, new_mask))) {
|
2008-06-05 19:57:11 +00:00
|
|
|
ret = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:07 +00:00
|
|
|
if (p->sched_class->set_cpus_allowed)
|
2008-03-26 21:23:49 +00:00
|
|
|
p->sched_class->set_cpus_allowed(p, new_mask);
|
2008-01-25 20:08:07 +00:00
|
|
|
else {
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_copy(&p->cpus_allowed, new_mask);
|
|
|
|
p->rt.nr_cpus_allowed = cpumask_weight(new_mask);
|
2008-01-25 20:08:07 +00:00
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Can the task run on the task's current CPU? If so, we're done */
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_test_cpu(task_cpu(p), new_mask))
|
2005-04-16 22:20:36 +00:00
|
|
|
goto out;
|
|
|
|
|
2008-11-24 16:05:03 +00:00
|
|
|
if (migrate_task(p, cpumask_any_and(cpu_online_mask, new_mask), &req)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Need help from migration thread: drop lock and wait. */
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
wake_up_process(rq->migration_thread);
|
|
|
|
wait_for_completion(&req.done);
|
|
|
|
tlb_migrate_finish(p->mm);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
task_rq_unlock(rq, &flags);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return ret;
|
|
|
|
}
|
2008-03-26 21:23:49 +00:00
|
|
|
EXPORT_SYMBOL_GPL(set_cpus_allowed_ptr);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2007-12-05 14:46:09 +00:00
|
|
|
* Move (not current) task off this cpu, onto dest cpu. We're doing
|
2005-04-16 22:20:36 +00:00
|
|
|
* this because either it can't run here any more (set_cpus_allowed()
|
|
|
|
* away from this CPU, or CPU going down), or because we're
|
|
|
|
* attempting to rebalance this task on exec (sched_exec).
|
|
|
|
*
|
|
|
|
* So we race with normal scheduler movements, but that's OK, as long
|
|
|
|
* as the task is no longer on this CPU.
|
2006-06-27 09:54:32 +00:00
|
|
|
*
|
|
|
|
* Returns non-zero if task was successfully migrated.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2006-06-27 09:54:32 +00:00
|
|
|
static int __migrate_task(struct task_struct *p, int src_cpu, int dest_cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq_dest, *rq_src;
|
2007-07-09 16:51:59 +00:00
|
|
|
int ret = 0, on_rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-07-15 11:43:49 +00:00
|
|
|
if (unlikely(!cpu_active(dest_cpu)))
|
2006-06-27 09:54:32 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq_src = cpu_rq(src_cpu);
|
|
|
|
rq_dest = cpu_rq(dest_cpu);
|
|
|
|
|
|
|
|
double_rq_lock(rq_src, rq_dest);
|
|
|
|
/* Already moved. */
|
|
|
|
if (task_cpu(p) != src_cpu)
|
2008-07-10 18:25:03 +00:00
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Affinity changed (again). */
|
2008-11-24 16:05:14 +00:00
|
|
|
if (!cpumask_test_cpu(dest_cpu, &p->cpus_allowed))
|
2008-07-10 18:25:03 +00:00
|
|
|
goto fail;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
on_rq = p->se.on_rq;
|
2007-08-09 09:16:51 +00:00
|
|
|
if (on_rq)
|
2007-08-09 09:16:49 +00:00
|
|
|
deactivate_task(rq_src, p, 0);
|
2007-08-09 09:16:51 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
set_task_cpu(p, dest_cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (on_rq) {
|
|
|
|
activate_task(rq_dest, p, 0);
|
2008-09-20 21:38:02 +00:00
|
|
|
check_preempt_curr(rq_dest, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-07-10 18:25:03 +00:00
|
|
|
done:
|
2006-06-27 09:54:32 +00:00
|
|
|
ret = 1;
|
2008-07-10 18:25:03 +00:00
|
|
|
fail:
|
2005-04-16 22:20:36 +00:00
|
|
|
double_rq_unlock(rq_src, rq_dest);
|
2006-06-27 09:54:32 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* migration_thread - this is a highprio system thread that performs
|
|
|
|
* thread migration by bumping thread off CPU then 'pushing' onto
|
|
|
|
* another runqueue.
|
|
|
|
*/
|
2005-09-10 07:26:11 +00:00
|
|
|
static int migration_thread(void *data)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int cpu = (long)data;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
BUG_ON(rq->migration_thread != current);
|
|
|
|
|
|
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
|
|
while (!kthread_should_stop()) {
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req *req;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct list_head *head;
|
|
|
|
|
|
|
|
spin_lock_irq(&rq->lock);
|
|
|
|
|
|
|
|
if (cpu_is_offline(cpu)) {
|
|
|
|
spin_unlock_irq(&rq->lock);
|
2009-06-17 23:27:45 +00:00
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (rq->active_balance) {
|
|
|
|
active_load_balance(rq, cpu);
|
|
|
|
rq->active_balance = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
head = &rq->migration_queue;
|
|
|
|
|
|
|
|
if (list_empty(head)) {
|
|
|
|
spin_unlock_irq(&rq->lock);
|
|
|
|
schedule();
|
|
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
|
|
continue;
|
|
|
|
}
|
2006-07-03 07:25:42 +00:00
|
|
|
req = list_entry(head->next, struct migration_req, list);
|
2005-04-16 22:20:36 +00:00
|
|
|
list_del_init(head->next);
|
|
|
|
|
2005-06-25 21:57:27 +00:00
|
|
|
spin_unlock(&rq->lock);
|
|
|
|
__migrate_task(req->task, cpu, req->dest_cpu);
|
|
|
|
local_irq_enable();
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
complete(&req->done);
|
|
|
|
}
|
|
|
|
__set_current_state(TASK_RUNNING);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
2007-10-17 06:30:56 +00:00
|
|
|
|
|
|
|
static int __migrate_task_irq(struct task_struct *p, int src_cpu, int dest_cpu)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
local_irq_disable();
|
|
|
|
ret = __migrate_task(p, src_cpu, dest_cpu);
|
|
|
|
local_irq_enable();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2006-12-10 10:20:11 +00:00
|
|
|
/*
|
2007-10-19 21:10:43 +00:00
|
|
|
* Figure out where task on dead CPU should go, use force if necessary.
|
2006-12-10 10:20:11 +00:00
|
|
|
*/
|
2006-07-03 07:25:40 +00:00
|
|
|
static void move_task_off_dead_cpu(int dead_cpu, struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
int dest_cpu;
|
2009-01-01 02:08:45 +00:00
|
|
|
const struct cpumask *nodemask = cpumask_of_node(cpu_to_node(dead_cpu));
|
2008-11-24 16:05:11 +00:00
|
|
|
|
|
|
|
again:
|
|
|
|
/* Look for allowed, online CPU in same node. */
|
|
|
|
for_each_cpu_and(dest_cpu, nodemask, cpu_online_mask)
|
|
|
|
if (cpumask_test_cpu(dest_cpu, &p->cpus_allowed))
|
|
|
|
goto move;
|
|
|
|
|
|
|
|
/* Any allowed, online CPU? */
|
|
|
|
dest_cpu = cpumask_any_and(&p->cpus_allowed, cpu_online_mask);
|
|
|
|
if (dest_cpu < nr_cpu_ids)
|
|
|
|
goto move;
|
|
|
|
|
|
|
|
/* No more Mr. Nice Guy. */
|
|
|
|
if (dest_cpu >= nr_cpu_ids) {
|
|
|
|
cpuset_cpus_allowed_locked(p, &p->cpus_allowed);
|
|
|
|
dest_cpu = cpumask_any_and(cpu_online_mask, &p->cpus_allowed);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:11 +00:00
|
|
|
/*
|
|
|
|
* Don't tell them about moving exiting tasks or
|
|
|
|
* kernel threads (both mm NULL), since they never
|
|
|
|
* leave kernel.
|
|
|
|
*/
|
|
|
|
if (p->mm && printk_ratelimit()) {
|
|
|
|
printk(KERN_INFO "process %d (%s) no "
|
|
|
|
"longer affine to cpu%d\n",
|
|
|
|
task_pid_nr(p), p->comm, dead_cpu);
|
2007-10-15 15:00:14 +00:00
|
|
|
}
|
2008-11-24 16:05:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
move:
|
|
|
|
/* It can have affinity changed while we were choosing. */
|
|
|
|
if (unlikely(!__migrate_task_irq(p, dead_cpu, dest_cpu)))
|
|
|
|
goto again;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* While a dead CPU has no uninterruptible tasks queued at this point,
|
|
|
|
* it might still have a nonzero ->nr_uninterruptible counter, because
|
|
|
|
* for performance reasons the counter is not stricly tracking tasks to
|
|
|
|
* their home CPUs. So we just add the counter to another CPU's counter,
|
|
|
|
* to keep the global sum constant after CPU-down:
|
|
|
|
*/
|
2006-07-03 07:25:42 +00:00
|
|
|
static void migrate_nr_uninterruptible(struct rq *rq_src)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-11-24 16:05:03 +00:00
|
|
|
struct rq *rq_dest = cpu_rq(cpumask_any(cpu_online_mask));
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
local_irq_save(flags);
|
|
|
|
double_rq_lock(rq_src, rq_dest);
|
|
|
|
rq_dest->nr_uninterruptible += rq_src->nr_uninterruptible;
|
|
|
|
rq_src->nr_uninterruptible = 0;
|
|
|
|
double_rq_unlock(rq_src, rq_dest);
|
|
|
|
local_irq_restore(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run through task list and migrate tasks from the dead cpu. */
|
|
|
|
static void migrate_live_tasks(int src_cpu)
|
|
|
|
{
|
2006-07-03 07:25:40 +00:00
|
|
|
struct task_struct *p, *t;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:56 +00:00
|
|
|
read_lock(&tasklist_lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
do_each_thread(t, p) {
|
|
|
|
if (p == current)
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
if (task_cpu(p) == src_cpu)
|
|
|
|
move_task_off_dead_cpu(src_cpu, p);
|
|
|
|
} while_each_thread(t, p);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:56 +00:00
|
|
|
read_unlock(&tasklist_lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* Schedules idle task to be the next runnable task on current CPU.
|
2007-11-15 19:57:40 +00:00
|
|
|
* It does so by boosting its priority to highest possible.
|
|
|
|
* Used by CPU offline code.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
void sched_idle_next(void)
|
|
|
|
{
|
2006-07-03 07:25:40 +00:00
|
|
|
int this_cpu = smp_processor_id();
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(this_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
struct task_struct *p = rq->idle;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
/* cpu has to be offline */
|
2006-07-03 07:25:40 +00:00
|
|
|
BUG_ON(cpu_online(this_cpu));
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
/*
|
|
|
|
* Strictly not necessary since rest of the CPUs are stopped by now
|
|
|
|
* and interrupts disabled on the current cpu.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
__setscheduler(rq, p, SCHED_FIFO, MAX_RT_PRIO-1);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2007-11-15 19:57:40 +00:00
|
|
|
update_rq_clock(rq);
|
|
|
|
activate_task(rq, p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
/*
|
|
|
|
* Ensures that the idle task is using init_mm right before its cpu goes
|
2005-04-16 22:20:36 +00:00
|
|
|
* offline.
|
|
|
|
*/
|
|
|
|
void idle_task_exit(void)
|
|
|
|
{
|
|
|
|
struct mm_struct *mm = current->active_mm;
|
|
|
|
|
|
|
|
BUG_ON(cpu_online(smp_processor_id()));
|
|
|
|
|
|
|
|
if (mm != &init_mm)
|
|
|
|
switch_mm(mm, &init_mm, current);
|
|
|
|
mmdrop(mm);
|
|
|
|
}
|
|
|
|
|
2006-12-10 10:20:11 +00:00
|
|
|
/* called under rq->lock with disabled interrupts */
|
2006-07-03 07:25:41 +00:00
|
|
|
static void migrate_dead(unsigned int dead_cpu, struct task_struct *p)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(dead_cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Must be exiting, otherwise would be on tasklist. */
|
2007-10-19 06:40:38 +00:00
|
|
|
BUG_ON(!p->exit_state);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Cannot have done final schedule yet: would have vanished. */
|
2006-09-29 09:01:11 +00:00
|
|
|
BUG_ON(p->state == TASK_DEAD);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
get_task_struct(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Drop lock around migration; if someone else moves it,
|
2007-12-05 14:46:09 +00:00
|
|
|
* that's OK. No task can be added to this CPU, so iteration is
|
2005-04-16 22:20:36 +00:00
|
|
|
* fine.
|
|
|
|
*/
|
2007-10-17 06:30:56 +00:00
|
|
|
spin_unlock_irq(&rq->lock);
|
2006-07-03 07:25:40 +00:00
|
|
|
move_task_off_dead_cpu(dead_cpu, p);
|
2007-10-17 06:30:56 +00:00
|
|
|
spin_lock_irq(&rq->lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
put_task_struct(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* release_task() removes task from tasklist, so we won't find dead tasks. */
|
|
|
|
static void migrate_dead_tasks(unsigned int dead_cpu)
|
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(dead_cpu);
|
2007-07-09 16:51:59 +00:00
|
|
|
struct task_struct *next;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
for ( ; ; ) {
|
|
|
|
if (!rq->nr_running)
|
|
|
|
break;
|
2007-08-09 09:16:47 +00:00
|
|
|
update_rq_clock(rq);
|
2009-03-02 05:55:26 +00:00
|
|
|
next = pick_next_task(rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
if (!next)
|
|
|
|
break;
|
2008-06-28 22:16:56 +00:00
|
|
|
next->sched_class->put_prev_task(rq, next);
|
2007-07-09 16:51:59 +00:00
|
|
|
migrate_dead(dead_cpu, next);
|
2007-07-26 11:40:43 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
2009-04-11 08:43:41 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* remove the tasks which were accounted by rq from calc_load_tasks.
|
|
|
|
*/
|
|
|
|
static void calc_global_load_remove(struct rq *rq)
|
|
|
|
{
|
|
|
|
atomic_long_sub(rq->calc_load_active, &calc_load_tasks);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif /* CONFIG_HOTPLUG_CPU */
|
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
#if defined(CONFIG_SCHED_DEBUG) && defined(CONFIG_SYSCTL)
|
|
|
|
|
|
|
|
static struct ctl_table sd_ctl_dir[] = {
|
2007-08-09 09:16:46 +00:00
|
|
|
{
|
|
|
|
.procname = "sched_domain",
|
2007-08-23 13:18:02 +00:00
|
|
|
.mode = 0555,
|
2007-08-09 09:16:46 +00:00
|
|
|
},
|
2007-10-29 20:18:11 +00:00
|
|
|
{0, },
|
2007-07-26 11:40:43 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct ctl_table sd_ctl_root[] = {
|
2007-08-09 09:16:46 +00:00
|
|
|
{
|
2007-08-23 13:18:02 +00:00
|
|
|
.ctl_name = CTL_KERN,
|
2007-08-09 09:16:46 +00:00
|
|
|
.procname = "kernel",
|
2007-08-23 13:18:02 +00:00
|
|
|
.mode = 0555,
|
2007-08-09 09:16:46 +00:00
|
|
|
.child = sd_ctl_dir,
|
|
|
|
},
|
2007-10-29 20:18:11 +00:00
|
|
|
{0, },
|
2007-07-26 11:40:43 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct ctl_table *sd_alloc_ctl_entry(int n)
|
|
|
|
{
|
|
|
|
struct ctl_table *entry =
|
2007-10-15 15:00:19 +00:00
|
|
|
kcalloc(n, sizeof(struct ctl_table), GFP_KERNEL);
|
2007-07-26 11:40:43 +00:00
|
|
|
|
|
|
|
return entry;
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:19 +00:00
|
|
|
static void sd_free_ctl_entry(struct ctl_table **tablep)
|
|
|
|
{
|
2007-10-17 14:55:11 +00:00
|
|
|
struct ctl_table *entry;
|
2007-10-15 15:00:19 +00:00
|
|
|
|
2007-10-17 14:55:11 +00:00
|
|
|
/*
|
|
|
|
* In the intermediate directories, both the child directory and
|
|
|
|
* procname are dynamically allocated and could fail but the mode
|
2007-12-05 14:46:09 +00:00
|
|
|
* will always be set. In the lowest directory the names are
|
2007-10-17 14:55:11 +00:00
|
|
|
* static strings and all have proc handlers.
|
|
|
|
*/
|
|
|
|
for (entry = *tablep; entry->mode; entry++) {
|
2007-10-15 15:00:19 +00:00
|
|
|
if (entry->child)
|
|
|
|
sd_free_ctl_entry(&entry->child);
|
2007-10-17 14:55:11 +00:00
|
|
|
if (entry->proc_handler == NULL)
|
|
|
|
kfree(entry->procname);
|
|
|
|
}
|
2007-10-15 15:00:19 +00:00
|
|
|
|
|
|
|
kfree(*tablep);
|
|
|
|
*tablep = NULL;
|
|
|
|
}
|
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
static void
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(struct ctl_table *entry,
|
2007-07-26 11:40:43 +00:00
|
|
|
const char *procname, void *data, int maxlen,
|
|
|
|
mode_t mode, proc_handler *proc_handler)
|
|
|
|
{
|
|
|
|
entry->procname = procname;
|
|
|
|
entry->data = data;
|
|
|
|
entry->maxlen = maxlen;
|
|
|
|
entry->mode = mode;
|
|
|
|
entry->proc_handler = proc_handler;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct ctl_table *
|
|
|
|
sd_alloc_ctl_domain_table(struct sched_domain *sd)
|
|
|
|
{
|
2008-10-09 09:35:51 +00:00
|
|
|
struct ctl_table *table = sd_alloc_ctl_entry(13);
|
2007-07-26 11:40:43 +00:00
|
|
|
|
2007-10-15 15:00:19 +00:00
|
|
|
if (table == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[0], "min_interval", &sd->min_interval,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(long), 0644, proc_doulongvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[1], "max_interval", &sd->max_interval,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(long), 0644, proc_doulongvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[2], "busy_idx", &sd->busy_idx,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[3], "idle_idx", &sd->idle_idx,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[4], "newidle_idx", &sd->newidle_idx,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[5], "wake_idx", &sd->wake_idx,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[6], "forkexec_idx", &sd->forkexec_idx,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[7], "busy_factor", &sd->busy_factor,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-08-09 09:16:46 +00:00
|
|
|
set_table_entry(&table[8], "imbalance_pct", &sd->imbalance_pct,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-10-15 15:00:14 +00:00
|
|
|
set_table_entry(&table[9], "cache_nice_tries",
|
2007-07-26 11:40:43 +00:00
|
|
|
&sd->cache_nice_tries,
|
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2007-10-15 15:00:14 +00:00
|
|
|
set_table_entry(&table[10], "flags", &sd->flags,
|
2007-07-26 11:40:43 +00:00
|
|
|
sizeof(int), 0644, proc_dointvec_minmax);
|
2008-10-09 09:35:51 +00:00
|
|
|
set_table_entry(&table[11], "name", sd->name,
|
|
|
|
CORENAME_MAX_SIZE, 0444, proc_dostring);
|
|
|
|
/* &table[12] is terminator */
|
2007-07-26 11:40:43 +00:00
|
|
|
|
|
|
|
return table;
|
|
|
|
}
|
|
|
|
|
2007-11-28 14:52:56 +00:00
|
|
|
static ctl_table *sd_alloc_ctl_cpu_table(int cpu)
|
2007-07-26 11:40:43 +00:00
|
|
|
{
|
|
|
|
struct ctl_table *entry, *table;
|
|
|
|
struct sched_domain *sd;
|
|
|
|
int domain_num = 0, i;
|
|
|
|
char buf[32];
|
|
|
|
|
|
|
|
for_each_domain(cpu, sd)
|
|
|
|
domain_num++;
|
|
|
|
entry = table = sd_alloc_ctl_entry(domain_num + 1);
|
2007-10-15 15:00:19 +00:00
|
|
|
if (table == NULL)
|
|
|
|
return NULL;
|
2007-07-26 11:40:43 +00:00
|
|
|
|
|
|
|
i = 0;
|
|
|
|
for_each_domain(cpu, sd) {
|
|
|
|
snprintf(buf, 32, "domain%d", i);
|
|
|
|
entry->procname = kstrdup(buf, GFP_KERNEL);
|
2007-08-23 13:18:02 +00:00
|
|
|
entry->mode = 0555;
|
2007-07-26 11:40:43 +00:00
|
|
|
entry->child = sd_alloc_ctl_domain_table(sd);
|
|
|
|
entry++;
|
|
|
|
i++;
|
|
|
|
}
|
|
|
|
return table;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct ctl_table_header *sd_sysctl_header;
|
2007-10-15 15:00:19 +00:00
|
|
|
static void register_sched_domain_sysctl(void)
|
2007-07-26 11:40:43 +00:00
|
|
|
{
|
|
|
|
int i, cpu_num = num_online_cpus();
|
|
|
|
struct ctl_table *entry = sd_alloc_ctl_entry(cpu_num + 1);
|
|
|
|
char buf[32];
|
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
WARN_ON(sd_ctl_dir[0].child);
|
|
|
|
sd_ctl_dir[0].child = entry;
|
|
|
|
|
2007-10-15 15:00:19 +00:00
|
|
|
if (entry == NULL)
|
|
|
|
return;
|
|
|
|
|
2007-10-15 15:00:19 +00:00
|
|
|
for_each_online_cpu(i) {
|
2007-07-26 11:40:43 +00:00
|
|
|
snprintf(buf, 32, "cpu%d", i);
|
|
|
|
entry->procname = kstrdup(buf, GFP_KERNEL);
|
2007-08-23 13:18:02 +00:00
|
|
|
entry->mode = 0555;
|
2007-07-26 11:40:43 +00:00
|
|
|
entry->child = sd_alloc_ctl_cpu_table(i);
|
2007-10-15 15:00:19 +00:00
|
|
|
entry++;
|
2007-07-26 11:40:43 +00:00
|
|
|
}
|
2007-10-24 16:23:48 +00:00
|
|
|
|
|
|
|
WARN_ON(sd_sysctl_header);
|
2007-07-26 11:40:43 +00:00
|
|
|
sd_sysctl_header = register_sysctl_table(sd_ctl_root);
|
|
|
|
}
|
2007-10-15 15:00:19 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
/* may be called multiple times per register */
|
2007-10-15 15:00:19 +00:00
|
|
|
static void unregister_sched_domain_sysctl(void)
|
|
|
|
{
|
2007-10-24 16:23:48 +00:00
|
|
|
if (sd_sysctl_header)
|
|
|
|
unregister_sysctl_table(sd_sysctl_header);
|
2007-10-15 15:00:19 +00:00
|
|
|
sd_sysctl_header = NULL;
|
2007-10-24 16:23:48 +00:00
|
|
|
if (sd_ctl_dir[0].child)
|
|
|
|
sd_free_ctl_entry(&sd_ctl_dir[0].child);
|
2007-10-15 15:00:19 +00:00
|
|
|
}
|
2007-07-26 11:40:43 +00:00
|
|
|
#else
|
2007-10-15 15:00:19 +00:00
|
|
|
static void register_sched_domain_sysctl(void)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
static void unregister_sched_domain_sysctl(void)
|
2007-07-26 11:40:43 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-06-04 19:04:05 +00:00
|
|
|
static void set_rq_online(struct rq *rq)
|
|
|
|
{
|
|
|
|
if (!rq->online) {
|
|
|
|
const struct sched_class *class;
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_set_cpu(rq->cpu, rq->rd->online);
|
2008-06-04 19:04:05 +00:00
|
|
|
rq->online = 1;
|
|
|
|
|
|
|
|
for_each_class(class) {
|
|
|
|
if (class->rq_online)
|
|
|
|
class->rq_online(rq);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void set_rq_offline(struct rq *rq)
|
|
|
|
{
|
|
|
|
if (rq->online) {
|
|
|
|
const struct sched_class *class;
|
|
|
|
|
|
|
|
for_each_class(class) {
|
|
|
|
if (class->rq_offline)
|
|
|
|
class->rq_offline(rq);
|
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_clear_cpu(rq->cpu, rq->rd->online);
|
2008-06-04 19:04:05 +00:00
|
|
|
rq->online = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* migration_call - callback that gets triggered when a CPU is added.
|
|
|
|
* Here we can start up the necessary migration thread for the new CPU.
|
|
|
|
*/
|
2006-07-03 07:25:40 +00:00
|
|
|
static int __cpuinit
|
|
|
|
migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct task_struct *p;
|
2006-07-03 07:25:40 +00:00
|
|
|
int cpu = (long)hcpu;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
switch (action) {
|
2007-05-09 09:34:04 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
case CPU_UP_PREPARE:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_UP_PREPARE_FROZEN:
|
2007-07-09 16:51:59 +00:00
|
|
|
p = kthread_create(migration_thread, hcpu, "migration/%d", cpu);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (IS_ERR(p))
|
|
|
|
return NOTIFY_BAD;
|
|
|
|
kthread_bind(p, cpu);
|
|
|
|
/* Must be high prio: stop_machine expects to yield to it. */
|
|
|
|
rq = task_rq_lock(p, &flags);
|
2007-07-09 16:51:59 +00:00
|
|
|
__setscheduler(rq, p, SCHED_FIFO, MAX_RT_PRIO-1);
|
2005-04-16 22:20:36 +00:00
|
|
|
task_rq_unlock(rq, &flags);
|
2009-06-17 23:27:45 +00:00
|
|
|
get_task_struct(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
cpu_rq(cpu)->migration_thread = p;
|
|
|
|
break;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
case CPU_ONLINE:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_ONLINE_FROZEN:
|
2007-10-19 21:10:43 +00:00
|
|
|
/* Strictly unnecessary, as first user will wake it. */
|
2005-04-16 22:20:36 +00:00
|
|
|
wake_up_process(cpu_rq(cpu)->migration_thread);
|
2008-03-10 20:52:41 +00:00
|
|
|
|
|
|
|
/* Update our root-domain */
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
2009-04-11 08:43:41 +00:00
|
|
|
rq->calc_load_update = calc_load_update;
|
|
|
|
rq->calc_load_active = 0;
|
2008-03-10 20:52:41 +00:00
|
|
|
if (rq->rd) {
|
2008-11-24 16:05:05 +00:00
|
|
|
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
|
2008-06-04 19:04:05 +00:00
|
|
|
|
|
|
|
set_rq_online(rq);
|
2008-03-10 20:52:41 +00:00
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
case CPU_UP_CANCELED:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_UP_CANCELED_FROZEN:
|
2006-06-25 12:49:10 +00:00
|
|
|
if (!cpu_rq(cpu)->migration_thread)
|
|
|
|
break;
|
2007-12-05 14:46:09 +00:00
|
|
|
/* Unbind it from offline cpu so it can run. Fall thru. */
|
2005-11-07 08:58:38 +00:00
|
|
|
kthread_bind(cpu_rq(cpu)->migration_thread,
|
2008-11-24 16:05:03 +00:00
|
|
|
cpumask_any(cpu_online_mask));
|
2005-04-16 22:20:36 +00:00
|
|
|
kthread_stop(cpu_rq(cpu)->migration_thread);
|
2009-06-17 23:27:45 +00:00
|
|
|
put_task_struct(cpu_rq(cpu)->migration_thread);
|
2005-04-16 22:20:36 +00:00
|
|
|
cpu_rq(cpu)->migration_thread = NULL;
|
|
|
|
break;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
case CPU_DEAD:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_DEAD_FROZEN:
|
2007-10-19 06:40:46 +00:00
|
|
|
cpuset_lock(); /* around calls to cpuset_cpus_allowed_lock() */
|
2005-04-16 22:20:36 +00:00
|
|
|
migrate_live_tasks(cpu);
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
kthread_stop(rq->migration_thread);
|
2009-06-17 23:27:45 +00:00
|
|
|
put_task_struct(rq->migration_thread);
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->migration_thread = NULL;
|
|
|
|
/* Idle task back to normal (off runqueue, low prio) */
|
2007-10-17 06:30:56 +00:00
|
|
|
spin_lock_irq(&rq->lock);
|
2007-08-09 09:16:47 +00:00
|
|
|
update_rq_clock(rq);
|
2007-08-09 09:16:49 +00:00
|
|
|
deactivate_task(rq, rq->idle, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->idle->static_prio = MAX_PRIO;
|
2007-07-09 16:51:59 +00:00
|
|
|
__setscheduler(rq, rq->idle, SCHED_NORMAL, 0);
|
|
|
|
rq->idle->sched_class = &idle_sched_class;
|
2005-04-16 22:20:36 +00:00
|
|
|
migrate_dead_tasks(cpu);
|
2007-10-17 06:30:56 +00:00
|
|
|
spin_unlock_irq(&rq->lock);
|
2007-10-19 06:40:46 +00:00
|
|
|
cpuset_unlock();
|
2005-04-16 22:20:36 +00:00
|
|
|
migrate_nr_uninterruptible(rq);
|
|
|
|
BUG_ON(rq->nr_running != 0);
|
2009-04-11 08:43:41 +00:00
|
|
|
calc_global_load_remove(rq);
|
2007-12-05 14:46:09 +00:00
|
|
|
/*
|
|
|
|
* No need to migrate the tasks: it was best-effort if
|
|
|
|
* they didn't take sched_hotcpu_mutex. Just wake up
|
|
|
|
* the requestors.
|
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
spin_lock_irq(&rq->lock);
|
|
|
|
while (!list_empty(&rq->migration_queue)) {
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req *req;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
req = list_entry(rq->migration_queue.next,
|
2006-07-03 07:25:42 +00:00
|
|
|
struct migration_req, list);
|
2005-04-16 22:20:36 +00:00
|
|
|
list_del_init(&req->list);
|
2008-12-09 14:47:00 +00:00
|
|
|
spin_unlock_irq(&rq->lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
complete(&req->done);
|
2008-12-09 14:47:00 +00:00
|
|
|
spin_lock_irq(&rq->lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
spin_unlock_irq(&rq->lock);
|
|
|
|
break;
|
2008-01-25 20:08:18 +00:00
|
|
|
|
2008-03-10 21:59:11 +00:00
|
|
|
case CPU_DYING:
|
|
|
|
case CPU_DYING_FROZEN:
|
2008-01-25 20:08:18 +00:00
|
|
|
/* Update our root-domain */
|
|
|
|
rq = cpu_rq(cpu);
|
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
|
|
|
if (rq->rd) {
|
2008-11-24 16:05:05 +00:00
|
|
|
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
|
2008-06-04 19:04:05 +00:00
|
|
|
set_rq_offline(rq);
|
2008-01-25 20:08:18 +00:00
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
|
|
}
|
|
|
|
|
2009-06-02 11:05:16 +00:00
|
|
|
/*
|
|
|
|
* Register at high priority so that task migration (migrate_all_tasks)
|
|
|
|
* happens before everything else. This has to be lower priority than
|
|
|
|
* the notifier in the perf_counter subsystem, though.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2006-06-27 09:54:10 +00:00
|
|
|
static struct notifier_block __cpuinitdata migration_notifier = {
|
2005-04-16 22:20:36 +00:00
|
|
|
.notifier_call = migration_call,
|
|
|
|
.priority = 10
|
|
|
|
};
|
|
|
|
|
2008-07-26 02:45:11 +00:00
|
|
|
static int __init migration_init(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
void *cpu = (void *)(long)smp_processor_id();
|
2006-09-29 09:00:22 +00:00
|
|
|
int err;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
|
|
|
/* Start one for the boot CPU: */
|
2006-09-29 09:00:22 +00:00
|
|
|
err = migration_call(&migration_notifier, CPU_UP_PREPARE, cpu);
|
|
|
|
BUG_ON(err == NOTIFY_BAD);
|
2005-04-16 22:20:36 +00:00
|
|
|
migration_call(&migration_notifier, CPU_ONLINE, cpu);
|
|
|
|
register_cpu_notifier(&migration_notifier);
|
2008-07-26 02:45:11 +00:00
|
|
|
|
|
|
|
return err;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-07-26 02:45:11 +00:00
|
|
|
early_initcall(migration_init);
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
2007-05-06 21:48:58 +00:00
|
|
|
|
2007-10-15 15:00:13 +00:00
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
2007-10-24 16:23:48 +00:00
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level,
|
2008-11-24 16:05:14 +00:00
|
|
|
struct cpumask *groupmask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-24 16:23:48 +00:00
|
|
|
struct sched_group *group = sd->groups;
|
2008-04-05 01:11:04 +00:00
|
|
|
char str[256];
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-12-13 11:25:51 +00:00
|
|
|
cpulist_scnprintf(str, sizeof(str), sched_domain_span(sd));
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_clear(groupmask);
|
2007-10-24 16:23:48 +00:00
|
|
|
|
|
|
|
printk(KERN_DEBUG "%*s domain %d: ", level, "", level);
|
|
|
|
|
|
|
|
if (!(sd->flags & SD_LOAD_BALANCE)) {
|
|
|
|
printk("does not load-balance\n");
|
|
|
|
if (sd->parent)
|
|
|
|
printk(KERN_ERR "ERROR: !SD_LOAD_BALANCE domain"
|
|
|
|
" has parent");
|
|
|
|
return -1;
|
2005-06-25 21:57:24 +00:00
|
|
|
}
|
|
|
|
|
2008-11-04 08:15:37 +00:00
|
|
|
printk(KERN_CONT "span %s level %s\n", str, sd->name);
|
2007-10-24 16:23:48 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_test_cpu(cpu, sched_domain_span(sd))) {
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_ERR "ERROR: domain->span does not contain "
|
|
|
|
"CPU%d\n", cpu);
|
|
|
|
}
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_test_cpu(cpu, sched_group_cpus(group))) {
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_ERR "ERROR: domain->groups does not contain"
|
|
|
|
" CPU%d\n", cpu);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_DEBUG "%*s groups:", level + 1, "");
|
2005-04-16 22:20:36 +00:00
|
|
|
do {
|
2007-10-24 16:23:48 +00:00
|
|
|
if (!group) {
|
|
|
|
printk("\n");
|
|
|
|
printk(KERN_ERR "ERROR: group is NULL\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
if (!group->__cpu_power) {
|
|
|
|
printk(KERN_CONT "\n");
|
|
|
|
printk(KERN_ERR "ERROR: domain->cpu_power not "
|
|
|
|
"set\n");
|
|
|
|
break;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_weight(sched_group_cpus(group))) {
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_CONT "\n");
|
|
|
|
printk(KERN_ERR "ERROR: empty group\n");
|
|
|
|
break;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (cpumask_intersects(groupmask, sched_group_cpus(group))) {
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_CONT "\n");
|
|
|
|
printk(KERN_ERR "ERROR: repeated CPUs\n");
|
|
|
|
break;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_or(groupmask, groupmask, sched_group_cpus(group));
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-12-13 11:25:51 +00:00
|
|
|
cpulist_scnprintf(str, sizeof(str), sched_group_cpus(group));
|
2009-04-14 03:39:36 +00:00
|
|
|
|
|
|
|
printk(KERN_CONT " %s", str);
|
|
|
|
if (group->__cpu_power != SCHED_LOAD_SCALE) {
|
|
|
|
printk(KERN_CONT " (__cpu_power = %d)",
|
|
|
|
group->__cpu_power);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
group = group->next;
|
|
|
|
} while (group != sd->groups);
|
|
|
|
printk(KERN_CONT "\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_equal(sched_domain_span(sd), groupmask))
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_ERR "ERROR: groups don't span domain->span\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (sd->parent &&
|
|
|
|
!cpumask_subset(groupmask, sched_domain_span(sd->parent)))
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_ERR "ERROR: parent span is not a superset "
|
|
|
|
"of domain->span\n");
|
|
|
|
return 0;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
static void sched_domain_debug(struct sched_domain *sd, int cpu)
|
|
|
|
{
|
2008-11-24 16:05:12 +00:00
|
|
|
cpumask_var_t groupmask;
|
2007-10-24 16:23:48 +00:00
|
|
|
int level = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
if (!sd) {
|
|
|
|
printk(KERN_DEBUG "CPU%d attaching NULL sched-domain.\n", cpu);
|
|
|
|
return;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
printk(KERN_DEBUG "CPU%d attaching sched-domain:\n", cpu);
|
|
|
|
|
2008-11-24 16:05:12 +00:00
|
|
|
if (!alloc_cpumask_var(&groupmask, GFP_KERNEL)) {
|
2008-04-05 01:11:11 +00:00
|
|
|
printk(KERN_DEBUG "Cannot load-balance (out of memory)\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
for (;;) {
|
2008-04-05 01:11:11 +00:00
|
|
|
if (sched_domain_debug_one(sd, cpu, level, groupmask))
|
2007-10-24 16:23:48 +00:00
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
level++;
|
|
|
|
sd = sd->parent;
|
2006-12-10 10:20:38 +00:00
|
|
|
if (!sd)
|
2007-10-24 16:23:48 +00:00
|
|
|
break;
|
|
|
|
}
|
2008-11-24 16:05:12 +00:00
|
|
|
free_cpumask_var(groupmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_SCHED_DEBUG */
|
2006-07-03 07:25:40 +00:00
|
|
|
# define sched_domain_debug(sd, cpu) do { } while (0)
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SCHED_DEBUG */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-06-25 21:57:33 +00:00
|
|
|
static int sd_degenerate(struct sched_domain *sd)
|
2005-06-25 21:57:25 +00:00
|
|
|
{
|
2008-11-24 16:05:04 +00:00
|
|
|
if (cpumask_weight(sched_domain_span(sd)) == 1)
|
2005-06-25 21:57:25 +00:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
/* Following flags need at least 2 groups */
|
|
|
|
if (sd->flags & (SD_LOAD_BALANCE |
|
|
|
|
SD_BALANCE_NEWIDLE |
|
|
|
|
SD_BALANCE_FORK |
|
2006-10-03 08:14:09 +00:00
|
|
|
SD_BALANCE_EXEC |
|
|
|
|
SD_SHARE_CPUPOWER |
|
|
|
|
SD_SHARE_PKG_RESOURCES)) {
|
2005-06-25 21:57:25 +00:00
|
|
|
if (sd->groups != sd->groups->next)
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Following flags don't use groups */
|
|
|
|
if (sd->flags & (SD_WAKE_IDLE |
|
|
|
|
SD_WAKE_AFFINE |
|
|
|
|
SD_WAKE_BALANCE))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
static int
|
|
|
|
sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent)
|
2005-06-25 21:57:25 +00:00
|
|
|
{
|
|
|
|
unsigned long cflags = sd->flags, pflags = parent->flags;
|
|
|
|
|
|
|
|
if (sd_degenerate(parent))
|
|
|
|
return 1;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (!cpumask_equal(sched_domain_span(sd), sched_domain_span(parent)))
|
2005-06-25 21:57:25 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Does parent contain flags not in child? */
|
|
|
|
/* WAKE_BALANCE is a subset of WAKE_AFFINE */
|
|
|
|
if (cflags & SD_WAKE_AFFINE)
|
|
|
|
pflags &= ~SD_WAKE_BALANCE;
|
|
|
|
/* Flags needing groups don't count if only 1 group in parent */
|
|
|
|
if (parent->groups == parent->groups->next) {
|
|
|
|
pflags &= ~(SD_LOAD_BALANCE |
|
|
|
|
SD_BALANCE_NEWIDLE |
|
|
|
|
SD_BALANCE_FORK |
|
2006-10-03 08:14:09 +00:00
|
|
|
SD_BALANCE_EXEC |
|
|
|
|
SD_SHARE_CPUPOWER |
|
|
|
|
SD_SHARE_PKG_RESOURCES);
|
2008-12-08 02:47:37 +00:00
|
|
|
if (nr_node_ids == 1)
|
|
|
|
pflags &= ~SD_SERIALIZE;
|
2005-06-25 21:57:25 +00:00
|
|
|
}
|
|
|
|
if (~cflags & pflags)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
static void free_rootdomain(struct root_domain *rd)
|
|
|
|
{
|
2008-11-24 16:05:13 +00:00
|
|
|
cpupri_cleanup(&rd->cpupri);
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
free_cpumask_var(rd->rto_mask);
|
|
|
|
free_cpumask_var(rd->online);
|
|
|
|
free_cpumask_var(rd->span);
|
|
|
|
kfree(rd);
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
static void rq_attach_root(struct rq *rq, struct root_domain *rd)
|
|
|
|
{
|
2009-02-12 10:35:40 +00:00
|
|
|
struct root_domain *old_rd = NULL;
|
2008-01-25 20:08:18 +00:00
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
|
|
|
|
|
|
|
if (rq->rd) {
|
2009-02-12 10:35:40 +00:00
|
|
|
old_rd = rq->rd;
|
2008-01-25 20:08:18 +00:00
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
if (cpumask_test_cpu(rq->cpu, old_rd->online))
|
2008-06-04 19:04:05 +00:00
|
|
|
set_rq_offline(rq);
|
2008-01-25 20:08:18 +00:00
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_clear_cpu(rq->cpu, old_rd->span);
|
2008-01-25 20:08:26 +00:00
|
|
|
|
2009-02-12 10:35:40 +00:00
|
|
|
/*
|
|
|
|
* If we dont want to free the old_rt yet then
|
|
|
|
* set old_rd to NULL to skip the freeing later
|
|
|
|
* in this function:
|
|
|
|
*/
|
|
|
|
if (!atomic_dec_and_test(&old_rd->refcount))
|
|
|
|
old_rd = NULL;
|
2008-01-25 20:08:18 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
atomic_inc(&rd->refcount);
|
|
|
|
rq->rd = rd;
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
cpumask_set_cpu(rq->cpu, rd->span);
|
|
|
|
if (cpumask_test_cpu(rq->cpu, cpu_online_mask))
|
2008-06-04 19:04:05 +00:00
|
|
|
set_rq_online(rq);
|
2008-01-25 20:08:18 +00:00
|
|
|
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
2009-02-12 10:35:40 +00:00
|
|
|
|
|
|
|
if (old_rd)
|
|
|
|
free_rootdomain(old_rd);
|
2008-01-25 20:08:18 +00:00
|
|
|
}
|
|
|
|
|
2009-06-15 05:34:19 +00:00
|
|
|
static int init_rootdomain(struct root_domain *rd, bool bootmem)
|
2008-01-25 20:08:18 +00:00
|
|
|
{
|
2009-06-10 20:42:36 +00:00
|
|
|
gfp_t gfp = GFP_KERNEL;
|
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
memset(rd, 0, sizeof(*rd));
|
|
|
|
|
2009-06-10 20:42:36 +00:00
|
|
|
if (bootmem)
|
|
|
|
gfp = GFP_NOWAIT;
|
2008-11-24 16:05:05 +00:00
|
|
|
|
2009-06-10 20:42:36 +00:00
|
|
|
if (!alloc_cpumask_var(&rd->span, gfp))
|
2009-01-06 09:39:06 +00:00
|
|
|
goto out;
|
2009-06-10 20:42:36 +00:00
|
|
|
if (!alloc_cpumask_var(&rd->online, gfp))
|
2008-11-24 16:05:05 +00:00
|
|
|
goto free_span;
|
2009-06-10 20:42:36 +00:00
|
|
|
if (!alloc_cpumask_var(&rd->rto_mask, gfp))
|
2008-11-24 16:05:05 +00:00
|
|
|
goto free_online;
|
2008-05-12 19:21:01 +00:00
|
|
|
|
2009-06-11 05:41:22 +00:00
|
|
|
if (cpupri_init(&rd->cpupri, bootmem) != 0)
|
2008-11-24 16:05:13 +00:00
|
|
|
goto free_rto_mask;
|
2008-11-24 16:05:05 +00:00
|
|
|
return 0;
|
2008-05-12 19:21:01 +00:00
|
|
|
|
2008-11-24 16:05:13 +00:00
|
|
|
free_rto_mask:
|
|
|
|
free_cpumask_var(rd->rto_mask);
|
2008-11-24 16:05:05 +00:00
|
|
|
free_online:
|
|
|
|
free_cpumask_var(rd->online);
|
|
|
|
free_span:
|
|
|
|
free_cpumask_var(rd->span);
|
2009-01-06 09:39:06 +00:00
|
|
|
out:
|
2008-11-24 16:05:05 +00:00
|
|
|
return -ENOMEM;
|
2008-01-25 20:08:18 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void init_defrootdomain(void)
|
|
|
|
{
|
2008-11-24 16:05:05 +00:00
|
|
|
init_rootdomain(&def_root_domain, true);
|
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
atomic_set(&def_root_domain.refcount, 1);
|
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:26 +00:00
|
|
|
static struct root_domain *alloc_rootdomain(void)
|
2008-01-25 20:08:18 +00:00
|
|
|
{
|
|
|
|
struct root_domain *rd;
|
|
|
|
|
|
|
|
rd = kmalloc(sizeof(*rd), GFP_KERNEL);
|
|
|
|
if (!rd)
|
|
|
|
return NULL;
|
|
|
|
|
2008-11-24 16:05:05 +00:00
|
|
|
if (init_rootdomain(rd, false) != 0) {
|
|
|
|
kfree(rd);
|
|
|
|
return NULL;
|
|
|
|
}
|
2008-01-25 20:08:18 +00:00
|
|
|
|
|
|
|
return rd;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2008-01-25 20:08:19 +00:00
|
|
|
* Attach the domain 'sd' to 'cpu' as its base domain. Callers must
|
2005-04-16 22:20:36 +00:00
|
|
|
* hold the hotplug lock.
|
|
|
|
*/
|
2008-01-25 20:08:19 +00:00
|
|
|
static void
|
|
|
|
cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2005-06-25 21:57:25 +00:00
|
|
|
struct sched_domain *tmp;
|
|
|
|
|
|
|
|
/* Remove the sched domains which do not contribute to scheduling. */
|
2008-11-06 01:45:16 +00:00
|
|
|
for (tmp = sd; tmp; ) {
|
2005-06-25 21:57:25 +00:00
|
|
|
struct sched_domain *parent = tmp->parent;
|
|
|
|
if (!parent)
|
|
|
|
break;
|
2008-11-06 01:45:16 +00:00
|
|
|
|
2006-10-03 08:14:08 +00:00
|
|
|
if (sd_parent_degenerate(tmp, parent)) {
|
2005-06-25 21:57:25 +00:00
|
|
|
tmp->parent = parent->parent;
|
2006-10-03 08:14:08 +00:00
|
|
|
if (parent->parent)
|
|
|
|
parent->parent->child = tmp;
|
2008-11-06 01:45:16 +00:00
|
|
|
} else
|
|
|
|
tmp = tmp->parent;
|
2005-06-25 21:57:25 +00:00
|
|
|
}
|
|
|
|
|
2006-10-03 08:14:08 +00:00
|
|
|
if (sd && sd_degenerate(sd)) {
|
2005-06-25 21:57:25 +00:00
|
|
|
sd = sd->parent;
|
2006-10-03 08:14:08 +00:00
|
|
|
if (sd)
|
|
|
|
sd->child = NULL;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
sched_domain_debug(sd, cpu);
|
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
rq_attach_root(rq, rd);
|
2005-06-25 21:57:27 +00:00
|
|
|
rcu_assign_pointer(rq->sd, sd);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* cpus with isolated domains */
|
2008-11-24 16:05:12 +00:00
|
|
|
static cpumask_var_t cpu_isolated_map;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Setup the mask of cpus configured for isolated domains */
|
|
|
|
static int __init isolated_cpu_setup(char *str)
|
|
|
|
{
|
2008-12-13 11:25:51 +00:00
|
|
|
cpulist_parse(str, cpu_isolated_map);
|
2005-04-16 22:20:36 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:13 +00:00
|
|
|
__setup("isolcpus=", isolated_cpu_setup);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2006-12-10 10:20:07 +00:00
|
|
|
* init_sched_build_groups takes the cpumask we wish to span, and a pointer
|
|
|
|
* to a function which identifies what group(along with sched group) a CPU
|
2008-11-24 16:05:14 +00:00
|
|
|
* belongs to. The return value of group_fn must be a >= 0 and < nr_cpu_ids
|
|
|
|
* (due to the fact that we keep track of groups covered with a struct cpumask).
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* init_sched_build_groups will build a circular linked list of the groups
|
|
|
|
* covered by the given span, and will set each group's ->cpumask correctly,
|
|
|
|
* and ->cpu_power to 0.
|
|
|
|
*/
|
2006-10-03 08:14:06 +00:00
|
|
|
static void
|
2008-11-24 16:05:14 +00:00
|
|
|
init_sched_build_groups(const struct cpumask *span,
|
|
|
|
const struct cpumask *cpu_map,
|
|
|
|
int (*group_fn)(int cpu, const struct cpumask *cpu_map,
|
2008-04-05 01:11:11 +00:00
|
|
|
struct sched_group **sg,
|
2008-11-24 16:05:14 +00:00
|
|
|
struct cpumask *tmpmask),
|
|
|
|
struct cpumask *covered, struct cpumask *tmpmask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sched_group *first = NULL, *last = NULL;
|
|
|
|
int i;
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_clear(covered);
|
2008-04-05 01:11:11 +00:00
|
|
|
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, span) {
|
2006-12-10 10:20:07 +00:00
|
|
|
struct sched_group *sg;
|
2008-04-05 01:11:11 +00:00
|
|
|
int group = group_fn(i, cpu_map, &sg, tmpmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
int j;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
if (cpumask_test_cpu(i, covered))
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_clear(sched_group_cpus(sg));
|
2007-05-08 07:32:57 +00:00
|
|
|
sg->__cpu_power = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(j, span) {
|
2008-04-05 01:11:11 +00:00
|
|
|
if (group_fn(j, cpu_map, NULL, tmpmask) != group)
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_set_cpu(j, covered);
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_set_cpu(j, sched_group_cpus(sg));
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
if (!first)
|
|
|
|
first = sg;
|
|
|
|
if (last)
|
|
|
|
last->next = sg;
|
|
|
|
last = sg;
|
|
|
|
}
|
|
|
|
last->next = first;
|
|
|
|
}
|
|
|
|
|
2005-09-06 22:18:14 +00:00
|
|
|
#define SD_NODES_PER_DOMAIN 16
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-09-06 22:18:14 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
[PATCH] scheduler cache-hot-autodetect
)
From: Ingo Molnar <mingo@elte.hu>
This is the latest version of the scheduler cache-hot-auto-tune patch.
The first problem was that detection time scaled with O(N^2), which is
unacceptable on larger SMP and NUMA systems. To solve this:
- I've added a 'domain distance' function, which is used to cache
measurement results. Each distance is only measured once. This means
that e.g. on NUMA distances of 0, 1 and 2 might be measured, on HT
distances 0 and 1, and on SMP distance 0 is measured. The code walks
the domain tree to determine the distance, so it automatically follows
whatever hierarchy an architecture sets up. This cuts down on the boot
time significantly and removes the O(N^2) limit. The only assumption
is that migration costs can be expressed as a function of domain
distance - this covers the overwhelming majority of existing systems,
and is a good guess even for more assymetric systems.
[ People hacking systems that have assymetries that break this
assumption (e.g. different CPU speeds) should experiment a bit with
the cpu_distance() function. Adding a ->migration_distance factor to
the domain structure would be one possible solution - but lets first
see the problem systems, if they exist at all. Lets not overdesign. ]
Another problem was that only a single cache-size was used for measuring
the cost of migration, and most architectures didnt set that variable
up. Furthermore, a single cache-size does not fit NUMA hierarchies with
L3 caches and does not fit HT setups, where different CPUs will often
have different 'effective cache sizes'. To solve this problem:
- Instead of relying on a single cache-size provided by the platform and
sticking to it, the code now auto-detects the 'effective migration
cost' between two measured CPUs, via iterating through a wide range of
cachesizes. The code searches for the maximum migration cost, which
occurs when the working set of the test-workload falls just below the
'effective cache size'. I.e. real-life optimized search is done for
the maximum migration cost, between two real CPUs.
This, amongst other things, has the positive effect hat if e.g. two
CPUs share a L2/L3 cache, a different (and accurate) migration cost
will be found than between two CPUs on the same system that dont share
any caches.
(The reliable measurement of migration costs is tricky - see the source
for details.)
Furthermore i've added various boot-time options to override/tune
migration behavior.
Firstly, there's a blanket override for autodetection:
migration_cost=1000,2000,3000
will override the depth 0/1/2 values with 1msec/2msec/3msec values.
Secondly, there's a global factor that can be used to increase (or
decrease) the autodetected values:
migration_factor=120
will increase the autodetected values by 20%. This option is useful to
tune things in a workload-dependent way - e.g. if a workload is
cache-insensitive then CPU utilization can be maximized by specifying
migration_factor=0.
I've tested the autodetection code quite extensively on x86, on 3
P3/Xeon/2MB, and the autodetected values look pretty good:
Dual Celeron (128K L2 cache):
---------------------
migration cost matrix (max_cache_size: 131072, cpu: 467 MHz):
---------------------
[00] [01]
[00]: - 1.7(1)
[01]: 1.7(1) -
---------------------
cacheflush times [2]: 0.0 (0) 1.7 (1784008)
---------------------
Here the slow memory subsystem dominates system performance, and even
though caches are small, the migration cost is 1.7 msecs.
Dual HT P4 (512K L2 cache):
---------------------
migration cost matrix (max_cache_size: 524288, cpu: 2379 MHz):
---------------------
[00] [01] [02] [03]
[00]: - 0.4(1) 0.0(0) 0.4(1)
[01]: 0.4(1) - 0.4(1) 0.0(0)
[02]: 0.0(0) 0.4(1) - 0.4(1)
[03]: 0.4(1) 0.0(0) 0.4(1) -
---------------------
cacheflush times [2]: 0.0 (33900) 0.4 (448514)
---------------------
Here it can be seen that there is no migration cost between two HT
siblings (CPU#0/2 and CPU#1/3 are separate physical CPUs). A fast memory
system makes inter-physical-CPU migration pretty cheap: 0.4 msecs.
8-way P3/Xeon [2MB L2 cache]:
---------------------
migration cost matrix (max_cache_size: 2097152, cpu: 700 MHz):
---------------------
[00] [01] [02] [03] [04] [05] [06] [07]
[00]: - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[01]: 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[02]: 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[03]: 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1) 19.2(1)
[04]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1) 19.2(1)
[05]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1) 19.2(1)
[06]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) - 19.2(1)
[07]: 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) 19.2(1) -
---------------------
cacheflush times [2]: 0.0 (0) 19.2 (19281756)
---------------------
This one has huge caches and a relatively slow memory subsystem - so the
migration cost is 19 msecs.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Ashok Raj <ashok.raj@intel.com>
Signed-off-by: Ken Chen <kenneth.w.chen@intel.com>
Cc: <wilder@us.ibm.com>
Signed-off-by: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-01-12 09:05:30 +00:00
|
|
|
|
2005-09-06 22:18:14 +00:00
|
|
|
/**
|
|
|
|
* find_next_best_node - find the next node to include in a sched_domain
|
|
|
|
* @node: node whose sched_domain we're building
|
|
|
|
* @used_nodes: nodes already in the sched_domain
|
|
|
|
*
|
2007-12-05 14:46:09 +00:00
|
|
|
* Find the next node to include in a given scheduling domain. Simply
|
2005-09-06 22:18:14 +00:00
|
|
|
* finds the closest node not already in the @used_nodes map.
|
|
|
|
*
|
|
|
|
* Should use nodemask_t.
|
|
|
|
*/
|
2008-04-05 01:11:10 +00:00
|
|
|
static int find_next_best_node(int node, nodemask_t *used_nodes)
|
2005-09-06 22:18:14 +00:00
|
|
|
{
|
|
|
|
int i, n, val, min_val, best_node = 0;
|
|
|
|
|
|
|
|
min_val = INT_MAX;
|
|
|
|
|
2008-05-12 19:21:12 +00:00
|
|
|
for (i = 0; i < nr_node_ids; i++) {
|
2005-09-06 22:18:14 +00:00
|
|
|
/* Start at @node */
|
2008-05-12 19:21:12 +00:00
|
|
|
n = (node + i) % nr_node_ids;
|
2005-09-06 22:18:14 +00:00
|
|
|
|
|
|
|
if (!nr_cpus_node(n))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Skip already used nodes */
|
2008-04-05 01:11:10 +00:00
|
|
|
if (node_isset(n, *used_nodes))
|
2005-09-06 22:18:14 +00:00
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Simple min distance search */
|
|
|
|
val = node_distance(node, n);
|
|
|
|
|
|
|
|
if (val < min_val) {
|
|
|
|
min_val = val;
|
|
|
|
best_node = n;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-04-05 01:11:10 +00:00
|
|
|
node_set(best_node, *used_nodes);
|
2005-09-06 22:18:14 +00:00
|
|
|
return best_node;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* sched_domain_node_span - get a cpumask for a node's sched_domain
|
|
|
|
* @node: node whose cpumask we're constructing
|
2008-04-22 17:07:22 +00:00
|
|
|
* @span: resulting cpumask
|
2005-09-06 22:18:14 +00:00
|
|
|
*
|
2007-12-05 14:46:09 +00:00
|
|
|
* Given a node, construct a good cpumask for its sched_domain to span. It
|
2005-09-06 22:18:14 +00:00
|
|
|
* should be one that prevents unnecessary balancing, but also spreads tasks
|
|
|
|
* out optimally.
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
static void sched_domain_node_span(int node, struct cpumask *span)
|
2005-09-06 22:18:14 +00:00
|
|
|
{
|
2008-04-05 01:11:10 +00:00
|
|
|
nodemask_t used_nodes;
|
2006-07-03 07:25:40 +00:00
|
|
|
int i;
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_clear(span);
|
2008-04-05 01:11:10 +00:00
|
|
|
nodes_clear(used_nodes);
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_or(span, span, cpumask_of_node(node));
|
2008-04-05 01:11:10 +00:00
|
|
|
node_set(node, used_nodes);
|
2005-09-06 22:18:14 +00:00
|
|
|
|
|
|
|
for (i = 1; i < SD_NODES_PER_DOMAIN; i++) {
|
2008-04-05 01:11:10 +00:00
|
|
|
int next_node = find_next_best_node(node, &used_nodes);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_or(span, span, cpumask_of_node(next_node));
|
2005-09-06 22:18:14 +00:00
|
|
|
}
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_NUMA */
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2006-06-27 09:54:42 +00:00
|
|
|
int sched_smt_power_savings = 0, sched_mc_power_savings = 0;
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
/*
|
|
|
|
* The cpus mask in sched_group and sched_domain hangs off the end.
|
2009-05-19 07:22:19 +00:00
|
|
|
*
|
|
|
|
* ( See the the comments in include/linux/sched.h:struct sched_group
|
|
|
|
* and struct sched_domain. )
|
2008-11-24 16:05:04 +00:00
|
|
|
*/
|
|
|
|
struct static_sched_group {
|
|
|
|
struct sched_group sg;
|
|
|
|
DECLARE_BITMAP(cpus, CONFIG_NR_CPUS);
|
|
|
|
};
|
|
|
|
|
|
|
|
struct static_sched_domain {
|
|
|
|
struct sched_domain sd;
|
|
|
|
DECLARE_BITMAP(span, CONFIG_NR_CPUS);
|
|
|
|
};
|
|
|
|
|
2005-09-06 22:18:14 +00:00
|
|
|
/*
|
2006-07-03 07:25:40 +00:00
|
|
|
* SMT sched-domains:
|
2005-09-06 22:18:14 +00:00
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SCHED_SMT
|
2008-11-24 16:05:04 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_domain, cpu_domains);
|
|
|
|
static DEFINE_PER_CPU(struct static_sched_group, sched_group_cpus);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2007-12-05 14:46:09 +00:00
|
|
|
static int
|
2008-11-24 16:05:14 +00:00
|
|
|
cpu_to_cpu_group(int cpu, const struct cpumask *cpu_map,
|
|
|
|
struct sched_group **sg, struct cpumask *unused)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-12-10 10:20:07 +00:00
|
|
|
if (sg)
|
2008-11-24 16:05:04 +00:00
|
|
|
*sg = &per_cpu(sched_group_cpus, cpu).sg;
|
2005-04-16 22:20:36 +00:00
|
|
|
return cpu;
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SCHED_SMT */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-07-03 07:25:40 +00:00
|
|
|
/*
|
|
|
|
* multi-core sched-domains:
|
|
|
|
*/
|
2006-03-27 09:15:22 +00:00
|
|
|
#ifdef CONFIG_SCHED_MC
|
2008-11-24 16:05:04 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_domain, core_domains);
|
|
|
|
static DEFINE_PER_CPU(struct static_sched_group, sched_group_core);
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SCHED_MC */
|
2006-03-27 09:15:22 +00:00
|
|
|
|
|
|
|
#if defined(CONFIG_SCHED_MC) && defined(CONFIG_SCHED_SMT)
|
2007-12-05 14:46:09 +00:00
|
|
|
static int
|
2008-11-24 16:05:14 +00:00
|
|
|
cpu_to_core_group(int cpu, const struct cpumask *cpu_map,
|
|
|
|
struct sched_group **sg, struct cpumask *mask)
|
2006-03-27 09:15:22 +00:00
|
|
|
{
|
2006-12-10 10:20:07 +00:00
|
|
|
int group;
|
2008-04-05 01:11:11 +00:00
|
|
|
|
2009-03-13 04:19:46 +00:00
|
|
|
cpumask_and(mask, topology_thread_cpumask(cpu), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
group = cpumask_first(mask);
|
2006-12-10 10:20:07 +00:00
|
|
|
if (sg)
|
2008-11-24 16:05:04 +00:00
|
|
|
*sg = &per_cpu(sched_group_core, group).sg;
|
2006-12-10 10:20:07 +00:00
|
|
|
return group;
|
2006-03-27 09:15:22 +00:00
|
|
|
}
|
|
|
|
#elif defined(CONFIG_SCHED_MC)
|
2007-12-05 14:46:09 +00:00
|
|
|
static int
|
2008-11-24 16:05:14 +00:00
|
|
|
cpu_to_core_group(int cpu, const struct cpumask *cpu_map,
|
|
|
|
struct sched_group **sg, struct cpumask *unused)
|
2006-03-27 09:15:22 +00:00
|
|
|
{
|
2006-12-10 10:20:07 +00:00
|
|
|
if (sg)
|
2008-11-24 16:05:04 +00:00
|
|
|
*sg = &per_cpu(sched_group_core, cpu).sg;
|
2006-03-27 09:15:22 +00:00
|
|
|
return cpu;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_domain, phys_domains);
|
|
|
|
static DEFINE_PER_CPU(struct static_sched_group, sched_group_phys);
|
2006-07-03 07:25:40 +00:00
|
|
|
|
2007-12-05 14:46:09 +00:00
|
|
|
static int
|
2008-11-24 16:05:14 +00:00
|
|
|
cpu_to_phys_group(int cpu, const struct cpumask *cpu_map,
|
|
|
|
struct sched_group **sg, struct cpumask *mask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2006-12-10 10:20:07 +00:00
|
|
|
int group;
|
2006-07-03 07:25:40 +00:00
|
|
|
#ifdef CONFIG_SCHED_MC
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(mask, cpu_coregroup_mask(cpu), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
group = cpumask_first(mask);
|
2006-03-27 09:15:22 +00:00
|
|
|
#elif defined(CONFIG_SCHED_SMT)
|
2009-03-13 04:19:46 +00:00
|
|
|
cpumask_and(mask, topology_thread_cpumask(cpu), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
group = cpumask_first(mask);
|
2005-04-16 22:20:36 +00:00
|
|
|
#else
|
2006-12-10 10:20:07 +00:00
|
|
|
group = cpu;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2006-12-10 10:20:07 +00:00
|
|
|
if (sg)
|
2008-11-24 16:05:04 +00:00
|
|
|
*sg = &per_cpu(sched_group_phys, group).sg;
|
2006-12-10 10:20:07 +00:00
|
|
|
return group;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_NUMA
|
|
|
|
/*
|
2005-09-06 22:18:14 +00:00
|
|
|
* The init_sched_build_groups can't handle what we want to do with node
|
|
|
|
* groups, so roll our own. Now each node has its own list of groups which
|
|
|
|
* gets dynamically allocated.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-01-11 00:04:16 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_domain, node_domains);
|
2008-04-05 01:11:04 +00:00
|
|
|
static struct sched_group ***sched_group_nodes_bycpu;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-01-11 00:04:16 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_domain, allnodes_domains);
|
2008-11-24 16:05:04 +00:00
|
|
|
static DEFINE_PER_CPU(struct static_sched_group, sched_group_allnodes);
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
static int cpu_to_allnodes_group(int cpu, const struct cpumask *cpu_map,
|
|
|
|
struct sched_group **sg,
|
|
|
|
struct cpumask *nodemask)
|
2005-09-06 22:18:14 +00:00
|
|
|
{
|
2006-12-10 10:20:07 +00:00
|
|
|
int group;
|
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(nodemask, cpumask_of_node(cpu_to_node(cpu)), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
group = cpumask_first(nodemask);
|
2006-12-10 10:20:07 +00:00
|
|
|
|
|
|
|
if (sg)
|
2008-11-24 16:05:04 +00:00
|
|
|
*sg = &per_cpu(sched_group_allnodes, group).sg;
|
2006-12-10 10:20:07 +00:00
|
|
|
return group;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-12-10 10:20:07 +00:00
|
|
|
|
2006-03-27 09:15:23 +00:00
|
|
|
static void init_numa_sched_groups_power(struct sched_group *group_head)
|
|
|
|
{
|
|
|
|
struct sched_group *sg = group_head;
|
|
|
|
int j;
|
|
|
|
|
|
|
|
if (!sg)
|
|
|
|
return;
|
2007-10-15 15:00:14 +00:00
|
|
|
do {
|
2008-11-24 16:05:04 +00:00
|
|
|
for_each_cpu(j, sched_group_cpus(sg)) {
|
2007-10-15 15:00:14 +00:00
|
|
|
struct sched_domain *sd;
|
2006-03-27 09:15:23 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(phys_domains, j).sd;
|
2009-04-15 01:59:10 +00:00
|
|
|
if (j != group_first_cpu(sd->groups)) {
|
2007-10-15 15:00:14 +00:00
|
|
|
/*
|
|
|
|
* Only add "power" once for each
|
|
|
|
* physical package.
|
|
|
|
*/
|
|
|
|
continue;
|
|
|
|
}
|
2006-03-27 09:15:23 +00:00
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
sg_inc_cpu_power(sg, sd->groups->__cpu_power);
|
|
|
|
}
|
|
|
|
sg = sg->next;
|
|
|
|
} while (sg != group_head);
|
2006-03-27 09:15:23 +00:00
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_NUMA */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-10-03 08:14:06 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
2006-06-27 09:54:38 +00:00
|
|
|
/* Free memory allocated for various sched_group structures */
|
2008-11-24 16:05:14 +00:00
|
|
|
static void free_sched_groups(const struct cpumask *cpu_map,
|
|
|
|
struct cpumask *nodemask)
|
2006-06-27 09:54:38 +00:00
|
|
|
{
|
2006-10-03 08:14:06 +00:00
|
|
|
int cpu, i;
|
2006-06-27 09:54:38 +00:00
|
|
|
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(cpu, cpu_map) {
|
2006-06-27 09:54:38 +00:00
|
|
|
struct sched_group **sched_group_nodes
|
|
|
|
= sched_group_nodes_bycpu[cpu];
|
|
|
|
|
|
|
|
if (!sched_group_nodes)
|
|
|
|
continue;
|
|
|
|
|
2008-05-12 19:21:12 +00:00
|
|
|
for (i = 0; i < nr_node_ids; i++) {
|
2006-06-27 09:54:38 +00:00
|
|
|
struct sched_group *oldsg, *sg = sched_group_nodes[i];
|
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(nodemask, cpumask_of_node(i), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_empty(nodemask))
|
2006-06-27 09:54:38 +00:00
|
|
|
continue;
|
|
|
|
|
|
|
|
if (sg == NULL)
|
|
|
|
continue;
|
|
|
|
sg = sg->next;
|
|
|
|
next_sg:
|
|
|
|
oldsg = sg;
|
|
|
|
sg = sg->next;
|
|
|
|
kfree(oldsg);
|
|
|
|
if (oldsg != sched_group_nodes[i])
|
|
|
|
goto next_sg;
|
|
|
|
}
|
|
|
|
kfree(sched_group_nodes);
|
|
|
|
sched_group_nodes_bycpu[cpu] = NULL;
|
|
|
|
}
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_NUMA */
|
2008-11-24 16:05:14 +00:00
|
|
|
static void free_sched_groups(const struct cpumask *cpu_map,
|
|
|
|
struct cpumask *nodemask)
|
2006-10-03 08:14:06 +00:00
|
|
|
{
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_NUMA */
|
2006-06-27 09:54:38 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
/*
|
|
|
|
* Initialize sched groups cpu_power.
|
|
|
|
*
|
|
|
|
* cpu_power indicates the capacity of sched group, which is used while
|
|
|
|
* distributing the load between different sched groups in a sched domain.
|
|
|
|
* Typically cpu_power for all the groups in a sched domain will be same unless
|
|
|
|
* there are asymmetries in the topology. If there are asymmetries, group
|
|
|
|
* having more cpu_power will pickup more load compared to the group having
|
|
|
|
* less cpu_power.
|
|
|
|
*
|
|
|
|
* cpu_power will be a multiple of SCHED_LOAD_SCALE. This multiple represents
|
|
|
|
* the maximum number of tasks a group can handle in the presence of other idle
|
|
|
|
* or lightly loaded groups in the same sched domain.
|
|
|
|
*/
|
|
|
|
static void init_sched_groups_power(int cpu, struct sched_domain *sd)
|
|
|
|
{
|
|
|
|
struct sched_domain *child;
|
|
|
|
struct sched_group *group;
|
|
|
|
|
|
|
|
WARN_ON(!sd || !sd->groups);
|
|
|
|
|
2009-04-15 01:59:10 +00:00
|
|
|
if (cpu != group_first_cpu(sd->groups))
|
2006-10-03 08:14:09 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
child = sd->child;
|
|
|
|
|
2007-05-08 07:32:57 +00:00
|
|
|
sd->groups->__cpu_power = 0;
|
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
/*
|
|
|
|
* For perf policy, if the groups in child domain share resources
|
|
|
|
* (for example cores sharing some portions of the cache hierarchy
|
|
|
|
* or SMT), then set this domain groups cpu_power such that each group
|
|
|
|
* can handle only one task, when there are other idle groups in the
|
|
|
|
* same sched domain.
|
|
|
|
*/
|
|
|
|
if (!child || (!(sd->flags & SD_POWERSAVINGS_BALANCE) &&
|
|
|
|
(child->flags &
|
|
|
|
(SD_SHARE_CPUPOWER | SD_SHARE_PKG_RESOURCES)))) {
|
2007-05-08 07:32:57 +00:00
|
|
|
sg_inc_cpu_power(sd->groups, SCHED_LOAD_SCALE);
|
2006-10-03 08:14:09 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* add cpu_power of each child group to this groups cpu_power
|
|
|
|
*/
|
|
|
|
group = child->groups;
|
|
|
|
do {
|
2007-05-08 07:32:57 +00:00
|
|
|
sg_inc_cpu_power(sd->groups, group->__cpu_power);
|
2006-10-03 08:14:09 +00:00
|
|
|
group = group->next;
|
|
|
|
} while (group != child->groups);
|
|
|
|
}
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
/*
|
|
|
|
* Initializers for schedule domains
|
|
|
|
* Non-inlined to reduce accumulated stack pressure in build_sched_domains()
|
|
|
|
*/
|
|
|
|
|
2008-10-09 09:35:51 +00:00
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
|
# define SD_INIT_NAME(sd, type) sd->name = #type
|
|
|
|
#else
|
|
|
|
# define SD_INIT_NAME(sd, type) do { } while (0)
|
|
|
|
#endif
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
#define SD_INIT(sd, type) sd_init_##type(sd)
|
2008-10-09 09:35:51 +00:00
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
#define SD_INIT_FUNC(type) \
|
|
|
|
static noinline void sd_init_##type(struct sched_domain *sd) \
|
|
|
|
{ \
|
|
|
|
memset(sd, 0, sizeof(*sd)); \
|
|
|
|
*sd = SD_##type##_INIT; \
|
2008-04-15 05:04:23 +00:00
|
|
|
sd->level = SD_LV_##type; \
|
2008-10-09 09:35:51 +00:00
|
|
|
SD_INIT_NAME(sd, type); \
|
2008-04-05 01:11:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
SD_INIT_FUNC(CPU)
|
|
|
|
#ifdef CONFIG_NUMA
|
|
|
|
SD_INIT_FUNC(ALLNODES)
|
|
|
|
SD_INIT_FUNC(NODE)
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_SCHED_SMT
|
|
|
|
SD_INIT_FUNC(SIBLING)
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_SCHED_MC
|
|
|
|
SD_INIT_FUNC(MC)
|
|
|
|
#endif
|
|
|
|
|
2008-04-15 05:04:23 +00:00
|
|
|
static int default_relax_domain_level = -1;
|
|
|
|
|
|
|
|
static int __init setup_relax_domain_level(char *str)
|
|
|
|
{
|
2008-05-13 02:27:17 +00:00
|
|
|
unsigned long val;
|
|
|
|
|
|
|
|
val = simple_strtoul(str, NULL, 0);
|
|
|
|
if (val < SD_LV_MAX)
|
|
|
|
default_relax_domain_level = val;
|
|
|
|
|
2008-04-15 05:04:23 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
__setup("relax_domain_level=", setup_relax_domain_level);
|
|
|
|
|
|
|
|
static void set_domain_attribute(struct sched_domain *sd,
|
|
|
|
struct sched_domain_attr *attr)
|
|
|
|
{
|
|
|
|
int request;
|
|
|
|
|
|
|
|
if (!attr || attr->relax_domain_level < 0) {
|
|
|
|
if (default_relax_domain_level < 0)
|
|
|
|
return;
|
|
|
|
else
|
|
|
|
request = default_relax_domain_level;
|
|
|
|
} else
|
|
|
|
request = attr->relax_domain_level;
|
|
|
|
if (request < sd->level) {
|
|
|
|
/* turn off idle balance on this domain */
|
|
|
|
sd->flags &= ~(SD_WAKE_IDLE|SD_BALANCE_NEWIDLE);
|
|
|
|
} else {
|
|
|
|
/* turn on idle balance on this domain */
|
|
|
|
sd->flags |= (SD_WAKE_IDLE_FAR|SD_BALANCE_NEWIDLE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2005-06-25 21:57:33 +00:00
|
|
|
* Build sched domains for a given set of cpus and attach the sched domains
|
|
|
|
* to the individual cpus
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
static int __build_sched_domains(const struct cpumask *cpu_map,
|
2008-04-15 05:04:23 +00:00
|
|
|
struct sched_domain_attr *attr)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-11-24 16:05:03 +00:00
|
|
|
int i, err = -ENOMEM;
|
2008-01-25 20:08:18 +00:00
|
|
|
struct root_domain *rd;
|
2008-11-24 16:05:03 +00:00
|
|
|
cpumask_var_t nodemask, this_sibling_map, this_core_map, send_covered,
|
|
|
|
tmpmask;
|
2005-09-06 22:18:14 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
2008-11-24 16:05:03 +00:00
|
|
|
cpumask_var_t domainspan, covered, notcovered;
|
2005-09-06 22:18:14 +00:00
|
|
|
struct sched_group **sched_group_nodes = NULL;
|
2006-12-10 10:20:07 +00:00
|
|
|
int sd_allnodes = 0;
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-11-24 16:05:03 +00:00
|
|
|
if (!alloc_cpumask_var(&domainspan, GFP_KERNEL))
|
|
|
|
goto out;
|
|
|
|
if (!alloc_cpumask_var(&covered, GFP_KERNEL))
|
|
|
|
goto free_domainspan;
|
|
|
|
if (!alloc_cpumask_var(¬covered, GFP_KERNEL))
|
|
|
|
goto free_covered;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
if (!alloc_cpumask_var(&nodemask, GFP_KERNEL))
|
|
|
|
goto free_notcovered;
|
|
|
|
if (!alloc_cpumask_var(&this_sibling_map, GFP_KERNEL))
|
|
|
|
goto free_nodemask;
|
|
|
|
if (!alloc_cpumask_var(&this_core_map, GFP_KERNEL))
|
|
|
|
goto free_this_sibling_map;
|
|
|
|
if (!alloc_cpumask_var(&send_covered, GFP_KERNEL))
|
|
|
|
goto free_this_core_map;
|
|
|
|
if (!alloc_cpumask_var(&tmpmask, GFP_KERNEL))
|
|
|
|
goto free_send_covered;
|
|
|
|
|
|
|
|
#ifdef CONFIG_NUMA
|
2005-09-06 22:18:14 +00:00
|
|
|
/*
|
|
|
|
* Allocate the per-node list of sched groups
|
|
|
|
*/
|
2008-05-12 19:21:12 +00:00
|
|
|
sched_group_nodes = kcalloc(nr_node_ids, sizeof(struct sched_group *),
|
2007-12-05 14:46:09 +00:00
|
|
|
GFP_KERNEL);
|
2005-09-06 22:18:14 +00:00
|
|
|
if (!sched_group_nodes) {
|
|
|
|
printk(KERN_WARNING "Can not alloc sched group node list\n");
|
2008-11-24 16:05:03 +00:00
|
|
|
goto free_tmpmask;
|
2005-09-06 22:18:14 +00:00
|
|
|
}
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-01-25 20:08:26 +00:00
|
|
|
rd = alloc_rootdomain();
|
2008-01-25 20:08:18 +00:00
|
|
|
if (!rd) {
|
|
|
|
printk(KERN_WARNING "Cannot alloc root domain\n");
|
2008-11-24 16:05:03 +00:00
|
|
|
goto free_sched_groups;
|
2008-01-25 20:08:18 +00:00
|
|
|
}
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
2008-11-24 16:05:14 +00:00
|
|
|
sched_group_nodes_bycpu[cpumask_first(cpu_map)] = sched_group_nodes;
|
2008-04-05 01:11:11 +00:00
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2005-06-25 21:57:33 +00:00
|
|
|
* Set up domains for cpus specified by the cpu_map.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sched_domain *sd = NULL, *p;
|
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(nodemask, cpumask_of_node(cpu_to_node(i)), cpu_map);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_NUMA
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_weight(cpu_map) >
|
|
|
|
SD_NODES_PER_DOMAIN*cpumask_weight(nodemask)) {
|
2009-01-11 00:04:16 +00:00
|
|
|
sd = &per_cpu(allnodes_domains, i).sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
SD_INIT(sd, ALLNODES);
|
2008-04-15 05:04:23 +00:00
|
|
|
set_domain_attribute(sd, attr);
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_copy(sched_domain_span(sd), cpu_map);
|
2008-04-05 01:11:11 +00:00
|
|
|
cpu_to_allnodes_group(i, cpu_map, &sd->groups, tmpmask);
|
2005-09-06 22:18:14 +00:00
|
|
|
p = sd;
|
2006-12-10 10:20:07 +00:00
|
|
|
sd_allnodes = 1;
|
2005-09-06 22:18:14 +00:00
|
|
|
} else
|
|
|
|
p = NULL;
|
|
|
|
|
2009-01-11 00:04:16 +00:00
|
|
|
sd = &per_cpu(node_domains, i).sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
SD_INIT(sd, NODE);
|
2008-04-15 05:04:23 +00:00
|
|
|
set_domain_attribute(sd, attr);
|
2008-11-24 16:05:04 +00:00
|
|
|
sched_domain_node_span(cpu_to_node(i), sched_domain_span(sd));
|
2005-09-06 22:18:14 +00:00
|
|
|
sd->parent = p;
|
2006-10-03 08:14:08 +00:00
|
|
|
if (p)
|
|
|
|
p->child = sd;
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_and(sched_domain_span(sd),
|
|
|
|
sched_domain_span(sd), cpu_map);
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
p = sd;
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(phys_domains, i).sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
SD_INIT(sd, CPU);
|
2008-04-15 05:04:23 +00:00
|
|
|
set_domain_attribute(sd, attr);
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_copy(sched_domain_span(sd), nodemask);
|
2005-04-16 22:20:36 +00:00
|
|
|
sd->parent = p;
|
2006-10-03 08:14:08 +00:00
|
|
|
if (p)
|
|
|
|
p->child = sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
cpu_to_phys_group(i, cpu_map, &sd->groups, tmpmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-03-27 09:15:22 +00:00
|
|
|
#ifdef CONFIG_SCHED_MC
|
|
|
|
p = sd;
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(core_domains, i).sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
SD_INIT(sd, MC);
|
2008-04-15 05:04:23 +00:00
|
|
|
set_domain_attribute(sd, attr);
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(sched_domain_span(sd), cpu_map,
|
|
|
|
cpu_coregroup_mask(i));
|
2006-03-27 09:15:22 +00:00
|
|
|
sd->parent = p;
|
2006-10-03 08:14:08 +00:00
|
|
|
p->child = sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
cpu_to_core_group(i, cpu_map, &sd->groups, tmpmask);
|
2006-03-27 09:15:22 +00:00
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SCHED_SMT
|
|
|
|
p = sd;
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(cpu_domains, i).sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
SD_INIT(sd, SIBLING);
|
2008-04-15 05:04:23 +00:00
|
|
|
set_domain_attribute(sd, attr);
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_and(sched_domain_span(sd),
|
2009-03-13 04:19:46 +00:00
|
|
|
topology_thread_cpumask(i), cpu_map);
|
2005-04-16 22:20:36 +00:00
|
|
|
sd->parent = p;
|
2006-10-03 08:14:08 +00:00
|
|
|
p->child = sd;
|
2008-04-05 01:11:11 +00:00
|
|
|
cpu_to_cpu_group(i, cpu_map, &sd->groups, tmpmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCHED_SMT
|
|
|
|
/* Set up CPU (sibling) groups */
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_and(this_sibling_map,
|
2009-03-13 04:19:46 +00:00
|
|
|
topology_thread_cpumask(i), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (i != cpumask_first(this_sibling_map))
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
init_sched_build_groups(this_sibling_map, cpu_map,
|
2008-04-05 01:11:11 +00:00
|
|
|
&cpu_to_cpu_group,
|
|
|
|
send_covered, tmpmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2006-03-27 09:15:22 +00:00
|
|
|
#ifdef CONFIG_SCHED_MC
|
|
|
|
/* Set up multi-core groups */
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(this_core_map, cpu_coregroup_mask(i), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (i != cpumask_first(this_core_map))
|
2006-03-27 09:15:22 +00:00
|
|
|
continue;
|
2008-04-05 01:11:11 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
init_sched_build_groups(this_core_map, cpu_map,
|
2008-04-05 01:11:11 +00:00
|
|
|
&cpu_to_core_group,
|
|
|
|
send_covered, tmpmask);
|
2006-03-27 09:15:22 +00:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Set up physical groups */
|
2008-05-12 19:21:12 +00:00
|
|
|
for (i = 0; i < nr_node_ids; i++) {
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(nodemask, cpumask_of_node(i), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_empty(nodemask))
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
|
|
|
|
2008-04-05 01:11:11 +00:00
|
|
|
init_sched_build_groups(nodemask, cpu_map,
|
|
|
|
&cpu_to_phys_group,
|
|
|
|
send_covered, tmpmask);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_NUMA
|
|
|
|
/* Set up node groups */
|
2008-04-05 01:11:11 +00:00
|
|
|
if (sd_allnodes) {
|
|
|
|
init_sched_build_groups(cpu_map, cpu_map,
|
|
|
|
&cpu_to_allnodes_group,
|
|
|
|
send_covered, tmpmask);
|
|
|
|
}
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-05-12 19:21:12 +00:00
|
|
|
for (i = 0; i < nr_node_ids; i++) {
|
2005-09-06 22:18:14 +00:00
|
|
|
/* Set up node groups */
|
|
|
|
struct sched_group *sg, *prev;
|
|
|
|
int j;
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_clear(covered);
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(nodemask, cpumask_of_node(i), cpu_map);
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_empty(nodemask)) {
|
2005-09-06 22:18:14 +00:00
|
|
|
sched_group_nodes[i] = NULL;
|
2005-09-06 22:18:14 +00:00
|
|
|
continue;
|
2005-09-06 22:18:14 +00:00
|
|
|
}
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-04-15 23:35:52 +00:00
|
|
|
sched_domain_node_span(i, domainspan);
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_and(domainspan, domainspan, cpu_map);
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
sg = kmalloc_node(sizeof(struct sched_group) + cpumask_size(),
|
|
|
|
GFP_KERNEL, i);
|
2006-06-27 09:54:38 +00:00
|
|
|
if (!sg) {
|
|
|
|
printk(KERN_WARNING "Can not alloc domain group for "
|
|
|
|
"node %d\n", i);
|
|
|
|
goto error;
|
|
|
|
}
|
2005-09-06 22:18:14 +00:00
|
|
|
sched_group_nodes[i] = sg;
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(j, nodemask) {
|
2005-09-06 22:18:14 +00:00
|
|
|
struct sched_domain *sd;
|
2007-07-09 16:52:00 +00:00
|
|
|
|
2009-01-11 00:04:16 +00:00
|
|
|
sd = &per_cpu(node_domains, j).sd;
|
2005-09-06 22:18:14 +00:00
|
|
|
sd->groups = sg;
|
|
|
|
}
|
2007-05-08 07:32:57 +00:00
|
|
|
sg->__cpu_power = 0;
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_copy(sched_group_cpus(sg), nodemask);
|
2006-06-27 09:54:38 +00:00
|
|
|
sg->next = sg;
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_or(covered, covered, nodemask);
|
2005-09-06 22:18:14 +00:00
|
|
|
prev = sg;
|
|
|
|
|
2008-05-12 19:21:12 +00:00
|
|
|
for (j = 0; j < nr_node_ids; j++) {
|
|
|
|
int n = (i + j) % nr_node_ids;
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_complement(notcovered, covered);
|
|
|
|
cpumask_and(tmpmask, notcovered, cpu_map);
|
|
|
|
cpumask_and(tmpmask, tmpmask, domainspan);
|
|
|
|
if (cpumask_empty(tmpmask))
|
2005-09-06 22:18:14 +00:00
|
|
|
break;
|
|
|
|
|
2009-01-01 02:08:45 +00:00
|
|
|
cpumask_and(tmpmask, tmpmask, cpumask_of_node(n));
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_empty(tmpmask))
|
2005-09-06 22:18:14 +00:00
|
|
|
continue;
|
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
sg = kmalloc_node(sizeof(struct sched_group) +
|
|
|
|
cpumask_size(),
|
2006-06-27 09:54:40 +00:00
|
|
|
GFP_KERNEL, i);
|
2005-09-06 22:18:14 +00:00
|
|
|
if (!sg) {
|
|
|
|
printk(KERN_WARNING
|
|
|
|
"Can not alloc domain group for node %d\n", j);
|
2006-06-27 09:54:38 +00:00
|
|
|
goto error;
|
2005-09-06 22:18:14 +00:00
|
|
|
}
|
2007-05-08 07:32:57 +00:00
|
|
|
sg->__cpu_power = 0;
|
2008-11-24 16:05:04 +00:00
|
|
|
cpumask_copy(sched_group_cpus(sg), tmpmask);
|
2006-06-27 09:54:38 +00:00
|
|
|
sg->next = prev->next;
|
2008-11-24 16:05:14 +00:00
|
|
|
cpumask_or(covered, covered, tmpmask);
|
2005-09-06 22:18:14 +00:00
|
|
|
prev->next = sg;
|
|
|
|
prev = sg;
|
|
|
|
}
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/* Calculate CPU power for physical packages and nodes */
|
2006-06-27 09:54:42 +00:00
|
|
|
#ifdef CONFIG_SCHED_SMT
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2008-11-24 16:05:04 +00:00
|
|
|
struct sched_domain *sd = &per_cpu(cpu_domains, i).sd;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
init_sched_groups_power(i, sd);
|
2006-06-27 09:54:42 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2006-03-27 09:15:22 +00:00
|
|
|
#ifdef CONFIG_SCHED_MC
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2008-11-24 16:05:04 +00:00
|
|
|
struct sched_domain *sd = &per_cpu(core_domains, i).sd;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
init_sched_groups_power(i, sd);
|
2006-06-27 09:54:42 +00:00
|
|
|
}
|
|
|
|
#endif
|
2006-03-27 09:15:22 +00:00
|
|
|
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2008-11-24 16:05:04 +00:00
|
|
|
struct sched_domain *sd = &per_cpu(phys_domains, i).sd;
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2006-10-03 08:14:09 +00:00
|
|
|
init_sched_groups_power(i, sd);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 22:18:14 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
2008-05-12 19:21:12 +00:00
|
|
|
for (i = 0; i < nr_node_ids; i++)
|
2006-03-27 09:15:23 +00:00
|
|
|
init_numa_sched_groups_power(sched_group_nodes[i]);
|
2005-09-06 22:18:14 +00:00
|
|
|
|
2006-12-10 10:20:07 +00:00
|
|
|
if (sd_allnodes) {
|
|
|
|
struct sched_group *sg;
|
2006-07-30 10:02:59 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
cpu_to_allnodes_group(cpumask_first(cpu_map), cpu_map, &sg,
|
2008-04-05 01:11:11 +00:00
|
|
|
tmpmask);
|
2006-07-30 10:02:59 +00:00
|
|
|
init_numa_sched_groups_power(sg);
|
|
|
|
}
|
2005-09-06 22:18:14 +00:00
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Attach the domains */
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map) {
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sched_domain *sd;
|
|
|
|
#ifdef CONFIG_SCHED_SMT
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(cpu_domains, i).sd;
|
2006-03-27 09:15:22 +00:00
|
|
|
#elif defined(CONFIG_SCHED_MC)
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(core_domains, i).sd;
|
2005-04-16 22:20:36 +00:00
|
|
|
#else
|
2008-11-24 16:05:04 +00:00
|
|
|
sd = &per_cpu(phys_domains, i).sd;
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2008-01-25 20:08:18 +00:00
|
|
|
cpu_attach_domain(sd, rd, i);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2006-06-27 09:54:38 +00:00
|
|
|
|
2008-11-24 16:05:03 +00:00
|
|
|
err = 0;
|
|
|
|
|
|
|
|
free_tmpmask:
|
|
|
|
free_cpumask_var(tmpmask);
|
|
|
|
free_send_covered:
|
|
|
|
free_cpumask_var(send_covered);
|
|
|
|
free_this_core_map:
|
|
|
|
free_cpumask_var(this_core_map);
|
|
|
|
free_this_sibling_map:
|
|
|
|
free_cpumask_var(this_sibling_map);
|
|
|
|
free_nodemask:
|
|
|
|
free_cpumask_var(nodemask);
|
|
|
|
free_notcovered:
|
|
|
|
#ifdef CONFIG_NUMA
|
|
|
|
free_cpumask_var(notcovered);
|
|
|
|
free_covered:
|
|
|
|
free_cpumask_var(covered);
|
|
|
|
free_domainspan:
|
|
|
|
free_cpumask_var(domainspan);
|
|
|
|
out:
|
|
|
|
#endif
|
|
|
|
return err;
|
|
|
|
|
|
|
|
free_sched_groups:
|
|
|
|
#ifdef CONFIG_NUMA
|
|
|
|
kfree(sched_group_nodes);
|
|
|
|
#endif
|
|
|
|
goto free_tmpmask;
|
2006-06-27 09:54:38 +00:00
|
|
|
|
2006-10-03 08:14:06 +00:00
|
|
|
#ifdef CONFIG_NUMA
|
2006-06-27 09:54:38 +00:00
|
|
|
error:
|
2008-04-05 01:11:11 +00:00
|
|
|
free_sched_groups(cpu_map, tmpmask);
|
2008-11-24 16:05:05 +00:00
|
|
|
free_rootdomain(rd);
|
2008-11-24 16:05:03 +00:00
|
|
|
goto free_tmpmask;
|
2006-10-03 08:14:06 +00:00
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-10-19 06:40:20 +00:00
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
static int build_sched_domains(const struct cpumask *cpu_map)
|
2008-04-15 05:04:23 +00:00
|
|
|
{
|
|
|
|
return __build_sched_domains(cpu_map, NULL);
|
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
static struct cpumask *doms_cur; /* current sched domains */
|
2007-10-19 06:40:20 +00:00
|
|
|
static int ndoms_cur; /* number of sched domains in 'doms_cur' */
|
2008-05-16 15:47:14 +00:00
|
|
|
static struct sched_domain_attr *dattr_cur;
|
|
|
|
/* attribues of custom domains in 'doms_cur' */
|
2007-10-19 06:40:20 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Special case: If a kmalloc of a doms_cur partition (array of
|
2008-11-24 16:05:12 +00:00
|
|
|
* cpumask) fails, then fallback to a single sched domain,
|
|
|
|
* as determined by the single cpumask fallback_doms.
|
2007-10-19 06:40:20 +00:00
|
|
|
*/
|
2008-11-24 16:05:12 +00:00
|
|
|
static cpumask_var_t fallback_doms;
|
2007-10-19 06:40:20 +00:00
|
|
|
|
2008-12-09 17:49:50 +00:00
|
|
|
/*
|
|
|
|
* arch_update_cpu_topology lets virtualized architectures update the
|
|
|
|
* cpu core maps. It is supposed to return 1 if the topology changed
|
|
|
|
* or 0 if it stayed the same.
|
|
|
|
*/
|
|
|
|
int __attribute__((weak)) arch_update_cpu_topology(void)
|
2008-03-12 17:31:59 +00:00
|
|
|
{
|
2008-12-09 17:49:50 +00:00
|
|
|
return 0;
|
2008-03-12 17:31:59 +00:00
|
|
|
}
|
|
|
|
|
2005-06-25 21:57:33 +00:00
|
|
|
/*
|
2007-12-05 14:46:09 +00:00
|
|
|
* Set up scheduler domains and groups. Callers must hold the hotplug lock.
|
2007-10-19 06:40:20 +00:00
|
|
|
* For now this just excludes isolated cpus, but could be used to
|
|
|
|
* exclude other special cases in the future.
|
2005-06-25 21:57:33 +00:00
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
static int arch_init_sched_domains(const struct cpumask *cpu_map)
|
2005-06-25 21:57:33 +00:00
|
|
|
{
|
2007-10-24 16:23:48 +00:00
|
|
|
int err;
|
|
|
|
|
2008-03-12 17:31:59 +00:00
|
|
|
arch_update_cpu_topology();
|
2007-10-19 06:40:20 +00:00
|
|
|
ndoms_cur = 1;
|
2008-11-24 16:05:14 +00:00
|
|
|
doms_cur = kmalloc(cpumask_size(), GFP_KERNEL);
|
2007-10-19 06:40:20 +00:00
|
|
|
if (!doms_cur)
|
2008-11-24 16:05:12 +00:00
|
|
|
doms_cur = fallback_doms;
|
2008-11-24 16:05:12 +00:00
|
|
|
cpumask_andnot(doms_cur, cpu_map, cpu_isolated_map);
|
2008-04-15 05:04:23 +00:00
|
|
|
dattr_cur = NULL;
|
2007-10-24 16:23:48 +00:00
|
|
|
err = build_sched_domains(doms_cur);
|
2007-10-15 15:00:19 +00:00
|
|
|
register_sched_domain_sysctl();
|
2007-10-24 16:23:48 +00:00
|
|
|
|
|
|
|
return err;
|
2005-06-25 21:57:33 +00:00
|
|
|
}
|
|
|
|
|
2008-11-24 16:05:14 +00:00
|
|
|
static void arch_destroy_sched_domains(const struct cpumask *cpu_map,
|
|
|
|
struct cpumask *tmpmask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-04-05 01:11:11 +00:00
|
|
|
free_sched_groups(cpu_map, tmpmask);
|
2005-09-06 22:18:14 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-06-25 21:57:33 +00:00
|
|
|
/*
|
|
|
|
* Detach sched domains from a group of cpus specified in cpu_map
|
|
|
|
* These cpus will now be attached to the NULL domain
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
static void detach_destroy_domains(const struct cpumask *cpu_map)
|
2005-06-25 21:57:33 +00:00
|
|
|
{
|
2008-11-24 16:05:14 +00:00
|
|
|
/* Save because hotplug lock held. */
|
|
|
|
static DECLARE_BITMAP(tmpmask, CONFIG_NR_CPUS);
|
2005-06-25 21:57:33 +00:00
|
|
|
int i;
|
|
|
|
|
2008-11-24 16:05:02 +00:00
|
|
|
for_each_cpu(i, cpu_map)
|
2008-01-25 20:08:18 +00:00
|
|
|
cpu_attach_domain(NULL, &def_root_domain, i);
|
2005-06-25 21:57:33 +00:00
|
|
|
synchronize_sched();
|
2008-11-24 16:05:14 +00:00
|
|
|
arch_destroy_sched_domains(cpu_map, to_cpumask(tmpmask));
|
2005-06-25 21:57:33 +00:00
|
|
|
}
|
|
|
|
|
2008-04-15 05:04:23 +00:00
|
|
|
/* handle null as "default" */
|
|
|
|
static int dattrs_equal(struct sched_domain_attr *cur, int idx_cur,
|
|
|
|
struct sched_domain_attr *new, int idx_new)
|
|
|
|
{
|
|
|
|
struct sched_domain_attr tmp;
|
|
|
|
|
|
|
|
/* fast path */
|
|
|
|
if (!new && !cur)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
tmp = SD_ATTR_INIT;
|
|
|
|
return !memcmp(cur ? (cur + idx_cur) : &tmp,
|
|
|
|
new ? (new + idx_new) : &tmp,
|
|
|
|
sizeof(struct sched_domain_attr));
|
|
|
|
}
|
|
|
|
|
2007-10-19 06:40:20 +00:00
|
|
|
/*
|
|
|
|
* Partition sched domains as specified by the 'ndoms_new'
|
2007-12-05 14:46:09 +00:00
|
|
|
* cpumasks in the array doms_new[] of cpumasks. This compares
|
2007-10-19 06:40:20 +00:00
|
|
|
* doms_new[] to the current sched domain partitioning, doms_cur[].
|
|
|
|
* It destroys each deleted domain and builds each new domain.
|
|
|
|
*
|
2008-11-24 16:05:14 +00:00
|
|
|
* 'doms_new' is an array of cpumask's of length 'ndoms_new'.
|
2007-12-05 14:46:09 +00:00
|
|
|
* The masks don't intersect (don't overlap.) We should setup one
|
|
|
|
* sched domain for each mask. CPUs not in any of the cpumasks will
|
|
|
|
* not be load balanced. If the same cpumask appears both in the
|
2007-10-19 06:40:20 +00:00
|
|
|
* current 'doms_cur' domains and in the new 'doms_new', we can leave
|
|
|
|
* it as it is.
|
|
|
|
*
|
2007-12-05 14:46:09 +00:00
|
|
|
* The passed in 'doms_new' should be kmalloc'd. This routine takes
|
|
|
|
* ownership of it and will kfree it when done with it. If the caller
|
2008-11-18 06:02:03 +00:00
|
|
|
* failed the kmalloc call, then it can pass in doms_new == NULL &&
|
|
|
|
* ndoms_new == 1, and partition_sched_domains() will fallback to
|
|
|
|
* the single partition 'fallback_doms', it also forces the domains
|
|
|
|
* to be rebuilt.
|
2007-10-19 06:40:20 +00:00
|
|
|
*
|
2008-11-24 16:05:14 +00:00
|
|
|
* If doms_new == NULL it will be replaced with cpu_online_mask.
|
2008-11-18 06:02:03 +00:00
|
|
|
* ndoms_new == 0 is a special case for destroying existing domains,
|
|
|
|
* and it will not create the default domain.
|
2008-08-29 20:11:41 +00:00
|
|
|
*
|
2007-10-19 06:40:20 +00:00
|
|
|
* Call with hotplug lock held
|
|
|
|
*/
|
2008-11-24 16:05:14 +00:00
|
|
|
/* FIXME: Change to struct cpumask *doms_new[] */
|
|
|
|
void partition_sched_domains(int ndoms_new, struct cpumask *doms_new,
|
2008-04-15 05:04:23 +00:00
|
|
|
struct sched_domain_attr *dattr_new)
|
2007-10-19 06:40:20 +00:00
|
|
|
{
|
2008-08-29 20:11:41 +00:00
|
|
|
int i, j, n;
|
2008-12-09 17:49:51 +00:00
|
|
|
int new_topology;
|
2007-10-19 06:40:20 +00:00
|
|
|
|
2008-04-28 09:33:07 +00:00
|
|
|
mutex_lock(&sched_domains_mutex);
|
2008-01-25 20:08:00 +00:00
|
|
|
|
2007-10-24 16:23:48 +00:00
|
|
|
/* always unregister in case we don't destroy any domains */
|
|
|
|
unregister_sched_domain_sysctl();
|
|
|
|
|
2008-12-09 17:49:51 +00:00
|
|
|
/* Let architecture update cpu core mappings. */
|
|
|
|
new_topology = arch_update_cpu_topology();
|
|
|
|
|
2008-08-29 20:11:41 +00:00
|
|
|
n = doms_new ? ndoms_new : 0;
|
2007-10-19 06:40:20 +00:00
|
|
|
|
|
|
|
/* Destroy deleted domains */
|
|
|
|
for (i = 0; i < ndoms_cur; i++) {
|
2008-12-09 17:49:51 +00:00
|
|
|
for (j = 0; j < n && !new_topology; j++) {
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_equal(&doms_cur[i], &doms_new[j])
|
2008-04-15 05:04:23 +00:00
|
|
|
&& dattrs_equal(dattr_cur, i, dattr_new, j))
|
2007-10-19 06:40:20 +00:00
|
|
|
goto match1;
|
|
|
|
}
|
|
|
|
/* no match - a current sched domain not in new doms_new[] */
|
|
|
|
detach_destroy_domains(doms_cur + i);
|
|
|
|
match1:
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
2008-07-15 11:43:49 +00:00
|
|
|
if (doms_new == NULL) {
|
|
|
|
ndoms_cur = 0;
|
2008-11-24 16:05:12 +00:00
|
|
|
doms_new = fallback_doms;
|
2008-11-24 16:05:12 +00:00
|
|
|
cpumask_andnot(&doms_new[0], cpu_online_mask, cpu_isolated_map);
|
2008-11-04 08:20:23 +00:00
|
|
|
WARN_ON_ONCE(dattr_new);
|
2008-07-15 11:43:49 +00:00
|
|
|
}
|
|
|
|
|
2007-10-19 06:40:20 +00:00
|
|
|
/* Build new domains */
|
|
|
|
for (i = 0; i < ndoms_new; i++) {
|
2008-12-09 17:49:51 +00:00
|
|
|
for (j = 0; j < ndoms_cur && !new_topology; j++) {
|
2008-11-24 16:05:14 +00:00
|
|
|
if (cpumask_equal(&doms_new[i], &doms_cur[j])
|
2008-04-15 05:04:23 +00:00
|
|
|
&& dattrs_equal(dattr_new, i, dattr_cur, j))
|
2007-10-19 06:40:20 +00:00
|
|
|
goto match2;
|
|
|
|
}
|
|
|
|
/* no match - add a new doms_new */
|
2008-04-15 05:04:23 +00:00
|
|
|
__build_sched_domains(doms_new + i,
|
|
|
|
dattr_new ? dattr_new + i : NULL);
|
2007-10-19 06:40:20 +00:00
|
|
|
match2:
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Remember the new sched domains */
|
2008-11-24 16:05:12 +00:00
|
|
|
if (doms_cur != fallback_doms)
|
2007-10-19 06:40:20 +00:00
|
|
|
kfree(doms_cur);
|
2008-04-15 05:04:23 +00:00
|
|
|
kfree(dattr_cur); /* kfree(NULL) is safe */
|
2007-10-19 06:40:20 +00:00
|
|
|
doms_cur = doms_new;
|
2008-04-15 05:04:23 +00:00
|
|
|
dattr_cur = dattr_new;
|
2007-10-19 06:40:20 +00:00
|
|
|
ndoms_cur = ndoms_new;
|
2007-10-24 16:23:48 +00:00
|
|
|
|
|
|
|
register_sched_domain_sysctl();
|
2008-01-25 20:08:00 +00:00
|
|
|
|
2008-04-28 09:33:07 +00:00
|
|
|
mutex_unlock(&sched_domains_mutex);
|
2007-10-19 06:40:20 +00:00
|
|
|
}
|
|
|
|
|
2006-06-27 09:54:42 +00:00
|
|
|
#if defined(CONFIG_SCHED_MC) || defined(CONFIG_SCHED_SMT)
|
2009-01-05 11:07:50 +00:00
|
|
|
static void arch_reinit_sched_domains(void)
|
2006-06-27 09:54:42 +00:00
|
|
|
{
|
2008-01-25 20:08:02 +00:00
|
|
|
get_online_cpus();
|
2008-08-29 20:11:41 +00:00
|
|
|
|
|
|
|
/* Destroy domains first to force the rebuild */
|
|
|
|
partition_sched_domains(0, NULL, NULL);
|
|
|
|
|
2008-07-15 11:43:49 +00:00
|
|
|
rebuild_sched_domains();
|
2008-01-25 20:08:02 +00:00
|
|
|
put_online_cpus();
|
2006-06-27 09:54:42 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t sched_power_savings_store(const char *buf, size_t count, int smt)
|
|
|
|
{
|
2008-12-18 17:56:09 +00:00
|
|
|
unsigned int level = 0;
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2008-12-18 17:56:09 +00:00
|
|
|
if (sscanf(buf, "%u", &level) != 1)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* level is always be positive so don't check for
|
|
|
|
* level < POWERSAVINGS_BALANCE_NONE which is 0
|
|
|
|
* What happens on 0 or 1 byte write,
|
|
|
|
* need to check for count as well?
|
|
|
|
*/
|
|
|
|
|
|
|
|
if (level >= MAX_POWERSAVINGS_BALANCE_LEVELS)
|
2006-06-27 09:54:42 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (smt)
|
2008-12-18 17:56:09 +00:00
|
|
|
sched_smt_power_savings = level;
|
2006-06-27 09:54:42 +00:00
|
|
|
else
|
2008-12-18 17:56:09 +00:00
|
|
|
sched_mc_power_savings = level;
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-01-05 11:07:50 +00:00
|
|
|
arch_reinit_sched_domains();
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2009-01-05 11:07:50 +00:00
|
|
|
return count;
|
2006-06-27 09:54:42 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCHED_MC
|
2008-07-30 05:33:52 +00:00
|
|
|
static ssize_t sched_mc_power_savings_show(struct sysdev_class *class,
|
|
|
|
char *page)
|
2006-06-27 09:54:42 +00:00
|
|
|
{
|
|
|
|
return sprintf(page, "%u\n", sched_mc_power_savings);
|
|
|
|
}
|
2008-07-30 05:33:52 +00:00
|
|
|
static ssize_t sched_mc_power_savings_store(struct sysdev_class *class,
|
2006-07-03 07:25:40 +00:00
|
|
|
const char *buf, size_t count)
|
2006-06-27 09:54:42 +00:00
|
|
|
{
|
|
|
|
return sched_power_savings_store(buf, count, 0);
|
|
|
|
}
|
2008-07-30 05:33:52 +00:00
|
|
|
static SYSDEV_CLASS_ATTR(sched_mc_power_savings, 0644,
|
|
|
|
sched_mc_power_savings_show,
|
|
|
|
sched_mc_power_savings_store);
|
2006-06-27 09:54:42 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCHED_SMT
|
2008-07-30 05:33:52 +00:00
|
|
|
static ssize_t sched_smt_power_savings_show(struct sysdev_class *dev,
|
|
|
|
char *page)
|
2006-06-27 09:54:42 +00:00
|
|
|
{
|
|
|
|
return sprintf(page, "%u\n", sched_smt_power_savings);
|
|
|
|
}
|
2008-07-30 05:33:52 +00:00
|
|
|
static ssize_t sched_smt_power_savings_store(struct sysdev_class *dev,
|
2006-07-03 07:25:40 +00:00
|
|
|
const char *buf, size_t count)
|
2006-06-27 09:54:42 +00:00
|
|
|
{
|
|
|
|
return sched_power_savings_store(buf, count, 1);
|
|
|
|
}
|
2008-07-30 05:33:52 +00:00
|
|
|
static SYSDEV_CLASS_ATTR(sched_smt_power_savings, 0644,
|
|
|
|
sched_smt_power_savings_show,
|
2007-08-12 16:08:19 +00:00
|
|
|
sched_smt_power_savings_store);
|
|
|
|
#endif
|
|
|
|
|
2009-01-05 11:18:02 +00:00
|
|
|
int __init sched_create_sysfs_power_savings_entries(struct sysdev_class *cls)
|
2007-08-12 16:08:19 +00:00
|
|
|
{
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCHED_SMT
|
|
|
|
if (smt_capable())
|
|
|
|
err = sysfs_create_file(&cls->kset.kobj,
|
|
|
|
&attr_sched_smt_power_savings.attr);
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_SCHED_MC
|
|
|
|
if (!err && mc_capable())
|
|
|
|
err = sysfs_create_file(&cls->kset.kobj,
|
|
|
|
&attr_sched_mc_power_savings.attr);
|
|
|
|
#endif
|
|
|
|
return err;
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_SCHED_MC || CONFIG_SCHED_SMT */
|
2006-06-27 09:54:42 +00:00
|
|
|
|
2008-07-15 11:43:49 +00:00
|
|
|
#ifndef CONFIG_CPUSETS
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2008-07-15 11:43:49 +00:00
|
|
|
* Add online and remove offline CPUs from the scheduler domains.
|
|
|
|
* When cpusets are enabled they take over this function.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
static int update_sched_domains(struct notifier_block *nfb,
|
|
|
|
unsigned long action, void *hcpu)
|
2008-07-15 11:43:49 +00:00
|
|
|
{
|
|
|
|
switch (action) {
|
|
|
|
case CPU_ONLINE:
|
|
|
|
case CPU_ONLINE_FROZEN:
|
|
|
|
case CPU_DEAD:
|
|
|
|
case CPU_DEAD_FROZEN:
|
2008-08-29 20:11:41 +00:00
|
|
|
partition_sched_domains(1, NULL, NULL);
|
2008-07-15 11:43:49 +00:00
|
|
|
return NOTIFY_OK;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static int update_runtime(struct notifier_block *nfb,
|
|
|
|
unsigned long action, void *hcpu)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-06-05 12:49:58 +00:00
|
|
|
int cpu = (int)(long)hcpu;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
switch (action) {
|
|
|
|
case CPU_DOWN_PREPARE:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_DOWN_PREPARE_FROZEN:
|
2008-06-05 12:49:58 +00:00
|
|
|
disable_runtime(cpu_rq(cpu));
|
2005-04-16 22:20:36 +00:00
|
|
|
return NOTIFY_OK;
|
|
|
|
|
|
|
|
case CPU_DOWN_FAILED:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_DOWN_FAILED_FROZEN:
|
2005-04-16 22:20:36 +00:00
|
|
|
case CPU_ONLINE:
|
2007-05-09 09:35:10 +00:00
|
|
|
case CPU_ONLINE_FROZEN:
|
2008-06-05 12:49:58 +00:00
|
|
|
enable_runtime(cpu_rq(cpu));
|
2008-07-15 11:43:49 +00:00
|
|
|
return NOTIFY_OK;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
default:
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void __init sched_init_smp(void)
|
|
|
|
{
|
2008-11-24 16:05:12 +00:00
|
|
|
cpumask_var_t non_isolated_cpus;
|
|
|
|
|
|
|
|
alloc_cpumask_var(&non_isolated_cpus, GFP_KERNEL);
|
2006-10-03 08:14:04 +00:00
|
|
|
|
2008-04-05 01:11:04 +00:00
|
|
|
#if defined(CONFIG_NUMA)
|
|
|
|
sched_group_nodes_bycpu = kzalloc(nr_cpu_ids * sizeof(void **),
|
|
|
|
GFP_KERNEL);
|
|
|
|
BUG_ON(sched_group_nodes_bycpu == NULL);
|
|
|
|
#endif
|
2008-01-25 20:08:02 +00:00
|
|
|
get_online_cpus();
|
2008-04-28 09:33:07 +00:00
|
|
|
mutex_lock(&sched_domains_mutex);
|
2008-11-24 16:05:12 +00:00
|
|
|
arch_init_sched_domains(cpu_online_mask);
|
|
|
|
cpumask_andnot(non_isolated_cpus, cpu_possible_mask, cpu_isolated_map);
|
|
|
|
if (cpumask_empty(non_isolated_cpus))
|
|
|
|
cpumask_set_cpu(smp_processor_id(), non_isolated_cpus);
|
2008-04-28 09:33:07 +00:00
|
|
|
mutex_unlock(&sched_domains_mutex);
|
2008-01-25 20:08:02 +00:00
|
|
|
put_online_cpus();
|
2008-07-15 11:43:49 +00:00
|
|
|
|
|
|
|
#ifndef CONFIG_CPUSETS
|
2005-04-16 22:20:36 +00:00
|
|
|
/* XXX: Theoretical race here - CPU may be hotplugged now */
|
|
|
|
hotcpu_notifier(update_sched_domains, 0);
|
2008-07-15 11:43:49 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/* RT runtime code needs to handle some hotplug events */
|
|
|
|
hotcpu_notifier(update_runtime, 0);
|
|
|
|
|
2008-04-29 08:02:46 +00:00
|
|
|
init_hrtick();
|
2006-10-03 08:14:04 +00:00
|
|
|
|
|
|
|
/* Move init over to a non-isolated CPU */
|
2008-11-24 16:05:12 +00:00
|
|
|
if (set_cpus_allowed_ptr(current, non_isolated_cpus) < 0)
|
2006-10-03 08:14:04 +00:00
|
|
|
BUG();
|
2007-11-09 21:39:38 +00:00
|
|
|
sched_init_granularity();
|
2008-11-24 16:05:12 +00:00
|
|
|
free_cpumask_var(non_isolated_cpus);
|
2008-11-24 16:05:12 +00:00
|
|
|
|
|
|
|
alloc_cpumask_var(&fallback_doms, GFP_KERNEL);
|
2008-11-24 16:05:13 +00:00
|
|
|
init_sched_rt_class();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
void __init sched_init_smp(void)
|
|
|
|
{
|
2007-11-09 21:39:38 +00:00
|
|
|
sched_init_granularity();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
#endif /* CONFIG_SMP */
|
|
|
|
|
2009-04-16 06:45:34 +00:00
|
|
|
const_debug unsigned int sysctl_timer_migration = 1;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
int in_sched_functions(unsigned long addr)
|
|
|
|
{
|
|
|
|
return in_lock_functions(addr) ||
|
|
|
|
(addr >= (unsigned long)__sched_text_start
|
|
|
|
&& addr < (unsigned long)__sched_text_end);
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:13 +00:00
|
|
|
static void init_cfs_rq(struct cfs_rq *cfs_rq, struct rq *rq)
|
2007-07-09 16:51:59 +00:00
|
|
|
{
|
|
|
|
cfs_rq->tasks_timeline = RB_ROOT;
|
2008-04-19 17:45:00 +00:00
|
|
|
INIT_LIST_HEAD(&cfs_rq->tasks);
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
cfs_rq->rq = rq;
|
|
|
|
#endif
|
2007-10-15 15:00:10 +00:00
|
|
|
cfs_rq->min_vruntime = (u64)(-(1LL << 20));
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:29 +00:00
|
|
|
static void init_rt_rq(struct rt_rq *rt_rq, struct rq *rq)
|
|
|
|
{
|
|
|
|
struct rt_prio_array *array;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
array = &rt_rq->active;
|
|
|
|
for (i = 0; i < MAX_RT_PRIO; i++) {
|
|
|
|
INIT_LIST_HEAD(array->queue + i);
|
|
|
|
__clear_bit(i, array->bitmap);
|
|
|
|
}
|
|
|
|
/* delimiter for bitsearch: */
|
|
|
|
__set_bit(MAX_RT_PRIO, array->bitmap);
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#if defined CONFIG_SMP || defined CONFIG_RT_GROUP_SCHED
|
2008-12-29 14:39:49 +00:00
|
|
|
rt_rq->highest_prio.curr = MAX_RT_PRIO;
|
2009-01-14 14:10:04 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-12-29 14:39:49 +00:00
|
|
|
rt_rq->highest_prio.next = MAX_RT_PRIO;
|
2008-01-25 20:08:31 +00:00
|
|
|
#endif
|
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
rt_rq->rt_nr_migratory = 0;
|
|
|
|
rt_rq->overloaded = 0;
|
sched: create "pushable_tasks" list to limit pushing to one attempt
The RT scheduler employs a "push/pull" design to actively balance tasks
within the system (on a per disjoint cpuset basis). When a task is
awoken, it is immediately determined if there are any lower priority
cpus which should be preempted. This is opposed to the way normal
SCHED_OTHER tasks behave, which will wait for a periodic rebalancing
operation to occur before spreading out load.
When a particular RQ has more than 1 active RT task, it is said to
be in an "overloaded" state. Once this occurs, the system enters
the active balancing mode, where it will try to push the task away,
or persuade a different cpu to pull it over. The system will stay
in this state until the system falls back below the <= 1 queued RT
task per RQ.
However, the current implementation suffers from a limitation in the
push logic. Once overloaded, all tasks (other than current) on the
RQ are analyzed on every push operation, even if it was previously
unpushable (due to affinity, etc). Whats more, the operation stops
at the first task that is unpushable and will not look at items
lower in the queue. This causes two problems:
1) We can have the same tasks analyzed over and over again during each
push, which extends out the fast path in the scheduler for no
gain. Consider a RQ that has dozens of tasks that are bound to a
core. Each one of those tasks will be encountered and skipped
for each push operation while they are queued.
2) There may be lower-priority tasks under the unpushable task that
could have been successfully pushed, but will never be considered
until either the unpushable task is cleared, or a pull operation
succeeds. The net result is a potential latency source for mid
priority tasks.
This patch aims to rectify these two conditions by introducing a new
priority sorted list: "pushable_tasks". A task is added to the list
each time a task is activated or preempted. It is removed from the
list any time it is deactivated, made current, or fails to push.
This works because a task only needs to be attempted to push once.
After an initial failure to push, the other cpus will eventually try to
pull the task when the conditions are proper. This also solves the
problem that we don't completely analyze all tasks due to encountering
an unpushable tasks. Now every task will have a push attempted (when
appropriate).
This reduces latency both by shorting the critical section of the
rq->lock for certain workloads, and by making sure the algorithm
considers all eligible tasks in the system.
[ rostedt: added a couple more BUG_ONs ]
Signed-off-by: Gregory Haskins <ghaskins@novell.com>
Acked-by: Steven Rostedt <srostedt@redhat.com>
2008-12-29 14:39:53 +00:00
|
|
|
plist_head_init(&rq->rt.pushable_tasks, &rq->lock);
|
2008-01-25 20:08:29 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
rt_rq->rt_time = 0;
|
|
|
|
rt_rq->rt_throttled = 0;
|
2008-04-19 17:44:58 +00:00
|
|
|
rt_rq->rt_runtime = 0;
|
|
|
|
spin_lock_init(&rt_rq->rt_runtime_lock);
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-02-13 14:45:39 +00:00
|
|
|
rt_rq->rt_nr_boosted = 0;
|
2008-01-25 20:08:30 +00:00
|
|
|
rt_rq->rq = rq;
|
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
}
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2008-04-19 17:44:59 +00:00
|
|
|
static void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq,
|
|
|
|
struct sched_entity *se, int cpu, int add,
|
|
|
|
struct sched_entity *parent)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-04-19 17:44:59 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
2008-01-25 20:08:30 +00:00
|
|
|
tg->cfs_rq[cpu] = cfs_rq;
|
|
|
|
init_cfs_rq(cfs_rq, rq);
|
|
|
|
cfs_rq->tg = tg;
|
|
|
|
if (add)
|
|
|
|
list_add(&cfs_rq->leaf_cfs_rq_list, &rq->leaf_cfs_rq_list);
|
|
|
|
|
|
|
|
tg->se[cpu] = se;
|
2008-04-19 17:44:59 +00:00
|
|
|
/* se could be NULL for init_task_group */
|
|
|
|
if (!se)
|
|
|
|
return;
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
if (!parent)
|
|
|
|
se->cfs_rq = &rq->cfs;
|
|
|
|
else
|
|
|
|
se->cfs_rq = parent->my_q;
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
se->my_q = cfs_rq;
|
|
|
|
se->load.weight = tg->shares;
|
2008-05-05 21:56:17 +00:00
|
|
|
se->load.inv_weight = 0;
|
2008-04-19 17:44:59 +00:00
|
|
|
se->parent = parent;
|
2008-01-25 20:08:30 +00:00
|
|
|
}
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-04-19 17:44:59 +00:00
|
|
|
static void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq,
|
|
|
|
struct sched_rt_entity *rt_se, int cpu, int add,
|
|
|
|
struct sched_rt_entity *parent)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-04-19 17:44:59 +00:00
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
tg->rt_rq[cpu] = rt_rq;
|
|
|
|
init_rt_rq(rt_rq, rq);
|
|
|
|
rt_rq->tg = tg;
|
|
|
|
rt_rq->rt_se = rt_se;
|
2008-04-19 17:44:58 +00:00
|
|
|
rt_rq->rt_runtime = tg->rt_bandwidth.rt_runtime;
|
2008-01-25 20:08:30 +00:00
|
|
|
if (add)
|
|
|
|
list_add(&rt_rq->leaf_rt_rq_list, &rq->leaf_rt_rq_list);
|
|
|
|
|
|
|
|
tg->rt_se[cpu] = rt_se;
|
2008-04-19 17:44:59 +00:00
|
|
|
if (!rt_se)
|
|
|
|
return;
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
if (!parent)
|
|
|
|
rt_se->rt_rq = &rq->rt;
|
|
|
|
else
|
|
|
|
rt_se->rt_rq = parent->my_q;
|
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
rt_se->my_q = rt_rq;
|
2008-04-19 17:44:59 +00:00
|
|
|
rt_se->parent = parent;
|
2008-01-25 20:08:30 +00:00
|
|
|
INIT_LIST_HEAD(&rt_se->run_list);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
void __init sched_init(void)
|
|
|
|
{
|
2007-07-09 16:51:59 +00:00
|
|
|
int i, j;
|
2008-04-05 01:11:04 +00:00
|
|
|
unsigned long alloc_size = 0, ptr;
|
|
|
|
|
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
alloc_size += 2 * nr_cpu_ids * sizeof(void **);
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
alloc_size += 2 * nr_cpu_ids * sizeof(void **);
|
2008-04-19 17:45:00 +00:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
alloc_size *= 2;
|
2009-03-19 04:52:20 +00:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPUMASK_OFFSTACK
|
2009-03-19 04:52:20 +00:00
|
|
|
alloc_size += num_possible_cpus() * cpumask_size();
|
2008-04-05 01:11:04 +00:00
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* As sched_init() is called before page_alloc is setup,
|
|
|
|
* we use alloc_bootmem().
|
|
|
|
*/
|
|
|
|
if (alloc_size) {
|
2009-06-10 20:42:36 +00:00
|
|
|
ptr = (unsigned long)kzalloc(alloc_size, GFP_NOWAIT);
|
2008-04-05 01:11:04 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
init_task_group.se = (struct sched_entity **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
|
|
|
|
|
|
|
init_task_group.cfs_rq = (struct cfs_rq **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
2008-04-19 17:45:00 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
root_task_group.se = (struct sched_entity **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
|
|
|
|
|
|
|
root_task_group.cfs_rq = (struct cfs_rq **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
2008-04-05 01:11:04 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
init_task_group.rt_se = (struct sched_rt_entity **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
|
|
|
|
|
|
|
init_task_group.rt_rq = (struct rt_rq **)ptr;
|
2008-04-19 17:45:00 +00:00
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
|
|
|
|
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
root_task_group.rt_se = (struct sched_rt_entity **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
|
|
|
|
|
|
|
root_task_group.rt_rq = (struct rt_rq **)ptr;
|
|
|
|
ptr += nr_cpu_ids * sizeof(void **);
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2009-03-19 04:52:20 +00:00
|
|
|
#ifdef CONFIG_CPUMASK_OFFSTACK
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
per_cpu(load_balance_tmpmask, i) = (void *)ptr;
|
|
|
|
ptr += cpumask_size();
|
|
|
|
}
|
|
|
|
#endif /* CONFIG_CPUMASK_OFFSTACK */
|
2008-04-05 01:11:04 +00:00
|
|
|
}
|
2007-07-09 16:51:59 +00:00
|
|
|
|
2008-01-25 20:08:18 +00:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
init_defrootdomain();
|
|
|
|
#endif
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
init_rt_bandwidth(&def_rt_bandwidth,
|
|
|
|
global_rt_period(), global_rt_runtime());
|
|
|
|
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
|
|
|
init_rt_bandwidth(&init_task_group.rt_bandwidth,
|
|
|
|
global_rt_period(), global_rt_runtime());
|
2008-04-19 17:45:00 +00:00
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
init_rt_bandwidth(&root_task_group.rt_bandwidth,
|
|
|
|
global_rt_period(), RUNTIME_INF);
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2008-04-19 17:44:57 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
list_add(&init_task_group.list, &task_groups);
|
2008-04-19 17:45:00 +00:00
|
|
|
INIT_LIST_HEAD(&init_task_group.children);
|
|
|
|
|
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
INIT_LIST_HEAD(&root_task_group.children);
|
|
|
|
init_task_group.parent = &root_task_group;
|
|
|
|
list_add(&init_task_group.siblings, &root_task_group.children);
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_USER_SCHED */
|
|
|
|
#endif /* CONFIG_GROUP_SCHED */
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2006-03-28 09:56:37 +00:00
|
|
|
for_each_possible_cpu(i) {
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
rq = cpu_rq(i);
|
|
|
|
spin_lock_init(&rq->lock);
|
2005-06-25 21:57:13 +00:00
|
|
|
rq->nr_running = 0;
|
2009-04-11 08:43:41 +00:00
|
|
|
rq->calc_load_active = 0;
|
|
|
|
rq->calc_load_update = jiffies + LOAD_FREQ;
|
2007-07-09 16:51:59 +00:00
|
|
|
init_cfs_rq(&rq->cfs, rq);
|
2008-01-25 20:08:30 +00:00
|
|
|
init_rt_rq(&rq->rt, rq);
|
2007-07-09 16:51:59 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-10-15 15:00:14 +00:00
|
|
|
init_task_group.shares = init_task_group_load;
|
2008-01-25 20:08:30 +00:00
|
|
|
INIT_LIST_HEAD(&rq->leaf_cfs_rq_list);
|
2008-04-19 17:44:59 +00:00
|
|
|
#ifdef CONFIG_CGROUP_SCHED
|
|
|
|
/*
|
|
|
|
* How much cpu bandwidth does init_task_group get?
|
|
|
|
*
|
|
|
|
* In case of task-groups formed thr' the cgroup filesystem, it
|
|
|
|
* gets 100% of the cpu resources in the system. This overall
|
|
|
|
* system cpu resource is divided among the tasks of
|
|
|
|
* init_task_group and its child task-groups in a fair manner,
|
|
|
|
* based on each entity's (task or task-group's) weight
|
|
|
|
* (se->load.weight).
|
|
|
|
*
|
|
|
|
* In other words, if init_task_group has 10 tasks of weight
|
|
|
|
* 1024) and two child groups A0 and A1 (of weight 1024 each),
|
|
|
|
* then A0's share of the cpu resource is:
|
|
|
|
*
|
2009-05-04 17:13:30 +00:00
|
|
|
* A0's bandwidth = 1024 / (10*1024 + 1024 + 1024) = 8.33%
|
2008-04-19 17:44:59 +00:00
|
|
|
*
|
|
|
|
* We achieve this by letting init_task_group's tasks sit
|
|
|
|
* directly in rq->cfs (i.e init_task_group->se[] = NULL).
|
|
|
|
*/
|
2008-04-19 17:44:59 +00:00
|
|
|
init_tg_cfs_entry(&init_task_group, &rq->cfs, NULL, i, 1, NULL);
|
2008-04-19 17:44:59 +00:00
|
|
|
#elif defined CONFIG_USER_SCHED
|
2008-04-19 17:45:00 +00:00
|
|
|
root_task_group.shares = NICE_0_LOAD;
|
|
|
|
init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, 0, NULL);
|
2008-04-19 17:44:59 +00:00
|
|
|
/*
|
|
|
|
* In case of task-groups formed thr' the user id of tasks,
|
|
|
|
* init_task_group represents tasks belonging to root user.
|
|
|
|
* Hence it forms a sibling of all subsequent groups formed.
|
|
|
|
* In this case, init_task_group gets only a fraction of overall
|
|
|
|
* system cpu resource, based on the weight assigned to root
|
|
|
|
* user's cpu share (INIT_TASK_GROUP_LOAD). This is accomplished
|
|
|
|
* by letting tasks of init_task_group sit in a separate cfs_rq
|
|
|
|
* (init_cfs_rq) and having one entity represent this group of
|
|
|
|
* tasks in rq->cfs (i.e init_task_group->se[] != NULL).
|
|
|
|
*/
|
2008-04-19 17:44:59 +00:00
|
|
|
init_tg_cfs_entry(&init_task_group,
|
2008-01-25 20:08:30 +00:00
|
|
|
&per_cpu(init_cfs_rq, i),
|
2008-04-19 17:45:00 +00:00
|
|
|
&per_cpu(init_sched_entity, i), i, 1,
|
|
|
|
root_task_group.se[i]);
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2008-04-19 17:44:59 +00:00
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
|
|
|
|
|
|
|
rq->rt.rt_runtime = def_rt_bandwidth.rt_runtime;
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
INIT_LIST_HEAD(&rq->leaf_rt_rq_list);
|
2008-04-19 17:44:59 +00:00
|
|
|
#ifdef CONFIG_CGROUP_SCHED
|
2008-04-19 17:44:59 +00:00
|
|
|
init_tg_rt_entry(&init_task_group, &rq->rt, NULL, i, 1, NULL);
|
2008-04-19 17:44:59 +00:00
|
|
|
#elif defined CONFIG_USER_SCHED
|
2008-04-19 17:45:00 +00:00
|
|
|
init_tg_rt_entry(&root_task_group, &rq->rt, NULL, i, 0, NULL);
|
2008-04-19 17:44:59 +00:00
|
|
|
init_tg_rt_entry(&init_task_group,
|
2008-01-25 20:08:30 +00:00
|
|
|
&per_cpu(init_rt_rq, i),
|
2008-04-19 17:45:00 +00:00
|
|
|
&per_cpu(init_sched_rt_entity, i), i, 1,
|
|
|
|
root_task_group.rt_se[i]);
|
2008-04-19 17:44:59 +00:00
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
for (j = 0; j < CPU_LOAD_IDX_MAX; j++)
|
|
|
|
rq->cpu_load[j] = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2005-06-25 21:57:24 +00:00
|
|
|
rq->sd = NULL;
|
2008-01-25 20:08:18 +00:00
|
|
|
rq->rd = NULL;
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->active_balance = 0;
|
2007-07-09 16:51:59 +00:00
|
|
|
rq->next_balance = jiffies;
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->push_cpu = 0;
|
2006-09-26 06:30:51 +00:00
|
|
|
rq->cpu = i;
|
2008-06-04 19:04:05 +00:00
|
|
|
rq->online = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
rq->migration_thread = NULL;
|
|
|
|
INIT_LIST_HEAD(&rq->migration_queue);
|
2008-01-25 20:08:26 +00:00
|
|
|
rq_attach_root(rq, &def_root_domain);
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2008-01-25 20:08:29 +00:00
|
|
|
init_rq_hrtick(rq);
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_set(&rq->nr_iowait, 0);
|
|
|
|
}
|
|
|
|
|
[PATCH] sched: implement smpnice
Problem:
The introduction of separate run queues per CPU has brought with it "nice"
enforcement problems that are best described by a simple example.
For the sake of argument suppose that on a single CPU machine with a
nice==19 hard spinner and a nice==0 hard spinner running that the nice==0
task gets 95% of the CPU and the nice==19 task gets 5% of the CPU. Now
suppose that there is a system with 2 CPUs and 2 nice==19 hard spinners and
2 nice==0 hard spinners running. The user of this system would be entitled
to expect that the nice==0 tasks each get 95% of a CPU and the nice==19
tasks only get 5% each. However, whether this expectation is met is pretty
much down to luck as there are four equally likely distributions of the
tasks to the CPUs that the load balancing code will consider to be balanced
with loads of 2.0 for each CPU. Two of these distributions involve one
nice==0 and one nice==19 task per CPU and in these circumstances the users
expectations will be met. The other two distributions both involve both
nice==0 tasks being on one CPU and both nice==19 being on the other CPU and
each task will get 50% of a CPU and the user's expectations will not be
met.
Solution:
The solution to this problem that is implemented in the attached patch is
to use weighted loads when determining if the system is balanced and, when
an imbalance is detected, to move an amount of weighted load between run
queues (as opposed to a number of tasks) to restore the balance. Once
again, the easiest way to explain why both of these measures are necessary
is to use a simple example. Suppose that (in a slight variation of the
above example) that we have a two CPU system with 4 nice==0 and 4 nice=19
hard spinning tasks running and that the 4 nice==0 tasks are on one CPU and
the 4 nice==19 tasks are on the other CPU. The weighted loads for the two
CPUs would be 4.0 and 0.2 respectively and the load balancing code would
move 2 tasks resulting in one CPU with a load of 2.0 and the other with
load of 2.2. If this was considered to be a big enough imbalance to
justify moving a task and that task was moved using the current
move_tasks() then it would move the highest priority task that it found and
this would result in one CPU with a load of 3.0 and the other with a load
of 1.2 which would result in the movement of a task in the opposite
direction and so on -- infinite loop. If, on the other hand, an amount of
load to be moved is calculated from the imbalance (in this case 0.1) and
move_tasks() skips tasks until it find ones whose contributions to the
weighted load are less than this amount it would move two of the nice==19
tasks resulting in a system with 2 nice==0 and 2 nice=19 on each CPU with
loads of 2.1 for each CPU.
One of the advantages of this mechanism is that on a system where all tasks
have nice==0 the load balancing calculations would be mathematically
identical to the current load balancing code.
Notes:
struct task_struct:
has a new field load_weight which (in a trade off of space for speed)
stores the contribution that this task makes to a CPU's weighted load when
it is runnable.
struct runqueue:
has a new field raw_weighted_load which is the sum of the load_weight
values for the currently runnable tasks on this run queue. This field
always needs to be updated when nr_running is updated so two new inline
functions inc_nr_running() and dec_nr_running() have been created to make
sure that this happens. This also offers a convenient way to optimize away
this part of the smpnice mechanism when CONFIG_SMP is not defined.
int try_to_wake_up():
in this function the value SCHED_LOAD_BALANCE is used to represent the load
contribution of a single task in various calculations in the code that
decides which CPU to put the waking task on. While this would be a valid
on a system where the nice values for the runnable tasks were distributed
evenly around zero it will lead to anomalous load balancing if the
distribution is skewed in either direction. To overcome this problem
SCHED_LOAD_SCALE has been replaced by the load_weight for the relevant task
or by the average load_weight per task for the queue in question (as
appropriate).
int move_tasks():
The modifications to this function were complicated by the fact that
active_load_balance() uses it to move exactly one task without checking
whether an imbalance actually exists. This precluded the simple
overloading of max_nr_move with max_load_move and necessitated the addition
of the latter as an extra argument to the function. The internal
implementation is then modified to move up to max_nr_move tasks and
max_load_move of weighted load. This slightly complicates the code where
move_tasks() is called and if ever active_load_balance() is changed to not
use move_tasks() the implementation of move_tasks() should be simplified
accordingly.
struct sched_group *find_busiest_group():
Similar to try_to_wake_up(), there are places in this function where
SCHED_LOAD_SCALE is used to represent the load contribution of a single
task and the same issues are created. A similar solution is adopted except
that it is now the average per task contribution to a group's load (as
opposed to a run queue) that is required. As this value is not directly
available from the group it is calculated on the fly as the queues in the
groups are visited when determining the busiest group.
A key change to this function is that it is no longer to scale down
*imbalance on exit as move_tasks() uses the load in its scaled form.
void set_user_nice():
has been modified to update the task's load_weight field when it's nice
value and also to ensure that its run queue's raw_weighted_load field is
updated if it was runnable.
From: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
With smpnice, sched groups with highest priority tasks can mask the imbalance
between the other sched groups with in the same domain. This patch fixes some
of the listed down scenarios by not considering the sched groups which are
lightly loaded.
a) on a simple 4-way MP system, if we have one high priority and 4 normal
priority tasks, with smpnice we would like to see the high priority task
scheduled on one cpu, two other cpus getting one normal task each and the
fourth cpu getting the remaining two normal tasks. but with current
smpnice extra normal priority task keeps jumping from one cpu to another
cpu having the normal priority task. This is because of the
busiest_has_loaded_cpus, nr_loaded_cpus logic.. We are not including the
cpu with high priority task in max_load calculations but including that in
total and avg_load calcuations.. leading to max_load < avg_load and load
balance between cpus running normal priority tasks(2 Vs 1) will always show
imbalanace as one normal priority and the extra normal priority task will
keep moving from one cpu to another cpu having normal priority task..
b) 4-way system with HT (8 logical processors). Package-P0 T0 has a
highest priority task, T1 is idle. Package-P1 Both T0 and T1 have 1 normal
priority task each.. P2 and P3 are idle. With this patch, one of the
normal priority tasks on P1 will be moved to P2 or P3..
c) With the current weighted smp nice calculations, it doesn't always make
sense to look at the highest weighted runqueue in the busy group..
Consider a load balance scenario on a DP with HT system, with Package-0
containing one high priority and one low priority, Package-1 containing one
low priority(with other thread being idle).. Package-1 thinks that it need
to take the low priority thread from Package-0. And find_busiest_queue()
returns the cpu thread with highest priority task.. And ultimately(with
help of active load balance) we move high priority task to Package-1. And
same continues with Package-0 now, moving high priority task from package-1
to package-0.. Even without the presence of active load balance, load
balance will fail to balance the above scenario.. Fix find_busiest_queue
to use "imbalance" when it is lightly loaded.
[kernel@kolivas.org: sched: store weighted load on up]
[kernel@kolivas.org: sched: add discrete weighted cpu load function]
[suresh.b.siddha@intel.com: sched: remove dead code]
Signed-off-by: Peter Williams <pwil3058@bigpond.com.au>
Cc: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: "Chen, Kenneth W" <kenneth.w.chen@intel.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Con Kolivas <kernel@kolivas.org>
Cc: John Hawkes <hawkes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-06-27 09:54:34 +00:00
|
|
|
set_load_weight(&init_task);
|
2006-07-30 10:03:52 +00:00
|
|
|
|
2007-07-26 11:40:43 +00:00
|
|
|
#ifdef CONFIG_PREEMPT_NOTIFIERS
|
|
|
|
INIT_HLIST_HEAD(&init_task.preempt_notifiers);
|
|
|
|
#endif
|
|
|
|
|
2006-12-10 10:20:25 +00:00
|
|
|
#ifdef CONFIG_SMP
|
Remove argument from open_softirq which is always NULL
As git-grep shows, open_softirq() is always called with the last argument
being NULL
block/blk-core.c: open_softirq(BLOCK_SOFTIRQ, blk_done_softirq, NULL);
kernel/hrtimer.c: open_softirq(HRTIMER_SOFTIRQ, run_hrtimer_softirq, NULL);
kernel/rcuclassic.c: open_softirq(RCU_SOFTIRQ, rcu_process_callbacks, NULL);
kernel/rcupreempt.c: open_softirq(RCU_SOFTIRQ, rcu_process_callbacks, NULL);
kernel/sched.c: open_softirq(SCHED_SOFTIRQ, run_rebalance_domains, NULL);
kernel/softirq.c: open_softirq(TASKLET_SOFTIRQ, tasklet_action, NULL);
kernel/softirq.c: open_softirq(HI_SOFTIRQ, tasklet_hi_action, NULL);
kernel/timer.c: open_softirq(TIMER_SOFTIRQ, run_timer_softirq, NULL);
net/core/dev.c: open_softirq(NET_TX_SOFTIRQ, net_tx_action, NULL);
net/core/dev.c: open_softirq(NET_RX_SOFTIRQ, net_rx_action, NULL);
This observation has already been made by Matthew Wilcox in June 2002
(http://www.cs.helsinki.fi/linux/linux-kernel/2002-25/0687.html)
"I notice that none of the current softirq routines use the data element
passed to them."
and the situation hasn't changed since them. So it appears we can safely
remove that extra argument to save 128 (54) bytes of kernel data (text).
Signed-off-by: Carlos R. Mafra <crmafra@ift.unesp.br>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
2008-05-15 14:15:37 +00:00
|
|
|
open_softirq(SCHED_SOFTIRQ, run_rebalance_domains);
|
2006-12-10 10:20:25 +00:00
|
|
|
#endif
|
|
|
|
|
2006-07-30 10:03:52 +00:00
|
|
|
#ifdef CONFIG_RT_MUTEXES
|
|
|
|
plist_head_init(&init_task.pi_waiters, &init_task.pi_lock);
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* The boot idle thread does lazy MMU switching as well:
|
|
|
|
*/
|
|
|
|
atomic_inc(&init_mm.mm_count);
|
|
|
|
enter_lazy_tlb(&init_mm, current);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Make us the idle thread. Technically, schedule() should not be
|
|
|
|
* called from this thread, however somewhere below it might be,
|
|
|
|
* but because we are the idle thread, we just pick up running again
|
|
|
|
* when this runqueue becomes "idle".
|
|
|
|
*/
|
|
|
|
init_idle(current, smp_processor_id());
|
2009-04-11 08:43:41 +00:00
|
|
|
|
|
|
|
calc_load_update = jiffies + LOAD_FREQ;
|
|
|
|
|
2007-07-09 16:51:59 +00:00
|
|
|
/*
|
|
|
|
* During early bootup we pretend to be a normal task:
|
|
|
|
*/
|
|
|
|
current->sched_class = &fair_sched_class;
|
2008-02-13 13:02:36 +00:00
|
|
|
|
2008-11-24 16:05:04 +00:00
|
|
|
/* Allocate the nohz_cpu_mask if CONFIG_CPUMASK_OFFSTACK */
|
2009-06-11 05:35:27 +00:00
|
|
|
alloc_cpumask_var(&nohz_cpu_mask, GFP_NOWAIT);
|
2008-11-24 23:27:51 +00:00
|
|
|
#ifdef CONFIG_SMP
|
2008-11-24 16:05:09 +00:00
|
|
|
#ifdef CONFIG_NO_HZ
|
2009-06-11 05:35:27 +00:00
|
|
|
alloc_cpumask_var(&nohz.cpu_mask, GFP_NOWAIT);
|
|
|
|
alloc_cpumask_var(&nohz.ilb_grp_nohz_mask, GFP_NOWAIT);
|
2008-11-24 16:05:09 +00:00
|
|
|
#endif
|
2009-06-11 05:35:27 +00:00
|
|
|
alloc_cpumask_var(&cpu_isolated_map, GFP_NOWAIT);
|
2008-11-24 23:27:51 +00:00
|
|
|
#endif /* SMP */
|
2008-11-24 16:05:04 +00:00
|
|
|
|
2009-05-04 17:13:30 +00:00
|
|
|
perf_counter_init();
|
|
|
|
|
2008-02-13 13:02:36 +00:00
|
|
|
scheduler_running = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_DEBUG_SPINLOCK_SLEEP
|
|
|
|
void __might_sleep(char *file, int line)
|
|
|
|
{
|
2006-07-03 07:25:40 +00:00
|
|
|
#ifdef in_atomic
|
2005-04-16 22:20:36 +00:00
|
|
|
static unsigned long prev_jiffy; /* ratelimiting */
|
|
|
|
|
2008-08-28 09:34:43 +00:00
|
|
|
if ((!in_atomic() && !irqs_disabled()) ||
|
|
|
|
system_state != SYSTEM_RUNNING || oops_in_progress)
|
|
|
|
return;
|
|
|
|
if (time_before(jiffies, prev_jiffy + HZ) && prev_jiffy)
|
|
|
|
return;
|
|
|
|
prev_jiffy = jiffies;
|
|
|
|
|
|
|
|
printk(KERN_ERR
|
|
|
|
"BUG: sleeping function called from invalid context at %s:%d\n",
|
|
|
|
file, line);
|
|
|
|
printk(KERN_ERR
|
|
|
|
"in_atomic(): %d, irqs_disabled(): %d, pid: %d, name: %s\n",
|
|
|
|
in_atomic(), irqs_disabled(),
|
|
|
|
current->pid, current->comm);
|
|
|
|
|
|
|
|
debug_show_held_locks(current);
|
|
|
|
if (irqs_disabled())
|
|
|
|
print_irqtrace_events(current);
|
|
|
|
dump_stack();
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__might_sleep);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_MAGIC_SYSRQ
|
2007-10-15 15:00:15 +00:00
|
|
|
static void normalize_task(struct rq *rq, struct task_struct *p)
|
|
|
|
{
|
|
|
|
int on_rq;
|
2008-05-03 16:29:28 +00:00
|
|
|
|
2007-10-15 15:00:15 +00:00
|
|
|
update_rq_clock(rq);
|
|
|
|
on_rq = p->se.on_rq;
|
|
|
|
if (on_rq)
|
|
|
|
deactivate_task(rq, p, 0);
|
|
|
|
__setscheduler(rq, p, SCHED_NORMAL, 0);
|
|
|
|
if (on_rq) {
|
|
|
|
activate_task(rq, p, 0);
|
|
|
|
resched_task(rq->curr);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
void normalize_rt_tasks(void)
|
|
|
|
{
|
2007-06-17 16:37:45 +00:00
|
|
|
struct task_struct *g, *p;
|
2005-04-16 22:20:36 +00:00
|
|
|
unsigned long flags;
|
2006-07-03 07:25:42 +00:00
|
|
|
struct rq *rq;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
read_lock_irqsave(&tasklist_lock, flags);
|
2007-06-17 16:37:45 +00:00
|
|
|
do_each_thread(g, p) {
|
2007-10-15 15:00:18 +00:00
|
|
|
/*
|
|
|
|
* Only normalize user tasks:
|
|
|
|
*/
|
|
|
|
if (!p->mm)
|
|
|
|
continue;
|
|
|
|
|
2007-08-02 15:41:40 +00:00
|
|
|
p->se.exec_start = 0;
|
|
|
|
#ifdef CONFIG_SCHEDSTATS
|
2007-07-09 16:51:59 +00:00
|
|
|
p->se.wait_start = 0;
|
|
|
|
p->se.sleep_start = 0;
|
|
|
|
p->se.block_start = 0;
|
2007-08-02 15:41:40 +00:00
|
|
|
#endif
|
2007-07-09 16:51:59 +00:00
|
|
|
|
|
|
|
if (!rt_task(p)) {
|
|
|
|
/*
|
|
|
|
* Renice negative nice level userspace
|
|
|
|
* tasks back to 0:
|
|
|
|
*/
|
|
|
|
if (TASK_NICE(p) < 0 && p->mm)
|
|
|
|
set_user_nice(p, 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
continue;
|
2007-07-09 16:51:59 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_lock(&p->pi_lock);
|
2006-06-27 09:54:51 +00:00
|
|
|
rq = __task_rq_lock(p);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-15 15:00:18 +00:00
|
|
|
normalize_task(rq, p);
|
2007-10-15 15:00:15 +00:00
|
|
|
|
2006-06-27 09:54:51 +00:00
|
|
|
__task_rq_unlock(rq);
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_unlock(&p->pi_lock);
|
2007-06-17 16:37:45 +00:00
|
|
|
} while_each_thread(g, p);
|
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
read_unlock_irqrestore(&tasklist_lock, flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_MAGIC_SYSRQ */
|
2005-09-12 14:59:21 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_IA64
|
|
|
|
/*
|
|
|
|
* These functions are only useful for the IA64 MCA handling.
|
|
|
|
*
|
|
|
|
* They can only be called when the whole system has been
|
|
|
|
* stopped - every CPU needs to be quiescent, and no scheduling
|
|
|
|
* activity can take place. Using them for anything else would
|
|
|
|
* be a serious bug, and as a result, they aren't even visible
|
|
|
|
* under any other configuration.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* curr_task - return the current task for a given cpu.
|
|
|
|
* @cpu: the processor in question.
|
|
|
|
*
|
|
|
|
* ONLY VALID WHEN THE WHOLE SYSTEM IS STOPPED!
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
struct task_struct *curr_task(int cpu)
|
2005-09-12 14:59:21 +00:00
|
|
|
{
|
|
|
|
return cpu_curr(cpu);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* set_curr_task - set the current task for a given cpu.
|
|
|
|
* @cpu: the processor in question.
|
|
|
|
* @p: the task pointer to set.
|
|
|
|
*
|
|
|
|
* Description: This function must only be used when non-maskable interrupts
|
2007-12-05 14:46:09 +00:00
|
|
|
* are serviced on a separate stack. It allows the architecture to switch the
|
|
|
|
* notion of the current task on a cpu in a non-blocking manner. This function
|
2005-09-12 14:59:21 +00:00
|
|
|
* must be called with all CPU's synchronized, and interrupts disabled, the
|
|
|
|
* and caller must save the original value of the current task (see
|
|
|
|
* curr_task() above) and restore that value before reenabling interrupts and
|
|
|
|
* re-starting the system.
|
|
|
|
*
|
|
|
|
* ONLY VALID WHEN THE WHOLE SYSTEM IS STOPPED!
|
|
|
|
*/
|
2006-07-03 07:25:41 +00:00
|
|
|
void set_curr_task(int cpu, struct task_struct *p)
|
2005-09-12 14:59:21 +00:00
|
|
|
{
|
|
|
|
cpu_curr(cpu) = p;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
static void free_fair_sched_group(struct task_group *tg)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
if (tg->cfs_rq)
|
|
|
|
kfree(tg->cfs_rq[i]);
|
|
|
|
if (tg->se)
|
|
|
|
kfree(tg->se[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
kfree(tg->cfs_rq);
|
|
|
|
kfree(tg->se);
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
static
|
|
|
|
int alloc_fair_sched_group(struct task_group *tg, struct task_group *parent)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
|
|
|
struct cfs_rq *cfs_rq;
|
2008-10-29 09:03:22 +00:00
|
|
|
struct sched_entity *se;
|
2007-10-15 15:00:09 +00:00
|
|
|
struct rq *rq;
|
2007-10-15 15:00:07 +00:00
|
|
|
int i;
|
|
|
|
|
2008-04-05 01:11:04 +00:00
|
|
|
tg->cfs_rq = kzalloc(sizeof(cfs_rq) * nr_cpu_ids, GFP_KERNEL);
|
2007-10-15 15:00:07 +00:00
|
|
|
if (!tg->cfs_rq)
|
|
|
|
goto err;
|
2008-04-05 01:11:04 +00:00
|
|
|
tg->se = kzalloc(sizeof(se) * nr_cpu_ids, GFP_KERNEL);
|
2007-10-15 15:00:07 +00:00
|
|
|
if (!tg->se)
|
|
|
|
goto err;
|
2008-02-13 14:45:40 +00:00
|
|
|
|
|
|
|
tg->shares = NICE_0_LOAD;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
2007-10-15 15:00:09 +00:00
|
|
|
rq = cpu_rq(i);
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
cfs_rq = kzalloc_node(sizeof(struct cfs_rq),
|
|
|
|
GFP_KERNEL, cpu_to_node(i));
|
2007-10-15 15:00:07 +00:00
|
|
|
if (!cfs_rq)
|
|
|
|
goto err;
|
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
se = kzalloc_node(sizeof(struct sched_entity),
|
|
|
|
GFP_KERNEL, cpu_to_node(i));
|
2007-10-15 15:00:07 +00:00
|
|
|
if (!se)
|
|
|
|
goto err;
|
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
init_tg_cfs_entry(tg, cfs_rq, se, i, 0, parent->se[i]);
|
2008-02-13 14:45:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
err:
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void register_fair_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
list_add_rcu(&tg->cfs_rq[cpu]->leaf_cfs_rq_list,
|
|
|
|
&cpu_rq(cpu)->leaf_cfs_rq_list);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void unregister_fair_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
list_del_rcu(&tg->cfs_rq[cpu]->leaf_cfs_rq_list);
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFG_FAIR_GROUP_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
static inline void free_fair_sched_group(struct task_group *tg)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
static inline
|
|
|
|
int alloc_fair_sched_group(struct task_group *tg, struct task_group *parent)
|
2008-02-13 14:45:40 +00:00
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void register_fair_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void unregister_fair_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-02-13 14:45:40 +00:00
|
|
|
static void free_rt_sched_group(struct task_group *tg)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
destroy_rt_bandwidth(&tg->rt_bandwidth);
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
if (tg->rt_rq)
|
|
|
|
kfree(tg->rt_rq[i]);
|
|
|
|
if (tg->rt_se)
|
|
|
|
kfree(tg->rt_se[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
kfree(tg->rt_rq);
|
|
|
|
kfree(tg->rt_se);
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
static
|
|
|
|
int alloc_rt_sched_group(struct task_group *tg, struct task_group *parent)
|
2008-02-13 14:45:40 +00:00
|
|
|
{
|
|
|
|
struct rt_rq *rt_rq;
|
2008-10-29 09:03:22 +00:00
|
|
|
struct sched_rt_entity *rt_se;
|
2008-02-13 14:45:40 +00:00
|
|
|
struct rq *rq;
|
|
|
|
int i;
|
|
|
|
|
2008-04-05 01:11:04 +00:00
|
|
|
tg->rt_rq = kzalloc(sizeof(rt_rq) * nr_cpu_ids, GFP_KERNEL);
|
2008-02-13 14:45:40 +00:00
|
|
|
if (!tg->rt_rq)
|
|
|
|
goto err;
|
2008-04-05 01:11:04 +00:00
|
|
|
tg->rt_se = kzalloc(sizeof(rt_se) * nr_cpu_ids, GFP_KERNEL);
|
2008-02-13 14:45:40 +00:00
|
|
|
if (!tg->rt_se)
|
|
|
|
goto err;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
init_rt_bandwidth(&tg->rt_bandwidth,
|
|
|
|
ktime_to_ns(def_rt_bandwidth.rt_period), 0);
|
2008-02-13 14:45:40 +00:00
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
rq = cpu_rq(i);
|
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
rt_rq = kzalloc_node(sizeof(struct rt_rq),
|
|
|
|
GFP_KERNEL, cpu_to_node(i));
|
2008-01-25 20:08:30 +00:00
|
|
|
if (!rt_rq)
|
|
|
|
goto err;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
rt_se = kzalloc_node(sizeof(struct sched_rt_entity),
|
|
|
|
GFP_KERNEL, cpu_to_node(i));
|
2008-01-25 20:08:30 +00:00
|
|
|
if (!rt_se)
|
|
|
|
goto err;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-10-29 09:03:22 +00:00
|
|
|
init_tg_rt_entry(tg, rt_rq, rt_se, i, 0, parent->rt_se[i]);
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
err:
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void register_rt_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
list_add_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list,
|
|
|
|
&cpu_rq(cpu)->leaf_rt_rq_list);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void unregister_rt_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
list_del_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list);
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_RT_GROUP_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
static inline void free_rt_sched_group(struct task_group *tg)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
static inline
|
|
|
|
int alloc_rt_sched_group(struct task_group *tg, struct task_group *parent)
|
2008-02-13 14:45:40 +00:00
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void register_rt_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void unregister_rt_sched_group(struct task_group *tg, int cpu)
|
|
|
|
{
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2008-02-13 14:45:40 +00:00
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
#ifdef CONFIG_GROUP_SCHED
|
2008-02-13 14:45:40 +00:00
|
|
|
static void free_sched_group(struct task_group *tg)
|
|
|
|
{
|
|
|
|
free_fair_sched_group(tg);
|
|
|
|
free_rt_sched_group(tg);
|
|
|
|
kfree(tg);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* allocate runqueue etc for a new task group */
|
2008-04-19 17:44:59 +00:00
|
|
|
struct task_group *sched_create_group(struct task_group *parent)
|
2008-02-13 14:45:40 +00:00
|
|
|
{
|
|
|
|
struct task_group *tg;
|
|
|
|
unsigned long flags;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
tg = kzalloc(sizeof(*tg), GFP_KERNEL);
|
|
|
|
if (!tg)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
if (!alloc_fair_sched_group(tg, parent))
|
2008-02-13 14:45:40 +00:00
|
|
|
goto err;
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
if (!alloc_rt_sched_group(tg, parent))
|
2008-02-13 14:45:40 +00:00
|
|
|
goto err;
|
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_lock_irqsave(&task_group_lock, flags);
|
2007-10-15 15:00:09 +00:00
|
|
|
for_each_possible_cpu(i) {
|
2008-02-13 14:45:40 +00:00
|
|
|
register_fair_sched_group(tg, i);
|
|
|
|
register_rt_sched_group(tg, i);
|
2007-10-15 15:00:09 +00:00
|
|
|
}
|
2008-01-25 20:08:30 +00:00
|
|
|
list_add_rcu(&tg->list, &task_groups);
|
2008-04-19 17:45:00 +00:00
|
|
|
|
|
|
|
WARN_ON(!parent); /* root should already exist */
|
|
|
|
|
|
|
|
tg->parent = parent;
|
|
|
|
INIT_LIST_HEAD(&tg->children);
|
2030-08-14 07:56:40 +00:00
|
|
|
list_add_rcu(&tg->siblings, &parent->children);
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_unlock_irqrestore(&task_group_lock, flags);
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2007-10-15 15:00:09 +00:00
|
|
|
return tg;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
|
|
|
err:
|
2008-01-25 20:08:30 +00:00
|
|
|
free_sched_group(tg);
|
2007-10-15 15:00:07 +00:00
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:09 +00:00
|
|
|
/* rcu callback to free various structures associated with a task group */
|
2008-01-25 20:08:30 +00:00
|
|
|
static void free_sched_group_rcu(struct rcu_head *rhp)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
|
|
|
/* now it should be safe to free those cfs_rqs */
|
2008-01-25 20:08:30 +00:00
|
|
|
free_sched_group(container_of(rhp, struct task_group, rcu));
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:09 +00:00
|
|
|
/* Destroy runqueue etc associated with a task group */
|
2007-10-15 15:00:14 +00:00
|
|
|
void sched_destroy_group(struct task_group *tg)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
2008-02-13 14:45:39 +00:00
|
|
|
unsigned long flags;
|
2007-10-15 15:00:09 +00:00
|
|
|
int i;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_lock_irqsave(&task_group_lock, flags);
|
2007-10-15 15:00:09 +00:00
|
|
|
for_each_possible_cpu(i) {
|
2008-02-13 14:45:40 +00:00
|
|
|
unregister_fair_sched_group(tg, i);
|
|
|
|
unregister_rt_sched_group(tg, i);
|
2007-10-15 15:00:09 +00:00
|
|
|
}
|
2008-01-25 20:08:30 +00:00
|
|
|
list_del_rcu(&tg->list);
|
2008-04-19 17:45:00 +00:00
|
|
|
list_del_rcu(&tg->siblings);
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_unlock_irqrestore(&task_group_lock, flags);
|
2007-10-15 15:00:09 +00:00
|
|
|
|
|
|
|
/* wait for possible concurrent references to cfs_rqs complete */
|
2008-01-25 20:08:30 +00:00
|
|
|
call_rcu(&tg->rcu, free_sched_group_rcu);
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:09 +00:00
|
|
|
/* change task's runqueue when it moves between groups.
|
2007-10-15 15:00:12 +00:00
|
|
|
* The caller of this function should have put the task in its new group
|
|
|
|
* by now. This function just updates tsk->se.cfs_rq and tsk->se.parent to
|
|
|
|
* reflect its new group.
|
2007-10-15 15:00:09 +00:00
|
|
|
*/
|
|
|
|
void sched_move_task(struct task_struct *tsk)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
|
|
|
int on_rq, running;
|
|
|
|
unsigned long flags;
|
|
|
|
struct rq *rq;
|
|
|
|
|
|
|
|
rq = task_rq_lock(tsk, &flags);
|
|
|
|
|
|
|
|
update_rq_clock(rq);
|
|
|
|
|
2007-12-18 14:21:13 +00:00
|
|
|
running = task_current(rq, tsk);
|
2007-10-15 15:00:07 +00:00
|
|
|
on_rq = tsk->se.on_rq;
|
|
|
|
|
2008-03-10 18:01:20 +00:00
|
|
|
if (on_rq)
|
2007-10-15 15:00:07 +00:00
|
|
|
dequeue_task(rq, tsk, 0);
|
2008-03-10 18:01:20 +00:00
|
|
|
if (unlikely(running))
|
|
|
|
tsk->sched_class->put_prev_task(rq, tsk);
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-01-25 20:08:30 +00:00
|
|
|
set_task_rq(tsk, task_cpu(tsk));
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-29 20:21:01 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
|
if (tsk->sched_class->moved_group)
|
|
|
|
tsk->sched_class->moved_group(tsk);
|
|
|
|
#endif
|
|
|
|
|
2008-03-10 18:01:20 +00:00
|
|
|
if (unlikely(running))
|
|
|
|
tsk->sched_class->set_curr_task(rq);
|
|
|
|
if (on_rq)
|
2007-10-15 15:00:07 +00:00
|
|
|
enqueue_task(rq, tsk, 0);
|
2007-10-15 15:00:07 +00:00
|
|
|
|
|
|
|
task_rq_unlock(rq, &flags);
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_GROUP_SCHED */
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2008-06-27 11:41:14 +00:00
|
|
|
static void __set_se_shares(struct sched_entity *se, unsigned long shares)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
|
|
|
struct cfs_rq *cfs_rq = se->cfs_rq;
|
|
|
|
int on_rq;
|
|
|
|
|
|
|
|
on_rq = se->on_rq;
|
2008-02-25 16:34:02 +00:00
|
|
|
if (on_rq)
|
2007-10-15 15:00:07 +00:00
|
|
|
dequeue_entity(cfs_rq, se, 0);
|
|
|
|
|
|
|
|
se->load.weight = shares;
|
2008-05-05 21:56:17 +00:00
|
|
|
se->load.inv_weight = 0;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-25 16:34:02 +00:00
|
|
|
if (on_rq)
|
2007-10-15 15:00:07 +00:00
|
|
|
enqueue_entity(cfs_rq, se, 0);
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
2008-02-25 16:34:02 +00:00
|
|
|
|
2008-06-27 11:41:14 +00:00
|
|
|
static void set_se_shares(struct sched_entity *se, unsigned long shares)
|
|
|
|
{
|
|
|
|
struct cfs_rq *cfs_rq = se->cfs_rq;
|
|
|
|
struct rq *rq = cfs_rq->rq;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rq->lock, flags);
|
|
|
|
__set_se_shares(se, shares);
|
|
|
|
spin_unlock_irqrestore(&rq->lock, flags);
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
static DEFINE_MUTEX(shares_mutex);
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
int sched_group_set_shares(struct task_group *tg, unsigned long shares)
|
2007-10-15 15:00:07 +00:00
|
|
|
{
|
|
|
|
int i;
|
2008-02-13 14:45:39 +00:00
|
|
|
unsigned long flags;
|
2008-01-22 10:24:58 +00:00
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
/*
|
|
|
|
* We can't change the weight of the root cgroup.
|
|
|
|
*/
|
|
|
|
if (!tg->se[0])
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-04-19 17:45:00 +00:00
|
|
|
if (shares < MIN_SHARES)
|
|
|
|
shares = MIN_SHARES;
|
sched: fair-group: fix a Div0 error of the fair group scheduler
When I echoed 0 into the "cpu.shares" file, a Div0 error occured.
We found it is caused by the following calling.
sched_group_set_shares(tg, shares)
set_se_shares(tg->se[i], shares/nr_cpu_ids)
__set_se_shares(se, shares)
div64_64((1ULL<<32), shares)
When the echoed value was less than the number of processores, the result of the
sentence "shares/nr_cpu_ids" was 0, and then the system called div64() to divide
the result, the Div0 error occured.
It is unnecessary that the shares value is divided by nr_cpu_ids, I think.
Because in the function __update_group_shares_cpu() and init_tg_cfs_entry(),
the shares value isn't divided by nr_cpu_ids when setting shares of the sched
entity.
This patch fixes this bug. And echoing ULONG_MAX value into cpu.shares also
causes Div0 error, so we set a macro MAX_SHARES to limit the max value of
shares.
Signed-off-by: Miao Xie <miaox@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-04-28 04:54:56 +00:00
|
|
|
else if (shares > MAX_SHARES)
|
|
|
|
shares = MAX_SHARES;
|
2008-02-25 16:34:02 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
mutex_lock(&shares_mutex);
|
2007-10-15 15:00:09 +00:00
|
|
|
if (tg->shares == shares)
|
2007-10-15 15:00:14 +00:00
|
|
|
goto done;
|
2007-10-15 15:00:07 +00:00
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_lock_irqsave(&task_group_lock, flags);
|
2008-02-13 14:45:40 +00:00
|
|
|
for_each_possible_cpu(i)
|
|
|
|
unregister_fair_sched_group(tg, i);
|
2008-04-19 17:45:00 +00:00
|
|
|
list_del_rcu(&tg->siblings);
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_unlock_irqrestore(&task_group_lock, flags);
|
sched: group scheduler, fix fairness of cpu bandwidth allocation for task groups
The current load balancing scheme isn't good enough for precise
group fairness.
For example: on a 8-cpu system, I created 3 groups as under:
a = 8 tasks (cpu.shares = 1024)
b = 4 tasks (cpu.shares = 1024)
c = 3 tasks (cpu.shares = 1024)
a, b and c are task groups that have equal weight. We would expect each
of the groups to receive 33.33% of cpu bandwidth under a fair scheduler.
This is what I get with the latest scheduler git tree:
Signed-off-by: Ingo Molnar <mingo@elte.hu>
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 277.676 | 57.8% | 54.1% 54.1% 54.1% 54.2% 56.7% 62.2% 62.8% 64.5%
b | 116.108 | 24.2% | 47.4% 48.1% 48.7% 49.3%
c | 86.326 | 18.0% | 47.5% 47.9% 48.5%
--------------------------------------------------------------------------------
Explanation of o/p:
Col1 -> Group name
Col2 -> Cumulative execution time (in seconds) received by all tasks of that
group in a 60sec window across 8 cpus
Col3 -> CPU bandwidth received by the group in the 60sec window, expressed in
percentage. Col3 data is derived as:
Col3 = 100 * Col2 / (NR_CPUS * 60)
Col4 -> CPU bandwidth received by each individual task of the group.
Col4 = 100 * cpu_time_recd_by_task / 60
[I can share the test case that produces a similar o/p if reqd]
The deviation from desired group fairness is as below:
a = +24.47%
b = -9.13%
c = -15.33%
which is quite high.
After the patch below is applied, here are the results:
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 163.112 | 34.0% | 33.2% 33.4% 33.5% 33.5% 33.7% 34.4% 34.8% 35.3%
b | 156.220 | 32.5% | 63.3% 64.5% 66.1% 66.5%
c | 160.653 | 33.5% | 85.8% 90.6% 91.4%
--------------------------------------------------------------------------------
Deviation from desired group fairness is as below:
a = +0.67%
b = -0.83%
c = +0.17%
which is far better IMO. Most of other runs have yielded a deviation within
+-2% at the most, which is good.
Why do we see bad (group) fairness with current scheuler?
=========================================================
Currently cpu's weight is just the summation of individual task weights.
This can yield incorrect results. For ex: consider three groups as below
on a 2-cpu system:
CPU0 CPU1
---------------------------
A (10) B(5)
C(5)
---------------------------
Group A has 10 tasks, all on CPU0, Group B and C have 5 tasks each all
of which are on CPU1. Each task has the same weight (NICE_0_LOAD =
1024).
The current scheme would yield a cpu weight of 10240 (10*1024) for each cpu and
the load balancer will think both CPUs are perfectly balanced and won't
move around any tasks. This, however, would yield this bandwidth:
A = 50%
B = 25%
C = 25%
which is not the desired result.
What's changing in the patch?
=============================
- How cpu weights are calculated when CONFIF_FAIR_GROUP_SCHED is
defined (see below)
- API Change
- Two tunables introduced in sysfs (under SCHED_DEBUG) to
control the frequency at which the load balance monitor
thread runs.
The basic change made in this patch is how cpu weight (rq->load.weight) is
calculated. Its now calculated as the summation of group weights on a cpu,
rather than summation of task weights. Weight exerted by a group on a
cpu is dependent on the shares allocated to it and also the number of
tasks the group has on that cpu compared to the total number of
(runnable) tasks the group has in the system.
Let,
W(K,i) = Weight of group K on cpu i
T(K,i) = Task load present in group K's cfs_rq on cpu i
T(K) = Total task load of group K across various cpus
S(K) = Shares allocated to group K
NRCPUS = Number of online cpus in the scheduler domain to
which group K is assigned.
Then,
W(K,i) = S(K) * NRCPUS * T(K,i) / T(K)
A load balance monitor thread is created at bootup, which periodically
runs and adjusts group's weight on each cpu. To avoid its overhead, two
min/max tunables are introduced (under SCHED_DEBUG) to control the rate
at which it runs.
Fixes from: Peter Zijlstra <a.p.zijlstra@chello.nl>
- don't start the load_balance_monitor when there is only a single cpu.
- rename the kthread because its currently longer than TASK_COMM_LEN
Signed-off-by: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-01-25 20:08:00 +00:00
|
|
|
|
|
|
|
/* wait for any ongoing reference to this group to finish */
|
|
|
|
synchronize_sched();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Now we are free to modify the group's share on each cpu
|
|
|
|
* w/o tripping rebalance_share or load_balance_fair.
|
|
|
|
*/
|
2007-10-15 15:00:09 +00:00
|
|
|
tg->shares = shares;
|
2008-06-27 11:41:14 +00:00
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
/*
|
|
|
|
* force a rebalance
|
|
|
|
*/
|
|
|
|
cfs_rq_set_shares(tg->cfs_rq[i], 0);
|
sched: fair-group: fix a Div0 error of the fair group scheduler
When I echoed 0 into the "cpu.shares" file, a Div0 error occured.
We found it is caused by the following calling.
sched_group_set_shares(tg, shares)
set_se_shares(tg->se[i], shares/nr_cpu_ids)
__set_se_shares(se, shares)
div64_64((1ULL<<32), shares)
When the echoed value was less than the number of processores, the result of the
sentence "shares/nr_cpu_ids" was 0, and then the system called div64() to divide
the result, the Div0 error occured.
It is unnecessary that the shares value is divided by nr_cpu_ids, I think.
Because in the function __update_group_shares_cpu() and init_tg_cfs_entry(),
the shares value isn't divided by nr_cpu_ids when setting shares of the sched
entity.
This patch fixes this bug. And echoing ULONG_MAX value into cpu.shares also
causes Div0 error, so we set a macro MAX_SHARES to limit the max value of
shares.
Signed-off-by: Miao Xie <miaox@cn.fujitsu.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-04-28 04:54:56 +00:00
|
|
|
set_se_shares(tg->se[i], shares);
|
2008-06-27 11:41:14 +00:00
|
|
|
}
|
2007-10-15 15:00:07 +00:00
|
|
|
|
sched: group scheduler, fix fairness of cpu bandwidth allocation for task groups
The current load balancing scheme isn't good enough for precise
group fairness.
For example: on a 8-cpu system, I created 3 groups as under:
a = 8 tasks (cpu.shares = 1024)
b = 4 tasks (cpu.shares = 1024)
c = 3 tasks (cpu.shares = 1024)
a, b and c are task groups that have equal weight. We would expect each
of the groups to receive 33.33% of cpu bandwidth under a fair scheduler.
This is what I get with the latest scheduler git tree:
Signed-off-by: Ingo Molnar <mingo@elte.hu>
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 277.676 | 57.8% | 54.1% 54.1% 54.1% 54.2% 56.7% 62.2% 62.8% 64.5%
b | 116.108 | 24.2% | 47.4% 48.1% 48.7% 49.3%
c | 86.326 | 18.0% | 47.5% 47.9% 48.5%
--------------------------------------------------------------------------------
Explanation of o/p:
Col1 -> Group name
Col2 -> Cumulative execution time (in seconds) received by all tasks of that
group in a 60sec window across 8 cpus
Col3 -> CPU bandwidth received by the group in the 60sec window, expressed in
percentage. Col3 data is derived as:
Col3 = 100 * Col2 / (NR_CPUS * 60)
Col4 -> CPU bandwidth received by each individual task of the group.
Col4 = 100 * cpu_time_recd_by_task / 60
[I can share the test case that produces a similar o/p if reqd]
The deviation from desired group fairness is as below:
a = +24.47%
b = -9.13%
c = -15.33%
which is quite high.
After the patch below is applied, here are the results:
--------------------------------------------------------------------------------
Col1 | Col2 | Col3 | Col4
------|---------|-------|-------------------------------------------------------
a | 163.112 | 34.0% | 33.2% 33.4% 33.5% 33.5% 33.7% 34.4% 34.8% 35.3%
b | 156.220 | 32.5% | 63.3% 64.5% 66.1% 66.5%
c | 160.653 | 33.5% | 85.8% 90.6% 91.4%
--------------------------------------------------------------------------------
Deviation from desired group fairness is as below:
a = +0.67%
b = -0.83%
c = +0.17%
which is far better IMO. Most of other runs have yielded a deviation within
+-2% at the most, which is good.
Why do we see bad (group) fairness with current scheuler?
=========================================================
Currently cpu's weight is just the summation of individual task weights.
This can yield incorrect results. For ex: consider three groups as below
on a 2-cpu system:
CPU0 CPU1
---------------------------
A (10) B(5)
C(5)
---------------------------
Group A has 10 tasks, all on CPU0, Group B and C have 5 tasks each all
of which are on CPU1. Each task has the same weight (NICE_0_LOAD =
1024).
The current scheme would yield a cpu weight of 10240 (10*1024) for each cpu and
the load balancer will think both CPUs are perfectly balanced and won't
move around any tasks. This, however, would yield this bandwidth:
A = 50%
B = 25%
C = 25%
which is not the desired result.
What's changing in the patch?
=============================
- How cpu weights are calculated when CONFIF_FAIR_GROUP_SCHED is
defined (see below)
- API Change
- Two tunables introduced in sysfs (under SCHED_DEBUG) to
control the frequency at which the load balance monitor
thread runs.
The basic change made in this patch is how cpu weight (rq->load.weight) is
calculated. Its now calculated as the summation of group weights on a cpu,
rather than summation of task weights. Weight exerted by a group on a
cpu is dependent on the shares allocated to it and also the number of
tasks the group has on that cpu compared to the total number of
(runnable) tasks the group has in the system.
Let,
W(K,i) = Weight of group K on cpu i
T(K,i) = Task load present in group K's cfs_rq on cpu i
T(K) = Total task load of group K across various cpus
S(K) = Shares allocated to group K
NRCPUS = Number of online cpus in the scheduler domain to
which group K is assigned.
Then,
W(K,i) = S(K) * NRCPUS * T(K,i) / T(K)
A load balance monitor thread is created at bootup, which periodically
runs and adjusts group's weight on each cpu. To avoid its overhead, two
min/max tunables are introduced (under SCHED_DEBUG) to control the rate
at which it runs.
Fixes from: Peter Zijlstra <a.p.zijlstra@chello.nl>
- don't start the load_balance_monitor when there is only a single cpu.
- rename the kthread because its currently longer than TASK_COMM_LEN
Signed-off-by: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-01-25 20:08:00 +00:00
|
|
|
/*
|
|
|
|
* Enable load balance activity on this group, by inserting it back on
|
|
|
|
* each cpu's rq->leaf_cfs_rq_list.
|
|
|
|
*/
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_lock_irqsave(&task_group_lock, flags);
|
2008-02-13 14:45:40 +00:00
|
|
|
for_each_possible_cpu(i)
|
|
|
|
register_fair_sched_group(tg, i);
|
2008-04-19 17:45:00 +00:00
|
|
|
list_add_rcu(&tg->siblings, &tg->parent->children);
|
2008-02-13 14:45:39 +00:00
|
|
|
spin_unlock_irqrestore(&task_group_lock, flags);
|
2007-10-15 15:00:14 +00:00
|
|
|
done:
|
2008-02-13 14:45:39 +00:00
|
|
|
mutex_unlock(&shares_mutex);
|
2007-10-15 15:00:09 +00:00
|
|
|
return 0;
|
2007-10-15 15:00:07 +00:00
|
|
|
}
|
|
|
|
|
2007-10-15 15:00:14 +00:00
|
|
|
unsigned long sched_group_shares(struct task_group *tg)
|
|
|
|
{
|
|
|
|
return tg->shares;
|
|
|
|
}
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2007-10-15 15:00:14 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
/*
|
2008-02-13 14:45:39 +00:00
|
|
|
* Ensure that the real time constraints are schedulable.
|
2008-01-25 20:08:30 +00:00
|
|
|
*/
|
2008-02-13 14:45:39 +00:00
|
|
|
static DEFINE_MUTEX(rt_constraints_mutex);
|
|
|
|
|
|
|
|
static unsigned long to_ratio(u64 period, u64 runtime)
|
|
|
|
{
|
|
|
|
if (runtime == RUNTIME_INF)
|
2008-08-19 10:33:06 +00:00
|
|
|
return 1ULL << 20;
|
2008-02-13 14:45:39 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
return div64_u64(runtime << 20, period);
|
2008-02-13 14:45:39 +00:00
|
|
|
}
|
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
/* Must be called with tasklist_lock held */
|
|
|
|
static inline int tg_has_rt_tasks(struct task_group *tg)
|
2008-04-19 17:45:00 +00:00
|
|
|
{
|
2008-08-19 10:33:06 +00:00
|
|
|
struct task_struct *g, *p;
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
do_each_thread(g, p) {
|
|
|
|
if (rt_task(p) && rt_rq_of_se(&p->rt)->tg == tg)
|
|
|
|
return 1;
|
|
|
|
} while_each_thread(g, p);
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
struct rt_schedulable_data {
|
|
|
|
struct task_group *tg;
|
|
|
|
u64 rt_period;
|
|
|
|
u64 rt_runtime;
|
|
|
|
};
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
static int tg_schedulable(struct task_group *tg, void *data)
|
|
|
|
{
|
|
|
|
struct rt_schedulable_data *d = data;
|
|
|
|
struct task_group *child;
|
|
|
|
unsigned long total, sum = 0;
|
|
|
|
u64 period, runtime;
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
period = ktime_to_ns(tg->rt_bandwidth.rt_period);
|
|
|
|
runtime = tg->rt_bandwidth.rt_runtime;
|
2008-04-19 17:45:00 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
if (tg == d->tg) {
|
|
|
|
period = d->rt_period;
|
|
|
|
runtime = d->rt_runtime;
|
2008-04-19 17:45:00 +00:00
|
|
|
}
|
|
|
|
|
2009-01-14 09:56:32 +00:00
|
|
|
#ifdef CONFIG_USER_SCHED
|
|
|
|
if (tg == &root_task_group) {
|
|
|
|
period = global_rt_period();
|
|
|
|
runtime = global_rt_runtime();
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2008-09-23 13:33:44 +00:00
|
|
|
/*
|
|
|
|
* Cannot have more runtime than the period.
|
|
|
|
*/
|
|
|
|
if (runtime > period && runtime != RUNTIME_INF)
|
|
|
|
return -EINVAL;
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-09-23 13:33:44 +00:00
|
|
|
/*
|
|
|
|
* Ensure we don't starve existing RT tasks.
|
|
|
|
*/
|
2008-08-19 10:33:06 +00:00
|
|
|
if (rt_bandwidth_enabled() && !runtime && tg_has_rt_tasks(tg))
|
|
|
|
return -EBUSY;
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
total = to_ratio(period, runtime);
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-09-23 13:33:44 +00:00
|
|
|
/*
|
|
|
|
* Nobody can have more than the global setting allows.
|
|
|
|
*/
|
|
|
|
if (total > to_ratio(global_rt_period(), global_rt_runtime()))
|
|
|
|
return -EINVAL;
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-09-23 13:33:44 +00:00
|
|
|
/*
|
|
|
|
* The sum of our children's runtime should not exceed our own.
|
|
|
|
*/
|
2008-08-19 10:33:06 +00:00
|
|
|
list_for_each_entry_rcu(child, &tg->children, siblings) {
|
|
|
|
period = ktime_to_ns(child->rt_bandwidth.rt_period);
|
|
|
|
runtime = child->rt_bandwidth.rt_runtime;
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
if (child == d->tg) {
|
|
|
|
period = d->rt_period;
|
|
|
|
runtime = d->rt_runtime;
|
|
|
|
}
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
sum += to_ratio(period, runtime);
|
2008-02-13 14:45:39 +00:00
|
|
|
}
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
if (sum > total)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return 0;
|
2008-01-25 20:08:30 +00:00
|
|
|
}
|
|
|
|
|
2008-08-19 10:33:06 +00:00
|
|
|
static int __rt_schedulable(struct task_group *tg, u64 period, u64 runtime)
|
2008-02-28 09:51:56 +00:00
|
|
|
{
|
2008-08-19 10:33:06 +00:00
|
|
|
struct rt_schedulable_data data = {
|
|
|
|
.tg = tg,
|
|
|
|
.rt_period = period,
|
|
|
|
.rt_runtime = runtime,
|
|
|
|
};
|
|
|
|
|
|
|
|
return walk_tg_tree(tg_schedulable, tg_nop, &data);
|
2008-02-28 09:51:56 +00:00
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
static int tg_set_bandwidth(struct task_group *tg,
|
|
|
|
u64 rt_period, u64 rt_runtime)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-04-19 17:44:58 +00:00
|
|
|
int i, err = 0;
|
2008-02-13 14:45:39 +00:00
|
|
|
|
|
|
|
mutex_lock(&rt_constraints_mutex);
|
2008-02-28 09:51:56 +00:00
|
|
|
read_lock(&tasklist_lock);
|
2008-08-19 10:33:06 +00:00
|
|
|
err = __rt_schedulable(tg, rt_period, rt_runtime);
|
|
|
|
if (err)
|
2008-02-13 14:45:39 +00:00
|
|
|
goto unlock;
|
2008-04-19 17:44:58 +00:00
|
|
|
|
|
|
|
spin_lock_irq(&tg->rt_bandwidth.rt_runtime_lock);
|
2008-04-19 17:44:57 +00:00
|
|
|
tg->rt_bandwidth.rt_period = ns_to_ktime(rt_period);
|
|
|
|
tg->rt_bandwidth.rt_runtime = rt_runtime;
|
2008-04-19 17:44:58 +00:00
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
struct rt_rq *rt_rq = tg->rt_rq[i];
|
|
|
|
|
|
|
|
spin_lock(&rt_rq->rt_runtime_lock);
|
|
|
|
rt_rq->rt_runtime = rt_runtime;
|
|
|
|
spin_unlock(&rt_rq->rt_runtime_lock);
|
|
|
|
}
|
|
|
|
spin_unlock_irq(&tg->rt_bandwidth.rt_runtime_lock);
|
2008-02-13 14:45:39 +00:00
|
|
|
unlock:
|
2008-02-28 09:51:56 +00:00
|
|
|
read_unlock(&tasklist_lock);
|
2008-02-13 14:45:39 +00:00
|
|
|
mutex_unlock(&rt_constraints_mutex);
|
|
|
|
|
|
|
|
return err;
|
2008-01-25 20:08:30 +00:00
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
int sched_group_set_rt_runtime(struct task_group *tg, long rt_runtime_us)
|
|
|
|
{
|
|
|
|
u64 rt_runtime, rt_period;
|
|
|
|
|
|
|
|
rt_period = ktime_to_ns(tg->rt_bandwidth.rt_period);
|
|
|
|
rt_runtime = (u64)rt_runtime_us * NSEC_PER_USEC;
|
|
|
|
if (rt_runtime_us < 0)
|
|
|
|
rt_runtime = RUNTIME_INF;
|
|
|
|
|
|
|
|
return tg_set_bandwidth(tg, rt_period, rt_runtime);
|
|
|
|
}
|
|
|
|
|
2008-02-13 14:45:39 +00:00
|
|
|
long sched_group_rt_runtime(struct task_group *tg)
|
|
|
|
{
|
|
|
|
u64 rt_runtime_us;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
if (tg->rt_bandwidth.rt_runtime == RUNTIME_INF)
|
2008-02-13 14:45:39 +00:00
|
|
|
return -1;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
rt_runtime_us = tg->rt_bandwidth.rt_runtime;
|
2008-02-13 14:45:39 +00:00
|
|
|
do_div(rt_runtime_us, NSEC_PER_USEC);
|
|
|
|
return rt_runtime_us;
|
|
|
|
}
|
2008-04-19 17:44:57 +00:00
|
|
|
|
|
|
|
int sched_group_set_rt_period(struct task_group *tg, long rt_period_us)
|
|
|
|
{
|
|
|
|
u64 rt_runtime, rt_period;
|
|
|
|
|
|
|
|
rt_period = (u64)rt_period_us * NSEC_PER_USEC;
|
|
|
|
rt_runtime = tg->rt_bandwidth.rt_runtime;
|
|
|
|
|
2008-06-26 16:54:09 +00:00
|
|
|
if (rt_period == 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
return tg_set_bandwidth(tg, rt_period, rt_runtime);
|
|
|
|
}
|
|
|
|
|
|
|
|
long sched_group_rt_period(struct task_group *tg)
|
|
|
|
{
|
|
|
|
u64 rt_period_us;
|
|
|
|
|
|
|
|
rt_period_us = ktime_to_ns(tg->rt_bandwidth.rt_period);
|
|
|
|
do_div(rt_period_us, NSEC_PER_USEC);
|
|
|
|
return rt_period_us;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int sched_rt_global_constraints(void)
|
|
|
|
{
|
2008-09-23 13:33:44 +00:00
|
|
|
u64 runtime, period;
|
2008-04-19 17:44:57 +00:00
|
|
|
int ret = 0;
|
|
|
|
|
2008-09-11 00:00:19 +00:00
|
|
|
if (sysctl_sched_rt_period <= 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-09-23 13:33:44 +00:00
|
|
|
runtime = global_rt_runtime();
|
|
|
|
period = global_rt_period();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sanity check on the sysctl variables.
|
|
|
|
*/
|
|
|
|
if (runtime > period && runtime != RUNTIME_INF)
|
|
|
|
return -EINVAL;
|
2008-06-19 12:22:27 +00:00
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
mutex_lock(&rt_constraints_mutex);
|
2008-08-19 10:33:06 +00:00
|
|
|
read_lock(&tasklist_lock);
|
2008-09-23 13:33:44 +00:00
|
|
|
ret = __rt_schedulable(NULL, 0, 0);
|
2008-08-19 10:33:06 +00:00
|
|
|
read_unlock(&tasklist_lock);
|
2008-04-19 17:44:57 +00:00
|
|
|
mutex_unlock(&rt_constraints_mutex);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2009-02-27 09:43:54 +00:00
|
|
|
|
|
|
|
int sched_rt_can_attach(struct task_group *tg, struct task_struct *tsk)
|
|
|
|
{
|
|
|
|
/* Don't accept realtime tasks when there is no way for them to run */
|
|
|
|
if (rt_task(tsk) && tg->rt_bandwidth.rt_runtime == 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2008-05-30 12:23:45 +00:00
|
|
|
#else /* !CONFIG_RT_GROUP_SCHED */
|
2008-04-19 17:44:57 +00:00
|
|
|
static int sched_rt_global_constraints(void)
|
|
|
|
{
|
2008-04-19 17:44:58 +00:00
|
|
|
unsigned long flags;
|
|
|
|
int i;
|
|
|
|
|
2008-09-11 00:00:19 +00:00
|
|
|
if (sysctl_sched_rt_period <= 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2009-05-05 15:50:21 +00:00
|
|
|
/*
|
|
|
|
* There's always some RT tasks in the root group
|
|
|
|
* -- migration, kstopmachine etc..
|
|
|
|
*/
|
|
|
|
if (sysctl_sched_rt_runtime == 0)
|
|
|
|
return -EBUSY;
|
|
|
|
|
2008-04-19 17:44:58 +00:00
|
|
|
spin_lock_irqsave(&def_rt_bandwidth.rt_runtime_lock, flags);
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
struct rt_rq *rt_rq = &cpu_rq(i)->rt;
|
|
|
|
|
|
|
|
spin_lock(&rt_rq->rt_runtime_lock);
|
|
|
|
rt_rq->rt_runtime = global_rt_runtime();
|
|
|
|
spin_unlock(&rt_rq->rt_runtime_lock);
|
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&def_rt_bandwidth.rt_runtime_lock, flags);
|
|
|
|
|
2008-04-19 17:44:57 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2008-04-19 17:44:57 +00:00
|
|
|
|
|
|
|
int sched_rt_handler(struct ctl_table *table, int write,
|
|
|
|
struct file *filp, void __user *buffer, size_t *lenp,
|
|
|
|
loff_t *ppos)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
int old_period, old_runtime;
|
|
|
|
static DEFINE_MUTEX(mutex);
|
|
|
|
|
|
|
|
mutex_lock(&mutex);
|
|
|
|
old_period = sysctl_sched_rt_period;
|
|
|
|
old_runtime = sysctl_sched_rt_runtime;
|
|
|
|
|
|
|
|
ret = proc_dointvec(table, write, filp, buffer, lenp, ppos);
|
|
|
|
|
|
|
|
if (!ret && write) {
|
|
|
|
ret = sched_rt_global_constraints();
|
|
|
|
if (ret) {
|
|
|
|
sysctl_sched_rt_period = old_period;
|
|
|
|
sysctl_sched_rt_runtime = old_runtime;
|
|
|
|
} else {
|
|
|
|
def_rt_bandwidth.rt_runtime = global_rt_runtime();
|
|
|
|
def_rt_bandwidth.rt_period =
|
|
|
|
ns_to_ktime(global_rt_period());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&mutex);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2007-10-19 06:41:03 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_CGROUP_SCHED
|
2007-10-19 06:41:03 +00:00
|
|
|
|
|
|
|
/* return corresponding task_group object of a cgroup */
|
2007-10-24 16:23:50 +00:00
|
|
|
static inline struct task_group *cgroup_tg(struct cgroup *cgrp)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2007-10-24 16:23:50 +00:00
|
|
|
return container_of(cgroup_subsys_state(cgrp, cpu_cgroup_subsys_id),
|
|
|
|
struct task_group, css);
|
2007-10-19 06:41:03 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct cgroup_subsys_state *
|
2007-10-24 16:23:50 +00:00
|
|
|
cpu_cgroup_create(struct cgroup_subsys *ss, struct cgroup *cgrp)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2008-04-19 17:44:59 +00:00
|
|
|
struct task_group *tg, *parent;
|
2007-10-19 06:41:03 +00:00
|
|
|
|
2007-10-24 16:23:50 +00:00
|
|
|
if (!cgrp->parent) {
|
2007-10-19 06:41:03 +00:00
|
|
|
/* This is early initialization for the top cgroup */
|
|
|
|
return &init_task_group.css;
|
|
|
|
}
|
|
|
|
|
2008-04-19 17:44:59 +00:00
|
|
|
parent = cgroup_tg(cgrp->parent);
|
|
|
|
tg = sched_create_group(parent);
|
2007-10-19 06:41:03 +00:00
|
|
|
if (IS_ERR(tg))
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
return &tg->css;
|
|
|
|
}
|
|
|
|
|
2007-12-05 14:46:09 +00:00
|
|
|
static void
|
|
|
|
cpu_cgroup_destroy(struct cgroup_subsys *ss, struct cgroup *cgrp)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2007-10-24 16:23:50 +00:00
|
|
|
struct task_group *tg = cgroup_tg(cgrp);
|
2007-10-19 06:41:03 +00:00
|
|
|
|
|
|
|
sched_destroy_group(tg);
|
|
|
|
}
|
|
|
|
|
2007-12-05 14:46:09 +00:00
|
|
|
static int
|
|
|
|
cpu_cgroup_can_attach(struct cgroup_subsys *ss, struct cgroup *cgrp,
|
|
|
|
struct task_struct *tsk)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2009-02-27 09:43:54 +00:00
|
|
|
if (!sched_rt_can_attach(cgroup_tg(cgrp), tsk))
|
2008-02-13 14:45:40 +00:00
|
|
|
return -EINVAL;
|
|
|
|
#else
|
2007-10-19 06:41:03 +00:00
|
|
|
/* We don't support RT-tasks being in separate groups */
|
|
|
|
if (tsk->sched_class != &fair_sched_class)
|
|
|
|
return -EINVAL;
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2007-10-19 06:41:03 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2007-10-24 16:23:50 +00:00
|
|
|
cpu_cgroup_attach(struct cgroup_subsys *ss, struct cgroup *cgrp,
|
2007-10-19 06:41:03 +00:00
|
|
|
struct cgroup *old_cont, struct task_struct *tsk)
|
|
|
|
{
|
|
|
|
sched_move_task(tsk);
|
|
|
|
}
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2008-04-29 07:59:56 +00:00
|
|
|
static int cpu_shares_write_u64(struct cgroup *cgrp, struct cftype *cftype,
|
2007-10-24 16:23:50 +00:00
|
|
|
u64 shareval)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2007-10-24 16:23:50 +00:00
|
|
|
return sched_group_set_shares(cgroup_tg(cgrp), shareval);
|
2007-10-19 06:41:03 +00:00
|
|
|
}
|
|
|
|
|
2008-04-29 07:59:56 +00:00
|
|
|
static u64 cpu_shares_read_u64(struct cgroup *cgrp, struct cftype *cft)
|
2007-10-19 06:41:03 +00:00
|
|
|
{
|
2007-10-24 16:23:50 +00:00
|
|
|
struct task_group *tg = cgroup_tg(cgrp);
|
2007-10-19 06:41:03 +00:00
|
|
|
|
|
|
|
return (u64) tg->shares;
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
2007-10-19 06:41:03 +00:00
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-05-14 23:05:46 +00:00
|
|
|
static int cpu_rt_runtime_write(struct cgroup *cgrp, struct cftype *cft,
|
2008-04-29 08:00:06 +00:00
|
|
|
s64 val)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-04-29 08:00:06 +00:00
|
|
|
return sched_group_set_rt_runtime(cgroup_tg(cgrp), val);
|
2008-01-25 20:08:30 +00:00
|
|
|
}
|
|
|
|
|
2008-04-29 08:00:06 +00:00
|
|
|
static s64 cpu_rt_runtime_read(struct cgroup *cgrp, struct cftype *cft)
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-04-29 08:00:06 +00:00
|
|
|
return sched_group_rt_runtime(cgroup_tg(cgrp));
|
2008-01-25 20:08:30 +00:00
|
|
|
}
|
2008-04-19 17:44:57 +00:00
|
|
|
|
|
|
|
static int cpu_rt_period_write_uint(struct cgroup *cgrp, struct cftype *cftype,
|
|
|
|
u64 rt_period_us)
|
|
|
|
{
|
|
|
|
return sched_group_set_rt_period(cgroup_tg(cgrp), rt_period_us);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u64 cpu_rt_period_read_uint(struct cgroup *cgrp, struct cftype *cft)
|
|
|
|
{
|
|
|
|
return sched_group_rt_period(cgroup_tg(cgrp));
|
|
|
|
}
|
2008-05-30 12:23:45 +00:00
|
|
|
#endif /* CONFIG_RT_GROUP_SCHED */
|
2008-01-25 20:08:30 +00:00
|
|
|
|
2007-10-29 20:18:11 +00:00
|
|
|
static struct cftype cpu_files[] = {
|
2008-02-13 14:45:40 +00:00
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
2007-10-29 20:18:11 +00:00
|
|
|
{
|
|
|
|
.name = "shares",
|
2008-04-29 07:59:56 +00:00
|
|
|
.read_u64 = cpu_shares_read_u64,
|
|
|
|
.write_u64 = cpu_shares_write_u64,
|
2007-10-29 20:18:11 +00:00
|
|
|
},
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_RT_GROUP_SCHED
|
2008-01-25 20:08:30 +00:00
|
|
|
{
|
2008-02-13 14:45:39 +00:00
|
|
|
.name = "rt_runtime_us",
|
2008-04-29 08:00:06 +00:00
|
|
|
.read_s64 = cpu_rt_runtime_read,
|
|
|
|
.write_s64 = cpu_rt_runtime_write,
|
2008-01-25 20:08:30 +00:00
|
|
|
},
|
2008-04-19 17:44:57 +00:00
|
|
|
{
|
|
|
|
.name = "rt_period_us",
|
2008-04-29 07:59:56 +00:00
|
|
|
.read_u64 = cpu_rt_period_read_uint,
|
|
|
|
.write_u64 = cpu_rt_period_write_uint,
|
2008-04-19 17:44:57 +00:00
|
|
|
},
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif
|
2007-10-19 06:41:03 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static int cpu_cgroup_populate(struct cgroup_subsys *ss, struct cgroup *cont)
|
|
|
|
{
|
2007-10-29 20:18:11 +00:00
|
|
|
return cgroup_add_files(cont, ss, cpu_files, ARRAY_SIZE(cpu_files));
|
2007-10-19 06:41:03 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
struct cgroup_subsys cpu_cgroup_subsys = {
|
2007-10-29 20:18:11 +00:00
|
|
|
.name = "cpu",
|
|
|
|
.create = cpu_cgroup_create,
|
|
|
|
.destroy = cpu_cgroup_destroy,
|
|
|
|
.can_attach = cpu_cgroup_can_attach,
|
|
|
|
.attach = cpu_cgroup_attach,
|
|
|
|
.populate = cpu_cgroup_populate,
|
|
|
|
.subsys_id = cpu_cgroup_subsys_id,
|
2007-10-19 06:41:03 +00:00
|
|
|
.early_init = 1,
|
|
|
|
};
|
|
|
|
|
2008-02-13 14:45:40 +00:00
|
|
|
#endif /* CONFIG_CGROUP_SCHED */
|
2007-12-02 19:04:49 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_CGROUP_CPUACCT
|
|
|
|
|
|
|
|
/*
|
|
|
|
* CPU accounting code for task groups.
|
|
|
|
*
|
|
|
|
* Based on the work by Paul Menage (menage@google.com) and Balbir Singh
|
|
|
|
* (balbir@in.ibm.com).
|
|
|
|
*/
|
|
|
|
|
2008-11-10 15:11:13 +00:00
|
|
|
/* track cpu usage of a group of tasks and its child groups */
|
2007-12-02 19:04:49 +00:00
|
|
|
struct cpuacct {
|
|
|
|
struct cgroup_subsys_state css;
|
|
|
|
/* cpuusage holds pointer to a u64-type object on every cpu */
|
|
|
|
u64 *cpuusage;
|
2009-03-31 04:32:22 +00:00
|
|
|
struct percpu_counter cpustat[CPUACCT_STAT_NSTATS];
|
2008-11-10 15:11:13 +00:00
|
|
|
struct cpuacct *parent;
|
2007-12-02 19:04:49 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct cgroup_subsys cpuacct_subsys;
|
|
|
|
|
|
|
|
/* return cpu accounting group corresponding to this container */
|
2008-02-29 04:32:43 +00:00
|
|
|
static inline struct cpuacct *cgroup_ca(struct cgroup *cgrp)
|
2007-12-02 19:04:49 +00:00
|
|
|
{
|
2008-02-29 04:32:43 +00:00
|
|
|
return container_of(cgroup_subsys_state(cgrp, cpuacct_subsys_id),
|
2007-12-02 19:04:49 +00:00
|
|
|
struct cpuacct, css);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* return cpu accounting group to which this task belongs */
|
|
|
|
static inline struct cpuacct *task_ca(struct task_struct *tsk)
|
|
|
|
{
|
|
|
|
return container_of(task_subsys_state(tsk, cpuacct_subsys_id),
|
|
|
|
struct cpuacct, css);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* create a new cpu accounting group */
|
|
|
|
static struct cgroup_subsys_state *cpuacct_create(
|
2008-02-29 04:32:43 +00:00
|
|
|
struct cgroup_subsys *ss, struct cgroup *cgrp)
|
2007-12-02 19:04:49 +00:00
|
|
|
{
|
|
|
|
struct cpuacct *ca = kzalloc(sizeof(*ca), GFP_KERNEL);
|
2009-03-31 04:32:22 +00:00
|
|
|
int i;
|
2007-12-02 19:04:49 +00:00
|
|
|
|
|
|
|
if (!ca)
|
2009-03-31 04:32:22 +00:00
|
|
|
goto out;
|
2007-12-02 19:04:49 +00:00
|
|
|
|
|
|
|
ca->cpuusage = alloc_percpu(u64);
|
2009-03-31 04:32:22 +00:00
|
|
|
if (!ca->cpuusage)
|
|
|
|
goto out_free_ca;
|
|
|
|
|
|
|
|
for (i = 0; i < CPUACCT_STAT_NSTATS; i++)
|
|
|
|
if (percpu_counter_init(&ca->cpustat[i], 0))
|
|
|
|
goto out_free_counters;
|
2007-12-02 19:04:49 +00:00
|
|
|
|
2008-11-10 15:11:13 +00:00
|
|
|
if (cgrp->parent)
|
|
|
|
ca->parent = cgroup_ca(cgrp->parent);
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
return &ca->css;
|
2009-03-31 04:32:22 +00:00
|
|
|
|
|
|
|
out_free_counters:
|
|
|
|
while (--i >= 0)
|
|
|
|
percpu_counter_destroy(&ca->cpustat[i]);
|
|
|
|
free_percpu(ca->cpuusage);
|
|
|
|
out_free_ca:
|
|
|
|
kfree(ca);
|
|
|
|
out:
|
|
|
|
return ERR_PTR(-ENOMEM);
|
2007-12-02 19:04:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* destroy an existing cpu accounting group */
|
2007-12-05 14:46:09 +00:00
|
|
|
static void
|
2008-02-29 04:32:43 +00:00
|
|
|
cpuacct_destroy(struct cgroup_subsys *ss, struct cgroup *cgrp)
|
2007-12-02 19:04:49 +00:00
|
|
|
{
|
2008-02-29 04:32:43 +00:00
|
|
|
struct cpuacct *ca = cgroup_ca(cgrp);
|
2009-03-31 04:32:22 +00:00
|
|
|
int i;
|
2007-12-02 19:04:49 +00:00
|
|
|
|
2009-03-31 04:32:22 +00:00
|
|
|
for (i = 0; i < CPUACCT_STAT_NSTATS; i++)
|
|
|
|
percpu_counter_destroy(&ca->cpustat[i]);
|
2007-12-02 19:04:49 +00:00
|
|
|
free_percpu(ca->cpuusage);
|
|
|
|
kfree(ca);
|
|
|
|
}
|
|
|
|
|
2008-12-16 06:02:01 +00:00
|
|
|
static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu)
|
|
|
|
{
|
2009-02-20 07:29:08 +00:00
|
|
|
u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
|
2008-12-16 06:02:01 +00:00
|
|
|
u64 data;
|
|
|
|
|
|
|
|
#ifndef CONFIG_64BIT
|
|
|
|
/*
|
|
|
|
* Take rq->lock to make 64-bit read safe on 32-bit platforms.
|
|
|
|
*/
|
|
|
|
spin_lock_irq(&cpu_rq(cpu)->lock);
|
|
|
|
data = *cpuusage;
|
|
|
|
spin_unlock_irq(&cpu_rq(cpu)->lock);
|
|
|
|
#else
|
|
|
|
data = *cpuusage;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu, u64 val)
|
|
|
|
{
|
2009-02-20 07:29:08 +00:00
|
|
|
u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
|
2008-12-16 06:02:01 +00:00
|
|
|
|
|
|
|
#ifndef CONFIG_64BIT
|
|
|
|
/*
|
|
|
|
* Take rq->lock to make 64-bit write safe on 32-bit platforms.
|
|
|
|
*/
|
|
|
|
spin_lock_irq(&cpu_rq(cpu)->lock);
|
|
|
|
*cpuusage = val;
|
|
|
|
spin_unlock_irq(&cpu_rq(cpu)->lock);
|
|
|
|
#else
|
|
|
|
*cpuusage = val;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
/* return total cpu usage (in nanoseconds) of a group */
|
2008-02-29 04:32:43 +00:00
|
|
|
static u64 cpuusage_read(struct cgroup *cgrp, struct cftype *cft)
|
2007-12-02 19:04:49 +00:00
|
|
|
{
|
2008-02-29 04:32:43 +00:00
|
|
|
struct cpuacct *ca = cgroup_ca(cgrp);
|
2007-12-02 19:04:49 +00:00
|
|
|
u64 totalcpuusage = 0;
|
|
|
|
int i;
|
|
|
|
|
2008-12-16 06:02:01 +00:00
|
|
|
for_each_present_cpu(i)
|
|
|
|
totalcpuusage += cpuacct_cpuusage_read(ca, i);
|
2007-12-02 19:04:49 +00:00
|
|
|
|
|
|
|
return totalcpuusage;
|
|
|
|
}
|
|
|
|
|
2008-02-29 04:32:44 +00:00
|
|
|
static int cpuusage_write(struct cgroup *cgrp, struct cftype *cftype,
|
|
|
|
u64 reset)
|
|
|
|
{
|
|
|
|
struct cpuacct *ca = cgroup_ca(cgrp);
|
|
|
|
int err = 0;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (reset) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2008-12-16 06:02:01 +00:00
|
|
|
for_each_present_cpu(i)
|
|
|
|
cpuacct_cpuusage_write(ca, i, 0);
|
2008-02-29 04:32:44 +00:00
|
|
|
|
|
|
|
out:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2008-12-16 06:04:15 +00:00
|
|
|
static int cpuacct_percpu_seq_read(struct cgroup *cgroup, struct cftype *cft,
|
|
|
|
struct seq_file *m)
|
|
|
|
{
|
|
|
|
struct cpuacct *ca = cgroup_ca(cgroup);
|
|
|
|
u64 percpu;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for_each_present_cpu(i) {
|
|
|
|
percpu = cpuacct_cpuusage_read(ca, i);
|
|
|
|
seq_printf(m, "%llu ", (unsigned long long) percpu);
|
|
|
|
}
|
|
|
|
seq_printf(m, "\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-03-31 04:32:22 +00:00
|
|
|
static const char *cpuacct_stat_desc[] = {
|
|
|
|
[CPUACCT_STAT_USER] = "user",
|
|
|
|
[CPUACCT_STAT_SYSTEM] = "system",
|
|
|
|
};
|
|
|
|
|
|
|
|
static int cpuacct_stats_show(struct cgroup *cgrp, struct cftype *cft,
|
|
|
|
struct cgroup_map_cb *cb)
|
|
|
|
{
|
|
|
|
struct cpuacct *ca = cgroup_ca(cgrp);
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < CPUACCT_STAT_NSTATS; i++) {
|
|
|
|
s64 val = percpu_counter_read(&ca->cpustat[i]);
|
|
|
|
val = cputime64_to_clock_t(val);
|
|
|
|
cb->fill(cb, cpuacct_stat_desc[i], val);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
static struct cftype files[] = {
|
|
|
|
{
|
|
|
|
.name = "usage",
|
2008-04-29 07:59:56 +00:00
|
|
|
.read_u64 = cpuusage_read,
|
|
|
|
.write_u64 = cpuusage_write,
|
2007-12-02 19:04:49 +00:00
|
|
|
},
|
2008-12-16 06:04:15 +00:00
|
|
|
{
|
|
|
|
.name = "usage_percpu",
|
|
|
|
.read_seq_string = cpuacct_percpu_seq_read,
|
|
|
|
},
|
2009-03-31 04:32:22 +00:00
|
|
|
{
|
|
|
|
.name = "stat",
|
|
|
|
.read_map = cpuacct_stats_show,
|
|
|
|
},
|
2007-12-02 19:04:49 +00:00
|
|
|
};
|
|
|
|
|
2008-02-29 04:32:43 +00:00
|
|
|
static int cpuacct_populate(struct cgroup_subsys *ss, struct cgroup *cgrp)
|
2007-12-02 19:04:49 +00:00
|
|
|
{
|
2008-02-29 04:32:43 +00:00
|
|
|
return cgroup_add_files(cgrp, ss, files, ARRAY_SIZE(files));
|
2007-12-02 19:04:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* charge this task's execution time to its accounting group.
|
|
|
|
*
|
|
|
|
* called with rq->lock held.
|
|
|
|
*/
|
|
|
|
static void cpuacct_charge(struct task_struct *tsk, u64 cputime)
|
|
|
|
{
|
|
|
|
struct cpuacct *ca;
|
2008-11-10 15:11:13 +00:00
|
|
|
int cpu;
|
2007-12-02 19:04:49 +00:00
|
|
|
|
2009-02-26 07:40:15 +00:00
|
|
|
if (unlikely(!cpuacct_subsys.active))
|
2007-12-02 19:04:49 +00:00
|
|
|
return;
|
|
|
|
|
2008-11-10 15:11:13 +00:00
|
|
|
cpu = task_cpu(tsk);
|
2009-03-23 04:32:53 +00:00
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
ca = task_ca(tsk);
|
|
|
|
|
2008-11-10 15:11:13 +00:00
|
|
|
for (; ca; ca = ca->parent) {
|
2009-02-20 07:29:08 +00:00
|
|
|
u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
|
2007-12-02 19:04:49 +00:00
|
|
|
*cpuusage += cputime;
|
|
|
|
}
|
2009-03-23 04:32:53 +00:00
|
|
|
|
|
|
|
rcu_read_unlock();
|
2007-12-02 19:04:49 +00:00
|
|
|
}
|
|
|
|
|
2009-03-31 04:32:22 +00:00
|
|
|
/*
|
|
|
|
* Charge the system/user time to the task's accounting group.
|
|
|
|
*/
|
|
|
|
static void cpuacct_update_stats(struct task_struct *tsk,
|
|
|
|
enum cpuacct_stat_index idx, cputime_t val)
|
|
|
|
{
|
|
|
|
struct cpuacct *ca;
|
|
|
|
|
|
|
|
if (unlikely(!cpuacct_subsys.active))
|
|
|
|
return;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
ca = task_ca(tsk);
|
|
|
|
|
|
|
|
do {
|
|
|
|
percpu_counter_add(&ca->cpustat[idx], val);
|
|
|
|
ca = ca->parent;
|
|
|
|
} while (ca);
|
|
|
|
rcu_read_unlock();
|
|
|
|
}
|
|
|
|
|
2007-12-02 19:04:49 +00:00
|
|
|
struct cgroup_subsys cpuacct_subsys = {
|
|
|
|
.name = "cpuacct",
|
|
|
|
.create = cpuacct_create,
|
|
|
|
.destroy = cpuacct_destroy,
|
|
|
|
.populate = cpuacct_populate,
|
|
|
|
.subsys_id = cpuacct_subsys_id,
|
|
|
|
};
|
|
|
|
#endif /* CONFIG_CGROUP_CPUACCT */
|