mirror of https://gitee.com/openkylin/linux.git
[PATCH] OOM kill: children accounting
In the badness() calculation, there's currently this piece of code: /* * Processes which fork a lot of child processes are likely * a good choice. We add the vmsize of the children if they * have an own mm. This prevents forking servers to flood the * machine with an endless amount of children */ list_for_each(tsk, &p->children) { struct task_struct *chld; chld = list_entry(tsk, struct task_struct, sibling); if (chld->mm = p->mm && chld->mm) points += chld->mm->total_vm; } The intention is clear: If some server (apache) keeps spawning new children and we run OOM, we want to kill the father rather than picking a child. This -- to some degree -- also helps a bit with getting fork bombs under control, though I'd consider this a desirable side-effect rather than a feature. There's one problem with this: No matter how many or few children there are, if just one of them misbehaves, and all others (including the father) do everything right, we still always kill the whole family. This hits in real life; whether it's javascript in konqueror resulting in kdeinit (and thus the whole KDE session) being hit or just a classical server that spawns children. Sidenote: The killer does kill all direct children as well, not only the selected father, see oom_kill_process(). The idea in attached patch is that we do want to account the memory consumption of the (direct) children to the father -- however not fully. This maintains the property that fathers with too many children will still very likely be picked, whereas a single misbehaving child has the chance to be picked by the OOM killer. In the patch I account only half (rounded up) of the children's vm_size to the parent. This means that if one child eats more mem than the rest of the family, it will be picked, otherwise it's still the father and thus the whole family that gets selected. This is heuristics -- we could debate whether accounting for a fourth would be better than for half of it. Or -- if people would consider it worth the trouble -- make it a sysctl. For now I sticked to accounting for half, which should IMHO be a significant improvement. The patch does one more thing: As users tend to be irritated by the choice of killed processes (mainly because the children are killed first, despite some of them having a very low OOM score), I added some more output: The selected (father) process will be reported first and it's oom_score printed to syslog. Description: Only account for half of children's vm size in oom score calculation This should still give the parent enough point in case of fork bombs. If any child however has more than 50% of the vm size of all children together, it'll get a higher score and be elected. This patch also makes the kernel display the oom_score. Signed-off-by: Kurt Garloff <garloff@suse.de> Cc: Rik van Riel <riel@redhat.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
bd71c2b174
commit
9827b781f2
|
@ -58,15 +58,17 @@ unsigned long badness(struct task_struct *p, unsigned long uptime)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Processes which fork a lot of child processes are likely
|
* Processes which fork a lot of child processes are likely
|
||||||
* a good choice. We add the vmsize of the children if they
|
* a good choice. We add half the vmsize of the children if they
|
||||||
* have an own mm. This prevents forking servers to flood the
|
* have an own mm. This prevents forking servers to flood the
|
||||||
* machine with an endless amount of children
|
* machine with an endless amount of children. In case a single
|
||||||
|
* child is eating the vast majority of memory, adding only half
|
||||||
|
* to the parents will make the child our kill candidate of choice.
|
||||||
*/
|
*/
|
||||||
list_for_each(tsk, &p->children) {
|
list_for_each(tsk, &p->children) {
|
||||||
struct task_struct *chld;
|
struct task_struct *chld;
|
||||||
chld = list_entry(tsk, struct task_struct, sibling);
|
chld = list_entry(tsk, struct task_struct, sibling);
|
||||||
if (chld->mm != p->mm && chld->mm)
|
if (chld->mm != p->mm && chld->mm)
|
||||||
points += chld->mm->total_vm;
|
points += chld->mm->total_vm/2 + 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -136,12 +138,12 @@ unsigned long badness(struct task_struct *p, unsigned long uptime)
|
||||||
*
|
*
|
||||||
* (not docbooked, we don't want this one cluttering up the manual)
|
* (not docbooked, we don't want this one cluttering up the manual)
|
||||||
*/
|
*/
|
||||||
static struct task_struct * select_bad_process(void)
|
static struct task_struct *select_bad_process(unsigned long *ppoints)
|
||||||
{
|
{
|
||||||
unsigned long maxpoints = 0;
|
|
||||||
struct task_struct *g, *p;
|
struct task_struct *g, *p;
|
||||||
struct task_struct *chosen = NULL;
|
struct task_struct *chosen = NULL;
|
||||||
struct timespec uptime;
|
struct timespec uptime;
|
||||||
|
*ppoints = 0;
|
||||||
|
|
||||||
do_posix_clock_monotonic_gettime(&uptime);
|
do_posix_clock_monotonic_gettime(&uptime);
|
||||||
do_each_thread(g, p) {
|
do_each_thread(g, p) {
|
||||||
|
@ -169,9 +171,9 @@ static struct task_struct * select_bad_process(void)
|
||||||
return p;
|
return p;
|
||||||
|
|
||||||
points = badness(p, uptime.tv_sec);
|
points = badness(p, uptime.tv_sec);
|
||||||
if (points > maxpoints || !chosen) {
|
if (points > *ppoints || !chosen) {
|
||||||
chosen = p;
|
chosen = p;
|
||||||
maxpoints = points;
|
*ppoints = points;
|
||||||
}
|
}
|
||||||
} while_each_thread(g, p);
|
} while_each_thread(g, p);
|
||||||
return chosen;
|
return chosen;
|
||||||
|
@ -237,12 +239,15 @@ static struct mm_struct *oom_kill_task(task_t *p)
|
||||||
return mm;
|
return mm;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct mm_struct *oom_kill_process(struct task_struct *p)
|
static struct mm_struct *oom_kill_process(struct task_struct *p,
|
||||||
|
unsigned long points)
|
||||||
{
|
{
|
||||||
struct mm_struct *mm;
|
struct mm_struct *mm;
|
||||||
struct task_struct *c;
|
struct task_struct *c;
|
||||||
struct list_head *tsk;
|
struct list_head *tsk;
|
||||||
|
|
||||||
|
printk(KERN_ERR "Out of Memory: Kill process %d (%s) score %li and "
|
||||||
|
"children.\n", p->pid, p->comm, points);
|
||||||
/* Try to kill a child first */
|
/* Try to kill a child first */
|
||||||
list_for_each(tsk, &p->children) {
|
list_for_each(tsk, &p->children) {
|
||||||
c = list_entry(tsk, struct task_struct, sibling);
|
c = list_entry(tsk, struct task_struct, sibling);
|
||||||
|
@ -267,6 +272,7 @@ void out_of_memory(gfp_t gfp_mask, int order)
|
||||||
{
|
{
|
||||||
struct mm_struct *mm = NULL;
|
struct mm_struct *mm = NULL;
|
||||||
task_t * p;
|
task_t * p;
|
||||||
|
unsigned long points;
|
||||||
|
|
||||||
if (printk_ratelimit()) {
|
if (printk_ratelimit()) {
|
||||||
printk("oom-killer: gfp_mask=0x%x, order=%d\n",
|
printk("oom-killer: gfp_mask=0x%x, order=%d\n",
|
||||||
|
@ -278,7 +284,7 @@ void out_of_memory(gfp_t gfp_mask, int order)
|
||||||
cpuset_lock();
|
cpuset_lock();
|
||||||
read_lock(&tasklist_lock);
|
read_lock(&tasklist_lock);
|
||||||
retry:
|
retry:
|
||||||
p = select_bad_process();
|
p = select_bad_process(&points);
|
||||||
|
|
||||||
if (PTR_ERR(p) == -1UL)
|
if (PTR_ERR(p) == -1UL)
|
||||||
goto out;
|
goto out;
|
||||||
|
@ -290,7 +296,7 @@ void out_of_memory(gfp_t gfp_mask, int order)
|
||||||
panic("Out of memory and no killable processes...\n");
|
panic("Out of memory and no killable processes...\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
mm = oom_kill_process(p);
|
mm = oom_kill_process(p, points);
|
||||||
if (!mm)
|
if (!mm)
|
||||||
goto retry;
|
goto retry;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue