Files
linux/kernel
Peter Zijlstra 1dc0fffc48 sched/core: Robustify preemption leak checks
When we warn about a preempt_count leak; reset the preempt_count to
the known good value such that the problem does not ripple forward.

This is most important on x86 which has a per cpu preempt_count that is
not saved/restored (after this series). So if you schedule with an
invalid (!2*PREEMPT_DISABLE_OFFSET) preempt_count the next task is
messed up too.

Enforcing this invariant limits the borkage to just the one task.

Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Frederic Weisbecker <fweisbec@gmail.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Reviewed-by: Steven Rostedt <rostedt@goodmis.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2015-10-06 17:08:17 +02:00
..
2015-06-30 19:44:57 -07:00
2015-04-11 22:27:55 -04:00
2014-10-09 22:26:04 -04:00
2014-08-08 15:57:18 -07:00
2015-04-12 21:03:31 +02:00
2015-07-20 21:43:54 +02:00
2015-08-03 11:34:16 +02:00
2014-09-10 15:42:12 -07:00
2015-08-27 19:40:58 -04:00
2014-12-04 14:34:54 -05:00
2015-01-17 10:02:23 +13:00
2014-12-10 17:41:10 -08:00