summaryrefslogtreecommitdiffstats
path: root/debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch10
1 files changed, 5 insertions, 5 deletions
diff --git a/debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch b/debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch
index 4aa84f292..5fb98b103 100644
--- a/debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch
+++ b/debian/patches-rt/0257-workqueue-Prevent-deadlock-stall-on-RT.patch
@@ -1,7 +1,7 @@
From: Thomas Gleixner <tglx@linutronix.de>
Date: Fri, 27 Jun 2014 16:24:52 +0200
-Subject: [PATCH 257/351] workqueue: Prevent deadlock/stall on RT
-Origin: https://git.kernel.org/cgit/linux/kernel/git/rt/linux-stable-rt.git/commit?id=87d9d53971f4e410e1ad22c2c9a80c3b1de7b58c
+Subject: [PATCH 257/353] workqueue: Prevent deadlock/stall on RT
+Origin: https://git.kernel.org/cgit/linux/kernel/git/rt/linux-stable-rt.git/commit?id=e621f4bb1ebb318d4bac120c74aefd215cc67662
Austin reported a XFS deadlock/stall on RT where scheduled work gets
never exececuted and tasks are waiting for each other for ever.
@@ -42,10 +42,10 @@ Cc: Steven Rostedt <rostedt@goodmis.org>
2 files changed, 51 insertions(+), 15 deletions(-)
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
-index cf60731667bf..f9babcec1afc 100644
+index e581ce1edc12..fd1a7d1ff9f4 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
-@@ -3615,9 +3615,8 @@ void __noreturn do_task_dead(void)
+@@ -3617,9 +3617,8 @@ void __noreturn do_task_dead(void)
static inline void sched_submit_work(struct task_struct *tsk)
{
@@ -56,7 +56,7 @@ index cf60731667bf..f9babcec1afc 100644
/*
* If a worker went to sleep, notify and ask workqueue whether
* it wants to wake up a task to maintain concurrency.
-@@ -3631,6 +3630,9 @@ static inline void sched_submit_work(struct task_struct *tsk)
+@@ -3633,6 +3632,9 @@ static inline void sched_submit_work(struct task_struct *tsk)
preempt_enable_no_resched();
}