summaryrefslogtreecommitdiffstats
path: root/debian/patches-rt/0020-printk-nbcon-Provide-function-to-flush-using-write_a.patch
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--debian/patches-rt/0020-printk-nbcon-Provide-function-to-flush-using-write_a.patch (renamed from debian/patches-rt/0024-printk-nbcon-Provide-function-to-flush-using-write_a.patch)129
1 files changed, 68 insertions, 61 deletions
diff --git a/debian/patches-rt/0024-printk-nbcon-Provide-function-to-flush-using-write_a.patch b/debian/patches-rt/0020-printk-nbcon-Provide-function-to-flush-using-write_a.patch
index 9f8a3e3765..26c68ed229 100644
--- a/debian/patches-rt/0024-printk-nbcon-Provide-function-to-flush-using-write_a.patch
+++ b/debian/patches-rt/0020-printk-nbcon-Provide-function-to-flush-using-write_a.patch
@@ -1,26 +1,29 @@
From: Thomas Gleixner <tglx@linutronix.de>
Date: Tue, 12 Sep 2023 12:00:08 +0000
-Subject: [PATCH 24/50] printk: nbcon: Provide function to flush using
+Subject: [PATCH 20/48] printk: nbcon: Provide function to flush using
write_atomic()
-Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/6.7/older/patches-6.7-rt6.tar.xz
+Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/6.8/older/patches-6.8.2-rt11.tar.xz
-Provide nbcon_atomic_flush_all() to perform flushing of all
+Provide nbcon_atomic_flush_pending() to perform flushing of all
registered nbcon consoles using their write_atomic() callback.
-Like with legacy consoles, the nbcon consoles are flushed one
-record per console. This allows all nbcon consoles to print
-lines pseudo-simultaneously, rather than one console waiting
-for the full ringbuffer to dump to another console before
-printing anything.
-Unlike console_flush_all(), nbcon_atomic_flush_all() will only
-flush up through the newest record at the time of the call.
-This prevents a CPU from printing unbounded when other CPUs are
-adding records.
+Unlike console_flush_all(), nbcon_atomic_flush_pending() will
+only flush up through the newest record at the time of the
+call. This prevents a CPU from printing unbounded when other
+CPUs are adding records.
-Perform nbcon console atomic flushing in
-console_flush_on_panic(). This function is not only used in
-panic() but also other locations where there may be stored
-messages that need to be flushed.
+Also unlike console_flush_all(), nbcon_atomic_flush_pending()
+will fully flush one console before flushing the next. This
+helps to guarantee that a block of pending records (such as
+a stack trace in an emergency situation) can be printed
+atomically at once before releasing console ownership.
+
+nbcon_atomic_flush_pending() is safe in any context because it
+uses write_atomic() and acquires with unsafe_takeover disabled.
+
+Use it in console_flush_on_panic() before flushing legacy
+consoles. The legacy write() callbacks are not fully safe when
+oops_in_progress is set.
Co-developed-by: John Ogness <john.ogness@linutronix.de>
Signed-off-by: John Ogness <john.ogness@linutronix.de>
@@ -28,31 +31,31 @@ Signed-off-by: Thomas Gleixner (Intel) <tglx@linutronix.de>
Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
---
kernel/printk/internal.h | 2
- kernel/printk/nbcon.c | 100 ++++++++++++++++++++++++++++++++++++++++++++++-
+ kernel/printk/nbcon.c | 104 ++++++++++++++++++++++++++++++++++++++++++++++-
kernel/printk/printk.c | 2
- 3 files changed, 102 insertions(+), 2 deletions(-)
+ 3 files changed, 106 insertions(+), 2 deletions(-)
--- a/kernel/printk/internal.h
+++ b/kernel/printk/internal.h
-@@ -77,6 +77,7 @@ void nbcon_seq_force(struct console *con
+@@ -84,6 +84,7 @@ void nbcon_seq_force(struct console *con
bool nbcon_alloc(struct console *con);
void nbcon_init(struct console *con);
void nbcon_free(struct console *con);
-+void nbcon_atomic_flush_all(void);
++void nbcon_atomic_flush_pending(void);
/*
* Check if the given console is currently capable and allowed to print
-@@ -131,6 +132,7 @@ static inline void nbcon_seq_force(struc
+@@ -138,6 +139,7 @@ static inline void nbcon_seq_force(struc
static inline bool nbcon_alloc(struct console *con) { return false; }
static inline void nbcon_init(struct console *con) { }
static inline void nbcon_free(struct console *con) { }
-+static inline void nbcon_atomic_flush_all(void) { }
++static inline void nbcon_atomic_flush_pending(void) { }
static inline bool console_is_usable(struct console *con, short flags) { return false; }
--- a/kernel/printk/nbcon.c
+++ b/kernel/printk/nbcon.c
-@@ -539,7 +539,6 @@ static struct printk_buffers panic_nbcon
+@@ -548,7 +548,6 @@ static struct printk_buffers panic_nbcon
* in an unsafe state. Otherwise, on success the caller may assume
* the console is not in an unsafe state.
*/
@@ -60,7 +63,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
static bool nbcon_context_try_acquire(struct nbcon_context *ctxt)
{
unsigned int cpu = smp_processor_id();
-@@ -841,7 +840,6 @@ EXPORT_SYMBOL_GPL(nbcon_exit_unsafe);
+@@ -850,7 +849,6 @@ EXPORT_SYMBOL_GPL(nbcon_exit_unsafe);
* When true is returned, @wctxt->ctxt.backlog indicates whether there are
* still records pending in the ringbuffer,
*/
@@ -68,56 +71,65 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
static bool nbcon_emit_next_record(struct nbcon_write_context *wctxt)
{
struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
-@@ -931,6 +929,104 @@ static bool nbcon_emit_next_record(struc
+@@ -938,6 +936,108 @@ static bool nbcon_emit_next_record(struc
}
/**
-+ * nbcon_atomic_emit_one - Print one record for an nbcon console using the
-+ * write_atomic() callback
-+ * @wctxt: An initialized write context struct to use
-+ * for this context
++ * __nbcon_atomic_flush_pending_con - Flush specified nbcon console using its
++ * write_atomic() callback
++ * @con: The nbcon console to flush
++ * @stop_seq: Flush up until this record
+ *
-+ * Return: False if the given console could not print a record or there
-+ * are no more records to print, otherwise true.
++ * Return: True if taken over while printing. Otherwise false.
+ *
-+ * This is an internal helper to handle the locking of the console before
-+ * calling nbcon_emit_next_record().
++ * If flushing up to @stop_seq was not successful, it only makes sense for the
++ * caller to try again when true was returned. When false is returned, either
++ * there are no more records available to read or this context is not allowed
++ * to acquire the console.
+ */
-+static bool nbcon_atomic_emit_one(struct nbcon_write_context *wctxt)
++static bool __nbcon_atomic_flush_pending_con(struct console *con, u64 stop_seq)
+{
-+ struct nbcon_context *ctxt = &ACCESS_PRIVATE(wctxt, ctxt);
++ struct nbcon_write_context wctxt = { };
++ struct nbcon_context *ctxt = &ACCESS_PRIVATE(&wctxt, ctxt);
++
++ ctxt->console = con;
++ ctxt->spinwait_max_us = 2000;
++ ctxt->prio = NBCON_PRIO_NORMAL;
+
+ if (!nbcon_context_try_acquire(ctxt))
+ return false;
+
-+ /*
-+ * nbcon_emit_next_record() returns false when the console was
-+ * handed over or taken over. In both cases the context is no
-+ * longer valid.
-+ */
-+ if (!nbcon_emit_next_record(wctxt))
-+ return false;
++ while (nbcon_seq_read(con) < stop_seq) {
++ /*
++ * nbcon_emit_next_record() returns false when the console was
++ * handed over or taken over. In both cases the context is no
++ * longer valid.
++ */
++ if (!nbcon_emit_next_record(&wctxt))
++ return true;
++
++ if (!ctxt->backlog)
++ break;
++ }
+
+ nbcon_context_release(ctxt);
+
-+ return ctxt->backlog;
++ return false;
+}
+
+/**
-+ * __nbcon_atomic_flush_all - Flush all nbcon consoles using their
++ * __nbcon_atomic_flush_pending - Flush all nbcon consoles using their
+ * write_atomic() callback
+ * @stop_seq: Flush up until this record
+ */
-+static void __nbcon_atomic_flush_all(u64 stop_seq)
++static void __nbcon_atomic_flush_pending(u64 stop_seq)
+{
-+ struct nbcon_write_context wctxt = { };
-+ struct nbcon_context *ctxt = &ACCESS_PRIVATE(&wctxt, ctxt);
+ struct console *con;
-+ bool any_progress;
++ bool should_retry;
+ int cookie;
+
+ do {
-+ any_progress = false;
++ should_retry = false;
+
+ cookie = console_srcu_read_lock();
+ for_each_console_srcu(con) {
@@ -133,40 +145,35 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
+ if (nbcon_seq_read(con) >= stop_seq)
+ continue;
+
-+ memset(ctxt, 0, sizeof(*ctxt));
-+ ctxt->console = con;
-+ ctxt->spinwait_max_us = 2000;
-+ ctxt->prio = NBCON_PRIO_NORMAL;
-+
+ /*
+ * Atomic flushing does not use console driver
+ * synchronization (i.e. it does not hold the port
+ * lock for uart consoles). Therefore IRQs must be
+ * disabled to avoid being interrupted and then
+ * calling into a driver that will deadlock trying
-+ * acquire console ownership.
++ * to acquire console ownership.
+ */
+ local_irq_save(irq_flags);
+
-+ any_progress |= nbcon_atomic_emit_one(&wctxt);
++ should_retry |= __nbcon_atomic_flush_pending_con(con, stop_seq);
+
+ local_irq_restore(irq_flags);
+ }
+ console_srcu_read_unlock(cookie);
-+ } while (any_progress);
++ } while (should_retry);
+}
+
+/**
-+ * nbcon_atomic_flush_all - Flush all nbcon consoles using their
++ * nbcon_atomic_flush_pending - Flush all nbcon consoles using their
+ * write_atomic() callback
+ *
+ * Flush the backlog up through the currently newest record. Any new
+ * records added while flushing will not be flushed. This is to avoid
+ * one CPU printing unbounded because other CPUs continue to add records.
+ */
-+void nbcon_atomic_flush_all(void)
++void nbcon_atomic_flush_pending(void)
+{
-+ __nbcon_atomic_flush_all(prb_next_reserve_seq(prb));
++ __nbcon_atomic_flush_pending(prb_next_reserve_seq(prb));
+}
+
+/**
@@ -179,7 +186,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
console_srcu_read_unlock(cookie);
}
-+ nbcon_atomic_flush_all();
++ nbcon_atomic_flush_pending();
+
console_flush_all(false, &next_seq, &handover);
}