summaryrefslogtreecommitdiffstats
path: root/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch
diff options
context:
space:
mode:
Diffstat (limited to 'debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch')
-rw-r--r--debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch116
1 files changed, 58 insertions, 58 deletions
diff --git a/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch b/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch
index 8cd083750..8b89a55ed 100644
--- a/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch
+++ b/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch
@@ -1,11 +1,11 @@
-From 1ab76ef52a7db4158a5d86604c2a318cfe847e9d Mon Sep 17 00:00:00 2001
+From a740baf29593f6257f2b1790a1f659514167e2da Mon Sep 17 00:00:00 2001
From: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Date: Wed, 3 Feb 2021 11:05:23 -0500
Subject: [PATCH 076/323] tracing: Merge irqflags + preempt counter.
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8
Content-Transfer-Encoding: 8bit
-Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.204-rt100.tar.xz
+Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.215-rt107.tar.xz
The state of the interrupts (irqflags) and the preemption counter are
both passed down to tracing_generic_entry_update(). Only one bit of
@@ -79,7 +79,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
17 files changed, 286 insertions(+), 307 deletions(-)
diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h
-index f7ed0471d5a8..2a98c40526a0 100644
+index f7ed0471d5a85..2a98c40526a0a 100644
--- a/include/linux/trace_events.h
+++ b/include/linux/trace_events.h
@@ -152,17 +152,29 @@ enum print_line_t {
@@ -128,7 +128,7 @@ index f7ed0471d5a8..2a98c40526a0 100644
};
diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c
-index ab912cc60760..a95a2027eefd 100644
+index ab912cc60760a..a95a2027eefda 100644
--- a/kernel/trace/blktrace.c
+++ b/kernel/trace/blktrace.c
@@ -72,17 +72,17 @@ static void trace_note(struct blk_trace *bt, pid_t pid, int action,
@@ -195,10 +195,10 @@ index ab912cc60760..a95a2027eefd 100644
}
}
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
-index 4e0411b19ef9..376eb8a1c913 100644
+index 22e1e57118698..348cbbb9a39b5 100644
--- a/kernel/trace/trace.c
+++ b/kernel/trace/trace.c
-@@ -176,7 +176,7 @@ static union trace_eval_map_item *trace_eval_maps;
+@@ -177,7 +177,7 @@ static union trace_eval_map_item *trace_eval_maps;
int tracing_set_tracer(struct trace_array *tr, const char *buf);
static void ftrace_trace_userstack(struct trace_array *tr,
struct trace_buffer *buffer,
@@ -207,7 +207,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
#define MAX_TRACER_SIZE 100
static char bootup_tracer_buf[MAX_TRACER_SIZE] __initdata;
-@@ -909,23 +909,23 @@ static inline void trace_access_lock_init(void)
+@@ -910,23 +910,23 @@ static inline void trace_access_lock_init(void)
#ifdef CONFIG_STACKTRACE
static void __ftrace_trace_stack(struct trace_buffer *buffer,
@@ -239,7 +239,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
}
-@@ -933,24 +933,24 @@ static inline void ftrace_trace_stack(struct trace_array *tr,
+@@ -934,24 +934,24 @@ static inline void ftrace_trace_stack(struct trace_array *tr,
static __always_inline void
trace_event_setup(struct ring_buffer_event *event,
@@ -268,7 +268,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
return event;
}
-@@ -1011,25 +1011,22 @@ int __trace_puts(unsigned long ip, const char *str, int size)
+@@ -1012,25 +1012,22 @@ int __trace_puts(unsigned long ip, const char *str, int size)
struct ring_buffer_event *event;
struct trace_buffer *buffer;
struct print_entry *entry;
@@ -298,7 +298,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event) {
size = 0;
goto out;
-@@ -1048,7 +1045,7 @@ int __trace_puts(unsigned long ip, const char *str, int size)
+@@ -1049,7 +1046,7 @@ int __trace_puts(unsigned long ip, const char *str, int size)
entry->buf[size] = '\0';
__buffer_unlock_commit(buffer, event);
@@ -307,7 +307,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
out:
ring_buffer_nest_end(buffer);
return size;
-@@ -1065,25 +1062,22 @@ int __trace_bputs(unsigned long ip, const char *str)
+@@ -1066,25 +1063,22 @@ int __trace_bputs(unsigned long ip, const char *str)
struct ring_buffer_event *event;
struct trace_buffer *buffer;
struct bputs_entry *entry;
@@ -336,7 +336,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
goto out;
-@@ -1092,7 +1086,7 @@ int __trace_bputs(unsigned long ip, const char *str)
+@@ -1093,7 +1087,7 @@ int __trace_bputs(unsigned long ip, const char *str)
entry->str = str;
__buffer_unlock_commit(buffer, event);
@@ -345,7 +345,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
ret = 1;
out:
-@@ -2581,36 +2575,69 @@ enum print_line_t trace_handle_return(struct trace_seq *s)
+@@ -2609,36 +2603,69 @@ enum print_line_t trace_handle_return(struct trace_seq *s)
}
EXPORT_SYMBOL_GPL(trace_handle_return);
@@ -433,7 +433,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
DEFINE_PER_CPU(struct ring_buffer_event *, trace_buffered_event);
-@@ -2729,7 +2756,7 @@ struct ring_buffer_event *
+@@ -2757,7 +2784,7 @@ struct ring_buffer_event *
trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
struct trace_event_file *trace_file,
int type, unsigned long len,
@@ -442,7 +442,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
struct ring_buffer_event *entry;
int val;
-@@ -2742,7 +2769,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
+@@ -2770,7 +2797,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
/* Try to use the per cpu buffer first */
val = this_cpu_inc_return(trace_buffered_event_cnt);
if ((len < (PAGE_SIZE - sizeof(*entry) - sizeof(entry->array[0]))) && val == 1) {
@@ -451,7 +451,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
entry->array[0] = len;
return entry;
}
-@@ -2750,7 +2777,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
+@@ -2778,7 +2805,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
}
entry = __trace_buffer_lock_reserve(*current_rb,
@@ -460,7 +460,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
/*
* If tracing is off, but we have triggers enabled
* we still need to look at the event data. Use the temp_buffer
-@@ -2759,8 +2786,8 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
+@@ -2787,8 +2814,8 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
*/
if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) {
*current_rb = temp_buffer;
@@ -471,7 +471,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
return entry;
}
-@@ -2846,7 +2873,7 @@ void trace_event_buffer_commit(struct trace_event_buffer *fbuffer)
+@@ -2874,7 +2901,7 @@ void trace_event_buffer_commit(struct trace_event_buffer *fbuffer)
ftrace_exports(fbuffer->event, TRACE_EXPORT_EVENT);
event_trigger_unlock_commit_regs(fbuffer->trace_file, fbuffer->buffer,
fbuffer->event, fbuffer->entry,
@@ -480,7 +480,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
EXPORT_SYMBOL_GPL(trace_event_buffer_commit);
-@@ -2862,7 +2889,7 @@ EXPORT_SYMBOL_GPL(trace_event_buffer_commit);
+@@ -2890,7 +2917,7 @@ EXPORT_SYMBOL_GPL(trace_event_buffer_commit);
void trace_buffer_unlock_commit_regs(struct trace_array *tr,
struct trace_buffer *buffer,
struct ring_buffer_event *event,
@@ -489,7 +489,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
struct pt_regs *regs)
{
__buffer_unlock_commit(buffer, event);
-@@ -2873,8 +2900,8 @@ void trace_buffer_unlock_commit_regs(struct trace_array *tr,
+@@ -2901,8 +2928,8 @@ void trace_buffer_unlock_commit_regs(struct trace_array *tr,
* and mmiotrace, but that's ok if they lose a function or
* two. They are not that meaningful.
*/
@@ -500,7 +500,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
/*
-@@ -2888,9 +2915,8 @@ trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer,
+@@ -2916,9 +2943,8 @@ trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer,
}
void
@@ -512,7 +512,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
struct trace_event_call *call = &event_function;
struct trace_buffer *buffer = tr->array_buffer.buffer;
-@@ -2898,7 +2924,7 @@ trace_function(struct trace_array *tr,
+@@ -2926,7 +2952,7 @@ trace_function(struct trace_array *tr,
struct ftrace_entry *entry;
event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry),
@@ -521,7 +521,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
return;
entry = ring_buffer_event_data(event);
-@@ -2932,8 +2958,8 @@ static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks);
+@@ -2960,8 +2986,8 @@ static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks);
static DEFINE_PER_CPU(int, ftrace_stack_reserve);
static void __ftrace_trace_stack(struct trace_buffer *buffer,
@@ -532,7 +532,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
struct trace_event_call *call = &event_kernel_stack;
struct ring_buffer_event *event;
-@@ -2981,7 +3007,7 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer,
+@@ -3009,7 +3035,7 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer,
size = nr_entries * sizeof(unsigned long);
event = __trace_buffer_lock_reserve(buffer, TRACE_STACK,
(sizeof(*entry) - sizeof(entry->caller)) + size,
@@ -541,7 +541,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
goto out;
entry = ring_buffer_event_data(event);
-@@ -3002,22 +3028,22 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer,
+@@ -3030,22 +3056,22 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer,
static inline void ftrace_trace_stack(struct trace_array *tr,
struct trace_buffer *buffer,
@@ -570,7 +570,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
return;
}
-@@ -3031,7 +3057,7 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip,
+@@ -3059,7 +3085,7 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip,
return;
rcu_irq_enter_irqson();
@@ -579,7 +579,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
rcu_irq_exit_irqson();
}
-@@ -3041,19 +3067,15 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip,
+@@ -3069,19 +3095,15 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip,
*/
void trace_dump_stack(int skip)
{
@@ -600,7 +600,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
EXPORT_SYMBOL_GPL(trace_dump_stack);
-@@ -3062,7 +3084,7 @@ static DEFINE_PER_CPU(int, user_stack_count);
+@@ -3090,7 +3112,7 @@ static DEFINE_PER_CPU(int, user_stack_count);
static void
ftrace_trace_userstack(struct trace_array *tr,
@@ -609,7 +609,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
struct trace_event_call *call = &event_user_stack;
struct ring_buffer_event *event;
-@@ -3089,7 +3111,7 @@ ftrace_trace_userstack(struct trace_array *tr,
+@@ -3117,7 +3139,7 @@ ftrace_trace_userstack(struct trace_array *tr,
__this_cpu_inc(user_stack_count);
event = __trace_buffer_lock_reserve(buffer, TRACE_USER_STACK,
@@ -618,7 +618,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
goto out_drop_count;
entry = ring_buffer_event_data(event);
-@@ -3109,7 +3131,7 @@ ftrace_trace_userstack(struct trace_array *tr,
+@@ -3137,7 +3159,7 @@ ftrace_trace_userstack(struct trace_array *tr,
#else /* CONFIG_USER_STACKTRACE_SUPPORT */
static void ftrace_trace_userstack(struct trace_array *tr,
struct trace_buffer *buffer,
@@ -627,7 +627,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
{
}
#endif /* !CONFIG_USER_STACKTRACE_SUPPORT */
-@@ -3239,9 +3261,9 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
+@@ -3267,9 +3289,9 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
struct trace_buffer *buffer;
struct trace_array *tr = &global_trace;
struct bprint_entry *entry;
@@ -639,7 +639,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (unlikely(tracing_selftest_running || tracing_disabled))
return 0;
-@@ -3249,7 +3271,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
+@@ -3277,7 +3299,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
/* Don't pollute graph traces with trace_vprintk internals */
pause_graph_tracing();
@@ -648,7 +648,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
preempt_disable_notrace();
tbuffer = get_trace_buf();
-@@ -3263,12 +3285,11 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
+@@ -3291,12 +3313,11 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
if (len > TRACE_BUF_SIZE/sizeof(int) || len < 0)
goto out_put;
@@ -662,7 +662,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
goto out;
entry = ring_buffer_event_data(event);
-@@ -3278,7 +3299,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
+@@ -3306,7 +3327,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
memcpy(entry->buf, tbuffer, sizeof(u32) * len);
if (!call_filter_check_discard(call, entry, buffer, event)) {
__buffer_unlock_commit(buffer, event);
@@ -671,7 +671,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
out:
-@@ -3301,9 +3322,9 @@ __trace_array_vprintk(struct trace_buffer *buffer,
+@@ -3329,9 +3350,9 @@ __trace_array_vprintk(struct trace_buffer *buffer,
{
struct trace_event_call *call = &event_print;
struct ring_buffer_event *event;
@@ -683,7 +683,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
char *tbuffer;
if (tracing_disabled || tracing_selftest_running)
-@@ -3312,7 +3333,7 @@ __trace_array_vprintk(struct trace_buffer *buffer,
+@@ -3340,7 +3361,7 @@ __trace_array_vprintk(struct trace_buffer *buffer,
/* Don't pollute graph traces with trace_vprintk internals */
pause_graph_tracing();
@@ -692,7 +692,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
preempt_disable_notrace();
-@@ -3324,11 +3345,10 @@ __trace_array_vprintk(struct trace_buffer *buffer,
+@@ -3352,11 +3373,10 @@ __trace_array_vprintk(struct trace_buffer *buffer,
len = vscnprintf(tbuffer, TRACE_BUF_SIZE, fmt, args);
@@ -705,7 +705,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (!event)
goto out;
entry = ring_buffer_event_data(event);
-@@ -3337,7 +3357,7 @@ __trace_array_vprintk(struct trace_buffer *buffer,
+@@ -3365,7 +3385,7 @@ __trace_array_vprintk(struct trace_buffer *buffer,
memcpy(&entry->buf, tbuffer, len + 1);
if (!call_filter_check_discard(call, entry, buffer, event)) {
__buffer_unlock_commit(buffer, event);
@@ -714,7 +714,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
}
out:
-@@ -6831,7 +6851,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
+@@ -6851,7 +6871,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
enum event_trigger_type tt = ETT_NONE;
struct trace_buffer *buffer;
struct print_entry *entry;
@@ -722,7 +722,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
ssize_t written;
int size;
int len;
-@@ -6851,7 +6870,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
+@@ -6871,7 +6890,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE);
@@ -730,7 +730,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */
/* If less than "<faulted>", then make sure we can still add that */
-@@ -6860,7 +6878,7 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
+@@ -6880,7 +6898,7 @@ tracing_mark_write(struct file *filp, const char __user *ubuf,
buffer = tr->array_buffer.buffer;
event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size,
@@ -739,7 +739,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
if (unlikely(!event))
/* Ring buffer disabled, return as if not open for write */
return -EBADF;
-@@ -6912,7 +6930,6 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf,
+@@ -6932,7 +6950,6 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf,
struct ring_buffer_event *event;
struct trace_buffer *buffer;
struct raw_data_entry *entry;
@@ -747,7 +747,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
ssize_t written;
int size;
int len;
-@@ -6934,14 +6951,13 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf,
+@@ -6954,14 +6971,13 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf,
BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE);
@@ -764,7 +764,7 @@ index 4e0411b19ef9..376eb8a1c913 100644
/* Ring buffer disabled, return as if not open for write */
return -EBADF;
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
-index 7c90872f2435..27ed42bccd7f 100644
+index 7c90872f2435d..27ed42bccd7f7 100644
--- a/kernel/trace/trace.h
+++ b/kernel/trace/trace.h
@@ -750,8 +750,7 @@ struct ring_buffer_event *
@@ -891,7 +891,7 @@ index 7c90872f2435..27ed42bccd7f 100644
if (tt)
event_triggers_post_call(file, tt);
diff --git a/kernel/trace/trace_branch.c b/kernel/trace/trace_branch.c
-index eff099123aa2..e47fdb4c92fb 100644
+index eff099123aa27..e47fdb4c92fbc 100644
--- a/kernel/trace/trace_branch.c
+++ b/kernel/trace/trace_branch.c
@@ -37,7 +37,7 @@ probe_likely_condition(struct ftrace_likely_data *f, int val, int expect)
@@ -917,7 +917,7 @@ index eff099123aa2..e47fdb4c92fb 100644
goto out;
diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c
-index 643e0b19920d..0443dd61667b 100644
+index 643e0b19920d2..0443dd61667b9 100644
--- a/kernel/trace/trace_event_perf.c
+++ b/kernel/trace/trace_event_perf.c
@@ -421,11 +421,8 @@ NOKPROBE_SYMBOL(perf_trace_buf_alloc);
@@ -934,7 +934,7 @@ index 643e0b19920d..0443dd61667b 100644
NOKPROBE_SYMBOL(perf_trace_buf_update);
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c
-index 4b5a8d7275be..df64b92c5edc 100644
+index 4b5a8d7275be7..df64b92c5edc3 100644
--- a/kernel/trace/trace_events.c
+++ b/kernel/trace/trace_events.c
@@ -259,22 +259,19 @@ void *trace_event_buffer_reserve(struct trace_event_buffer *fbuffer,
@@ -1000,7 +1000,7 @@ index 4b5a8d7275be..df64b92c5edc 100644
atomic_dec(&per_cpu(ftrace_test_event_disable, cpu));
preempt_enable_notrace();
diff --git a/kernel/trace/trace_events_inject.c b/kernel/trace/trace_events_inject.c
-index 149c7dc6a447..b1fce64e126c 100644
+index 149c7dc6a4473..b1fce64e126c0 100644
--- a/kernel/trace/trace_events_inject.c
+++ b/kernel/trace/trace_events_inject.c
@@ -192,7 +192,6 @@ static void *trace_alloc_entry(struct trace_event_call *call, int *size)
@@ -1024,7 +1024,7 @@ index 149c7dc6a447..b1fce64e126c 100644
while ((len = parse_field(str, call, &field, &val)) > 0) {
if (is_function_field(field))
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c
-index 93e20ed642e5..8606cb73341e 100644
+index 93e20ed642e53..8606cb73341ed 100644
--- a/kernel/trace/trace_functions.c
+++ b/kernel/trace/trace_functions.c
@@ -133,15 +133,14 @@ function_trace_call(unsigned long ip, unsigned long parent_ip,
@@ -1099,7 +1099,7 @@ index 93e20ed642e5..8606cb73341e 100644
static void
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
-index 60d66278aa0d..b086ba8bb3d6 100644
+index 60d66278aa0d8..b086ba8bb3d65 100644
--- a/kernel/trace/trace_functions_graph.c
+++ b/kernel/trace/trace_functions_graph.c
@@ -96,8 +96,7 @@ print_graph_duration(struct trace_array *tr, unsigned long long duration,
@@ -1213,7 +1213,7 @@ index 60d66278aa0d..b086ba8bb3d6 100644
atomic_dec(&data->disabled);
local_irq_restore(flags);
diff --git a/kernel/trace/trace_hwlat.c b/kernel/trace/trace_hwlat.c
-index d071fc271eef..4c01c5d8b9a7 100644
+index d071fc271eef7..4c01c5d8b9a7a 100644
--- a/kernel/trace/trace_hwlat.c
+++ b/kernel/trace/trace_hwlat.c
@@ -108,14 +108,9 @@ static void trace_hwlat_sample(struct hwlat_sample *sample)
@@ -1233,7 +1233,7 @@ index d071fc271eef..4c01c5d8b9a7 100644
return;
entry = ring_buffer_event_data(event);
diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c
-index 619a60944bb6..4a11967c8daa 100644
+index 619a60944bb6d..4a11967c8daa1 100644
--- a/kernel/trace/trace_irqsoff.c
+++ b/kernel/trace/trace_irqsoff.c
@@ -143,11 +143,14 @@ irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip,
@@ -1467,7 +1467,7 @@ index 619a60944bb6..4a11967c8daa 100644
static int preemptoff_tracer_init(struct trace_array *tr)
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c
-index 718357289899..a2478605e761 100644
+index 7183572898998..a2478605e7617 100644
--- a/kernel/trace/trace_kprobe.c
+++ b/kernel/trace/trace_kprobe.c
@@ -1394,8 +1394,7 @@ __kprobe_trace_func(struct trace_kprobe *tk, struct pt_regs *regs,
@@ -1509,7 +1509,7 @@ index 718357289899..a2478605e761 100644
return;
diff --git a/kernel/trace/trace_mmiotrace.c b/kernel/trace/trace_mmiotrace.c
-index 84582bf1ed5f..7221ae0b4c47 100644
+index 84582bf1ed5fb..7221ae0b4c477 100644
--- a/kernel/trace/trace_mmiotrace.c
+++ b/kernel/trace/trace_mmiotrace.c
@@ -300,10 +300,11 @@ static void __trace_mmiotrace_rw(struct trace_array *tr,
@@ -1559,7 +1559,7 @@ index 84582bf1ed5f..7221ae0b4c47 100644
void mmio_trace_mapping(struct mmiotrace_map *map)
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c
-index 037e1e863b17..c1f582e0e97f 100644
+index 037e1e863b17f..c1f582e0e97f0 100644
--- a/kernel/trace/trace_sched_wakeup.c
+++ b/kernel/trace/trace_sched_wakeup.c
@@ -67,7 +67,7 @@ static bool function_enabled;
@@ -1800,7 +1800,7 @@ index 037e1e863b17..c1f582e0e97f 100644
out_locked:
arch_spin_unlock(&wakeup_lock);
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c
-index d85a2f0f316b..8bfcd3b09422 100644
+index d85a2f0f316b3..8bfcd3b094226 100644
--- a/kernel/trace/trace_syscalls.c
+++ b/kernel/trace/trace_syscalls.c
@@ -298,9 +298,8 @@ static void ftrace_syscall_enter(void *data, struct pt_regs *regs, long id)
@@ -1874,7 +1874,7 @@ index d85a2f0f316b..8bfcd3b09422 100644
static int reg_event_syscall_enter(struct trace_event_file *file,
diff --git a/kernel/trace/trace_uprobe.c b/kernel/trace/trace_uprobe.c
-index 60ff36f5d7f9..0b07bb07127d 100644
+index 60ff36f5d7f9e..0b07bb07127d8 100644
--- a/kernel/trace/trace_uprobe.c
+++ b/kernel/trace/trace_uprobe.c
@@ -966,7 +966,7 @@ static void __uprobe_trace_func(struct trace_uprobe *tu,
@@ -1896,5 +1896,5 @@ index 60ff36f5d7f9..0b07bb07127d 100644
/* uprobe handler */
--
-2.43.0
+2.44.0