diff options
Diffstat (limited to 'debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch')
-rw-r--r-- | debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch | 116 |
1 files changed, 58 insertions, 58 deletions
diff --git a/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch b/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch index 8cd083750..8b89a55ed 100644 --- a/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch +++ b/debian/patches-rt/0076-tracing-Merge-irqflags-preempt-counter.patch @@ -1,11 +1,11 @@ -From 1ab76ef52a7db4158a5d86604c2a318cfe847e9d Mon Sep 17 00:00:00 2001 +From a740baf29593f6257f2b1790a1f659514167e2da Mon Sep 17 00:00:00 2001 From: Sebastian Andrzej Siewior <bigeasy@linutronix.de> Date: Wed, 3 Feb 2021 11:05:23 -0500 Subject: [PATCH 076/323] tracing: Merge irqflags + preempt counter. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit -Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.204-rt100.tar.xz +Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.215-rt107.tar.xz The state of the interrupts (irqflags) and the preemption counter are both passed down to tracing_generic_entry_update(). Only one bit of @@ -79,7 +79,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de> 17 files changed, 286 insertions(+), 307 deletions(-) diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h -index f7ed0471d5a8..2a98c40526a0 100644 +index f7ed0471d5a85..2a98c40526a0a 100644 --- a/include/linux/trace_events.h +++ b/include/linux/trace_events.h @@ -152,17 +152,29 @@ enum print_line_t { @@ -128,7 +128,7 @@ index f7ed0471d5a8..2a98c40526a0 100644 }; diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c -index ab912cc60760..a95a2027eefd 100644 +index ab912cc60760a..a95a2027eefda 100644 --- a/kernel/trace/blktrace.c +++ b/kernel/trace/blktrace.c @@ -72,17 +72,17 @@ static void trace_note(struct blk_trace *bt, pid_t pid, int action, @@ -195,10 +195,10 @@ index ab912cc60760..a95a2027eefd 100644 } } diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c -index 4e0411b19ef9..376eb8a1c913 100644 +index 22e1e57118698..348cbbb9a39b5 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c -@@ -176,7 +176,7 @@ static union trace_eval_map_item *trace_eval_maps; +@@ -177,7 +177,7 @@ static union trace_eval_map_item *trace_eval_maps; int tracing_set_tracer(struct trace_array *tr, const char *buf); static void ftrace_trace_userstack(struct trace_array *tr, struct trace_buffer *buffer, @@ -207,7 +207,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 #define MAX_TRACER_SIZE 100 static char bootup_tracer_buf[MAX_TRACER_SIZE] __initdata; -@@ -909,23 +909,23 @@ static inline void trace_access_lock_init(void) +@@ -910,23 +910,23 @@ static inline void trace_access_lock_init(void) #ifdef CONFIG_STACKTRACE static void __ftrace_trace_stack(struct trace_buffer *buffer, @@ -239,7 +239,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { } -@@ -933,24 +933,24 @@ static inline void ftrace_trace_stack(struct trace_array *tr, +@@ -934,24 +934,24 @@ static inline void ftrace_trace_stack(struct trace_array *tr, static __always_inline void trace_event_setup(struct ring_buffer_event *event, @@ -268,7 +268,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 return event; } -@@ -1011,25 +1011,22 @@ int __trace_puts(unsigned long ip, const char *str, int size) +@@ -1012,25 +1012,22 @@ int __trace_puts(unsigned long ip, const char *str, int size) struct ring_buffer_event *event; struct trace_buffer *buffer; struct print_entry *entry; @@ -298,7 +298,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) { size = 0; goto out; -@@ -1048,7 +1045,7 @@ int __trace_puts(unsigned long ip, const char *str, int size) +@@ -1049,7 +1046,7 @@ int __trace_puts(unsigned long ip, const char *str, int size) entry->buf[size] = '\0'; __buffer_unlock_commit(buffer, event); @@ -307,7 +307,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 out: ring_buffer_nest_end(buffer); return size; -@@ -1065,25 +1062,22 @@ int __trace_bputs(unsigned long ip, const char *str) +@@ -1066,25 +1063,22 @@ int __trace_bputs(unsigned long ip, const char *str) struct ring_buffer_event *event; struct trace_buffer *buffer; struct bputs_entry *entry; @@ -336,7 +336,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) goto out; -@@ -1092,7 +1086,7 @@ int __trace_bputs(unsigned long ip, const char *str) +@@ -1093,7 +1087,7 @@ int __trace_bputs(unsigned long ip, const char *str) entry->str = str; __buffer_unlock_commit(buffer, event); @@ -345,7 +345,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 ret = 1; out: -@@ -2581,36 +2575,69 @@ enum print_line_t trace_handle_return(struct trace_seq *s) +@@ -2609,36 +2603,69 @@ enum print_line_t trace_handle_return(struct trace_seq *s) } EXPORT_SYMBOL_GPL(trace_handle_return); @@ -433,7 +433,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } DEFINE_PER_CPU(struct ring_buffer_event *, trace_buffered_event); -@@ -2729,7 +2756,7 @@ struct ring_buffer_event * +@@ -2757,7 +2784,7 @@ struct ring_buffer_event * trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, struct trace_event_file *trace_file, int type, unsigned long len, @@ -442,7 +442,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { struct ring_buffer_event *entry; int val; -@@ -2742,7 +2769,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, +@@ -2770,7 +2797,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, /* Try to use the per cpu buffer first */ val = this_cpu_inc_return(trace_buffered_event_cnt); if ((len < (PAGE_SIZE - sizeof(*entry) - sizeof(entry->array[0]))) && val == 1) { @@ -451,7 +451,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 entry->array[0] = len; return entry; } -@@ -2750,7 +2777,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, +@@ -2778,7 +2805,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, } entry = __trace_buffer_lock_reserve(*current_rb, @@ -460,7 +460,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 /* * If tracing is off, but we have triggers enabled * we still need to look at the event data. Use the temp_buffer -@@ -2759,8 +2786,8 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, +@@ -2787,8 +2814,8 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, */ if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) { *current_rb = temp_buffer; @@ -471,7 +471,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } return entry; } -@@ -2846,7 +2873,7 @@ void trace_event_buffer_commit(struct trace_event_buffer *fbuffer) +@@ -2874,7 +2901,7 @@ void trace_event_buffer_commit(struct trace_event_buffer *fbuffer) ftrace_exports(fbuffer->event, TRACE_EXPORT_EVENT); event_trigger_unlock_commit_regs(fbuffer->trace_file, fbuffer->buffer, fbuffer->event, fbuffer->entry, @@ -480,7 +480,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } EXPORT_SYMBOL_GPL(trace_event_buffer_commit); -@@ -2862,7 +2889,7 @@ EXPORT_SYMBOL_GPL(trace_event_buffer_commit); +@@ -2890,7 +2917,7 @@ EXPORT_SYMBOL_GPL(trace_event_buffer_commit); void trace_buffer_unlock_commit_regs(struct trace_array *tr, struct trace_buffer *buffer, struct ring_buffer_event *event, @@ -489,7 +489,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 struct pt_regs *regs) { __buffer_unlock_commit(buffer, event); -@@ -2873,8 +2900,8 @@ void trace_buffer_unlock_commit_regs(struct trace_array *tr, +@@ -2901,8 +2928,8 @@ void trace_buffer_unlock_commit_regs(struct trace_array *tr, * and mmiotrace, but that's ok if they lose a function or * two. They are not that meaningful. */ @@ -500,7 +500,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } /* -@@ -2888,9 +2915,8 @@ trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer, +@@ -2916,9 +2943,8 @@ trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer, } void @@ -512,7 +512,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { struct trace_event_call *call = &event_function; struct trace_buffer *buffer = tr->array_buffer.buffer; -@@ -2898,7 +2924,7 @@ trace_function(struct trace_array *tr, +@@ -2926,7 +2952,7 @@ trace_function(struct trace_array *tr, struct ftrace_entry *entry; event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry), @@ -521,7 +521,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) return; entry = ring_buffer_event_data(event); -@@ -2932,8 +2958,8 @@ static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks); +@@ -2960,8 +2986,8 @@ static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks); static DEFINE_PER_CPU(int, ftrace_stack_reserve); static void __ftrace_trace_stack(struct trace_buffer *buffer, @@ -532,7 +532,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { struct trace_event_call *call = &event_kernel_stack; struct ring_buffer_event *event; -@@ -2981,7 +3007,7 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer, +@@ -3009,7 +3035,7 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer, size = nr_entries * sizeof(unsigned long); event = __trace_buffer_lock_reserve(buffer, TRACE_STACK, (sizeof(*entry) - sizeof(entry->caller)) + size, @@ -541,7 +541,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) goto out; entry = ring_buffer_event_data(event); -@@ -3002,22 +3028,22 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer, +@@ -3030,22 +3056,22 @@ static void __ftrace_trace_stack(struct trace_buffer *buffer, static inline void ftrace_trace_stack(struct trace_array *tr, struct trace_buffer *buffer, @@ -570,7 +570,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 return; } -@@ -3031,7 +3057,7 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, +@@ -3059,7 +3085,7 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, return; rcu_irq_enter_irqson(); @@ -579,7 +579,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 rcu_irq_exit_irqson(); } -@@ -3041,19 +3067,15 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, +@@ -3069,19 +3095,15 @@ void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, */ void trace_dump_stack(int skip) { @@ -600,7 +600,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } EXPORT_SYMBOL_GPL(trace_dump_stack); -@@ -3062,7 +3084,7 @@ static DEFINE_PER_CPU(int, user_stack_count); +@@ -3090,7 +3112,7 @@ static DEFINE_PER_CPU(int, user_stack_count); static void ftrace_trace_userstack(struct trace_array *tr, @@ -609,7 +609,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { struct trace_event_call *call = &event_user_stack; struct ring_buffer_event *event; -@@ -3089,7 +3111,7 @@ ftrace_trace_userstack(struct trace_array *tr, +@@ -3117,7 +3139,7 @@ ftrace_trace_userstack(struct trace_array *tr, __this_cpu_inc(user_stack_count); event = __trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, @@ -618,7 +618,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) goto out_drop_count; entry = ring_buffer_event_data(event); -@@ -3109,7 +3131,7 @@ ftrace_trace_userstack(struct trace_array *tr, +@@ -3137,7 +3159,7 @@ ftrace_trace_userstack(struct trace_array *tr, #else /* CONFIG_USER_STACKTRACE_SUPPORT */ static void ftrace_trace_userstack(struct trace_array *tr, struct trace_buffer *buffer, @@ -627,7 +627,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 { } #endif /* !CONFIG_USER_STACKTRACE_SUPPORT */ -@@ -3239,9 +3261,9 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) +@@ -3267,9 +3289,9 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) struct trace_buffer *buffer; struct trace_array *tr = &global_trace; struct bprint_entry *entry; @@ -639,7 +639,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (unlikely(tracing_selftest_running || tracing_disabled)) return 0; -@@ -3249,7 +3271,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) +@@ -3277,7 +3299,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) /* Don't pollute graph traces with trace_vprintk internals */ pause_graph_tracing(); @@ -648,7 +648,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 preempt_disable_notrace(); tbuffer = get_trace_buf(); -@@ -3263,12 +3285,11 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) +@@ -3291,12 +3313,11 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) if (len > TRACE_BUF_SIZE/sizeof(int) || len < 0) goto out_put; @@ -662,7 +662,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) goto out; entry = ring_buffer_event_data(event); -@@ -3278,7 +3299,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) +@@ -3306,7 +3327,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) memcpy(entry->buf, tbuffer, sizeof(u32) * len); if (!call_filter_check_discard(call, entry, buffer, event)) { __buffer_unlock_commit(buffer, event); @@ -671,7 +671,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } out: -@@ -3301,9 +3322,9 @@ __trace_array_vprintk(struct trace_buffer *buffer, +@@ -3329,9 +3350,9 @@ __trace_array_vprintk(struct trace_buffer *buffer, { struct trace_event_call *call = &event_print; struct ring_buffer_event *event; @@ -683,7 +683,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 char *tbuffer; if (tracing_disabled || tracing_selftest_running) -@@ -3312,7 +3333,7 @@ __trace_array_vprintk(struct trace_buffer *buffer, +@@ -3340,7 +3361,7 @@ __trace_array_vprintk(struct trace_buffer *buffer, /* Don't pollute graph traces with trace_vprintk internals */ pause_graph_tracing(); @@ -692,7 +692,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 preempt_disable_notrace(); -@@ -3324,11 +3345,10 @@ __trace_array_vprintk(struct trace_buffer *buffer, +@@ -3352,11 +3373,10 @@ __trace_array_vprintk(struct trace_buffer *buffer, len = vscnprintf(tbuffer, TRACE_BUF_SIZE, fmt, args); @@ -705,7 +705,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (!event) goto out; entry = ring_buffer_event_data(event); -@@ -3337,7 +3357,7 @@ __trace_array_vprintk(struct trace_buffer *buffer, +@@ -3365,7 +3385,7 @@ __trace_array_vprintk(struct trace_buffer *buffer, memcpy(&entry->buf, tbuffer, len + 1); if (!call_filter_check_discard(call, entry, buffer, event)) { __buffer_unlock_commit(buffer, event); @@ -714,7 +714,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 } out: -@@ -6831,7 +6851,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, +@@ -6851,7 +6871,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, enum event_trigger_type tt = ETT_NONE; struct trace_buffer *buffer; struct print_entry *entry; @@ -722,7 +722,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 ssize_t written; int size; int len; -@@ -6851,7 +6870,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, +@@ -6871,7 +6890,6 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE); @@ -730,7 +730,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */ /* If less than "<faulted>", then make sure we can still add that */ -@@ -6860,7 +6878,7 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, +@@ -6880,7 +6898,7 @@ tracing_mark_write(struct file *filp, const char __user *ubuf, buffer = tr->array_buffer.buffer; event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, @@ -739,7 +739,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 if (unlikely(!event)) /* Ring buffer disabled, return as if not open for write */ return -EBADF; -@@ -6912,7 +6930,6 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf, +@@ -6932,7 +6950,6 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf, struct ring_buffer_event *event; struct trace_buffer *buffer; struct raw_data_entry *entry; @@ -747,7 +747,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 ssize_t written; int size; int len; -@@ -6934,14 +6951,13 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf, +@@ -6954,14 +6971,13 @@ tracing_mark_raw_write(struct file *filp, const char __user *ubuf, BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE); @@ -764,7 +764,7 @@ index 4e0411b19ef9..376eb8a1c913 100644 /* Ring buffer disabled, return as if not open for write */ return -EBADF; diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h -index 7c90872f2435..27ed42bccd7f 100644 +index 7c90872f2435d..27ed42bccd7f7 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h @@ -750,8 +750,7 @@ struct ring_buffer_event * @@ -891,7 +891,7 @@ index 7c90872f2435..27ed42bccd7f 100644 if (tt) event_triggers_post_call(file, tt); diff --git a/kernel/trace/trace_branch.c b/kernel/trace/trace_branch.c -index eff099123aa2..e47fdb4c92fb 100644 +index eff099123aa27..e47fdb4c92fbc 100644 --- a/kernel/trace/trace_branch.c +++ b/kernel/trace/trace_branch.c @@ -37,7 +37,7 @@ probe_likely_condition(struct ftrace_likely_data *f, int val, int expect) @@ -917,7 +917,7 @@ index eff099123aa2..e47fdb4c92fb 100644 goto out; diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c -index 643e0b19920d..0443dd61667b 100644 +index 643e0b19920d2..0443dd61667b9 100644 --- a/kernel/trace/trace_event_perf.c +++ b/kernel/trace/trace_event_perf.c @@ -421,11 +421,8 @@ NOKPROBE_SYMBOL(perf_trace_buf_alloc); @@ -934,7 +934,7 @@ index 643e0b19920d..0443dd61667b 100644 NOKPROBE_SYMBOL(perf_trace_buf_update); diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c -index 4b5a8d7275be..df64b92c5edc 100644 +index 4b5a8d7275be7..df64b92c5edc3 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c @@ -259,22 +259,19 @@ void *trace_event_buffer_reserve(struct trace_event_buffer *fbuffer, @@ -1000,7 +1000,7 @@ index 4b5a8d7275be..df64b92c5edc 100644 atomic_dec(&per_cpu(ftrace_test_event_disable, cpu)); preempt_enable_notrace(); diff --git a/kernel/trace/trace_events_inject.c b/kernel/trace/trace_events_inject.c -index 149c7dc6a447..b1fce64e126c 100644 +index 149c7dc6a4473..b1fce64e126c0 100644 --- a/kernel/trace/trace_events_inject.c +++ b/kernel/trace/trace_events_inject.c @@ -192,7 +192,6 @@ static void *trace_alloc_entry(struct trace_event_call *call, int *size) @@ -1024,7 +1024,7 @@ index 149c7dc6a447..b1fce64e126c 100644 while ((len = parse_field(str, call, &field, &val)) > 0) { if (is_function_field(field)) diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c -index 93e20ed642e5..8606cb73341e 100644 +index 93e20ed642e53..8606cb73341ed 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c @@ -133,15 +133,14 @@ function_trace_call(unsigned long ip, unsigned long parent_ip, @@ -1099,7 +1099,7 @@ index 93e20ed642e5..8606cb73341e 100644 static void diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c -index 60d66278aa0d..b086ba8bb3d6 100644 +index 60d66278aa0d8..b086ba8bb3d65 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c @@ -96,8 +96,7 @@ print_graph_duration(struct trace_array *tr, unsigned long long duration, @@ -1213,7 +1213,7 @@ index 60d66278aa0d..b086ba8bb3d6 100644 atomic_dec(&data->disabled); local_irq_restore(flags); diff --git a/kernel/trace/trace_hwlat.c b/kernel/trace/trace_hwlat.c -index d071fc271eef..4c01c5d8b9a7 100644 +index d071fc271eef7..4c01c5d8b9a7a 100644 --- a/kernel/trace/trace_hwlat.c +++ b/kernel/trace/trace_hwlat.c @@ -108,14 +108,9 @@ static void trace_hwlat_sample(struct hwlat_sample *sample) @@ -1233,7 +1233,7 @@ index d071fc271eef..4c01c5d8b9a7 100644 return; entry = ring_buffer_event_data(event); diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c -index 619a60944bb6..4a11967c8daa 100644 +index 619a60944bb6d..4a11967c8daa1 100644 --- a/kernel/trace/trace_irqsoff.c +++ b/kernel/trace/trace_irqsoff.c @@ -143,11 +143,14 @@ irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip, @@ -1467,7 +1467,7 @@ index 619a60944bb6..4a11967c8daa 100644 static int preemptoff_tracer_init(struct trace_array *tr) diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c -index 718357289899..a2478605e761 100644 +index 7183572898998..a2478605e7617 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c @@ -1394,8 +1394,7 @@ __kprobe_trace_func(struct trace_kprobe *tk, struct pt_regs *regs, @@ -1509,7 +1509,7 @@ index 718357289899..a2478605e761 100644 return; diff --git a/kernel/trace/trace_mmiotrace.c b/kernel/trace/trace_mmiotrace.c -index 84582bf1ed5f..7221ae0b4c47 100644 +index 84582bf1ed5fb..7221ae0b4c477 100644 --- a/kernel/trace/trace_mmiotrace.c +++ b/kernel/trace/trace_mmiotrace.c @@ -300,10 +300,11 @@ static void __trace_mmiotrace_rw(struct trace_array *tr, @@ -1559,7 +1559,7 @@ index 84582bf1ed5f..7221ae0b4c47 100644 void mmio_trace_mapping(struct mmiotrace_map *map) diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c -index 037e1e863b17..c1f582e0e97f 100644 +index 037e1e863b17f..c1f582e0e97f0 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c @@ -67,7 +67,7 @@ static bool function_enabled; @@ -1800,7 +1800,7 @@ index 037e1e863b17..c1f582e0e97f 100644 out_locked: arch_spin_unlock(&wakeup_lock); diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c -index d85a2f0f316b..8bfcd3b09422 100644 +index d85a2f0f316b3..8bfcd3b094226 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c @@ -298,9 +298,8 @@ static void ftrace_syscall_enter(void *data, struct pt_regs *regs, long id) @@ -1874,7 +1874,7 @@ index d85a2f0f316b..8bfcd3b09422 100644 static int reg_event_syscall_enter(struct trace_event_file *file, diff --git a/kernel/trace/trace_uprobe.c b/kernel/trace/trace_uprobe.c -index 60ff36f5d7f9..0b07bb07127d 100644 +index 60ff36f5d7f9e..0b07bb07127d8 100644 --- a/kernel/trace/trace_uprobe.c +++ b/kernel/trace/trace_uprobe.c @@ -966,7 +966,7 @@ static void __uprobe_trace_func(struct trace_uprobe *tu, @@ -1896,5 +1896,5 @@ index 60ff36f5d7f9..0b07bb07127d 100644 /* uprobe handler */ -- -2.43.0 +2.44.0 |