aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
authorSteven Rostedt <srostedt@redhat.com>2009-02-18 18:33:57 -0500
committerSteven Rostedt <srostedt@redhat.com>2009-02-18 22:04:01 -0500
commit0c5119c1e655e0719a69601b1049acdd5ec1c125 (patch)
treee808e36e274afc7c6521f69194e6fc2597e189bd /kernel
parent5b058bcde961bf28678a70e44c079107313543b6 (diff)
downloadkernel_samsung_smdk4412-0c5119c1e655e0719a69601b1049acdd5ec1c125.tar.gz
kernel_samsung_smdk4412-0c5119c1e655e0719a69601b1049acdd5ec1c125.tar.bz2
kernel_samsung_smdk4412-0c5119c1e655e0719a69601b1049acdd5ec1c125.zip
tracing: disable tracing while testing ring buffer
Impact: fix to prevent hard lockup on self tests If one of the tracers are broken and is constantly filling the ring buffer while the test of the ring buffer is running, it will hang the box. The reason is that the test is a consumer that will not stop till the ring buffer is empty. But if the tracer is broken and is constantly producing input to the buffer, this test will never end. The result is a lockup of the box. This happened when KALLSYMS was not defined and the dynamic ftrace test constantly filled the ring buffer, because the filter failed and all functions were being traced. Something was being called that constantly filled the buffer. Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/trace/trace_selftest.c9
1 files changed, 9 insertions, 0 deletions
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c
index 88c8eb70f54..a7e0ef662f9 100644
--- a/kernel/trace/trace_selftest.c
+++ b/kernel/trace/trace_selftest.c
@@ -57,11 +57,20 @@ static int trace_test_buffer(struct trace_array *tr, unsigned long *count)
cnt = ring_buffer_entries(tr->buffer);
+ /*
+ * The trace_test_buffer_cpu runs a while loop to consume all data.
+ * If the calling tracer is broken, and is constantly filling
+ * the buffer, this will run forever, and hard lock the box.
+ * We disable the ring buffer while we do this test to prevent
+ * a hard lock up.
+ */
+ tracing_off();
for_each_possible_cpu(cpu) {
ret = trace_test_buffer_cpu(tr, cpu);
if (ret)
break;
}
+ tracing_on();
__raw_spin_unlock(&ftrace_max_lock);
local_irq_restore(flags);