2008-11-12 13:14:39 +08:00
|
|
|
/*
|
|
|
|
* unlikely profiler
|
|
|
|
*
|
|
|
|
* Copyright (C) 2008 Steven Rostedt <srostedt@redhat.com>
|
|
|
|
*/
|
|
|
|
#include <linux/kallsyms.h>
|
|
|
|
#include <linux/seq_file.h>
|
|
|
|
#include <linux/spinlock.h>
|
2008-11-29 11:12:46 +08:00
|
|
|
#include <linux/irqflags.h>
|
2008-11-12 13:14:39 +08:00
|
|
|
#include <linux/uaccess.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/ftrace.h>
|
|
|
|
#include <linux/hash.h>
|
|
|
|
#include <linux/fs.h>
|
|
|
|
#include <asm/local.h>
|
2008-12-24 12:24:13 +08:00
|
|
|
|
2008-11-12 13:14:39 +08:00
|
|
|
#include "trace.h"
|
2009-01-11 03:34:13 +08:00
|
|
|
#include "trace_stat.h"
|
2008-12-24 12:24:13 +08:00
|
|
|
#include "trace_output.h"
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
#ifdef CONFIG_BRANCH_TRACER
|
2008-11-12 13:14:40 +08:00
|
|
|
|
2009-01-11 03:34:13 +08:00
|
|
|
static struct tracer branch_trace;
|
2008-11-13 04:24:24 +08:00
|
|
|
static int branch_tracing_enabled __read_mostly;
|
|
|
|
static DEFINE_MUTEX(branch_tracing_mutex);
|
2008-12-28 06:25:38 +08:00
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
static struct trace_array *branch_tracer;
|
2008-11-12 13:14:40 +08:00
|
|
|
|
|
|
|
static void
|
2017-01-19 21:57:41 +08:00
|
|
|
probe_likely_condition(struct ftrace_likely_data *f, int val, int expect)
|
2008-11-12 13:14:40 +08:00
|
|
|
{
|
2015-05-05 23:45:27 +08:00
|
|
|
struct trace_event_call *call = &event_branch;
|
2008-11-13 04:24:24 +08:00
|
|
|
struct trace_array *tr = branch_tracer;
|
2012-08-07 04:24:11 +08:00
|
|
|
struct trace_array_cpu *data;
|
2008-11-12 13:14:40 +08:00
|
|
|
struct ring_buffer_event *event;
|
2008-11-13 04:24:24 +08:00
|
|
|
struct trace_branch *entry;
|
2009-10-08 09:53:41 +08:00
|
|
|
struct ring_buffer *buffer;
|
2009-02-06 02:12:56 +08:00
|
|
|
unsigned long flags;
|
2015-07-08 03:05:03 +08:00
|
|
|
int pc;
|
2008-11-12 13:14:40 +08:00
|
|
|
const char *p;
|
|
|
|
|
2015-07-08 03:05:03 +08:00
|
|
|
if (current->trace_recursion & TRACE_BRANCH_BIT)
|
|
|
|
return;
|
|
|
|
|
2008-11-12 13:14:40 +08:00
|
|
|
/*
|
|
|
|
* I would love to save just the ftrace_likely_data pointer, but
|
|
|
|
* this code can also be used by modules. Ugly things can happen
|
|
|
|
* if the module is unloaded, and then we go and read the
|
|
|
|
* pointer. This is slower, but much safer.
|
|
|
|
*/
|
|
|
|
|
|
|
|
if (unlikely(!tr))
|
|
|
|
return;
|
|
|
|
|
2015-07-08 03:05:03 +08:00
|
|
|
raw_local_irq_save(flags);
|
|
|
|
current->trace_recursion |= TRACE_BRANCH_BIT;
|
|
|
|
data = this_cpu_ptr(tr->trace_buffer.data);
|
|
|
|
if (atomic_read(&data->disabled))
|
2008-11-12 13:14:40 +08:00
|
|
|
goto out;
|
|
|
|
|
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-06 02:14:13 +08:00
|
|
|
pc = preempt_count();
|
2013-03-08 23:40:07 +08:00
|
|
|
buffer = tr->trace_buffer.buffer;
|
2009-10-08 09:53:41 +08:00
|
|
|
event = trace_buffer_lock_reserve(buffer, TRACE_BRANCH,
|
tracing: Introduce trace_buffer_{lock_reserve,unlock_commit}
Impact: new API
These new functions do what previously was being open coded, reducing
the number of details ftrace plugin writers have to worry about.
It also standardizes the handling of stacktrace, userstacktrace and
other trace options we may introduce in the future.
With this patch, for instance, the blk tracer (and some others already
in the tree) can use the "userstacktrace" /d/tracing/trace_options
facility.
$ codiff /tmp/vmlinux.before /tmp/vmlinux.after
linux-2.6-tip/kernel/trace/trace.c:
trace_vprintk | -5
trace_graph_return | -22
trace_graph_entry | -26
trace_function | -45
__ftrace_trace_stack | -27
ftrace_trace_userstack | -29
tracing_sched_switch_trace | -66
tracing_stop | +1
trace_seq_to_user | -1
ftrace_trace_special | -63
ftrace_special | +1
tracing_sched_wakeup_trace | -70
tracing_reset_online_cpus | -1
13 functions changed, 2 bytes added, 355 bytes removed, diff: -353
linux-2.6-tip/block/blktrace.c:
__blk_add_trace | -58
1 function changed, 58 bytes removed, diff: -58
linux-2.6-tip/kernel/trace/trace.c:
trace_buffer_lock_reserve | +88
trace_buffer_unlock_commit | +86
2 functions changed, 174 bytes added, diff: +174
/tmp/vmlinux.after:
16 functions changed, 176 bytes added, 413 bytes removed, diff: -237
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Acked-by: Frédéric Weisbecker <fweisbec@gmail.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-02-06 02:14:13 +08:00
|
|
|
sizeof(*entry), flags, pc);
|
2008-11-12 13:14:40 +08:00
|
|
|
if (!event)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
entry = ring_buffer_event_data(event);
|
|
|
|
|
|
|
|
/* Strip off the path, only save the file */
|
2017-01-19 21:57:41 +08:00
|
|
|
p = f->data.file + strlen(f->data.file);
|
|
|
|
while (p >= f->data.file && *p != '/')
|
2008-11-12 13:14:40 +08:00
|
|
|
p--;
|
|
|
|
p++;
|
|
|
|
|
2017-01-19 21:57:41 +08:00
|
|
|
strncpy(entry->func, f->data.func, TRACE_FUNC_SIZE);
|
2008-11-12 13:14:40 +08:00
|
|
|
strncpy(entry->file, p, TRACE_FILE_SIZE);
|
|
|
|
entry->func[TRACE_FUNC_SIZE] = 0;
|
|
|
|
entry->file[TRACE_FILE_SIZE] = 0;
|
2017-01-19 21:57:41 +08:00
|
|
|
entry->constant = f->constant;
|
|
|
|
entry->line = f->data.line;
|
2008-11-12 13:14:40 +08:00
|
|
|
entry->correct = val == expect;
|
|
|
|
|
2013-10-24 21:34:17 +08:00
|
|
|
if (!call_filter_check_discard(call, entry, buffer, event))
|
2016-11-24 09:28:38 +08:00
|
|
|
trace_buffer_unlock_commit_nostack(buffer, event);
|
2008-11-12 13:14:40 +08:00
|
|
|
|
|
|
|
out:
|
2015-07-08 03:05:03 +08:00
|
|
|
current->trace_recursion &= ~TRACE_BRANCH_BIT;
|
|
|
|
raw_local_irq_restore(flags);
|
2008-11-12 13:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline
|
2017-01-19 21:57:41 +08:00
|
|
|
void trace_likely_condition(struct ftrace_likely_data *f, int val, int expect)
|
2008-11-12 13:14:40 +08:00
|
|
|
{
|
2008-11-13 04:24:24 +08:00
|
|
|
if (!branch_tracing_enabled)
|
2008-11-12 13:14:40 +08:00
|
|
|
return;
|
|
|
|
|
|
|
|
probe_likely_condition(f, val, expect);
|
|
|
|
}
|
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
int enable_branch_tracing(struct trace_array *tr)
|
2008-11-12 13:14:40 +08:00
|
|
|
{
|
2008-11-13 04:24:24 +08:00
|
|
|
mutex_lock(&branch_tracing_mutex);
|
|
|
|
branch_tracer = tr;
|
2008-11-12 13:14:40 +08:00
|
|
|
/*
|
|
|
|
* Must be seen before enabling. The reader is a condition
|
|
|
|
* where we do not need a matching rmb()
|
|
|
|
*/
|
|
|
|
smp_wmb();
|
2008-11-13 04:24:24 +08:00
|
|
|
branch_tracing_enabled++;
|
|
|
|
mutex_unlock(&branch_tracing_mutex);
|
2008-11-12 13:14:40 +08:00
|
|
|
|
2009-02-10 14:02:46 +08:00
|
|
|
return 0;
|
2008-11-12 13:14:40 +08:00
|
|
|
}
|
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
void disable_branch_tracing(void)
|
2008-11-12 13:14:40 +08:00
|
|
|
{
|
2008-11-13 04:24:24 +08:00
|
|
|
mutex_lock(&branch_tracing_mutex);
|
2008-11-12 13:14:40 +08:00
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
if (!branch_tracing_enabled)
|
2008-11-12 13:14:40 +08:00
|
|
|
goto out_unlock;
|
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
branch_tracing_enabled--;
|
2008-11-12 13:14:40 +08:00
|
|
|
|
|
|
|
out_unlock:
|
2008-11-13 04:24:24 +08:00
|
|
|
mutex_unlock(&branch_tracing_mutex);
|
2008-11-12 13:14:40 +08:00
|
|
|
}
|
2008-11-13 04:24:24 +08:00
|
|
|
|
2008-11-16 12:57:26 +08:00
|
|
|
static int branch_trace_init(struct trace_array *tr)
|
2008-11-13 04:24:24 +08:00
|
|
|
{
|
2015-10-16 21:04:49 +08:00
|
|
|
return enable_branch_tracing(tr);
|
2008-11-13 04:24:24 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void branch_trace_reset(struct trace_array *tr)
|
|
|
|
{
|
2015-10-16 21:04:49 +08:00
|
|
|
disable_branch_tracing();
|
2008-11-13 04:24:24 +08:00
|
|
|
}
|
|
|
|
|
2009-02-04 08:05:50 +08:00
|
|
|
static enum print_line_t trace_branch_print(struct trace_iterator *iter,
|
2010-04-23 06:46:14 +08:00
|
|
|
int flags, struct trace_event *event)
|
2008-12-24 12:24:13 +08:00
|
|
|
{
|
|
|
|
struct trace_branch *field;
|
|
|
|
|
2009-02-03 06:30:12 +08:00
|
|
|
trace_assign_type(field, iter->ent);
|
2008-12-24 12:24:13 +08:00
|
|
|
|
2014-11-13 02:19:06 +08:00
|
|
|
trace_seq_printf(&iter->seq, "[%s] %s:%s:%d\n",
|
|
|
|
field->correct ? " ok " : " MISS ",
|
|
|
|
field->func,
|
|
|
|
field->file,
|
|
|
|
field->line);
|
|
|
|
|
|
|
|
return trace_handle_return(&iter->seq);
|
2008-12-24 12:24:13 +08:00
|
|
|
}
|
|
|
|
|
2009-04-13 16:02:34 +08:00
|
|
|
static void branch_print_header(struct seq_file *s)
|
|
|
|
{
|
|
|
|
seq_puts(s, "# TASK-PID CPU# TIMESTAMP CORRECT"
|
2014-11-09 04:42:11 +08:00
|
|
|
" FUNC:FILE:LINE\n"
|
|
|
|
"# | | | | | "
|
|
|
|
" |\n");
|
2009-04-13 16:02:34 +08:00
|
|
|
}
|
2008-12-28 06:25:38 +08:00
|
|
|
|
2010-04-23 06:46:14 +08:00
|
|
|
static struct trace_event_functions trace_branch_funcs = {
|
|
|
|
.trace = trace_branch_print,
|
|
|
|
};
|
|
|
|
|
2008-12-24 12:24:13 +08:00
|
|
|
static struct trace_event trace_branch_event = {
|
2009-03-11 02:10:56 +08:00
|
|
|
.type = TRACE_BRANCH,
|
2010-04-23 06:46:14 +08:00
|
|
|
.funcs = &trace_branch_funcs,
|
2008-12-24 12:24:13 +08:00
|
|
|
};
|
|
|
|
|
2009-01-11 03:34:13 +08:00
|
|
|
static struct tracer branch_trace __read_mostly =
|
|
|
|
{
|
|
|
|
.name = "branch",
|
|
|
|
.init = branch_trace_init,
|
|
|
|
.reset = branch_trace_reset,
|
|
|
|
#ifdef CONFIG_FTRACE_SELFTEST
|
|
|
|
.selftest = trace_selftest_startup_branch,
|
|
|
|
#endif /* CONFIG_FTRACE_SELFTEST */
|
2009-04-13 16:02:34 +08:00
|
|
|
.print_header = branch_print_header,
|
2009-01-11 03:34:13 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
__init static int init_branch_tracer(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2015-05-05 21:39:12 +08:00
|
|
|
ret = register_trace_event(&trace_branch_event);
|
2009-01-11 03:34:13 +08:00
|
|
|
if (!ret) {
|
|
|
|
printk(KERN_WARNING "Warning: could not register "
|
|
|
|
"branch events\n");
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return register_tracer(&branch_trace);
|
|
|
|
}
|
2012-10-06 00:13:07 +08:00
|
|
|
core_initcall(init_branch_tracer);
|
2009-01-11 03:34:13 +08:00
|
|
|
|
2008-11-12 13:14:40 +08:00
|
|
|
#else
|
|
|
|
static inline
|
2017-01-19 21:57:41 +08:00
|
|
|
void trace_likely_condition(struct ftrace_likely_data *f, int val, int expect)
|
2008-11-12 13:14:40 +08:00
|
|
|
{
|
|
|
|
}
|
2008-11-13 04:24:24 +08:00
|
|
|
#endif /* CONFIG_BRANCH_TRACER */
|
2008-11-12 13:14:40 +08:00
|
|
|
|
2017-01-19 21:57:14 +08:00
|
|
|
void ftrace_likely_update(struct ftrace_likely_data *f, int val,
|
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-18 01:29:35 +08:00
|
|
|
int expect, int is_constant)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-18 01:29:35 +08:00
|
|
|
/* A constant is always correct */
|
2017-01-19 21:57:14 +08:00
|
|
|
if (is_constant) {
|
|
|
|
f->constant++;
|
tracing: Process constants for (un)likely() profiler
When running the likely/unlikely profiler, one of the results did not look
accurate. It noted that the unlikely() in link_path_walk() was 100%
incorrect. When I added a trace_printk() to see what was happening there, it
became 80% correct! Looking deeper into what whas happening, I found that
gcc split that if statement into two paths. One where the if statement
became a constant, the other path a variable. The other path had the if
statement always hit (making the unlikely there, always false), but since
the #define unlikely() has:
#define unlikely() (__builtin_constant_p(x) ? !!(x) : __branch_check__(x, 0))
Where constants are ignored by the branch profiler, the "constant" path
made by the compiler was ignored, even though it was hit 80% of the time.
By just passing the constant value to the __branch_check__() function and
tracing it out of line (as always correct, as likely/unlikely isn't a factor
for constants), then we get back the accurate readings of branches that were
optimized by gcc causing part of the execution to become constant.
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
2017-01-18 01:29:35 +08:00
|
|
|
val = expect;
|
2017-01-19 21:57:14 +08:00
|
|
|
}
|
2008-11-12 13:14:40 +08:00
|
|
|
/*
|
|
|
|
* I would love to have a trace point here instead, but the
|
|
|
|
* trace point code is so inundated with unlikely and likely
|
|
|
|
* conditions that the recursive nightmare that exists is too
|
|
|
|
* much to try to get working. At least for now.
|
|
|
|
*/
|
2017-01-19 21:57:41 +08:00
|
|
|
trace_likely_condition(f, val, expect);
|
2008-11-12 13:14:40 +08:00
|
|
|
|
2008-11-12 13:14:39 +08:00
|
|
|
/* FIXME: Make this atomic! */
|
|
|
|
if (val == expect)
|
2017-01-19 21:57:14 +08:00
|
|
|
f->data.correct++;
|
2008-11-12 13:14:39 +08:00
|
|
|
else
|
2017-01-19 21:57:14 +08:00
|
|
|
f->data.incorrect++;
|
2008-11-12 13:14:39 +08:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(ftrace_likely_update);
|
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
extern unsigned long __start_annotated_branch_profile[];
|
|
|
|
extern unsigned long __stop_annotated_branch_profile[];
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
static int annotated_branch_stat_headers(struct seq_file *m)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
2014-11-09 04:42:11 +08:00
|
|
|
seq_puts(m, " correct incorrect % "
|
|
|
|
" Function "
|
2014-11-09 04:42:10 +08:00
|
|
|
" File Line\n"
|
|
|
|
" ------- --------- - "
|
|
|
|
" -------- "
|
|
|
|
" ---- ----\n");
|
2008-12-28 06:25:38 +08:00
|
|
|
return 0;
|
2008-11-12 13:14:39 +08:00
|
|
|
}
|
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
static inline long get_incorrect_percent(struct ftrace_branch_data *p)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
2008-12-28 06:25:38 +08:00
|
|
|
long percent;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
if (p->correct) {
|
|
|
|
percent = p->incorrect * 100;
|
|
|
|
percent /= p->correct + p->incorrect;
|
|
|
|
} else
|
|
|
|
percent = p->incorrect ? 100 : -1;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
return percent;
|
2008-11-12 13:14:39 +08:00
|
|
|
}
|
|
|
|
|
2017-01-19 21:57:14 +08:00
|
|
|
static const char *branch_stat_process_file(struct ftrace_branch_data *p)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
|
|
|
const char *f;
|
|
|
|
|
|
|
|
/* Only print the file, not the path */
|
|
|
|
f = p->file + strlen(p->file);
|
|
|
|
while (f >= p->file && *f != '/')
|
|
|
|
f--;
|
2017-01-19 21:57:14 +08:00
|
|
|
return ++f;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void branch_stat_show(struct seq_file *m,
|
|
|
|
struct ftrace_branch_data *p, const char *f)
|
|
|
|
{
|
|
|
|
long percent;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-11-21 14:30:54 +08:00
|
|
|
/*
|
|
|
|
* The miss is overlayed on correct, and hit on incorrect.
|
|
|
|
*/
|
2008-12-28 06:25:38 +08:00
|
|
|
percent = get_incorrect_percent(p);
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-11-21 14:51:53 +08:00
|
|
|
if (percent < 0)
|
2014-11-09 04:42:10 +08:00
|
|
|
seq_puts(m, " X ");
|
2008-11-21 14:51:53 +08:00
|
|
|
else
|
|
|
|
seq_printf(m, "%3ld ", percent);
|
2017-01-19 21:57:14 +08:00
|
|
|
|
2008-11-12 13:14:39 +08:00
|
|
|
seq_printf(m, "%-30.30s %-20.20s %d\n", p->func, f, p->line);
|
2017-01-19 21:57:14 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int branch_stat_show_normal(struct seq_file *m,
|
|
|
|
struct ftrace_branch_data *p, const char *f)
|
|
|
|
{
|
|
|
|
seq_printf(m, "%8lu %8lu ", p->correct, p->incorrect);
|
|
|
|
branch_stat_show(m, p, f);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int annotate_branch_stat_show(struct seq_file *m, void *v)
|
|
|
|
{
|
|
|
|
struct ftrace_likely_data *p = v;
|
|
|
|
const char *f;
|
|
|
|
int l;
|
|
|
|
|
|
|
|
f = branch_stat_process_file(&p->data);
|
|
|
|
|
|
|
|
if (!p->constant)
|
|
|
|
return branch_stat_show_normal(m, &p->data, f);
|
|
|
|
|
|
|
|
l = snprintf(NULL, 0, "/%lu", p->constant);
|
|
|
|
l = l > 8 ? 0 : 8 - l;
|
|
|
|
|
|
|
|
seq_printf(m, "%8lu/%lu %*lu ",
|
|
|
|
p->data.correct, p->constant, l, p->data.incorrect);
|
|
|
|
branch_stat_show(m, &p->data, f);
|
2008-11-12 13:14:39 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-03-25 01:38:36 +08:00
|
|
|
static void *annotated_branch_stat_start(struct tracer_stat *trace)
|
2008-12-28 06:25:38 +08:00
|
|
|
{
|
|
|
|
return __start_annotated_branch_profile;
|
|
|
|
}
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
static void *
|
|
|
|
annotated_branch_stat_next(void *v, int idx)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
2017-01-19 21:57:14 +08:00
|
|
|
struct ftrace_likely_data *p = v;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
++p;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
if ((void *)p >= (void *)__stop_annotated_branch_profile)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return p;
|
2008-11-12 13:14:39 +08:00
|
|
|
}
|
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
static int annotated_branch_stat_cmp(void *p1, void *p2)
|
|
|
|
{
|
|
|
|
struct ftrace_branch_data *a = p1;
|
|
|
|
struct ftrace_branch_data *b = p2;
|
|
|
|
|
|
|
|
long percent_a, percent_b;
|
|
|
|
|
|
|
|
percent_a = get_incorrect_percent(a);
|
|
|
|
percent_b = get_incorrect_percent(b);
|
|
|
|
|
|
|
|
if (percent_a < percent_b)
|
|
|
|
return -1;
|
|
|
|
if (percent_a > percent_b)
|
|
|
|
return 1;
|
2010-01-28 00:25:54 +08:00
|
|
|
|
|
|
|
if (a->incorrect < b->incorrect)
|
|
|
|
return -1;
|
|
|
|
if (a->incorrect > b->incorrect)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Since the above shows worse (incorrect) cases
|
|
|
|
* first, we continue that by showing best (correct)
|
|
|
|
* cases last.
|
|
|
|
*/
|
|
|
|
if (a->correct > b->correct)
|
|
|
|
return -1;
|
|
|
|
if (a->correct < b->correct)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
return 0;
|
2008-12-28 06:25:38 +08:00
|
|
|
}
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2009-01-11 03:34:13 +08:00
|
|
|
static struct tracer_stat annotated_branch_stats = {
|
|
|
|
.name = "branch_annotated",
|
|
|
|
.stat_start = annotated_branch_stat_start,
|
|
|
|
.stat_next = annotated_branch_stat_next,
|
|
|
|
.stat_cmp = annotated_branch_stat_cmp,
|
|
|
|
.stat_headers = annotated_branch_stat_headers,
|
2017-01-19 21:57:14 +08:00
|
|
|
.stat_show = annotate_branch_stat_show
|
2009-01-11 03:34:13 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
__init static int init_annotated_branch_stats(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = register_stat_tracer(&annotated_branch_stats);
|
|
|
|
if (!ret) {
|
|
|
|
printk(KERN_WARNING "Warning: could not register "
|
|
|
|
"annotated branches stats\n");
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
fs_initcall(init_annotated_branch_stats);
|
|
|
|
|
2008-11-21 14:30:54 +08:00
|
|
|
#ifdef CONFIG_PROFILE_ALL_BRANCHES
|
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
extern unsigned long __start_branch_profile[];
|
|
|
|
extern unsigned long __stop_branch_profile[];
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
static int all_branch_stat_headers(struct seq_file *m)
|
|
|
|
{
|
2014-11-09 04:42:11 +08:00
|
|
|
seq_puts(m, " miss hit % "
|
|
|
|
" Function "
|
2014-11-09 04:42:10 +08:00
|
|
|
" File Line\n"
|
|
|
|
" ------- --------- - "
|
|
|
|
" -------- "
|
|
|
|
" ---- ----\n");
|
2008-12-28 06:25:38 +08:00
|
|
|
return 0;
|
|
|
|
}
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2009-03-25 01:38:36 +08:00
|
|
|
static void *all_branch_stat_start(struct tracer_stat *trace)
|
2008-11-12 13:14:39 +08:00
|
|
|
{
|
2008-12-28 06:25:38 +08:00
|
|
|
return __start_branch_profile;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *
|
|
|
|
all_branch_stat_next(void *v, int idx)
|
|
|
|
{
|
|
|
|
struct ftrace_branch_data *p = v;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
++p;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
if ((void *)p >= (void *)__stop_branch_profile)
|
|
|
|
return NULL;
|
2008-11-12 13:14:39 +08:00
|
|
|
|
2008-12-28 06:25:38 +08:00
|
|
|
return p;
|
|
|
|
}
|
2008-11-21 14:30:54 +08:00
|
|
|
|
2017-01-19 21:57:14 +08:00
|
|
|
static int all_branch_stat_show(struct seq_file *m, void *v)
|
|
|
|
{
|
|
|
|
struct ftrace_branch_data *p = v;
|
|
|
|
const char *f;
|
|
|
|
|
|
|
|
f = branch_stat_process_file(p);
|
|
|
|
return branch_stat_show_normal(m, p, f);
|
|
|
|
}
|
|
|
|
|
2009-01-11 03:34:13 +08:00
|
|
|
static struct tracer_stat all_branch_stats = {
|
|
|
|
.name = "branch_all",
|
2009-01-09 02:03:56 +08:00
|
|
|
.stat_start = all_branch_stat_start,
|
|
|
|
.stat_next = all_branch_stat_next,
|
|
|
|
.stat_headers = all_branch_stat_headers,
|
2017-01-19 21:57:14 +08:00
|
|
|
.stat_show = all_branch_stat_show
|
2009-01-09 02:03:56 +08:00
|
|
|
};
|
2008-12-28 06:25:38 +08:00
|
|
|
|
2009-01-11 03:34:13 +08:00
|
|
|
__init static int all_annotated_branch_stats(void)
|
2008-12-28 06:25:38 +08:00
|
|
|
{
|
|
|
|
int ret;
|
2009-01-11 03:34:13 +08:00
|
|
|
|
|
|
|
ret = register_stat_tracer(&all_branch_stats);
|
2008-12-28 06:25:38 +08:00
|
|
|
if (!ret) {
|
2009-01-11 03:34:13 +08:00
|
|
|
printk(KERN_WARNING "Warning: could not register "
|
|
|
|
"all branches stats\n");
|
2008-12-28 06:25:38 +08:00
|
|
|
return 1;
|
|
|
|
}
|
2009-01-11 03:34:13 +08:00
|
|
|
return 0;
|
2008-12-28 06:25:38 +08:00
|
|
|
}
|
2009-01-11 03:34:13 +08:00
|
|
|
fs_initcall(all_annotated_branch_stats);
|
|
|
|
#endif /* CONFIG_PROFILE_ALL_BRANCHES */
|