perf annotate: Move compute_ipc() to annotation library
Out of the TUI code, as it has nothing specific to that UI and should be used in the other output modes as well. Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: David Ahern <dsahern@gmail.com> Cc: Jin Yao <yao.jin@linux.intel.com> Cc: Jiri Olsa <jolsa@kernel.org> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Wang Nan <wangnan0@huawei.com> Link: https://lkml.kernel.org/n/tip-0jahghvqdodb8vu2591pkv3d@git.kernel.org Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
9d6bb41d1c
commit
f56c083bc4
|
@ -967,73 +967,6 @@ int hist_entry__tui_annotate(struct hist_entry *he, struct perf_evsel *evsel,
|
|||
return map_symbol__tui_annotate(&he->ms, evsel, hbt);
|
||||
}
|
||||
|
||||
|
||||
static unsigned count_insn(struct annotation *notes, u64 start, u64 end)
|
||||
{
|
||||
unsigned n_insn = 0;
|
||||
u64 offset;
|
||||
|
||||
for (offset = start; offset <= end; offset++) {
|
||||
if (notes->offsets[offset])
|
||||
n_insn++;
|
||||
}
|
||||
return n_insn;
|
||||
}
|
||||
|
||||
static void count_and_fill(struct annotation *notes, u64 start, u64 end,
|
||||
struct cyc_hist *ch)
|
||||
{
|
||||
unsigned n_insn;
|
||||
u64 offset;
|
||||
|
||||
n_insn = count_insn(notes, start, end);
|
||||
if (n_insn && ch->num && ch->cycles) {
|
||||
float ipc = n_insn / ((double)ch->cycles / (double)ch->num);
|
||||
|
||||
/* Hide data when there are too many overlaps. */
|
||||
if (ch->reset >= 0x7fff || ch->reset >= ch->num / 2)
|
||||
return;
|
||||
|
||||
for (offset = start; offset <= end; offset++) {
|
||||
struct annotation_line *al = notes->offsets[offset];
|
||||
|
||||
if (al)
|
||||
al->ipc = ipc;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* This should probably be in util/annotate.c to share with the tty
|
||||
* annotate, but right now we need the per byte offsets arrays,
|
||||
* which are only here.
|
||||
*/
|
||||
static void annotate__compute_ipc(struct annotation *notes, size_t size)
|
||||
{
|
||||
u64 offset;
|
||||
|
||||
if (!notes->src || !notes->src->cycles_hist)
|
||||
return;
|
||||
|
||||
pthread_mutex_lock(¬es->lock);
|
||||
for (offset = 0; offset < size; ++offset) {
|
||||
struct cyc_hist *ch;
|
||||
|
||||
ch = ¬es->src->cycles_hist[offset];
|
||||
if (ch && ch->cycles) {
|
||||
struct annotation_line *al;
|
||||
|
||||
if (ch->have_start)
|
||||
count_and_fill(notes, ch->start, offset, ch);
|
||||
al = notes->offsets[offset];
|
||||
if (al && ch->num_aggr)
|
||||
al->cycles = ch->cycles_aggr / ch->num_aggr;
|
||||
notes->have_cycles = true;
|
||||
}
|
||||
}
|
||||
pthread_mutex_unlock(¬es->lock);
|
||||
}
|
||||
|
||||
static void annotate_browser__mark_jump_targets(struct annotate_browser *browser,
|
||||
size_t size)
|
||||
{
|
||||
|
@ -1161,7 +1094,7 @@ int symbol__tui_annotate(struct symbol *sym, struct map *map,
|
|||
}
|
||||
|
||||
annotate_browser__mark_jump_targets(&browser, size);
|
||||
annotate__compute_ipc(notes, size);
|
||||
annotation__compute_ipc(notes, size);
|
||||
|
||||
browser.addr_width = browser.target_width = browser.min_addr_width = hex_width(size);
|
||||
browser.max_addr_width = hex_width(sym->end);
|
||||
|
|
|
@ -833,6 +833,66 @@ int addr_map_symbol__account_cycles(struct addr_map_symbol *ams,
|
|||
return err;
|
||||
}
|
||||
|
||||
static unsigned annotation__count_insn(struct annotation *notes, u64 start, u64 end)
|
||||
{
|
||||
unsigned n_insn = 0;
|
||||
u64 offset;
|
||||
|
||||
for (offset = start; offset <= end; offset++) {
|
||||
if (notes->offsets[offset])
|
||||
n_insn++;
|
||||
}
|
||||
return n_insn;
|
||||
}
|
||||
|
||||
static void annotation__count_and_fill(struct annotation *notes, u64 start, u64 end, struct cyc_hist *ch)
|
||||
{
|
||||
unsigned n_insn;
|
||||
u64 offset;
|
||||
|
||||
n_insn = annotation__count_insn(notes, start, end);
|
||||
if (n_insn && ch->num && ch->cycles) {
|
||||
float ipc = n_insn / ((double)ch->cycles / (double)ch->num);
|
||||
|
||||
/* Hide data when there are too many overlaps. */
|
||||
if (ch->reset >= 0x7fff || ch->reset >= ch->num / 2)
|
||||
return;
|
||||
|
||||
for (offset = start; offset <= end; offset++) {
|
||||
struct annotation_line *al = notes->offsets[offset];
|
||||
|
||||
if (al)
|
||||
al->ipc = ipc;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void annotation__compute_ipc(struct annotation *notes, size_t size)
|
||||
{
|
||||
u64 offset;
|
||||
|
||||
if (!notes->src || !notes->src->cycles_hist)
|
||||
return;
|
||||
|
||||
pthread_mutex_lock(¬es->lock);
|
||||
for (offset = 0; offset < size; ++offset) {
|
||||
struct cyc_hist *ch;
|
||||
|
||||
ch = ¬es->src->cycles_hist[offset];
|
||||
if (ch && ch->cycles) {
|
||||
struct annotation_line *al;
|
||||
|
||||
if (ch->have_start)
|
||||
annotation__count_and_fill(notes, ch->start, offset, ch);
|
||||
al = notes->offsets[offset];
|
||||
if (al && ch->num_aggr)
|
||||
al->cycles = ch->cycles_aggr / ch->num_aggr;
|
||||
notes->have_cycles = true;
|
||||
}
|
||||
}
|
||||
pthread_mutex_unlock(¬es->lock);
|
||||
}
|
||||
|
||||
int addr_map_symbol__inc_samples(struct addr_map_symbol *ams, struct perf_sample *sample,
|
||||
int evidx)
|
||||
{
|
||||
|
|
|
@ -174,6 +174,8 @@ static inline int annotation__cycles_width(struct annotation *notes)
|
|||
return notes->have_cycles ? ANNOTATION__IPC_WIDTH + ANNOTATION__CYCLES_WIDTH : 0;
|
||||
}
|
||||
|
||||
void annotation__compute_ipc(struct annotation *notes, size_t size);
|
||||
|
||||
static inline struct sym_hist *annotation__histogram(struct annotation *notes, int idx)
|
||||
{
|
||||
return (((void *)¬es->src->histograms) +
|
||||
|
|
Loading…
Reference in New Issue