Merge tag 'perf-urgent-for-mingo' of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux into perf/urgent
Pull perf fixes from Arnaldo Carvalho de Melo: * Endianness fixes from Jiri Olsa * Fixes for make perf tarball * Fix for DSO name in perf script callchains, from David Ahern * Segfault fixes for perf top --callchain, from Namhyung Kim * Minor function result fixes from Srikar Dronamraju * Add missing 3rd ioctl parameter, from Namhyung Kim * Fix pager usage in minimal embedded systems, from Avik Sil Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com> Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
commit
02e03040a3
|
@ -555,6 +555,8 @@ enum perf_event_type {
|
|||
PERF_RECORD_MAX, /* non-ABI */
|
||||
};
|
||||
|
||||
#define PERF_MAX_STACK_DEPTH 255
|
||||
|
||||
enum perf_callchain_context {
|
||||
PERF_CONTEXT_HV = (__u64)-32,
|
||||
PERF_CONTEXT_KERNEL = (__u64)-128,
|
||||
|
@ -609,8 +611,6 @@ struct perf_guest_info_callbacks {
|
|||
#include <linux/sysfs.h>
|
||||
#include <asm/local.h>
|
||||
|
||||
#define PERF_MAX_STACK_DEPTH 255
|
||||
|
||||
struct perf_callchain_entry {
|
||||
__u64 nr;
|
||||
__u64 ip[PERF_MAX_STACK_DEPTH];
|
||||
|
|
|
@ -3181,7 +3181,6 @@ static void perf_event_for_each(struct perf_event *event,
|
|||
event = event->group_leader;
|
||||
|
||||
perf_event_for_each_child(event, func);
|
||||
func(event);
|
||||
list_for_each_entry(sibling, &event->sibling_list, group_entry)
|
||||
perf_event_for_each_child(sibling, func);
|
||||
mutex_unlock(&ctx->mutex);
|
||||
|
|
|
@ -1,4 +1,6 @@
|
|||
tools/perf
|
||||
tools/scripts
|
||||
tools/lib/traceevent
|
||||
include/linux/const.h
|
||||
include/linux/perf_event.h
|
||||
include/linux/rbtree.h
|
||||
|
|
|
@ -152,7 +152,7 @@ static int perf_evsel__add_hist_entry(struct perf_evsel *evsel,
|
|||
|
||||
if (symbol_conf.use_callchain) {
|
||||
err = callchain_append(he->callchain,
|
||||
&evsel->hists.callchain_cursor,
|
||||
&callchain_cursor,
|
||||
sample->period);
|
||||
if (err)
|
||||
return err;
|
||||
|
@ -162,7 +162,7 @@ static int perf_evsel__add_hist_entry(struct perf_evsel *evsel,
|
|||
* so we don't allocated the extra space needed because the stdio
|
||||
* code will not use it.
|
||||
*/
|
||||
if (al->sym != NULL && use_browser > 0) {
|
||||
if (he->ms.sym != NULL && use_browser > 0) {
|
||||
struct annotation *notes = symbol__annotation(he->ms.sym);
|
||||
|
||||
assert(evsel != NULL);
|
||||
|
|
|
@ -1129,7 +1129,7 @@ static int add_default_attributes(void)
|
|||
return 0;
|
||||
|
||||
if (!evsel_list->nr_entries) {
|
||||
if (perf_evlist__add_attrs_array(evsel_list, default_attrs) < 0)
|
||||
if (perf_evlist__add_default_attrs(evsel_list, default_attrs) < 0)
|
||||
return -1;
|
||||
}
|
||||
|
||||
|
@ -1139,21 +1139,21 @@ static int add_default_attributes(void)
|
|||
return 0;
|
||||
|
||||
/* Append detailed run extra attributes: */
|
||||
if (perf_evlist__add_attrs_array(evsel_list, detailed_attrs) < 0)
|
||||
if (perf_evlist__add_default_attrs(evsel_list, detailed_attrs) < 0)
|
||||
return -1;
|
||||
|
||||
if (detailed_run < 2)
|
||||
return 0;
|
||||
|
||||
/* Append very detailed run extra attributes: */
|
||||
if (perf_evlist__add_attrs_array(evsel_list, very_detailed_attrs) < 0)
|
||||
if (perf_evlist__add_default_attrs(evsel_list, very_detailed_attrs) < 0)
|
||||
return -1;
|
||||
|
||||
if (detailed_run < 3)
|
||||
return 0;
|
||||
|
||||
/* Append very, very detailed run extra attributes: */
|
||||
return perf_evlist__add_attrs_array(evsel_list, very_very_detailed_attrs);
|
||||
return perf_evlist__add_default_attrs(evsel_list, very_very_detailed_attrs);
|
||||
}
|
||||
|
||||
int cmd_stat(int argc, const char **argv, const char *prefix __used)
|
||||
|
|
|
@ -787,7 +787,7 @@ static void perf_event__process_sample(struct perf_tool *tool,
|
|||
}
|
||||
|
||||
if (symbol_conf.use_callchain) {
|
||||
err = callchain_append(he->callchain, &evsel->hists.callchain_cursor,
|
||||
err = callchain_append(he->callchain, &callchain_cursor,
|
||||
sample->period);
|
||||
if (err)
|
||||
return;
|
||||
|
|
|
@ -409,14 +409,15 @@ Counters can be enabled and disabled in two ways: via ioctl and via
|
|||
prctl. When a counter is disabled, it doesn't count or generate
|
||||
events but does continue to exist and maintain its count value.
|
||||
|
||||
An individual counter or counter group can be enabled with
|
||||
An individual counter can be enabled with
|
||||
|
||||
ioctl(fd, PERF_EVENT_IOC_ENABLE);
|
||||
ioctl(fd, PERF_EVENT_IOC_ENABLE, 0);
|
||||
|
||||
or disabled with
|
||||
|
||||
ioctl(fd, PERF_EVENT_IOC_DISABLE);
|
||||
ioctl(fd, PERF_EVENT_IOC_DISABLE, 0);
|
||||
|
||||
For a counter group, pass PERF_IOC_FLAG_GROUP as the third argument.
|
||||
Enabling or disabling the leader of a group enables or disables the
|
||||
whole group; that is, while the group leader is disabled, none of the
|
||||
counters in the group will count. Enabling or disabling a member of a
|
||||
|
|
|
@ -668,7 +668,7 @@ static int annotate_browser__run(struct annotate_browser *browser, int evidx,
|
|||
"q/ESC/CTRL+C Exit\n\n"
|
||||
"-> Go to target\n"
|
||||
"<- Exit\n"
|
||||
"h Cycle thru hottest instructions\n"
|
||||
"H Cycle thru hottest instructions\n"
|
||||
"j Toggle showing jump to target arrows\n"
|
||||
"J Toggle showing number of jump sources on targets\n"
|
||||
"n Search next string\n"
|
||||
|
|
|
@ -12,7 +12,7 @@ LF='
|
|||
# First check if there is a .git to get the version from git describe
|
||||
# otherwise try to get the version from the kernel makefile
|
||||
if test -d ../../.git -o -f ../../.git &&
|
||||
VN=$(git describe --abbrev=4 HEAD 2>/dev/null) &&
|
||||
VN=$(git describe --match 'v[0-9].[0-9]*' --abbrev=4 HEAD 2>/dev/null) &&
|
||||
case "$VN" in
|
||||
*$LF*) (exit 1) ;;
|
||||
v[0-9]*)
|
||||
|
|
|
@ -18,6 +18,8 @@
|
|||
#include "util.h"
|
||||
#include "callchain.h"
|
||||
|
||||
__thread struct callchain_cursor callchain_cursor;
|
||||
|
||||
bool ip_callchain__valid(struct ip_callchain *chain,
|
||||
const union perf_event *event)
|
||||
{
|
||||
|
|
|
@ -76,6 +76,8 @@ struct callchain_cursor {
|
|||
struct callchain_cursor_node *curr;
|
||||
};
|
||||
|
||||
extern __thread struct callchain_cursor callchain_cursor;
|
||||
|
||||
static inline void callchain_init(struct callchain_root *root)
|
||||
{
|
||||
INIT_LIST_HEAD(&root->node.siblings);
|
||||
|
|
|
@ -159,6 +159,17 @@ out_delete_partial_list:
|
|||
return -1;
|
||||
}
|
||||
|
||||
int __perf_evlist__add_default_attrs(struct perf_evlist *evlist,
|
||||
struct perf_event_attr *attrs, size_t nr_attrs)
|
||||
{
|
||||
size_t i;
|
||||
|
||||
for (i = 0; i < nr_attrs; i++)
|
||||
event_attr_init(attrs + i);
|
||||
|
||||
return perf_evlist__add_attrs(evlist, attrs, nr_attrs);
|
||||
}
|
||||
|
||||
static int trace_event__id(const char *evname)
|
||||
{
|
||||
char *filename, *colon;
|
||||
|
@ -263,7 +274,8 @@ void perf_evlist__disable(struct perf_evlist *evlist)
|
|||
for (cpu = 0; cpu < evlist->cpus->nr; cpu++) {
|
||||
list_for_each_entry(pos, &evlist->entries, node) {
|
||||
for (thread = 0; thread < evlist->threads->nr; thread++)
|
||||
ioctl(FD(pos, cpu, thread), PERF_EVENT_IOC_DISABLE);
|
||||
ioctl(FD(pos, cpu, thread),
|
||||
PERF_EVENT_IOC_DISABLE, 0);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -276,7 +288,8 @@ void perf_evlist__enable(struct perf_evlist *evlist)
|
|||
for (cpu = 0; cpu < evlist->cpus->nr; cpu++) {
|
||||
list_for_each_entry(pos, &evlist->entries, node) {
|
||||
for (thread = 0; thread < evlist->threads->nr; thread++)
|
||||
ioctl(FD(pos, cpu, thread), PERF_EVENT_IOC_ENABLE);
|
||||
ioctl(FD(pos, cpu, thread),
|
||||
PERF_EVENT_IOC_ENABLE, 0);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,6 +54,8 @@ void perf_evlist__add(struct perf_evlist *evlist, struct perf_evsel *entry);
|
|||
int perf_evlist__add_default(struct perf_evlist *evlist);
|
||||
int perf_evlist__add_attrs(struct perf_evlist *evlist,
|
||||
struct perf_event_attr *attrs, size_t nr_attrs);
|
||||
int __perf_evlist__add_default_attrs(struct perf_evlist *evlist,
|
||||
struct perf_event_attr *attrs, size_t nr_attrs);
|
||||
int perf_evlist__add_tracepoints(struct perf_evlist *evlist,
|
||||
const char *tracepoints[], size_t nr_tracepoints);
|
||||
int perf_evlist__set_tracepoints_handlers(struct perf_evlist *evlist,
|
||||
|
@ -62,6 +64,8 @@ int perf_evlist__set_tracepoints_handlers(struct perf_evlist *evlist,
|
|||
|
||||
#define perf_evlist__add_attrs_array(evlist, array) \
|
||||
perf_evlist__add_attrs(evlist, array, ARRAY_SIZE(array))
|
||||
#define perf_evlist__add_default_attrs(evlist, array) \
|
||||
__perf_evlist__add_default_attrs(evlist, array, ARRAY_SIZE(array))
|
||||
|
||||
#define perf_evlist__add_tracepoints_array(evlist, array) \
|
||||
perf_evlist__add_tracepoints(evlist, array, ARRAY_SIZE(array))
|
||||
|
|
|
@ -494,16 +494,24 @@ int perf_evsel__open_per_thread(struct perf_evsel *evsel,
|
|||
}
|
||||
|
||||
static int perf_event__parse_id_sample(const union perf_event *event, u64 type,
|
||||
struct perf_sample *sample)
|
||||
struct perf_sample *sample,
|
||||
bool swapped)
|
||||
{
|
||||
const u64 *array = event->sample.array;
|
||||
union u64_swap u;
|
||||
|
||||
array += ((event->header.size -
|
||||
sizeof(event->header)) / sizeof(u64)) - 1;
|
||||
|
||||
if (type & PERF_SAMPLE_CPU) {
|
||||
u32 *p = (u32 *)array;
|
||||
sample->cpu = *p;
|
||||
u.val64 = *array;
|
||||
if (swapped) {
|
||||
/* undo swap of u64, then swap on individual u32s */
|
||||
u.val64 = bswap_64(u.val64);
|
||||
u.val32[0] = bswap_32(u.val32[0]);
|
||||
}
|
||||
|
||||
sample->cpu = u.val32[0];
|
||||
array--;
|
||||
}
|
||||
|
||||
|
@ -523,9 +531,16 @@ static int perf_event__parse_id_sample(const union perf_event *event, u64 type,
|
|||
}
|
||||
|
||||
if (type & PERF_SAMPLE_TID) {
|
||||
u32 *p = (u32 *)array;
|
||||
sample->pid = p[0];
|
||||
sample->tid = p[1];
|
||||
u.val64 = *array;
|
||||
if (swapped) {
|
||||
/* undo swap of u64, then swap on individual u32s */
|
||||
u.val64 = bswap_64(u.val64);
|
||||
u.val32[0] = bswap_32(u.val32[0]);
|
||||
u.val32[1] = bswap_32(u.val32[1]);
|
||||
}
|
||||
|
||||
sample->pid = u.val32[0];
|
||||
sample->tid = u.val32[1];
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
@ -562,7 +577,7 @@ int perf_event__parse_sample(const union perf_event *event, u64 type,
|
|||
if (event->header.type != PERF_RECORD_SAMPLE) {
|
||||
if (!sample_id_all)
|
||||
return 0;
|
||||
return perf_event__parse_id_sample(event, type, data);
|
||||
return perf_event__parse_id_sample(event, type, data, swapped);
|
||||
}
|
||||
|
||||
array = event->sample.array;
|
||||
|
|
|
@ -378,7 +378,7 @@ void hist_entry__free(struct hist_entry *he)
|
|||
* collapse the histogram
|
||||
*/
|
||||
|
||||
static bool hists__collapse_insert_entry(struct hists *hists,
|
||||
static bool hists__collapse_insert_entry(struct hists *hists __used,
|
||||
struct rb_root *root,
|
||||
struct hist_entry *he)
|
||||
{
|
||||
|
@ -397,8 +397,9 @@ static bool hists__collapse_insert_entry(struct hists *hists,
|
|||
iter->period += he->period;
|
||||
iter->nr_events += he->nr_events;
|
||||
if (symbol_conf.use_callchain) {
|
||||
callchain_cursor_reset(&hists->callchain_cursor);
|
||||
callchain_merge(&hists->callchain_cursor, iter->callchain,
|
||||
callchain_cursor_reset(&callchain_cursor);
|
||||
callchain_merge(&callchain_cursor,
|
||||
iter->callchain,
|
||||
he->callchain);
|
||||
}
|
||||
hist_entry__free(he);
|
||||
|
|
|
@ -67,8 +67,6 @@ struct hists {
|
|||
struct events_stats stats;
|
||||
u64 event_stream;
|
||||
u16 col_len[HISTC_NR_COLS];
|
||||
/* Best would be to reuse the session callchain cursor */
|
||||
struct callchain_cursor callchain_cursor;
|
||||
};
|
||||
|
||||
struct hist_entry *__hists__add_entry(struct hists *self,
|
||||
|
|
|
@ -57,6 +57,10 @@ void setup_pager(void)
|
|||
}
|
||||
if (!pager)
|
||||
pager = getenv("PAGER");
|
||||
if (!pager) {
|
||||
if (!access("/usr/bin/pager", X_OK))
|
||||
pager = "/usr/bin/pager";
|
||||
}
|
||||
if (!pager)
|
||||
pager = "less";
|
||||
else if (!*pager || !strcmp(pager, "cat"))
|
||||
|
|
|
@ -2164,16 +2164,12 @@ int del_perf_probe_events(struct strlist *dellist)
|
|||
|
||||
error:
|
||||
if (kfd >= 0) {
|
||||
if (namelist)
|
||||
strlist__delete(namelist);
|
||||
|
||||
strlist__delete(namelist);
|
||||
close(kfd);
|
||||
}
|
||||
|
||||
if (ufd >= 0) {
|
||||
if (unamelist)
|
||||
strlist__delete(unamelist);
|
||||
|
||||
strlist__delete(unamelist);
|
||||
close(ufd);
|
||||
}
|
||||
|
||||
|
|
|
@ -288,7 +288,8 @@ struct branch_info *machine__resolve_bstack(struct machine *self,
|
|||
return bi;
|
||||
}
|
||||
|
||||
int machine__resolve_callchain(struct machine *self, struct perf_evsel *evsel,
|
||||
int machine__resolve_callchain(struct machine *self,
|
||||
struct perf_evsel *evsel __used,
|
||||
struct thread *thread,
|
||||
struct ip_callchain *chain,
|
||||
struct symbol **parent)
|
||||
|
@ -297,7 +298,12 @@ int machine__resolve_callchain(struct machine *self, struct perf_evsel *evsel,
|
|||
unsigned int i;
|
||||
int err;
|
||||
|
||||
callchain_cursor_reset(&evsel->hists.callchain_cursor);
|
||||
callchain_cursor_reset(&callchain_cursor);
|
||||
|
||||
if (chain->nr > PERF_MAX_STACK_DEPTH) {
|
||||
pr_warning("corrupted callchain. skipping...\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
for (i = 0; i < chain->nr; i++) {
|
||||
u64 ip;
|
||||
|
@ -317,7 +323,14 @@ int machine__resolve_callchain(struct machine *self, struct perf_evsel *evsel,
|
|||
case PERF_CONTEXT_USER:
|
||||
cpumode = PERF_RECORD_MISC_USER; break;
|
||||
default:
|
||||
break;
|
||||
pr_debug("invalid callchain context: "
|
||||
"%"PRId64"\n", (s64) ip);
|
||||
/*
|
||||
* It seems the callchain is corrupted.
|
||||
* Discard all.
|
||||
*/
|
||||
callchain_cursor_reset(&callchain_cursor);
|
||||
return 0;
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
@ -333,7 +346,7 @@ int machine__resolve_callchain(struct machine *self, struct perf_evsel *evsel,
|
|||
break;
|
||||
}
|
||||
|
||||
err = callchain_cursor_append(&evsel->hists.callchain_cursor,
|
||||
err = callchain_cursor_append(&callchain_cursor,
|
||||
ip, al.map, al.sym);
|
||||
if (err)
|
||||
return err;
|
||||
|
@ -441,37 +454,65 @@ void mem_bswap_64(void *src, int byte_size)
|
|||
}
|
||||
}
|
||||
|
||||
static void perf_event__all64_swap(union perf_event *event)
|
||||
static void swap_sample_id_all(union perf_event *event, void *data)
|
||||
{
|
||||
void *end = (void *) event + event->header.size;
|
||||
int size = end - data;
|
||||
|
||||
BUG_ON(size % sizeof(u64));
|
||||
mem_bswap_64(data, size);
|
||||
}
|
||||
|
||||
static void perf_event__all64_swap(union perf_event *event,
|
||||
bool sample_id_all __used)
|
||||
{
|
||||
struct perf_event_header *hdr = &event->header;
|
||||
mem_bswap_64(hdr + 1, event->header.size - sizeof(*hdr));
|
||||
}
|
||||
|
||||
static void perf_event__comm_swap(union perf_event *event)
|
||||
static void perf_event__comm_swap(union perf_event *event, bool sample_id_all)
|
||||
{
|
||||
event->comm.pid = bswap_32(event->comm.pid);
|
||||
event->comm.tid = bswap_32(event->comm.tid);
|
||||
|
||||
if (sample_id_all) {
|
||||
void *data = &event->comm.comm;
|
||||
|
||||
data += ALIGN(strlen(data) + 1, sizeof(u64));
|
||||
swap_sample_id_all(event, data);
|
||||
}
|
||||
}
|
||||
|
||||
static void perf_event__mmap_swap(union perf_event *event)
|
||||
static void perf_event__mmap_swap(union perf_event *event,
|
||||
bool sample_id_all)
|
||||
{
|
||||
event->mmap.pid = bswap_32(event->mmap.pid);
|
||||
event->mmap.tid = bswap_32(event->mmap.tid);
|
||||
event->mmap.start = bswap_64(event->mmap.start);
|
||||
event->mmap.len = bswap_64(event->mmap.len);
|
||||
event->mmap.pgoff = bswap_64(event->mmap.pgoff);
|
||||
|
||||
if (sample_id_all) {
|
||||
void *data = &event->mmap.filename;
|
||||
|
||||
data += ALIGN(strlen(data) + 1, sizeof(u64));
|
||||
swap_sample_id_all(event, data);
|
||||
}
|
||||
}
|
||||
|
||||
static void perf_event__task_swap(union perf_event *event)
|
||||
static void perf_event__task_swap(union perf_event *event, bool sample_id_all)
|
||||
{
|
||||
event->fork.pid = bswap_32(event->fork.pid);
|
||||
event->fork.tid = bswap_32(event->fork.tid);
|
||||
event->fork.ppid = bswap_32(event->fork.ppid);
|
||||
event->fork.ptid = bswap_32(event->fork.ptid);
|
||||
event->fork.time = bswap_64(event->fork.time);
|
||||
|
||||
if (sample_id_all)
|
||||
swap_sample_id_all(event, &event->fork + 1);
|
||||
}
|
||||
|
||||
static void perf_event__read_swap(union perf_event *event)
|
||||
static void perf_event__read_swap(union perf_event *event, bool sample_id_all)
|
||||
{
|
||||
event->read.pid = bswap_32(event->read.pid);
|
||||
event->read.tid = bswap_32(event->read.tid);
|
||||
|
@ -479,6 +520,9 @@ static void perf_event__read_swap(union perf_event *event)
|
|||
event->read.time_enabled = bswap_64(event->read.time_enabled);
|
||||
event->read.time_running = bswap_64(event->read.time_running);
|
||||
event->read.id = bswap_64(event->read.id);
|
||||
|
||||
if (sample_id_all)
|
||||
swap_sample_id_all(event, &event->read + 1);
|
||||
}
|
||||
|
||||
static u8 revbyte(u8 b)
|
||||
|
@ -530,7 +574,8 @@ void perf_event__attr_swap(struct perf_event_attr *attr)
|
|||
swap_bitfield((u8 *) (&attr->read_format + 1), sizeof(u64));
|
||||
}
|
||||
|
||||
static void perf_event__hdr_attr_swap(union perf_event *event)
|
||||
static void perf_event__hdr_attr_swap(union perf_event *event,
|
||||
bool sample_id_all __used)
|
||||
{
|
||||
size_t size;
|
||||
|
||||
|
@ -541,18 +586,21 @@ static void perf_event__hdr_attr_swap(union perf_event *event)
|
|||
mem_bswap_64(event->attr.id, size);
|
||||
}
|
||||
|
||||
static void perf_event__event_type_swap(union perf_event *event)
|
||||
static void perf_event__event_type_swap(union perf_event *event,
|
||||
bool sample_id_all __used)
|
||||
{
|
||||
event->event_type.event_type.event_id =
|
||||
bswap_64(event->event_type.event_type.event_id);
|
||||
}
|
||||
|
||||
static void perf_event__tracing_data_swap(union perf_event *event)
|
||||
static void perf_event__tracing_data_swap(union perf_event *event,
|
||||
bool sample_id_all __used)
|
||||
{
|
||||
event->tracing_data.size = bswap_32(event->tracing_data.size);
|
||||
}
|
||||
|
||||
typedef void (*perf_event__swap_op)(union perf_event *event);
|
||||
typedef void (*perf_event__swap_op)(union perf_event *event,
|
||||
bool sample_id_all);
|
||||
|
||||
static perf_event__swap_op perf_event__swap_ops[] = {
|
||||
[PERF_RECORD_MMAP] = perf_event__mmap_swap,
|
||||
|
@ -986,6 +1034,15 @@ static int perf_session__process_user_event(struct perf_session *session, union
|
|||
}
|
||||
}
|
||||
|
||||
static void event_swap(union perf_event *event, bool sample_id_all)
|
||||
{
|
||||
perf_event__swap_op swap;
|
||||
|
||||
swap = perf_event__swap_ops[event->header.type];
|
||||
if (swap)
|
||||
swap(event, sample_id_all);
|
||||
}
|
||||
|
||||
static int perf_session__process_event(struct perf_session *session,
|
||||
union perf_event *event,
|
||||
struct perf_tool *tool,
|
||||
|
@ -994,9 +1051,8 @@ static int perf_session__process_event(struct perf_session *session,
|
|||
struct perf_sample sample;
|
||||
int ret;
|
||||
|
||||
if (session->header.needs_swap &&
|
||||
perf_event__swap_ops[event->header.type])
|
||||
perf_event__swap_ops[event->header.type](event);
|
||||
if (session->header.needs_swap)
|
||||
event_swap(event, session->sample_id_all);
|
||||
|
||||
if (event->header.type >= PERF_RECORD_HEADER_MAX)
|
||||
return -EINVAL;
|
||||
|
@ -1428,7 +1484,6 @@ void perf_event__print_ip(union perf_event *event, struct perf_sample *sample,
|
|||
int print_sym, int print_dso, int print_symoffset)
|
||||
{
|
||||
struct addr_location al;
|
||||
struct callchain_cursor *cursor = &evsel->hists.callchain_cursor;
|
||||
struct callchain_cursor_node *node;
|
||||
|
||||
if (perf_event__preprocess_sample(event, machine, &al, sample,
|
||||
|
@ -1446,10 +1501,10 @@ void perf_event__print_ip(union perf_event *event, struct perf_sample *sample,
|
|||
error("Failed to resolve callchain. Skipping\n");
|
||||
return;
|
||||
}
|
||||
callchain_cursor_commit(cursor);
|
||||
callchain_cursor_commit(&callchain_cursor);
|
||||
|
||||
while (1) {
|
||||
node = callchain_cursor_current(cursor);
|
||||
node = callchain_cursor_current(&callchain_cursor);
|
||||
if (!node)
|
||||
break;
|
||||
|
||||
|
@ -1460,12 +1515,12 @@ void perf_event__print_ip(union perf_event *event, struct perf_sample *sample,
|
|||
}
|
||||
if (print_dso) {
|
||||
printf(" (");
|
||||
map__fprintf_dsoname(al.map, stdout);
|
||||
map__fprintf_dsoname(node->map, stdout);
|
||||
printf(")");
|
||||
}
|
||||
printf("\n");
|
||||
|
||||
callchain_cursor_advance(cursor);
|
||||
callchain_cursor_advance(&callchain_cursor);
|
||||
}
|
||||
|
||||
} else {
|
||||
|
|
|
@ -323,6 +323,7 @@ struct dso *dso__new(const char *name)
|
|||
dso->sorted_by_name = 0;
|
||||
dso->has_build_id = 0;
|
||||
dso->kernel = DSO_TYPE_USER;
|
||||
dso->needs_swap = DSO_SWAP__UNSET;
|
||||
INIT_LIST_HEAD(&dso->node);
|
||||
}
|
||||
|
||||
|
@ -1156,6 +1157,33 @@ static size_t elf_addr_to_index(Elf *elf, GElf_Addr addr)
|
|||
return -1;
|
||||
}
|
||||
|
||||
static int dso__swap_init(struct dso *dso, unsigned char eidata)
|
||||
{
|
||||
static unsigned int const endian = 1;
|
||||
|
||||
dso->needs_swap = DSO_SWAP__NO;
|
||||
|
||||
switch (eidata) {
|
||||
case ELFDATA2LSB:
|
||||
/* We are big endian, DSO is little endian. */
|
||||
if (*(unsigned char const *)&endian != 1)
|
||||
dso->needs_swap = DSO_SWAP__YES;
|
||||
break;
|
||||
|
||||
case ELFDATA2MSB:
|
||||
/* We are little endian, DSO is big endian. */
|
||||
if (*(unsigned char const *)&endian != 0)
|
||||
dso->needs_swap = DSO_SWAP__YES;
|
||||
break;
|
||||
|
||||
default:
|
||||
pr_err("unrecognized DSO data encoding %d\n", eidata);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int dso__load_sym(struct dso *dso, struct map *map, const char *name,
|
||||
int fd, symbol_filter_t filter, int kmodule,
|
||||
int want_symtab)
|
||||
|
@ -1187,6 +1215,9 @@ static int dso__load_sym(struct dso *dso, struct map *map, const char *name,
|
|||
goto out_elf_end;
|
||||
}
|
||||
|
||||
if (dso__swap_init(dso, ehdr.e_ident[EI_DATA]))
|
||||
goto out_elf_end;
|
||||
|
||||
/* Always reject images with a mismatched build-id: */
|
||||
if (dso->has_build_id) {
|
||||
u8 build_id[BUILD_ID_SIZE];
|
||||
|
@ -1272,7 +1303,7 @@ static int dso__load_sym(struct dso *dso, struct map *map, const char *name,
|
|||
if (opdsec && sym.st_shndx == opdidx) {
|
||||
u32 offset = sym.st_value - opdshdr.sh_addr;
|
||||
u64 *opd = opddata->d_buf + offset;
|
||||
sym.st_value = *opd;
|
||||
sym.st_value = DSO__SWAP(dso, u64, *opd);
|
||||
sym.st_shndx = elf_addr_to_index(elf, sym.st_value);
|
||||
}
|
||||
|
||||
|
@ -2786,8 +2817,11 @@ int machine__load_vmlinux_path(struct machine *machine, enum map_type type,
|
|||
|
||||
struct map *dso__new_map(const char *name)
|
||||
{
|
||||
struct map *map = NULL;
|
||||
struct dso *dso = dso__new(name);
|
||||
struct map *map = map__new2(0, dso, MAP__FUNCTION);
|
||||
|
||||
if (dso)
|
||||
map = map__new2(0, dso, MAP__FUNCTION);
|
||||
|
||||
return map;
|
||||
}
|
||||
|
|
|
@ -9,6 +9,7 @@
|
|||
#include <linux/list.h>
|
||||
#include <linux/rbtree.h>
|
||||
#include <stdio.h>
|
||||
#include <byteswap.h>
|
||||
|
||||
#ifdef HAVE_CPLUS_DEMANGLE
|
||||
extern char *cplus_demangle(const char *, int);
|
||||
|
@ -160,11 +161,18 @@ enum dso_kernel_type {
|
|||
DSO_TYPE_GUEST_KERNEL
|
||||
};
|
||||
|
||||
enum dso_swap_type {
|
||||
DSO_SWAP__UNSET,
|
||||
DSO_SWAP__NO,
|
||||
DSO_SWAP__YES,
|
||||
};
|
||||
|
||||
struct dso {
|
||||
struct list_head node;
|
||||
struct rb_root symbols[MAP__NR_TYPES];
|
||||
struct rb_root symbol_names[MAP__NR_TYPES];
|
||||
enum dso_kernel_type kernel;
|
||||
enum dso_swap_type needs_swap;
|
||||
u8 adjust_symbols:1;
|
||||
u8 has_build_id:1;
|
||||
u8 hit:1;
|
||||
|
@ -182,6 +190,28 @@ struct dso {
|
|||
char name[0];
|
||||
};
|
||||
|
||||
#define DSO__SWAP(dso, type, val) \
|
||||
({ \
|
||||
type ____r = val; \
|
||||
BUG_ON(dso->needs_swap == DSO_SWAP__UNSET); \
|
||||
if (dso->needs_swap == DSO_SWAP__YES) { \
|
||||
switch (sizeof(____r)) { \
|
||||
case 2: \
|
||||
____r = bswap_16(val); \
|
||||
break; \
|
||||
case 4: \
|
||||
____r = bswap_32(val); \
|
||||
break; \
|
||||
case 8: \
|
||||
____r = bswap_64(val); \
|
||||
break; \
|
||||
default: \
|
||||
BUG_ON(1); \
|
||||
} \
|
||||
} \
|
||||
____r; \
|
||||
})
|
||||
|
||||
struct dso *dso__new(const char *name);
|
||||
void dso__delete(struct dso *dso);
|
||||
|
||||
|
|
Loading…
Reference in New Issue