perf tools: Fix various typos in comments
Fix ~124 single-word typos and a few spelling errors in the perf tooling code, accumulated over the years. Signed-off-by: Ingo Molnar <mingo@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Link: https://lore.kernel.org/r/20210321113734.GA248990@gmail.com Link: http://lore.kernel.org/lkml/20210323160915.GA61903@gmail.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
a6cb06ff49
commit
4d39c89f0b
|
@ -57,7 +57,7 @@ OPTIONS
|
|||
-u::
|
||||
--update=::
|
||||
Update specified file of the cache. Note that this doesn't remove
|
||||
older entires since those may be still needed for annotating old
|
||||
older entries since those may be still needed for annotating old
|
||||
(or remote) perf.data. Only if there is already a cache which has
|
||||
exactly same build-id, that is replaced by new one. It can be used
|
||||
to update kallsyms and kernel dso to vmlinux in order to support
|
||||
|
|
|
@ -475,7 +475,7 @@ OPTIONS
|
|||
but probably we'll make the default not to show the switch-on/off events
|
||||
on the --group mode and if there is only one event besides the off/on ones,
|
||||
go straight to the histogram browser, just like 'perf report' with no events
|
||||
explicitely specified does.
|
||||
explicitly specified does.
|
||||
|
||||
--itrace::
|
||||
Options for decoding instruction tracing data. The options are:
|
||||
|
|
|
@ -317,7 +317,7 @@ Default is to monitor all CPUS.
|
|||
but probably we'll make the default not to show the switch-on/off events
|
||||
on the --group mode and if there is only one event besides the off/on ones,
|
||||
go straight to the histogram browser, just like 'perf top' with no events
|
||||
explicitely specified does.
|
||||
explicitly specified does.
|
||||
|
||||
--stitch-lbr::
|
||||
Show callgraph with stitched LBRs, which may have more complete
|
||||
|
|
|
@ -378,7 +378,7 @@ static int cs_etm_recording_options(struct auxtrace_record *itr,
|
|||
opts->auxtrace_mmap_pages = roundup_pow_of_two(sz);
|
||||
}
|
||||
|
||||
/* Snapshost size can't be bigger than the auxtrace area */
|
||||
/* Snapshot size can't be bigger than the auxtrace area */
|
||||
if (opts->auxtrace_snapshot_size >
|
||||
opts->auxtrace_mmap_pages * (size_t)page_size) {
|
||||
pr_err("Snapshot size %zu must not be greater than AUX area tracing mmap size %zu\n",
|
||||
|
|
|
@ -6,11 +6,11 @@
|
|||
#include "debug.h"
|
||||
#include "symbol.h"
|
||||
|
||||
/* On arm64, kernel text segment start at high memory address,
|
||||
/* On arm64, kernel text segment starts at high memory address,
|
||||
* for example 0xffff 0000 8xxx xxxx. Modules start at a low memory
|
||||
* address, like 0xffff 0000 00ax xxxx. When only samll amount of
|
||||
* address, like 0xffff 0000 00ax xxxx. When only small amount of
|
||||
* memory is used by modules, gap between end of module's text segment
|
||||
* and start of kernel text segment may be reach 2G.
|
||||
* and start of kernel text segment may reach 2G.
|
||||
* Therefore do not fill this gap and do not assign it to the kernel dso map.
|
||||
*/
|
||||
|
||||
|
|
|
@ -108,7 +108,7 @@ int arch_sdt_arg_parse_op(char *old_op, char **new_op)
|
|||
/* [sp], [sp, NUM] or [sp,NUM] */
|
||||
new_len = 7; /* + ( % s p ) NULL */
|
||||
|
||||
/* If the arugment is [sp], need to fill offset '0' */
|
||||
/* If the argument is [sp], need to fill offset '0' */
|
||||
if (rm[2].rm_so == -1)
|
||||
new_len += 1;
|
||||
else
|
||||
|
|
|
@ -176,7 +176,7 @@ int cpu_isa_init(struct perf_kvm_stat *kvm, const char *cpuid __maybe_unused)
|
|||
}
|
||||
|
||||
/*
|
||||
* Incase of powerpc architecture, pmu registers are programmable
|
||||
* In case of powerpc architecture, pmu registers are programmable
|
||||
* by guest kernel. So monitoring guest via host may not provide
|
||||
* valid samples with default 'cycles' event. It is better to use
|
||||
* 'trace_imc/trace_cycles' event for guest profiling, since it
|
||||
|
|
|
@ -10,6 +10,6 @@
|
|||
|
||||
#define SPRN_PVR 0x11F /* Processor Version Register */
|
||||
#define PVR_VER(pvr) (((pvr) >> 16) & 0xFFFF) /* Version field */
|
||||
#define PVR_REV(pvr) (((pvr) >> 0) & 0xFFFF) /* Revison field */
|
||||
#define PVR_REV(pvr) (((pvr) >> 0) & 0xFFFF) /* Revision field */
|
||||
|
||||
#endif /* __PERF_UTIL_HEADER_H */
|
||||
|
|
|
@ -73,7 +73,7 @@ static int bp_modify1(void)
|
|||
/*
|
||||
* The parent does following steps:
|
||||
* - creates a new breakpoint (id 0) for bp_2 function
|
||||
* - changes that breakponit to bp_1 function
|
||||
* - changes that breakpoint to bp_1 function
|
||||
* - waits for the breakpoint to hit and checks
|
||||
* it has proper rip of bp_1 function
|
||||
* - detaches the child
|
||||
|
|
|
@ -165,7 +165,7 @@ static int sdt_init_op_regex(void)
|
|||
/*
|
||||
* Max x86 register name length is 5(ex: %r15d). So, 6th char
|
||||
* should always contain NULL. This helps to find register name
|
||||
* length using strlen, insted of maintaing one more variable.
|
||||
* length using strlen, instead of maintaining one more variable.
|
||||
*/
|
||||
#define SDT_REG_NAME_SIZE 6
|
||||
|
||||
|
@ -207,7 +207,7 @@ int arch_sdt_arg_parse_op(char *old_op, char **new_op)
|
|||
* and displacement 0 (Both sign and displacement 0 are
|
||||
* optional so it may be empty). Use one more character
|
||||
* to hold last NULL so that strlen can be used to find
|
||||
* prefix length, instead of maintaing one more variable.
|
||||
* prefix length, instead of maintaining one more variable.
|
||||
*/
|
||||
char prefix[3] = {0};
|
||||
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
* While the second model, enabled via --multiq option, uses multiple
|
||||
* queueing (which refers to one epoll instance per worker). For example,
|
||||
* short lived tcp connections in a high throughput httpd server will
|
||||
* ditribute the accept()'ing connections across CPUs. In this case each
|
||||
* distribute the accept()'ing connections across CPUs. In this case each
|
||||
* worker does a limited amount of processing.
|
||||
*
|
||||
* [queue A] ---> [worker]
|
||||
|
@ -198,7 +198,7 @@ static void *workerfn(void *arg)
|
|||
|
||||
do {
|
||||
/*
|
||||
* Block undefinitely waiting for the IN event.
|
||||
* Block indefinitely waiting for the IN event.
|
||||
* In order to stress the epoll_wait(2) syscall,
|
||||
* call it event per event, instead of a larger
|
||||
* batch (max)limit.
|
||||
|
|
|
@ -42,7 +42,7 @@
|
|||
#endif
|
||||
|
||||
/*
|
||||
* Regular printout to the terminal, supressed if -q is specified:
|
||||
* Regular printout to the terminal, suppressed if -q is specified:
|
||||
*/
|
||||
#define tprintf(x...) do { if (g && g->p.show_details >= 0) printf(x); } while (0)
|
||||
|
||||
|
|
|
@ -239,7 +239,7 @@ static int evsel__add_sample(struct evsel *evsel, struct perf_sample *sample,
|
|||
}
|
||||
|
||||
/*
|
||||
* XXX filtered samples can still have branch entires pointing into our
|
||||
* XXX filtered samples can still have branch entries pointing into our
|
||||
* symbol and are missed.
|
||||
*/
|
||||
process_branch_stack(sample->branch_stack, al, sample);
|
||||
|
|
|
@ -1796,7 +1796,7 @@ static int ui_init(void)
|
|||
data__for_each_file(i, d) {
|
||||
|
||||
/*
|
||||
* Baseline or compute realted columns:
|
||||
* Baseline or compute related columns:
|
||||
*
|
||||
* PERF_HPP_DIFF__BASELINE
|
||||
* PERF_HPP_DIFF__DELTA
|
||||
|
|
|
@ -49,7 +49,7 @@ struct lock_stat {
|
|||
|
||||
/*
|
||||
* FIXME: evsel__intval() returns u64,
|
||||
* so address of lockdep_map should be dealed as 64bit.
|
||||
* so address of lockdep_map should be treated as 64bit.
|
||||
* Is there more better solution?
|
||||
*/
|
||||
void *addr; /* address of lockdep_map, used as ID */
|
||||
|
|
|
@ -1712,7 +1712,7 @@ static int perf_sched__process_fork_event(struct perf_tool *tool,
|
|||
{
|
||||
struct perf_sched *sched = container_of(tool, struct perf_sched, tool);
|
||||
|
||||
/* run the fork event through the perf machineruy */
|
||||
/* run the fork event through the perf machinery */
|
||||
perf_event__process_fork(tool, event, sample, machine);
|
||||
|
||||
/* and then run additional processing needed for this command */
|
||||
|
|
|
@ -2486,7 +2486,7 @@ static int perf_script__fopen_per_event_dump(struct perf_script *script)
|
|||
/*
|
||||
* Already setup? I.e. we may be called twice in cases like
|
||||
* Intel PT, one for the intel_pt// and dummy events, then
|
||||
* for the evsels syntheized from the auxtrace info.
|
||||
* for the evsels synthesized from the auxtrace info.
|
||||
*
|
||||
* Ses perf_script__process_auxtrace_info.
|
||||
*/
|
||||
|
@ -3083,7 +3083,7 @@ static int list_available_scripts(const struct option *opt __maybe_unused,
|
|||
*
|
||||
* Fixme: All existing "xxx-record" are all in good formats "-e event ",
|
||||
* which is covered well now. And new parsing code should be added to
|
||||
* cover the future complexing formats like event groups etc.
|
||||
* cover the future complex formats like event groups etc.
|
||||
*/
|
||||
static int check_ev_match(char *dir_name, char *scriptname,
|
||||
struct perf_session *session)
|
||||
|
|
|
@ -1705,7 +1705,7 @@ static int add_default_attributes(void)
|
|||
bzero(&errinfo, sizeof(errinfo));
|
||||
if (transaction_run) {
|
||||
/* Handle -T as -M transaction. Once platform specific metrics
|
||||
* support has been added to the json files, all archictures
|
||||
* support has been added to the json files, all architectures
|
||||
* will use this approach. To determine transaction support
|
||||
* on an architecture test for such a metric name.
|
||||
*/
|
||||
|
@ -2459,7 +2459,7 @@ int cmd_stat(int argc, const char **argv)
|
|||
/*
|
||||
* We synthesize the kernel mmap record just so that older tools
|
||||
* don't emit warnings about not being able to resolve symbols
|
||||
* due to /proc/sys/kernel/kptr_restrict settings and instear provide
|
||||
* due to /proc/sys/kernel/kptr_restrict settings and instead provide
|
||||
* a saner message about no samples being in the perf.data file.
|
||||
*
|
||||
* This also serves to suppress a warning about f_header.data.size == 0
|
||||
|
|
|
@ -1607,7 +1607,7 @@ int cmd_top(int argc, const char **argv)
|
|||
if (status) {
|
||||
/*
|
||||
* Some arches do not provide a get_cpuid(), so just use pr_debug, otherwise
|
||||
* warn the user explicitely.
|
||||
* warn the user explicitly.
|
||||
*/
|
||||
eprintf(status == ENOSYS ? 1 : 0, verbose,
|
||||
"Couldn't read the cpuid for this machine: %s\n",
|
||||
|
|
|
@ -262,7 +262,7 @@ int sys_enter(struct syscall_enter_args *args)
|
|||
/*
|
||||
* Jump to syscall specific augmenter, even if the default one,
|
||||
* "!raw_syscalls:unaugmented" that will just return 1 to return the
|
||||
* unagmented tracepoint payload.
|
||||
* unaugmented tracepoint payload.
|
||||
*/
|
||||
bpf_tail_call(args, &syscalls_sys_enter, augmented_args->args.syscall_nr);
|
||||
|
||||
|
@ -282,7 +282,7 @@ int sys_exit(struct syscall_exit_args *args)
|
|||
/*
|
||||
* Jump to syscall specific return augmenter, even if the default one,
|
||||
* "!raw_syscalls:unaugmented" that will just return 1 to return the
|
||||
* unagmented tracepoint payload.
|
||||
* unaugmented tracepoint payload.
|
||||
*/
|
||||
bpf_tail_call(args, &syscalls_sys_exit, exit_args.syscall_nr);
|
||||
/*
|
||||
|
|
|
@ -390,7 +390,7 @@ jvmti_write_code(void *agent, char const *sym,
|
|||
rec.p.total_size += size;
|
||||
|
||||
/*
|
||||
* If JVM is multi-threaded, nultiple concurrent calls to agent
|
||||
* If JVM is multi-threaded, multiple concurrent calls to agent
|
||||
* may be possible, so protect file writes
|
||||
*/
|
||||
flockfile(fp);
|
||||
|
@ -457,7 +457,7 @@ jvmti_write_debug_info(void *agent, uint64_t code,
|
|||
rec.p.total_size = size;
|
||||
|
||||
/*
|
||||
* If JVM is multi-threaded, nultiple concurrent calls to agent
|
||||
* If JVM is multi-threaded, multiple concurrent calls to agent
|
||||
* may be possible, so protect file writes
|
||||
*/
|
||||
flockfile(fp);
|
||||
|
|
|
@ -885,37 +885,37 @@
|
|||
"MetricName": "flush_rate_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization (11 to 14) as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization (11 to 14) as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_11_14_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_11to14_slots_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization (15 to 17) as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization (15 to 17) as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_15_17_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_15to17_slots_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization 18+ as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization 18+ as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_18_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_18plus_slots_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization (1 to 2) as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization (1 to 2) as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_1_2_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_1to2_slots_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization (3 to 6) as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization (3 to 6) as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_3_6_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_3to6_slots_percent"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "GCT slot utilization (7 to 10) as a % of cycles this thread had atleast 1 slot valid",
|
||||
"BriefDescription": "GCT slot utilization (7 to 10) as a % of cycles this thread had at least 1 slot valid",
|
||||
"MetricExpr": "PM_GCT_UTIL_7_10_ENTRIES / ( PM_RUN_CYC - PM_GCT_NOSLOT_CYC) * 100",
|
||||
"MetricGroup": "general",
|
||||
"MetricName": "gct_util_7to10_slots_percent"
|
||||
|
|
|
@ -1691,7 +1691,7 @@
|
|||
"MetricName": "custom_secs"
|
||||
},
|
||||
{
|
||||
"BriefDescription": "Percentage Cycles atleast one instruction dispatched",
|
||||
"BriefDescription": "Percentage Cycles at least one instruction dispatched",
|
||||
"MetricExpr": "PM_1PLUS_PPC_DISP / PM_CYC * 100",
|
||||
"MetricName": "cycles_atleast_one_inst_dispatched_percent"
|
||||
},
|
||||
|
|
|
@ -1149,7 +1149,7 @@ static int process_one_file(const char *fpath, const struct stat *sb,
|
|||
* and directory tree could result in build failure due to table
|
||||
* names not being found.
|
||||
*
|
||||
* Atleast for now, be strict with processing JSON file names.
|
||||
* At least for now, be strict with processing JSON file names.
|
||||
* i.e. if JSON file name cannot be mapped to C-style table name,
|
||||
* fail.
|
||||
*/
|
||||
|
|
|
@ -356,7 +356,7 @@ def handle_irq_softirq_exit(event_info):
|
|||
return
|
||||
rec_data = {'sirq_ent_t':sirq_ent_t, 'sirq_ext_t':time,
|
||||
'irq_list':irq_list, 'event_list':event_list}
|
||||
# merge information realted to a NET_RX softirq
|
||||
# merge information related to a NET_RX softirq
|
||||
receive_hunk_list.append(rec_data)
|
||||
|
||||
def handle_napi_poll(event_info):
|
||||
|
|
|
@ -225,11 +225,11 @@ int test__bp_signal(struct test *test __maybe_unused, int subtest __maybe_unused
|
|||
*
|
||||
* The test case check following error conditions:
|
||||
* - we get stuck in signal handler because of debug
|
||||
* exception being triggered receursively due to
|
||||
* exception being triggered recursively due to
|
||||
* the wrong RF EFLAG management
|
||||
*
|
||||
* - we never trigger the sig_handler breakpoint due
|
||||
* to the rong RF EFLAG management
|
||||
* to the wrong RF EFLAG management
|
||||
*
|
||||
*/
|
||||
|
||||
|
@ -242,7 +242,7 @@ int test__bp_signal(struct test *test __maybe_unused, int subtest __maybe_unused
|
|||
ioctl(fd3, PERF_EVENT_IOC_ENABLE, 0);
|
||||
|
||||
/*
|
||||
* Kick off the test by trigering 'fd1'
|
||||
* Kick off the test by triggering 'fd1'
|
||||
* breakpoint.
|
||||
*/
|
||||
test_function();
|
||||
|
|
|
@ -658,7 +658,7 @@ static int do_test_code_reading(bool try_kcore)
|
|||
/*
|
||||
* Both cpus and threads are now owned by evlist
|
||||
* and will be freed by following perf_evlist__set_maps
|
||||
* call. Getting refference to keep them alive.
|
||||
* call. Getting reference to keep them alive.
|
||||
*/
|
||||
perf_cpu_map__get(cpus);
|
||||
perf_thread_map__get(threads);
|
||||
|
|
|
@ -47,7 +47,7 @@ static struct sample fake_samples[] = {
|
|||
};
|
||||
|
||||
/*
|
||||
* Will be casted to struct ip_callchain which has all 64 bit entries
|
||||
* Will be cast to struct ip_callchain which has all 64 bit entries
|
||||
* of nr and ips[].
|
||||
*/
|
||||
static u64 fake_callchains[][10] = {
|
||||
|
@ -297,7 +297,7 @@ out:
|
|||
return err;
|
||||
}
|
||||
|
||||
/* callcain + NO children */
|
||||
/* callchain + NO children */
|
||||
static int test2(struct evsel *evsel, struct machine *machine)
|
||||
{
|
||||
int err;
|
||||
|
|
|
@ -20,7 +20,7 @@
|
|||
|
||||
#if defined(__s390x__)
|
||||
/* Return true if kvm module is available and loaded. Test this
|
||||
* and retun success when trace point kvm_s390_create_vm
|
||||
* and return success when trace point kvm_s390_create_vm
|
||||
* exists. Otherwise this test always fails.
|
||||
*/
|
||||
static bool kvm_s390_create_vm_valid(void)
|
||||
|
|
|
@ -186,7 +186,7 @@ static int __compute_metric(const char *name, struct value *vals,
|
|||
*ratio2 = compute_single(&metric_events, evlist, &st, name2);
|
||||
|
||||
out:
|
||||
/* ... clenup. */
|
||||
/* ... cleanup. */
|
||||
metricgroup__rblist_exit(&metric_events);
|
||||
runtime_stat__exit(&st);
|
||||
evlist__free_stats(evlist);
|
||||
|
|
|
@ -80,7 +80,7 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map)
|
|||
* CPU 1 is on core_id 1 and physical_package_id 3
|
||||
*
|
||||
* Core_id and physical_package_id are platform and architecture
|
||||
* dependend and might have higher numbers than the CPU id.
|
||||
* dependent and might have higher numbers than the CPU id.
|
||||
* This actually depends on the configuration.
|
||||
*
|
||||
* In this case process_cpu_topology() prints error message:
|
||||
|
|
|
@ -85,7 +85,7 @@ struct mmsghdr {
|
|||
|
||||
/*
|
||||
* POSIX 1003.1g - ancillary data object information
|
||||
* Ancillary data consits of a sequence of pairs of
|
||||
* Ancillary data consists of a sequence of pairs of
|
||||
* (cmsghdr, cmsg_data[])
|
||||
*/
|
||||
|
||||
|
|
|
@ -415,7 +415,7 @@ static int sym_title(struct symbol *sym, struct map *map, char *title,
|
|||
}
|
||||
|
||||
/*
|
||||
* This can be called from external jumps, i.e. jumps from one functon
|
||||
* This can be called from external jumps, i.e. jumps from one function
|
||||
* to another, like from the kernel's entry_SYSCALL_64 function to the
|
||||
* swapgs_restore_regs_and_return_to_usermode() function.
|
||||
*
|
||||
|
|
|
@ -117,7 +117,7 @@ static void hist_browser__update_rows(struct hist_browser *hb)
|
|||
browser->rows -= browser->extra_title_lines;
|
||||
/*
|
||||
* Verify if we were at the last line and that line isn't
|
||||
* visibe because we now show the header line(s).
|
||||
* visible because we now show the header line(s).
|
||||
*/
|
||||
index_row = browser->index - browser->top_idx;
|
||||
if (index_row >= browser->rows)
|
||||
|
|
|
@ -671,7 +671,7 @@ int bpf__probe(struct bpf_object *obj)
|
|||
* After probing, let's consider prologue, which
|
||||
* adds program fetcher to BPF programs.
|
||||
*
|
||||
* hook_load_preprocessorr() hooks pre-processor
|
||||
* hook_load_preprocessor() hooks pre-processor
|
||||
* to bpf_program, let it generate prologue
|
||||
* dynamically during loading.
|
||||
*/
|
||||
|
|
|
@ -23,7 +23,7 @@
|
|||
* @children: tree of call paths of functions called
|
||||
*
|
||||
* In combination with the call_return structure, the call_path structure
|
||||
* defines a context-sensitve call-graph.
|
||||
* defines a context-sensitive call-graph.
|
||||
*/
|
||||
struct call_path {
|
||||
struct call_path *parent;
|
||||
|
|
|
@ -877,7 +877,7 @@ append_chain_children(struct callchain_node *root,
|
|||
if (!node)
|
||||
return -1;
|
||||
|
||||
/* lookup in childrens */
|
||||
/* lookup in children */
|
||||
while (*p) {
|
||||
enum match_result ret;
|
||||
|
||||
|
|
|
@ -699,7 +699,7 @@ static int collect_config(const char *var, const char *value,
|
|||
/* perf_config_set can contain both user and system config items.
|
||||
* So we should know where each value is from.
|
||||
* The classification would be needed when a particular config file
|
||||
* is overwrited by setting feature i.e. set_config().
|
||||
* is overwritten by setting feature i.e. set_config().
|
||||
*/
|
||||
if (strcmp(config_file_name, perf_etc_perfconfig()) == 0) {
|
||||
section->from_system_config = true;
|
||||
|
|
|
@ -317,7 +317,7 @@ cs_etm_decoder__do_hard_timestamp(struct cs_etm_queue *etmq,
|
|||
* This is the first timestamp we've seen since the beginning of traces
|
||||
* or a discontinuity. Since timestamps packets are generated *after*
|
||||
* range packets have been generated, we need to estimate the time at
|
||||
* which instructions started by substracting the number of instructions
|
||||
* which instructions started by subtracting the number of instructions
|
||||
* executed to the timestamp.
|
||||
*/
|
||||
packet_queue->timestamp = elem->timestamp - packet_queue->instr_count;
|
||||
|
|
|
@ -202,7 +202,7 @@ void cs_etm__etmq_set_traceid_queue_timestamp(struct cs_etm_queue *etmq,
|
|||
u8 trace_chan_id)
|
||||
{
|
||||
/*
|
||||
* Wnen a timestamp packet is encountered the backend code
|
||||
* When a timestamp packet is encountered the backend code
|
||||
* is stopped so that the front end has time to process packets
|
||||
* that were accumulated in the traceID queue. Since there can
|
||||
* be more than one channel per cs_etm_queue, we need to specify
|
||||
|
@ -1697,7 +1697,7 @@ static bool cs_etm__is_svc_instr(struct cs_etm_queue *etmq, u8 trace_chan_id,
|
|||
* | 1 1 0 1 1 1 1 1 | imm8 |
|
||||
* +-----------------+--------+
|
||||
*
|
||||
* According to the specifiction, it only defines SVC for T32
|
||||
* According to the specification, it only defines SVC for T32
|
||||
* with 16 bits instruction and has no definition for 32bits;
|
||||
* so below only read 2 bytes as instruction size for T32.
|
||||
*/
|
||||
|
@ -1929,7 +1929,7 @@ static int cs_etm__set_sample_flags(struct cs_etm_queue *etmq,
|
|||
|
||||
/*
|
||||
* If the previous packet is an exception return packet
|
||||
* and the return address just follows SVC instuction,
|
||||
* and the return address just follows SVC instruction,
|
||||
* it needs to calibrate the previous packet sample flags
|
||||
* as PERF_IP_FLAG_SYSCALLRET.
|
||||
*/
|
||||
|
@ -2003,7 +2003,7 @@ static int cs_etm__set_sample_flags(struct cs_etm_queue *etmq,
|
|||
* contain exception type related info so we cannot decide
|
||||
* the exception type purely based on exception return packet.
|
||||
* If we record the exception number from exception packet and
|
||||
* reuse it for excpetion return packet, this is not reliable
|
||||
* reuse it for exception return packet, this is not reliable
|
||||
* due the trace can be discontinuity or the interrupt can
|
||||
* be nested, thus the recorded exception number cannot be
|
||||
* used for exception return packet for these two cases.
|
||||
|
|
|
@ -12,7 +12,8 @@
|
|||
|
||||
struct perf_session;
|
||||
|
||||
/* Versionning header in case things need tro change in the future. That way
|
||||
/*
|
||||
* Versioning header in case things need to change in the future. That way
|
||||
* decoding of old snapshot is still possible.
|
||||
*/
|
||||
enum {
|
||||
|
@ -77,7 +78,7 @@ enum {
|
|||
|
||||
/*
|
||||
* ETMv3 exception encoding number:
|
||||
* See Embedded Trace Macrocell spcification (ARM IHI 0014Q)
|
||||
* See Embedded Trace Macrocell specification (ARM IHI 0014Q)
|
||||
* table 7-12 Encoding of Exception[3:0] for non-ARMv7-M processors.
|
||||
*/
|
||||
enum {
|
||||
|
|
|
@ -949,7 +949,7 @@ static char *change_name(char *name, char *orig_name, int dup)
|
|||
/*
|
||||
* Add '_' prefix to potential keywork. According to
|
||||
* Mathieu Desnoyers (https://lore.kernel.org/lkml/1074266107.40857.1422045946295.JavaMail.zimbra@efficios.com),
|
||||
* futher CTF spec updating may require us to use '$'.
|
||||
* further CTF spec updating may require us to use '$'.
|
||||
*/
|
||||
if (dup < 0)
|
||||
len = strlen(name) + sizeof("_");
|
||||
|
|
|
@ -147,7 +147,7 @@ error:
|
|||
* Demangle Java function signature (openJDK, not GCJ)
|
||||
* input:
|
||||
* str: string to parse. String is not modified
|
||||
* flags: comobination of JAVA_DEMANGLE_* flags to modify demangling
|
||||
* flags: combination of JAVA_DEMANGLE_* flags to modify demangling
|
||||
* return:
|
||||
* if input can be demangled, then a newly allocated string is returned.
|
||||
* if input cannot be demangled, then NULL is returned
|
||||
|
@ -164,7 +164,7 @@ java_demangle_sym(const char *str, int flags)
|
|||
if (!str)
|
||||
return NULL;
|
||||
|
||||
/* find start of retunr type */
|
||||
/* find start of return type */
|
||||
p = strrchr(str, ')');
|
||||
if (!p)
|
||||
return NULL;
|
||||
|
|
|
@ -216,7 +216,7 @@ struct dso {
|
|||
|
||||
/* dso__for_each_symbol - iterate over the symbols of given type
|
||||
*
|
||||
* @dso: the 'struct dso *' in which symbols itereated
|
||||
* @dso: the 'struct dso *' in which symbols are iterated
|
||||
* @pos: the 'struct symbol *' to use as a loop cursor
|
||||
* @n: the 'struct rb_node *' to use as a temporary storage
|
||||
*/
|
||||
|
|
|
@ -91,7 +91,7 @@ static Dwarf_Line *cu_getsrc_die(Dwarf_Die *cu_die, Dwarf_Addr addr)
|
|||
return NULL;
|
||||
} while (laddr == addr);
|
||||
l++;
|
||||
/* Going foward to find the statement line */
|
||||
/* Going forward to find the statement line */
|
||||
do {
|
||||
line = dwarf_onesrcline(lines, l++);
|
||||
if (!line || dwarf_lineaddr(line, &laddr) != 0 ||
|
||||
|
@ -177,7 +177,7 @@ int cu_walk_functions_at(Dwarf_Die *cu_die, Dwarf_Addr addr,
|
|||
* die_get_linkage_name - Get the linkage name of the object
|
||||
* @dw_die: A DIE of the object
|
||||
*
|
||||
* Get the linkage name attiribute of given @dw_die.
|
||||
* Get the linkage name attribute of given @dw_die.
|
||||
* For C++ binary, the linkage name will be the mangled symbol.
|
||||
*/
|
||||
const char *die_get_linkage_name(Dwarf_Die *dw_die)
|
||||
|
@ -739,7 +739,7 @@ static int __die_walk_instances_cb(Dwarf_Die *inst, void *data)
|
|||
* @data: user data
|
||||
*
|
||||
* Walk on the instances of give @in_die. @in_die must be an inlined function
|
||||
* declartion. This returns the return value of @callback if it returns
|
||||
* declaration. This returns the return value of @callback if it returns
|
||||
* non-zero value, or -ENOENT if there is no instance.
|
||||
*/
|
||||
int die_walk_instances(Dwarf_Die *or_die, int (*callback)(Dwarf_Die *, void *),
|
||||
|
|
|
@ -22,7 +22,7 @@ const char *cu_get_comp_dir(Dwarf_Die *cu_die);
|
|||
int cu_find_lineinfo(Dwarf_Die *cudie, unsigned long addr,
|
||||
const char **fname, int *lineno);
|
||||
|
||||
/* Walk on funcitons at given address */
|
||||
/* Walk on functions at given address */
|
||||
int cu_walk_functions_at(Dwarf_Die *cu_die, Dwarf_Addr addr,
|
||||
int (*callback)(Dwarf_Die *, void *), void *data);
|
||||
|
||||
|
|
|
@ -21,7 +21,7 @@
|
|||
* all struct perf_record_lost_samples.lost fields reported.
|
||||
*
|
||||
* The total_period is needed because by default auto-freq is used, so
|
||||
* multipling nr_events[PERF_EVENT_SAMPLE] by a frequency isn't possible to get
|
||||
* multiplying nr_events[PERF_EVENT_SAMPLE] by a frequency isn't possible to get
|
||||
* the total number of low level events, it is necessary to to sum all struct
|
||||
* perf_record_sample.period and stash the result in total_period.
|
||||
*/
|
||||
|
|
|
@ -1210,7 +1210,7 @@ bool evlist__valid_read_format(struct evlist *evlist)
|
|||
}
|
||||
}
|
||||
|
||||
/* PERF_SAMPLE_READ imples PERF_FORMAT_ID. */
|
||||
/* PERF_SAMPLE_READ implies PERF_FORMAT_ID. */
|
||||
if ((sample_type & PERF_SAMPLE_READ) &&
|
||||
!(read_format & PERF_FORMAT_ID)) {
|
||||
return false;
|
||||
|
|
|
@ -621,7 +621,7 @@ const char *evsel__hw_cache_result[PERF_COUNT_HW_CACHE_RESULT_MAX][EVSEL__MAX_AL
|
|||
#define COP(x) (1 << x)
|
||||
|
||||
/*
|
||||
* cache operartion stat
|
||||
* cache operation stat
|
||||
* L1I : Read and prefetch only
|
||||
* ITLB and BPU : Read-only
|
||||
*/
|
||||
|
@ -2275,7 +2275,7 @@ int evsel__parse_sample(struct evsel *evsel, union perf_event *event,
|
|||
/*
|
||||
* Undo swap of u64, then swap on individual u32s,
|
||||
* get the size of the raw area and undo all of the
|
||||
* swap. The pevent interface handles endianity by
|
||||
* swap. The pevent interface handles endianness by
|
||||
* itself.
|
||||
*/
|
||||
if (swapped) {
|
||||
|
|
|
@ -3,7 +3,7 @@
|
|||
#define PARSE_CTX_H 1
|
||||
|
||||
// There are fixes that need to land upstream before we can use libbpf's headers,
|
||||
// for now use our copy uncoditionally, since the data structures at this point
|
||||
// for now use our copy unconditionally, since the data structures at this point
|
||||
// are exactly the same, no problem.
|
||||
//#ifdef HAVE_LIBBPF_SUPPORT
|
||||
//#include <bpf/hashmap.h>
|
||||
|
|
|
@ -127,7 +127,7 @@ static int __do_write_buf(struct feat_fd *ff, const void *buf, size_t size)
|
|||
return 0;
|
||||
}
|
||||
|
||||
/* Return: 0 if succeded, -ERR if failed. */
|
||||
/* Return: 0 if succeeded, -ERR if failed. */
|
||||
int do_write(struct feat_fd *ff, const void *buf, size_t size)
|
||||
{
|
||||
if (!ff->buf)
|
||||
|
@ -135,7 +135,7 @@ int do_write(struct feat_fd *ff, const void *buf, size_t size)
|
|||
return __do_write_buf(ff, buf, size);
|
||||
}
|
||||
|
||||
/* Return: 0 if succeded, -ERR if failed. */
|
||||
/* Return: 0 if succeeded, -ERR if failed. */
|
||||
static int do_write_bitmap(struct feat_fd *ff, unsigned long *set, u64 size)
|
||||
{
|
||||
u64 *p = (u64 *) set;
|
||||
|
@ -154,7 +154,7 @@ static int do_write_bitmap(struct feat_fd *ff, unsigned long *set, u64 size)
|
|||
return 0;
|
||||
}
|
||||
|
||||
/* Return: 0 if succeded, -ERR if failed. */
|
||||
/* Return: 0 if succeeded, -ERR if failed. */
|
||||
int write_padded(struct feat_fd *ff, const void *bf,
|
||||
size_t count, size_t count_aligned)
|
||||
{
|
||||
|
@ -170,7 +170,7 @@ int write_padded(struct feat_fd *ff, const void *bf,
|
|||
#define string_size(str) \
|
||||
(PERF_ALIGN((strlen(str) + 1), NAME_ALIGN) + sizeof(u32))
|
||||
|
||||
/* Return: 0 if succeded, -ERR if failed. */
|
||||
/* Return: 0 if succeeded, -ERR if failed. */
|
||||
static int do_write_string(struct feat_fd *ff, const char *str)
|
||||
{
|
||||
u32 len, olen;
|
||||
|
@ -266,7 +266,7 @@ static char *do_read_string(struct feat_fd *ff)
|
|||
return NULL;
|
||||
}
|
||||
|
||||
/* Return: 0 if succeded, -ERR if failed. */
|
||||
/* Return: 0 if succeeded, -ERR if failed. */
|
||||
static int do_read_bitmap(struct feat_fd *ff, unsigned long **pset, u64 *psize)
|
||||
{
|
||||
unsigned long *set;
|
||||
|
@ -2874,7 +2874,7 @@ static int process_bpf_prog_info(struct feat_fd *ff, void *data __maybe_unused)
|
|||
int err = -1;
|
||||
|
||||
if (ff->ph->needs_swap) {
|
||||
pr_warning("interpreting bpf_prog_info from systems with endianity is not yet supported\n");
|
||||
pr_warning("interpreting bpf_prog_info from systems with endianness is not yet supported\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -2942,7 +2942,7 @@ static int process_bpf_btf(struct feat_fd *ff, void *data __maybe_unused)
|
|||
int err = -1;
|
||||
|
||||
if (ff->ph->needs_swap) {
|
||||
pr_warning("interpreting btf from systems with endianity is not yet supported\n");
|
||||
pr_warning("interpreting btf from systems with endianness is not yet supported\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -3481,11 +3481,11 @@ static const size_t attr_pipe_abi_sizes[] = {
|
|||
};
|
||||
|
||||
/*
|
||||
* In the legacy pipe format, there is an implicit assumption that endiannesss
|
||||
* In the legacy pipe format, there is an implicit assumption that endianness
|
||||
* between host recording the samples, and host parsing the samples is the
|
||||
* same. This is not always the case given that the pipe output may always be
|
||||
* redirected into a file and analyzed on a different machine with possibly a
|
||||
* different endianness and perf_event ABI revsions in the perf tool itself.
|
||||
* different endianness and perf_event ABI revisions in the perf tool itself.
|
||||
*/
|
||||
static int try_all_pipe_abis(uint64_t hdr_sz, struct perf_header *ph)
|
||||
{
|
||||
|
|
|
@ -3569,7 +3569,7 @@ int intel_pt_process_auxtrace_info(union perf_event *event,
|
|||
/*
|
||||
* Since this thread will not be kept in any rbtree not in a
|
||||
* list, initialize its list node so that at thread__put() the
|
||||
* current thread lifetime assuption is kept and we don't segfault
|
||||
* current thread lifetime assumption is kept and we don't segfault
|
||||
* at list_del_init().
|
||||
*/
|
||||
INIT_LIST_HEAD(&pt->unknown_thread->node);
|
||||
|
|
|
@ -30,7 +30,7 @@
|
|||
*
|
||||
* It does so by calculating the costs of the path ending in characters
|
||||
* i (in string1) and j (in string2), respectively, given that the last
|
||||
* operation is a substition, a swap, a deletion, or an insertion.
|
||||
* operation is a substitution, a swap, a deletion, or an insertion.
|
||||
*
|
||||
* This implementation allows the costs to be weighted:
|
||||
*
|
||||
|
|
|
@ -4,7 +4,7 @@
|
|||
* generic one.
|
||||
*
|
||||
* The function 'LIBUNWIND__ARCH_REG_ID' name is set according to arch
|
||||
* name and the defination of this function is included directly from
|
||||
* name and the definition of this function is included directly from
|
||||
* 'arch/arm64/util/unwind-libunwind.c', to make sure that this function
|
||||
* is defined no matter what arch the host is.
|
||||
*
|
||||
|
|
|
@ -4,7 +4,7 @@
|
|||
* generic one.
|
||||
*
|
||||
* The function 'LIBUNWIND__ARCH_REG_ID' name is set according to arch
|
||||
* name and the defination of this function is included directly from
|
||||
* name and the definition of this function is included directly from
|
||||
* 'arch/x86/util/unwind-libunwind.c', to make sure that this function
|
||||
* is defined no matter what arch the host is.
|
||||
*
|
||||
|
|
|
@ -471,7 +471,7 @@ int llvm__compile_bpf(const char *path, void **p_obj_buf,
|
|||
|
||||
/*
|
||||
* This is an optional work. Even it fail we can continue our
|
||||
* work. Needn't to check error return.
|
||||
* work. Needn't check error return.
|
||||
*/
|
||||
llvm__get_kbuild_opts(&kbuild_dir, &kbuild_include_opts);
|
||||
|
||||
|
|
|
@ -905,7 +905,7 @@ static struct map *machine__addnew_module_map(struct machine *machine, u64 start
|
|||
|
||||
maps__insert(&machine->kmaps, map);
|
||||
|
||||
/* Put the map here because maps__insert alread got it */
|
||||
/* Put the map here because maps__insert already got it */
|
||||
map__put(map);
|
||||
out:
|
||||
/* put the dso here, corresponding to machine__findnew_module_dso */
|
||||
|
@ -1952,7 +1952,7 @@ int machine__process_fork_event(struct machine *machine, union perf_event *event
|
|||
* maps because that is what the kernel just did.
|
||||
*
|
||||
* But when synthesizing, this should not be done. If we do, we end up
|
||||
* with overlapping maps as we process the sythesized MMAP2 events that
|
||||
* with overlapping maps as we process the synthesized MMAP2 events that
|
||||
* get delivered shortly thereafter.
|
||||
*
|
||||
* Use the FORK event misc flags in an internal way to signal this
|
||||
|
@ -2518,7 +2518,7 @@ static bool has_stitched_lbr(struct thread *thread,
|
|||
|
||||
/*
|
||||
* Check if there are identical LBRs between two samples.
|
||||
* Identicall LBRs must have same from, to and flags values. Also,
|
||||
* Identical LBRs must have same from, to and flags values. Also,
|
||||
* they have to be saved in the same LBR registers (same physical
|
||||
* index).
|
||||
*
|
||||
|
@ -2588,7 +2588,7 @@ err:
|
|||
}
|
||||
|
||||
/*
|
||||
* Recolve LBR callstack chain sample
|
||||
* Resolve LBR callstack chain sample
|
||||
* Return:
|
||||
* 1 on success get LBR callchain information
|
||||
* 0 no available LBR callchain information, should try fp
|
||||
|
|
|
@ -75,7 +75,7 @@ struct thread;
|
|||
|
||||
/* map__for_each_symbol - iterate over the symbols in the given map
|
||||
*
|
||||
* @map: the 'struct map *' in which symbols itereated
|
||||
* @map: the 'struct map *' in which symbols are iterated
|
||||
* @pos: the 'struct symbol *' to use as a loop cursor
|
||||
* @n: the 'struct rb_node *' to use as a temporary storage
|
||||
* Note: caller must ensure map->dso is not NULL (map is loaded).
|
||||
|
@ -86,7 +86,7 @@ struct thread;
|
|||
/* map__for_each_symbol_with_name - iterate over the symbols in the given map
|
||||
* that have the given name
|
||||
*
|
||||
* @map: the 'struct map *' in which symbols itereated
|
||||
* @map: the 'struct map *' in which symbols are iterated
|
||||
* @sym_name: the symbol name
|
||||
* @pos: the 'struct symbol *' to use as a loop cursor
|
||||
*/
|
||||
|
|
|
@ -81,7 +81,7 @@ struct c2c_stats {
|
|||
u32 rmt_dram; /* count of loads miss to remote DRAM */
|
||||
u32 blk_data; /* count of loads blocked by data */
|
||||
u32 blk_addr; /* count of loads blocked by address conflict */
|
||||
u32 nomap; /* count of load/stores with no phys adrs */
|
||||
u32 nomap; /* count of load/stores with no phys addrs */
|
||||
u32 noparse; /* count of unparsable data sources */
|
||||
};
|
||||
|
||||
|
|
|
@ -181,7 +181,7 @@ static bool evsel_same_pmu(struct evsel *ev1, struct evsel *ev2)
|
|||
* @pctx: the parse context for the metric expression.
|
||||
* @metric_no_merge: don't attempt to share events for the metric with other
|
||||
* metrics.
|
||||
* @has_constraint: is there a contraint on the group of events? In which case
|
||||
* @has_constraint: is there a constraint on the group of events? In which case
|
||||
* the events won't be grouped.
|
||||
* @metric_events: out argument, null terminated array of evsel's associated
|
||||
* with the metric.
|
||||
|
|
|
@ -843,9 +843,9 @@ split_bpf_config_terms(struct list_head *evt_head_config,
|
|||
struct parse_events_term *term, *temp;
|
||||
|
||||
/*
|
||||
* Currectly, all possible user config term
|
||||
* Currently, all possible user config term
|
||||
* belong to bpf object. parse_events__is_hardcoded_term()
|
||||
* happends to be a good flag.
|
||||
* happens to be a good flag.
|
||||
*
|
||||
* See parse_events_config_bpf() and
|
||||
* config_term_tracepoint().
|
||||
|
@ -895,7 +895,7 @@ int parse_events_load_bpf(struct parse_events_state *parse_state,
|
|||
|
||||
/*
|
||||
* Caller doesn't know anything about obj_head_config,
|
||||
* so combine them together again before returnning.
|
||||
* so combine them together again before returning.
|
||||
*/
|
||||
if (head_config)
|
||||
list_splice_tail(&obj_head_config, head_config);
|
||||
|
@ -1182,10 +1182,10 @@ do { \
|
|||
}
|
||||
|
||||
/*
|
||||
* Check term availbility after basic checking so
|
||||
* Check term availability after basic checking so
|
||||
* PARSE_EVENTS__TERM_TYPE_USER can be found and filtered.
|
||||
*
|
||||
* If check availbility at the entry of this function,
|
||||
* If check availability at the entry of this function,
|
||||
* user will see "'<sysfs term>' is not usable in 'perf stat'"
|
||||
* if an invalid config term is provided for legacy events
|
||||
* (for example, instructions/badterm/...), which is confusing.
|
||||
|
|
|
@ -1069,7 +1069,7 @@ int perf_pmu__format_type(struct list_head *formats, const char *name)
|
|||
|
||||
/*
|
||||
* Sets value based on the format definition (format parameter)
|
||||
* and unformated value (value parameter).
|
||||
* and unformatted value (value parameter).
|
||||
*/
|
||||
static void pmu_format_value(unsigned long *format, __u64 value, __u64 *v,
|
||||
bool zero)
|
||||
|
@ -1408,7 +1408,7 @@ int perf_pmu__check_alias(struct perf_pmu *pmu, struct list_head *head_terms,
|
|||
}
|
||||
|
||||
/*
|
||||
* if no unit or scale foundin aliases, then
|
||||
* if no unit or scale found in aliases, then
|
||||
* set defaults as for evsel
|
||||
* unit cannot left to NULL
|
||||
*/
|
||||
|
|
|
@ -3228,7 +3228,7 @@ errout:
|
|||
return err;
|
||||
}
|
||||
|
||||
/* Concatinate two arrays */
|
||||
/* Concatenate two arrays */
|
||||
static void *memcat(void *a, size_t sz_a, void *b, size_t sz_b)
|
||||
{
|
||||
void *ret;
|
||||
|
@ -3258,7 +3258,7 @@ concat_probe_trace_events(struct probe_trace_event **tevs, int *ntevs,
|
|||
if (*ntevs + ntevs2 > probe_conf.max_probes)
|
||||
ret = -E2BIG;
|
||||
else {
|
||||
/* Concatinate the array of probe_trace_event */
|
||||
/* Concatenate the array of probe_trace_event */
|
||||
new_tevs = memcat(*tevs, (*ntevs) * sizeof(**tevs),
|
||||
*tevs2, ntevs2 * sizeof(**tevs2));
|
||||
if (!new_tevs)
|
||||
|
|
|
@ -164,7 +164,7 @@ static struct probe_trace_arg_ref *alloc_trace_arg_ref(long offs)
|
|||
/*
|
||||
* Convert a location into trace_arg.
|
||||
* If tvar == NULL, this just checks variable can be converted.
|
||||
* If fentry == true and vr_die is a parameter, do huristic search
|
||||
* If fentry == true and vr_die is a parameter, do heuristic search
|
||||
* for the location fuzzed by function entry mcount.
|
||||
*/
|
||||
static int convert_variable_location(Dwarf_Die *vr_die, Dwarf_Addr addr,
|
||||
|
@ -498,7 +498,7 @@ static int convert_variable_fields(Dwarf_Die *vr_die, const char *varname,
|
|||
" nor array.\n", varname);
|
||||
return -EINVAL;
|
||||
}
|
||||
/* While prcessing unnamed field, we don't care about this */
|
||||
/* While processing unnamed field, we don't care about this */
|
||||
if (field->ref && dwarf_diename(vr_die)) {
|
||||
pr_err("Semantic error: %s must be referred by '.'\n",
|
||||
field->name);
|
||||
|
@ -1832,7 +1832,7 @@ static int line_range_walk_cb(const char *fname, int lineno,
|
|||
(lf->lno_s > lineno || lf->lno_e < lineno))
|
||||
return 0;
|
||||
|
||||
/* Make sure this line can be reversable */
|
||||
/* Make sure this line can be reversible */
|
||||
if (cu_find_lineinfo(&lf->cu_die, addr, &__fname, &__lineno) > 0
|
||||
&& (lineno != __lineno || strcmp(fname, __fname)))
|
||||
return 0;
|
||||
|
|
|
@ -45,7 +45,7 @@
|
|||
* the data portion is mmap()'ed.
|
||||
*
|
||||
* To sort the queues in chronological order, all queue access is controlled
|
||||
* by the auxtrace_heap. This is basicly a stack, each stack element has two
|
||||
* by the auxtrace_heap. This is basically a stack, each stack element has two
|
||||
* entries, the queue number and a time stamp. However the stack is sorted by
|
||||
* the time stamps. The highest time stamp is at the bottom the lowest
|
||||
* (nearest) time stamp is at the top. That sort order is maintained at all
|
||||
|
@ -65,11 +65,11 @@
|
|||
* stamp of the last processed entry of the auxtrace_buffer replaces the
|
||||
* current auxtrace_heap top.
|
||||
*
|
||||
* 3. Auxtrace_queues might run of out data and are feeded by the
|
||||
* 3. Auxtrace_queues might run of out data and are fed by the
|
||||
* PERF_RECORD_AUXTRACE handling, see s390_cpumsf_process_auxtrace_event().
|
||||
*
|
||||
* Event Generation
|
||||
* Each sampling-data entry in the auxilary trace data generates a perf sample.
|
||||
* Each sampling-data entry in the auxiliary trace data generates a perf sample.
|
||||
* This sample is filled
|
||||
* with data from the auxtrace such as PID/TID, instruction address, CPU state,
|
||||
* etc. This sample is processed with perf_session__deliver_synth_event() to
|
||||
|
@ -575,7 +575,7 @@ static unsigned long long get_trailer_time(const unsigned char *buf)
|
|||
* pointer to the queue, the second parameter is the time stamp. This
|
||||
* is the time stamp:
|
||||
* - of the event that triggered this processing.
|
||||
* - or the time stamp when the last proccesing of this queue stopped.
|
||||
* - or the time stamp when the last processing of this queue stopped.
|
||||
* In this case it stopped at a 4KB page boundary and record the
|
||||
* position on where to continue processing on the next invocation
|
||||
* (see buffer->use_data and buffer->use_size).
|
||||
|
@ -640,7 +640,7 @@ static int s390_cpumsf_samples(struct s390_cpumsf_queue *sfq, u64 *ts)
|
|||
goto out;
|
||||
}
|
||||
|
||||
pos += dsdes; /* Skip diagnositic entry */
|
||||
pos += dsdes; /* Skip diagnostic entry */
|
||||
|
||||
/* Check for trailer entry */
|
||||
if (!s390_cpumsf_reached_trailer(bsdes + dsdes, pos)) {
|
||||
|
|
|
@ -1531,7 +1531,7 @@ static void set_table_handlers(struct tables *tables)
|
|||
* Attempt to use the call path root from the call return
|
||||
* processor, if the call return processor is in use. Otherwise,
|
||||
* we allocate a new call path root. This prevents exporting
|
||||
* duplicate call path ids when both are in use simultaniously.
|
||||
* duplicate call path ids when both are in use simultaneously.
|
||||
*/
|
||||
if (tables->dbe.crp)
|
||||
tables->dbe.cpr = tables->dbe.crp->cpr;
|
||||
|
|
|
@ -1069,7 +1069,7 @@ static void callchain__lbr_callstack_printf(struct perf_sample *sample)
|
|||
* in "to" register.
|
||||
* For example, there is a call stack
|
||||
* "A"->"B"->"C"->"D".
|
||||
* The LBR registers will recorde like
|
||||
* The LBR registers will be recorded like
|
||||
* "C"->"D", "B"->"C", "A"->"B".
|
||||
* So only the first "to" register and all "from"
|
||||
* registers are needed to construct the whole stack.
|
||||
|
@ -1584,7 +1584,7 @@ static s64 perf_session__process_user_event(struct perf_session *session,
|
|||
return tool->event_update(tool, event, &session->evlist);
|
||||
case PERF_RECORD_HEADER_EVENT_TYPE:
|
||||
/*
|
||||
* Depreceated, but we need to handle it for sake
|
||||
* Deprecated, but we need to handle it for sake
|
||||
* of old data files create in pipe mode.
|
||||
*/
|
||||
return 0;
|
||||
|
|
|
@ -12,7 +12,7 @@
|
|||
* build complex strings/buffers whose final size isn't easily known.
|
||||
*
|
||||
* It is NOT legal to copy the ->buf pointer away.
|
||||
* `strbuf_detach' is the operation that detachs a buffer from its shell
|
||||
* `strbuf_detach' is the operation that detaches a buffer from its shell
|
||||
* while keeping the shell valid wrt its invariants.
|
||||
*
|
||||
* 2. the ->buf member is a byte array that has at least ->len + 1 bytes
|
||||
|
|
|
@ -8,8 +8,8 @@
|
|||
|
||||
/* A node of string filter */
|
||||
struct strfilter_node {
|
||||
struct strfilter_node *l; /* Tree left branche (for &,|) */
|
||||
struct strfilter_node *r; /* Tree right branche (for !,&,|) */
|
||||
struct strfilter_node *l; /* Tree left branch (for &,|) */
|
||||
struct strfilter_node *r; /* Tree right branch (for !,&,|) */
|
||||
const char *p; /* Operator or rule */
|
||||
};
|
||||
|
||||
|
|
|
@ -1058,7 +1058,7 @@ static int dso__process_kernel_symbol(struct dso *dso, struct map *map,
|
|||
curr_dso->symtab_type = dso->symtab_type;
|
||||
maps__insert(kmaps, curr_map);
|
||||
/*
|
||||
* Add it before we drop the referece to curr_map, i.e. while
|
||||
* Add it before we drop the reference to curr_map, i.e. while
|
||||
* we still are sure to have a reference to this DSO via
|
||||
* *curr_map->dso.
|
||||
*/
|
||||
|
|
|
@ -1210,7 +1210,7 @@ static size_t mask_size(struct perf_cpu_map *map, int *max)
|
|||
*max = 0;
|
||||
|
||||
for (i = 0; i < map->nr; i++) {
|
||||
/* bit possition of the cpu is + 1 */
|
||||
/* bit position of the cpu is + 1 */
|
||||
int bit = map->map[i] + 1;
|
||||
|
||||
if (bit > *max)
|
||||
|
@ -1236,7 +1236,7 @@ void *cpu_map_data__alloc(struct perf_cpu_map *map, size_t *size, u16 *type, int
|
|||
* mask = size of 'struct perf_record_record_cpu_map' +
|
||||
* maximum cpu bit converted to size of longs
|
||||
*
|
||||
* and finaly + the size of 'struct perf_record_cpu_map_data'.
|
||||
* and finally + the size of 'struct perf_record_cpu_map_data'.
|
||||
*/
|
||||
size_cpus = cpus_size(map);
|
||||
size_mask = mask_size(map, max);
|
||||
|
|
|
@ -82,7 +82,7 @@ UNW_OBJ(dwarf_find_debug_frame) (int found, unw_dyn_info_t *di_debug,
|
|||
#define DW_EH_PE_funcrel 0x40 /* start-of-procedure-relative */
|
||||
#define DW_EH_PE_aligned 0x50 /* aligned pointer */
|
||||
|
||||
/* Flags intentionaly not handled, since they're not needed:
|
||||
/* Flags intentionally not handled, since they're not needed:
|
||||
* #define DW_EH_PE_indirect 0x80
|
||||
* #define DW_EH_PE_uleb128 0x01
|
||||
* #define DW_EH_PE_udata2 0x02
|
||||
|
|
Loading…
Reference in New Issue