Merge branch 'oprofile/perf' into oprofile/core
Conflicts: arch/arm/oprofile/common.c Signed-off-by: Robert Richter <robert.richter@amd.com>
This commit is contained in:
commit
0361e02342
|
@ -123,6 +123,12 @@ armpmu_get_max_events(void)
|
|||
}
|
||||
EXPORT_SYMBOL_GPL(armpmu_get_max_events);
|
||||
|
||||
int perf_num_counters(void)
|
||||
{
|
||||
return armpmu_get_max_events();
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(perf_num_counters);
|
||||
|
||||
#define HW_OP_UNSUPPORTED 0xFFFF
|
||||
|
||||
#define C(_x) \
|
||||
|
|
|
@ -6,4 +6,8 @@ DRIVER_OBJS = $(addprefix ../../../drivers/oprofile/, \
|
|||
oprofilefs.o oprofile_stats.o \
|
||||
timer_int.o )
|
||||
|
||||
ifeq ($(CONFIG_HW_PERF_EVENTS),y)
|
||||
DRIVER_OBJS += $(addprefix ../../../drivers/oprofile/, oprofile_perf.o)
|
||||
endif
|
||||
|
||||
oprofile-y := $(DRIVER_OBJS) common.o
|
||||
|
|
|
@ -25,138 +25,10 @@
|
|||
#include <asm/ptrace.h>
|
||||
|
||||
#ifdef CONFIG_HW_PERF_EVENTS
|
||||
/*
|
||||
* Per performance monitor configuration as set via oprofilefs.
|
||||
*/
|
||||
struct op_counter_config {
|
||||
unsigned long count;
|
||||
unsigned long enabled;
|
||||
unsigned long event;
|
||||
unsigned long unit_mask;
|
||||
unsigned long kernel;
|
||||
unsigned long user;
|
||||
struct perf_event_attr attr;
|
||||
};
|
||||
|
||||
static int op_arm_enabled;
|
||||
static DEFINE_MUTEX(op_arm_mutex);
|
||||
|
||||
static struct op_counter_config *counter_config;
|
||||
static struct perf_event **perf_events[nr_cpumask_bits];
|
||||
static int perf_num_counters;
|
||||
|
||||
/*
|
||||
* Overflow callback for oprofile.
|
||||
*/
|
||||
static void op_overflow_handler(struct perf_event *event, int unused,
|
||||
struct perf_sample_data *data, struct pt_regs *regs)
|
||||
char *op_name_from_perf_id(void)
|
||||
{
|
||||
int id;
|
||||
u32 cpu = smp_processor_id();
|
||||
enum arm_perf_pmu_ids id = armpmu_get_pmu_id();
|
||||
|
||||
for (id = 0; id < perf_num_counters; ++id)
|
||||
if (perf_events[cpu][id] == event)
|
||||
break;
|
||||
|
||||
if (id != perf_num_counters)
|
||||
oprofile_add_sample(regs, id);
|
||||
else
|
||||
pr_warning("oprofile: ignoring spurious overflow "
|
||||
"on cpu %u\n", cpu);
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by op_arm_setup to create perf attributes to mirror the oprofile
|
||||
* settings in counter_config. Attributes are created as `pinned' events and
|
||||
* so are permanently scheduled on the PMU.
|
||||
*/
|
||||
static void op_perf_setup(void)
|
||||
{
|
||||
int i;
|
||||
u32 size = sizeof(struct perf_event_attr);
|
||||
struct perf_event_attr *attr;
|
||||
|
||||
for (i = 0; i < perf_num_counters; ++i) {
|
||||
attr = &counter_config[i].attr;
|
||||
memset(attr, 0, size);
|
||||
attr->type = PERF_TYPE_RAW;
|
||||
attr->size = size;
|
||||
attr->config = counter_config[i].event;
|
||||
attr->sample_period = counter_config[i].count;
|
||||
attr->pinned = 1;
|
||||
}
|
||||
}
|
||||
|
||||
static int op_create_counter(int cpu, int event)
|
||||
{
|
||||
struct perf_event *pevent;
|
||||
|
||||
if (!counter_config[event].enabled || perf_events[cpu][event])
|
||||
return 0;
|
||||
|
||||
pevent = perf_event_create_kernel_counter(&counter_config[event].attr,
|
||||
cpu, -1,
|
||||
op_overflow_handler);
|
||||
|
||||
if (IS_ERR(pevent))
|
||||
return PTR_ERR(pevent);
|
||||
|
||||
if (pevent->state != PERF_EVENT_STATE_ACTIVE) {
|
||||
perf_event_release_kernel(pevent);
|
||||
pr_warning("oprofile: failed to enable event %d "
|
||||
"on CPU %d\n", event, cpu);
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
perf_events[cpu][event] = pevent;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void op_destroy_counter(int cpu, int event)
|
||||
{
|
||||
struct perf_event *pevent = perf_events[cpu][event];
|
||||
|
||||
if (pevent) {
|
||||
perf_event_release_kernel(pevent);
|
||||
perf_events[cpu][event] = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by op_arm_start to create active perf events based on the
|
||||
* perviously configured attributes.
|
||||
*/
|
||||
static int op_perf_start(void)
|
||||
{
|
||||
int cpu, event, ret = 0;
|
||||
|
||||
for_each_online_cpu(cpu) {
|
||||
for (event = 0; event < perf_num_counters; ++event) {
|
||||
ret = op_create_counter(cpu, event);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by op_arm_stop at the end of a profiling run.
|
||||
*/
|
||||
static void op_perf_stop(void)
|
||||
{
|
||||
int cpu, event;
|
||||
|
||||
for_each_online_cpu(cpu)
|
||||
for (event = 0; event < perf_num_counters; ++event)
|
||||
op_destroy_counter(cpu, event);
|
||||
}
|
||||
|
||||
|
||||
static char *op_name_from_perf_id(enum arm_perf_pmu_ids id)
|
||||
{
|
||||
switch (id) {
|
||||
case ARM_PERF_PMU_ID_XSCALE1:
|
||||
return "arm/xscale1";
|
||||
|
@ -175,115 +47,6 @@ static char *op_name_from_perf_id(enum arm_perf_pmu_ids id)
|
|||
}
|
||||
}
|
||||
|
||||
static int op_arm_create_files(struct super_block *sb, struct dentry *root)
|
||||
{
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < perf_num_counters; i++) {
|
||||
struct dentry *dir;
|
||||
char buf[4];
|
||||
|
||||
snprintf(buf, sizeof buf, "%d", i);
|
||||
dir = oprofilefs_mkdir(sb, root, buf);
|
||||
oprofilefs_create_ulong(sb, dir, "enabled", &counter_config[i].enabled);
|
||||
oprofilefs_create_ulong(sb, dir, "event", &counter_config[i].event);
|
||||
oprofilefs_create_ulong(sb, dir, "count", &counter_config[i].count);
|
||||
oprofilefs_create_ulong(sb, dir, "unit_mask", &counter_config[i].unit_mask);
|
||||
oprofilefs_create_ulong(sb, dir, "kernel", &counter_config[i].kernel);
|
||||
oprofilefs_create_ulong(sb, dir, "user", &counter_config[i].user);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int op_arm_setup(void)
|
||||
{
|
||||
spin_lock(&oprofilefs_lock);
|
||||
op_perf_setup();
|
||||
spin_unlock(&oprofilefs_lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int op_arm_start(void)
|
||||
{
|
||||
int ret = -EBUSY;
|
||||
|
||||
mutex_lock(&op_arm_mutex);
|
||||
if (!op_arm_enabled) {
|
||||
ret = 0;
|
||||
op_perf_start();
|
||||
op_arm_enabled = 1;
|
||||
}
|
||||
mutex_unlock(&op_arm_mutex);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void op_arm_stop(void)
|
||||
{
|
||||
mutex_lock(&op_arm_mutex);
|
||||
if (op_arm_enabled)
|
||||
op_perf_stop();
|
||||
op_arm_enabled = 0;
|
||||
mutex_unlock(&op_arm_mutex);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PM
|
||||
static int op_arm_suspend(struct platform_device *dev, pm_message_t state)
|
||||
{
|
||||
mutex_lock(&op_arm_mutex);
|
||||
if (op_arm_enabled)
|
||||
op_perf_stop();
|
||||
mutex_unlock(&op_arm_mutex);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int op_arm_resume(struct platform_device *dev)
|
||||
{
|
||||
mutex_lock(&op_arm_mutex);
|
||||
if (op_arm_enabled && op_perf_start())
|
||||
op_arm_enabled = 0;
|
||||
mutex_unlock(&op_arm_mutex);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct platform_driver oprofile_driver = {
|
||||
.driver = {
|
||||
.name = "arm-oprofile",
|
||||
},
|
||||
.resume = op_arm_resume,
|
||||
.suspend = op_arm_suspend,
|
||||
};
|
||||
|
||||
static struct platform_device *oprofile_pdev;
|
||||
|
||||
static int __init init_driverfs(void)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = platform_driver_register(&oprofile_driver);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
oprofile_pdev = platform_device_register_simple(
|
||||
oprofile_driver.driver.name, 0, NULL, 0);
|
||||
if (IS_ERR(oprofile_pdev)) {
|
||||
ret = PTR_ERR(oprofile_pdev);
|
||||
platform_driver_unregister(&oprofile_driver);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void __exit exit_driverfs(void)
|
||||
{
|
||||
platform_device_unregister(oprofile_pdev);
|
||||
platform_driver_unregister(&oprofile_driver);
|
||||
}
|
||||
#else
|
||||
static int __init init_driverfs(void) { return 0; }
|
||||
#define exit_driverfs() do { } while (0)
|
||||
#endif /* CONFIG_PM */
|
||||
|
||||
static int report_trace(struct stackframe *frame, void *d)
|
||||
{
|
||||
unsigned int *depth = d;
|
||||
|
@ -346,79 +109,17 @@ static void arm_backtrace(struct pt_regs * const regs, unsigned int depth)
|
|||
tail = user_backtrace(tail);
|
||||
}
|
||||
|
||||
void oprofile_arch_exit(void)
|
||||
{
|
||||
int cpu, id;
|
||||
struct perf_event *event;
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
for (id = 0; id < perf_num_counters; ++id) {
|
||||
event = perf_events[cpu][id];
|
||||
if (event)
|
||||
perf_event_release_kernel(event);
|
||||
}
|
||||
|
||||
kfree(perf_events[cpu]);
|
||||
}
|
||||
|
||||
kfree(counter_config);
|
||||
exit_driverfs();
|
||||
}
|
||||
|
||||
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
||||
{
|
||||
int cpu, ret = 0;
|
||||
|
||||
ret = init_driverfs();
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
memset(&perf_events, 0, sizeof(perf_events));
|
||||
|
||||
perf_num_counters = armpmu_get_max_events();
|
||||
|
||||
counter_config = kcalloc(perf_num_counters,
|
||||
sizeof(struct op_counter_config), GFP_KERNEL);
|
||||
|
||||
if (!counter_config) {
|
||||
pr_info("oprofile: failed to allocate %d "
|
||||
"counters\n", perf_num_counters);
|
||||
ret = -ENOMEM;
|
||||
perf_num_counters = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
perf_events[cpu] = kcalloc(perf_num_counters,
|
||||
sizeof(struct perf_event *), GFP_KERNEL);
|
||||
if (!perf_events[cpu]) {
|
||||
pr_info("oprofile: failed to allocate %d perf events "
|
||||
"for cpu %d\n", perf_num_counters, cpu);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
ops->backtrace = arm_backtrace;
|
||||
ops->create_files = op_arm_create_files;
|
||||
ops->setup = op_arm_setup;
|
||||
ops->start = op_arm_start;
|
||||
ops->stop = op_arm_stop;
|
||||
ops->shutdown = op_arm_stop;
|
||||
ops->cpu_type = op_name_from_perf_id(armpmu_get_pmu_id());
|
||||
|
||||
if (!ops->cpu_type)
|
||||
ret = -ENODEV;
|
||||
else
|
||||
pr_info("oprofile: using %s\n", ops->cpu_type);
|
||||
|
||||
out:
|
||||
if (ret)
|
||||
oprofile_arch_exit();
|
||||
|
||||
return ret;
|
||||
return oprofile_perf_init(ops);
|
||||
}
|
||||
|
||||
void __exit oprofile_arch_exit(void)
|
||||
{
|
||||
oprofile_perf_exit();
|
||||
}
|
||||
#else
|
||||
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
||||
{
|
||||
|
|
|
@ -249,6 +249,11 @@ config ARCH_SHMOBILE
|
|||
select PM
|
||||
select PM_RUNTIME
|
||||
|
||||
config CPU_HAS_PMU
|
||||
depends on CPU_SH4 || CPU_SH4A
|
||||
default y
|
||||
bool
|
||||
|
||||
if SUPERH32
|
||||
|
||||
choice
|
||||
|
@ -738,6 +743,14 @@ config GUSA_RB
|
|||
LLSC, this should be more efficient than the other alternative of
|
||||
disabling interrupts around the atomic sequence.
|
||||
|
||||
config HW_PERF_EVENTS
|
||||
bool "Enable hardware performance counter support for perf events"
|
||||
depends on PERF_EVENTS && CPU_HAS_PMU
|
||||
default y
|
||||
help
|
||||
Enable hardware performance counter support for perf events. If
|
||||
disabled, perf events will use software events only.
|
||||
|
||||
source "drivers/sh/Kconfig"
|
||||
|
||||
endmenu
|
||||
|
|
|
@ -59,6 +59,24 @@ static inline int sh_pmu_initialized(void)
|
|||
return !!sh_pmu;
|
||||
}
|
||||
|
||||
const char *perf_pmu_name(void)
|
||||
{
|
||||
if (!sh_pmu)
|
||||
return NULL;
|
||||
|
||||
return sh_pmu->name;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(perf_pmu_name);
|
||||
|
||||
int perf_num_counters(void)
|
||||
{
|
||||
if (!sh_pmu)
|
||||
return 0;
|
||||
|
||||
return sh_pmu->num_events;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(perf_num_counters);
|
||||
|
||||
/*
|
||||
* Release the PMU if this is the last perf_event.
|
||||
*/
|
||||
|
|
|
@ -6,4 +6,8 @@ DRIVER_OBJS = $(addprefix ../../../drivers/oprofile/, \
|
|||
oprofilefs.o oprofile_stats.o \
|
||||
timer_int.o )
|
||||
|
||||
ifeq ($(CONFIG_HW_PERF_EVENTS),y)
|
||||
DRIVER_OBJS += $(addprefix ../../../drivers/oprofile/, oprofile_perf.o)
|
||||
endif
|
||||
|
||||
oprofile-y := $(DRIVER_OBJS) common.o backtrace.o
|
||||
|
|
|
@ -17,114 +17,45 @@
|
|||
#include <linux/init.h>
|
||||
#include <linux/errno.h>
|
||||
#include <linux/smp.h>
|
||||
#include <linux/perf_event.h>
|
||||
#include <asm/processor.h>
|
||||
#include "op_impl.h"
|
||||
|
||||
static struct op_sh_model *model;
|
||||
|
||||
static struct op_counter_config ctr[20];
|
||||
|
||||
#ifdef CONFIG_HW_PERF_EVENTS
|
||||
extern void sh_backtrace(struct pt_regs * const regs, unsigned int depth);
|
||||
|
||||
static int op_sh_setup(void)
|
||||
char *op_name_from_perf_id(void)
|
||||
{
|
||||
/* Pre-compute the values to stuff in the hardware registers. */
|
||||
model->reg_setup(ctr);
|
||||
const char *pmu;
|
||||
char buf[20];
|
||||
int size;
|
||||
|
||||
/* Configure the registers on all cpus. */
|
||||
on_each_cpu(model->cpu_setup, NULL, 1);
|
||||
pmu = perf_pmu_name();
|
||||
if (!pmu)
|
||||
return NULL;
|
||||
|
||||
return 0;
|
||||
}
|
||||
size = snprintf(buf, sizeof(buf), "sh/%s", pmu);
|
||||
if (size > -1 && size < sizeof(buf))
|
||||
return buf;
|
||||
|
||||
static int op_sh_create_files(struct super_block *sb, struct dentry *root)
|
||||
{
|
||||
int i, ret = 0;
|
||||
|
||||
for (i = 0; i < model->num_counters; i++) {
|
||||
struct dentry *dir;
|
||||
char buf[4];
|
||||
|
||||
snprintf(buf, sizeof(buf), "%d", i);
|
||||
dir = oprofilefs_mkdir(sb, root, buf);
|
||||
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "enabled", &ctr[i].enabled);
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "event", &ctr[i].event);
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "kernel", &ctr[i].kernel);
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "user", &ctr[i].user);
|
||||
|
||||
if (model->create_files)
|
||||
ret |= model->create_files(sb, dir);
|
||||
else
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "count", &ctr[i].count);
|
||||
|
||||
/* Dummy entries */
|
||||
ret |= oprofilefs_create_ulong(sb, dir, "unit_mask", &ctr[i].unit_mask);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int op_sh_start(void)
|
||||
{
|
||||
/* Enable performance monitoring for all counters. */
|
||||
on_each_cpu(model->cpu_start, NULL, 1);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void op_sh_stop(void)
|
||||
{
|
||||
/* Disable performance monitoring for all counters. */
|
||||
on_each_cpu(model->cpu_stop, NULL, 1);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
||||
{
|
||||
struct op_sh_model *lmodel = NULL;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Always assign the backtrace op. If the counter initialization
|
||||
* fails, we fall back to the timer which will still make use of
|
||||
* this.
|
||||
*/
|
||||
ops->backtrace = sh_backtrace;
|
||||
|
||||
/*
|
||||
* XXX
|
||||
*
|
||||
* All of the SH7750/SH-4A counters have been converted to perf,
|
||||
* this infrastructure hook is left for other users until they've
|
||||
* had a chance to convert over, at which point all of this
|
||||
* will be deleted.
|
||||
*/
|
||||
|
||||
if (!lmodel)
|
||||
return -ENODEV;
|
||||
if (!(current_cpu_data.flags & CPU_HAS_PERF_COUNTER))
|
||||
return -ENODEV;
|
||||
|
||||
ret = lmodel->init();
|
||||
if (unlikely(ret != 0))
|
||||
return ret;
|
||||
|
||||
model = lmodel;
|
||||
|
||||
ops->setup = op_sh_setup;
|
||||
ops->create_files = op_sh_create_files;
|
||||
ops->start = op_sh_start;
|
||||
ops->stop = op_sh_stop;
|
||||
ops->cpu_type = lmodel->cpu_type;
|
||||
|
||||
printk(KERN_INFO "oprofile: using %s performance monitoring.\n",
|
||||
lmodel->cpu_type);
|
||||
|
||||
return 0;
|
||||
return oprofile_perf_init(ops);
|
||||
}
|
||||
|
||||
void oprofile_arch_exit(void)
|
||||
void __exit oprofile_arch_exit(void)
|
||||
{
|
||||
if (model && model->exit)
|
||||
model->exit();
|
||||
oprofile_perf_exit();
|
||||
}
|
||||
#else
|
||||
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
||||
{
|
||||
pr_info("oprofile: hardware counters not available\n");
|
||||
return -ENODEV;
|
||||
}
|
||||
void __exit oprofile_arch_exit(void) {}
|
||||
#endif /* CONFIG_HW_PERF_EVENTS */
|
||||
|
|
|
@ -1,33 +0,0 @@
|
|||
#ifndef __OP_IMPL_H
|
||||
#define __OP_IMPL_H
|
||||
|
||||
/* Per-counter configuration as set via oprofilefs. */
|
||||
struct op_counter_config {
|
||||
unsigned long enabled;
|
||||
unsigned long event;
|
||||
|
||||
unsigned long count;
|
||||
|
||||
/* Dummy values for userspace tool compliance */
|
||||
unsigned long kernel;
|
||||
unsigned long user;
|
||||
unsigned long unit_mask;
|
||||
};
|
||||
|
||||
/* Per-architecture configury and hooks. */
|
||||
struct op_sh_model {
|
||||
void (*reg_setup)(struct op_counter_config *);
|
||||
int (*create_files)(struct super_block *sb, struct dentry *dir);
|
||||
void (*cpu_setup)(void *dummy);
|
||||
int (*init)(void);
|
||||
void (*exit)(void);
|
||||
void (*cpu_start)(void *args);
|
||||
void (*cpu_stop)(void *args);
|
||||
char *cpu_type;
|
||||
unsigned char num_counters;
|
||||
};
|
||||
|
||||
/* arch/sh/oprofile/common.c */
|
||||
extern void sh_backtrace(struct pt_regs * const regs, unsigned int depth);
|
||||
|
||||
#endif /* __OP_IMPL_H */
|
|
@ -0,0 +1,323 @@
|
|||
/*
|
||||
* Copyright 2010 ARM Ltd.
|
||||
*
|
||||
* Perf-events backend for OProfile.
|
||||
*/
|
||||
#include <linux/perf_event.h>
|
||||
#include <linux/oprofile.h>
|
||||
#include <linux/slab.h>
|
||||
|
||||
/*
|
||||
* Per performance monitor configuration as set via oprofilefs.
|
||||
*/
|
||||
struct op_counter_config {
|
||||
unsigned long count;
|
||||
unsigned long enabled;
|
||||
unsigned long event;
|
||||
unsigned long unit_mask;
|
||||
unsigned long kernel;
|
||||
unsigned long user;
|
||||
struct perf_event_attr attr;
|
||||
};
|
||||
|
||||
static int oprofile_perf_enabled;
|
||||
static DEFINE_MUTEX(oprofile_perf_mutex);
|
||||
|
||||
static struct op_counter_config *counter_config;
|
||||
static struct perf_event **perf_events[nr_cpumask_bits];
|
||||
static int num_counters;
|
||||
|
||||
/*
|
||||
* Overflow callback for oprofile.
|
||||
*/
|
||||
static void op_overflow_handler(struct perf_event *event, int unused,
|
||||
struct perf_sample_data *data, struct pt_regs *regs)
|
||||
{
|
||||
int id;
|
||||
u32 cpu = smp_processor_id();
|
||||
|
||||
for (id = 0; id < num_counters; ++id)
|
||||
if (perf_events[cpu][id] == event)
|
||||
break;
|
||||
|
||||
if (id != num_counters)
|
||||
oprofile_add_sample(regs, id);
|
||||
else
|
||||
pr_warning("oprofile: ignoring spurious overflow "
|
||||
"on cpu %u\n", cpu);
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by oprofile_perf_setup to create perf attributes to mirror the oprofile
|
||||
* settings in counter_config. Attributes are created as `pinned' events and
|
||||
* so are permanently scheduled on the PMU.
|
||||
*/
|
||||
static void op_perf_setup(void)
|
||||
{
|
||||
int i;
|
||||
u32 size = sizeof(struct perf_event_attr);
|
||||
struct perf_event_attr *attr;
|
||||
|
||||
for (i = 0; i < num_counters; ++i) {
|
||||
attr = &counter_config[i].attr;
|
||||
memset(attr, 0, size);
|
||||
attr->type = PERF_TYPE_RAW;
|
||||
attr->size = size;
|
||||
attr->config = counter_config[i].event;
|
||||
attr->sample_period = counter_config[i].count;
|
||||
attr->pinned = 1;
|
||||
}
|
||||
}
|
||||
|
||||
static int op_create_counter(int cpu, int event)
|
||||
{
|
||||
struct perf_event *pevent;
|
||||
|
||||
if (!counter_config[event].enabled || perf_events[cpu][event])
|
||||
return 0;
|
||||
|
||||
pevent = perf_event_create_kernel_counter(&counter_config[event].attr,
|
||||
cpu, -1,
|
||||
op_overflow_handler);
|
||||
|
||||
if (IS_ERR(pevent))
|
||||
return PTR_ERR(pevent);
|
||||
|
||||
if (pevent->state != PERF_EVENT_STATE_ACTIVE) {
|
||||
perf_event_release_kernel(pevent);
|
||||
pr_warning("oprofile: failed to enable event %d "
|
||||
"on CPU %d\n", event, cpu);
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
perf_events[cpu][event] = pevent;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void op_destroy_counter(int cpu, int event)
|
||||
{
|
||||
struct perf_event *pevent = perf_events[cpu][event];
|
||||
|
||||
if (pevent) {
|
||||
perf_event_release_kernel(pevent);
|
||||
perf_events[cpu][event] = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by oprofile_perf_start to create active perf events based on the
|
||||
* perviously configured attributes.
|
||||
*/
|
||||
static int op_perf_start(void)
|
||||
{
|
||||
int cpu, event, ret = 0;
|
||||
|
||||
for_each_online_cpu(cpu) {
|
||||
for (event = 0; event < num_counters; ++event) {
|
||||
ret = op_create_counter(cpu, event);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Called by oprofile_perf_stop at the end of a profiling run.
|
||||
*/
|
||||
static void op_perf_stop(void)
|
||||
{
|
||||
int cpu, event;
|
||||
|
||||
for_each_online_cpu(cpu)
|
||||
for (event = 0; event < num_counters; ++event)
|
||||
op_destroy_counter(cpu, event);
|
||||
}
|
||||
|
||||
static int oprofile_perf_create_files(struct super_block *sb, struct dentry *root)
|
||||
{
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < num_counters; i++) {
|
||||
struct dentry *dir;
|
||||
char buf[4];
|
||||
|
||||
snprintf(buf, sizeof buf, "%d", i);
|
||||
dir = oprofilefs_mkdir(sb, root, buf);
|
||||
oprofilefs_create_ulong(sb, dir, "enabled", &counter_config[i].enabled);
|
||||
oprofilefs_create_ulong(sb, dir, "event", &counter_config[i].event);
|
||||
oprofilefs_create_ulong(sb, dir, "count", &counter_config[i].count);
|
||||
oprofilefs_create_ulong(sb, dir, "unit_mask", &counter_config[i].unit_mask);
|
||||
oprofilefs_create_ulong(sb, dir, "kernel", &counter_config[i].kernel);
|
||||
oprofilefs_create_ulong(sb, dir, "user", &counter_config[i].user);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int oprofile_perf_setup(void)
|
||||
{
|
||||
spin_lock(&oprofilefs_lock);
|
||||
op_perf_setup();
|
||||
spin_unlock(&oprofilefs_lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int oprofile_perf_start(void)
|
||||
{
|
||||
int ret = -EBUSY;
|
||||
|
||||
mutex_lock(&oprofile_perf_mutex);
|
||||
if (!oprofile_perf_enabled) {
|
||||
ret = 0;
|
||||
op_perf_start();
|
||||
oprofile_perf_enabled = 1;
|
||||
}
|
||||
mutex_unlock(&oprofile_perf_mutex);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void oprofile_perf_stop(void)
|
||||
{
|
||||
mutex_lock(&oprofile_perf_mutex);
|
||||
if (oprofile_perf_enabled)
|
||||
op_perf_stop();
|
||||
oprofile_perf_enabled = 0;
|
||||
mutex_unlock(&oprofile_perf_mutex);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PM
|
||||
static int oprofile_perf_suspend(struct platform_device *dev, pm_message_t state)
|
||||
{
|
||||
mutex_lock(&oprofile_perf_mutex);
|
||||
if (oprofile_perf_enabled)
|
||||
op_perf_stop();
|
||||
mutex_unlock(&oprofile_perf_mutex);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int oprofile_perf_resume(struct platform_device *dev)
|
||||
{
|
||||
mutex_lock(&oprofile_perf_mutex);
|
||||
if (oprofile_perf_enabled && op_perf_start())
|
||||
oprofile_perf_enabled = 0;
|
||||
mutex_unlock(&oprofile_perf_mutex);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct platform_driver oprofile_driver = {
|
||||
.driver = {
|
||||
.name = "oprofile-perf",
|
||||
},
|
||||
.resume = oprofile_perf_resume,
|
||||
.suspend = oprofile_perf_suspend,
|
||||
};
|
||||
|
||||
static struct platform_device *oprofile_pdev;
|
||||
|
||||
static int __init init_driverfs(void)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = platform_driver_register(&oprofile_driver);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
oprofile_pdev = platform_device_register_simple(
|
||||
oprofile_driver.driver.name, 0, NULL, 0);
|
||||
if (IS_ERR(oprofile_pdev)) {
|
||||
ret = PTR_ERR(oprofile_pdev);
|
||||
platform_driver_unregister(&oprofile_driver);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void __exit exit_driverfs(void)
|
||||
{
|
||||
platform_device_unregister(oprofile_pdev);
|
||||
platform_driver_unregister(&oprofile_driver);
|
||||
}
|
||||
#else
|
||||
static int __init init_driverfs(void) { return 0; }
|
||||
#define exit_driverfs() do { } while (0)
|
||||
#endif /* CONFIG_PM */
|
||||
|
||||
void oprofile_perf_exit(void)
|
||||
{
|
||||
int cpu, id;
|
||||
struct perf_event *event;
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
for (id = 0; id < num_counters; ++id) {
|
||||
event = perf_events[cpu][id];
|
||||
if (event)
|
||||
perf_event_release_kernel(event);
|
||||
}
|
||||
|
||||
kfree(perf_events[cpu]);
|
||||
}
|
||||
|
||||
kfree(counter_config);
|
||||
exit_driverfs();
|
||||
}
|
||||
|
||||
int __init oprofile_perf_init(struct oprofile_operations *ops)
|
||||
{
|
||||
int cpu, ret = 0;
|
||||
|
||||
ret = init_driverfs();
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
memset(&perf_events, 0, sizeof(perf_events));
|
||||
|
||||
num_counters = perf_num_counters();
|
||||
if (num_counters <= 0) {
|
||||
pr_info("oprofile: no performance counters\n");
|
||||
ret = -ENODEV;
|
||||
goto out;
|
||||
}
|
||||
|
||||
counter_config = kcalloc(num_counters,
|
||||
sizeof(struct op_counter_config), GFP_KERNEL);
|
||||
|
||||
if (!counter_config) {
|
||||
pr_info("oprofile: failed to allocate %d "
|
||||
"counters\n", num_counters);
|
||||
ret = -ENOMEM;
|
||||
num_counters = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
perf_events[cpu] = kcalloc(num_counters,
|
||||
sizeof(struct perf_event *), GFP_KERNEL);
|
||||
if (!perf_events[cpu]) {
|
||||
pr_info("oprofile: failed to allocate %d perf events "
|
||||
"for cpu %d\n", num_counters, cpu);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
ops->create_files = oprofile_perf_create_files;
|
||||
ops->setup = oprofile_perf_setup;
|
||||
ops->start = oprofile_perf_start;
|
||||
ops->stop = oprofile_perf_stop;
|
||||
ops->shutdown = oprofile_perf_stop;
|
||||
ops->cpu_type = op_name_from_perf_id();
|
||||
|
||||
if (!ops->cpu_type)
|
||||
ret = -ENODEV;
|
||||
else
|
||||
pr_info("oprofile: using %s\n", ops->cpu_type);
|
||||
|
||||
out:
|
||||
if (ret)
|
||||
oprofile_perf_exit();
|
||||
|
||||
return ret;
|
||||
}
|
|
@ -15,6 +15,7 @@
|
|||
|
||||
#include <linux/types.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/init.h>
|
||||
#include <asm/atomic.h>
|
||||
|
||||
/* Each escaped entry is prefixed by ESCAPE_CODE
|
||||
|
@ -185,4 +186,10 @@ int oprofile_add_data(struct op_entry *entry, unsigned long val);
|
|||
int oprofile_add_data64(struct op_entry *entry, u64 val);
|
||||
int oprofile_write_commit(struct op_entry *entry);
|
||||
|
||||
#ifdef CONFIG_PERF_EVENTS
|
||||
int __init oprofile_perf_init(struct oprofile_operations *ops);
|
||||
void __exit oprofile_perf_exit(void);
|
||||
char *op_name_from_perf_id(void);
|
||||
#endif /* CONFIG_PERF_EVENTS */
|
||||
|
||||
#endif /* OPROFILE_H */
|
||||
|
|
|
@ -849,6 +849,8 @@ extern int perf_max_events;
|
|||
|
||||
extern const struct pmu *hw_perf_event_init(struct perf_event *event);
|
||||
|
||||
extern int perf_num_counters(void);
|
||||
extern const char *perf_pmu_name(void);
|
||||
extern void perf_event_task_sched_in(struct task_struct *task);
|
||||
extern void perf_event_task_sched_out(struct task_struct *task, struct task_struct *next);
|
||||
extern void perf_event_task_tick(struct task_struct *task);
|
||||
|
|
|
@ -85,6 +85,11 @@ void __weak hw_perf_enable(void) { barrier(); }
|
|||
|
||||
void __weak perf_event_print_debug(void) { }
|
||||
|
||||
extern __weak const char *perf_pmu_name(void)
|
||||
{
|
||||
return "pmu";
|
||||
}
|
||||
|
||||
static DEFINE_PER_CPU(int, perf_disable_count);
|
||||
|
||||
void perf_disable(void)
|
||||
|
|
Loading…
Reference in New Issue