127 lines
3.0 KiB
C
127 lines
3.0 KiB
C
/**
|
|
* @file cpu_buffer.h
|
|
*
|
|
* @remark Copyright 2002 OProfile authors
|
|
* @remark Read the file COPYING
|
|
*
|
|
* @author John Levon <levon@movementarian.org>
|
|
*/
|
|
|
|
#ifndef OPROFILE_CPU_BUFFER_H
|
|
#define OPROFILE_CPU_BUFFER_H
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/cache.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/ring_buffer.h>
|
|
|
|
struct task_struct;
|
|
|
|
int alloc_cpu_buffers(void);
|
|
void free_cpu_buffers(void);
|
|
|
|
void start_cpu_work(void);
|
|
void end_cpu_work(void);
|
|
|
|
/* CPU buffer is composed of such entries (which are
|
|
* also used for context switch notes)
|
|
*/
|
|
struct op_sample {
|
|
unsigned long eip;
|
|
unsigned long event;
|
|
};
|
|
|
|
struct op_entry {
|
|
struct ring_buffer_event *event;
|
|
struct op_sample *sample;
|
|
unsigned long irq_flags;
|
|
};
|
|
|
|
struct oprofile_cpu_buffer {
|
|
volatile unsigned long head_pos;
|
|
volatile unsigned long tail_pos;
|
|
unsigned long buffer_size;
|
|
struct task_struct *last_task;
|
|
int last_is_kernel;
|
|
int tracing;
|
|
unsigned long sample_received;
|
|
unsigned long sample_lost_overflow;
|
|
unsigned long backtrace_aborted;
|
|
unsigned long sample_invalid_eip;
|
|
int cpu;
|
|
struct delayed_work work;
|
|
};
|
|
|
|
extern struct ring_buffer *op_ring_buffer_read;
|
|
extern struct ring_buffer *op_ring_buffer_write;
|
|
DECLARE_PER_CPU(struct oprofile_cpu_buffer, cpu_buffer);
|
|
|
|
/*
|
|
* Resets the cpu buffer to a sane state.
|
|
*
|
|
* reset these to invalid values; the next sample collected will
|
|
* populate the buffer with proper values to initialize the buffer
|
|
*/
|
|
static inline void cpu_buffer_reset(int cpu)
|
|
{
|
|
struct oprofile_cpu_buffer *cpu_buf = &per_cpu(cpu_buffer, cpu);
|
|
|
|
cpu_buf->last_is_kernel = -1;
|
|
cpu_buf->last_task = NULL;
|
|
}
|
|
|
|
static inline int cpu_buffer_write_entry(struct op_entry *entry)
|
|
{
|
|
entry->event = ring_buffer_lock_reserve(op_ring_buffer_write,
|
|
sizeof(struct op_sample),
|
|
&entry->irq_flags);
|
|
if (entry->event)
|
|
entry->sample = ring_buffer_event_data(entry->event);
|
|
else
|
|
entry->sample = NULL;
|
|
|
|
if (!entry->sample)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int cpu_buffer_write_commit(struct op_entry *entry)
|
|
{
|
|
return ring_buffer_unlock_commit(op_ring_buffer_write, entry->event,
|
|
entry->irq_flags);
|
|
}
|
|
|
|
static inline struct op_sample *cpu_buffer_read_entry(int cpu)
|
|
{
|
|
struct ring_buffer_event *e;
|
|
e = ring_buffer_consume(op_ring_buffer_read, cpu, NULL);
|
|
if (e)
|
|
return ring_buffer_event_data(e);
|
|
if (ring_buffer_swap_cpu(op_ring_buffer_read,
|
|
op_ring_buffer_write,
|
|
cpu))
|
|
return NULL;
|
|
e = ring_buffer_consume(op_ring_buffer_read, cpu, NULL);
|
|
if (e)
|
|
return ring_buffer_event_data(e);
|
|
return NULL;
|
|
}
|
|
|
|
/* "acquire" as many cpu buffer slots as we can */
|
|
static inline unsigned long cpu_buffer_entries(int cpu)
|
|
{
|
|
return ring_buffer_entries_cpu(op_ring_buffer_read, cpu)
|
|
+ ring_buffer_entries_cpu(op_ring_buffer_write, cpu);
|
|
}
|
|
|
|
/* transient events for the CPU buffer -> event buffer */
|
|
#define CPU_IS_KERNEL 1
|
|
#define CPU_TRACE_BEGIN 2
|
|
#define IBS_FETCH_BEGIN 3
|
|
#define IBS_OP_BEGIN 4
|
|
|
|
#endif /* OPROFILE_CPU_BUFFER_H */
|