libperf: Switch cpu to more accurate cpu_map_idx
Modify variable names and adopt perf_cpu_map__for_each_cpu() in perf_evsel__open(). Renaming is done by looking for consistency in API usage. Signed-off-by: Ian Rogers <irogers@google.com> Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: James Clark <james.clark@arm.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: John Garry <john.garry@huawei.com> Cc: Kajol Jain <kjain@linux.ibm.com> Cc: Kan Liang <kan.liang@linux.intel.com> Cc: Leo Yan <leo.yan@linaro.org> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Mathieu Poirier <mathieu.poirier@linaro.org> Cc: Mike Leach <mike.leach@linaro.org> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Paul Clarke <pc@us.ibm.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Riccardo Mancini <rickyman7@gmail.com> Cc: Stephane Eranian <eranian@google.com> Cc: Suzuki Poulouse <suzuki.poulose@arm.com> Cc: Vineet Singh <vineet.singh@intel.com> Cc: coresight@lists.linaro.org Cc: linux-arm-kernel@lists.infradead.org Cc: zhengjun.xing@intel.com Link: https://lore.kernel.org/r/20220105061351.120843-28-irogers@google.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
2ca0a3718d
commit
7e3d1784c8
|
@ -43,18 +43,22 @@ void perf_evsel__delete(struct perf_evsel *evsel)
|
|||
free(evsel);
|
||||
}
|
||||
|
||||
#define FD(e, x, y) ((int *) xyarray__entry(e->fd, x, y))
|
||||
#define MMAP(e, x, y) (e->mmap ? ((struct perf_mmap *) xyarray__entry(e->mmap, x, y)) : NULL)
|
||||
#define FD(_evsel, _cpu_map_idx, _thread) \
|
||||
((int *)xyarray__entry(_evsel->fd, _cpu_map_idx, _thread))
|
||||
#define MMAP(_evsel, _cpu_map_idx, _thread) \
|
||||
(_evsel->mmap ? ((struct perf_mmap *) xyarray__entry(_evsel->mmap, _cpu_map_idx, _thread)) \
|
||||
: NULL)
|
||||
|
||||
int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads)
|
||||
{
|
||||
evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int));
|
||||
|
||||
if (evsel->fd) {
|
||||
int cpu, thread;
|
||||
for (cpu = 0; cpu < ncpus; cpu++) {
|
||||
int idx, thread;
|
||||
|
||||
for (idx = 0; idx < ncpus; idx++) {
|
||||
for (thread = 0; thread < nthreads; thread++) {
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, idx, thread);
|
||||
|
||||
if (fd)
|
||||
*fd = -1;
|
||||
|
@ -80,7 +84,7 @@ sys_perf_event_open(struct perf_event_attr *attr,
|
|||
return syscall(__NR_perf_event_open, attr, pid, cpu, group_fd, flags);
|
||||
}
|
||||
|
||||
static int get_group_fd(struct perf_evsel *evsel, int cpu, int thread, int *group_fd)
|
||||
static int get_group_fd(struct perf_evsel *evsel, int cpu_map_idx, int thread, int *group_fd)
|
||||
{
|
||||
struct perf_evsel *leader = evsel->leader;
|
||||
int *fd;
|
||||
|
@ -97,7 +101,7 @@ static int get_group_fd(struct perf_evsel *evsel, int cpu, int thread, int *grou
|
|||
if (!leader->fd)
|
||||
return -ENOTCONN;
|
||||
|
||||
fd = FD(leader, cpu, thread);
|
||||
fd = FD(leader, cpu_map_idx, thread);
|
||||
if (fd == NULL || *fd == -1)
|
||||
return -EBADF;
|
||||
|
||||
|
@ -109,7 +113,7 @@ static int get_group_fd(struct perf_evsel *evsel, int cpu, int thread, int *grou
|
|||
int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
|
||||
struct perf_thread_map *threads)
|
||||
{
|
||||
int cpu, thread, err = 0;
|
||||
int cpu, idx, thread, err = 0;
|
||||
|
||||
if (cpus == NULL) {
|
||||
static struct perf_cpu_map *empty_cpu_map;
|
||||
|
@ -139,21 +143,21 @@ int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
|
|||
perf_evsel__alloc_fd(evsel, cpus->nr, threads->nr) < 0)
|
||||
return -ENOMEM;
|
||||
|
||||
for (cpu = 0; cpu < cpus->nr; cpu++) {
|
||||
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
|
||||
for (thread = 0; thread < threads->nr; thread++) {
|
||||
int fd, group_fd, *evsel_fd;
|
||||
|
||||
evsel_fd = FD(evsel, cpu, thread);
|
||||
evsel_fd = FD(evsel, idx, thread);
|
||||
if (evsel_fd == NULL)
|
||||
return -EINVAL;
|
||||
|
||||
err = get_group_fd(evsel, cpu, thread, &group_fd);
|
||||
err = get_group_fd(evsel, idx, thread, &group_fd);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
fd = sys_perf_event_open(&evsel->attr,
|
||||
threads->map[thread].pid,
|
||||
cpus->map[cpu], group_fd, 0);
|
||||
cpu, group_fd, 0);
|
||||
|
||||
if (fd < 0)
|
||||
return -errno;
|
||||
|
@ -165,12 +169,12 @@ int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
|
|||
return err;
|
||||
}
|
||||
|
||||
static void perf_evsel__close_fd_cpu(struct perf_evsel *evsel, int cpu)
|
||||
static void perf_evsel__close_fd_cpu(struct perf_evsel *evsel, int cpu_map_idx)
|
||||
{
|
||||
int thread;
|
||||
|
||||
for (thread = 0; thread < xyarray__max_y(evsel->fd); ++thread) {
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, cpu_map_idx, thread);
|
||||
|
||||
if (fd && *fd >= 0) {
|
||||
close(*fd);
|
||||
|
@ -181,10 +185,8 @@ static void perf_evsel__close_fd_cpu(struct perf_evsel *evsel, int cpu)
|
|||
|
||||
void perf_evsel__close_fd(struct perf_evsel *evsel)
|
||||
{
|
||||
int cpu;
|
||||
|
||||
for (cpu = 0; cpu < xyarray__max_x(evsel->fd); cpu++)
|
||||
perf_evsel__close_fd_cpu(evsel, cpu);
|
||||
for (int idx = 0; idx < xyarray__max_x(evsel->fd); idx++)
|
||||
perf_evsel__close_fd_cpu(evsel, idx);
|
||||
}
|
||||
|
||||
void perf_evsel__free_fd(struct perf_evsel *evsel)
|
||||
|
@ -202,29 +204,29 @@ void perf_evsel__close(struct perf_evsel *evsel)
|
|||
perf_evsel__free_fd(evsel);
|
||||
}
|
||||
|
||||
void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu)
|
||||
void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu_map_idx)
|
||||
{
|
||||
if (evsel->fd == NULL)
|
||||
return;
|
||||
|
||||
perf_evsel__close_fd_cpu(evsel, cpu);
|
||||
perf_evsel__close_fd_cpu(evsel, cpu_map_idx);
|
||||
}
|
||||
|
||||
void perf_evsel__munmap(struct perf_evsel *evsel)
|
||||
{
|
||||
int cpu, thread;
|
||||
int idx, thread;
|
||||
|
||||
if (evsel->fd == NULL || evsel->mmap == NULL)
|
||||
return;
|
||||
|
||||
for (cpu = 0; cpu < xyarray__max_x(evsel->fd); cpu++) {
|
||||
for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) {
|
||||
for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, idx, thread);
|
||||
|
||||
if (fd == NULL || *fd < 0)
|
||||
continue;
|
||||
|
||||
perf_mmap__munmap(MMAP(evsel, cpu, thread));
|
||||
perf_mmap__munmap(MMAP(evsel, idx, thread));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -234,7 +236,7 @@ void perf_evsel__munmap(struct perf_evsel *evsel)
|
|||
|
||||
int perf_evsel__mmap(struct perf_evsel *evsel, int pages)
|
||||
{
|
||||
int ret, cpu, thread;
|
||||
int ret, idx, thread;
|
||||
struct perf_mmap_param mp = {
|
||||
.prot = PROT_READ | PROT_WRITE,
|
||||
.mask = (pages * page_size) - 1,
|
||||
|
@ -246,18 +248,18 @@ int perf_evsel__mmap(struct perf_evsel *evsel, int pages)
|
|||
if (perf_evsel__alloc_mmap(evsel, xyarray__max_x(evsel->fd), xyarray__max_y(evsel->fd)) < 0)
|
||||
return -ENOMEM;
|
||||
|
||||
for (cpu = 0; cpu < xyarray__max_x(evsel->fd); cpu++) {
|
||||
for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) {
|
||||
for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, idx, thread);
|
||||
struct perf_mmap *map;
|
||||
|
||||
if (fd == NULL || *fd < 0)
|
||||
continue;
|
||||
|
||||
map = MMAP(evsel, cpu, thread);
|
||||
map = MMAP(evsel, idx, thread);
|
||||
perf_mmap__init(map, NULL, false, NULL);
|
||||
|
||||
ret = perf_mmap__mmap(map, &mp, *fd, cpu);
|
||||
ret = perf_mmap__mmap(map, &mp, *fd, idx);
|
||||
if (ret) {
|
||||
perf_evsel__munmap(evsel);
|
||||
return ret;
|
||||
|
@ -268,14 +270,14 @@ int perf_evsel__mmap(struct perf_evsel *evsel, int pages)
|
|||
return 0;
|
||||
}
|
||||
|
||||
void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu, int thread)
|
||||
void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread)
|
||||
{
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, cpu_map_idx, thread);
|
||||
|
||||
if (fd == NULL || *fd < 0 || MMAP(evsel, cpu, thread) == NULL)
|
||||
if (fd == NULL || *fd < 0 || MMAP(evsel, cpu_map_idx, thread) == NULL)
|
||||
return NULL;
|
||||
|
||||
return MMAP(evsel, cpu, thread)->base;
|
||||
return MMAP(evsel, cpu_map_idx, thread)->base;
|
||||
}
|
||||
|
||||
int perf_evsel__read_size(struct perf_evsel *evsel)
|
||||
|
@ -303,19 +305,19 @@ int perf_evsel__read_size(struct perf_evsel *evsel)
|
|||
return size;
|
||||
}
|
||||
|
||||
int perf_evsel__read(struct perf_evsel *evsel, int cpu, int thread,
|
||||
int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread,
|
||||
struct perf_counts_values *count)
|
||||
{
|
||||
size_t size = perf_evsel__read_size(evsel);
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, cpu_map_idx, thread);
|
||||
|
||||
memset(count, 0, sizeof(*count));
|
||||
|
||||
if (fd == NULL || *fd < 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (MMAP(evsel, cpu, thread) &&
|
||||
!perf_mmap__read_self(MMAP(evsel, cpu, thread), count))
|
||||
if (MMAP(evsel, cpu_map_idx, thread) &&
|
||||
!perf_mmap__read_self(MMAP(evsel, cpu_map_idx, thread), count))
|
||||
return 0;
|
||||
|
||||
if (readn(*fd, count->values, size) <= 0)
|
||||
|
@ -326,13 +328,13 @@ int perf_evsel__read(struct perf_evsel *evsel, int cpu, int thread,
|
|||
|
||||
static int perf_evsel__run_ioctl(struct perf_evsel *evsel,
|
||||
int ioc, void *arg,
|
||||
int cpu)
|
||||
int cpu_map_idx)
|
||||
{
|
||||
int thread;
|
||||
|
||||
for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
|
||||
int err;
|
||||
int *fd = FD(evsel, cpu, thread);
|
||||
int *fd = FD(evsel, cpu_map_idx, thread);
|
||||
|
||||
if (fd == NULL || *fd < 0)
|
||||
return -1;
|
||||
|
@ -346,9 +348,9 @@ static int perf_evsel__run_ioctl(struct perf_evsel *evsel,
|
|||
return 0;
|
||||
}
|
||||
|
||||
int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu)
|
||||
int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu_map_idx)
|
||||
{
|
||||
return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, cpu);
|
||||
return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, cpu_map_idx);
|
||||
}
|
||||
|
||||
int perf_evsel__enable(struct perf_evsel *evsel)
|
||||
|
@ -361,9 +363,9 @@ int perf_evsel__enable(struct perf_evsel *evsel)
|
|||
return err;
|
||||
}
|
||||
|
||||
int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu)
|
||||
int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu_map_idx)
|
||||
{
|
||||
return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, cpu);
|
||||
return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, cpu_map_idx);
|
||||
}
|
||||
|
||||
int perf_evsel__disable(struct perf_evsel *evsel)
|
||||
|
|
|
@ -28,16 +28,16 @@ LIBPERF_API void perf_evsel__delete(struct perf_evsel *evsel);
|
|||
LIBPERF_API int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
|
||||
struct perf_thread_map *threads);
|
||||
LIBPERF_API void perf_evsel__close(struct perf_evsel *evsel);
|
||||
LIBPERF_API void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu);
|
||||
LIBPERF_API void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu_map_idx);
|
||||
LIBPERF_API int perf_evsel__mmap(struct perf_evsel *evsel, int pages);
|
||||
LIBPERF_API void perf_evsel__munmap(struct perf_evsel *evsel);
|
||||
LIBPERF_API void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu, int thread);
|
||||
LIBPERF_API int perf_evsel__read(struct perf_evsel *evsel, int cpu, int thread,
|
||||
LIBPERF_API void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread);
|
||||
LIBPERF_API int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread,
|
||||
struct perf_counts_values *count);
|
||||
LIBPERF_API int perf_evsel__enable(struct perf_evsel *evsel);
|
||||
LIBPERF_API int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu);
|
||||
LIBPERF_API int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu_map_idx);
|
||||
LIBPERF_API int perf_evsel__disable(struct perf_evsel *evsel);
|
||||
LIBPERF_API int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu);
|
||||
LIBPERF_API int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu_map_idx);
|
||||
LIBPERF_API struct perf_cpu_map *perf_evsel__cpus(struct perf_evsel *evsel);
|
||||
LIBPERF_API struct perf_thread_map *perf_evsel__threads(struct perf_evsel *evsel);
|
||||
LIBPERF_API struct perf_event_attr *perf_evsel__attr(struct perf_evsel *evsel);
|
||||
|
|
Loading…
Reference in New Issue