2015-04-09 23:53:42 +08:00
|
|
|
/*
|
|
|
|
* auxtrace.c: AUX area trace support
|
|
|
|
* Copyright (c) 2013-2015, Intel Corporation.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms and conditions of the GNU General Public License,
|
|
|
|
* version 2, as published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
|
|
* more details.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/mman.h>
|
|
|
|
#include <stdbool.h>
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/perf_event.h>
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/bitops.h>
|
|
|
|
#include <linux/log2.h>
|
2015-04-21 17:21:51 +08:00
|
|
|
#include <linux/string.h>
|
2015-04-09 23:53:42 +08:00
|
|
|
|
2015-04-21 17:21:51 +08:00
|
|
|
#include <sys/param.h>
|
2015-04-09 23:53:44 +08:00
|
|
|
#include <stdlib.h>
|
2015-04-09 23:53:50 +08:00
|
|
|
#include <stdio.h>
|
2015-04-09 23:53:44 +08:00
|
|
|
#include <string.h>
|
2015-04-21 17:21:51 +08:00
|
|
|
#include <limits.h>
|
2015-04-09 23:53:44 +08:00
|
|
|
#include <errno.h>
|
2015-04-21 17:21:51 +08:00
|
|
|
#include <linux/list.h>
|
2015-04-09 23:53:44 +08:00
|
|
|
|
2015-04-09 23:53:42 +08:00
|
|
|
#include "../perf.h"
|
|
|
|
#include "util.h"
|
|
|
|
#include "evlist.h"
|
|
|
|
#include "cpumap.h"
|
|
|
|
#include "thread_map.h"
|
|
|
|
#include "asm/bug.h"
|
|
|
|
#include "auxtrace.h"
|
|
|
|
|
2015-04-09 23:53:54 +08:00
|
|
|
#include <linux/hash.h>
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
#include "event.h"
|
2015-04-09 23:53:50 +08:00
|
|
|
#include "session.h"
|
2015-04-09 23:53:44 +08:00
|
|
|
#include "debug.h"
|
2015-04-09 23:53:49 +08:00
|
|
|
#include "parse-options.h"
|
2015-04-09 23:53:44 +08:00
|
|
|
|
2015-07-18 00:33:42 +08:00
|
|
|
#include "intel-pt.h"
|
perf tools: Add Intel BTS support
Intel BTS support fits within the new auxtrace infrastructure. Recording is
supporting by identifying the Intel BTS PMU, parsing options and setting up
events.
Decoding is supported by queuing up trace data by thread and then decoding
synchronously delivering synthesized event samples into the session processing
for tools to consume.
Committer note:
E.g:
[root@felicio ~]# perf record --per-thread -e intel_bts// ls
anaconda-ks.cfg apctest.output bin kernel-rt-3.10.0-298.rt56.171.el7.x86_64.rpm libexec lock_page.bpf.c perf.data perf.data.old
[ perf record: Woken up 3 times to write data ]
[ perf record: Captured and wrote 4.367 MB perf.data ]
[root@felicio ~]# perf evlist -v
intel_bts//: type: 6, size: 112, { sample_period, sample_freq }: 1, sample_type: IP|TID|IDENTIFIER, read_format: ID, disabled: 1, enable_on_exec: 1, sample_id_all: 1, exclude_guest: 1
dummy:u: type: 1, size: 112, config: 0x9, { sample_period, sample_freq }: 1, sample_type: IP|TID|IDENTIFIER, read_format: ID, disabled: 1, exclude_kernel: 1, exclude_hv: 1, mmap: 1, comm: 1, enable_on_exec: 1, task: 1, sample_id_all: 1, mmap2: 1, comm_exec: 1
[root@felicio ~]# perf script # the navigate in the pager to some interesting place:
ls 1843 1 branches: ffffffff810a60cb flush_signal_handlers ([kernel.kallsyms]) => ffffffff8121a522 setup_new_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8121a529 setup_new_exec ([kernel.kallsyms]) => ffffffff8122fa30 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa5d do_close_on_exec ([kernel.kallsyms]) => ffffffff81767ae0 _raw_spin_lock ([kernel.kallsyms])
ls 1843 1 branches: ffffffff81767af4 _raw_spin_lock ([kernel.kallsyms]) => ffffffff8122fa62 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fac9 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fad2 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fadd do_close_on_exec ([kernel.kallsyms]) => ffffffff8120fc80 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcaf filp_close ([kernel.kallsyms]) => ffffffff8120fcb6 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcc2 filp_close ([kernel.kallsyms]) => ffffffff812547f0 dnotify_flush ([kernel.kallsyms])
ls 1843 1 branches: ffffffff81254823 dnotify_flush ([kernel.kallsyms]) => ffffffff8120fcc7 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fccd filp_close ([kernel.kallsyms]) => ffffffff81261790 locks_remove_posix ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812617a3 locks_remove_posix ([kernel.kallsyms]) => ffffffff812617b9 locks_remove_posix ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812617b9 locks_remove_posix ([kernel.kallsyms]) => ffffffff8120fcd2 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcd5 filp_close ([kernel.kallsyms]) => ffffffff812142c0 fput ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812142d6 fput ([kernel.kallsyms]) => ffffffff812142df fput ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8121430c fput ([kernel.kallsyms]) => ffffffff810b6580 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65ad task_work_add ([kernel.kallsyms]) => ffffffff810b65b1 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65c1 task_work_add ([kernel.kallsyms]) => ffffffff810bc710 kick_process ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810bc725 kick_process ([kernel.kallsyms]) => ffffffff810bc742 kick_process ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810bc742 kick_process ([kernel.kallsyms]) => ffffffff810b65c6 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65c9 task_work_add ([kernel.kallsyms]) => ffffffff81214311 fput ([kernel.kallsyms])
Signed-off-by: Adrian Hunter <adrian.hunter@intel.com>
Tested-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Link: http://lkml.kernel.org/r/1437150840-31811-9-git-send-email-adrian.hunter@intel.com
[ Merged sample->time fix for bug found after first round of testing on slightly older kernel ]
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2015-07-18 00:33:43 +08:00
|
|
|
#include "intel-bts.h"
|
2015-07-18 00:33:42 +08:00
|
|
|
|
2015-04-09 23:53:42 +08:00
|
|
|
int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
|
|
|
|
struct auxtrace_mmap_params *mp,
|
|
|
|
void *userpg, int fd)
|
|
|
|
{
|
|
|
|
struct perf_event_mmap_page *pc = userpg;
|
|
|
|
|
|
|
|
WARN_ONCE(mm->base, "Uninitialized auxtrace_mmap\n");
|
|
|
|
|
|
|
|
mm->userpg = userpg;
|
|
|
|
mm->mask = mp->mask;
|
|
|
|
mm->len = mp->len;
|
|
|
|
mm->prev = 0;
|
|
|
|
mm->idx = mp->idx;
|
|
|
|
mm->tid = mp->tid;
|
|
|
|
mm->cpu = mp->cpu;
|
|
|
|
|
|
|
|
if (!mp->len) {
|
|
|
|
mm->base = NULL;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-07-14 20:32:41 +08:00
|
|
|
#if BITS_PER_LONG != 64 && !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
|
|
|
|
pr_err("Cannot use AUX area tracing mmaps\n");
|
|
|
|
return -1;
|
|
|
|
#endif
|
|
|
|
|
2015-04-09 23:53:42 +08:00
|
|
|
pc->aux_offset = mp->offset;
|
|
|
|
pc->aux_size = mp->len;
|
|
|
|
|
|
|
|
mm->base = mmap(NULL, mp->len, mp->prot, MAP_SHARED, fd, mp->offset);
|
|
|
|
if (mm->base == MAP_FAILED) {
|
|
|
|
pr_debug2("failed to mmap AUX area\n");
|
|
|
|
mm->base = NULL;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_mmap__munmap(struct auxtrace_mmap *mm)
|
|
|
|
{
|
|
|
|
if (mm->base) {
|
|
|
|
munmap(mm->base, mm->len);
|
|
|
|
mm->base = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
|
|
|
|
off_t auxtrace_offset,
|
|
|
|
unsigned int auxtrace_pages,
|
|
|
|
bool auxtrace_overwrite)
|
|
|
|
{
|
|
|
|
if (auxtrace_pages) {
|
|
|
|
mp->offset = auxtrace_offset;
|
|
|
|
mp->len = auxtrace_pages * (size_t)page_size;
|
|
|
|
mp->mask = is_power_of_2(mp->len) ? mp->len - 1 : 0;
|
|
|
|
mp->prot = PROT_READ | (auxtrace_overwrite ? 0 : PROT_WRITE);
|
|
|
|
pr_debug2("AUX area mmap length %zu\n", mp->len);
|
|
|
|
} else {
|
|
|
|
mp->len = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
|
|
|
|
struct perf_evlist *evlist, int idx,
|
|
|
|
bool per_cpu)
|
|
|
|
{
|
|
|
|
mp->idx = idx;
|
|
|
|
|
|
|
|
if (per_cpu) {
|
|
|
|
mp->cpu = evlist->cpus->map[idx];
|
|
|
|
if (evlist->threads)
|
2015-06-23 06:36:02 +08:00
|
|
|
mp->tid = thread_map__pid(evlist->threads, 0);
|
2015-04-09 23:53:42 +08:00
|
|
|
else
|
|
|
|
mp->tid = -1;
|
|
|
|
} else {
|
|
|
|
mp->cpu = -1;
|
2015-06-23 06:36:02 +08:00
|
|
|
mp->tid = thread_map__pid(evlist->threads, idx);
|
2015-04-09 23:53:42 +08:00
|
|
|
}
|
|
|
|
}
|
2015-04-09 23:53:44 +08:00
|
|
|
|
2015-04-21 17:21:51 +08:00
|
|
|
#define AUXTRACE_INIT_NR_QUEUES 32
|
|
|
|
|
|
|
|
static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues)
|
|
|
|
{
|
|
|
|
struct auxtrace_queue *queue_array;
|
|
|
|
unsigned int max_nr_queues, i;
|
|
|
|
|
|
|
|
max_nr_queues = UINT_MAX / sizeof(struct auxtrace_queue);
|
|
|
|
if (nr_queues > max_nr_queues)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue));
|
|
|
|
if (!queue_array)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
for (i = 0; i < nr_queues; i++) {
|
|
|
|
INIT_LIST_HEAD(&queue_array[i].head);
|
|
|
|
queue_array[i].priv = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return queue_array;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_queues__init(struct auxtrace_queues *queues)
|
|
|
|
{
|
|
|
|
queues->nr_queues = AUXTRACE_INIT_NR_QUEUES;
|
|
|
|
queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues);
|
|
|
|
if (!queues->queue_array)
|
|
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_queues__grow(struct auxtrace_queues *queues,
|
|
|
|
unsigned int new_nr_queues)
|
|
|
|
{
|
|
|
|
unsigned int nr_queues = queues->nr_queues;
|
|
|
|
struct auxtrace_queue *queue_array;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
if (!nr_queues)
|
|
|
|
nr_queues = AUXTRACE_INIT_NR_QUEUES;
|
|
|
|
|
|
|
|
while (nr_queues && nr_queues < new_nr_queues)
|
|
|
|
nr_queues <<= 1;
|
|
|
|
|
|
|
|
if (nr_queues < queues->nr_queues || nr_queues < new_nr_queues)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
queue_array = auxtrace_alloc_queue_array(nr_queues);
|
|
|
|
if (!queue_array)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
for (i = 0; i < queues->nr_queues; i++) {
|
|
|
|
list_splice_tail(&queues->queue_array[i].head,
|
|
|
|
&queue_array[i].head);
|
|
|
|
queue_array[i].priv = queues->queue_array[i].priv;
|
|
|
|
}
|
|
|
|
|
|
|
|
queues->nr_queues = nr_queues;
|
|
|
|
queues->queue_array = queue_array;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *auxtrace_copy_data(u64 size, struct perf_session *session)
|
|
|
|
{
|
|
|
|
int fd = perf_data_file__fd(session->file);
|
|
|
|
void *p;
|
|
|
|
ssize_t ret;
|
|
|
|
|
|
|
|
if (size > SSIZE_MAX)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
p = malloc(size);
|
|
|
|
if (!p)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
ret = readn(fd, p, size);
|
|
|
|
if (ret != (ssize_t)size) {
|
|
|
|
free(p);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_queues__add_buffer(struct auxtrace_queues *queues,
|
|
|
|
unsigned int idx,
|
|
|
|
struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
struct auxtrace_queue *queue;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
if (idx >= queues->nr_queues) {
|
|
|
|
err = auxtrace_queues__grow(queues, idx + 1);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
queue = &queues->queue_array[idx];
|
|
|
|
|
|
|
|
if (!queue->set) {
|
|
|
|
queue->set = true;
|
|
|
|
queue->tid = buffer->tid;
|
|
|
|
queue->cpu = buffer->cpu;
|
|
|
|
} else if (buffer->cpu != queue->cpu || buffer->tid != queue->tid) {
|
|
|
|
pr_err("auxtrace queue conflict: cpu %d, tid %d vs cpu %d, tid %d\n",
|
|
|
|
queue->cpu, queue->tid, buffer->cpu, buffer->tid);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
buffer->buffer_nr = queues->next_buffer_nr++;
|
|
|
|
|
|
|
|
list_add_tail(&buffer->list, &queue->head);
|
|
|
|
|
|
|
|
queues->new_data = true;
|
|
|
|
queues->populated = true;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Limit buffers to 32MiB on 32-bit */
|
|
|
|
#define BUFFER_LIMIT_FOR_32_BIT (32 * 1024 * 1024)
|
|
|
|
|
|
|
|
static int auxtrace_queues__split_buffer(struct auxtrace_queues *queues,
|
|
|
|
unsigned int idx,
|
|
|
|
struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
u64 sz = buffer->size;
|
|
|
|
bool consecutive = false;
|
|
|
|
struct auxtrace_buffer *b;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
while (sz > BUFFER_LIMIT_FOR_32_BIT) {
|
|
|
|
b = memdup(buffer, sizeof(struct auxtrace_buffer));
|
|
|
|
if (!b)
|
|
|
|
return -ENOMEM;
|
|
|
|
b->size = BUFFER_LIMIT_FOR_32_BIT;
|
|
|
|
b->consecutive = consecutive;
|
|
|
|
err = auxtrace_queues__add_buffer(queues, idx, b);
|
|
|
|
if (err) {
|
|
|
|
auxtrace_buffer__free(b);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
buffer->data_offset += BUFFER_LIMIT_FOR_32_BIT;
|
|
|
|
sz -= BUFFER_LIMIT_FOR_32_BIT;
|
|
|
|
consecutive = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
buffer->size = sz;
|
|
|
|
buffer->consecutive = consecutive;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_queues__add_event_buffer(struct auxtrace_queues *queues,
|
|
|
|
struct perf_session *session,
|
|
|
|
unsigned int idx,
|
|
|
|
struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
if (session->one_mmap) {
|
|
|
|
buffer->data = buffer->data_offset - session->one_mmap_offset +
|
|
|
|
session->one_mmap_addr;
|
|
|
|
} else if (perf_data_file__is_pipe(session->file)) {
|
|
|
|
buffer->data = auxtrace_copy_data(buffer->size, session);
|
|
|
|
if (!buffer->data)
|
|
|
|
return -ENOMEM;
|
|
|
|
buffer->data_needs_freeing = true;
|
|
|
|
} else if (BITS_PER_LONG == 32 &&
|
|
|
|
buffer->size > BUFFER_LIMIT_FOR_32_BIT) {
|
|
|
|
int err;
|
|
|
|
|
|
|
|
err = auxtrace_queues__split_buffer(queues, idx, buffer);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
return auxtrace_queues__add_buffer(queues, idx, buffer);
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_queues__add_event(struct auxtrace_queues *queues,
|
|
|
|
struct perf_session *session,
|
|
|
|
union perf_event *event, off_t data_offset,
|
|
|
|
struct auxtrace_buffer **buffer_ptr)
|
|
|
|
{
|
|
|
|
struct auxtrace_buffer *buffer;
|
|
|
|
unsigned int idx;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
buffer = zalloc(sizeof(struct auxtrace_buffer));
|
|
|
|
if (!buffer)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
buffer->pid = -1;
|
|
|
|
buffer->tid = event->auxtrace.tid;
|
|
|
|
buffer->cpu = event->auxtrace.cpu;
|
|
|
|
buffer->data_offset = data_offset;
|
|
|
|
buffer->offset = event->auxtrace.offset;
|
|
|
|
buffer->reference = event->auxtrace.reference;
|
|
|
|
buffer->size = event->auxtrace.size;
|
|
|
|
idx = event->auxtrace.idx;
|
|
|
|
|
|
|
|
err = auxtrace_queues__add_event_buffer(queues, session, idx, buffer);
|
|
|
|
if (err)
|
|
|
|
goto out_err;
|
|
|
|
|
|
|
|
if (buffer_ptr)
|
|
|
|
*buffer_ptr = buffer;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_err:
|
|
|
|
auxtrace_buffer__free(buffer);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2015-04-30 22:37:25 +08:00
|
|
|
static int auxtrace_queues__add_indexed_event(struct auxtrace_queues *queues,
|
|
|
|
struct perf_session *session,
|
|
|
|
off_t file_offset, size_t sz)
|
|
|
|
{
|
|
|
|
union perf_event *event;
|
|
|
|
int err;
|
|
|
|
char buf[PERF_SAMPLE_MAX_SIZE];
|
|
|
|
|
|
|
|
err = perf_session__peek_event(session, file_offset, buf,
|
|
|
|
PERF_SAMPLE_MAX_SIZE, &event, NULL);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
if (event->header.type == PERF_RECORD_AUXTRACE) {
|
|
|
|
if (event->header.size < sizeof(struct auxtrace_event) ||
|
|
|
|
event->header.size != sz) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
file_offset += event->header.size;
|
|
|
|
err = auxtrace_queues__add_event(queues, session, event,
|
|
|
|
file_offset, NULL);
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2015-04-21 17:21:51 +08:00
|
|
|
void auxtrace_queues__free(struct auxtrace_queues *queues)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < queues->nr_queues; i++) {
|
|
|
|
while (!list_empty(&queues->queue_array[i].head)) {
|
|
|
|
struct auxtrace_buffer *buffer;
|
|
|
|
|
|
|
|
buffer = list_entry(queues->queue_array[i].head.next,
|
|
|
|
struct auxtrace_buffer, list);
|
|
|
|
list_del(&buffer->list);
|
|
|
|
auxtrace_buffer__free(buffer);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
zfree(&queues->queue_array);
|
|
|
|
queues->nr_queues = 0;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:52 +08:00
|
|
|
static void auxtrace_heapify(struct auxtrace_heap_item *heap_array,
|
|
|
|
unsigned int pos, unsigned int queue_nr,
|
|
|
|
u64 ordinal)
|
|
|
|
{
|
|
|
|
unsigned int parent;
|
|
|
|
|
|
|
|
while (pos) {
|
|
|
|
parent = (pos - 1) >> 1;
|
|
|
|
if (heap_array[parent].ordinal <= ordinal)
|
|
|
|
break;
|
|
|
|
heap_array[pos] = heap_array[parent];
|
|
|
|
pos = parent;
|
|
|
|
}
|
|
|
|
heap_array[pos].queue_nr = queue_nr;
|
|
|
|
heap_array[pos].ordinal = ordinal;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
|
|
|
|
u64 ordinal)
|
|
|
|
{
|
|
|
|
struct auxtrace_heap_item *heap_array;
|
|
|
|
|
|
|
|
if (queue_nr >= heap->heap_sz) {
|
|
|
|
unsigned int heap_sz = AUXTRACE_INIT_NR_QUEUES;
|
|
|
|
|
|
|
|
while (heap_sz <= queue_nr)
|
|
|
|
heap_sz <<= 1;
|
|
|
|
heap_array = realloc(heap->heap_array,
|
|
|
|
heap_sz * sizeof(struct auxtrace_heap_item));
|
|
|
|
if (!heap_array)
|
|
|
|
return -ENOMEM;
|
|
|
|
heap->heap_array = heap_array;
|
|
|
|
heap->heap_sz = heap_sz;
|
|
|
|
}
|
|
|
|
|
|
|
|
auxtrace_heapify(heap->heap_array, heap->heap_cnt++, queue_nr, ordinal);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_heap__free(struct auxtrace_heap *heap)
|
|
|
|
{
|
|
|
|
zfree(&heap->heap_array);
|
|
|
|
heap->heap_cnt = 0;
|
|
|
|
heap->heap_sz = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_heap__pop(struct auxtrace_heap *heap)
|
|
|
|
{
|
|
|
|
unsigned int pos, last, heap_cnt = heap->heap_cnt;
|
|
|
|
struct auxtrace_heap_item *heap_array;
|
|
|
|
|
|
|
|
if (!heap_cnt)
|
|
|
|
return;
|
|
|
|
|
|
|
|
heap->heap_cnt -= 1;
|
|
|
|
|
|
|
|
heap_array = heap->heap_array;
|
|
|
|
|
|
|
|
pos = 0;
|
|
|
|
while (1) {
|
|
|
|
unsigned int left, right;
|
|
|
|
|
|
|
|
left = (pos << 1) + 1;
|
|
|
|
if (left >= heap_cnt)
|
|
|
|
break;
|
|
|
|
right = left + 1;
|
|
|
|
if (right >= heap_cnt) {
|
|
|
|
heap_array[pos] = heap_array[left];
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (heap_array[left].ordinal < heap_array[right].ordinal) {
|
|
|
|
heap_array[pos] = heap_array[left];
|
|
|
|
pos = left;
|
|
|
|
} else {
|
|
|
|
heap_array[pos] = heap_array[right];
|
|
|
|
pos = right;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
last = heap_cnt - 1;
|
|
|
|
auxtrace_heapify(heap_array, pos, heap_array[last].queue_nr,
|
|
|
|
heap_array[last].ordinal);
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr)
|
|
|
|
{
|
|
|
|
if (itr)
|
|
|
|
return itr->info_priv_size(itr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_not_supported(void)
|
|
|
|
{
|
|
|
|
pr_err("AUX area tracing is not supported on this architecture\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_record__info_fill(struct auxtrace_record *itr,
|
|
|
|
struct perf_session *session,
|
|
|
|
struct auxtrace_info_event *auxtrace_info,
|
|
|
|
size_t priv_size)
|
|
|
|
{
|
|
|
|
if (itr)
|
|
|
|
return itr->info_fill(itr, session, auxtrace_info, priv_size);
|
|
|
|
return auxtrace_not_supported();
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_record__free(struct auxtrace_record *itr)
|
|
|
|
{
|
|
|
|
if (itr)
|
|
|
|
itr->free(itr);
|
|
|
|
}
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
int auxtrace_record__snapshot_start(struct auxtrace_record *itr)
|
|
|
|
{
|
|
|
|
if (itr && itr->snapshot_start)
|
|
|
|
return itr->snapshot_start(itr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_record__snapshot_finish(struct auxtrace_record *itr)
|
|
|
|
{
|
|
|
|
if (itr && itr->snapshot_finish)
|
|
|
|
return itr->snapshot_finish(itr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
|
|
|
|
struct auxtrace_mmap *mm,
|
|
|
|
unsigned char *data, u64 *head, u64 *old)
|
|
|
|
{
|
|
|
|
if (itr && itr->find_snapshot)
|
|
|
|
return itr->find_snapshot(itr, idx, mm, data, head, old);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
int auxtrace_record__options(struct auxtrace_record *itr,
|
|
|
|
struct perf_evlist *evlist,
|
|
|
|
struct record_opts *opts)
|
|
|
|
{
|
|
|
|
if (itr)
|
|
|
|
return itr->recording_options(itr, evlist, opts);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
u64 auxtrace_record__reference(struct auxtrace_record *itr)
|
|
|
|
{
|
|
|
|
if (itr)
|
|
|
|
return itr->reference(itr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
|
|
|
|
struct record_opts *opts, const char *str)
|
|
|
|
{
|
|
|
|
if (!str)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (itr)
|
|
|
|
return itr->parse_snapshot_options(itr, opts, str);
|
|
|
|
|
|
|
|
pr_err("No AUX area tracing to snapshot\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
struct auxtrace_record *__weak
|
|
|
|
auxtrace_record__init(struct perf_evlist *evlist __maybe_unused, int *err)
|
|
|
|
{
|
|
|
|
*err = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2015-04-30 22:37:25 +08:00
|
|
|
static int auxtrace_index__alloc(struct list_head *head)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
|
|
|
|
auxtrace_index = malloc(sizeof(struct auxtrace_index));
|
|
|
|
if (!auxtrace_index)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
auxtrace_index->nr = 0;
|
|
|
|
INIT_LIST_HEAD(&auxtrace_index->list);
|
|
|
|
|
|
|
|
list_add_tail(&auxtrace_index->list, head);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_index__free(struct list_head *head)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index, *n;
|
|
|
|
|
|
|
|
list_for_each_entry_safe(auxtrace_index, n, head, list) {
|
|
|
|
list_del(&auxtrace_index->list);
|
|
|
|
free(auxtrace_index);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct auxtrace_index *auxtrace_index__last(struct list_head *head)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
if (list_empty(head)) {
|
|
|
|
err = auxtrace_index__alloc(head);
|
|
|
|
if (err)
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
auxtrace_index = list_entry(head->prev, struct auxtrace_index, list);
|
|
|
|
|
|
|
|
if (auxtrace_index->nr >= PERF_AUXTRACE_INDEX_ENTRY_COUNT) {
|
|
|
|
err = auxtrace_index__alloc(head);
|
|
|
|
if (err)
|
|
|
|
return NULL;
|
|
|
|
auxtrace_index = list_entry(head->prev, struct auxtrace_index,
|
|
|
|
list);
|
|
|
|
}
|
|
|
|
|
|
|
|
return auxtrace_index;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_index__auxtrace_event(struct list_head *head,
|
|
|
|
union perf_event *event, off_t file_offset)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
size_t nr;
|
|
|
|
|
|
|
|
auxtrace_index = auxtrace_index__last(head);
|
|
|
|
if (!auxtrace_index)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
nr = auxtrace_index->nr;
|
|
|
|
auxtrace_index->entries[nr].file_offset = file_offset;
|
|
|
|
auxtrace_index->entries[nr].sz = event->header.size;
|
|
|
|
auxtrace_index->nr += 1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_index__do_write(int fd,
|
|
|
|
struct auxtrace_index *auxtrace_index)
|
|
|
|
{
|
|
|
|
struct auxtrace_index_entry ent;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; i < auxtrace_index->nr; i++) {
|
|
|
|
ent.file_offset = auxtrace_index->entries[i].file_offset;
|
|
|
|
ent.sz = auxtrace_index->entries[i].sz;
|
|
|
|
if (writen(fd, &ent, sizeof(ent)) != sizeof(ent))
|
|
|
|
return -errno;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_index__write(int fd, struct list_head *head)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
u64 total = 0;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
list_for_each_entry(auxtrace_index, head, list)
|
|
|
|
total += auxtrace_index->nr;
|
|
|
|
|
|
|
|
if (writen(fd, &total, sizeof(total)) != sizeof(total))
|
|
|
|
return -errno;
|
|
|
|
|
|
|
|
list_for_each_entry(auxtrace_index, head, list) {
|
|
|
|
err = auxtrace_index__do_write(fd, auxtrace_index);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_index__process_entry(int fd, struct list_head *head,
|
|
|
|
bool needs_swap)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
struct auxtrace_index_entry ent;
|
|
|
|
size_t nr;
|
|
|
|
|
|
|
|
if (readn(fd, &ent, sizeof(ent)) != sizeof(ent))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
auxtrace_index = auxtrace_index__last(head);
|
|
|
|
if (!auxtrace_index)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
nr = auxtrace_index->nr;
|
|
|
|
if (needs_swap) {
|
|
|
|
auxtrace_index->entries[nr].file_offset =
|
|
|
|
bswap_64(ent.file_offset);
|
|
|
|
auxtrace_index->entries[nr].sz = bswap_64(ent.sz);
|
|
|
|
} else {
|
|
|
|
auxtrace_index->entries[nr].file_offset = ent.file_offset;
|
|
|
|
auxtrace_index->entries[nr].sz = ent.sz;
|
|
|
|
}
|
|
|
|
|
|
|
|
auxtrace_index->nr = nr + 1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
|
|
|
|
bool needs_swap)
|
|
|
|
{
|
|
|
|
struct list_head *head = &session->auxtrace_index;
|
|
|
|
u64 nr;
|
|
|
|
|
|
|
|
if (readn(fd, &nr, sizeof(u64)) != sizeof(u64))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
if (needs_swap)
|
|
|
|
nr = bswap_64(nr);
|
|
|
|
|
|
|
|
if (sizeof(u64) + nr * sizeof(struct auxtrace_index_entry) > size)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
while (nr--) {
|
|
|
|
int err;
|
|
|
|
|
|
|
|
err = auxtrace_index__process_entry(fd, head, needs_swap);
|
|
|
|
if (err)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int auxtrace_queues__process_index_entry(struct auxtrace_queues *queues,
|
|
|
|
struct perf_session *session,
|
|
|
|
struct auxtrace_index_entry *ent)
|
|
|
|
{
|
|
|
|
return auxtrace_queues__add_indexed_event(queues, session,
|
|
|
|
ent->file_offset, ent->sz);
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_queues__process_index(struct auxtrace_queues *queues,
|
|
|
|
struct perf_session *session)
|
|
|
|
{
|
|
|
|
struct auxtrace_index *auxtrace_index;
|
|
|
|
struct auxtrace_index_entry *ent;
|
|
|
|
size_t i;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
list_for_each_entry(auxtrace_index, &session->auxtrace_index, list) {
|
|
|
|
for (i = 0; i < auxtrace_index->nr; i++) {
|
|
|
|
ent = &auxtrace_index->entries[i];
|
|
|
|
err = auxtrace_queues__process_index_entry(queues,
|
|
|
|
session,
|
|
|
|
ent);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-04-21 17:21:51 +08:00
|
|
|
struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
|
|
|
|
struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
if (buffer) {
|
|
|
|
if (list_is_last(&buffer->list, &queue->head))
|
|
|
|
return NULL;
|
|
|
|
return list_entry(buffer->list.next, struct auxtrace_buffer,
|
|
|
|
list);
|
|
|
|
} else {
|
|
|
|
if (list_empty(&queue->head))
|
|
|
|
return NULL;
|
|
|
|
return list_entry(queue->head.next, struct auxtrace_buffer,
|
|
|
|
list);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd)
|
|
|
|
{
|
|
|
|
size_t adj = buffer->data_offset & (page_size - 1);
|
|
|
|
size_t size = buffer->size + adj;
|
|
|
|
off_t file_offset = buffer->data_offset - adj;
|
|
|
|
void *addr;
|
|
|
|
|
|
|
|
if (buffer->data)
|
|
|
|
return buffer->data;
|
|
|
|
|
|
|
|
addr = mmap(NULL, size, PROT_READ, MAP_SHARED, fd, file_offset);
|
|
|
|
if (addr == MAP_FAILED)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
buffer->mmap_addr = addr;
|
|
|
|
buffer->mmap_size = size;
|
|
|
|
|
|
|
|
buffer->data = addr + adj;
|
|
|
|
|
|
|
|
return buffer->data;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
if (!buffer->data || !buffer->mmap_addr)
|
|
|
|
return;
|
|
|
|
munmap(buffer->mmap_addr, buffer->mmap_size);
|
|
|
|
buffer->mmap_addr = NULL;
|
|
|
|
buffer->mmap_size = 0;
|
|
|
|
buffer->data = NULL;
|
|
|
|
buffer->use_data = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
auxtrace_buffer__put_data(buffer);
|
|
|
|
if (buffer->data_needs_freeing) {
|
|
|
|
buffer->data_needs_freeing = false;
|
|
|
|
zfree(&buffer->data);
|
|
|
|
buffer->use_data = NULL;
|
|
|
|
buffer->size = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_buffer__free(struct auxtrace_buffer *buffer)
|
|
|
|
{
|
|
|
|
auxtrace_buffer__drop_data(buffer);
|
|
|
|
free(buffer);
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:50 +08:00
|
|
|
void auxtrace_synth_error(struct auxtrace_error_event *auxtrace_error, int type,
|
|
|
|
int code, int cpu, pid_t pid, pid_t tid, u64 ip,
|
|
|
|
const char *msg)
|
|
|
|
{
|
|
|
|
size_t size;
|
|
|
|
|
|
|
|
memset(auxtrace_error, 0, sizeof(struct auxtrace_error_event));
|
|
|
|
|
|
|
|
auxtrace_error->header.type = PERF_RECORD_AUXTRACE_ERROR;
|
|
|
|
auxtrace_error->type = type;
|
|
|
|
auxtrace_error->code = code;
|
|
|
|
auxtrace_error->cpu = cpu;
|
|
|
|
auxtrace_error->pid = pid;
|
|
|
|
auxtrace_error->tid = tid;
|
|
|
|
auxtrace_error->ip = ip;
|
|
|
|
strlcpy(auxtrace_error->msg, msg, MAX_AUXTRACE_ERROR_MSG);
|
|
|
|
|
|
|
|
size = (void *)auxtrace_error->msg - (void *)auxtrace_error +
|
|
|
|
strlen(auxtrace_error->msg) + 1;
|
|
|
|
auxtrace_error->header.size = PERF_ALIGN(size, sizeof(u64));
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
int perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr,
|
|
|
|
struct perf_tool *tool,
|
|
|
|
struct perf_session *session,
|
|
|
|
perf_event__handler_t process)
|
|
|
|
{
|
|
|
|
union perf_event *ev;
|
|
|
|
size_t priv_size;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
pr_debug2("Synthesizing auxtrace information\n");
|
|
|
|
priv_size = auxtrace_record__info_priv_size(itr);
|
|
|
|
ev = zalloc(sizeof(struct auxtrace_info_event) + priv_size);
|
|
|
|
if (!ev)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
ev->auxtrace_info.header.type = PERF_RECORD_AUXTRACE_INFO;
|
|
|
|
ev->auxtrace_info.header.size = sizeof(struct auxtrace_info_event) +
|
|
|
|
priv_size;
|
|
|
|
err = auxtrace_record__info_fill(itr, session, &ev->auxtrace_info,
|
|
|
|
priv_size);
|
|
|
|
if (err)
|
|
|
|
goto out_free;
|
|
|
|
|
|
|
|
err = process(tool, ev, NULL, NULL);
|
|
|
|
out_free:
|
|
|
|
free(ev);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:53 +08:00
|
|
|
static bool auxtrace__dont_decode(struct perf_session *session)
|
|
|
|
{
|
|
|
|
return !session->itrace_synth_opts ||
|
|
|
|
session->itrace_synth_opts->dont_decode;
|
|
|
|
}
|
|
|
|
|
|
|
|
int perf_event__process_auxtrace_info(struct perf_tool *tool __maybe_unused,
|
|
|
|
union perf_event *event,
|
2015-07-18 00:33:42 +08:00
|
|
|
struct perf_session *session)
|
2015-04-09 23:53:53 +08:00
|
|
|
{
|
|
|
|
enum auxtrace_type type = event->auxtrace_info.type;
|
|
|
|
|
|
|
|
if (dump_trace)
|
|
|
|
fprintf(stdout, " type: %u\n", type);
|
|
|
|
|
|
|
|
switch (type) {
|
2015-07-18 00:33:36 +08:00
|
|
|
case PERF_AUXTRACE_INTEL_PT:
|
2015-07-18 00:33:42 +08:00
|
|
|
return intel_pt_process_auxtrace_info(event, session);
|
perf tools: Add Intel BTS support
Intel BTS support fits within the new auxtrace infrastructure. Recording is
supporting by identifying the Intel BTS PMU, parsing options and setting up
events.
Decoding is supported by queuing up trace data by thread and then decoding
synchronously delivering synthesized event samples into the session processing
for tools to consume.
Committer note:
E.g:
[root@felicio ~]# perf record --per-thread -e intel_bts// ls
anaconda-ks.cfg apctest.output bin kernel-rt-3.10.0-298.rt56.171.el7.x86_64.rpm libexec lock_page.bpf.c perf.data perf.data.old
[ perf record: Woken up 3 times to write data ]
[ perf record: Captured and wrote 4.367 MB perf.data ]
[root@felicio ~]# perf evlist -v
intel_bts//: type: 6, size: 112, { sample_period, sample_freq }: 1, sample_type: IP|TID|IDENTIFIER, read_format: ID, disabled: 1, enable_on_exec: 1, sample_id_all: 1, exclude_guest: 1
dummy:u: type: 1, size: 112, config: 0x9, { sample_period, sample_freq }: 1, sample_type: IP|TID|IDENTIFIER, read_format: ID, disabled: 1, exclude_kernel: 1, exclude_hv: 1, mmap: 1, comm: 1, enable_on_exec: 1, task: 1, sample_id_all: 1, mmap2: 1, comm_exec: 1
[root@felicio ~]# perf script # the navigate in the pager to some interesting place:
ls 1843 1 branches: ffffffff810a60cb flush_signal_handlers ([kernel.kallsyms]) => ffffffff8121a522 setup_new_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8121a529 setup_new_exec ([kernel.kallsyms]) => ffffffff8122fa30 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa5d do_close_on_exec ([kernel.kallsyms]) => ffffffff81767ae0 _raw_spin_lock ([kernel.kallsyms])
ls 1843 1 branches: ffffffff81767af4 _raw_spin_lock ([kernel.kallsyms]) => ffffffff8122fa62 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fa8e do_close_on_exec ([kernel.kallsyms]) => ffffffff8122faf0 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122faf7 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fa8b do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fac9 do_close_on_exec ([kernel.kallsyms]) => ffffffff8122fad2 do_close_on_exec ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8122fadd do_close_on_exec ([kernel.kallsyms]) => ffffffff8120fc80 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcaf filp_close ([kernel.kallsyms]) => ffffffff8120fcb6 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcc2 filp_close ([kernel.kallsyms]) => ffffffff812547f0 dnotify_flush ([kernel.kallsyms])
ls 1843 1 branches: ffffffff81254823 dnotify_flush ([kernel.kallsyms]) => ffffffff8120fcc7 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fccd filp_close ([kernel.kallsyms]) => ffffffff81261790 locks_remove_posix ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812617a3 locks_remove_posix ([kernel.kallsyms]) => ffffffff812617b9 locks_remove_posix ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812617b9 locks_remove_posix ([kernel.kallsyms]) => ffffffff8120fcd2 filp_close ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8120fcd5 filp_close ([kernel.kallsyms]) => ffffffff812142c0 fput ([kernel.kallsyms])
ls 1843 1 branches: ffffffff812142d6 fput ([kernel.kallsyms]) => ffffffff812142df fput ([kernel.kallsyms])
ls 1843 1 branches: ffffffff8121430c fput ([kernel.kallsyms]) => ffffffff810b6580 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65ad task_work_add ([kernel.kallsyms]) => ffffffff810b65b1 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65c1 task_work_add ([kernel.kallsyms]) => ffffffff810bc710 kick_process ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810bc725 kick_process ([kernel.kallsyms]) => ffffffff810bc742 kick_process ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810bc742 kick_process ([kernel.kallsyms]) => ffffffff810b65c6 task_work_add ([kernel.kallsyms])
ls 1843 1 branches: ffffffff810b65c9 task_work_add ([kernel.kallsyms]) => ffffffff81214311 fput ([kernel.kallsyms])
Signed-off-by: Adrian Hunter <adrian.hunter@intel.com>
Tested-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Link: http://lkml.kernel.org/r/1437150840-31811-9-git-send-email-adrian.hunter@intel.com
[ Merged sample->time fix for bug found after first round of testing on slightly older kernel ]
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2015-07-18 00:33:43 +08:00
|
|
|
case PERF_AUXTRACE_INTEL_BTS:
|
|
|
|
return intel_bts_process_auxtrace_info(event, session);
|
2015-04-09 23:53:53 +08:00
|
|
|
case PERF_AUXTRACE_UNKNOWN:
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
s64 perf_event__process_auxtrace(struct perf_tool *tool,
|
|
|
|
union perf_event *event,
|
|
|
|
struct perf_session *session)
|
|
|
|
{
|
|
|
|
s64 err;
|
|
|
|
|
|
|
|
if (dump_trace)
|
|
|
|
fprintf(stdout, " size: %#"PRIx64" offset: %#"PRIx64" ref: %#"PRIx64" idx: %u tid: %d cpu: %d\n",
|
|
|
|
event->auxtrace.size, event->auxtrace.offset,
|
|
|
|
event->auxtrace.reference, event->auxtrace.idx,
|
|
|
|
event->auxtrace.tid, event->auxtrace.cpu);
|
|
|
|
|
|
|
|
if (auxtrace__dont_decode(session))
|
|
|
|
return event->auxtrace.size;
|
|
|
|
|
|
|
|
if (!session->auxtrace || event->header.type != PERF_RECORD_AUXTRACE)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
err = session->auxtrace->process_auxtrace_event(session, event, tool);
|
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
return event->auxtrace.size;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:49 +08:00
|
|
|
#define PERF_ITRACE_DEFAULT_PERIOD_TYPE PERF_ITRACE_PERIOD_NANOSECS
|
|
|
|
#define PERF_ITRACE_DEFAULT_PERIOD 100000
|
|
|
|
#define PERF_ITRACE_DEFAULT_CALLCHAIN_SZ 16
|
|
|
|
#define PERF_ITRACE_MAX_CALLCHAIN_SZ 1024
|
|
|
|
|
|
|
|
void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts)
|
|
|
|
{
|
|
|
|
synth_opts->instructions = true;
|
|
|
|
synth_opts->branches = true;
|
2015-04-30 22:37:28 +08:00
|
|
|
synth_opts->transactions = true;
|
2015-04-09 23:53:49 +08:00
|
|
|
synth_opts->errors = true;
|
|
|
|
synth_opts->period_type = PERF_ITRACE_DEFAULT_PERIOD_TYPE;
|
|
|
|
synth_opts->period = PERF_ITRACE_DEFAULT_PERIOD;
|
|
|
|
synth_opts->callchain_sz = PERF_ITRACE_DEFAULT_CALLCHAIN_SZ;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Please check tools/perf/Documentation/perf-script.txt for information
|
|
|
|
* about the options parsed here, which is introduced after this cset,
|
|
|
|
* when support in 'perf script' for these options is introduced.
|
|
|
|
*/
|
|
|
|
int itrace_parse_synth_opts(const struct option *opt, const char *str,
|
|
|
|
int unset)
|
|
|
|
{
|
|
|
|
struct itrace_synth_opts *synth_opts = opt->value;
|
|
|
|
const char *p;
|
|
|
|
char *endptr;
|
2015-07-18 00:33:46 +08:00
|
|
|
bool period_type_set = false;
|
2015-04-09 23:53:49 +08:00
|
|
|
|
|
|
|
synth_opts->set = true;
|
|
|
|
|
|
|
|
if (unset) {
|
|
|
|
synth_opts->dont_decode = true;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!str) {
|
|
|
|
itrace_synth_opts__set_default(synth_opts);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (p = str; *p;) {
|
|
|
|
switch (*p++) {
|
|
|
|
case 'i':
|
|
|
|
synth_opts->instructions = true;
|
|
|
|
while (*p == ' ' || *p == ',')
|
|
|
|
p += 1;
|
|
|
|
if (isdigit(*p)) {
|
|
|
|
synth_opts->period = strtoull(p, &endptr, 10);
|
|
|
|
p = endptr;
|
|
|
|
while (*p == ' ' || *p == ',')
|
|
|
|
p += 1;
|
|
|
|
switch (*p++) {
|
|
|
|
case 'i':
|
|
|
|
synth_opts->period_type =
|
|
|
|
PERF_ITRACE_PERIOD_INSTRUCTIONS;
|
2015-07-18 00:33:46 +08:00
|
|
|
period_type_set = true;
|
2015-04-09 23:53:49 +08:00
|
|
|
break;
|
|
|
|
case 't':
|
|
|
|
synth_opts->period_type =
|
|
|
|
PERF_ITRACE_PERIOD_TICKS;
|
2015-07-18 00:33:46 +08:00
|
|
|
period_type_set = true;
|
2015-04-09 23:53:49 +08:00
|
|
|
break;
|
|
|
|
case 'm':
|
|
|
|
synth_opts->period *= 1000;
|
|
|
|
/* Fall through */
|
|
|
|
case 'u':
|
|
|
|
synth_opts->period *= 1000;
|
|
|
|
/* Fall through */
|
|
|
|
case 'n':
|
|
|
|
if (*p++ != 's')
|
|
|
|
goto out_err;
|
|
|
|
synth_opts->period_type =
|
|
|
|
PERF_ITRACE_PERIOD_NANOSECS;
|
2015-07-18 00:33:46 +08:00
|
|
|
period_type_set = true;
|
2015-04-09 23:53:49 +08:00
|
|
|
break;
|
|
|
|
case '\0':
|
|
|
|
goto out;
|
|
|
|
default:
|
|
|
|
goto out_err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case 'b':
|
|
|
|
synth_opts->branches = true;
|
|
|
|
break;
|
2015-04-30 22:37:28 +08:00
|
|
|
case 'x':
|
|
|
|
synth_opts->transactions = true;
|
|
|
|
break;
|
2015-04-09 23:53:49 +08:00
|
|
|
case 'e':
|
|
|
|
synth_opts->errors = true;
|
|
|
|
break;
|
|
|
|
case 'd':
|
|
|
|
synth_opts->log = true;
|
|
|
|
break;
|
|
|
|
case 'c':
|
|
|
|
synth_opts->branches = true;
|
|
|
|
synth_opts->calls = true;
|
|
|
|
break;
|
|
|
|
case 'r':
|
|
|
|
synth_opts->branches = true;
|
|
|
|
synth_opts->returns = true;
|
|
|
|
break;
|
|
|
|
case 'g':
|
|
|
|
synth_opts->callchain = true;
|
|
|
|
synth_opts->callchain_sz =
|
|
|
|
PERF_ITRACE_DEFAULT_CALLCHAIN_SZ;
|
|
|
|
while (*p == ' ' || *p == ',')
|
|
|
|
p += 1;
|
|
|
|
if (isdigit(*p)) {
|
|
|
|
unsigned int val;
|
|
|
|
|
|
|
|
val = strtoul(p, &endptr, 10);
|
|
|
|
p = endptr;
|
|
|
|
if (!val || val > PERF_ITRACE_MAX_CALLCHAIN_SZ)
|
|
|
|
goto out_err;
|
|
|
|
synth_opts->callchain_sz = val;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case ' ':
|
|
|
|
case ',':
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
goto out_err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
if (synth_opts->instructions) {
|
2015-07-18 00:33:46 +08:00
|
|
|
if (!period_type_set)
|
2015-04-09 23:53:49 +08:00
|
|
|
synth_opts->period_type =
|
|
|
|
PERF_ITRACE_DEFAULT_PERIOD_TYPE;
|
|
|
|
if (!synth_opts->period)
|
|
|
|
synth_opts->period = PERF_ITRACE_DEFAULT_PERIOD;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_err:
|
|
|
|
pr_err("Bad Instruction Tracing options '%s'\n", str);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:50 +08:00
|
|
|
static const char * const auxtrace_error_type_name[] = {
|
|
|
|
[PERF_AUXTRACE_ERROR_ITRACE] = "instruction trace",
|
|
|
|
};
|
|
|
|
|
|
|
|
static const char *auxtrace_error_name(int type)
|
|
|
|
{
|
|
|
|
const char *error_type_name = NULL;
|
|
|
|
|
|
|
|
if (type < PERF_AUXTRACE_ERROR_MAX)
|
|
|
|
error_type_name = auxtrace_error_type_name[type];
|
|
|
|
if (!error_type_name)
|
|
|
|
error_type_name = "unknown AUX";
|
|
|
|
return error_type_name;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp)
|
|
|
|
{
|
|
|
|
struct auxtrace_error_event *e = &event->auxtrace_error;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = fprintf(fp, " %s error type %u",
|
|
|
|
auxtrace_error_name(e->type), e->type);
|
|
|
|
ret += fprintf(fp, " cpu %d pid %d tid %d ip %#"PRIx64" code %u: %s\n",
|
|
|
|
e->cpu, e->pid, e->tid, e->ip, e->code, e->msg);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
void perf_session__auxtrace_error_inc(struct perf_session *session,
|
|
|
|
union perf_event *event)
|
|
|
|
{
|
|
|
|
struct auxtrace_error_event *e = &event->auxtrace_error;
|
|
|
|
|
|
|
|
if (e->type < PERF_AUXTRACE_ERROR_MAX)
|
|
|
|
session->evlist->stats.nr_auxtrace_errors[e->type] += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
void events_stats__auxtrace_error_warn(const struct events_stats *stats)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < PERF_AUXTRACE_ERROR_MAX; i++) {
|
|
|
|
if (!stats->nr_auxtrace_errors[i])
|
|
|
|
continue;
|
|
|
|
ui__warning("%u %s errors\n",
|
|
|
|
stats->nr_auxtrace_errors[i],
|
|
|
|
auxtrace_error_name(i));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
int perf_event__process_auxtrace_error(struct perf_tool *tool __maybe_unused,
|
|
|
|
union perf_event *event,
|
2015-04-09 23:53:53 +08:00
|
|
|
struct perf_session *session)
|
2015-04-09 23:53:50 +08:00
|
|
|
{
|
2015-04-09 23:53:53 +08:00
|
|
|
if (auxtrace__dont_decode(session))
|
|
|
|
return 0;
|
|
|
|
|
2015-04-09 23:53:50 +08:00
|
|
|
perf_event__fprintf_auxtrace_error(event, stdout);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
static int __auxtrace_mmap__read(struct auxtrace_mmap *mm,
|
|
|
|
struct auxtrace_record *itr,
|
|
|
|
struct perf_tool *tool, process_auxtrace_t fn,
|
|
|
|
bool snapshot, size_t snapshot_size)
|
2015-04-09 23:53:44 +08:00
|
|
|
{
|
2015-04-30 22:37:31 +08:00
|
|
|
u64 head, old = mm->prev, offset, ref;
|
2015-04-09 23:53:44 +08:00
|
|
|
unsigned char *data = mm->base;
|
|
|
|
size_t size, head_off, old_off, len1, len2, padding;
|
|
|
|
union perf_event ev;
|
|
|
|
void *data1, *data2;
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
if (snapshot) {
|
|
|
|
head = auxtrace_mmap__read_snapshot_head(mm);
|
|
|
|
if (auxtrace_record__find_snapshot(itr, mm->idx, mm, data,
|
|
|
|
&head, &old))
|
|
|
|
return -1;
|
|
|
|
} else {
|
|
|
|
head = auxtrace_mmap__read_head(mm);
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
if (old == head)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
pr_debug3("auxtrace idx %d old %#"PRIx64" head %#"PRIx64" diff %#"PRIx64"\n",
|
|
|
|
mm->idx, old, head, head - old);
|
|
|
|
|
|
|
|
if (mm->mask) {
|
|
|
|
head_off = head & mm->mask;
|
|
|
|
old_off = old & mm->mask;
|
|
|
|
} else {
|
|
|
|
head_off = head % mm->len;
|
|
|
|
old_off = old % mm->len;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (head_off > old_off)
|
|
|
|
size = head_off - old_off;
|
|
|
|
else
|
|
|
|
size = mm->len - (old_off - head_off);
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
if (snapshot && size > snapshot_size)
|
|
|
|
size = snapshot_size;
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
ref = auxtrace_record__reference(itr);
|
|
|
|
|
|
|
|
if (head > old || size <= head || mm->mask) {
|
|
|
|
offset = head - size;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* When the buffer size is not a power of 2, 'head' wraps at the
|
|
|
|
* highest multiple of the buffer size, so we have to subtract
|
|
|
|
* the remainder here.
|
|
|
|
*/
|
|
|
|
u64 rem = (0ULL - mm->len) % mm->len;
|
|
|
|
|
|
|
|
offset = head - size - rem;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size > head_off) {
|
|
|
|
len1 = size - head_off;
|
|
|
|
data1 = &data[mm->len - len1];
|
|
|
|
len2 = head_off;
|
|
|
|
data2 = &data[0];
|
|
|
|
} else {
|
|
|
|
len1 = size;
|
|
|
|
data1 = &data[head_off - len1];
|
|
|
|
len2 = 0;
|
|
|
|
data2 = NULL;
|
|
|
|
}
|
|
|
|
|
2015-05-29 21:33:38 +08:00
|
|
|
if (itr->alignment) {
|
|
|
|
unsigned int unwanted = len1 % itr->alignment;
|
|
|
|
|
|
|
|
len1 -= unwanted;
|
|
|
|
size -= unwanted;
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:44 +08:00
|
|
|
/* padding must be written by fn() e.g. record__process_auxtrace() */
|
|
|
|
padding = size & 7;
|
|
|
|
if (padding)
|
|
|
|
padding = 8 - padding;
|
|
|
|
|
|
|
|
memset(&ev, 0, sizeof(ev));
|
|
|
|
ev.auxtrace.header.type = PERF_RECORD_AUXTRACE;
|
|
|
|
ev.auxtrace.header.size = sizeof(ev.auxtrace);
|
|
|
|
ev.auxtrace.size = size + padding;
|
|
|
|
ev.auxtrace.offset = offset;
|
|
|
|
ev.auxtrace.reference = ref;
|
|
|
|
ev.auxtrace.idx = mm->idx;
|
|
|
|
ev.auxtrace.tid = mm->tid;
|
|
|
|
ev.auxtrace.cpu = mm->cpu;
|
|
|
|
|
|
|
|
if (fn(tool, &ev, data1, len1, data2, len2))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
mm->prev = head;
|
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
if (!snapshot) {
|
|
|
|
auxtrace_mmap__write_tail(mm, head);
|
|
|
|
if (itr->read_finish) {
|
|
|
|
int err;
|
2015-04-09 23:53:44 +08:00
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
err = itr->read_finish(itr, mm->idx);
|
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
}
|
2015-04-09 23:53:44 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
2015-04-09 23:53:54 +08:00
|
|
|
|
2015-04-30 22:37:31 +08:00
|
|
|
int auxtrace_mmap__read(struct auxtrace_mmap *mm, struct auxtrace_record *itr,
|
|
|
|
struct perf_tool *tool, process_auxtrace_t fn)
|
|
|
|
{
|
|
|
|
return __auxtrace_mmap__read(mm, itr, tool, fn, false, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_mmap__read_snapshot(struct auxtrace_mmap *mm,
|
|
|
|
struct auxtrace_record *itr,
|
|
|
|
struct perf_tool *tool, process_auxtrace_t fn,
|
|
|
|
size_t snapshot_size)
|
|
|
|
{
|
|
|
|
return __auxtrace_mmap__read(mm, itr, tool, fn, true, snapshot_size);
|
|
|
|
}
|
|
|
|
|
2015-04-09 23:53:54 +08:00
|
|
|
/**
|
|
|
|
* struct auxtrace_cache - hash table to implement a cache
|
|
|
|
* @hashtable: the hashtable
|
|
|
|
* @sz: hashtable size (number of hlists)
|
|
|
|
* @entry_size: size of an entry
|
|
|
|
* @limit: limit the number of entries to this maximum, when reached the cache
|
|
|
|
* is dropped and caching begins again with an empty cache
|
|
|
|
* @cnt: current number of entries
|
|
|
|
* @bits: hashtable size (@sz = 2^@bits)
|
|
|
|
*/
|
|
|
|
struct auxtrace_cache {
|
|
|
|
struct hlist_head *hashtable;
|
|
|
|
size_t sz;
|
|
|
|
size_t entry_size;
|
|
|
|
size_t limit;
|
|
|
|
size_t cnt;
|
|
|
|
unsigned int bits;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
|
|
|
|
unsigned int limit_percent)
|
|
|
|
{
|
|
|
|
struct auxtrace_cache *c;
|
|
|
|
struct hlist_head *ht;
|
|
|
|
size_t sz, i;
|
|
|
|
|
|
|
|
c = zalloc(sizeof(struct auxtrace_cache));
|
|
|
|
if (!c)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
sz = 1UL << bits;
|
|
|
|
|
|
|
|
ht = calloc(sz, sizeof(struct hlist_head));
|
|
|
|
if (!ht)
|
|
|
|
goto out_free;
|
|
|
|
|
|
|
|
for (i = 0; i < sz; i++)
|
|
|
|
INIT_HLIST_HEAD(&ht[i]);
|
|
|
|
|
|
|
|
c->hashtable = ht;
|
|
|
|
c->sz = sz;
|
|
|
|
c->entry_size = entry_size;
|
|
|
|
c->limit = (c->sz * limit_percent) / 100;
|
|
|
|
c->bits = bits;
|
|
|
|
|
|
|
|
return c;
|
|
|
|
|
|
|
|
out_free:
|
|
|
|
free(c);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void auxtrace_cache__drop(struct auxtrace_cache *c)
|
|
|
|
{
|
|
|
|
struct auxtrace_cache_entry *entry;
|
|
|
|
struct hlist_node *tmp;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
if (!c)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (i = 0; i < c->sz; i++) {
|
|
|
|
hlist_for_each_entry_safe(entry, tmp, &c->hashtable[i], hash) {
|
|
|
|
hlist_del(&entry->hash);
|
|
|
|
auxtrace_cache__free_entry(c, entry);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c->cnt = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_cache__free(struct auxtrace_cache *c)
|
|
|
|
{
|
|
|
|
if (!c)
|
|
|
|
return;
|
|
|
|
|
|
|
|
auxtrace_cache__drop(c);
|
|
|
|
free(c->hashtable);
|
|
|
|
free(c);
|
|
|
|
}
|
|
|
|
|
|
|
|
void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c)
|
|
|
|
{
|
|
|
|
return malloc(c->entry_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
void auxtrace_cache__free_entry(struct auxtrace_cache *c __maybe_unused,
|
|
|
|
void *entry)
|
|
|
|
{
|
|
|
|
free(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
|
|
|
|
struct auxtrace_cache_entry *entry)
|
|
|
|
{
|
|
|
|
if (c->limit && ++c->cnt > c->limit)
|
|
|
|
auxtrace_cache__drop(c);
|
|
|
|
|
|
|
|
entry->key = key;
|
|
|
|
hlist_add_head(&entry->hash, &c->hashtable[hash_32(key, c->bits)]);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key)
|
|
|
|
{
|
|
|
|
struct auxtrace_cache_entry *entry;
|
|
|
|
struct hlist_head *hlist;
|
|
|
|
|
|
|
|
if (!c)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
hlist = &c->hashtable[hash_32(key, c->bits)];
|
|
|
|
hlist_for_each_entry(entry, hlist, hash) {
|
|
|
|
if (entry->key == key)
|
|
|
|
return entry;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|