2009-12-12 07:24:02 +08:00
|
|
|
#include <linux/kernel.h>
|
|
|
|
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <sys/types.h>
|
|
|
|
|
|
|
|
#include "session.h"
|
2009-12-15 00:22:59 +08:00
|
|
|
#include "sort.h"
|
2009-12-12 07:24:02 +08:00
|
|
|
#include "util.h"
|
|
|
|
|
|
|
|
static int perf_session__open(struct perf_session *self, bool force)
|
|
|
|
{
|
|
|
|
struct stat input_stat;
|
|
|
|
|
|
|
|
self->fd = open(self->filename, O_RDONLY);
|
|
|
|
if (self->fd < 0) {
|
|
|
|
pr_err("failed to open file: %s", self->filename);
|
|
|
|
if (!strcmp(self->filename, "perf.data"))
|
|
|
|
pr_err(" (try 'perf record' first)");
|
|
|
|
pr_err("\n");
|
|
|
|
return -errno;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (fstat(self->fd, &input_stat) < 0)
|
|
|
|
goto out_close;
|
|
|
|
|
|
|
|
if (!force && input_stat.st_uid && (input_stat.st_uid != geteuid())) {
|
|
|
|
pr_err("file %s not owned by current user or root\n",
|
|
|
|
self->filename);
|
|
|
|
goto out_close;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!input_stat.st_size) {
|
|
|
|
pr_info("zero-sized file (%s), nothing to do!\n",
|
|
|
|
self->filename);
|
|
|
|
goto out_close;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (perf_header__read(&self->header, self->fd) < 0) {
|
|
|
|
pr_err("incompatible file format");
|
|
|
|
goto out_close;
|
|
|
|
}
|
|
|
|
|
|
|
|
self->size = input_stat.st_size;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_close:
|
|
|
|
close(self->fd);
|
|
|
|
self->fd = -1;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
2009-12-16 06:04:39 +08:00
|
|
|
struct perf_session *perf_session__new(const char *filename, int mode, bool force)
|
2009-12-12 07:24:02 +08:00
|
|
|
{
|
2009-12-14 05:50:28 +08:00
|
|
|
size_t len = filename ? strlen(filename) + 1 : 0;
|
2009-12-12 07:24:02 +08:00
|
|
|
struct perf_session *self = zalloc(sizeof(*self) + len);
|
|
|
|
|
|
|
|
if (self == NULL)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if (perf_header__init(&self->header) < 0)
|
perf session: Move kmaps to perf_session
There is still some more work to do to disentangle map creation
from DSO loading, but this happens only for the kernel, and for
the early adopters of perf diff, where this disentanglement
matters most, we'll be testing different kernels, so no problem
here.
Further clarification: right now we create the kernel maps for
the various modules and discontiguous kernel text maps when
loading the DSO, we should do it as a two step process, first
creating the maps, for multiple mappings with the same DSO
store, then doing the dso load just once, for the first hit on
one of the maps sharing this DSO backing store.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1260741029-4430-6-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-12-14 05:50:29 +08:00
|
|
|
goto out_free;
|
2009-12-12 07:24:02 +08:00
|
|
|
|
|
|
|
memcpy(self->filename, filename, len);
|
2009-12-14 05:50:28 +08:00
|
|
|
self->threads = RB_ROOT;
|
|
|
|
self->last_match = NULL;
|
2009-12-14 05:50:27 +08:00
|
|
|
self->mmap_window = 32;
|
|
|
|
self->cwd = NULL;
|
|
|
|
self->cwdlen = 0;
|
2009-12-28 07:37:03 +08:00
|
|
|
self->unknown_events = 0;
|
perf session: Move kmaps to perf_session
There is still some more work to do to disentangle map creation
from DSO loading, but this happens only for the kernel, and for
the early adopters of perf diff, where this disentanglement
matters most, we'll be testing different kernels, so no problem
here.
Further clarification: right now we create the kernel maps for
the various modules and discontiguous kernel text maps when
loading the DSO, we should do it as a two step process, first
creating the maps, for multiple mappings with the same DSO
store, then doing the dso load just once, for the first hit on
one of the maps sharing this DSO backing store.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1260741029-4430-6-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-12-14 05:50:29 +08:00
|
|
|
map_groups__init(&self->kmaps);
|
2009-12-12 07:24:02 +08:00
|
|
|
|
perf session: Move kmaps to perf_session
There is still some more work to do to disentangle map creation
from DSO loading, but this happens only for the kernel, and for
the early adopters of perf diff, where this disentanglement
matters most, we'll be testing different kernels, so no problem
here.
Further clarification: right now we create the kernel maps for
the various modules and discontiguous kernel text maps when
loading the DSO, we should do it as a two step process, first
creating the maps, for multiple mappings with the same DSO
store, then doing the dso load just once, for the first hit on
one of the maps sharing this DSO backing store.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1260741029-4430-6-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-12-14 05:50:29 +08:00
|
|
|
if (mode == O_RDONLY && perf_session__open(self, force) < 0)
|
|
|
|
goto out_delete;
|
2009-12-28 07:37:02 +08:00
|
|
|
|
|
|
|
self->sample_type = perf_header__sample_type(&self->header);
|
2009-12-12 07:24:02 +08:00
|
|
|
out:
|
|
|
|
return self;
|
perf session: Move kmaps to perf_session
There is still some more work to do to disentangle map creation
from DSO loading, but this happens only for the kernel, and for
the early adopters of perf diff, where this disentanglement
matters most, we'll be testing different kernels, so no problem
here.
Further clarification: right now we create the kernel maps for
the various modules and discontiguous kernel text maps when
loading the DSO, we should do it as a two step process, first
creating the maps, for multiple mappings with the same DSO
store, then doing the dso load just once, for the first hit on
one of the maps sharing this DSO backing store.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1260741029-4430-6-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-12-14 05:50:29 +08:00
|
|
|
out_free:
|
2009-12-12 07:24:02 +08:00
|
|
|
free(self);
|
|
|
|
return NULL;
|
perf session: Move kmaps to perf_session
There is still some more work to do to disentangle map creation
from DSO loading, but this happens only for the kernel, and for
the early adopters of perf diff, where this disentanglement
matters most, we'll be testing different kernels, so no problem
here.
Further clarification: right now we create the kernel maps for
the various modules and discontiguous kernel text maps when
loading the DSO, we should do it as a two step process, first
creating the maps, for multiple mappings with the same DSO
store, then doing the dso load just once, for the first hit on
one of the maps sharing this DSO backing store.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1260741029-4430-6-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2009-12-14 05:50:29 +08:00
|
|
|
out_delete:
|
|
|
|
perf_session__delete(self);
|
|
|
|
return NULL;
|
2009-12-12 07:24:02 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void perf_session__delete(struct perf_session *self)
|
|
|
|
{
|
|
|
|
perf_header__exit(&self->header);
|
|
|
|
close(self->fd);
|
2009-12-14 05:50:27 +08:00
|
|
|
free(self->cwd);
|
2009-12-12 07:24:02 +08:00
|
|
|
free(self);
|
|
|
|
}
|
2009-12-15 00:22:59 +08:00
|
|
|
|
|
|
|
static bool symbol__match_parent_regex(struct symbol *sym)
|
|
|
|
{
|
|
|
|
if (sym->name && !regexec(&parent_regex, sym->name, 0, NULL, 0))
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct symbol **perf_session__resolve_callchain(struct perf_session *self,
|
|
|
|
struct thread *thread,
|
|
|
|
struct ip_callchain *chain,
|
|
|
|
struct symbol **parent)
|
|
|
|
{
|
|
|
|
u8 cpumode = PERF_RECORD_MISC_USER;
|
|
|
|
struct symbol **syms = NULL;
|
|
|
|
unsigned int i;
|
|
|
|
|
2009-12-16 06:04:42 +08:00
|
|
|
if (symbol_conf.use_callchain) {
|
2009-12-15 00:22:59 +08:00
|
|
|
syms = calloc(chain->nr, sizeof(*syms));
|
|
|
|
if (!syms) {
|
|
|
|
fprintf(stderr, "Can't allocate memory for symbols\n");
|
|
|
|
exit(-1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < chain->nr; i++) {
|
|
|
|
u64 ip = chain->ips[i];
|
|
|
|
struct addr_location al;
|
|
|
|
|
|
|
|
if (ip >= PERF_CONTEXT_MAX) {
|
|
|
|
switch (ip) {
|
|
|
|
case PERF_CONTEXT_HV:
|
|
|
|
cpumode = PERF_RECORD_MISC_HYPERVISOR; break;
|
|
|
|
case PERF_CONTEXT_KERNEL:
|
|
|
|
cpumode = PERF_RECORD_MISC_KERNEL; break;
|
|
|
|
case PERF_CONTEXT_USER:
|
|
|
|
cpumode = PERF_RECORD_MISC_USER; break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
thread__find_addr_location(thread, self, cpumode,
|
|
|
|
MAP__FUNCTION, ip, &al, NULL);
|
|
|
|
if (al.sym != NULL) {
|
|
|
|
if (sort__has_parent && !*parent &&
|
|
|
|
symbol__match_parent_regex(al.sym))
|
|
|
|
*parent = al.sym;
|
2009-12-16 06:04:42 +08:00
|
|
|
if (!symbol_conf.use_callchain)
|
2009-12-15 00:22:59 +08:00
|
|
|
break;
|
|
|
|
syms[i] = al.sym;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return syms;
|
|
|
|
}
|
2009-12-28 07:36:59 +08:00
|
|
|
|
|
|
|
static int process_event_stub(event_t *event __used,
|
|
|
|
struct perf_session *session __used)
|
|
|
|
{
|
|
|
|
dump_printf(": unhandled!\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void perf_event_ops__fill_defaults(struct perf_event_ops *handler)
|
|
|
|
{
|
2009-12-28 07:37:05 +08:00
|
|
|
if (handler->sample == NULL)
|
|
|
|
handler->sample = process_event_stub;
|
|
|
|
if (handler->mmap == NULL)
|
|
|
|
handler->mmap = process_event_stub;
|
|
|
|
if (handler->comm == NULL)
|
|
|
|
handler->comm = process_event_stub;
|
|
|
|
if (handler->fork == NULL)
|
|
|
|
handler->fork = process_event_stub;
|
|
|
|
if (handler->exit == NULL)
|
|
|
|
handler->exit = process_event_stub;
|
|
|
|
if (handler->lost == NULL)
|
|
|
|
handler->lost = process_event_stub;
|
|
|
|
if (handler->read == NULL)
|
|
|
|
handler->read = process_event_stub;
|
|
|
|
if (handler->throttle == NULL)
|
|
|
|
handler->throttle = process_event_stub;
|
|
|
|
if (handler->unthrottle == NULL)
|
|
|
|
handler->unthrottle = process_event_stub;
|
2009-12-28 07:36:59 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static const char *event__name[] = {
|
|
|
|
[0] = "TOTAL",
|
|
|
|
[PERF_RECORD_MMAP] = "MMAP",
|
|
|
|
[PERF_RECORD_LOST] = "LOST",
|
|
|
|
[PERF_RECORD_COMM] = "COMM",
|
|
|
|
[PERF_RECORD_EXIT] = "EXIT",
|
|
|
|
[PERF_RECORD_THROTTLE] = "THROTTLE",
|
|
|
|
[PERF_RECORD_UNTHROTTLE] = "UNTHROTTLE",
|
|
|
|
[PERF_RECORD_FORK] = "FORK",
|
|
|
|
[PERF_RECORD_READ] = "READ",
|
|
|
|
[PERF_RECORD_SAMPLE] = "SAMPLE",
|
|
|
|
};
|
|
|
|
|
|
|
|
unsigned long event__total[PERF_RECORD_MAX];
|
|
|
|
|
|
|
|
void event__print_totals(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
for (i = 0; i < PERF_RECORD_MAX; ++i)
|
|
|
|
pr_info("%10s events: %10ld\n",
|
|
|
|
event__name[i], event__total[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int perf_session__process_event(struct perf_session *self,
|
|
|
|
event_t *event,
|
|
|
|
struct perf_event_ops *ops,
|
|
|
|
unsigned long offset, unsigned long head)
|
|
|
|
{
|
|
|
|
trace_event(event);
|
|
|
|
|
|
|
|
if (event->header.type < PERF_RECORD_MAX) {
|
|
|
|
dump_printf("%p [%p]: PERF_RECORD_%s",
|
|
|
|
(void *)(offset + head),
|
|
|
|
(void *)(long)(event->header.size),
|
|
|
|
event__name[event->header.type]);
|
|
|
|
++event__total[0];
|
|
|
|
++event__total[event->header.type];
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (event->header.type) {
|
|
|
|
case PERF_RECORD_SAMPLE:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->sample(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_MMAP:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->mmap(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_COMM:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->comm(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_FORK:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->fork(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_EXIT:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->exit(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_LOST:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->lost(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_READ:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->read(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_THROTTLE:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->throttle(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
case PERF_RECORD_UNTHROTTLE:
|
2009-12-28 07:37:05 +08:00
|
|
|
return ops->unthrottle(event, self);
|
2009-12-28 07:36:59 +08:00
|
|
|
default:
|
2009-12-28 07:37:03 +08:00
|
|
|
self->unknown_events++;
|
2009-12-28 07:36:59 +08:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
int perf_header__read_build_ids(int input, u64 offset, u64 size)
|
|
|
|
{
|
|
|
|
struct build_id_event bev;
|
|
|
|
char filename[PATH_MAX];
|
|
|
|
u64 limit = offset + size;
|
|
|
|
int err = -1;
|
|
|
|
|
|
|
|
while (offset < limit) {
|
|
|
|
struct dso *dso;
|
|
|
|
ssize_t len;
|
2010-01-08 05:59:39 +08:00
|
|
|
struct list_head *head = &dsos__user;
|
2009-12-28 07:36:59 +08:00
|
|
|
|
|
|
|
if (read(input, &bev, sizeof(bev)) != sizeof(bev))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
len = bev.header.size - sizeof(bev);
|
|
|
|
if (read(input, filename, len) != len)
|
|
|
|
goto out;
|
|
|
|
|
2010-01-08 05:59:39 +08:00
|
|
|
if (bev.header.misc & PERF_RECORD_MISC_KERNEL)
|
|
|
|
head = &dsos__kernel;
|
|
|
|
|
|
|
|
dso = __dsos__findnew(head, filename);
|
perf tools: Encode kernel module mappings in perf.data
We were always looking at the running machine /proc/modules,
even when processing a perf.data file, which only makes sense
when we're doing 'perf record' and 'perf report' on the same
machine, and in close sucession, or if we don't use modules at
all, right Peter? ;-)
Now, at 'perf record' time we read /proc/modules, find the long
path for modules, and put them as PERF_MMAP events, just like we
did to encode the reloc reference symbol for vmlinux. Talking
about that now it is encoded in .pgoff, so that we can use
.{start,len} to store the address boundaries for the kernel so
that when we reconstruct the kmaps tree we can do lookups right
away, without having to fixup the end of the kernel maps like we
did in the past (and now only in perf record).
One more step in the 'perf archive' direction when we'll finally
be able to collect data in one machine and analyse in another.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1263396139-4798-1-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2010-01-13 23:22:17 +08:00
|
|
|
if (dso != NULL) {
|
2009-12-28 07:36:59 +08:00
|
|
|
dso__set_build_id(dso, &bev.build_id);
|
perf tools: Encode kernel module mappings in perf.data
We were always looking at the running machine /proc/modules,
even when processing a perf.data file, which only makes sense
when we're doing 'perf record' and 'perf report' on the same
machine, and in close sucession, or if we don't use modules at
all, right Peter? ;-)
Now, at 'perf record' time we read /proc/modules, find the long
path for modules, and put them as PERF_MMAP events, just like we
did to encode the reloc reference symbol for vmlinux. Talking
about that now it is encoded in .pgoff, so that we can use
.{start,len} to store the address boundaries for the kernel so
that when we reconstruct the kmaps tree we can do lookups right
away, without having to fixup the end of the kernel maps like we
did in the past (and now only in perf record).
One more step in the 'perf archive' direction when we'll finally
be able to collect data in one machine and analyse in another.
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frédéric Weisbecker <fweisbec@gmail.com>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Paul Mackerras <paulus@samba.org>
LKML-Reference: <1263396139-4798-1-git-send-email-acme@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2010-01-13 23:22:17 +08:00
|
|
|
if (head == &dsos__kernel && filename[0] == '[')
|
|
|
|
dso->kernel = 1;
|
|
|
|
}
|
2009-12-28 07:36:59 +08:00
|
|
|
|
|
|
|
offset += bev.header.size;
|
|
|
|
}
|
|
|
|
err = 0;
|
|
|
|
out:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct thread *perf_session__register_idle_thread(struct perf_session *self)
|
|
|
|
{
|
|
|
|
struct thread *thread = perf_session__findnew(self, 0);
|
|
|
|
|
|
|
|
if (thread == NULL || thread__set_comm(thread, "swapper")) {
|
|
|
|
pr_err("problem inserting idle task.\n");
|
|
|
|
thread = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return thread;
|
|
|
|
}
|
|
|
|
|
|
|
|
int perf_session__process_events(struct perf_session *self,
|
|
|
|
struct perf_event_ops *ops)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
unsigned long head, shift;
|
|
|
|
unsigned long offset = 0;
|
|
|
|
size_t page_size;
|
|
|
|
event_t *event;
|
|
|
|
uint32_t size;
|
|
|
|
char *buf;
|
|
|
|
|
|
|
|
if (perf_session__register_idle_thread(self) == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
perf_event_ops__fill_defaults(ops);
|
|
|
|
|
|
|
|
page_size = getpagesize();
|
|
|
|
|
|
|
|
head = self->header.data_offset;
|
|
|
|
|
2009-12-28 07:37:04 +08:00
|
|
|
if (!symbol_conf.full_paths) {
|
2009-12-28 07:36:59 +08:00
|
|
|
char bf[PATH_MAX];
|
|
|
|
|
|
|
|
if (getcwd(bf, sizeof(bf)) == NULL) {
|
|
|
|
err = -errno;
|
|
|
|
out_getcwd_err:
|
|
|
|
pr_err("failed to get the current directory\n");
|
|
|
|
goto out_err;
|
|
|
|
}
|
|
|
|
self->cwd = strdup(bf);
|
|
|
|
if (self->cwd == NULL) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto out_getcwd_err;
|
|
|
|
}
|
|
|
|
self->cwdlen = strlen(self->cwd);
|
|
|
|
}
|
|
|
|
|
|
|
|
shift = page_size * (head / page_size);
|
|
|
|
offset += shift;
|
|
|
|
head -= shift;
|
|
|
|
|
|
|
|
remap:
|
|
|
|
buf = mmap(NULL, page_size * self->mmap_window, PROT_READ,
|
|
|
|
MAP_SHARED, self->fd, offset);
|
|
|
|
if (buf == MAP_FAILED) {
|
|
|
|
pr_err("failed to mmap file\n");
|
|
|
|
err = -errno;
|
|
|
|
goto out_err;
|
|
|
|
}
|
|
|
|
|
|
|
|
more:
|
|
|
|
event = (event_t *)(buf + head);
|
|
|
|
|
|
|
|
size = event->header.size;
|
|
|
|
if (size == 0)
|
|
|
|
size = 8;
|
|
|
|
|
|
|
|
if (head + event->header.size >= page_size * self->mmap_window) {
|
|
|
|
int munmap_ret;
|
|
|
|
|
|
|
|
shift = page_size * (head / page_size);
|
|
|
|
|
|
|
|
munmap_ret = munmap(buf, page_size * self->mmap_window);
|
|
|
|
assert(munmap_ret == 0);
|
|
|
|
|
|
|
|
offset += shift;
|
|
|
|
head -= shift;
|
|
|
|
goto remap;
|
|
|
|
}
|
|
|
|
|
|
|
|
size = event->header.size;
|
|
|
|
|
|
|
|
dump_printf("\n%p [%p]: event: %d\n",
|
|
|
|
(void *)(offset + head),
|
|
|
|
(void *)(long)event->header.size,
|
|
|
|
event->header.type);
|
|
|
|
|
|
|
|
if (size == 0 ||
|
|
|
|
perf_session__process_event(self, event, ops, offset, head) < 0) {
|
|
|
|
dump_printf("%p [%p]: skipping unknown header type: %d\n",
|
|
|
|
(void *)(offset + head),
|
|
|
|
(void *)(long)(event->header.size),
|
|
|
|
event->header.type);
|
|
|
|
/*
|
|
|
|
* assume we lost track of the stream, check alignment, and
|
|
|
|
* increment a single u64 in the hope to catch on again 'soon'.
|
|
|
|
*/
|
|
|
|
if (unlikely(head & 7))
|
|
|
|
head &= ~7ULL;
|
|
|
|
|
|
|
|
size = 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
head += size;
|
|
|
|
|
|
|
|
if (offset + head >= self->header.data_offset + self->header.data_size)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
if (offset + head < self->size)
|
|
|
|
goto more;
|
|
|
|
done:
|
|
|
|
err = 0;
|
|
|
|
out_err:
|
|
|
|
return err;
|
|
|
|
}
|
2009-12-28 07:37:01 +08:00
|
|
|
|
2009-12-28 07:37:02 +08:00
|
|
|
bool perf_session__has_traces(struct perf_session *self, const char *msg)
|
2009-12-28 07:37:01 +08:00
|
|
|
{
|
|
|
|
if (!(self->sample_type & PERF_SAMPLE_RAW)) {
|
2009-12-28 07:37:02 +08:00
|
|
|
pr_err("No trace sample to read. Did you call 'perf %s'?\n", msg);
|
|
|
|
return false;
|
2009-12-28 07:37:01 +08:00
|
|
|
}
|
|
|
|
|
2009-12-28 07:37:02 +08:00
|
|
|
return true;
|
2009-12-28 07:37:01 +08:00
|
|
|
}
|
2010-01-06 02:50:31 +08:00
|
|
|
|
|
|
|
int perf_session__set_kallsyms_ref_reloc_sym(struct perf_session *self,
|
|
|
|
const char *symbol_name,
|
|
|
|
u64 addr)
|
|
|
|
{
|
|
|
|
char *bracket;
|
|
|
|
|
|
|
|
self->ref_reloc_sym.name = strdup(symbol_name);
|
|
|
|
if (self->ref_reloc_sym.name == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
bracket = strchr(self->ref_reloc_sym.name, ']');
|
|
|
|
if (bracket)
|
|
|
|
*bracket = '\0';
|
|
|
|
|
|
|
|
self->ref_reloc_sym.addr = addr;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static u64 map__reloc_map_ip(struct map *map, u64 ip)
|
|
|
|
{
|
|
|
|
return ip + (s64)map->pgoff;
|
|
|
|
}
|
|
|
|
|
|
|
|
static u64 map__reloc_unmap_ip(struct map *map, u64 ip)
|
|
|
|
{
|
|
|
|
return ip - (s64)map->pgoff;
|
|
|
|
}
|
|
|
|
|
|
|
|
void perf_session__reloc_vmlinux_maps(struct perf_session *self,
|
|
|
|
u64 unrelocated_addr)
|
|
|
|
{
|
|
|
|
enum map_type type;
|
|
|
|
s64 reloc = unrelocated_addr - self->ref_reloc_sym.addr;
|
|
|
|
|
|
|
|
if (!reloc)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (type = 0; type < MAP__NR_TYPES; ++type) {
|
|
|
|
struct map *map = self->vmlinux_maps[type];
|
|
|
|
|
|
|
|
map->map_ip = map__reloc_map_ip;
|
|
|
|
map->unmap_ip = map__reloc_unmap_ip;
|
|
|
|
map->pgoff = reloc;
|
|
|
|
}
|
|
|
|
}
|