mirror of https://gitee.com/openkylin/linux.git
This includes the final patch to clean up and fix the issue with the
design of tracepoints and how a user could register a tracepoint and have that tracepoint not be activated but no error was shown. The design was for an out of tree module but broke in tree users. The clean up was to remove the saving of the hash table of tracepoint names such that they can be enabled before they exist (enabling a module tracepoint before that module is loaded). This added more complexity than needed. The clean up was to remove that code and just enable tracepoints that exist or fail if they do not. This removed a lot of code as well as the complexity that it brought. As a side effect, instead of registering a tracepoint by its name, the tracepoint needs to be registered with the tracepoint descriptor. This removes having to duplicate the tracepoint names that are enabled. The second patch was added that simplified the way modules were searched for. This cleanup required changes that were in the 3.15 queue as well as some changes that were added late in the 3.14-rc cycle. This final change waited till the two were merged in upstream and then the change was added and full tests were run. Unfortunately, the test found some errors, but after it was already submitted to the for-next branch and not to be rebased. Sparse errors were detected by Fengguang Wu's bot tests, and my internal tests discovered that the anonymous union initialization triggered a bug in older gcc compilers. Luckily, there was a bugzilla for the gcc bug which gave a work around to the problem. The third and fourth patch handled the sparse error and the gcc bug respectively. A final patch was tagged along to fix a missing documentation for the README file. -----BEGIN PGP SIGNATURE----- Version: GnuPG v1 iQEcBAABAgAGBQJTR+pwAAoJEKQekfcNnQGuvfoH/A4XZu4/1h2ZuKhzGi6lrrWr +zHUQ+JmGiAYRziQFwr2t/gqJ2vmDfHJnbDjKi6Emx8JcxesHas6CQOWps4zEic0 dwYSQjvuGNGFIFt+7I0K1OxfVVdt2PQ2lVrB5WgYdbash5J4Bi+09QBv0RbUKheo 37dKSeN3pbsuQsR70OTVP8laG3dA9IbHW7PsKnxIEB5zeIUHUBME/QdPPj/CuJwk wxZjXC2dbc3rdRlQjTVtWV3ZkGgZJB0k+JxjvZTA0N6u8Hj8LiFPuNawzf7ceBHx gc++57+WuMW0f0X/ar5/+3UPGFQKMSvKmdxIQCnWXQz5seTYYKDEx7mTH22fxgg= =OgeQ -----END PGP SIGNATURE----- Merge tag 'trace-3.15-v2' of git://git.kernel.org/pub/scm/linux/kernel/git/rostedt/linux-trace Pull more tracing updates from Steven Rostedt: "This includes the final patch to clean up and fix the issue with the design of tracepoints and how a user could register a tracepoint and have that tracepoint not be activated but no error was shown. The design was for an out of tree module but broke in tree users. The clean up was to remove the saving of the hash table of tracepoint names such that they can be enabled before they exist (enabling a module tracepoint before that module is loaded). This added more complexity than needed. The clean up was to remove that code and just enable tracepoints that exist or fail if they do not. This removed a lot of code as well as the complexity that it brought. As a side effect, instead of registering a tracepoint by its name, the tracepoint needs to be registered with the tracepoint descriptor. This removes having to duplicate the tracepoint names that are enabled. The second patch was added that simplified the way modules were searched for. This cleanup required changes that were in the 3.15 queue as well as some changes that were added late in the 3.14-rc cycle. This final change waited till the two were merged in upstream and then the change was added and full tests were run. Unfortunately, the test found some errors, but after it was already submitted to the for-next branch and not to be rebased. Sparse errors were detected by Fengguang Wu's bot tests, and my internal tests discovered that the anonymous union initialization triggered a bug in older gcc compilers. Luckily, there was a bugzilla for the gcc bug which gave a work around to the problem. The third and fourth patch handled the sparse error and the gcc bug respectively. A final patch was tagged along to fix a missing documentation for the README file" * tag 'trace-3.15-v2' of git://git.kernel.org/pub/scm/linux/kernel/git/rostedt/linux-trace: tracing: Add missing function triggers dump and cpudump to README tracing: Fix anonymous unions in struct ftrace_event_call tracepoint: Fix sparse warnings in tracepoint.c tracepoint: Simplify tracepoint module search tracepoint: Use struct pointer instead of name hash for reg/unreg tracepoints
This commit is contained in:
commit
0a7418f5f5
|
@ -7,6 +7,7 @@
|
|||
#include <linux/percpu.h>
|
||||
#include <linux/hardirq.h>
|
||||
#include <linux/perf_event.h>
|
||||
#include <linux/tracepoint.h>
|
||||
|
||||
struct trace_array;
|
||||
struct trace_buffer;
|
||||
|
@ -232,6 +233,7 @@ enum {
|
|||
TRACE_EVENT_FL_IGNORE_ENABLE_BIT,
|
||||
TRACE_EVENT_FL_WAS_ENABLED_BIT,
|
||||
TRACE_EVENT_FL_USE_CALL_FILTER_BIT,
|
||||
TRACE_EVENT_FL_TRACEPOINT_BIT,
|
||||
};
|
||||
|
||||
/*
|
||||
|
@ -244,6 +246,7 @@ enum {
|
|||
* (used for module unloading, if a module event is enabled,
|
||||
* it is best to clear the buffers that used it).
|
||||
* USE_CALL_FILTER - For ftrace internal events, don't use file filter
|
||||
* TRACEPOINT - Event is a tracepoint
|
||||
*/
|
||||
enum {
|
||||
TRACE_EVENT_FL_FILTERED = (1 << TRACE_EVENT_FL_FILTERED_BIT),
|
||||
|
@ -252,12 +255,17 @@ enum {
|
|||
TRACE_EVENT_FL_IGNORE_ENABLE = (1 << TRACE_EVENT_FL_IGNORE_ENABLE_BIT),
|
||||
TRACE_EVENT_FL_WAS_ENABLED = (1 << TRACE_EVENT_FL_WAS_ENABLED_BIT),
|
||||
TRACE_EVENT_FL_USE_CALL_FILTER = (1 << TRACE_EVENT_FL_USE_CALL_FILTER_BIT),
|
||||
TRACE_EVENT_FL_TRACEPOINT = (1 << TRACE_EVENT_FL_TRACEPOINT_BIT),
|
||||
};
|
||||
|
||||
struct ftrace_event_call {
|
||||
struct list_head list;
|
||||
struct ftrace_event_class *class;
|
||||
char *name;
|
||||
union {
|
||||
char *name;
|
||||
/* Set TRACE_EVENT_FL_TRACEPOINT flag when using "tp" */
|
||||
struct tracepoint *tp;
|
||||
};
|
||||
struct trace_event event;
|
||||
const char *print_fmt;
|
||||
struct event_filter *filter;
|
||||
|
@ -271,6 +279,7 @@ struct ftrace_event_call {
|
|||
* bit 3: ftrace internal event (do not enable)
|
||||
* bit 4: Event was enabled by module
|
||||
* bit 5: use call filter rather than file filter
|
||||
* bit 6: Event is a tracepoint
|
||||
*/
|
||||
int flags; /* static flags of different events */
|
||||
|
||||
|
@ -283,6 +292,15 @@ struct ftrace_event_call {
|
|||
#endif
|
||||
};
|
||||
|
||||
static inline const char *
|
||||
ftrace_event_name(struct ftrace_event_call *call)
|
||||
{
|
||||
if (call->flags & TRACE_EVENT_FL_TRACEPOINT)
|
||||
return call->tp ? call->tp->name : NULL;
|
||||
else
|
||||
return call->name;
|
||||
}
|
||||
|
||||
struct trace_array;
|
||||
struct ftrace_subsystem_dir;
|
||||
|
||||
|
@ -353,7 +371,7 @@ struct ftrace_event_file {
|
|||
#define __TRACE_EVENT_FLAGS(name, value) \
|
||||
static int __init trace_init_flags_##name(void) \
|
||||
{ \
|
||||
event_##name.flags = value; \
|
||||
event_##name.flags |= value; \
|
||||
return 0; \
|
||||
} \
|
||||
early_initcall(trace_init_flags_##name);
|
||||
|
|
|
@ -119,8 +119,10 @@ extern struct trace_event_functions exit_syscall_print_funcs;
|
|||
static struct syscall_metadata __syscall_meta_##sname; \
|
||||
static struct ftrace_event_call __used \
|
||||
event_enter_##sname = { \
|
||||
.name = "sys_enter"#sname, \
|
||||
.class = &event_class_syscall_enter, \
|
||||
{ \
|
||||
.name = "sys_enter"#sname, \
|
||||
}, \
|
||||
.event.funcs = &enter_syscall_print_funcs, \
|
||||
.data = (void *)&__syscall_meta_##sname,\
|
||||
.flags = TRACE_EVENT_FL_CAP_ANY, \
|
||||
|
@ -133,8 +135,10 @@ extern struct trace_event_functions exit_syscall_print_funcs;
|
|||
static struct syscall_metadata __syscall_meta_##sname; \
|
||||
static struct ftrace_event_call __used \
|
||||
event_exit_##sname = { \
|
||||
.name = "sys_exit"#sname, \
|
||||
.class = &event_class_syscall_exit, \
|
||||
{ \
|
||||
.name = "sys_exit"#sname, \
|
||||
}, \
|
||||
.event.funcs = &exit_syscall_print_funcs, \
|
||||
.data = (void *)&__syscall_meta_##sname,\
|
||||
.flags = TRACE_EVENT_FL_CAP_ANY, \
|
||||
|
|
|
@ -6,7 +6,7 @@
|
|||
*
|
||||
* See Documentation/trace/tracepoints.txt.
|
||||
*
|
||||
* (C) Copyright 2008 Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
|
||||
* Copyright (C) 2008-2014 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
|
||||
*
|
||||
* Heavily inspired from the Linux Kernel Markers.
|
||||
*
|
||||
|
@ -21,6 +21,7 @@
|
|||
|
||||
struct module;
|
||||
struct tracepoint;
|
||||
struct notifier_block;
|
||||
|
||||
struct tracepoint_func {
|
||||
void *func;
|
||||
|
@ -35,31 +36,38 @@ struct tracepoint {
|
|||
struct tracepoint_func __rcu *funcs;
|
||||
};
|
||||
|
||||
/*
|
||||
* Connect a probe to a tracepoint.
|
||||
* Internal API, should not be used directly.
|
||||
*/
|
||||
extern int tracepoint_probe_register(const char *name, void *probe, void *data);
|
||||
|
||||
/*
|
||||
* Disconnect a probe from a tracepoint.
|
||||
* Internal API, should not be used directly.
|
||||
*/
|
||||
extern int
|
||||
tracepoint_probe_unregister(const char *name, void *probe, void *data);
|
||||
tracepoint_probe_register(struct tracepoint *tp, void *probe, void *data);
|
||||
extern int
|
||||
tracepoint_probe_unregister(struct tracepoint *tp, void *probe, void *data);
|
||||
extern void
|
||||
for_each_kernel_tracepoint(void (*fct)(struct tracepoint *tp, void *priv),
|
||||
void *priv);
|
||||
|
||||
#ifdef CONFIG_MODULES
|
||||
struct tp_module {
|
||||
struct list_head list;
|
||||
unsigned int num_tracepoints;
|
||||
struct tracepoint * const *tracepoints_ptrs;
|
||||
struct module *mod;
|
||||
};
|
||||
|
||||
bool trace_module_has_bad_taint(struct module *mod);
|
||||
extern int register_tracepoint_module_notifier(struct notifier_block *nb);
|
||||
extern int unregister_tracepoint_module_notifier(struct notifier_block *nb);
|
||||
#else
|
||||
static inline bool trace_module_has_bad_taint(struct module *mod)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
static inline
|
||||
int register_tracepoint_module_notifier(struct notifier_block *nb)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline
|
||||
int unregister_tracepoint_module_notifier(struct notifier_block *nb)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
#endif /* CONFIG_MODULES */
|
||||
|
||||
/*
|
||||
|
@ -72,6 +80,11 @@ static inline void tracepoint_synchronize_unregister(void)
|
|||
synchronize_sched();
|
||||
}
|
||||
|
||||
#ifdef CONFIG_HAVE_SYSCALL_TRACEPOINTS
|
||||
extern void syscall_regfunc(void);
|
||||
extern void syscall_unregfunc(void);
|
||||
#endif /* CONFIG_HAVE_SYSCALL_TRACEPOINTS */
|
||||
|
||||
#define PARAMS(args...) args
|
||||
|
||||
#endif /* _LINUX_TRACEPOINT_H */
|
||||
|
@ -160,14 +173,14 @@ static inline void tracepoint_synchronize_unregister(void)
|
|||
static inline int \
|
||||
register_trace_##name(void (*probe)(data_proto), void *data) \
|
||||
{ \
|
||||
return tracepoint_probe_register(#name, (void *)probe, \
|
||||
data); \
|
||||
return tracepoint_probe_register(&__tracepoint_##name, \
|
||||
(void *)probe, data); \
|
||||
} \
|
||||
static inline int \
|
||||
unregister_trace_##name(void (*probe)(data_proto), void *data) \
|
||||
{ \
|
||||
return tracepoint_probe_unregister(#name, (void *)probe, \
|
||||
data); \
|
||||
return tracepoint_probe_unregister(&__tracepoint_##name,\
|
||||
(void *)probe, data); \
|
||||
} \
|
||||
static inline void \
|
||||
check_trace_callback_type_##name(void (*cb)(data_proto)) \
|
||||
|
|
|
@ -13,9 +13,6 @@
|
|||
|
||||
#ifdef CONFIG_HAVE_SYSCALL_TRACEPOINTS
|
||||
|
||||
extern void syscall_regfunc(void);
|
||||
extern void syscall_unregfunc(void);
|
||||
|
||||
TRACE_EVENT_FN(sys_enter,
|
||||
|
||||
TP_PROTO(struct pt_regs *regs, long id),
|
||||
|
|
|
@ -470,10 +470,13 @@ static inline notrace int ftrace_get_offsets_##call( \
|
|||
* };
|
||||
*
|
||||
* static struct ftrace_event_call event_<call> = {
|
||||
* .name = "<call>",
|
||||
* .class = event_class_<template>,
|
||||
* {
|
||||
* .tp = &__tracepoint_<call>,
|
||||
* },
|
||||
* .event = &ftrace_event_type_<call>,
|
||||
* .print_fmt = print_fmt_<call>,
|
||||
* .flags = TRACE_EVENT_FL_TRACEPOINT,
|
||||
* };
|
||||
* // its only safe to use pointers when doing linker tricks to
|
||||
* // create an array.
|
||||
|
@ -605,10 +608,13 @@ static struct ftrace_event_class __used __refdata event_class_##call = { \
|
|||
#define DEFINE_EVENT(template, call, proto, args) \
|
||||
\
|
||||
static struct ftrace_event_call __used event_##call = { \
|
||||
.name = #call, \
|
||||
.class = &event_class_##template, \
|
||||
{ \
|
||||
.tp = &__tracepoint_##call, \
|
||||
}, \
|
||||
.event.funcs = &ftrace_event_type_funcs_##template, \
|
||||
.print_fmt = print_fmt_##template, \
|
||||
.flags = TRACE_EVENT_FL_TRACEPOINT, \
|
||||
}; \
|
||||
static struct ftrace_event_call __used \
|
||||
__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call
|
||||
|
@ -619,10 +625,13 @@ __attribute__((section("_ftrace_events"))) *__event_##call = &event_##call
|
|||
static const char print_fmt_##call[] = print; \
|
||||
\
|
||||
static struct ftrace_event_call __used event_##call = { \
|
||||
.name = #call, \
|
||||
.class = &event_class_##template, \
|
||||
{ \
|
||||
.tp = &__tracepoint_##call, \
|
||||
}, \
|
||||
.event.funcs = &ftrace_event_type_funcs_##call, \
|
||||
.print_fmt = print_fmt_##call, \
|
||||
.flags = TRACE_EVENT_FL_TRACEPOINT, \
|
||||
}; \
|
||||
static struct ftrace_event_call __used \
|
||||
__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call
|
||||
|
|
|
@ -3611,6 +3611,8 @@ static const char readme_msg[] =
|
|||
#ifdef CONFIG_TRACER_SNAPSHOT
|
||||
"\t\t snapshot\n"
|
||||
#endif
|
||||
"\t\t dump\n"
|
||||
"\t\t cpudump\n"
|
||||
"\t example: echo do_fault:traceoff > set_ftrace_filter\n"
|
||||
"\t echo do_trap:traceoff:3 > set_ftrace_filter\n"
|
||||
"\t The first one will disable tracing every time do_fault is hit\n"
|
||||
|
|
|
@ -223,24 +223,25 @@ int ftrace_event_reg(struct ftrace_event_call *call,
|
|||
{
|
||||
struct ftrace_event_file *file = data;
|
||||
|
||||
WARN_ON(!(call->flags & TRACE_EVENT_FL_TRACEPOINT));
|
||||
switch (type) {
|
||||
case TRACE_REG_REGISTER:
|
||||
return tracepoint_probe_register(call->name,
|
||||
return tracepoint_probe_register(call->tp,
|
||||
call->class->probe,
|
||||
file);
|
||||
case TRACE_REG_UNREGISTER:
|
||||
tracepoint_probe_unregister(call->name,
|
||||
tracepoint_probe_unregister(call->tp,
|
||||
call->class->probe,
|
||||
file);
|
||||
return 0;
|
||||
|
||||
#ifdef CONFIG_PERF_EVENTS
|
||||
case TRACE_REG_PERF_REGISTER:
|
||||
return tracepoint_probe_register(call->name,
|
||||
return tracepoint_probe_register(call->tp,
|
||||
call->class->perf_probe,
|
||||
call);
|
||||
case TRACE_REG_PERF_UNREGISTER:
|
||||
tracepoint_probe_unregister(call->name,
|
||||
tracepoint_probe_unregister(call->tp,
|
||||
call->class->perf_probe,
|
||||
call);
|
||||
return 0;
|
||||
|
@ -352,7 +353,7 @@ static int __ftrace_event_enable_disable(struct ftrace_event_file *file,
|
|||
if (ret) {
|
||||
tracing_stop_cmdline_record();
|
||||
pr_info("event trace: Could not enable event "
|
||||
"%s\n", call->name);
|
||||
"%s\n", ftrace_event_name(call));
|
||||
break;
|
||||
}
|
||||
set_bit(FTRACE_EVENT_FL_ENABLED_BIT, &file->flags);
|
||||
|
@ -481,27 +482,29 @@ __ftrace_set_clr_event_nolock(struct trace_array *tr, const char *match,
|
|||
{
|
||||
struct ftrace_event_file *file;
|
||||
struct ftrace_event_call *call;
|
||||
const char *name;
|
||||
int ret = -EINVAL;
|
||||
|
||||
list_for_each_entry(file, &tr->events, list) {
|
||||
|
||||
call = file->event_call;
|
||||
name = ftrace_event_name(call);
|
||||
|
||||
if (!call->name || !call->class || !call->class->reg)
|
||||
if (!name || !call->class || !call->class->reg)
|
||||
continue;
|
||||
|
||||
if (call->flags & TRACE_EVENT_FL_IGNORE_ENABLE)
|
||||
continue;
|
||||
|
||||
if (match &&
|
||||
strcmp(match, call->name) != 0 &&
|
||||
strcmp(match, name) != 0 &&
|
||||
strcmp(match, call->class->system) != 0)
|
||||
continue;
|
||||
|
||||
if (sub && strcmp(sub, call->class->system) != 0)
|
||||
continue;
|
||||
|
||||
if (event && strcmp(event, call->name) != 0)
|
||||
if (event && strcmp(event, name) != 0)
|
||||
continue;
|
||||
|
||||
ftrace_event_enable_disable(file, set);
|
||||
|
@ -699,7 +702,7 @@ static int t_show(struct seq_file *m, void *v)
|
|||
|
||||
if (strcmp(call->class->system, TRACE_SYSTEM) != 0)
|
||||
seq_printf(m, "%s:", call->class->system);
|
||||
seq_printf(m, "%s\n", call->name);
|
||||
seq_printf(m, "%s\n", ftrace_event_name(call));
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -792,7 +795,7 @@ system_enable_read(struct file *filp, char __user *ubuf, size_t cnt,
|
|||
mutex_lock(&event_mutex);
|
||||
list_for_each_entry(file, &tr->events, list) {
|
||||
call = file->event_call;
|
||||
if (!call->name || !call->class || !call->class->reg)
|
||||
if (!ftrace_event_name(call) || !call->class || !call->class->reg)
|
||||
continue;
|
||||
|
||||
if (system && strcmp(call->class->system, system->name) != 0)
|
||||
|
@ -907,7 +910,7 @@ static int f_show(struct seq_file *m, void *v)
|
|||
|
||||
switch ((unsigned long)v) {
|
||||
case FORMAT_HEADER:
|
||||
seq_printf(m, "name: %s\n", call->name);
|
||||
seq_printf(m, "name: %s\n", ftrace_event_name(call));
|
||||
seq_printf(m, "ID: %d\n", call->event.type);
|
||||
seq_printf(m, "format:\n");
|
||||
return 0;
|
||||
|
@ -1527,6 +1530,7 @@ event_create_dir(struct dentry *parent, struct ftrace_event_file *file)
|
|||
struct trace_array *tr = file->tr;
|
||||
struct list_head *head;
|
||||
struct dentry *d_events;
|
||||
const char *name;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
|
@ -1540,10 +1544,11 @@ event_create_dir(struct dentry *parent, struct ftrace_event_file *file)
|
|||
} else
|
||||
d_events = parent;
|
||||
|
||||
file->dir = debugfs_create_dir(call->name, d_events);
|
||||
name = ftrace_event_name(call);
|
||||
file->dir = debugfs_create_dir(name, d_events);
|
||||
if (!file->dir) {
|
||||
pr_warning("Could not create debugfs '%s' directory\n",
|
||||
call->name);
|
||||
name);
|
||||
return -1;
|
||||
}
|
||||
|
||||
|
@ -1567,7 +1572,7 @@ event_create_dir(struct dentry *parent, struct ftrace_event_file *file)
|
|||
ret = call->class->define_fields(call);
|
||||
if (ret < 0) {
|
||||
pr_warning("Could not initialize trace point"
|
||||
" events/%s\n", call->name);
|
||||
" events/%s\n", name);
|
||||
return -1;
|
||||
}
|
||||
}
|
||||
|
@ -1631,15 +1636,17 @@ static void event_remove(struct ftrace_event_call *call)
|
|||
static int event_init(struct ftrace_event_call *call)
|
||||
{
|
||||
int ret = 0;
|
||||
const char *name;
|
||||
|
||||
if (WARN_ON(!call->name))
|
||||
name = ftrace_event_name(call);
|
||||
if (WARN_ON(!name))
|
||||
return -EINVAL;
|
||||
|
||||
if (call->class->raw_init) {
|
||||
ret = call->class->raw_init(call);
|
||||
if (ret < 0 && ret != -ENOSYS)
|
||||
pr_warn("Could not initialize trace events/%s\n",
|
||||
call->name);
|
||||
name);
|
||||
}
|
||||
|
||||
return ret;
|
||||
|
@ -1885,7 +1892,7 @@ __trace_add_event_dirs(struct trace_array *tr)
|
|||
ret = __trace_add_new_event(call, tr);
|
||||
if (ret < 0)
|
||||
pr_warning("Could not create directory for event %s\n",
|
||||
call->name);
|
||||
ftrace_event_name(call));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1894,18 +1901,20 @@ find_event_file(struct trace_array *tr, const char *system, const char *event)
|
|||
{
|
||||
struct ftrace_event_file *file;
|
||||
struct ftrace_event_call *call;
|
||||
const char *name;
|
||||
|
||||
list_for_each_entry(file, &tr->events, list) {
|
||||
|
||||
call = file->event_call;
|
||||
name = ftrace_event_name(call);
|
||||
|
||||
if (!call->name || !call->class || !call->class->reg)
|
||||
if (!name || !call->class || !call->class->reg)
|
||||
continue;
|
||||
|
||||
if (call->flags & TRACE_EVENT_FL_IGNORE_ENABLE)
|
||||
continue;
|
||||
|
||||
if (strcmp(event, call->name) == 0 &&
|
||||
if (strcmp(event, name) == 0 &&
|
||||
strcmp(system, call->class->system) == 0)
|
||||
return file;
|
||||
}
|
||||
|
@ -1973,7 +1982,7 @@ event_enable_print(struct seq_file *m, unsigned long ip,
|
|||
seq_printf(m, "%s:%s:%s",
|
||||
data->enable ? ENABLE_EVENT_STR : DISABLE_EVENT_STR,
|
||||
data->file->event_call->class->system,
|
||||
data->file->event_call->name);
|
||||
ftrace_event_name(data->file->event_call));
|
||||
|
||||
if (data->count == -1)
|
||||
seq_printf(m, ":unlimited\n");
|
||||
|
@ -2193,7 +2202,7 @@ __trace_early_add_event_dirs(struct trace_array *tr)
|
|||
ret = event_create_dir(tr->event_dir, file);
|
||||
if (ret < 0)
|
||||
pr_warning("Could not create directory for event %s\n",
|
||||
file->event_call->name);
|
||||
ftrace_event_name(file->event_call));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -2217,7 +2226,7 @@ __trace_early_add_events(struct trace_array *tr)
|
|||
ret = __trace_early_add_new_event(call, tr);
|
||||
if (ret < 0)
|
||||
pr_warning("Could not create early event %s\n",
|
||||
call->name);
|
||||
ftrace_event_name(call));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -2549,7 +2558,7 @@ static __init void event_trace_self_tests(void)
|
|||
continue;
|
||||
#endif
|
||||
|
||||
pr_info("Testing event %s: ", call->name);
|
||||
pr_info("Testing event %s: ", ftrace_event_name(call));
|
||||
|
||||
/*
|
||||
* If an event is already enabled, someone is using
|
||||
|
|
|
@ -1095,7 +1095,7 @@ event_enable_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
|
|||
seq_printf(m, "%s:%s:%s",
|
||||
enable_data->enable ? ENABLE_EVENT_STR : DISABLE_EVENT_STR,
|
||||
enable_data->file->event_call->class->system,
|
||||
enable_data->file->event_call->name);
|
||||
ftrace_event_name(enable_data->file->event_call));
|
||||
|
||||
if (data->count == -1)
|
||||
seq_puts(m, ":unlimited");
|
||||
|
|
|
@ -173,9 +173,11 @@ struct ftrace_event_class __refdata event_class_ftrace_##call = { \
|
|||
}; \
|
||||
\
|
||||
struct ftrace_event_call __used event_##call = { \
|
||||
.name = #call, \
|
||||
.event.type = etype, \
|
||||
.class = &event_class_ftrace_##call, \
|
||||
{ \
|
||||
.name = #call, \
|
||||
}, \
|
||||
.event.type = etype, \
|
||||
.print_fmt = print, \
|
||||
.flags = TRACE_EVENT_FL_IGNORE_ENABLE | TRACE_EVENT_FL_USE_CALL_FILTER, \
|
||||
}; \
|
||||
|
|
|
@ -341,7 +341,7 @@ static struct trace_kprobe *find_trace_kprobe(const char *event,
|
|||
struct trace_kprobe *tk;
|
||||
|
||||
list_for_each_entry(tk, &probe_list, list)
|
||||
if (strcmp(tk->tp.call.name, event) == 0 &&
|
||||
if (strcmp(ftrace_event_name(&tk->tp.call), event) == 0 &&
|
||||
strcmp(tk->tp.call.class->system, group) == 0)
|
||||
return tk;
|
||||
return NULL;
|
||||
|
@ -516,7 +516,8 @@ static int register_trace_kprobe(struct trace_kprobe *tk)
|
|||
mutex_lock(&probe_lock);
|
||||
|
||||
/* Delete old (same name) event if exist */
|
||||
old_tk = find_trace_kprobe(tk->tp.call.name, tk->tp.call.class->system);
|
||||
old_tk = find_trace_kprobe(ftrace_event_name(&tk->tp.call),
|
||||
tk->tp.call.class->system);
|
||||
if (old_tk) {
|
||||
ret = unregister_trace_kprobe(old_tk);
|
||||
if (ret < 0)
|
||||
|
@ -564,7 +565,8 @@ static int trace_kprobe_module_callback(struct notifier_block *nb,
|
|||
if (ret)
|
||||
pr_warning("Failed to re-register probe %s on"
|
||||
"%s: %d\n",
|
||||
tk->tp.call.name, mod->name, ret);
|
||||
ftrace_event_name(&tk->tp.call),
|
||||
mod->name, ret);
|
||||
}
|
||||
}
|
||||
mutex_unlock(&probe_lock);
|
||||
|
@ -818,7 +820,8 @@ static int probes_seq_show(struct seq_file *m, void *v)
|
|||
int i;
|
||||
|
||||
seq_printf(m, "%c", trace_kprobe_is_return(tk) ? 'r' : 'p');
|
||||
seq_printf(m, ":%s/%s", tk->tp.call.class->system, tk->tp.call.name);
|
||||
seq_printf(m, ":%s/%s", tk->tp.call.class->system,
|
||||
ftrace_event_name(&tk->tp.call));
|
||||
|
||||
if (!tk->symbol)
|
||||
seq_printf(m, " 0x%p", tk->rp.kp.addr);
|
||||
|
@ -876,7 +879,8 @@ static int probes_profile_seq_show(struct seq_file *m, void *v)
|
|||
{
|
||||
struct trace_kprobe *tk = v;
|
||||
|
||||
seq_printf(m, " %-44s %15lu %15lu\n", tk->tp.call.name, tk->nhit,
|
||||
seq_printf(m, " %-44s %15lu %15lu\n",
|
||||
ftrace_event_name(&tk->tp.call), tk->nhit,
|
||||
tk->rp.kp.nmissed);
|
||||
|
||||
return 0;
|
||||
|
@ -1011,7 +1015,7 @@ print_kprobe_event(struct trace_iterator *iter, int flags,
|
|||
field = (struct kprobe_trace_entry_head *)iter->ent;
|
||||
tp = container_of(event, struct trace_probe, call.event);
|
||||
|
||||
if (!trace_seq_printf(s, "%s: (", tp->call.name))
|
||||
if (!trace_seq_printf(s, "%s: (", ftrace_event_name(&tp->call)))
|
||||
goto partial;
|
||||
|
||||
if (!seq_print_ip_sym(s, field->ip, flags | TRACE_ITER_SYM_OFFSET))
|
||||
|
@ -1047,7 +1051,7 @@ print_kretprobe_event(struct trace_iterator *iter, int flags,
|
|||
field = (struct kretprobe_trace_entry_head *)iter->ent;
|
||||
tp = container_of(event, struct trace_probe, call.event);
|
||||
|
||||
if (!trace_seq_printf(s, "%s: (", tp->call.name))
|
||||
if (!trace_seq_printf(s, "%s: (", ftrace_event_name(&tp->call)))
|
||||
goto partial;
|
||||
|
||||
if (!seq_print_ip_sym(s, field->ret_ip, flags | TRACE_ITER_SYM_OFFSET))
|
||||
|
@ -1286,7 +1290,8 @@ static int register_kprobe_event(struct trace_kprobe *tk)
|
|||
call->data = tk;
|
||||
ret = trace_add_event_call(call);
|
||||
if (ret) {
|
||||
pr_info("Failed to register kprobe event: %s\n", call->name);
|
||||
pr_info("Failed to register kprobe event: %s\n",
|
||||
ftrace_event_name(call));
|
||||
kfree(call->print_fmt);
|
||||
unregister_ftrace_event(&call->event);
|
||||
}
|
||||
|
|
|
@ -431,7 +431,7 @@ int ftrace_raw_output_prep(struct trace_iterator *iter,
|
|||
}
|
||||
|
||||
trace_seq_init(p);
|
||||
ret = trace_seq_printf(s, "%s: ", event->name);
|
||||
ret = trace_seq_printf(s, "%s: ", ftrace_event_name(event));
|
||||
if (!ret)
|
||||
return TRACE_TYPE_PARTIAL_LINE;
|
||||
|
||||
|
|
|
@ -294,7 +294,7 @@ static struct trace_uprobe *find_probe_event(const char *event, const char *grou
|
|||
struct trace_uprobe *tu;
|
||||
|
||||
list_for_each_entry(tu, &uprobe_list, list)
|
||||
if (strcmp(tu->tp.call.name, event) == 0 &&
|
||||
if (strcmp(ftrace_event_name(&tu->tp.call), event) == 0 &&
|
||||
strcmp(tu->tp.call.class->system, group) == 0)
|
||||
return tu;
|
||||
|
||||
|
@ -324,7 +324,8 @@ static int register_trace_uprobe(struct trace_uprobe *tu)
|
|||
mutex_lock(&uprobe_lock);
|
||||
|
||||
/* register as an event */
|
||||
old_tu = find_probe_event(tu->tp.call.name, tu->tp.call.class->system);
|
||||
old_tu = find_probe_event(ftrace_event_name(&tu->tp.call),
|
||||
tu->tp.call.class->system);
|
||||
if (old_tu) {
|
||||
/* delete old event */
|
||||
ret = unregister_trace_uprobe(old_tu);
|
||||
|
@ -599,7 +600,8 @@ static int probes_seq_show(struct seq_file *m, void *v)
|
|||
char c = is_ret_probe(tu) ? 'r' : 'p';
|
||||
int i;
|
||||
|
||||
seq_printf(m, "%c:%s/%s", c, tu->tp.call.class->system, tu->tp.call.name);
|
||||
seq_printf(m, "%c:%s/%s", c, tu->tp.call.class->system,
|
||||
ftrace_event_name(&tu->tp.call));
|
||||
seq_printf(m, " %s:0x%p", tu->filename, (void *)tu->offset);
|
||||
|
||||
for (i = 0; i < tu->tp.nr_args; i++)
|
||||
|
@ -649,7 +651,8 @@ static int probes_profile_seq_show(struct seq_file *m, void *v)
|
|||
{
|
||||
struct trace_uprobe *tu = v;
|
||||
|
||||
seq_printf(m, " %s %-44s %15lu\n", tu->filename, tu->tp.call.name, tu->nhit);
|
||||
seq_printf(m, " %s %-44s %15lu\n", tu->filename,
|
||||
ftrace_event_name(&tu->tp.call), tu->nhit);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -844,12 +847,14 @@ print_uprobe_event(struct trace_iterator *iter, int flags, struct trace_event *e
|
|||
tu = container_of(event, struct trace_uprobe, tp.call.event);
|
||||
|
||||
if (is_ret_probe(tu)) {
|
||||
if (!trace_seq_printf(s, "%s: (0x%lx <- 0x%lx)", tu->tp.call.name,
|
||||
if (!trace_seq_printf(s, "%s: (0x%lx <- 0x%lx)",
|
||||
ftrace_event_name(&tu->tp.call),
|
||||
entry->vaddr[1], entry->vaddr[0]))
|
||||
goto partial;
|
||||
data = DATAOF_TRACE_ENTRY(entry, true);
|
||||
} else {
|
||||
if (!trace_seq_printf(s, "%s: (0x%lx)", tu->tp.call.name,
|
||||
if (!trace_seq_printf(s, "%s: (0x%lx)",
|
||||
ftrace_event_name(&tu->tp.call),
|
||||
entry->vaddr[0]))
|
||||
goto partial;
|
||||
data = DATAOF_TRACE_ENTRY(entry, false);
|
||||
|
@ -1275,7 +1280,8 @@ static int register_uprobe_event(struct trace_uprobe *tu)
|
|||
ret = trace_add_event_call(call);
|
||||
|
||||
if (ret) {
|
||||
pr_info("Failed to register uprobe event: %s\n", call->name);
|
||||
pr_info("Failed to register uprobe event: %s\n",
|
||||
ftrace_event_name(call));
|
||||
kfree(call->print_fmt);
|
||||
unregister_ftrace_event(&call->event);
|
||||
}
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
/*
|
||||
* Copyright (C) 2008 Mathieu Desnoyers
|
||||
* Copyright (C) 2008-2014 Mathieu Desnoyers
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify
|
||||
* it under the terms of the GNU General Public License as published by
|
||||
|
@ -33,39 +33,27 @@ extern struct tracepoint * const __stop___tracepoints_ptrs[];
|
|||
/* Set to 1 to enable tracepoint debug output */
|
||||
static const int tracepoint_debug;
|
||||
|
||||
/*
|
||||
* Tracepoints mutex protects the builtin and module tracepoints and the hash
|
||||
* table, as well as the local module list.
|
||||
*/
|
||||
static DEFINE_MUTEX(tracepoints_mutex);
|
||||
|
||||
#ifdef CONFIG_MODULES
|
||||
/* Local list of struct module */
|
||||
/*
|
||||
* Tracepoint module list mutex protects the local module list.
|
||||
*/
|
||||
static DEFINE_MUTEX(tracepoint_module_list_mutex);
|
||||
|
||||
/* Local list of struct tp_module */
|
||||
static LIST_HEAD(tracepoint_module_list);
|
||||
#endif /* CONFIG_MODULES */
|
||||
|
||||
/*
|
||||
* Tracepoint hash table, containing the active tracepoints.
|
||||
* Protected by tracepoints_mutex.
|
||||
* tracepoints_mutex protects the builtin and module tracepoints.
|
||||
* tracepoints_mutex nests inside tracepoint_module_list_mutex.
|
||||
*/
|
||||
#define TRACEPOINT_HASH_BITS 6
|
||||
#define TRACEPOINT_TABLE_SIZE (1 << TRACEPOINT_HASH_BITS)
|
||||
static struct hlist_head tracepoint_table[TRACEPOINT_TABLE_SIZE];
|
||||
static DEFINE_MUTEX(tracepoints_mutex);
|
||||
|
||||
/*
|
||||
* Note about RCU :
|
||||
* It is used to delay the free of multiple probes array until a quiescent
|
||||
* state is reached.
|
||||
* Tracepoint entries modifications are protected by the tracepoints_mutex.
|
||||
*/
|
||||
struct tracepoint_entry {
|
||||
struct hlist_node hlist;
|
||||
struct tracepoint_func *funcs;
|
||||
int refcount; /* Number of times armed. 0 if disarmed. */
|
||||
int enabled; /* Tracepoint enabled */
|
||||
char name[0];
|
||||
};
|
||||
|
||||
struct tp_probes {
|
||||
struct rcu_head rcu;
|
||||
struct tracepoint_func probes[0];
|
||||
|
@ -92,34 +80,33 @@ static inline void release_probes(struct tracepoint_func *old)
|
|||
}
|
||||
}
|
||||
|
||||
static void debug_print_probes(struct tracepoint_entry *entry)
|
||||
static void debug_print_probes(struct tracepoint_func *funcs)
|
||||
{
|
||||
int i;
|
||||
|
||||
if (!tracepoint_debug || !entry->funcs)
|
||||
if (!tracepoint_debug || !funcs)
|
||||
return;
|
||||
|
||||
for (i = 0; entry->funcs[i].func; i++)
|
||||
printk(KERN_DEBUG "Probe %d : %p\n", i, entry->funcs[i].func);
|
||||
for (i = 0; funcs[i].func; i++)
|
||||
printk(KERN_DEBUG "Probe %d : %p\n", i, funcs[i].func);
|
||||
}
|
||||
|
||||
static struct tracepoint_func *
|
||||
tracepoint_entry_add_probe(struct tracepoint_entry *entry,
|
||||
void *probe, void *data)
|
||||
static struct tracepoint_func *func_add(struct tracepoint_func **funcs,
|
||||
struct tracepoint_func *tp_func)
|
||||
{
|
||||
int nr_probes = 0;
|
||||
struct tracepoint_func *old, *new;
|
||||
|
||||
if (WARN_ON(!probe))
|
||||
if (WARN_ON(!tp_func->func))
|
||||
return ERR_PTR(-EINVAL);
|
||||
|
||||
debug_print_probes(entry);
|
||||
old = entry->funcs;
|
||||
debug_print_probes(*funcs);
|
||||
old = *funcs;
|
||||
if (old) {
|
||||
/* (N -> N+1), (N != 0, 1) probes */
|
||||
for (nr_probes = 0; old[nr_probes].func; nr_probes++)
|
||||
if (old[nr_probes].func == probe &&
|
||||
old[nr_probes].data == data)
|
||||
if (old[nr_probes].func == tp_func->func &&
|
||||
old[nr_probes].data == tp_func->data)
|
||||
return ERR_PTR(-EEXIST);
|
||||
}
|
||||
/* + 2 : one for new probe, one for NULL func */
|
||||
|
@ -128,33 +115,30 @@ tracepoint_entry_add_probe(struct tracepoint_entry *entry,
|
|||
return ERR_PTR(-ENOMEM);
|
||||
if (old)
|
||||
memcpy(new, old, nr_probes * sizeof(struct tracepoint_func));
|
||||
new[nr_probes].func = probe;
|
||||
new[nr_probes].data = data;
|
||||
new[nr_probes] = *tp_func;
|
||||
new[nr_probes + 1].func = NULL;
|
||||
entry->refcount = nr_probes + 1;
|
||||
entry->funcs = new;
|
||||
debug_print_probes(entry);
|
||||
*funcs = new;
|
||||
debug_print_probes(*funcs);
|
||||
return old;
|
||||
}
|
||||
|
||||
static void *
|
||||
tracepoint_entry_remove_probe(struct tracepoint_entry *entry,
|
||||
void *probe, void *data)
|
||||
static void *func_remove(struct tracepoint_func **funcs,
|
||||
struct tracepoint_func *tp_func)
|
||||
{
|
||||
int nr_probes = 0, nr_del = 0, i;
|
||||
struct tracepoint_func *old, *new;
|
||||
|
||||
old = entry->funcs;
|
||||
old = *funcs;
|
||||
|
||||
if (!old)
|
||||
return ERR_PTR(-ENOENT);
|
||||
|
||||
debug_print_probes(entry);
|
||||
debug_print_probes(*funcs);
|
||||
/* (N -> M), (N > 1, M >= 0) probes */
|
||||
if (probe) {
|
||||
if (tp_func->func) {
|
||||
for (nr_probes = 0; old[nr_probes].func; nr_probes++) {
|
||||
if (old[nr_probes].func == probe &&
|
||||
old[nr_probes].data == data)
|
||||
if (old[nr_probes].func == tp_func->func &&
|
||||
old[nr_probes].data == tp_func->data)
|
||||
nr_del++;
|
||||
}
|
||||
}
|
||||
|
@ -165,9 +149,8 @@ tracepoint_entry_remove_probe(struct tracepoint_entry *entry,
|
|||
*/
|
||||
if (nr_probes - nr_del == 0) {
|
||||
/* N -> 0, (N > 1) */
|
||||
entry->funcs = NULL;
|
||||
entry->refcount = 0;
|
||||
debug_print_probes(entry);
|
||||
*funcs = NULL;
|
||||
debug_print_probes(*funcs);
|
||||
return old;
|
||||
} else {
|
||||
int j = 0;
|
||||
|
@ -177,91 +160,35 @@ tracepoint_entry_remove_probe(struct tracepoint_entry *entry,
|
|||
if (new == NULL)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
for (i = 0; old[i].func; i++)
|
||||
if (old[i].func != probe || old[i].data != data)
|
||||
if (old[i].func != tp_func->func
|
||||
|| old[i].data != tp_func->data)
|
||||
new[j++] = old[i];
|
||||
new[nr_probes - nr_del].func = NULL;
|
||||
entry->refcount = nr_probes - nr_del;
|
||||
entry->funcs = new;
|
||||
*funcs = new;
|
||||
}
|
||||
debug_print_probes(entry);
|
||||
debug_print_probes(*funcs);
|
||||
return old;
|
||||
}
|
||||
|
||||
/*
|
||||
* Get tracepoint if the tracepoint is present in the tracepoint hash table.
|
||||
* Must be called with tracepoints_mutex held.
|
||||
* Returns NULL if not present.
|
||||
* Add the probe function to a tracepoint.
|
||||
*/
|
||||
static struct tracepoint_entry *get_tracepoint(const char *name)
|
||||
static int tracepoint_add_func(struct tracepoint *tp,
|
||||
struct tracepoint_func *func)
|
||||
{
|
||||
struct hlist_head *head;
|
||||
struct tracepoint_entry *e;
|
||||
u32 hash = jhash(name, strlen(name), 0);
|
||||
struct tracepoint_func *old, *tp_funcs;
|
||||
|
||||
head = &tracepoint_table[hash & (TRACEPOINT_TABLE_SIZE - 1)];
|
||||
hlist_for_each_entry(e, head, hlist) {
|
||||
if (!strcmp(name, e->name))
|
||||
return e;
|
||||
if (tp->regfunc && !static_key_enabled(&tp->key))
|
||||
tp->regfunc();
|
||||
|
||||
tp_funcs = rcu_dereference_protected(tp->funcs,
|
||||
lockdep_is_held(&tracepoints_mutex));
|
||||
old = func_add(&tp_funcs, func);
|
||||
if (IS_ERR(old)) {
|
||||
WARN_ON_ONCE(1);
|
||||
return PTR_ERR(old);
|
||||
}
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* Add the tracepoint to the tracepoint hash table. Must be called with
|
||||
* tracepoints_mutex held.
|
||||
*/
|
||||
static struct tracepoint_entry *add_tracepoint(const char *name)
|
||||
{
|
||||
struct hlist_head *head;
|
||||
struct tracepoint_entry *e;
|
||||
size_t name_len = strlen(name) + 1;
|
||||
u32 hash = jhash(name, name_len-1, 0);
|
||||
|
||||
head = &tracepoint_table[hash & (TRACEPOINT_TABLE_SIZE - 1)];
|
||||
hlist_for_each_entry(e, head, hlist) {
|
||||
if (!strcmp(name, e->name)) {
|
||||
printk(KERN_NOTICE
|
||||
"tracepoint %s busy\n", name);
|
||||
return ERR_PTR(-EEXIST); /* Already there */
|
||||
}
|
||||
}
|
||||
/*
|
||||
* Using kmalloc here to allocate a variable length element. Could
|
||||
* cause some memory fragmentation if overused.
|
||||
*/
|
||||
e = kmalloc(sizeof(struct tracepoint_entry) + name_len, GFP_KERNEL);
|
||||
if (!e)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
memcpy(&e->name[0], name, name_len);
|
||||
e->funcs = NULL;
|
||||
e->refcount = 0;
|
||||
e->enabled = 0;
|
||||
hlist_add_head(&e->hlist, head);
|
||||
return e;
|
||||
}
|
||||
|
||||
/*
|
||||
* Remove the tracepoint from the tracepoint hash table. Must be called with
|
||||
* mutex_lock held.
|
||||
*/
|
||||
static inline void remove_tracepoint(struct tracepoint_entry *e)
|
||||
{
|
||||
hlist_del(&e->hlist);
|
||||
kfree(e);
|
||||
}
|
||||
|
||||
/*
|
||||
* Sets the probe callback corresponding to one tracepoint.
|
||||
*/
|
||||
static void set_tracepoint(struct tracepoint_entry **entry,
|
||||
struct tracepoint *elem, int active)
|
||||
{
|
||||
WARN_ON(strcmp((*entry)->name, elem->name) != 0);
|
||||
|
||||
if (elem->regfunc && !static_key_enabled(&elem->key) && active)
|
||||
elem->regfunc();
|
||||
else if (elem->unregfunc && static_key_enabled(&elem->key) && !active)
|
||||
elem->unregfunc();
|
||||
release_probes(old);
|
||||
|
||||
/*
|
||||
* rcu_assign_pointer has a smp_wmb() which makes sure that the new
|
||||
|
@ -270,193 +197,90 @@ static void set_tracepoint(struct tracepoint_entry **entry,
|
|||
* include/linux/tracepoints.h. A matching smp_read_barrier_depends()
|
||||
* is used.
|
||||
*/
|
||||
rcu_assign_pointer(elem->funcs, (*entry)->funcs);
|
||||
if (active && !static_key_enabled(&elem->key))
|
||||
static_key_slow_inc(&elem->key);
|
||||
else if (!active && static_key_enabled(&elem->key))
|
||||
static_key_slow_dec(&elem->key);
|
||||
rcu_assign_pointer(tp->funcs, tp_funcs);
|
||||
if (!static_key_enabled(&tp->key))
|
||||
static_key_slow_inc(&tp->key);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Disable a tracepoint and its probe callback.
|
||||
* Remove a probe function from a tracepoint.
|
||||
* Note: only waiting an RCU period after setting elem->call to the empty
|
||||
* function insures that the original callback is not used anymore. This insured
|
||||
* by preempt_disable around the call site.
|
||||
*/
|
||||
static void disable_tracepoint(struct tracepoint *elem)
|
||||
static int tracepoint_remove_func(struct tracepoint *tp,
|
||||
struct tracepoint_func *func)
|
||||
{
|
||||
if (elem->unregfunc && static_key_enabled(&elem->key))
|
||||
elem->unregfunc();
|
||||
struct tracepoint_func *old, *tp_funcs;
|
||||
|
||||
if (static_key_enabled(&elem->key))
|
||||
static_key_slow_dec(&elem->key);
|
||||
rcu_assign_pointer(elem->funcs, NULL);
|
||||
}
|
||||
|
||||
/**
|
||||
* tracepoint_update_probe_range - Update a probe range
|
||||
* @begin: beginning of the range
|
||||
* @end: end of the range
|
||||
*
|
||||
* Updates the probe callback corresponding to a range of tracepoints.
|
||||
* Called with tracepoints_mutex held.
|
||||
*/
|
||||
static void tracepoint_update_probe_range(struct tracepoint * const *begin,
|
||||
struct tracepoint * const *end)
|
||||
{
|
||||
struct tracepoint * const *iter;
|
||||
struct tracepoint_entry *mark_entry;
|
||||
|
||||
if (!begin)
|
||||
return;
|
||||
|
||||
for (iter = begin; iter < end; iter++) {
|
||||
mark_entry = get_tracepoint((*iter)->name);
|
||||
if (mark_entry) {
|
||||
set_tracepoint(&mark_entry, *iter,
|
||||
!!mark_entry->refcount);
|
||||
mark_entry->enabled = !!mark_entry->refcount;
|
||||
} else {
|
||||
disable_tracepoint(*iter);
|
||||
}
|
||||
tp_funcs = rcu_dereference_protected(tp->funcs,
|
||||
lockdep_is_held(&tracepoints_mutex));
|
||||
old = func_remove(&tp_funcs, func);
|
||||
if (IS_ERR(old)) {
|
||||
WARN_ON_ONCE(1);
|
||||
return PTR_ERR(old);
|
||||
}
|
||||
}
|
||||
release_probes(old);
|
||||
|
||||
#ifdef CONFIG_MODULES
|
||||
void module_update_tracepoints(void)
|
||||
{
|
||||
struct tp_module *tp_mod;
|
||||
if (!tp_funcs) {
|
||||
/* Removed last function */
|
||||
if (tp->unregfunc && static_key_enabled(&tp->key))
|
||||
tp->unregfunc();
|
||||
|
||||
list_for_each_entry(tp_mod, &tracepoint_module_list, list)
|
||||
tracepoint_update_probe_range(tp_mod->tracepoints_ptrs,
|
||||
tp_mod->tracepoints_ptrs + tp_mod->num_tracepoints);
|
||||
}
|
||||
#else /* CONFIG_MODULES */
|
||||
void module_update_tracepoints(void)
|
||||
{
|
||||
}
|
||||
#endif /* CONFIG_MODULES */
|
||||
|
||||
|
||||
/*
|
||||
* Update probes, removing the faulty probes.
|
||||
* Called with tracepoints_mutex held.
|
||||
*/
|
||||
static void tracepoint_update_probes(void)
|
||||
{
|
||||
/* Core kernel tracepoints */
|
||||
tracepoint_update_probe_range(__start___tracepoints_ptrs,
|
||||
__stop___tracepoints_ptrs);
|
||||
/* tracepoints in modules. */
|
||||
module_update_tracepoints();
|
||||
}
|
||||
|
||||
static struct tracepoint_func *
|
||||
tracepoint_add_probe(const char *name, void *probe, void *data)
|
||||
{
|
||||
struct tracepoint_entry *entry;
|
||||
struct tracepoint_func *old;
|
||||
|
||||
entry = get_tracepoint(name);
|
||||
if (!entry) {
|
||||
entry = add_tracepoint(name);
|
||||
if (IS_ERR(entry))
|
||||
return (struct tracepoint_func *)entry;
|
||||
if (static_key_enabled(&tp->key))
|
||||
static_key_slow_dec(&tp->key);
|
||||
}
|
||||
old = tracepoint_entry_add_probe(entry, probe, data);
|
||||
if (IS_ERR(old) && !entry->refcount)
|
||||
remove_tracepoint(entry);
|
||||
return old;
|
||||
rcu_assign_pointer(tp->funcs, tp_funcs);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* tracepoint_probe_register - Connect a probe to a tracepoint
|
||||
* @name: tracepoint name
|
||||
* @tp: tracepoint
|
||||
* @probe: probe handler
|
||||
* @data: probe private data
|
||||
*
|
||||
* Returns:
|
||||
* - 0 if the probe was successfully registered, and tracepoint
|
||||
* callsites are currently loaded for that probe,
|
||||
* - -ENODEV if the probe was successfully registered, but no tracepoint
|
||||
* callsite is currently loaded for that probe,
|
||||
* - other negative error value on error.
|
||||
*
|
||||
* When tracepoint_probe_register() returns either 0 or -ENODEV,
|
||||
* parameters @name, @probe, and @data may be used by the tracepoint
|
||||
* infrastructure until the probe is unregistered.
|
||||
*
|
||||
* The probe address must at least be aligned on the architecture pointer size.
|
||||
* Returns 0 if ok, error value on error.
|
||||
* Note: if @tp is within a module, the caller is responsible for
|
||||
* unregistering the probe before the module is gone. This can be
|
||||
* performed either with a tracepoint module going notifier, or from
|
||||
* within module exit functions.
|
||||
*/
|
||||
int tracepoint_probe_register(const char *name, void *probe, void *data)
|
||||
int tracepoint_probe_register(struct tracepoint *tp, void *probe, void *data)
|
||||
{
|
||||
struct tracepoint_func *old;
|
||||
struct tracepoint_entry *entry;
|
||||
int ret = 0;
|
||||
struct tracepoint_func tp_func;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&tracepoints_mutex);
|
||||
old = tracepoint_add_probe(name, probe, data);
|
||||
if (IS_ERR(old)) {
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
return PTR_ERR(old);
|
||||
}
|
||||
tracepoint_update_probes(); /* may update entry */
|
||||
entry = get_tracepoint(name);
|
||||
/* Make sure the entry was enabled */
|
||||
if (!entry || !entry->enabled)
|
||||
ret = -ENODEV;
|
||||
tp_func.func = probe;
|
||||
tp_func.data = data;
|
||||
ret = tracepoint_add_func(tp, &tp_func);
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
release_probes(old);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(tracepoint_probe_register);
|
||||
|
||||
static struct tracepoint_func *
|
||||
tracepoint_remove_probe(const char *name, void *probe, void *data)
|
||||
{
|
||||
struct tracepoint_entry *entry;
|
||||
struct tracepoint_func *old;
|
||||
|
||||
entry = get_tracepoint(name);
|
||||
if (!entry)
|
||||
return ERR_PTR(-ENOENT);
|
||||
old = tracepoint_entry_remove_probe(entry, probe, data);
|
||||
if (IS_ERR(old))
|
||||
return old;
|
||||
if (!entry->refcount)
|
||||
remove_tracepoint(entry);
|
||||
return old;
|
||||
}
|
||||
|
||||
/**
|
||||
* tracepoint_probe_unregister - Disconnect a probe from a tracepoint
|
||||
* @name: tracepoint name
|
||||
* @tp: tracepoint
|
||||
* @probe: probe function pointer
|
||||
* @data: probe private data
|
||||
*
|
||||
* We do not need to call a synchronize_sched to make sure the probes have
|
||||
* finished running before doing a module unload, because the module unload
|
||||
* itself uses stop_machine(), which insures that every preempt disabled section
|
||||
* have finished.
|
||||
* Returns 0 if ok, error value on error.
|
||||
*/
|
||||
int tracepoint_probe_unregister(const char *name, void *probe, void *data)
|
||||
int tracepoint_probe_unregister(struct tracepoint *tp, void *probe, void *data)
|
||||
{
|
||||
struct tracepoint_func *old;
|
||||
struct tracepoint_func tp_func;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&tracepoints_mutex);
|
||||
old = tracepoint_remove_probe(name, probe, data);
|
||||
if (IS_ERR(old)) {
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
return PTR_ERR(old);
|
||||
}
|
||||
tracepoint_update_probes(); /* may update entry */
|
||||
tp_func.func = probe;
|
||||
tp_func.data = data;
|
||||
ret = tracepoint_remove_func(tp, &tp_func);
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
release_probes(old);
|
||||
return 0;
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(tracepoint_probe_unregister);
|
||||
|
||||
|
||||
#ifdef CONFIG_MODULES
|
||||
bool trace_module_has_bad_taint(struct module *mod)
|
||||
{
|
||||
|
@ -464,6 +288,74 @@ bool trace_module_has_bad_taint(struct module *mod)
|
|||
(1 << TAINT_UNSIGNED_MODULE));
|
||||
}
|
||||
|
||||
static BLOCKING_NOTIFIER_HEAD(tracepoint_notify_list);
|
||||
|
||||
/**
|
||||
* register_tracepoint_notifier - register tracepoint coming/going notifier
|
||||
* @nb: notifier block
|
||||
*
|
||||
* Notifiers registered with this function are called on module
|
||||
* coming/going with the tracepoint_module_list_mutex held.
|
||||
* The notifier block callback should expect a "struct tp_module" data
|
||||
* pointer.
|
||||
*/
|
||||
int register_tracepoint_module_notifier(struct notifier_block *nb)
|
||||
{
|
||||
struct tp_module *tp_mod;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&tracepoint_module_list_mutex);
|
||||
ret = blocking_notifier_chain_register(&tracepoint_notify_list, nb);
|
||||
if (ret)
|
||||
goto end;
|
||||
list_for_each_entry(tp_mod, &tracepoint_module_list, list)
|
||||
(void) nb->notifier_call(nb, MODULE_STATE_COMING, tp_mod);
|
||||
end:
|
||||
mutex_unlock(&tracepoint_module_list_mutex);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(register_tracepoint_module_notifier);
|
||||
|
||||
/**
|
||||
* unregister_tracepoint_notifier - unregister tracepoint coming/going notifier
|
||||
* @nb: notifier block
|
||||
*
|
||||
* The notifier block callback should expect a "struct tp_module" data
|
||||
* pointer.
|
||||
*/
|
||||
int unregister_tracepoint_module_notifier(struct notifier_block *nb)
|
||||
{
|
||||
struct tp_module *tp_mod;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&tracepoint_module_list_mutex);
|
||||
ret = blocking_notifier_chain_unregister(&tracepoint_notify_list, nb);
|
||||
if (ret)
|
||||
goto end;
|
||||
list_for_each_entry(tp_mod, &tracepoint_module_list, list)
|
||||
(void) nb->notifier_call(nb, MODULE_STATE_GOING, tp_mod);
|
||||
end:
|
||||
mutex_unlock(&tracepoint_module_list_mutex);
|
||||
return ret;
|
||||
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(unregister_tracepoint_module_notifier);
|
||||
|
||||
/*
|
||||
* Ensure the tracer unregistered the module's probes before the module
|
||||
* teardown is performed. Prevents leaks of probe and data pointers.
|
||||
*/
|
||||
static void tp_module_going_check_quiescent(struct tracepoint * const *begin,
|
||||
struct tracepoint * const *end)
|
||||
{
|
||||
struct tracepoint * const *iter;
|
||||
|
||||
if (!begin)
|
||||
return;
|
||||
for (iter = begin; iter < end; iter++)
|
||||
WARN_ON_ONCE((*iter)->funcs);
|
||||
}
|
||||
|
||||
static int tracepoint_module_coming(struct module *mod)
|
||||
{
|
||||
struct tp_module *tp_mod;
|
||||
|
@ -479,36 +371,41 @@ static int tracepoint_module_coming(struct module *mod)
|
|||
*/
|
||||
if (trace_module_has_bad_taint(mod))
|
||||
return 0;
|
||||
mutex_lock(&tracepoints_mutex);
|
||||
mutex_lock(&tracepoint_module_list_mutex);
|
||||
tp_mod = kmalloc(sizeof(struct tp_module), GFP_KERNEL);
|
||||
if (!tp_mod) {
|
||||
ret = -ENOMEM;
|
||||
goto end;
|
||||
}
|
||||
tp_mod->num_tracepoints = mod->num_tracepoints;
|
||||
tp_mod->tracepoints_ptrs = mod->tracepoints_ptrs;
|
||||
tp_mod->mod = mod;
|
||||
list_add_tail(&tp_mod->list, &tracepoint_module_list);
|
||||
tracepoint_update_probe_range(mod->tracepoints_ptrs,
|
||||
mod->tracepoints_ptrs + mod->num_tracepoints);
|
||||
blocking_notifier_call_chain(&tracepoint_notify_list,
|
||||
MODULE_STATE_COMING, tp_mod);
|
||||
end:
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
mutex_unlock(&tracepoint_module_list_mutex);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int tracepoint_module_going(struct module *mod)
|
||||
static void tracepoint_module_going(struct module *mod)
|
||||
{
|
||||
struct tp_module *pos;
|
||||
struct tp_module *tp_mod;
|
||||
|
||||
if (!mod->num_tracepoints)
|
||||
return 0;
|
||||
return;
|
||||
|
||||
mutex_lock(&tracepoints_mutex);
|
||||
tracepoint_update_probe_range(mod->tracepoints_ptrs,
|
||||
mod->tracepoints_ptrs + mod->num_tracepoints);
|
||||
list_for_each_entry(pos, &tracepoint_module_list, list) {
|
||||
if (pos->tracepoints_ptrs == mod->tracepoints_ptrs) {
|
||||
list_del(&pos->list);
|
||||
kfree(pos);
|
||||
mutex_lock(&tracepoint_module_list_mutex);
|
||||
list_for_each_entry(tp_mod, &tracepoint_module_list, list) {
|
||||
if (tp_mod->mod == mod) {
|
||||
blocking_notifier_call_chain(&tracepoint_notify_list,
|
||||
MODULE_STATE_GOING, tp_mod);
|
||||
list_del(&tp_mod->list);
|
||||
kfree(tp_mod);
|
||||
/*
|
||||
* Called the going notifier before checking for
|
||||
* quiescence.
|
||||
*/
|
||||
tp_module_going_check_quiescent(mod->tracepoints_ptrs,
|
||||
mod->tracepoints_ptrs + mod->num_tracepoints);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
@ -518,12 +415,11 @@ static int tracepoint_module_going(struct module *mod)
|
|||
* flag on "going", in case a module taints the kernel only after being
|
||||
* loaded.
|
||||
*/
|
||||
mutex_unlock(&tracepoints_mutex);
|
||||
return 0;
|
||||
mutex_unlock(&tracepoint_module_list_mutex);
|
||||
}
|
||||
|
||||
int tracepoint_module_notify(struct notifier_block *self,
|
||||
unsigned long val, void *data)
|
||||
static int tracepoint_module_notify(struct notifier_block *self,
|
||||
unsigned long val, void *data)
|
||||
{
|
||||
struct module *mod = data;
|
||||
int ret = 0;
|
||||
|
@ -535,24 +431,58 @@ int tracepoint_module_notify(struct notifier_block *self,
|
|||
case MODULE_STATE_LIVE:
|
||||
break;
|
||||
case MODULE_STATE_GOING:
|
||||
ret = tracepoint_module_going(mod);
|
||||
tracepoint_module_going(mod);
|
||||
break;
|
||||
case MODULE_STATE_UNFORMED:
|
||||
break;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
struct notifier_block tracepoint_module_nb = {
|
||||
static struct notifier_block tracepoint_module_nb = {
|
||||
.notifier_call = tracepoint_module_notify,
|
||||
.priority = 0,
|
||||
};
|
||||
|
||||
static int init_tracepoints(void)
|
||||
static __init int init_tracepoints(void)
|
||||
{
|
||||
return register_module_notifier(&tracepoint_module_nb);
|
||||
int ret;
|
||||
|
||||
ret = register_module_notifier(&tracepoint_module_nb);
|
||||
if (ret)
|
||||
pr_warning("Failed to register tracepoint module enter notifier\n");
|
||||
|
||||
return ret;
|
||||
}
|
||||
__initcall(init_tracepoints);
|
||||
#endif /* CONFIG_MODULES */
|
||||
|
||||
static void for_each_tracepoint_range(struct tracepoint * const *begin,
|
||||
struct tracepoint * const *end,
|
||||
void (*fct)(struct tracepoint *tp, void *priv),
|
||||
void *priv)
|
||||
{
|
||||
struct tracepoint * const *iter;
|
||||
|
||||
if (!begin)
|
||||
return;
|
||||
for (iter = begin; iter < end; iter++)
|
||||
fct(*iter, priv);
|
||||
}
|
||||
|
||||
/**
|
||||
* for_each_kernel_tracepoint - iteration on all kernel tracepoints
|
||||
* @fct: callback
|
||||
* @priv: private data
|
||||
*/
|
||||
void for_each_kernel_tracepoint(void (*fct)(struct tracepoint *tp, void *priv),
|
||||
void *priv)
|
||||
{
|
||||
for_each_tracepoint_range(__start___tracepoints_ptrs,
|
||||
__stop___tracepoints_ptrs, fct, priv);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(for_each_kernel_tracepoint);
|
||||
|
||||
#ifdef CONFIG_HAVE_SYSCALL_TRACEPOINTS
|
||||
|
||||
/* NB: reg/unreg are called while guarded with the tracepoints_mutex */
|
||||
|
|
Loading…
Reference in New Issue