2015-04-30 01:11:00 +08:00
|
|
|
|
|
|
|
#undef TRACE_SYSTEM_VAR
|
|
|
|
|
|
|
|
#ifdef CONFIG_PERF_EVENTS
|
|
|
|
|
|
|
|
#undef __entry
|
|
|
|
#define __entry entry
|
|
|
|
|
|
|
|
#undef __get_dynamic_array
|
|
|
|
#define __get_dynamic_array(field) \
|
|
|
|
((void *)__entry + (__entry->__data_loc_##field & 0xffff))
|
|
|
|
|
|
|
|
#undef __get_dynamic_array_len
|
|
|
|
#define __get_dynamic_array_len(field) \
|
|
|
|
((__entry->__data_loc_##field >> 16) & 0xffff)
|
|
|
|
|
|
|
|
#undef __get_str
|
2016-07-02 07:44:34 +08:00
|
|
|
#define __get_str(field) ((char *)__get_dynamic_array(field))
|
2015-04-30 01:11:00 +08:00
|
|
|
|
|
|
|
#undef __get_bitmask
|
|
|
|
#define __get_bitmask(field) (char *)__get_dynamic_array(field)
|
|
|
|
|
|
|
|
#undef __perf_count
|
|
|
|
#define __perf_count(c) (__count = (c))
|
|
|
|
|
|
|
|
#undef __perf_task
|
|
|
|
#define __perf_task(t) (__task = (t))
|
|
|
|
|
|
|
|
#undef DECLARE_EVENT_CLASS
|
|
|
|
#define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \
|
|
|
|
static notrace void \
|
|
|
|
perf_trace_##call(void *__data, proto) \
|
|
|
|
{ \
|
2015-05-05 23:45:27 +08:00
|
|
|
struct trace_event_call *event_call = __data; \
|
2015-05-14 03:33:52 +08:00
|
|
|
struct trace_event_data_offsets_##call __maybe_unused __data_offsets;\
|
2015-05-14 03:27:47 +08:00
|
|
|
struct trace_event_raw_##call *entry; \
|
2016-04-07 09:43:25 +08:00
|
|
|
struct bpf_prog *prog = event_call->prog; \
|
2015-04-30 01:11:00 +08:00
|
|
|
struct pt_regs *__regs; \
|
2016-04-07 09:43:23 +08:00
|
|
|
u64 __count = 1; \
|
2015-04-30 01:11:00 +08:00
|
|
|
struct task_struct *__task = NULL; \
|
|
|
|
struct hlist_head *head; \
|
|
|
|
int __entry_size; \
|
|
|
|
int __data_size; \
|
|
|
|
int rctx; \
|
|
|
|
\
|
2015-05-14 03:40:23 +08:00
|
|
|
__data_size = trace_event_get_offsets_##call(&__data_offsets, args); \
|
2015-04-30 01:11:00 +08:00
|
|
|
\
|
|
|
|
head = this_cpu_ptr(event_call->perf_events); \
|
2016-04-07 09:43:25 +08:00
|
|
|
if (!prog && __builtin_constant_p(!__task) && !__task && \
|
2015-04-30 01:11:00 +08:00
|
|
|
hlist_empty(head)) \
|
|
|
|
return; \
|
|
|
|
\
|
|
|
|
__entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\
|
|
|
|
sizeof(u64)); \
|
|
|
|
__entry_size -= sizeof(u32); \
|
|
|
|
\
|
2016-04-07 09:43:24 +08:00
|
|
|
entry = perf_trace_buf_alloc(__entry_size, &__regs, &rctx); \
|
2015-04-30 01:11:00 +08:00
|
|
|
if (!entry) \
|
|
|
|
return; \
|
|
|
|
\
|
|
|
|
perf_fetch_caller_regs(__regs); \
|
|
|
|
\
|
|
|
|
tstruct \
|
|
|
|
\
|
|
|
|
{ assign; } \
|
|
|
|
\
|
2016-04-19 11:11:50 +08:00
|
|
|
perf_trace_run_bpf_submit(entry, __entry_size, rctx, \
|
|
|
|
event_call, __count, __regs, \
|
|
|
|
head, __task); \
|
2015-04-30 01:11:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This part is compiled out, it is only here as a build time check
|
|
|
|
* to make sure that if the tracepoint handling changes, the
|
|
|
|
* perf probe will fail to compile unless it too is updated.
|
|
|
|
*/
|
|
|
|
#undef DEFINE_EVENT
|
|
|
|
#define DEFINE_EVENT(template, call, proto, args) \
|
|
|
|
static inline void perf_test_probe_##call(void) \
|
|
|
|
{ \
|
|
|
|
check_trace_callback_type_##call(perf_trace_##template); \
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#undef DEFINE_EVENT_PRINT
|
|
|
|
#define DEFINE_EVENT_PRINT(template, name, proto, args, print) \
|
|
|
|
DEFINE_EVENT(template, name, PARAMS(proto), PARAMS(args))
|
|
|
|
|
|
|
|
#include TRACE_INCLUDE(TRACE_INCLUDE_FILE)
|
|
|
|
#endif /* CONFIG_PERF_EVENTS */
|