perf: Add API for PMUs to write to the AUX area
For pmus that wish to write data to ring buffer's AUX area, provide perf_aux_output_{begin,end}() calls to initiate/commit data writes, similarly to perf_output_{begin,end}. These also use the same output handle structure. Also, similarly to software counterparts, these will direct inherited events' output to parents' ring buffers. After the perf_aux_output_begin() returns successfully, handle->size is set to the maximum amount of data that can be written wrt aux_tail pointer, so that no data that the user hasn't seen will be overwritten, therefore this should always be called before hardware writing is enabled. On success, this will return the pointer to pmu driver's private structure allocated for this aux area by pmu::setup_aux. Same pointer can also be retrieved using perf_get_aux() while hardware writing is enabled. PMU driver should pass the actual amount of data written as a parameter to perf_aux_output_end(). All hardware writes should be completed and visible before this one is called. Additionally, perf_aux_output_skip() will adjust output handle and aux_head in case some part of the buffer has to be skipped over to maintain hardware's alignment constraints. Nested writers are forbidden and guards are in place to catch such attempts. Signed-off-by: Alexander Shishkin <alexander.shishkin@linux.intel.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Borislav Petkov <bp@alien8.de> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: H. Peter Anvin <hpa@zytor.com> Cc: Kaixu Xia <kaixu.xia@linaro.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Mike Galbraith <efault@gmx.de> Cc: Paul Mackerras <paulus@samba.org> Cc: Robert Richter <rric@kernel.org> Cc: Stephane Eranian <eranian@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: acme@infradead.org Cc: adrian.hunter@intel.com Cc: kan.liang@intel.com Cc: markus.t.metzger@intel.com Cc: mathieu.poirier@linaro.org Link: http://lkml.kernel.org/r/1421237903-181015-8-git-send-email-alexander.shishkin@linux.intel.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
parent
68db7e98c3
commit
fdc2670666
|
@ -573,7 +573,10 @@ struct perf_output_handle {
|
|||
struct ring_buffer *rb;
|
||||
unsigned long wakeup;
|
||||
unsigned long size;
|
||||
void *addr;
|
||||
union {
|
||||
void *addr;
|
||||
unsigned long head;
|
||||
};
|
||||
int page;
|
||||
};
|
||||
|
||||
|
@ -608,6 +611,14 @@ perf_cgroup_from_task(struct task_struct *task)
|
|||
|
||||
#ifdef CONFIG_PERF_EVENTS
|
||||
|
||||
extern void *perf_aux_output_begin(struct perf_output_handle *handle,
|
||||
struct perf_event *event);
|
||||
extern void perf_aux_output_end(struct perf_output_handle *handle,
|
||||
unsigned long size, bool truncated);
|
||||
extern int perf_aux_output_skip(struct perf_output_handle *handle,
|
||||
unsigned long size);
|
||||
extern void *perf_get_aux(struct perf_output_handle *handle);
|
||||
|
||||
extern int perf_pmu_register(struct pmu *pmu, const char *name, int type);
|
||||
extern void perf_pmu_unregister(struct pmu *pmu);
|
||||
|
||||
|
@ -898,6 +909,17 @@ extern void perf_event_disable(struct perf_event *event);
|
|||
extern int __perf_event_disable(void *info);
|
||||
extern void perf_event_task_tick(void);
|
||||
#else /* !CONFIG_PERF_EVENTS: */
|
||||
static inline void *
|
||||
perf_aux_output_begin(struct perf_output_handle *handle,
|
||||
struct perf_event *event) { return NULL; }
|
||||
static inline void
|
||||
perf_aux_output_end(struct perf_output_handle *handle, unsigned long size,
|
||||
bool truncated) { }
|
||||
static inline int
|
||||
perf_aux_output_skip(struct perf_output_handle *handle,
|
||||
unsigned long size) { return -EINVAL; }
|
||||
static inline void *
|
||||
perf_get_aux(struct perf_output_handle *handle) { return NULL; }
|
||||
static inline void
|
||||
perf_event_task_sched_in(struct task_struct *prev,
|
||||
struct task_struct *task) { }
|
||||
|
|
|
@ -3423,7 +3423,6 @@ static void free_event_rcu(struct rcu_head *head)
|
|||
kfree(event);
|
||||
}
|
||||
|
||||
static void ring_buffer_put(struct ring_buffer *rb);
|
||||
static void ring_buffer_attach(struct perf_event *event,
|
||||
struct ring_buffer *rb);
|
||||
|
||||
|
@ -4361,7 +4360,7 @@ static void rb_free_rcu(struct rcu_head *rcu_head)
|
|||
rb_free(rb);
|
||||
}
|
||||
|
||||
static struct ring_buffer *ring_buffer_get(struct perf_event *event)
|
||||
struct ring_buffer *ring_buffer_get(struct perf_event *event)
|
||||
{
|
||||
struct ring_buffer *rb;
|
||||
|
||||
|
@ -4376,7 +4375,7 @@ static struct ring_buffer *ring_buffer_get(struct perf_event *event)
|
|||
return rb;
|
||||
}
|
||||
|
||||
static void ring_buffer_put(struct ring_buffer *rb)
|
||||
void ring_buffer_put(struct ring_buffer *rb)
|
||||
{
|
||||
if (!atomic_dec_and_test(&rb->refcount))
|
||||
return;
|
||||
|
|
|
@ -36,6 +36,8 @@ struct ring_buffer {
|
|||
struct user_struct *mmap_user;
|
||||
|
||||
/* AUX area */
|
||||
local_t aux_head;
|
||||
local_t aux_nest;
|
||||
unsigned long aux_pgoff;
|
||||
int aux_nr_pages;
|
||||
atomic_t aux_mmap_count;
|
||||
|
@ -56,6 +58,8 @@ extern void perf_event_wakeup(struct perf_event *event);
|
|||
extern int rb_alloc_aux(struct ring_buffer *rb, struct perf_event *event,
|
||||
pgoff_t pgoff, int nr_pages, int flags);
|
||||
extern void rb_free_aux(struct ring_buffer *rb);
|
||||
extern struct ring_buffer *ring_buffer_get(struct perf_event *event);
|
||||
extern void ring_buffer_put(struct ring_buffer *rb);
|
||||
|
||||
static inline bool rb_has_aux(struct ring_buffer *rb)
|
||||
{
|
||||
|
|
|
@ -243,6 +243,145 @@ ring_buffer_init(struct ring_buffer *rb, long watermark, int flags)
|
|||
spin_lock_init(&rb->event_lock);
|
||||
}
|
||||
|
||||
/*
|
||||
* This is called before hardware starts writing to the AUX area to
|
||||
* obtain an output handle and make sure there's room in the buffer.
|
||||
* When the capture completes, call perf_aux_output_end() to commit
|
||||
* the recorded data to the buffer.
|
||||
*
|
||||
* The ordering is similar to that of perf_output_{begin,end}, with
|
||||
* the exception of (B), which should be taken care of by the pmu
|
||||
* driver, since ordering rules will differ depending on hardware.
|
||||
*/
|
||||
void *perf_aux_output_begin(struct perf_output_handle *handle,
|
||||
struct perf_event *event)
|
||||
{
|
||||
struct perf_event *output_event = event;
|
||||
unsigned long aux_head, aux_tail;
|
||||
struct ring_buffer *rb;
|
||||
|
||||
if (output_event->parent)
|
||||
output_event = output_event->parent;
|
||||
|
||||
/*
|
||||
* Since this will typically be open across pmu::add/pmu::del, we
|
||||
* grab ring_buffer's refcount instead of holding rcu read lock
|
||||
* to make sure it doesn't disappear under us.
|
||||
*/
|
||||
rb = ring_buffer_get(output_event);
|
||||
if (!rb)
|
||||
return NULL;
|
||||
|
||||
if (!rb_has_aux(rb) || !atomic_inc_not_zero(&rb->aux_refcount))
|
||||
goto err;
|
||||
|
||||
/*
|
||||
* Nesting is not supported for AUX area, make sure nested
|
||||
* writers are caught early
|
||||
*/
|
||||
if (WARN_ON_ONCE(local_xchg(&rb->aux_nest, 1)))
|
||||
goto err_put;
|
||||
|
||||
aux_head = local_read(&rb->aux_head);
|
||||
aux_tail = ACCESS_ONCE(rb->user_page->aux_tail);
|
||||
|
||||
handle->rb = rb;
|
||||
handle->event = event;
|
||||
handle->head = aux_head;
|
||||
if (aux_head - aux_tail < perf_aux_size(rb))
|
||||
handle->size = CIRC_SPACE(aux_head, aux_tail, perf_aux_size(rb));
|
||||
else
|
||||
handle->size = 0;
|
||||
|
||||
/*
|
||||
* handle->size computation depends on aux_tail load; this forms a
|
||||
* control dependency barrier separating aux_tail load from aux data
|
||||
* store that will be enabled on successful return
|
||||
*/
|
||||
if (!handle->size) { /* A, matches D */
|
||||
event->pending_disable = 1;
|
||||
perf_output_wakeup(handle);
|
||||
local_set(&rb->aux_nest, 0);
|
||||
goto err_put;
|
||||
}
|
||||
|
||||
return handle->rb->aux_priv;
|
||||
|
||||
err_put:
|
||||
rb_free_aux(rb);
|
||||
|
||||
err:
|
||||
ring_buffer_put(rb);
|
||||
handle->event = NULL;
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* Commit the data written by hardware into the ring buffer by adjusting
|
||||
* aux_head and posting a PERF_RECORD_AUX into the perf buffer. It is the
|
||||
* pmu driver's responsibility to observe ordering rules of the hardware,
|
||||
* so that all the data is externally visible before this is called.
|
||||
*/
|
||||
void perf_aux_output_end(struct perf_output_handle *handle, unsigned long size,
|
||||
bool truncated)
|
||||
{
|
||||
struct ring_buffer *rb = handle->rb;
|
||||
unsigned long aux_head = local_read(&rb->aux_head);
|
||||
u64 flags = 0;
|
||||
|
||||
if (truncated)
|
||||
flags |= PERF_AUX_FLAG_TRUNCATED;
|
||||
|
||||
local_add(size, &rb->aux_head);
|
||||
|
||||
if (size || flags) {
|
||||
/*
|
||||
* Only send RECORD_AUX if we have something useful to communicate
|
||||
*/
|
||||
|
||||
perf_event_aux_event(handle->event, aux_head, size, flags);
|
||||
}
|
||||
|
||||
rb->user_page->aux_head = local_read(&rb->aux_head);
|
||||
|
||||
perf_output_wakeup(handle);
|
||||
handle->event = NULL;
|
||||
|
||||
local_set(&rb->aux_nest, 0);
|
||||
rb_free_aux(rb);
|
||||
ring_buffer_put(rb);
|
||||
}
|
||||
|
||||
/*
|
||||
* Skip over a given number of bytes in the AUX buffer, due to, for example,
|
||||
* hardware's alignment constraints.
|
||||
*/
|
||||
int perf_aux_output_skip(struct perf_output_handle *handle, unsigned long size)
|
||||
{
|
||||
struct ring_buffer *rb = handle->rb;
|
||||
unsigned long aux_head;
|
||||
|
||||
if (size > handle->size)
|
||||
return -ENOSPC;
|
||||
|
||||
local_add(size, &rb->aux_head);
|
||||
|
||||
handle->head = aux_head;
|
||||
handle->size -= size;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void *perf_get_aux(struct perf_output_handle *handle)
|
||||
{
|
||||
/* this is only valid between perf_aux_output_begin and *_end */
|
||||
if (!handle->event)
|
||||
return NULL;
|
||||
|
||||
return handle->rb->aux_priv;
|
||||
}
|
||||
|
||||
#define PERF_AUX_GFP (GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY)
|
||||
|
||||
static struct page *rb_alloc_aux_page(int node, int order)
|
||||
|
|
Loading…
Reference in New Issue