mirror of https://gitee.com/openkylin/linux.git
perf stat: Introduce --repeat forever
The following patch causes 'perf stat --repeat 0' to be interpreted as 'forever', displaying the stats for every run. We act as if a single run was asked, and reset the stats in each iteration. In this mode SIGINT is passed to perf to be able to stop the loop with Ctrl+C. Signed-off-by: Frederik Deweerdt <frederik.deweerdt@gmail.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Paul Mackerras <paulus@samba.org> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Link: http://lkml.kernel.org/r/20130301180227.GA24385@ks398093.ip-192-95-24.net Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
d723a55096
commit
a7e191c376
|
@ -52,7 +52,7 @@ OPTIONS
|
||||||
|
|
||||||
-r::
|
-r::
|
||||||
--repeat=<n>::
|
--repeat=<n>::
|
||||||
repeat command and print average + stddev (max: 100)
|
repeat command and print average + stddev (max: 100). 0 means forever.
|
||||||
|
|
||||||
-B::
|
-B::
|
||||||
--big-num::
|
--big-num::
|
||||||
|
|
|
@ -94,6 +94,7 @@ static const char *pre_cmd = NULL;
|
||||||
static const char *post_cmd = NULL;
|
static const char *post_cmd = NULL;
|
||||||
static bool sync_run = false;
|
static bool sync_run = false;
|
||||||
static unsigned int interval = 0;
|
static unsigned int interval = 0;
|
||||||
|
static bool forever = false;
|
||||||
static struct timespec ref_time;
|
static struct timespec ref_time;
|
||||||
static struct cpu_map *sock_map;
|
static struct cpu_map *sock_map;
|
||||||
|
|
||||||
|
@ -125,6 +126,11 @@ static inline int perf_evsel__nr_cpus(struct perf_evsel *evsel)
|
||||||
return perf_evsel__cpus(evsel)->nr;
|
return perf_evsel__cpus(evsel)->nr;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void perf_evsel__reset_stat_priv(struct perf_evsel *evsel)
|
||||||
|
{
|
||||||
|
memset(evsel->priv, 0, sizeof(struct perf_stat));
|
||||||
|
}
|
||||||
|
|
||||||
static int perf_evsel__alloc_stat_priv(struct perf_evsel *evsel)
|
static int perf_evsel__alloc_stat_priv(struct perf_evsel *evsel)
|
||||||
{
|
{
|
||||||
evsel->priv = zalloc(sizeof(struct perf_stat));
|
evsel->priv = zalloc(sizeof(struct perf_stat));
|
||||||
|
@ -173,6 +179,22 @@ static struct stats runtime_itlb_cache_stats[MAX_NR_CPUS];
|
||||||
static struct stats runtime_dtlb_cache_stats[MAX_NR_CPUS];
|
static struct stats runtime_dtlb_cache_stats[MAX_NR_CPUS];
|
||||||
static struct stats walltime_nsecs_stats;
|
static struct stats walltime_nsecs_stats;
|
||||||
|
|
||||||
|
static void reset_stats(void)
|
||||||
|
{
|
||||||
|
memset(runtime_nsecs_stats, 0, sizeof(runtime_nsecs_stats));
|
||||||
|
memset(runtime_cycles_stats, 0, sizeof(runtime_cycles_stats));
|
||||||
|
memset(runtime_stalled_cycles_front_stats, 0, sizeof(runtime_stalled_cycles_front_stats));
|
||||||
|
memset(runtime_stalled_cycles_back_stats, 0, sizeof(runtime_stalled_cycles_back_stats));
|
||||||
|
memset(runtime_branches_stats, 0, sizeof(runtime_branches_stats));
|
||||||
|
memset(runtime_cacherefs_stats, 0, sizeof(runtime_cacherefs_stats));
|
||||||
|
memset(runtime_l1_dcache_stats, 0, sizeof(runtime_l1_dcache_stats));
|
||||||
|
memset(runtime_l1_icache_stats, 0, sizeof(runtime_l1_icache_stats));
|
||||||
|
memset(runtime_ll_cache_stats, 0, sizeof(runtime_ll_cache_stats));
|
||||||
|
memset(runtime_itlb_cache_stats, 0, sizeof(runtime_itlb_cache_stats));
|
||||||
|
memset(runtime_dtlb_cache_stats, 0, sizeof(runtime_dtlb_cache_stats));
|
||||||
|
memset(&walltime_nsecs_stats, 0, sizeof(walltime_nsecs_stats));
|
||||||
|
}
|
||||||
|
|
||||||
static int create_perf_stat_counter(struct perf_evsel *evsel)
|
static int create_perf_stat_counter(struct perf_evsel *evsel)
|
||||||
{
|
{
|
||||||
struct perf_event_attr *attr = &evsel->attr;
|
struct perf_event_attr *attr = &evsel->attr;
|
||||||
|
@ -1252,7 +1274,7 @@ int cmd_stat(int argc, const char **argv, const char *prefix __maybe_unused)
|
||||||
OPT_INCR('v', "verbose", &verbose,
|
OPT_INCR('v', "verbose", &verbose,
|
||||||
"be more verbose (show counter open errors, etc)"),
|
"be more verbose (show counter open errors, etc)"),
|
||||||
OPT_INTEGER('r', "repeat", &run_count,
|
OPT_INTEGER('r', "repeat", &run_count,
|
||||||
"repeat command and print average + stddev (max: 100)"),
|
"repeat command and print average + stddev (max: 100, forever: 0)"),
|
||||||
OPT_BOOLEAN('n', "null", &null_run,
|
OPT_BOOLEAN('n', "null", &null_run,
|
||||||
"null run - dont start any counters"),
|
"null run - dont start any counters"),
|
||||||
OPT_INCR('d', "detailed", &detailed_run,
|
OPT_INCR('d', "detailed", &detailed_run,
|
||||||
|
@ -1355,8 +1377,12 @@ int cmd_stat(int argc, const char **argv, const char *prefix __maybe_unused)
|
||||||
|
|
||||||
if (!argc && !perf_target__has_task(&target))
|
if (!argc && !perf_target__has_task(&target))
|
||||||
usage_with_options(stat_usage, options);
|
usage_with_options(stat_usage, options);
|
||||||
if (run_count <= 0)
|
if (run_count < 0) {
|
||||||
usage_with_options(stat_usage, options);
|
usage_with_options(stat_usage, options);
|
||||||
|
} else if (run_count == 0) {
|
||||||
|
forever = true;
|
||||||
|
run_count = 1;
|
||||||
|
}
|
||||||
|
|
||||||
/* no_aggr, cgroup are for system-wide only */
|
/* no_aggr, cgroup are for system-wide only */
|
||||||
if ((no_aggr || nr_cgroups) && !perf_target__has_cpu(&target)) {
|
if ((no_aggr || nr_cgroups) && !perf_target__has_cpu(&target)) {
|
||||||
|
@ -1413,21 +1439,30 @@ int cmd_stat(int argc, const char **argv, const char *prefix __maybe_unused)
|
||||||
* task, but being ignored by perf stat itself:
|
* task, but being ignored by perf stat itself:
|
||||||
*/
|
*/
|
||||||
atexit(sig_atexit);
|
atexit(sig_atexit);
|
||||||
signal(SIGINT, skip_signal);
|
if (!forever)
|
||||||
|
signal(SIGINT, skip_signal);
|
||||||
signal(SIGCHLD, skip_signal);
|
signal(SIGCHLD, skip_signal);
|
||||||
signal(SIGALRM, skip_signal);
|
signal(SIGALRM, skip_signal);
|
||||||
signal(SIGABRT, skip_signal);
|
signal(SIGABRT, skip_signal);
|
||||||
|
|
||||||
status = 0;
|
status = 0;
|
||||||
for (run_idx = 0; run_idx < run_count; run_idx++) {
|
for (run_idx = 0; forever || run_idx < run_count; run_idx++) {
|
||||||
if (run_count != 1 && verbose)
|
if (run_count != 1 && verbose)
|
||||||
fprintf(output, "[ perf stat: executing run #%d ... ]\n",
|
fprintf(output, "[ perf stat: executing run #%d ... ]\n",
|
||||||
run_idx + 1);
|
run_idx + 1);
|
||||||
|
|
||||||
status = run_perf_stat(argc, argv);
|
status = run_perf_stat(argc, argv);
|
||||||
|
if (forever && status != -1) {
|
||||||
|
print_stat(argc, argv);
|
||||||
|
list_for_each_entry(pos, &evsel_list->entries, node) {
|
||||||
|
perf_evsel__reset_stat_priv(pos);
|
||||||
|
perf_evsel__reset_counts(pos, perf_evsel__nr_cpus(pos));
|
||||||
|
}
|
||||||
|
reset_stats();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (status != -1 && !interval)
|
if (!forever && status != -1 && !interval)
|
||||||
print_stat(argc, argv);
|
print_stat(argc, argv);
|
||||||
out_free_fd:
|
out_free_fd:
|
||||||
list_for_each_entry(pos, &evsel_list->entries, node) {
|
list_for_each_entry(pos, &evsel_list->entries, node) {
|
||||||
|
|
|
@ -633,6 +633,12 @@ int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void perf_evsel__reset_counts(struct perf_evsel *evsel, int ncpus)
|
||||||
|
{
|
||||||
|
memset(evsel->counts, 0, (sizeof(*evsel->counts) +
|
||||||
|
(ncpus * sizeof(struct perf_counts_values))));
|
||||||
|
}
|
||||||
|
|
||||||
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus)
|
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus)
|
||||||
{
|
{
|
||||||
evsel->counts = zalloc((sizeof(*evsel->counts) +
|
evsel->counts = zalloc((sizeof(*evsel->counts) +
|
||||||
|
|
|
@ -121,6 +121,7 @@ int perf_evsel__group_desc(struct perf_evsel *evsel, char *buf, size_t size);
|
||||||
int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads);
|
int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads);
|
||||||
int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads);
|
int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads);
|
||||||
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus);
|
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus);
|
||||||
|
void perf_evsel__reset_counts(struct perf_evsel *evsel, int ncpus);
|
||||||
void perf_evsel__free_fd(struct perf_evsel *evsel);
|
void perf_evsel__free_fd(struct perf_evsel *evsel);
|
||||||
void perf_evsel__free_id(struct perf_evsel *evsel);
|
void perf_evsel__free_id(struct perf_evsel *evsel);
|
||||||
void perf_evsel__free_counts(struct perf_evsel *evsel);
|
void perf_evsel__free_counts(struct perf_evsel *evsel);
|
||||||
|
|
Loading…
Reference in New Issue