alistair23-linux/tools/perf/util/metricgroup.c
Jiri Olsa f6fb0960f9 perf metric: Add recursion check when processing nested metrics
Keeping the stack of nested metrics via 'struct expr_id' objects
and checking if we are in recursion via already processed metric.

The stack is implemented as static array within the struct egroup
with 100 entries, which should be enough nesting depth for any
metric we have or plan to have at the moment.

Adding test that simulates the recursion and checks we can
detect it.

Committer notes:

Bumped RECURSION_ID_MAX to 1000 as per Jiri's reply to Paul Clark on the
patch series e-mail discussion.

Fixed these:

  tests/parse-metric.c:308:7: error: missing field 'val' initializer [-Werror,-Wmissing-field-initializers]
                  { 0 },
                      ^

  util/metricgroup.c:924:28: error: missing field 'parent' initializer [-Werror,-Wmissing-field-initializers]
          struct expr_ids ids = { 0 };
                                    ^
  util/metricgroup.c:924:26: error: suggest braces around initialization of subobject [-Werror,-Wmissing-braces]
          struct expr_ids ids = { 0 };
                                  ^
                                  {}
  util/metricgroup.c:924:26: error: suggest braces around initialization of subobject [-Werror,-Wmissing-braces]
          struct expr_ids ids = { 0 };
                                  ^
                                  {}
  util/metricgroup.c:924:28: error: missing field 'cnt' initializer [-Werror,-Wmissing-field-initializers]
          struct expr_ids ids = { 0 };
                                    ^

Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Reviewed-by: Kajol Jain <kjain@linux.ibm.com>
Acked-by: Ian Rogers <irogers@google.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: John Garry <john.garry@huawei.com>
Cc: Michael Petlan <mpetlan@redhat.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Paul Clarke <pc@us.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Stephane Eranian <eranian@google.com>
Link: http://lore.kernel.org/lkml/20200719181320.785305-16-jolsa@kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2020-07-30 07:01:49 -03:00

1107 lines
25 KiB
C

// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright (c) 2017, Intel Corporation.
*/
/* Manage metrics and groups of metrics from JSON files */
#include "metricgroup.h"
#include "debug.h"
#include "evlist.h"
#include "evsel.h"
#include "strbuf.h"
#include "pmu.h"
#include "expr.h"
#include "rblist.h"
#include <string.h>
#include <errno.h>
#include "pmu-events/pmu-events.h"
#include "strlist.h"
#include <assert.h>
#include <linux/ctype.h>
#include <linux/string.h>
#include <linux/zalloc.h>
#include <subcmd/parse-options.h>
#include <api/fs/fs.h>
#include "util.h"
#include <asm/bug.h>
struct metric_event *metricgroup__lookup(struct rblist *metric_events,
struct evsel *evsel,
bool create)
{
struct rb_node *nd;
struct metric_event me = {
.evsel = evsel
};
if (!metric_events)
return NULL;
nd = rblist__find(metric_events, &me);
if (nd)
return container_of(nd, struct metric_event, nd);
if (create) {
rblist__add_node(metric_events, &me);
nd = rblist__find(metric_events, &me);
if (nd)
return container_of(nd, struct metric_event, nd);
}
return NULL;
}
static int metric_event_cmp(struct rb_node *rb_node, const void *entry)
{
struct metric_event *a = container_of(rb_node,
struct metric_event,
nd);
const struct metric_event *b = entry;
if (a->evsel == b->evsel)
return 0;
if ((char *)a->evsel < (char *)b->evsel)
return -1;
return +1;
}
static struct rb_node *metric_event_new(struct rblist *rblist __maybe_unused,
const void *entry)
{
struct metric_event *me = malloc(sizeof(struct metric_event));
if (!me)
return NULL;
memcpy(me, entry, sizeof(struct metric_event));
me->evsel = ((struct metric_event *)entry)->evsel;
INIT_LIST_HEAD(&me->head);
return &me->nd;
}
static void metric_event_delete(struct rblist *rblist __maybe_unused,
struct rb_node *rb_node)
{
struct metric_event *me = container_of(rb_node, struct metric_event, nd);
struct metric_expr *expr, *tmp;
list_for_each_entry_safe(expr, tmp, &me->head, nd) {
free(expr->metric_refs);
free(expr);
}
free(me);
}
static void metricgroup__rblist_init(struct rblist *metric_events)
{
rblist__init(metric_events);
metric_events->node_cmp = metric_event_cmp;
metric_events->node_new = metric_event_new;
metric_events->node_delete = metric_event_delete;
}
void metricgroup__rblist_exit(struct rblist *metric_events)
{
rblist__exit(metric_events);
}
/*
* A node in the list of referenced metrics. metric_expr
* is held as a convenience to avoid a search through the
* metric list.
*/
struct metric_ref_node {
const char *metric_name;
const char *metric_expr;
struct list_head list;
};
struct egroup {
struct list_head nd;
struct expr_parse_ctx pctx;
const char *metric_name;
const char *metric_expr;
const char *metric_unit;
struct list_head metric_refs;
int metric_refs_cnt;
int runtime;
bool has_constraint;
};
#define RECURSION_ID_MAX 1000
struct expr_ids {
struct expr_id id[RECURSION_ID_MAX];
int cnt;
};
static struct expr_id *expr_ids__alloc(struct expr_ids *ids)
{
if (ids->cnt >= RECURSION_ID_MAX)
return NULL;
return &ids->id[ids->cnt++];
}
static void expr_ids__exit(struct expr_ids *ids)
{
int i;
for (i = 0; i < ids->cnt; i++)
free(ids->id[i].id);
}
/**
* Find a group of events in perf_evlist that correpond to those from a parsed
* metric expression. Note, as find_evsel_group is called in the same order as
* perf_evlist was constructed, metric_no_merge doesn't need to test for
* underfilling a group.
* @perf_evlist: a list of events something like: {metric1 leader, metric1
* sibling, metric1 sibling}:W,duration_time,{metric2 leader, metric2 sibling,
* metric2 sibling}:W,duration_time
* @pctx: the parse context for the metric expression.
* @metric_no_merge: don't attempt to share events for the metric with other
* metrics.
* @has_constraint: is there a contraint on the group of events? In which case
* the events won't be grouped.
* @metric_events: out argument, null terminated array of evsel's associated
* with the metric.
* @evlist_used: in/out argument, bitmap tracking which evlist events are used.
* @return the first metric event or NULL on failure.
*/
static struct evsel *find_evsel_group(struct evlist *perf_evlist,
struct expr_parse_ctx *pctx,
bool metric_no_merge,
bool has_constraint,
struct evsel **metric_events,
unsigned long *evlist_used)
{
struct evsel *ev, *current_leader = NULL;
struct expr_id_data *val_ptr;
int i = 0, matched_events = 0, events_to_match;
const int idnum = (int)hashmap__size(&pctx->ids);
/* duration_time is grouped separately. */
if (!has_constraint &&
hashmap__find(&pctx->ids, "duration_time", (void **)&val_ptr))
events_to_match = idnum - 1;
else
events_to_match = idnum;
evlist__for_each_entry (perf_evlist, ev) {
/*
* Events with a constraint aren't grouped and match the first
* events available.
*/
if (has_constraint && ev->weak_group)
continue;
/* Ignore event if already used and merging is disabled. */
if (metric_no_merge && test_bit(ev->idx, evlist_used))
continue;
if (!has_constraint && ev->leader != current_leader) {
/*
* Start of a new group, discard the whole match and
* start again.
*/
matched_events = 0;
memset(metric_events, 0,
sizeof(struct evsel *) * idnum);
current_leader = ev->leader;
}
if (hashmap__find(&pctx->ids, ev->name, (void **)&val_ptr)) {
if (has_constraint) {
/*
* Events aren't grouped, ensure the same event
* isn't matched from two groups.
*/
for (i = 0; i < matched_events; i++) {
if (!strcmp(ev->name,
metric_events[i]->name)) {
break;
}
}
if (i != matched_events)
continue;
}
metric_events[matched_events++] = ev;
}
if (matched_events == events_to_match)
break;
}
if (events_to_match != idnum) {
/* Add the first duration_time. */
evlist__for_each_entry(perf_evlist, ev) {
if (!strcmp(ev->name, "duration_time")) {
metric_events[matched_events++] = ev;
break;
}
}
}
if (matched_events != idnum) {
/* Not whole match */
return NULL;
}
metric_events[idnum] = NULL;
for (i = 0; i < idnum; i++) {
ev = metric_events[i];
ev->metric_leader = ev;
set_bit(ev->idx, evlist_used);
}
return metric_events[0];
}
static int metricgroup__setup_events(struct list_head *groups,
bool metric_no_merge,
struct evlist *perf_evlist,
struct rblist *metric_events_list)
{
struct metric_event *me;
struct metric_expr *expr;
int i = 0;
int ret = 0;
struct egroup *eg;
struct evsel *evsel, *tmp;
unsigned long *evlist_used;
evlist_used = bitmap_alloc(perf_evlist->core.nr_entries);
if (!evlist_used)
return -ENOMEM;
list_for_each_entry (eg, groups, nd) {
struct evsel **metric_events;
struct metric_ref *metric_refs = NULL;
metric_events = calloc(sizeof(void *),
hashmap__size(&eg->pctx.ids) + 1);
if (!metric_events) {
ret = -ENOMEM;
break;
}
evsel = find_evsel_group(perf_evlist, &eg->pctx,
metric_no_merge,
eg->has_constraint, metric_events,
evlist_used);
if (!evsel) {
pr_debug("Cannot resolve %s: %s\n",
eg->metric_name, eg->metric_expr);
free(metric_events);
continue;
}
for (i = 0; metric_events[i]; i++)
metric_events[i]->collect_stat = true;
me = metricgroup__lookup(metric_events_list, evsel, true);
if (!me) {
ret = -ENOMEM;
free(metric_events);
break;
}
expr = malloc(sizeof(struct metric_expr));
if (!expr) {
ret = -ENOMEM;
free(metric_events);
break;
}
/*
* Collect and store collected nested expressions
* for metric processing.
*/
if (eg->metric_refs_cnt) {
struct metric_ref_node *ref;
metric_refs = zalloc(sizeof(struct metric_ref) * (eg->metric_refs_cnt + 1));
if (!metric_refs) {
ret = -ENOMEM;
free(metric_events);
break;
}
i = 0;
list_for_each_entry(ref, &eg->metric_refs, list) {
/*
* Intentionally passing just const char pointers,
* originally from 'struct pmu_event' object.
* We don't need to change them, so there's no
* need to create our own copy.
*/
metric_refs[i].metric_name = ref->metric_name;
metric_refs[i].metric_expr = ref->metric_expr;
i++;
}
};
expr->metric_refs = metric_refs;
expr->metric_expr = eg->metric_expr;
expr->metric_name = eg->metric_name;
expr->metric_unit = eg->metric_unit;
expr->metric_events = metric_events;
expr->runtime = eg->runtime;
list_add(&expr->nd, &me->head);
}
evlist__for_each_entry_safe(perf_evlist, tmp, evsel) {
if (!test_bit(evsel->idx, evlist_used)) {
evlist__remove(perf_evlist, evsel);
evsel__delete(evsel);
}
}
bitmap_free(evlist_used);
return ret;
}
static bool match_metric(const char *n, const char *list)
{
int len;
char *m;
if (!list)
return false;
if (!strcmp(list, "all"))
return true;
if (!n)
return !strcasecmp(list, "No_group");
len = strlen(list);
m = strcasestr(n, list);
if (!m)
return false;
if ((m == n || m[-1] == ';' || m[-1] == ' ') &&
(m[len] == 0 || m[len] == ';'))
return true;
return false;
}
struct mep {
struct rb_node nd;
const char *name;
struct strlist *metrics;
};
static int mep_cmp(struct rb_node *rb_node, const void *entry)
{
struct mep *a = container_of(rb_node, struct mep, nd);
struct mep *b = (struct mep *)entry;
return strcmp(a->name, b->name);
}
static struct rb_node *mep_new(struct rblist *rl __maybe_unused,
const void *entry)
{
struct mep *me = malloc(sizeof(struct mep));
if (!me)
return NULL;
memcpy(me, entry, sizeof(struct mep));
me->name = strdup(me->name);
if (!me->name)
goto out_me;
me->metrics = strlist__new(NULL, NULL);
if (!me->metrics)
goto out_name;
return &me->nd;
out_name:
zfree(&me->name);
out_me:
free(me);
return NULL;
}
static struct mep *mep_lookup(struct rblist *groups, const char *name)
{
struct rb_node *nd;
struct mep me = {
.name = name
};
nd = rblist__find(groups, &me);
if (nd)
return container_of(nd, struct mep, nd);
rblist__add_node(groups, &me);
nd = rblist__find(groups, &me);
if (nd)
return container_of(nd, struct mep, nd);
return NULL;
}
static void mep_delete(struct rblist *rl __maybe_unused,
struct rb_node *nd)
{
struct mep *me = container_of(nd, struct mep, nd);
strlist__delete(me->metrics);
zfree(&me->name);
free(me);
}
static void metricgroup__print_strlist(struct strlist *metrics, bool raw)
{
struct str_node *sn;
int n = 0;
strlist__for_each_entry (sn, metrics) {
if (raw)
printf("%s%s", n > 0 ? " " : "", sn->s);
else
printf(" %s\n", sn->s);
n++;
}
if (raw)
putchar('\n');
}
void metricgroup__print(bool metrics, bool metricgroups, char *filter,
bool raw, bool details)
{
struct pmu_events_map *map = perf_pmu__find_map(NULL);
struct pmu_event *pe;
int i;
struct rblist groups;
struct rb_node *node, *next;
struct strlist *metriclist = NULL;
if (!map)
return;
if (!metricgroups) {
metriclist = strlist__new(NULL, NULL);
if (!metriclist)
return;
}
rblist__init(&groups);
groups.node_new = mep_new;
groups.node_cmp = mep_cmp;
groups.node_delete = mep_delete;
for (i = 0; ; i++) {
const char *g;
pe = &map->table[i];
if (!pe->name && !pe->metric_group && !pe->metric_name)
break;
if (!pe->metric_expr)
continue;
g = pe->metric_group;
if (!g && pe->metric_name) {
if (pe->name)
continue;
g = "No_group";
}
if (g) {
char *omg;
char *mg = strdup(g);
if (!mg)
return;
omg = mg;
while ((g = strsep(&mg, ";")) != NULL) {
struct mep *me;
char *s;
g = skip_spaces(g);
if (*g == 0)
g = "No_group";
if (filter && !strstr(g, filter))
continue;
if (raw)
s = (char *)pe->metric_name;
else {
if (asprintf(&s, "%s\n%*s%s]",
pe->metric_name, 8, "[", pe->desc) < 0)
return;
if (details) {
if (asprintf(&s, "%s\n%*s%s]",
s, 8, "[", pe->metric_expr) < 0)
return;
}
}
if (!s)
continue;
if (!metricgroups) {
strlist__add(metriclist, s);
} else {
me = mep_lookup(&groups, g);
if (!me)
continue;
strlist__add(me->metrics, s);
}
}
free(omg);
}
}
if (metricgroups && !raw)
printf("\nMetric Groups:\n\n");
else if (metrics && !raw)
printf("\nMetrics:\n\n");
for (node = rb_first_cached(&groups.entries); node; node = next) {
struct mep *me = container_of(node, struct mep, nd);
if (metricgroups)
printf("%s%s%s", me->name, metrics && !raw ? ":" : "", raw ? " " : "\n");
if (metrics)
metricgroup__print_strlist(me->metrics, raw);
next = rb_next(node);
rblist__remove_node(&groups, node);
}
if (!metricgroups)
metricgroup__print_strlist(metriclist, raw);
strlist__delete(metriclist);
}
static void metricgroup__add_metric_weak_group(struct strbuf *events,
struct expr_parse_ctx *ctx)
{
struct hashmap_entry *cur;
size_t bkt;
bool no_group = true, has_duration = false;
hashmap__for_each_entry((&ctx->ids), cur, bkt) {
pr_debug("found event %s\n", (const char *)cur->key);
/*
* Duration time maps to a software event and can make
* groups not count. Always use it outside a
* group.
*/
if (!strcmp(cur->key, "duration_time")) {
has_duration = true;
continue;
}
strbuf_addf(events, "%s%s",
no_group ? "{" : ",",
(const char *)cur->key);
no_group = false;
}
if (!no_group) {
strbuf_addf(events, "}:W");
if (has_duration)
strbuf_addf(events, ",duration_time");
} else if (has_duration)
strbuf_addf(events, "duration_time");
}
static void metricgroup__add_metric_non_group(struct strbuf *events,
struct expr_parse_ctx *ctx)
{
struct hashmap_entry *cur;
size_t bkt;
bool first = true;
hashmap__for_each_entry((&ctx->ids), cur, bkt) {
if (!first)
strbuf_addf(events, ",");
strbuf_addf(events, "%s", (const char *)cur->key);
first = false;
}
}
static void metricgroup___watchdog_constraint_hint(const char *name, bool foot)
{
static bool violate_nmi_constraint;
if (!foot) {
pr_warning("Splitting metric group %s into standalone metrics.\n", name);
violate_nmi_constraint = true;
return;
}
if (!violate_nmi_constraint)
return;
pr_warning("Try disabling the NMI watchdog to comply NO_NMI_WATCHDOG metric constraint:\n"
" echo 0 > /proc/sys/kernel/nmi_watchdog\n"
" perf stat ...\n"
" echo 1 > /proc/sys/kernel/nmi_watchdog\n");
}
static bool metricgroup__has_constraint(struct pmu_event *pe)
{
if (!pe->metric_constraint)
return false;
if (!strcmp(pe->metric_constraint, "NO_NMI_WATCHDOG") &&
sysctl__nmi_watchdog_enabled()) {
metricgroup___watchdog_constraint_hint(pe->metric_name, false);
return true;
}
return false;
}
int __weak arch_get_runtimeparam(void)
{
return 1;
}
static int __add_metric(struct list_head *group_list,
struct pmu_event *pe,
bool metric_no_group,
int runtime,
struct egroup **egp,
struct expr_id *parent,
struct expr_ids *ids)
{
struct metric_ref_node *ref;
struct egroup *eg;
if (*egp == NULL) {
/*
* We got in here for the parent group,
* allocate it and put it on the list.
*/
eg = zalloc(sizeof(*eg));
if (!eg)
return -ENOMEM;
expr__ctx_init(&eg->pctx);
eg->metric_name = pe->metric_name;
eg->metric_expr = pe->metric_expr;
eg->metric_unit = pe->unit;
eg->runtime = runtime;
eg->has_constraint = metric_no_group || metricgroup__has_constraint(pe);
INIT_LIST_HEAD(&eg->metric_refs);
eg->metric_refs_cnt = 0;
*egp = eg;
parent = expr_ids__alloc(ids);
if (!parent) {
free(eg);
return -EINVAL;
}
parent->id = strdup(pe->metric_name);
if (!parent->id) {
free(eg);
return -ENOMEM;
}
} else {
/*
* We got here for the referenced metric, via the
* recursive metricgroup__add_metric call, add
* it to the parent group.
*/
eg = *egp;
ref = malloc(sizeof(*ref));
if (!ref)
return -ENOMEM;
/*
* Intentionally passing just const char pointers,
* from 'pe' object, so they never go away. We don't
* need to change them, so there's no need to create
* our own copy.
*/
ref->metric_name = pe->metric_name;
ref->metric_expr = pe->metric_expr;
list_add(&ref->list, &eg->metric_refs);
eg->metric_refs_cnt++;
}
/* Force all found IDs in metric to have us as parent ID. */
WARN_ON_ONCE(!parent);
eg->pctx.parent = parent;
/*
* For both the parent and referenced metrics, we parse
* all the metric's IDs and add it to the parent context.
*/
if (expr__find_other(pe->metric_expr, NULL, &eg->pctx, runtime) < 0) {
expr__ctx_clear(&eg->pctx);
free(eg);
return -EINVAL;
}
/*
* We add new group only in the 'parent' call,
* so bail out for referenced metric case.
*/
if (eg->metric_refs_cnt)
return 0;
if (list_empty(group_list))
list_add(&eg->nd, group_list);
else {
struct list_head *pos;
/* Place the largest groups at the front. */
list_for_each_prev(pos, group_list) {
struct egroup *old = list_entry(pos, struct egroup, nd);
if (hashmap__size(&eg->pctx.ids) <=
hashmap__size(&old->pctx.ids))
break;
}
list_add(&eg->nd, pos);
}
return 0;
}
#define map_for_each_event(__pe, __idx, __map) \
for (__idx = 0, __pe = &__map->table[__idx]; \
__pe->name || __pe->metric_group || __pe->metric_name; \
__pe = &__map->table[++__idx])
#define map_for_each_metric(__pe, __idx, __map, __metric) \
map_for_each_event(__pe, __idx, __map) \
if (__pe->metric_expr && \
(match_metric(__pe->metric_group, __metric) || \
match_metric(__pe->metric_name, __metric)))
static struct pmu_event *find_metric(const char *metric, struct pmu_events_map *map)
{
struct pmu_event *pe;
int i;
map_for_each_event(pe, i, map) {
if (match_metric(pe->metric_name, metric))
return pe;
}
return NULL;
}
static int recursion_check(struct egroup *eg, const char *id, struct expr_id **parent,
struct expr_ids *ids)
{
struct expr_id_data *data;
struct expr_id *p;
int ret;
/*
* We get the parent referenced by 'id' argument and
* traverse through all the parent object IDs to check
* if we already processed 'id', if we did, it's recursion
* and we fail.
*/
ret = expr__get_id(&eg->pctx, id, &data);
if (ret)
return ret;
p = data->parent;
while (p->parent) {
if (!strcmp(p->id, id)) {
pr_err("failed: recursion detected for %s\n", id);
return -1;
}
p = p->parent;
}
/*
* If we are over the limit of static entris, the metric
* is too difficult/nested to process, fail as well.
*/
p = expr_ids__alloc(ids);
if (!p) {
pr_err("failed: too many nested metrics\n");
return -EINVAL;
}
p->id = strdup(id);
p->parent = data->parent;
*parent = p;
return p->id ? 0 : -ENOMEM;
}
static int add_metric(struct list_head *group_list,
struct pmu_event *pe,
bool metric_no_group,
struct egroup **egp,
struct expr_id *parent,
struct expr_ids *ids);
static int __resolve_metric(struct egroup *eg,
bool metric_no_group,
struct list_head *group_list,
struct pmu_events_map *map,
struct expr_ids *ids)
{
struct hashmap_entry *cur;
size_t bkt;
bool all;
int ret;
/*
* Iterate all the parsed IDs and if there's metric,
* add it to the context.
*/
do {
all = true;
hashmap__for_each_entry((&eg->pctx.ids), cur, bkt) {
struct expr_id *parent;
struct pmu_event *pe;
pe = find_metric(cur->key, map);
if (!pe)
continue;
ret = recursion_check(eg, cur->key, &parent, ids);
if (ret)
return ret;
all = false;
/* The metric key itself needs to go out.. */
expr__del_id(&eg->pctx, cur->key);
/* ... and it gets resolved to the parent context. */
ret = add_metric(group_list, pe, metric_no_group, &eg, parent, ids);
if (ret)
return ret;
/*
* We added new metric to hashmap, so we need
* to break the iteration and start over.
*/
break;
}
} while (!all);
return 0;
}
static int resolve_metric(bool metric_no_group,
struct list_head *metric_list,
struct pmu_events_map *map,
struct expr_ids *ids)
{
struct egroup *eg;
int err;
list_for_each_entry(eg, metric_list, nd) {
err = __resolve_metric(eg, metric_no_group, metric_list, map, ids);
if (err)
return err;
}
return 0;
}
static int add_metric(struct list_head *group_list,
struct pmu_event *pe,
bool metric_no_group,
struct egroup **egp,
struct expr_id *parent,
struct expr_ids *ids)
{
struct egroup *orig = *egp;
int ret = 0;
pr_debug("metric expr %s for %s\n", pe->metric_expr, pe->metric_name);
if (!strstr(pe->metric_expr, "?")) {
ret = __add_metric(group_list, pe, metric_no_group, 1, egp, parent, ids);
} else {
int j, count;
count = arch_get_runtimeparam();
/* This loop is added to create multiple
* events depend on count value and add
* those events to group_list.
*/
for (j = 0; j < count && !ret; j++, *egp = orig)
ret = __add_metric(group_list, pe, metric_no_group, j, egp, parent, ids);
}
return ret;
}
static int metricgroup__add_metric(const char *metric, bool metric_no_group,
struct strbuf *events,
struct list_head *group_list,
struct pmu_events_map *map)
{
struct expr_ids ids = { .cnt = 0, };
struct pmu_event *pe;
struct egroup *eg;
LIST_HEAD(list);
int i, ret;
bool has_match = false;
map_for_each_metric(pe, i, map, metric) {
has_match = true;
eg = NULL;
ret = add_metric(&list, pe, metric_no_group, &eg, NULL, &ids);
if (ret)
return ret;
/*
* Process any possible referenced metrics
* included in the expression.
*/
ret = resolve_metric(metric_no_group,
&list, map, &ids);
if (ret)
return ret;
}
/* End of pmu events. */
if (!has_match)
return -EINVAL;
list_for_each_entry(eg, &list, nd) {
if (events->len > 0)
strbuf_addf(events, ",");
if (eg->has_constraint) {
metricgroup__add_metric_non_group(events,
&eg->pctx);
} else {
metricgroup__add_metric_weak_group(events,
&eg->pctx);
}
}
list_splice(&list, group_list);
expr_ids__exit(&ids);
return 0;
}
static int metricgroup__add_metric_list(const char *list, bool metric_no_group,
struct strbuf *events,
struct list_head *group_list,
struct pmu_events_map *map)
{
char *llist, *nlist, *p;
int ret = -EINVAL;
nlist = strdup(list);
if (!nlist)
return -ENOMEM;
llist = nlist;
strbuf_init(events, 100);
strbuf_addf(events, "%s", "");
while ((p = strsep(&llist, ",")) != NULL) {
ret = metricgroup__add_metric(p, metric_no_group, events,
group_list, map);
if (ret == -EINVAL) {
fprintf(stderr, "Cannot find metric or group `%s'\n",
p);
break;
}
}
free(nlist);
if (!ret)
metricgroup___watchdog_constraint_hint(NULL, true);
return ret;
}
static void egroup__free_refs(struct egroup *egroup)
{
struct metric_ref_node *ref, *tmp;
list_for_each_entry_safe(ref, tmp, &egroup->metric_refs, list) {
list_del(&ref->list);
free(ref);
}
}
static void metricgroup__free_egroups(struct list_head *group_list)
{
struct egroup *eg, *egtmp;
list_for_each_entry_safe (eg, egtmp, group_list, nd) {
egroup__free_refs(eg);
expr__ctx_clear(&eg->pctx);
list_del_init(&eg->nd);
free(eg);
}
}
static int parse_groups(struct evlist *perf_evlist, const char *str,
bool metric_no_group,
bool metric_no_merge,
struct perf_pmu *fake_pmu,
struct rblist *metric_events,
struct pmu_events_map *map)
{
struct parse_events_error parse_error;
struct strbuf extra_events;
LIST_HEAD(group_list);
int ret;
if (metric_events->nr_entries == 0)
metricgroup__rblist_init(metric_events);
ret = metricgroup__add_metric_list(str, metric_no_group,
&extra_events, &group_list, map);
if (ret)
return ret;
pr_debug("adding %s\n", extra_events.buf);
bzero(&parse_error, sizeof(parse_error));
ret = __parse_events(perf_evlist, extra_events.buf, &parse_error, fake_pmu);
if (ret) {
parse_events_print_error(&parse_error, extra_events.buf);
goto out;
}
strbuf_release(&extra_events);
ret = metricgroup__setup_events(&group_list, metric_no_merge,
perf_evlist, metric_events);
out:
metricgroup__free_egroups(&group_list);
return ret;
}
int metricgroup__parse_groups(const struct option *opt,
const char *str,
bool metric_no_group,
bool metric_no_merge,
struct rblist *metric_events)
{
struct evlist *perf_evlist = *(struct evlist **)opt->value;
struct pmu_events_map *map = perf_pmu__find_map(NULL);
if (!map)
return 0;
return parse_groups(perf_evlist, str, metric_no_group,
metric_no_merge, NULL, metric_events, map);
}
int metricgroup__parse_groups_test(struct evlist *evlist,
struct pmu_events_map *map,
const char *str,
bool metric_no_group,
bool metric_no_merge,
struct rblist *metric_events)
{
return parse_groups(evlist, str, metric_no_group,
metric_no_merge, &perf_pmu__fake, metric_events, map);
}
bool metricgroup__has_metric(const char *metric)
{
struct pmu_events_map *map = perf_pmu__find_map(NULL);
struct pmu_event *pe;
int i;
if (!map)
return false;
for (i = 0; ; i++) {
pe = &map->table[i];
if (!pe->name && !pe->metric_group && !pe->metric_name)
break;
if (!pe->metric_expr)
continue;
if (match_metric(pe->metric_name, metric))
return true;
}
return false;
}