perf header: Support HYBRID_CPU_PMU_CAPS feature

Perf has supported the CPU_PMU_CAPS feature to display a list of CPU PMU
capabilities. But on a hybrid platform, it may have several CPU PMUs (such
as "cpu_core" and "cpu_atom"). The CPU_PMU_CAPS feature is hard to extend
to support multiple CPU PMUs well if it needs to be compatible for the case
of old perf data file + new perf tool.

So for better compatibility we now create a new feature HYBRID_CPU_PMU_CAPS
in the header.

For the perf.data generated on hybrid platform,

  root@otcpl-adl-s-2:~# perf report --header-only -I

  # cpu_core pmu capabilities: branches=32, max_precise=3, pmu_name=alderlake_hybrid
  # cpu_atom pmu capabilities: branches=32, max_precise=3, pmu_name=alderlake_hybrid
  # missing features: TRACING_DATA BRANCH_STACK GROUP_DESC AUXTRACE STAT CLOCKID DIR_FORMAT COMPRESSED CPU_PMU_CAPS CLOCK_DATA

For the perf.data generated on non-hybrid platform

  root@kbl-ppc:~# perf report --header-only -I

  # cpu pmu capabilities: branches=32, max_precise=3, pmu_name=skylake
  # missing features: TRACING_DATA BRANCH_STACK GROUP_DESC AUXTRACE STAT CLOCKID DIR_FORMAT COMPRESSED CLOCK_DATA HYBRID_TOPOLOGY HYBRID_CPU_PMU_CAPS

Signed-off-by: Jin Yao <yao.jin@linux.intel.com>
Acked-by: Jiri Olsa <jolsa@redhat.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Jin Yao <yao.jin@intel.com>
Cc: Kan Liang <kan.liang@linux.intel.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Link: http://lore.kernel.org/lkml/20210514122948.9472-3-yao.jin@linux.intel.com
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
Jin Yao 2021-05-14 20:29:48 +08:00 коммит произвёл Arnaldo Carvalho de Melo
Родитель f7d74ce32f
Коммит e119083bab
5 изменённых файлов: 175 добавлений и 19 удалений

Просмотреть файл

@ -419,6 +419,22 @@ Example:
cpu_core cpu list : 0-15
cpu_atom cpu list : 16-23
HEADER_HYBRID_CPU_PMU_CAPS = 31,
A list of hybrid CPU PMU capabilities.
struct {
u32 nr_pmu;
struct {
u32 nr_cpu_pmu_caps;
{
char name[];
char value[];
} [nr_cpu_pmu_caps];
char pmu_name[];
} [nr_pmu];
};
other bits are reserved and should ignored for now
HEADER_FEAT_BITS = 256,

Просмотреть файл

@ -208,6 +208,12 @@ void perf_env__exit(struct perf_env *env)
zfree(&env->hybrid_nodes[i].cpus);
}
zfree(&env->hybrid_nodes);
for (i = 0; i < env->nr_hybrid_cpc_nodes; i++) {
zfree(&env->hybrid_cpc_nodes[i].cpu_pmu_caps);
zfree(&env->hybrid_cpc_nodes[i].pmu_name);
}
zfree(&env->hybrid_cpc_nodes);
}
void perf_env__init(struct perf_env *env __maybe_unused)

Просмотреть файл

@ -42,6 +42,13 @@ struct hybrid_node {
char *cpus;
};
struct hybrid_cpc_node {
int nr_cpu_pmu_caps;
unsigned int max_branches;
char *cpu_pmu_caps;
char *pmu_name;
};
struct perf_env {
char *hostname;
char *os_release;
@ -65,6 +72,7 @@ struct perf_env {
int nr_groups;
int nr_cpu_pmu_caps;
int nr_hybrid_nodes;
int nr_hybrid_cpc_nodes;
char *cmdline;
const char **cmdline_argv;
char *sibling_cores;
@ -84,6 +92,7 @@ struct perf_env {
struct memory_node *memory_nodes;
unsigned long long memory_bsize;
struct hybrid_node *hybrid_nodes;
struct hybrid_cpc_node *hybrid_cpc_nodes;
#ifdef HAVE_LIBBPF_SUPPORT
/*
* bpf_info_lock protects bpf rbtrees. This is needed because the

Просмотреть файл

@ -49,6 +49,7 @@
#include "cputopo.h"
#include "bpf-event.h"
#include "clockid.h"
#include "pmu-hybrid.h"
#include <linux/ctype.h>
#include <internal/lib.h>
@ -1459,18 +1460,14 @@ static int write_compressed(struct feat_fd *ff __maybe_unused,
return do_write(ff, &(ff->ph->env.comp_mmap_len), sizeof(ff->ph->env.comp_mmap_len));
}
static int write_cpu_pmu_caps(struct feat_fd *ff,
struct evlist *evlist __maybe_unused)
static int write_per_cpu_pmu_caps(struct feat_fd *ff, struct perf_pmu *pmu,
bool write_pmu)
{
struct perf_pmu *cpu_pmu = perf_pmu__find("cpu");
struct perf_pmu_caps *caps = NULL;
int nr_caps;
int ret;
if (!cpu_pmu)
return -ENOENT;
nr_caps = perf_pmu__caps_parse(cpu_pmu);
nr_caps = perf_pmu__caps_parse(pmu);
if (nr_caps < 0)
return nr_caps;
@ -1478,7 +1475,7 @@ static int write_cpu_pmu_caps(struct feat_fd *ff,
if (ret < 0)
return ret;
list_for_each_entry(caps, &cpu_pmu->caps, list) {
list_for_each_entry(caps, &pmu->caps, list) {
ret = do_write_string(ff, caps->name);
if (ret < 0)
return ret;
@ -1488,7 +1485,47 @@ static int write_cpu_pmu_caps(struct feat_fd *ff,
return ret;
}
if (write_pmu) {
ret = do_write_string(ff, pmu->name);
if (ret < 0)
return ret;
}
return ret;
}
static int write_cpu_pmu_caps(struct feat_fd *ff,
struct evlist *evlist __maybe_unused)
{
struct perf_pmu *cpu_pmu = perf_pmu__find("cpu");
if (!cpu_pmu)
return -ENOENT;
return write_per_cpu_pmu_caps(ff, cpu_pmu, false);
}
static int write_hybrid_cpu_pmu_caps(struct feat_fd *ff,
struct evlist *evlist __maybe_unused)
{
struct perf_pmu *pmu;
u32 nr_pmu = perf_pmu__hybrid_pmu_num();
int ret;
if (nr_pmu == 0)
return -ENOENT;
ret = do_write(ff, &nr_pmu, sizeof(nr_pmu));
if (ret < 0)
return ret;
perf_pmu__for_each_hybrid_pmu(pmu) {
ret = write_per_cpu_pmu_caps(ff, pmu, true);
if (ret < 0)
return ret;
}
return 0;
}
static void print_hostname(struct feat_fd *ff, FILE *fp)
@ -1962,18 +1999,28 @@ static void print_compressed(struct feat_fd *ff, FILE *fp)
ff->ph->env.comp_level, ff->ph->env.comp_ratio);
}
static void print_cpu_pmu_caps(struct feat_fd *ff, FILE *fp)
static void print_per_cpu_pmu_caps(FILE *fp, int nr_caps, char *cpu_pmu_caps,
char *pmu_name)
{
const char *delimiter = "# cpu pmu capabilities: ";
u32 nr_caps = ff->ph->env.nr_cpu_pmu_caps;
char *str;
const char *delimiter;
char *str, buf[128];
if (!nr_caps) {
if (!pmu_name)
fprintf(fp, "# cpu pmu capabilities: not available\n");
else
fprintf(fp, "# %s pmu capabilities: not available\n", pmu_name);
return;
}
str = ff->ph->env.cpu_pmu_caps;
if (!pmu_name)
scnprintf(buf, sizeof(buf), "# cpu pmu capabilities: ");
else
scnprintf(buf, sizeof(buf), "# %s pmu capabilities: ", pmu_name);
delimiter = buf;
str = cpu_pmu_caps;
while (nr_caps--) {
fprintf(fp, "%s%s", delimiter, str);
delimiter = ", ";
@ -1983,6 +2030,24 @@ static void print_cpu_pmu_caps(struct feat_fd *ff, FILE *fp)
fprintf(fp, "\n");
}
static void print_cpu_pmu_caps(struct feat_fd *ff, FILE *fp)
{
print_per_cpu_pmu_caps(fp, ff->ph->env.nr_cpu_pmu_caps,
ff->ph->env.cpu_pmu_caps, NULL);
}
static void print_hybrid_cpu_pmu_caps(struct feat_fd *ff, FILE *fp)
{
struct hybrid_cpc_node *n;
for (int i = 0; i < ff->ph->env.nr_hybrid_cpc_nodes; i++) {
n = &ff->ph->env.hybrid_cpc_nodes[i];
print_per_cpu_pmu_caps(fp, n->nr_cpu_pmu_caps,
n->cpu_pmu_caps,
n->pmu_name);
}
}
static void print_pmu_mappings(struct feat_fd *ff, FILE *fp)
{
const char *delimiter = "# pmu mappings: ";
@ -3088,8 +3153,9 @@ static int process_compressed(struct feat_fd *ff,
return 0;
}
static int process_cpu_pmu_caps(struct feat_fd *ff,
void *data __maybe_unused)
static int process_per_cpu_pmu_caps(struct feat_fd *ff, int *nr_cpu_pmu_caps,
char **cpu_pmu_caps,
unsigned int *max_branches)
{
char *name, *value;
struct strbuf sb;
@ -3103,7 +3169,7 @@ static int process_cpu_pmu_caps(struct feat_fd *ff,
return 0;
}
ff->ph->env.nr_cpu_pmu_caps = nr_caps;
*nr_cpu_pmu_caps = nr_caps;
if (strbuf_init(&sb, 128) < 0)
return -1;
@ -3125,12 +3191,12 @@ static int process_cpu_pmu_caps(struct feat_fd *ff,
goto free_value;
if (!strcmp(name, "branches"))
ff->ph->env.max_branches = atoi(value);
*max_branches = atoi(value);
free(value);
free(name);
}
ff->ph->env.cpu_pmu_caps = strbuf_detach(&sb, NULL);
*cpu_pmu_caps = strbuf_detach(&sb, NULL);
return 0;
free_value:
@ -3142,6 +3208,63 @@ error:
return -1;
}
static int process_cpu_pmu_caps(struct feat_fd *ff,
void *data __maybe_unused)
{
return process_per_cpu_pmu_caps(ff, &ff->ph->env.nr_cpu_pmu_caps,
&ff->ph->env.cpu_pmu_caps,
&ff->ph->env.max_branches);
}
static int process_hybrid_cpu_pmu_caps(struct feat_fd *ff,
void *data __maybe_unused)
{
struct hybrid_cpc_node *nodes;
u32 nr_pmu, i;
int ret;
if (do_read_u32(ff, &nr_pmu))
return -1;
if (!nr_pmu) {
pr_debug("hybrid cpu pmu capabilities not available\n");
return 0;
}
nodes = zalloc(sizeof(*nodes) * nr_pmu);
if (!nodes)
return -ENOMEM;
for (i = 0; i < nr_pmu; i++) {
struct hybrid_cpc_node *n = &nodes[i];
ret = process_per_cpu_pmu_caps(ff, &n->nr_cpu_pmu_caps,
&n->cpu_pmu_caps,
&n->max_branches);
if (ret)
goto err;
n->pmu_name = do_read_string(ff);
if (!n->pmu_name) {
ret = -1;
goto err;
}
}
ff->ph->env.nr_hybrid_cpc_nodes = nr_pmu;
ff->ph->env.hybrid_cpc_nodes = nodes;
return 0;
err:
for (i = 0; i < nr_pmu; i++) {
free(nodes[i].cpu_pmu_caps);
free(nodes[i].pmu_name);
}
free(nodes);
return ret;
}
#define FEAT_OPR(n, func, __full_only) \
[HEADER_##n] = { \
.name = __stringify(n), \
@ -3204,6 +3327,7 @@ const struct perf_header_feature_ops feat_ops[HEADER_LAST_FEATURE] = {
FEAT_OPR(CPU_PMU_CAPS, cpu_pmu_caps, false),
FEAT_OPR(CLOCK_DATA, clock_data, false),
FEAT_OPN(HYBRID_TOPOLOGY, hybrid_topology, true),
FEAT_OPR(HYBRID_CPU_PMU_CAPS, hybrid_cpu_pmu_caps, false),
};
struct header_print_data {

Просмотреть файл

@ -46,6 +46,7 @@ enum {
HEADER_CPU_PMU_CAPS,
HEADER_CLOCK_DATA,
HEADER_HYBRID_TOPOLOGY,
HEADER_HYBRID_CPU_PMU_CAPS,
HEADER_LAST_FEATURE,
HEADER_FEAT_BITS = 256,
};