perf lock: Use BPF for lock contention analysis
Add -b/--use-bpf option to use BPF to collect lock contention stats. For simplicity it now runs system-wide and requires C-c to stop. Upcoming changes will add the usual filtering. $ sudo perf lock con -b ^C contended total wait max wait avg wait type caller 42 192.67 us 13.64 us 4.59 us spinlock queue_work_on+0x20 23 85.54 us 10.28 us 3.72 us spinlock worker_thread+0x14a 6 13.92 us 6.51 us 2.32 us mutex kernfs_iop_permission+0x30 3 11.59 us 10.04 us 3.86 us mutex kernfs_dop_revalidate+0x3c 1 7.52 us 7.52 us 7.52 us spinlock kthread+0x115 1 7.24 us 7.24 us 7.24 us rwlock:W sys_epoll_wait+0x148 2 7.08 us 3.99 us 3.54 us spinlock delayed_work_timer_fn+0x1b 1 6.41 us 6.41 us 6.41 us spinlock idle_balance+0xa06 2 2.50 us 1.83 us 1.25 us mutex kernfs_iop_lookup+0x2f 1 1.71 us 1.71 us 1.71 us mutex kernfs_iop_getattr+0x2c Signed-off-by: Namhyung Kim <namhyung@kernel.org> Cc: Blake Jones <blakejones@google.com> Cc: Boqun Feng <boqun.feng@gmail.com> Cc: Davidlohr Bueso <dave@stgolabs.net> Cc: Ian Rogers <irogers@google.com> Cc: Ingo Molnar <mingo@kernel.org> Cc: Jiri Olsa <jolsa@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Stephane Eranian <eranian@google.com> Cc: Waiman Long <longman@redhat.com> Cc: Will Deacon <will@kernel.org> Link: https://lore.kernel.org/r/20220729200756.666106-3-namhyung@kernel.org Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
Родитель
77d54a2cd6
Коммит
407b36f69e
|
@ -123,6 +123,11 @@ CONTENTION OPTIONS
|
|||
--threads::
|
||||
Show per-thread lock contention stat
|
||||
|
||||
-b::
|
||||
--use-bpf::
|
||||
Use BPF program to collect lock contention stats instead of
|
||||
using the input data.
|
||||
|
||||
|
||||
SEE ALSO
|
||||
--------
|
||||
|
|
|
@ -1028,7 +1028,7 @@ SKEL_TMP_OUT := $(abspath $(SKEL_OUT)/.tmp)
|
|||
SKELETONS := $(SKEL_OUT)/bpf_prog_profiler.skel.h
|
||||
SKELETONS += $(SKEL_OUT)/bperf_leader.skel.h $(SKEL_OUT)/bperf_follower.skel.h
|
||||
SKELETONS += $(SKEL_OUT)/bperf_cgroup.skel.h $(SKEL_OUT)/func_latency.skel.h
|
||||
SKELETONS += $(SKEL_OUT)/off_cpu.skel.h
|
||||
SKELETONS += $(SKEL_OUT)/off_cpu.skel.h $(SKEL_OUT)/lock_contention.skel.h
|
||||
SKELETONS += $(SKEL_OUT)/kwork_trace.skel.h
|
||||
|
||||
$(SKEL_TMP_OUT) $(LIBBPF_OUTPUT):
|
||||
|
|
|
@ -10,6 +10,7 @@
|
|||
#include "util/thread.h"
|
||||
#include "util/header.h"
|
||||
#include "util/callchain.h"
|
||||
#include "util/lock-contention.h"
|
||||
|
||||
#include <subcmd/pager.h>
|
||||
#include <subcmd/parse-options.h>
|
||||
|
@ -47,84 +48,11 @@ static struct hlist_head lockhash_table[LOCKHASH_SIZE];
|
|||
#define __lockhashfn(key) hash_long((unsigned long)key, LOCKHASH_BITS)
|
||||
#define lockhashentry(key) (lockhash_table + __lockhashfn((key)))
|
||||
|
||||
struct lock_stat {
|
||||
struct hlist_node hash_entry;
|
||||
struct rb_node rb; /* used for sorting */
|
||||
|
||||
u64 addr; /* address of lockdep_map, used as ID */
|
||||
char *name; /* for strcpy(), we cannot use const */
|
||||
|
||||
unsigned int nr_acquire;
|
||||
unsigned int nr_acquired;
|
||||
unsigned int nr_contended;
|
||||
unsigned int nr_release;
|
||||
|
||||
union {
|
||||
unsigned int nr_readlock;
|
||||
unsigned int flags;
|
||||
};
|
||||
unsigned int nr_trylock;
|
||||
|
||||
/* these times are in nano sec. */
|
||||
u64 avg_wait_time;
|
||||
u64 wait_time_total;
|
||||
u64 wait_time_min;
|
||||
u64 wait_time_max;
|
||||
|
||||
int broken; /* flag of blacklist */
|
||||
int combined;
|
||||
};
|
||||
|
||||
/*
|
||||
* States of lock_seq_stat
|
||||
*
|
||||
* UNINITIALIZED is required for detecting first event of acquire.
|
||||
* As the nature of lock events, there is no guarantee
|
||||
* that the first event for the locks are acquire,
|
||||
* it can be acquired, contended or release.
|
||||
*/
|
||||
#define SEQ_STATE_UNINITIALIZED 0 /* initial state */
|
||||
#define SEQ_STATE_RELEASED 1
|
||||
#define SEQ_STATE_ACQUIRING 2
|
||||
#define SEQ_STATE_ACQUIRED 3
|
||||
#define SEQ_STATE_READ_ACQUIRED 4
|
||||
#define SEQ_STATE_CONTENDED 5
|
||||
|
||||
/*
|
||||
* MAX_LOCK_DEPTH
|
||||
* Imported from include/linux/sched.h.
|
||||
* Should this be synchronized?
|
||||
*/
|
||||
#define MAX_LOCK_DEPTH 48
|
||||
|
||||
/*
|
||||
* struct lock_seq_stat:
|
||||
* Place to put on state of one lock sequence
|
||||
* 1) acquire -> acquired -> release
|
||||
* 2) acquire -> contended -> acquired -> release
|
||||
* 3) acquire (with read or try) -> release
|
||||
* 4) Are there other patterns?
|
||||
*/
|
||||
struct lock_seq_stat {
|
||||
struct list_head list;
|
||||
int state;
|
||||
u64 prev_event_time;
|
||||
u64 addr;
|
||||
|
||||
int read_count;
|
||||
};
|
||||
|
||||
struct thread_stat {
|
||||
struct rb_node rb;
|
||||
|
||||
u32 tid;
|
||||
struct list_head seq_list;
|
||||
};
|
||||
|
||||
static struct rb_root thread_stats;
|
||||
|
||||
static bool combine_locks;
|
||||
static bool show_thread_stats;
|
||||
static bool use_bpf;
|
||||
|
||||
static enum {
|
||||
LOCK_AGGR_ADDR,
|
||||
|
@ -132,31 +60,6 @@ static enum {
|
|||
LOCK_AGGR_CALLER,
|
||||
} aggr_mode = LOCK_AGGR_ADDR;
|
||||
|
||||
/*
|
||||
* CONTENTION_STACK_DEPTH
|
||||
* Number of stack trace entries to find callers
|
||||
*/
|
||||
#define CONTENTION_STACK_DEPTH 8
|
||||
|
||||
/*
|
||||
* CONTENTION_STACK_SKIP
|
||||
* Number of stack trace entries to skip when finding callers.
|
||||
* The first few entries belong to the locking implementation itself.
|
||||
*/
|
||||
#define CONTENTION_STACK_SKIP 3
|
||||
|
||||
/*
|
||||
* flags for lock:contention_begin
|
||||
* Imported from include/trace/events/lock.h.
|
||||
*/
|
||||
#define LCB_F_SPIN (1U << 0)
|
||||
#define LCB_F_READ (1U << 1)
|
||||
#define LCB_F_WRITE (1U << 2)
|
||||
#define LCB_F_RT (1U << 3)
|
||||
#define LCB_F_PERCPU (1U << 4)
|
||||
#define LCB_F_MUTEX (1U << 5)
|
||||
|
||||
|
||||
static u64 sched_text_start;
|
||||
static u64 sched_text_end;
|
||||
static u64 lock_text_start;
|
||||
|
@ -947,7 +850,7 @@ end:
|
|||
return 0;
|
||||
}
|
||||
|
||||
static bool is_lock_function(struct machine *machine, u64 addr)
|
||||
bool is_lock_function(struct machine *machine, u64 addr)
|
||||
{
|
||||
if (!sched_text_start) {
|
||||
struct map *kmap;
|
||||
|
@ -1671,6 +1574,10 @@ out_delete:
|
|||
return err;
|
||||
}
|
||||
|
||||
static void sighandler(int sig __maybe_unused)
|
||||
{
|
||||
}
|
||||
|
||||
static int __cmd_contention(void)
|
||||
{
|
||||
int err = -EINVAL;
|
||||
|
@ -1686,7 +1593,7 @@ static int __cmd_contention(void)
|
|||
.force = force,
|
||||
};
|
||||
|
||||
session = perf_session__new(&data, &eops);
|
||||
session = perf_session__new(use_bpf ? NULL : &data, &eops);
|
||||
if (IS_ERR(session)) {
|
||||
pr_err("Initializing perf session failed\n");
|
||||
return PTR_ERR(session);
|
||||
|
@ -1696,17 +1603,30 @@ static int __cmd_contention(void)
|
|||
symbol_conf.sort_by_name = true;
|
||||
symbol__init(&session->header.env);
|
||||
|
||||
if (!perf_session__has_traces(session, "lock record"))
|
||||
goto out_delete;
|
||||
if (use_bpf) {
|
||||
if (lock_contention_prepare() < 0) {
|
||||
pr_err("lock contention BPF setup failed\n");
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (!evlist__find_evsel_by_str(session->evlist, "lock:contention_begin")) {
|
||||
pr_err("lock contention evsel not found\n");
|
||||
goto out_delete;
|
||||
}
|
||||
signal(SIGINT, sighandler);
|
||||
signal(SIGCHLD, sighandler);
|
||||
signal(SIGTERM, sighandler);
|
||||
} else {
|
||||
if (!perf_session__has_traces(session, "lock record"))
|
||||
goto out_delete;
|
||||
|
||||
if (perf_session__set_tracepoints_handlers(session, contention_tracepoints)) {
|
||||
pr_err("Initializing perf session tracepoint handlers failed\n");
|
||||
goto out_delete;
|
||||
if (!evlist__find_evsel_by_str(session->evlist,
|
||||
"lock:contention_begin")) {
|
||||
pr_err("lock contention evsel not found\n");
|
||||
goto out_delete;
|
||||
}
|
||||
|
||||
if (perf_session__set_tracepoints_handlers(session,
|
||||
contention_tracepoints)) {
|
||||
pr_err("Initializing perf session tracepoint handlers failed\n");
|
||||
goto out_delete;
|
||||
}
|
||||
}
|
||||
|
||||
if (setup_output_field(true, output_fields))
|
||||
|
@ -1720,9 +1640,19 @@ static int __cmd_contention(void)
|
|||
else
|
||||
aggr_mode = LOCK_AGGR_CALLER;
|
||||
|
||||
err = perf_session__process_events(session);
|
||||
if (err)
|
||||
goto out_delete;
|
||||
if (use_bpf) {
|
||||
lock_contention_start();
|
||||
|
||||
/* wait for signal */
|
||||
pause();
|
||||
|
||||
lock_contention_stop();
|
||||
lock_contention_read(&session->machines.host, &lockhash_table[0]);
|
||||
} else {
|
||||
err = perf_session__process_events(session);
|
||||
if (err)
|
||||
goto out_delete;
|
||||
}
|
||||
|
||||
setup_pager();
|
||||
|
||||
|
@ -1730,6 +1660,7 @@ static int __cmd_contention(void)
|
|||
print_contention_result();
|
||||
|
||||
out_delete:
|
||||
lock_contention_finish();
|
||||
perf_session__delete(session);
|
||||
return err;
|
||||
}
|
||||
|
@ -1853,13 +1784,14 @@ int cmd_lock(int argc, const char **argv)
|
|||
OPT_PARENT(lock_options)
|
||||
};
|
||||
|
||||
const struct option contention_options[] = {
|
||||
struct option contention_options[] = {
|
||||
OPT_STRING('k', "key", &sort_key, "wait_total",
|
||||
"key for sorting (contended / wait_total / wait_max / wait_min / avg_wait)"),
|
||||
OPT_STRING('F', "field", &output_fields, "contended,wait_total,wait_max,avg_wait",
|
||||
"output fields (contended / wait_total / wait_max / wait_min / avg_wait)"),
|
||||
OPT_BOOLEAN('t', "threads", &show_thread_stats,
|
||||
"show per-thread lock stats"),
|
||||
OPT_BOOLEAN('b', "use-bpf", &use_bpf, "use BPF program to collect lock contention stats"),
|
||||
OPT_PARENT(lock_options)
|
||||
};
|
||||
|
||||
|
@ -1922,6 +1854,10 @@ int cmd_lock(int argc, const char **argv)
|
|||
sort_key = "wait_total";
|
||||
output_fields = "contended,wait_total,wait_max,avg_wait";
|
||||
|
||||
#ifndef HAVE_BPF_SKEL
|
||||
set_option_nobuild(contention_options, 'b', "use-bpf",
|
||||
"no BUILD_BPF_SKEL=1", false);
|
||||
#endif
|
||||
if (argc) {
|
||||
argc = parse_options(argc, argv, contention_options,
|
||||
contention_usage, 0);
|
||||
|
|
|
@ -149,6 +149,7 @@ perf-$(CONFIG_PERF_BPF_SKEL) += bpf_counter_cgroup.o
|
|||
perf-$(CONFIG_PERF_BPF_SKEL) += bpf_ftrace.o
|
||||
perf-$(CONFIG_PERF_BPF_SKEL) += bpf_off_cpu.o
|
||||
perf-$(CONFIG_PERF_BPF_SKEL) += bpf_kwork.o
|
||||
perf-$(CONFIG_PERF_BPF_SKEL) += bpf_lock_contention.o
|
||||
perf-$(CONFIG_BPF_PROLOGUE) += bpf-prologue.o
|
||||
perf-$(CONFIG_LIBELF) += symbol-elf.o
|
||||
perf-$(CONFIG_LIBELF) += probe-file.o
|
||||
|
|
|
@ -0,0 +1,132 @@
|
|||
// SPDX-License-Identifier: GPL-2.0
|
||||
#include "util/debug.h"
|
||||
#include "util/machine.h"
|
||||
#include "util/map.h"
|
||||
#include "util/symbol.h"
|
||||
#include "util/lock-contention.h"
|
||||
#include <linux/zalloc.h>
|
||||
#include <bpf/bpf.h>
|
||||
|
||||
#include "bpf_skel/lock_contention.skel.h"
|
||||
|
||||
static struct lock_contention_bpf *skel;
|
||||
|
||||
/* should be same as bpf_skel/lock_contention.bpf.c */
|
||||
struct lock_contention_key {
|
||||
u32 stack_id;
|
||||
};
|
||||
|
||||
struct lock_contention_data {
|
||||
u64 total_time;
|
||||
u64 min_time;
|
||||
u64 max_time;
|
||||
u32 count;
|
||||
u32 flags;
|
||||
};
|
||||
|
||||
int lock_contention_prepare(void)
|
||||
{
|
||||
skel = lock_contention_bpf__open();
|
||||
if (!skel) {
|
||||
pr_err("Failed to open lock-contention BPF skeleton\n");
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (lock_contention_bpf__load(skel) < 0) {
|
||||
pr_err("Failed to load lock-contention BPF skeleton\n");
|
||||
return -1;
|
||||
}
|
||||
|
||||
lock_contention_bpf__attach(skel);
|
||||
return 0;
|
||||
}
|
||||
|
||||
int lock_contention_start(void)
|
||||
{
|
||||
skel->bss->enabled = 1;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int lock_contention_stop(void)
|
||||
{
|
||||
skel->bss->enabled = 0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int lock_contention_read(struct machine *machine, struct hlist_head *head)
|
||||
{
|
||||
int fd, stack;
|
||||
u32 prev_key, key;
|
||||
struct lock_contention_data data;
|
||||
struct lock_stat *st;
|
||||
u64 stack_trace[CONTENTION_STACK_DEPTH];
|
||||
|
||||
fd = bpf_map__fd(skel->maps.lock_stat);
|
||||
stack = bpf_map__fd(skel->maps.stacks);
|
||||
|
||||
prev_key = 0;
|
||||
while (!bpf_map_get_next_key(fd, &prev_key, &key)) {
|
||||
struct map *kmap;
|
||||
struct symbol *sym;
|
||||
int idx;
|
||||
|
||||
bpf_map_lookup_elem(fd, &key, &data);
|
||||
st = zalloc(sizeof(*st));
|
||||
if (st == NULL)
|
||||
return -1;
|
||||
|
||||
st->nr_contended = data.count;
|
||||
st->wait_time_total = data.total_time;
|
||||
st->wait_time_max = data.max_time;
|
||||
st->wait_time_min = data.min_time;
|
||||
|
||||
if (data.count)
|
||||
st->avg_wait_time = data.total_time / data.count;
|
||||
|
||||
st->flags = data.flags;
|
||||
|
||||
bpf_map_lookup_elem(stack, &key, stack_trace);
|
||||
|
||||
/* skip BPF + lock internal functions */
|
||||
idx = CONTENTION_STACK_SKIP;
|
||||
while (is_lock_function(machine, stack_trace[idx]) &&
|
||||
idx < CONTENTION_STACK_DEPTH - 1)
|
||||
idx++;
|
||||
|
||||
st->addr = stack_trace[idx];
|
||||
sym = machine__find_kernel_symbol(machine, st->addr, &kmap);
|
||||
|
||||
if (sym) {
|
||||
unsigned long offset;
|
||||
int ret = 0;
|
||||
|
||||
offset = kmap->map_ip(kmap, st->addr) - sym->start;
|
||||
|
||||
if (offset)
|
||||
ret = asprintf(&st->name, "%s+%#lx", sym->name, offset);
|
||||
else
|
||||
st->name = strdup(sym->name);
|
||||
|
||||
if (ret < 0 || st->name == NULL)
|
||||
return -1;
|
||||
} else if (asprintf(&st->name, "%#lx", (unsigned long)st->addr) < 0) {
|
||||
free(st);
|
||||
return -1;
|
||||
}
|
||||
|
||||
hlist_add_head(&st->hash_entry, head);
|
||||
prev_key = key;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int lock_contention_finish(void)
|
||||
{
|
||||
if (skel) {
|
||||
skel->bss->enabled = 0;
|
||||
lock_contention_bpf__destroy(skel);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
|
@ -0,0 +1,131 @@
|
|||
// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
|
||||
// Copyright (c) 2022 Google
|
||||
#include "vmlinux.h"
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_tracing.h>
|
||||
#include <bpf/bpf_core_read.h>
|
||||
|
||||
/* maximum stack trace depth */
|
||||
#define MAX_STACKS 8
|
||||
|
||||
/* default buffer size */
|
||||
#define MAX_ENTRIES 10240
|
||||
|
||||
struct contention_key {
|
||||
__u32 stack_id;
|
||||
};
|
||||
|
||||
struct contention_data {
|
||||
__u64 total_time;
|
||||
__u64 min_time;
|
||||
__u64 max_time;
|
||||
__u32 count;
|
||||
__u32 flags;
|
||||
};
|
||||
|
||||
struct tstamp_data {
|
||||
__u64 timestamp;
|
||||
__u64 lock;
|
||||
__u32 flags;
|
||||
__u32 stack_id;
|
||||
};
|
||||
|
||||
/* callstack storage */
|
||||
struct {
|
||||
__uint(type, BPF_MAP_TYPE_STACK_TRACE);
|
||||
__uint(key_size, sizeof(__u32));
|
||||
__uint(value_size, MAX_STACKS * sizeof(__u64));
|
||||
__uint(max_entries, MAX_ENTRIES);
|
||||
} stacks SEC(".maps");
|
||||
|
||||
/* maintain timestamp at the beginning of contention */
|
||||
struct {
|
||||
__uint(type, BPF_MAP_TYPE_TASK_STORAGE);
|
||||
__uint(map_flags, BPF_F_NO_PREALLOC);
|
||||
__type(key, int);
|
||||
__type(value, struct tstamp_data);
|
||||
} tstamp SEC(".maps");
|
||||
|
||||
/* actual lock contention statistics */
|
||||
struct {
|
||||
__uint(type, BPF_MAP_TYPE_HASH);
|
||||
__uint(key_size, sizeof(struct contention_key));
|
||||
__uint(value_size, sizeof(struct contention_data));
|
||||
__uint(max_entries, MAX_ENTRIES);
|
||||
} lock_stat SEC(".maps");
|
||||
|
||||
/* control flags */
|
||||
int enabled;
|
||||
|
||||
SEC("tp_btf/contention_begin")
|
||||
int contention_begin(u64 *ctx)
|
||||
{
|
||||
struct task_struct *curr;
|
||||
struct tstamp_data *pelem;
|
||||
|
||||
if (!enabled)
|
||||
return 0;
|
||||
|
||||
curr = bpf_get_current_task_btf();
|
||||
pelem = bpf_task_storage_get(&tstamp, curr, NULL,
|
||||
BPF_LOCAL_STORAGE_GET_F_CREATE);
|
||||
if (!pelem || pelem->lock)
|
||||
return 0;
|
||||
|
||||
pelem->timestamp = bpf_ktime_get_ns();
|
||||
pelem->lock = (__u64)ctx[0];
|
||||
pelem->flags = (__u32)ctx[1];
|
||||
pelem->stack_id = bpf_get_stackid(ctx, &stacks, BPF_F_FAST_STACK_CMP);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
SEC("tp_btf/contention_end")
|
||||
int contention_end(u64 *ctx)
|
||||
{
|
||||
struct task_struct *curr;
|
||||
struct tstamp_data *pelem;
|
||||
struct contention_key key;
|
||||
struct contention_data *data;
|
||||
__u64 duration;
|
||||
|
||||
if (!enabled)
|
||||
return 0;
|
||||
|
||||
curr = bpf_get_current_task_btf();
|
||||
pelem = bpf_task_storage_get(&tstamp, curr, NULL, 0);
|
||||
if (!pelem || pelem->lock != ctx[0])
|
||||
return 0;
|
||||
|
||||
duration = bpf_ktime_get_ns() - pelem->timestamp;
|
||||
|
||||
key.stack_id = pelem->stack_id;
|
||||
data = bpf_map_lookup_elem(&lock_stat, &key);
|
||||
if (!data) {
|
||||
struct contention_data first = {
|
||||
.total_time = duration,
|
||||
.max_time = duration,
|
||||
.min_time = duration,
|
||||
.count = 1,
|
||||
.flags = pelem->flags,
|
||||
};
|
||||
|
||||
bpf_map_update_elem(&lock_stat, &key, &first, BPF_NOEXIST);
|
||||
pelem->lock = 0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
__sync_fetch_and_add(&data->total_time, duration);
|
||||
__sync_fetch_and_add(&data->count, 1);
|
||||
|
||||
/* FIXME: need atomic operations */
|
||||
if (data->max_time < duration)
|
||||
data->max_time = duration;
|
||||
if (data->min_time > duration)
|
||||
data->min_time = duration;
|
||||
|
||||
pelem->lock = 0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
char LICENSE[] SEC("license") = "Dual BSD/GPL";
|
|
@ -0,0 +1,133 @@
|
|||
// SPDX-License-Identifier: GPL-2.0
|
||||
#ifndef PERF_LOCK_CONTENTION_H
|
||||
#define PERF_LOCK_CONTENTION_H
|
||||
|
||||
#include <linux/list.h>
|
||||
#include <linux/rbtree.h>
|
||||
|
||||
struct lock_stat {
|
||||
struct hlist_node hash_entry;
|
||||
struct rb_node rb; /* used for sorting */
|
||||
|
||||
u64 addr; /* address of lockdep_map, used as ID */
|
||||
char *name; /* for strcpy(), we cannot use const */
|
||||
|
||||
unsigned int nr_acquire;
|
||||
unsigned int nr_acquired;
|
||||
unsigned int nr_contended;
|
||||
unsigned int nr_release;
|
||||
|
||||
union {
|
||||
unsigned int nr_readlock;
|
||||
unsigned int flags;
|
||||
};
|
||||
unsigned int nr_trylock;
|
||||
|
||||
/* these times are in nano sec. */
|
||||
u64 avg_wait_time;
|
||||
u64 wait_time_total;
|
||||
u64 wait_time_min;
|
||||
u64 wait_time_max;
|
||||
|
||||
int broken; /* flag of blacklist */
|
||||
int combined;
|
||||
};
|
||||
|
||||
/*
|
||||
* States of lock_seq_stat
|
||||
*
|
||||
* UNINITIALIZED is required for detecting first event of acquire.
|
||||
* As the nature of lock events, there is no guarantee
|
||||
* that the first event for the locks are acquire,
|
||||
* it can be acquired, contended or release.
|
||||
*/
|
||||
#define SEQ_STATE_UNINITIALIZED 0 /* initial state */
|
||||
#define SEQ_STATE_RELEASED 1
|
||||
#define SEQ_STATE_ACQUIRING 2
|
||||
#define SEQ_STATE_ACQUIRED 3
|
||||
#define SEQ_STATE_READ_ACQUIRED 4
|
||||
#define SEQ_STATE_CONTENDED 5
|
||||
|
||||
/*
|
||||
* MAX_LOCK_DEPTH
|
||||
* Imported from include/linux/sched.h.
|
||||
* Should this be synchronized?
|
||||
*/
|
||||
#define MAX_LOCK_DEPTH 48
|
||||
|
||||
/*
|
||||
* struct lock_seq_stat:
|
||||
* Place to put on state of one lock sequence
|
||||
* 1) acquire -> acquired -> release
|
||||
* 2) acquire -> contended -> acquired -> release
|
||||
* 3) acquire (with read or try) -> release
|
||||
* 4) Are there other patterns?
|
||||
*/
|
||||
struct lock_seq_stat {
|
||||
struct list_head list;
|
||||
int state;
|
||||
u64 prev_event_time;
|
||||
u64 addr;
|
||||
|
||||
int read_count;
|
||||
};
|
||||
|
||||
struct thread_stat {
|
||||
struct rb_node rb;
|
||||
|
||||
u32 tid;
|
||||
struct list_head seq_list;
|
||||
};
|
||||
|
||||
/*
|
||||
* CONTENTION_STACK_DEPTH
|
||||
* Number of stack trace entries to find callers
|
||||
*/
|
||||
#define CONTENTION_STACK_DEPTH 8
|
||||
|
||||
/*
|
||||
* CONTENTION_STACK_SKIP
|
||||
* Number of stack trace entries to skip when finding callers.
|
||||
* The first few entries belong to the locking implementation itself.
|
||||
*/
|
||||
#define CONTENTION_STACK_SKIP 3
|
||||
|
||||
/*
|
||||
* flags for lock:contention_begin
|
||||
* Imported from include/trace/events/lock.h.
|
||||
*/
|
||||
#define LCB_F_SPIN (1U << 0)
|
||||
#define LCB_F_READ (1U << 1)
|
||||
#define LCB_F_WRITE (1U << 2)
|
||||
#define LCB_F_RT (1U << 3)
|
||||
#define LCB_F_PERCPU (1U << 4)
|
||||
#define LCB_F_MUTEX (1U << 5)
|
||||
|
||||
struct machine;
|
||||
|
||||
#ifdef HAVE_BPF_SKEL
|
||||
|
||||
int lock_contention_prepare(void);
|
||||
int lock_contention_start(void);
|
||||
int lock_contention_stop(void);
|
||||
int lock_contention_read(struct machine *machine, struct hlist_head *head);
|
||||
int lock_contention_finish(void);
|
||||
|
||||
#else /* !HAVE_BPF_SKEL */
|
||||
|
||||
static inline int lock_contention_prepare(void) { return 0; }
|
||||
static inline int lock_contention_start(void) { return 0; }
|
||||
static inline int lock_contention_stop(void) { return 0; }
|
||||
static inline int lock_contention_finish(void) { return 0; }
|
||||
|
||||
static inline int lock_contention_read(struct machine *machine __maybe_unused,
|
||||
struct hlist_head *head __maybe_unused)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
#endif /* HAVE_BPF_SKEL */
|
||||
|
||||
bool is_lock_function(struct machine *machine, u64 addr);
|
||||
|
||||
#endif /* PERF_LOCK_CONTENTION_H */
|
Загрузка…
Ссылка в новой задаче