mirror of
git://git.yoctoproject.org/linux-yocto.git
synced 2025-07-05 21:35:46 +02:00

Hi, fix some spelling errors in selftest, the details are as follows: -in the codes: test_bpf_sk_stoarge_map_iter_fd(void) ->test_bpf_sk_storage_map_iter_fd(void) load BTF from btf_data.o->load BTF from btf_data.bpf.o -in the code comments: preample->preamble multi-contollers->multi-controllers errono->errno unsighed/unsinged->unsigned egree->egress shoud->should regsiter->register assummed->assumed conditiona->conditional rougly->roughly timetamp->timestamp ingores->ignores null-termainted->null-terminated slepable->sleepable implemenation->implementation veriables->variables timetamps->timestamps substitue a costant->substitute a constant secton->section unreferened->unreferenced verifer->verifier libppf->libbpf ... Signed-off-by: Lin Yikai <yikai.lin@vivo.com> Link: https://lore.kernel.org/r/20240905110354.3274546-1-yikai.lin@vivo.com Signed-off-by: Alexei Starovoitov <ast@kernel.org>
1116 lines
29 KiB
C
1116 lines
29 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
#include <unistd.h>
|
|
#include <pthread.h>
|
|
#include <test_progs.h>
|
|
#include "uprobe_multi.skel.h"
|
|
#include "uprobe_multi_bench.skel.h"
|
|
#include "uprobe_multi_usdt.skel.h"
|
|
#include "uprobe_multi_consumers.skel.h"
|
|
#include "uprobe_multi_pid_filter.skel.h"
|
|
#include "bpf/libbpf_internal.h"
|
|
#include "testing_helpers.h"
|
|
#include "../sdt.h"
|
|
|
|
static char test_data[] = "test_data";
|
|
|
|
noinline void uprobe_multi_func_1(void)
|
|
{
|
|
asm volatile ("");
|
|
}
|
|
|
|
noinline void uprobe_multi_func_2(void)
|
|
{
|
|
asm volatile ("");
|
|
}
|
|
|
|
noinline void uprobe_multi_func_3(void)
|
|
{
|
|
asm volatile ("");
|
|
}
|
|
|
|
noinline void usdt_trigger(void)
|
|
{
|
|
STAP_PROBE(test, pid_filter_usdt);
|
|
}
|
|
|
|
struct child {
|
|
int go[2];
|
|
int c2p[2]; /* child -> parent channel */
|
|
int pid;
|
|
int tid;
|
|
pthread_t thread;
|
|
char stack[65536];
|
|
};
|
|
|
|
static void release_child(struct child *child)
|
|
{
|
|
int child_status;
|
|
|
|
if (!child)
|
|
return;
|
|
close(child->go[1]);
|
|
close(child->go[0]);
|
|
if (child->thread)
|
|
pthread_join(child->thread, NULL);
|
|
close(child->c2p[0]);
|
|
close(child->c2p[1]);
|
|
if (child->pid > 0)
|
|
waitpid(child->pid, &child_status, 0);
|
|
}
|
|
|
|
static void kick_child(struct child *child)
|
|
{
|
|
char c = 1;
|
|
|
|
if (child) {
|
|
write(child->go[1], &c, 1);
|
|
release_child(child);
|
|
}
|
|
fflush(NULL);
|
|
}
|
|
|
|
static int child_func(void *arg)
|
|
{
|
|
struct child *child = arg;
|
|
int err, c;
|
|
|
|
close(child->go[1]);
|
|
|
|
/* wait for parent's kick */
|
|
err = read(child->go[0], &c, 1);
|
|
if (err != 1)
|
|
exit(err);
|
|
|
|
uprobe_multi_func_1();
|
|
uprobe_multi_func_2();
|
|
uprobe_multi_func_3();
|
|
usdt_trigger();
|
|
|
|
exit(errno);
|
|
}
|
|
|
|
static int spawn_child_flag(struct child *child, bool clone_vm)
|
|
{
|
|
/* pipe to notify child to execute the trigger functions */
|
|
if (pipe(child->go))
|
|
return -1;
|
|
|
|
if (clone_vm) {
|
|
child->pid = child->tid = clone(child_func, child->stack + sizeof(child->stack)/2,
|
|
CLONE_VM|SIGCHLD, child);
|
|
} else {
|
|
child->pid = child->tid = fork();
|
|
}
|
|
if (child->pid < 0) {
|
|
release_child(child);
|
|
errno = EINVAL;
|
|
return -1;
|
|
}
|
|
|
|
/* fork-ed child */
|
|
if (!clone_vm && child->pid == 0)
|
|
child_func(child);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int spawn_child(struct child *child)
|
|
{
|
|
return spawn_child_flag(child, false);
|
|
}
|
|
|
|
static void *child_thread(void *ctx)
|
|
{
|
|
struct child *child = ctx;
|
|
int c = 0, err;
|
|
|
|
child->tid = syscall(SYS_gettid);
|
|
|
|
/* let parent know we are ready */
|
|
err = write(child->c2p[1], &c, 1);
|
|
if (err != 1)
|
|
pthread_exit(&err);
|
|
|
|
/* wait for parent's kick */
|
|
err = read(child->go[0], &c, 1);
|
|
if (err != 1)
|
|
pthread_exit(&err);
|
|
|
|
uprobe_multi_func_1();
|
|
uprobe_multi_func_2();
|
|
uprobe_multi_func_3();
|
|
usdt_trigger();
|
|
|
|
err = 0;
|
|
pthread_exit(&err);
|
|
}
|
|
|
|
static int spawn_thread(struct child *child)
|
|
{
|
|
int c, err;
|
|
|
|
/* pipe to notify child to execute the trigger functions */
|
|
if (pipe(child->go))
|
|
return -1;
|
|
/* pipe to notify parent that child thread is ready */
|
|
if (pipe(child->c2p)) {
|
|
close(child->go[0]);
|
|
close(child->go[1]);
|
|
return -1;
|
|
}
|
|
|
|
child->pid = getpid();
|
|
|
|
err = pthread_create(&child->thread, NULL, child_thread, child);
|
|
if (err) {
|
|
err = -errno;
|
|
close(child->go[0]);
|
|
close(child->go[1]);
|
|
close(child->c2p[0]);
|
|
close(child->c2p[1]);
|
|
errno = -err;
|
|
return -1;
|
|
}
|
|
|
|
err = read(child->c2p[0], &c, 1);
|
|
if (!ASSERT_EQ(err, 1, "child_thread_ready"))
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void uprobe_multi_test_run(struct uprobe_multi *skel, struct child *child)
|
|
{
|
|
skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1;
|
|
skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2;
|
|
skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3;
|
|
|
|
skel->bss->user_ptr = test_data;
|
|
|
|
/*
|
|
* Disable pid check in bpf program if we are pid filter test,
|
|
* because the probe should be executed only by child->pid
|
|
* passed at the probe attach.
|
|
*/
|
|
skel->bss->pid = child ? 0 : getpid();
|
|
skel->bss->expect_pid = child ? child->pid : 0;
|
|
|
|
/* trigger all probes, if we are testing child *process*, just to make
|
|
* sure that PID filtering doesn't let through activations from wrong
|
|
* PIDs; when we test child *thread*, we don't want to do this to
|
|
* avoid double counting number of triggering events
|
|
*/
|
|
if (!child || !child->thread) {
|
|
uprobe_multi_func_1();
|
|
uprobe_multi_func_2();
|
|
uprobe_multi_func_3();
|
|
usdt_trigger();
|
|
}
|
|
|
|
if (child)
|
|
kick_child(child);
|
|
|
|
/*
|
|
* There are 2 entry and 2 exit probe called for each uprobe_multi_func_[123]
|
|
* function and each sleepable probe (6) increments uprobe_multi_sleep_result.
|
|
*/
|
|
ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 2, "uprobe_multi_func_1_result");
|
|
ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 2, "uprobe_multi_func_2_result");
|
|
ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 2, "uprobe_multi_func_3_result");
|
|
|
|
ASSERT_EQ(skel->bss->uretprobe_multi_func_1_result, 2, "uretprobe_multi_func_1_result");
|
|
ASSERT_EQ(skel->bss->uretprobe_multi_func_2_result, 2, "uretprobe_multi_func_2_result");
|
|
ASSERT_EQ(skel->bss->uretprobe_multi_func_3_result, 2, "uretprobe_multi_func_3_result");
|
|
|
|
ASSERT_EQ(skel->bss->uprobe_multi_sleep_result, 6, "uprobe_multi_sleep_result");
|
|
|
|
ASSERT_FALSE(skel->bss->bad_pid_seen, "bad_pid_seen");
|
|
|
|
if (child) {
|
|
ASSERT_EQ(skel->bss->child_pid, child->pid, "uprobe_multi_child_pid");
|
|
ASSERT_EQ(skel->bss->child_tid, child->tid, "uprobe_multi_child_tid");
|
|
}
|
|
}
|
|
|
|
static void test_skel_api(void)
|
|
{
|
|
struct uprobe_multi *skel = NULL;
|
|
int err;
|
|
|
|
skel = uprobe_multi__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load"))
|
|
goto cleanup;
|
|
|
|
err = uprobe_multi__attach(skel);
|
|
if (!ASSERT_OK(err, "uprobe_multi__attach"))
|
|
goto cleanup;
|
|
|
|
uprobe_multi_test_run(skel, NULL);
|
|
|
|
cleanup:
|
|
uprobe_multi__destroy(skel);
|
|
}
|
|
|
|
static void
|
|
__test_attach_api(const char *binary, const char *pattern, struct bpf_uprobe_multi_opts *opts,
|
|
struct child *child)
|
|
{
|
|
pid_t pid = child ? child->pid : -1;
|
|
struct uprobe_multi *skel = NULL;
|
|
|
|
skel = uprobe_multi__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load"))
|
|
goto cleanup;
|
|
|
|
opts->retprobe = false;
|
|
skel->links.uprobe = bpf_program__attach_uprobe_multi(skel->progs.uprobe, pid,
|
|
binary, pattern, opts);
|
|
if (!ASSERT_OK_PTR(skel->links.uprobe, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
opts->retprobe = true;
|
|
skel->links.uretprobe = bpf_program__attach_uprobe_multi(skel->progs.uretprobe, pid,
|
|
binary, pattern, opts);
|
|
if (!ASSERT_OK_PTR(skel->links.uretprobe, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
opts->retprobe = false;
|
|
skel->links.uprobe_sleep = bpf_program__attach_uprobe_multi(skel->progs.uprobe_sleep, pid,
|
|
binary, pattern, opts);
|
|
if (!ASSERT_OK_PTR(skel->links.uprobe_sleep, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
opts->retprobe = true;
|
|
skel->links.uretprobe_sleep = bpf_program__attach_uprobe_multi(skel->progs.uretprobe_sleep,
|
|
pid, binary, pattern, opts);
|
|
if (!ASSERT_OK_PTR(skel->links.uretprobe_sleep, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
opts->retprobe = false;
|
|
skel->links.uprobe_extra = bpf_program__attach_uprobe_multi(skel->progs.uprobe_extra, -1,
|
|
binary, pattern, opts);
|
|
if (!ASSERT_OK_PTR(skel->links.uprobe_extra, "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
|
|
/* Attach (uprobe-backed) USDTs */
|
|
skel->links.usdt_pid = bpf_program__attach_usdt(skel->progs.usdt_pid, pid, binary,
|
|
"test", "pid_filter_usdt", NULL);
|
|
if (!ASSERT_OK_PTR(skel->links.usdt_pid, "attach_usdt_pid"))
|
|
goto cleanup;
|
|
|
|
skel->links.usdt_extra = bpf_program__attach_usdt(skel->progs.usdt_extra, -1, binary,
|
|
"test", "pid_filter_usdt", NULL);
|
|
if (!ASSERT_OK_PTR(skel->links.usdt_extra, "attach_usdt_extra"))
|
|
goto cleanup;
|
|
|
|
uprobe_multi_test_run(skel, child);
|
|
|
|
ASSERT_FALSE(skel->bss->bad_pid_seen_usdt, "bad_pid_seen_usdt");
|
|
if (child) {
|
|
ASSERT_EQ(skel->bss->child_pid_usdt, child->pid, "usdt_multi_child_pid");
|
|
ASSERT_EQ(skel->bss->child_tid_usdt, child->tid, "usdt_multi_child_tid");
|
|
}
|
|
cleanup:
|
|
uprobe_multi__destroy(skel);
|
|
}
|
|
|
|
static void
|
|
test_attach_api(const char *binary, const char *pattern, struct bpf_uprobe_multi_opts *opts)
|
|
{
|
|
static struct child child;
|
|
|
|
/* no pid filter */
|
|
__test_attach_api(binary, pattern, opts, NULL);
|
|
|
|
/* pid filter */
|
|
if (!ASSERT_OK(spawn_child(&child), "spawn_child"))
|
|
return;
|
|
|
|
__test_attach_api(binary, pattern, opts, &child);
|
|
|
|
/* pid filter (thread) */
|
|
if (!ASSERT_OK(spawn_thread(&child), "spawn_thread"))
|
|
return;
|
|
|
|
__test_attach_api(binary, pattern, opts, &child);
|
|
}
|
|
|
|
static void test_attach_api_pattern(void)
|
|
{
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
|
|
|
|
test_attach_api("/proc/self/exe", "uprobe_multi_func_*", &opts);
|
|
test_attach_api("/proc/self/exe", "uprobe_multi_func_?", &opts);
|
|
}
|
|
|
|
static void test_attach_api_syms(void)
|
|
{
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
|
|
const char *syms[3] = {
|
|
"uprobe_multi_func_1",
|
|
"uprobe_multi_func_2",
|
|
"uprobe_multi_func_3",
|
|
};
|
|
|
|
opts.syms = syms;
|
|
opts.cnt = ARRAY_SIZE(syms);
|
|
test_attach_api("/proc/self/exe", NULL, &opts);
|
|
}
|
|
|
|
static void test_attach_api_fails(void)
|
|
{
|
|
LIBBPF_OPTS(bpf_link_create_opts, opts);
|
|
const char *path = "/proc/self/exe";
|
|
struct uprobe_multi *skel = NULL;
|
|
int prog_fd, link_fd = -1;
|
|
unsigned long offset = 0;
|
|
|
|
skel = uprobe_multi__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load"))
|
|
goto cleanup;
|
|
|
|
prog_fd = bpf_program__fd(skel->progs.uprobe_extra);
|
|
|
|
/* abnormal cnt */
|
|
opts.uprobe_multi.path = path;
|
|
opts.uprobe_multi.offsets = &offset;
|
|
opts.uprobe_multi.cnt = INT_MAX;
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -E2BIG, "big cnt"))
|
|
goto cleanup;
|
|
|
|
/* cnt is 0 */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EINVAL, "cnt_is_zero"))
|
|
goto cleanup;
|
|
|
|
/* negative offset */
|
|
offset = -1;
|
|
opts.uprobe_multi.path = path;
|
|
opts.uprobe_multi.offsets = (unsigned long *) &offset;
|
|
opts.uprobe_multi.cnt = 1;
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EINVAL, "offset_is_negative"))
|
|
goto cleanup;
|
|
|
|
/* offsets is NULL */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EINVAL, "offsets_is_null"))
|
|
goto cleanup;
|
|
|
|
/* wrong offsets pointer */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.offsets = (unsigned long *) 1,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EFAULT, "offsets_is_wrong"))
|
|
goto cleanup;
|
|
|
|
/* path is NULL */
|
|
offset = 1;
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EINVAL, "path_is_null"))
|
|
goto cleanup;
|
|
|
|
/* wrong path pointer */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = (const char *) 1,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EFAULT, "path_is_wrong"))
|
|
goto cleanup;
|
|
|
|
/* wrong path type */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = "/",
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EBADF, "path_is_wrong_type"))
|
|
goto cleanup;
|
|
|
|
/* wrong cookies pointer */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cookies = (__u64 *) 1ULL,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EFAULT, "cookies_is_wrong"))
|
|
goto cleanup;
|
|
|
|
/* wrong ref_ctr_offsets pointer */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cookies = (__u64 *) &offset,
|
|
.uprobe_multi.ref_ctr_offsets = (unsigned long *) 1,
|
|
.uprobe_multi.cnt = 1,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EFAULT, "ref_ctr_offsets_is_wrong"))
|
|
goto cleanup;
|
|
|
|
/* wrong flags */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.flags = 1 << 31,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
if (!ASSERT_EQ(link_fd, -EINVAL, "wrong_flags"))
|
|
goto cleanup;
|
|
|
|
/* wrong pid */
|
|
LIBBPF_OPTS_RESET(opts,
|
|
.uprobe_multi.path = path,
|
|
.uprobe_multi.offsets = (unsigned long *) &offset,
|
|
.uprobe_multi.cnt = 1,
|
|
.uprobe_multi.pid = -2,
|
|
);
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
goto cleanup;
|
|
ASSERT_EQ(link_fd, -EINVAL, "pid_is_wrong");
|
|
|
|
cleanup:
|
|
if (link_fd >= 0)
|
|
close(link_fd);
|
|
uprobe_multi__destroy(skel);
|
|
}
|
|
|
|
#ifdef __x86_64__
|
|
noinline void uprobe_multi_error_func(void)
|
|
{
|
|
/*
|
|
* If --fcf-protection=branch is enabled the gcc generates endbr as
|
|
* first instruction, so marking the exact address of int3 with the
|
|
* symbol to be used in the attach_uprobe_fail_trap test below.
|
|
*/
|
|
asm volatile (
|
|
".globl uprobe_multi_error_func_int3; \n"
|
|
"uprobe_multi_error_func_int3: \n"
|
|
"int3 \n"
|
|
);
|
|
}
|
|
|
|
/*
|
|
* Attaching uprobe on uprobe_multi_error_func results in error
|
|
* because it already starts with int3 instruction.
|
|
*/
|
|
static void attach_uprobe_fail_trap(struct uprobe_multi *skel)
|
|
{
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
|
|
const char *syms[4] = {
|
|
"uprobe_multi_func_1",
|
|
"uprobe_multi_func_2",
|
|
"uprobe_multi_func_3",
|
|
"uprobe_multi_error_func_int3",
|
|
};
|
|
|
|
opts.syms = syms;
|
|
opts.cnt = ARRAY_SIZE(syms);
|
|
|
|
skel->links.uprobe = bpf_program__attach_uprobe_multi(skel->progs.uprobe, -1,
|
|
"/proc/self/exe", NULL, &opts);
|
|
if (!ASSERT_ERR_PTR(skel->links.uprobe, "bpf_program__attach_uprobe_multi")) {
|
|
bpf_link__destroy(skel->links.uprobe);
|
|
skel->links.uprobe = NULL;
|
|
}
|
|
}
|
|
#else
|
|
static void attach_uprobe_fail_trap(struct uprobe_multi *skel) { }
|
|
#endif
|
|
|
|
short sema_1 __used, sema_2 __used;
|
|
|
|
static void attach_uprobe_fail_refctr(struct uprobe_multi *skel)
|
|
{
|
|
unsigned long *tmp_offsets = NULL, *tmp_ref_ctr_offsets = NULL;
|
|
unsigned long offsets[3], ref_ctr_offsets[3];
|
|
LIBBPF_OPTS(bpf_link_create_opts, opts);
|
|
const char *path = "/proc/self/exe";
|
|
const char *syms[3] = {
|
|
"uprobe_multi_func_1",
|
|
"uprobe_multi_func_2",
|
|
};
|
|
const char *sema[3] = {
|
|
"sema_1",
|
|
"sema_2",
|
|
};
|
|
int prog_fd, link_fd, err;
|
|
|
|
prog_fd = bpf_program__fd(skel->progs.uprobe_extra);
|
|
|
|
err = elf_resolve_syms_offsets("/proc/self/exe", 2, (const char **) &syms,
|
|
&tmp_offsets, STT_FUNC);
|
|
if (!ASSERT_OK(err, "elf_resolve_syms_offsets_func"))
|
|
return;
|
|
|
|
err = elf_resolve_syms_offsets("/proc/self/exe", 2, (const char **) &sema,
|
|
&tmp_ref_ctr_offsets, STT_OBJECT);
|
|
if (!ASSERT_OK(err, "elf_resolve_syms_offsets_sema"))
|
|
goto cleanup;
|
|
|
|
/*
|
|
* We attach to 3 uprobes on 2 functions, so 2 uprobes share single function,
|
|
* but with different ref_ctr_offset which is not allowed and results in fail.
|
|
*/
|
|
offsets[0] = tmp_offsets[0]; /* uprobe_multi_func_1 */
|
|
offsets[1] = tmp_offsets[1]; /* uprobe_multi_func_2 */
|
|
offsets[2] = tmp_offsets[1]; /* uprobe_multi_func_2 */
|
|
|
|
ref_ctr_offsets[0] = tmp_ref_ctr_offsets[0]; /* sema_1 */
|
|
ref_ctr_offsets[1] = tmp_ref_ctr_offsets[1]; /* sema_2 */
|
|
ref_ctr_offsets[2] = tmp_ref_ctr_offsets[0]; /* sema_1, error */
|
|
|
|
opts.uprobe_multi.path = path;
|
|
opts.uprobe_multi.offsets = (const unsigned long *) &offsets;
|
|
opts.uprobe_multi.ref_ctr_offsets = (const unsigned long *) &ref_ctr_offsets;
|
|
opts.uprobe_multi.cnt = 3;
|
|
|
|
link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_ERR(link_fd, "link_fd"))
|
|
close(link_fd);
|
|
|
|
cleanup:
|
|
free(tmp_ref_ctr_offsets);
|
|
free(tmp_offsets);
|
|
}
|
|
|
|
static void test_attach_uprobe_fails(void)
|
|
{
|
|
struct uprobe_multi *skel = NULL;
|
|
|
|
skel = uprobe_multi__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load"))
|
|
return;
|
|
|
|
/* attach fails due to adding uprobe on trap instruction, x86_64 only */
|
|
attach_uprobe_fail_trap(skel);
|
|
|
|
/* attach fail due to wrong ref_ctr_offs on one of the uprobes */
|
|
attach_uprobe_fail_refctr(skel);
|
|
|
|
uprobe_multi__destroy(skel);
|
|
}
|
|
|
|
static void __test_link_api(struct child *child)
|
|
{
|
|
int prog_fd, link1_fd = -1, link2_fd = -1, link3_fd = -1, link4_fd = -1;
|
|
LIBBPF_OPTS(bpf_link_create_opts, opts);
|
|
const char *path = "/proc/self/exe";
|
|
struct uprobe_multi *skel = NULL;
|
|
unsigned long *offsets = NULL;
|
|
const char *syms[3] = {
|
|
"uprobe_multi_func_1",
|
|
"uprobe_multi_func_2",
|
|
"uprobe_multi_func_3",
|
|
};
|
|
int link_extra_fd = -1;
|
|
int err;
|
|
|
|
err = elf_resolve_syms_offsets(path, 3, syms, (unsigned long **) &offsets, STT_FUNC);
|
|
if (!ASSERT_OK(err, "elf_resolve_syms_offsets"))
|
|
return;
|
|
|
|
opts.uprobe_multi.path = path;
|
|
opts.uprobe_multi.offsets = offsets;
|
|
opts.uprobe_multi.cnt = ARRAY_SIZE(syms);
|
|
opts.uprobe_multi.pid = child ? child->pid : 0;
|
|
|
|
skel = uprobe_multi__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load"))
|
|
goto cleanup;
|
|
|
|
opts.kprobe_multi.flags = 0;
|
|
prog_fd = bpf_program__fd(skel->progs.uprobe);
|
|
link1_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_GE(link1_fd, 0, "link1_fd"))
|
|
goto cleanup;
|
|
|
|
opts.kprobe_multi.flags = BPF_F_UPROBE_MULTI_RETURN;
|
|
prog_fd = bpf_program__fd(skel->progs.uretprobe);
|
|
link2_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_GE(link2_fd, 0, "link2_fd"))
|
|
goto cleanup;
|
|
|
|
opts.kprobe_multi.flags = 0;
|
|
prog_fd = bpf_program__fd(skel->progs.uprobe_sleep);
|
|
link3_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_GE(link3_fd, 0, "link3_fd"))
|
|
goto cleanup;
|
|
|
|
opts.kprobe_multi.flags = BPF_F_UPROBE_MULTI_RETURN;
|
|
prog_fd = bpf_program__fd(skel->progs.uretprobe_sleep);
|
|
link4_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_GE(link4_fd, 0, "link4_fd"))
|
|
goto cleanup;
|
|
|
|
opts.kprobe_multi.flags = 0;
|
|
opts.uprobe_multi.pid = 0;
|
|
prog_fd = bpf_program__fd(skel->progs.uprobe_extra);
|
|
link_extra_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts);
|
|
if (!ASSERT_GE(link_extra_fd, 0, "link_extra_fd"))
|
|
goto cleanup;
|
|
|
|
uprobe_multi_test_run(skel, child);
|
|
|
|
cleanup:
|
|
if (link1_fd >= 0)
|
|
close(link1_fd);
|
|
if (link2_fd >= 0)
|
|
close(link2_fd);
|
|
if (link3_fd >= 0)
|
|
close(link3_fd);
|
|
if (link4_fd >= 0)
|
|
close(link4_fd);
|
|
if (link_extra_fd >= 0)
|
|
close(link_extra_fd);
|
|
|
|
uprobe_multi__destroy(skel);
|
|
free(offsets);
|
|
}
|
|
|
|
static void test_link_api(void)
|
|
{
|
|
static struct child child;
|
|
|
|
/* no pid filter */
|
|
__test_link_api(NULL);
|
|
|
|
/* pid filter */
|
|
if (!ASSERT_OK(spawn_child(&child), "spawn_child"))
|
|
return;
|
|
|
|
__test_link_api(&child);
|
|
|
|
/* pid filter (thread) */
|
|
if (!ASSERT_OK(spawn_thread(&child), "spawn_thread"))
|
|
return;
|
|
|
|
__test_link_api(&child);
|
|
}
|
|
|
|
static struct bpf_program *
|
|
get_program(struct uprobe_multi_consumers *skel, int prog)
|
|
{
|
|
switch (prog) {
|
|
case 0:
|
|
return skel->progs.uprobe_0;
|
|
case 1:
|
|
return skel->progs.uprobe_1;
|
|
case 2:
|
|
return skel->progs.uprobe_2;
|
|
case 3:
|
|
return skel->progs.uprobe_3;
|
|
default:
|
|
ASSERT_FAIL("get_program");
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static struct bpf_link **
|
|
get_link(struct uprobe_multi_consumers *skel, int link)
|
|
{
|
|
switch (link) {
|
|
case 0:
|
|
return &skel->links.uprobe_0;
|
|
case 1:
|
|
return &skel->links.uprobe_1;
|
|
case 2:
|
|
return &skel->links.uprobe_2;
|
|
case 3:
|
|
return &skel->links.uprobe_3;
|
|
default:
|
|
ASSERT_FAIL("get_link");
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static int uprobe_attach(struct uprobe_multi_consumers *skel, int idx)
|
|
{
|
|
struct bpf_program *prog = get_program(skel, idx);
|
|
struct bpf_link **link = get_link(skel, idx);
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
|
|
|
|
if (!prog || !link)
|
|
return -1;
|
|
|
|
/*
|
|
* bit/prog: 0,1 uprobe entry
|
|
* bit/prog: 2,3 uprobe return
|
|
*/
|
|
opts.retprobe = idx == 2 || idx == 3;
|
|
|
|
*link = bpf_program__attach_uprobe_multi(prog, 0, "/proc/self/exe",
|
|
"uprobe_consumer_test",
|
|
&opts);
|
|
if (!ASSERT_OK_PTR(*link, "bpf_program__attach_uprobe_multi"))
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
static void uprobe_detach(struct uprobe_multi_consumers *skel, int idx)
|
|
{
|
|
struct bpf_link **link = get_link(skel, idx);
|
|
|
|
bpf_link__destroy(*link);
|
|
*link = NULL;
|
|
}
|
|
|
|
static bool test_bit(int bit, unsigned long val)
|
|
{
|
|
return val & (1 << bit);
|
|
}
|
|
|
|
noinline int
|
|
uprobe_consumer_test(struct uprobe_multi_consumers *skel,
|
|
unsigned long before, unsigned long after)
|
|
{
|
|
int idx;
|
|
|
|
/* detach uprobe for each unset programs in 'before' state ... */
|
|
for (idx = 0; idx < 4; idx++) {
|
|
if (test_bit(idx, before) && !test_bit(idx, after))
|
|
uprobe_detach(skel, idx);
|
|
}
|
|
|
|
/* ... and attach all new programs in 'after' state */
|
|
for (idx = 0; idx < 4; idx++) {
|
|
if (!test_bit(idx, before) && test_bit(idx, after)) {
|
|
if (!ASSERT_OK(uprobe_attach(skel, idx), "uprobe_attach_after"))
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void consumer_test(struct uprobe_multi_consumers *skel,
|
|
unsigned long before, unsigned long after)
|
|
{
|
|
int err, idx;
|
|
|
|
printf("consumer_test before %lu after %lu\n", before, after);
|
|
|
|
/* 'before' is each, we attach uprobe for every set idx */
|
|
for (idx = 0; idx < 4; idx++) {
|
|
if (test_bit(idx, before)) {
|
|
if (!ASSERT_OK(uprobe_attach(skel, idx), "uprobe_attach_before"))
|
|
goto cleanup;
|
|
}
|
|
}
|
|
|
|
err = uprobe_consumer_test(skel, before, after);
|
|
if (!ASSERT_EQ(err, 0, "uprobe_consumer_test"))
|
|
goto cleanup;
|
|
|
|
for (idx = 0; idx < 4; idx++) {
|
|
const char *fmt = "BUG";
|
|
__u64 val = 0;
|
|
|
|
if (idx < 2) {
|
|
/*
|
|
* uprobe entry
|
|
* +1 if define in 'before'
|
|
*/
|
|
if (test_bit(idx, before))
|
|
val++;
|
|
fmt = "prog 0/1: uprobe";
|
|
} else {
|
|
/*
|
|
* uprobe return is tricky ;-)
|
|
*
|
|
* to trigger uretprobe consumer, the uretprobe needs to be installed,
|
|
* which means one of the 'return' uprobes was alive when probe was hit:
|
|
*
|
|
* idxs: 2/3 uprobe return in 'installed' mask
|
|
*
|
|
* in addition if 'after' state removes everything that was installed in
|
|
* 'before' state, then uprobe kernel object goes away and return uprobe
|
|
* is not installed and we won't hit it even if it's in 'after' state.
|
|
*/
|
|
unsigned long had_uretprobes = before & 0b1100; /* is uretprobe installed */
|
|
unsigned long probe_preserved = before & after; /* did uprobe go away */
|
|
|
|
if (had_uretprobes && probe_preserved && test_bit(idx, after))
|
|
val++;
|
|
fmt = "idx 2/3: uretprobe";
|
|
}
|
|
|
|
ASSERT_EQ(skel->bss->uprobe_result[idx], val, fmt);
|
|
skel->bss->uprobe_result[idx] = 0;
|
|
}
|
|
|
|
cleanup:
|
|
for (idx = 0; idx < 4; idx++)
|
|
uprobe_detach(skel, idx);
|
|
}
|
|
|
|
static void test_consumers(void)
|
|
{
|
|
struct uprobe_multi_consumers *skel;
|
|
int before, after;
|
|
|
|
skel = uprobe_multi_consumers__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi_consumers__open_and_load"))
|
|
return;
|
|
|
|
/*
|
|
* The idea of this test is to try all possible combinations of
|
|
* uprobes consumers attached on single function.
|
|
*
|
|
* - 2 uprobe entry consumer
|
|
* - 2 uprobe exit consumers
|
|
*
|
|
* The test uses 4 uprobes attached on single function, but that
|
|
* translates into single uprobe with 4 consumers in kernel.
|
|
*
|
|
* The before/after values present the state of attached consumers
|
|
* before and after the probed function:
|
|
*
|
|
* bit/prog 0,1 : uprobe entry
|
|
* bit/prog 2,3 : uprobe return
|
|
*
|
|
* For example for:
|
|
*
|
|
* before = 0b0101
|
|
* after = 0b0110
|
|
*
|
|
* it means that before we call 'uprobe_consumer_test' we attach
|
|
* uprobes defined in 'before' value:
|
|
*
|
|
* - bit/prog 0: uprobe entry
|
|
* - bit/prog 2: uprobe return
|
|
*
|
|
* uprobe_consumer_test is called and inside it we attach and detach
|
|
* uprobes based on 'after' value:
|
|
*
|
|
* - bit/prog 0: stays untouched
|
|
* - bit/prog 2: uprobe return is detached
|
|
*
|
|
* uprobe_consumer_test returns and we check counters values increased
|
|
* by bpf programs on each uprobe to match the expected count based on
|
|
* before/after bits.
|
|
*/
|
|
|
|
for (before = 0; before < 16; before++) {
|
|
for (after = 0; after < 16; after++)
|
|
consumer_test(skel, before, after);
|
|
}
|
|
|
|
uprobe_multi_consumers__destroy(skel);
|
|
}
|
|
|
|
static struct bpf_program *uprobe_multi_program(struct uprobe_multi_pid_filter *skel, int idx)
|
|
{
|
|
switch (idx) {
|
|
case 0: return skel->progs.uprobe_multi_0;
|
|
case 1: return skel->progs.uprobe_multi_1;
|
|
case 2: return skel->progs.uprobe_multi_2;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
#define TASKS 3
|
|
|
|
static void run_pid_filter(struct uprobe_multi_pid_filter *skel, bool clone_vm, bool retprobe)
|
|
{
|
|
LIBBPF_OPTS(bpf_uprobe_multi_opts, opts, .retprobe = retprobe);
|
|
struct bpf_link *link[TASKS] = {};
|
|
struct child child[TASKS] = {};
|
|
int i;
|
|
|
|
memset(skel->bss->test, 0, sizeof(skel->bss->test));
|
|
|
|
for (i = 0; i < TASKS; i++) {
|
|
if (!ASSERT_OK(spawn_child_flag(&child[i], clone_vm), "spawn_child"))
|
|
goto cleanup;
|
|
skel->bss->pids[i] = child[i].pid;
|
|
}
|
|
|
|
for (i = 0; i < TASKS; i++) {
|
|
link[i] = bpf_program__attach_uprobe_multi(uprobe_multi_program(skel, i),
|
|
child[i].pid, "/proc/self/exe",
|
|
"uprobe_multi_func_1", &opts);
|
|
if (!ASSERT_OK_PTR(link[i], "bpf_program__attach_uprobe_multi"))
|
|
goto cleanup;
|
|
}
|
|
|
|
for (i = 0; i < TASKS; i++)
|
|
kick_child(&child[i]);
|
|
|
|
for (i = 0; i < TASKS; i++) {
|
|
ASSERT_EQ(skel->bss->test[i][0], 1, "pid");
|
|
ASSERT_EQ(skel->bss->test[i][1], 0, "unknown");
|
|
}
|
|
|
|
cleanup:
|
|
for (i = 0; i < TASKS; i++)
|
|
bpf_link__destroy(link[i]);
|
|
for (i = 0; i < TASKS; i++)
|
|
release_child(&child[i]);
|
|
}
|
|
|
|
static void test_pid_filter_process(bool clone_vm)
|
|
{
|
|
struct uprobe_multi_pid_filter *skel;
|
|
|
|
skel = uprobe_multi_pid_filter__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi_pid_filter__open_and_load"))
|
|
return;
|
|
|
|
run_pid_filter(skel, clone_vm, false);
|
|
run_pid_filter(skel, clone_vm, true);
|
|
|
|
uprobe_multi_pid_filter__destroy(skel);
|
|
}
|
|
|
|
static void test_bench_attach_uprobe(void)
|
|
{
|
|
long attach_start_ns = 0, attach_end_ns = 0;
|
|
struct uprobe_multi_bench *skel = NULL;
|
|
long detach_start_ns, detach_end_ns;
|
|
double attach_delta, detach_delta;
|
|
int err;
|
|
|
|
skel = uprobe_multi_bench__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi_bench__open_and_load"))
|
|
goto cleanup;
|
|
|
|
attach_start_ns = get_time_ns();
|
|
|
|
err = uprobe_multi_bench__attach(skel);
|
|
if (!ASSERT_OK(err, "uprobe_multi_bench__attach"))
|
|
goto cleanup;
|
|
|
|
attach_end_ns = get_time_ns();
|
|
|
|
system("./uprobe_multi bench");
|
|
|
|
ASSERT_EQ(skel->bss->count, 50000, "uprobes_count");
|
|
|
|
cleanup:
|
|
detach_start_ns = get_time_ns();
|
|
uprobe_multi_bench__destroy(skel);
|
|
detach_end_ns = get_time_ns();
|
|
|
|
attach_delta = (attach_end_ns - attach_start_ns) / 1000000000.0;
|
|
detach_delta = (detach_end_ns - detach_start_ns) / 1000000000.0;
|
|
|
|
printf("%s: attached in %7.3lfs\n", __func__, attach_delta);
|
|
printf("%s: detached in %7.3lfs\n", __func__, detach_delta);
|
|
}
|
|
|
|
static void test_bench_attach_usdt(void)
|
|
{
|
|
long attach_start_ns = 0, attach_end_ns = 0;
|
|
struct uprobe_multi_usdt *skel = NULL;
|
|
long detach_start_ns, detach_end_ns;
|
|
double attach_delta, detach_delta;
|
|
|
|
skel = uprobe_multi_usdt__open_and_load();
|
|
if (!ASSERT_OK_PTR(skel, "uprobe_multi__open"))
|
|
goto cleanup;
|
|
|
|
attach_start_ns = get_time_ns();
|
|
|
|
skel->links.usdt0 = bpf_program__attach_usdt(skel->progs.usdt0, -1, "./uprobe_multi",
|
|
"test", "usdt", NULL);
|
|
if (!ASSERT_OK_PTR(skel->links.usdt0, "bpf_program__attach_usdt"))
|
|
goto cleanup;
|
|
|
|
attach_end_ns = get_time_ns();
|
|
|
|
system("./uprobe_multi usdt");
|
|
|
|
ASSERT_EQ(skel->bss->count, 50000, "usdt_count");
|
|
|
|
cleanup:
|
|
detach_start_ns = get_time_ns();
|
|
uprobe_multi_usdt__destroy(skel);
|
|
detach_end_ns = get_time_ns();
|
|
|
|
attach_delta = (attach_end_ns - attach_start_ns) / 1000000000.0;
|
|
detach_delta = (detach_end_ns - detach_start_ns) / 1000000000.0;
|
|
|
|
printf("%s: attached in %7.3lfs\n", __func__, attach_delta);
|
|
printf("%s: detached in %7.3lfs\n", __func__, detach_delta);
|
|
}
|
|
|
|
void test_uprobe_multi_test(void)
|
|
{
|
|
if (test__start_subtest("skel_api"))
|
|
test_skel_api();
|
|
if (test__start_subtest("attach_api_pattern"))
|
|
test_attach_api_pattern();
|
|
if (test__start_subtest("attach_api_syms"))
|
|
test_attach_api_syms();
|
|
if (test__start_subtest("link_api"))
|
|
test_link_api();
|
|
if (test__start_subtest("bench_uprobe"))
|
|
test_bench_attach_uprobe();
|
|
if (test__start_subtest("bench_usdt"))
|
|
test_bench_attach_usdt();
|
|
if (test__start_subtest("attach_api_fails"))
|
|
test_attach_api_fails();
|
|
if (test__start_subtest("attach_uprobe_fails"))
|
|
test_attach_uprobe_fails();
|
|
if (test__start_subtest("consumers"))
|
|
test_consumers();
|
|
if (test__start_subtest("filter_fork"))
|
|
test_pid_filter_process(false);
|
|
if (test__start_subtest("filter_clone_vm"))
|
|
test_pid_filter_process(true);
|
|
}
|