blob: 75a0dea511b3f2259fdf544d67e14b53df14b221 [file] [log] [blame]
// SPDX-License-Identifier: GPL-2.0
/* Copyright (c) 2024 Meta Platforms, Inc. and affiliates. */
#include <test_progs.h>
#include <time.h>
#include <sys/epoll.h>
#include "struct_ops_module.skel.h"
#include "struct_ops_nulled_out_cb.skel.h"
#include "struct_ops_forgotten_cb.skel.h"
#include "struct_ops_detach.skel.h"
#include "unsupported_ops.skel.h"
static void check_map_info(struct bpf_map_info *info)
{
struct bpf_btf_info btf_info;
char btf_name[256];
u32 btf_info_len = sizeof(btf_info);
int err, fd;
fd = bpf_btf_get_fd_by_id(info->btf_vmlinux_id);
if (!ASSERT_GE(fd, 0, "get_value_type_btf_obj_fd"))
return;
memset(&btf_info, 0, sizeof(btf_info));
btf_info.name = ptr_to_u64(btf_name);
btf_info.name_len = sizeof(btf_name);
err = bpf_btf_get_info_by_fd(fd, &btf_info, &btf_info_len);
if (!ASSERT_OK(err, "get_value_type_btf_obj_info"))
goto cleanup;
if (!ASSERT_EQ(strcmp(btf_name, "bpf_testmod"), 0, "get_value_type_btf_obj_name"))
goto cleanup;
cleanup:
close(fd);
}
static int attach_ops_and_check(struct struct_ops_module *skel,
struct bpf_map *map,
int expected_test_2_result)
{
struct bpf_link *link;
link = bpf_map__attach_struct_ops(map);
ASSERT_OK_PTR(link, "attach_test_mod_1");
if (!link)
return -1;
/* test_{1,2}() would be called from bpf_dummy_reg() in bpf_testmod.c */
ASSERT_EQ(skel->bss->test_1_result, 0xdeadbeef, "test_1_result");
ASSERT_EQ(skel->bss->test_2_result, expected_test_2_result, "test_2_result");
bpf_link__destroy(link);
return 0;
}
static void test_struct_ops_load(void)
{
struct struct_ops_module *skel;
struct bpf_map_info info = {};
int err;
u32 len;
skel = struct_ops_module__open();
if (!ASSERT_OK_PTR(skel, "struct_ops_module_open"))
return;
skel->struct_ops.testmod_1->data = 13;
skel->struct_ops.testmod_1->test_2 = skel->progs.test_3;
/* Since test_2() is not being used, it should be disabled from
* auto-loading, or it will fail to load.
*/
bpf_program__set_autoload(skel->progs.test_2, false);
bpf_map__set_autocreate(skel->maps.testmod_zeroed, false);
err = struct_ops_module__load(skel);
if (!ASSERT_OK(err, "struct_ops_module_load"))
goto cleanup;
len = sizeof(info);
err = bpf_map_get_info_by_fd(bpf_map__fd(skel->maps.testmod_1), &info,
&len);
if (!ASSERT_OK(err, "bpf_map_get_info_by_fd"))
goto cleanup;
check_map_info(&info);
/* test_3() will be called from bpf_dummy_reg() in bpf_testmod.c
*
* In bpf_testmod.c it will pass 4 and 13 (the value of data) to
* .test_2. So, the value of test_2_result should be 20 (4 + 13 +
* 3).
*/
if (!attach_ops_and_check(skel, skel->maps.testmod_1, 20))
goto cleanup;
if (!attach_ops_and_check(skel, skel->maps.testmod_2, 12))
goto cleanup;
cleanup:
struct_ops_module__destroy(skel);
}
static void test_struct_ops_not_zeroed(void)
{
struct struct_ops_module *skel;
int err;
/* zeroed is 0, and zeroed_op is null */
skel = struct_ops_module__open();
if (!ASSERT_OK_PTR(skel, "struct_ops_module_open"))
return;
skel->struct_ops.testmod_zeroed->zeroed = 0;
/* zeroed_op prog should be not loaded automatically now */
skel->struct_ops.testmod_zeroed->zeroed_op = NULL;
err = struct_ops_module__load(skel);
ASSERT_OK(err, "struct_ops_module_load");
struct_ops_module__destroy(skel);
/* zeroed is not 0 */
skel = struct_ops_module__open();
if (!ASSERT_OK_PTR(skel, "struct_ops_module_open_not_zeroed"))
return;
/* libbpf should reject the testmod_zeroed since struct
* bpf_testmod_ops in the kernel has no "zeroed" field and the
* value of "zeroed" is non-zero.
*/
skel->struct_ops.testmod_zeroed->zeroed = 0xdeadbeef;
skel->struct_ops.testmod_zeroed->zeroed_op = NULL;
err = struct_ops_module__load(skel);
ASSERT_ERR(err, "struct_ops_module_load_not_zeroed");
struct_ops_module__destroy(skel);
/* zeroed_op is not null */
skel = struct_ops_module__open();
if (!ASSERT_OK_PTR(skel, "struct_ops_module_open_not_zeroed_op"))
return;
/* libbpf should reject the testmod_zeroed since the value of its
* "zeroed_op" is not null.
*/
skel->struct_ops.testmod_zeroed->zeroed_op = skel->progs.test_3;
err = struct_ops_module__load(skel);
ASSERT_ERR(err, "struct_ops_module_load_not_zeroed_op");
struct_ops_module__destroy(skel);
}
/* The signature of an implementation might not match the signature of the
* function pointer prototype defined in the BPF program. This mismatch
* should be allowed as long as the behavior of the operator program
* adheres to the signature in the kernel. Libbpf should not enforce the
* signature; rather, let the kernel verifier handle the enforcement.
*/
static void test_struct_ops_incompatible(void)
{
struct struct_ops_module *skel;
struct bpf_link *link;
int err;
skel = struct_ops_module__open();
if (!ASSERT_OK_PTR(skel, "struct_ops_module_open"))
return;
bpf_map__set_autocreate(skel->maps.testmod_zeroed, false);
err = struct_ops_module__load(skel);
if (!ASSERT_OK(err, "skel_load"))
goto cleanup;
link = bpf_map__attach_struct_ops(skel->maps.testmod_incompatible);
if (ASSERT_OK_PTR(link, "attach_struct_ops"))
bpf_link__destroy(link);
cleanup:
struct_ops_module__destroy(skel);
}
/* validate that it's ok to "turn off" callback that kernel supports */
static void test_struct_ops_nulled_out_cb(void)
{
struct struct_ops_nulled_out_cb *skel;
int err;
skel = struct_ops_nulled_out_cb__open();
if (!ASSERT_OK_PTR(skel, "skel_open"))
return;
/* kernel knows about test_1, but we still null it out */
skel->struct_ops.ops->test_1 = NULL;
err = struct_ops_nulled_out_cb__load(skel);
if (!ASSERT_OK(err, "skel_load"))
goto cleanup;
ASSERT_FALSE(bpf_program__autoload(skel->progs.test_1_turn_off), "prog_autoload");
ASSERT_LT(bpf_program__fd(skel->progs.test_1_turn_off), 0, "prog_fd");
cleanup:
struct_ops_nulled_out_cb__destroy(skel);
}
/* validate that libbpf generates reasonable error message if struct_ops is
* not referenced in any struct_ops map
*/
static void test_struct_ops_forgotten_cb(void)
{
struct struct_ops_forgotten_cb *skel;
char *log;
int err;
skel = struct_ops_forgotten_cb__open();
if (!ASSERT_OK_PTR(skel, "skel_open"))
return;
start_libbpf_log_capture();
err = struct_ops_forgotten_cb__load(skel);
if (!ASSERT_ERR(err, "skel_load"))
goto cleanup;
log = stop_libbpf_log_capture();
ASSERT_HAS_SUBSTR(log,
"prog 'test_1_forgotten': SEC(\"struct_ops\") program isn't referenced anywhere, did you forget to use it?",
"libbpf_log");
free(log);
struct_ops_forgotten_cb__destroy(skel);
/* now let's programmatically use it, we should be fine now */
skel = struct_ops_forgotten_cb__open();
if (!ASSERT_OK_PTR(skel, "skel_open"))
return;
skel->struct_ops.ops->test_1 = skel->progs.test_1_forgotten; /* not anymore */
err = struct_ops_forgotten_cb__load(skel);
if (!ASSERT_OK(err, "skel_load"))
goto cleanup;
cleanup:
struct_ops_forgotten_cb__destroy(skel);
}
/* Detach a link from a user space program */
static void test_detach_link(void)
{
struct epoll_event ev, events[2];
struct struct_ops_detach *skel;
struct bpf_link *link = NULL;
int fd, epollfd = -1, nfds;
int err;
skel = struct_ops_detach__open_and_load();
if (!ASSERT_OK_PTR(skel, "struct_ops_detach__open_and_load"))
return;
link = bpf_map__attach_struct_ops(skel->maps.testmod_do_detach);
if (!ASSERT_OK_PTR(link, "attach_struct_ops"))
goto cleanup;
fd = bpf_link__fd(link);
if (!ASSERT_GE(fd, 0, "link_fd"))
goto cleanup;
epollfd = epoll_create1(0);
if (!ASSERT_GE(epollfd, 0, "epoll_create1"))
goto cleanup;
ev.events = EPOLLHUP;
ev.data.fd = fd;
err = epoll_ctl(epollfd, EPOLL_CTL_ADD, fd, &ev);
if (!ASSERT_OK(err, "epoll_ctl"))
goto cleanup;
err = bpf_link__detach(link);
if (!ASSERT_OK(err, "detach_link"))
goto cleanup;
/* Wait for EPOLLHUP */
nfds = epoll_wait(epollfd, events, 2, 500);
if (!ASSERT_EQ(nfds, 1, "epoll_wait"))
goto cleanup;
if (!ASSERT_EQ(events[0].data.fd, fd, "epoll_wait_fd"))
goto cleanup;
if (!ASSERT_TRUE(events[0].events & EPOLLHUP, "events[0].events"))
goto cleanup;
cleanup:
if (epollfd >= 0)
close(epollfd);
bpf_link__destroy(link);
struct_ops_detach__destroy(skel);
}
void serial_test_struct_ops_module(void)
{
if (test__start_subtest("struct_ops_load"))
test_struct_ops_load();
if (test__start_subtest("struct_ops_not_zeroed"))
test_struct_ops_not_zeroed();
if (test__start_subtest("struct_ops_incompatible"))
test_struct_ops_incompatible();
if (test__start_subtest("struct_ops_null_out_cb"))
test_struct_ops_nulled_out_cb();
if (test__start_subtest("struct_ops_forgotten_cb"))
test_struct_ops_forgotten_cb();
if (test__start_subtest("test_detach_link"))
test_detach_link();
RUN_TESTS(unsupported_ops);
}