| /* SPDX-License-Identifier: GPL-2.0-only */ |
| /* |
| * Copyright (c) 2012 The Chromium OS Authors. All rights reserved. |
| * |
| * kselftest_harness.h: simple C unit test helper. |
| * |
| * See documentation in Documentation/dev-tools/kselftest.rst |
| * |
| * API inspired by code.google.com/p/googletest |
| */ |
| |
| /** |
| * DOC: example |
| * |
| * .. code-block:: c |
| * |
| * #include "../kselftest_harness.h" |
| * |
| * TEST(standalone_test) { |
| * do_some_stuff; |
| * EXPECT_GT(10, stuff) { |
| * stuff_state_t state; |
| * enumerate_stuff_state(&state); |
| * TH_LOG("expectation failed with state: %s", state.msg); |
| * } |
| * more_stuff; |
| * ASSERT_NE(some_stuff, NULL) TH_LOG("how did it happen?!"); |
| * last_stuff; |
| * EXPECT_EQ(0, last_stuff); |
| * } |
| * |
| * FIXTURE(my_fixture) { |
| * mytype_t *data; |
| * int awesomeness_level; |
| * }; |
| * FIXTURE_SETUP(my_fixture) { |
| * self->data = mytype_new(); |
| * ASSERT_NE(NULL, self->data); |
| * } |
| * FIXTURE_TEARDOWN(my_fixture) { |
| * mytype_free(self->data); |
| * } |
| * TEST_F(my_fixture, data_is_good) { |
| * EXPECT_EQ(1, is_my_data_good(self->data)); |
| * } |
| * |
| * TEST_HARNESS_MAIN |
| */ |
| |
| #ifndef __KSELFTEST_HARNESS_H |
| #define __KSELFTEST_HARNESS_H |
| |
| #ifndef _GNU_SOURCE |
| #define _GNU_SOURCE |
| #endif |
| #include <asm/types.h> |
| #include <errno.h> |
| #include <stdbool.h> |
| #include <stdint.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <string.h> |
| #include <sys/mman.h> |
| #include <sys/types.h> |
| #include <sys/wait.h> |
| #include <unistd.h> |
| #include <setjmp.h> |
| |
| #include "kselftest.h" |
| |
| #define TEST_TIMEOUT_DEFAULT 30 |
| |
| /* Utilities exposed to the test definitions */ |
| #ifndef TH_LOG_STREAM |
| # define TH_LOG_STREAM stderr |
| #endif |
| |
| #ifndef TH_LOG_ENABLED |
| # define TH_LOG_ENABLED 1 |
| #endif |
| |
| /** |
| * TH_LOG() |
| * |
| * @fmt: format string |
| * @...: optional arguments |
| * |
| * .. code-block:: c |
| * |
| * TH_LOG(format, ...) |
| * |
| * Optional debug logging function available for use in tests. |
| * Logging may be enabled or disabled by defining TH_LOG_ENABLED. |
| * E.g., #define TH_LOG_ENABLED 1 |
| * |
| * If no definition is provided, logging is enabled by default. |
| * |
| * If there is no way to print an error message for the process running the |
| * test (e.g. not allowed to write to stderr), it is still possible to get the |
| * ASSERT_* number for which the test failed. This behavior can be enabled by |
| * writing `_metadata->no_print = true;` before the check sequence that is |
| * unable to print. When an error occur, instead of printing an error message |
| * and calling `abort(3)`, the test process call `_exit(2)` with the assert |
| * number as argument, which is then printed by the parent process. |
| */ |
| #define TH_LOG(fmt, ...) do { \ |
| if (TH_LOG_ENABLED) \ |
| __TH_LOG(fmt, ##__VA_ARGS__); \ |
| } while (0) |
| |
| /* Unconditional logger for internal use. */ |
| #define __TH_LOG(fmt, ...) \ |
| fprintf(TH_LOG_STREAM, "# %s:%d:%s:" fmt "\n", \ |
| __FILE__, __LINE__, _metadata->name, ##__VA_ARGS__) |
| |
| /** |
| * SKIP() |
| * |
| * @statement: statement to run after reporting SKIP |
| * @fmt: format string |
| * @...: optional arguments |
| * |
| * .. code-block:: c |
| * |
| * SKIP(statement, fmt, ...); |
| * |
| * This forces a "pass" after reporting why something is being skipped |
| * and runs "statement", which is usually "return" or "goto skip". |
| */ |
| #define SKIP(statement, fmt, ...) do { \ |
| snprintf(_metadata->results->reason, \ |
| sizeof(_metadata->results->reason), fmt, ##__VA_ARGS__); \ |
| if (TH_LOG_ENABLED) { \ |
| fprintf(TH_LOG_STREAM, "# SKIP %s\n", \ |
| _metadata->results->reason); \ |
| } \ |
| _metadata->passed = 1; \ |
| _metadata->skip = 1; \ |
| _metadata->trigger = 0; \ |
| statement; \ |
| } while (0) |
| |
| /** |
| * TEST() - Defines the test function and creates the registration |
| * stub |
| * |
| * @test_name: test name |
| * |
| * .. code-block:: c |
| * |
| * TEST(name) { implementation } |
| * |
| * Defines a test by name. |
| * Names must be unique and tests must not be run in parallel. The |
| * implementation containing block is a function and scoping should be treated |
| * as such. Returning early may be performed with a bare "return;" statement. |
| * |
| * EXPECT_* and ASSERT_* are valid in a TEST() { } context. |
| */ |
| #define TEST(test_name) __TEST_IMPL(test_name, -1) |
| |
| /** |
| * TEST_SIGNAL() |
| * |
| * @test_name: test name |
| * @signal: signal number |
| * |
| * .. code-block:: c |
| * |
| * TEST_SIGNAL(name, signal) { implementation } |
| * |
| * Defines a test by name and the expected term signal. |
| * Names must be unique and tests must not be run in parallel. The |
| * implementation containing block is a function and scoping should be treated |
| * as such. Returning early may be performed with a bare "return;" statement. |
| * |
| * EXPECT_* and ASSERT_* are valid in a TEST() { } context. |
| */ |
| #define TEST_SIGNAL(test_name, signal) __TEST_IMPL(test_name, signal) |
| |
| #define __TEST_IMPL(test_name, _signal) \ |
| static void test_name(struct __test_metadata *_metadata); \ |
| static inline void wrapper_##test_name( \ |
| struct __test_metadata *_metadata, \ |
| struct __fixture_variant_metadata *variant) \ |
| { \ |
| _metadata->setup_completed = true; \ |
| if (setjmp(_metadata->env) == 0) \ |
| test_name(_metadata); \ |
| __test_check_assert(_metadata); \ |
| } \ |
| static struct __test_metadata _##test_name##_object = \ |
| { .name = #test_name, \ |
| .fn = &wrapper_##test_name, \ |
| .fixture = &_fixture_global, \ |
| .termsig = _signal, \ |
| .timeout = TEST_TIMEOUT_DEFAULT, }; \ |
| static void __attribute__((constructor)) _register_##test_name(void) \ |
| { \ |
| __register_test(&_##test_name##_object); \ |
| } \ |
| static void test_name( \ |
| struct __test_metadata __attribute__((unused)) *_metadata) |
| |
| /** |
| * FIXTURE_DATA() - Wraps the struct name so we have one less |
| * argument to pass around |
| * |
| * @datatype_name: datatype name |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE_DATA(datatype_name) |
| * |
| * Almost always, you want just FIXTURE() instead (see below). |
| * This call may be used when the type of the fixture data |
| * is needed. In general, this should not be needed unless |
| * the *self* is being passed to a helper directly. |
| */ |
| #define FIXTURE_DATA(datatype_name) struct _test_data_##datatype_name |
| |
| /** |
| * FIXTURE() - Called once per fixture to setup the data and |
| * register |
| * |
| * @fixture_name: fixture name |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE(fixture_name) { |
| * type property1; |
| * ... |
| * }; |
| * |
| * Defines the data provided to TEST_F()-defined tests as *self*. It should be |
| * populated and cleaned up using FIXTURE_SETUP() and FIXTURE_TEARDOWN(). |
| */ |
| #define FIXTURE(fixture_name) \ |
| FIXTURE_VARIANT(fixture_name); \ |
| static struct __fixture_metadata _##fixture_name##_fixture_object = \ |
| { .name = #fixture_name, }; \ |
| static void __attribute__((constructor)) \ |
| _register_##fixture_name##_data(void) \ |
| { \ |
| __register_fixture(&_##fixture_name##_fixture_object); \ |
| } \ |
| FIXTURE_DATA(fixture_name) |
| |
| /** |
| * FIXTURE_SETUP() - Prepares the setup function for the fixture. |
| * *_metadata* is included so that EXPECT_* and ASSERT_* work correctly. |
| * |
| * @fixture_name: fixture name |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE_SETUP(fixture_name) { implementation } |
| * |
| * Populates the required "setup" function for a fixture. An instance of the |
| * datatype defined with FIXTURE_DATA() will be exposed as *self* for the |
| * implementation. |
| * |
| * ASSERT_* are valid for use in this context and will prempt the execution |
| * of any dependent fixture tests. |
| * |
| * A bare "return;" statement may be used to return early. |
| */ |
| #define FIXTURE_SETUP(fixture_name) \ |
| void fixture_name##_setup( \ |
| struct __test_metadata __attribute__((unused)) *_metadata, \ |
| FIXTURE_DATA(fixture_name) __attribute__((unused)) *self, \ |
| const FIXTURE_VARIANT(fixture_name) \ |
| __attribute__((unused)) *variant) |
| |
| /** |
| * FIXTURE_TEARDOWN() |
| * *_metadata* is included so that EXPECT_* and ASSERT_* work correctly. |
| * |
| * @fixture_name: fixture name |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE_TEARDOWN(fixture_name) { implementation } |
| * |
| * Populates the required "teardown" function for a fixture. An instance of the |
| * datatype defined with FIXTURE_DATA() will be exposed as *self* for the |
| * implementation to clean up. |
| * |
| * A bare "return;" statement may be used to return early. |
| */ |
| #define FIXTURE_TEARDOWN(fixture_name) \ |
| void fixture_name##_teardown( \ |
| struct __test_metadata __attribute__((unused)) *_metadata, \ |
| FIXTURE_DATA(fixture_name) __attribute__((unused)) *self, \ |
| const FIXTURE_VARIANT(fixture_name) \ |
| __attribute__((unused)) *variant) |
| |
| /** |
| * FIXTURE_VARIANT() - Optionally called once per fixture |
| * to declare fixture variant |
| * |
| * @fixture_name: fixture name |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE_VARIANT(fixture_name) { |
| * type property1; |
| * ... |
| * }; |
| * |
| * Defines type of constant parameters provided to FIXTURE_SETUP(), TEST_F() and |
| * FIXTURE_TEARDOWN as *variant*. Variants allow the same tests to be run with |
| * different arguments. |
| */ |
| #define FIXTURE_VARIANT(fixture_name) struct _fixture_variant_##fixture_name |
| |
| /** |
| * FIXTURE_VARIANT_ADD() - Called once per fixture |
| * variant to setup and register the data |
| * |
| * @fixture_name: fixture name |
| * @variant_name: name of the parameter set |
| * |
| * .. code-block:: c |
| * |
| * FIXTURE_VARIANT_ADD(fixture_name, variant_name) { |
| * .property1 = val1, |
| * ... |
| * }; |
| * |
| * Defines a variant of the test fixture, provided to FIXTURE_SETUP() and |
| * TEST_F() as *variant*. Tests of each fixture will be run once for each |
| * variant. |
| */ |
| #define FIXTURE_VARIANT_ADD(fixture_name, variant_name) \ |
| extern FIXTURE_VARIANT(fixture_name) \ |
| _##fixture_name##_##variant_name##_variant; \ |
| static struct __fixture_variant_metadata \ |
| _##fixture_name##_##variant_name##_object = \ |
| { .name = #variant_name, \ |
| .data = &_##fixture_name##_##variant_name##_variant}; \ |
| static void __attribute__((constructor)) \ |
| _register_##fixture_name##_##variant_name(void) \ |
| { \ |
| __register_fixture_variant(&_##fixture_name##_fixture_object, \ |
| &_##fixture_name##_##variant_name##_object); \ |
| } \ |
| FIXTURE_VARIANT(fixture_name) \ |
| _##fixture_name##_##variant_name##_variant = |
| |
| /** |
| * TEST_F() - Emits test registration and helpers for |
| * fixture-based test cases |
| * |
| * @fixture_name: fixture name |
| * @test_name: test name |
| * |
| * .. code-block:: c |
| * |
| * TEST_F(fixture, name) { implementation } |
| * |
| * Defines a test that depends on a fixture (e.g., is part of a test case). |
| * Very similar to TEST() except that *self* is the setup instance of fixture's |
| * datatype exposed for use by the implementation. |
| */ |
| #define TEST_F(fixture_name, test_name) \ |
| __TEST_F_IMPL(fixture_name, test_name, -1, TEST_TIMEOUT_DEFAULT) |
| |
| #define TEST_F_SIGNAL(fixture_name, test_name, signal) \ |
| __TEST_F_IMPL(fixture_name, test_name, signal, TEST_TIMEOUT_DEFAULT) |
| |
| #define TEST_F_TIMEOUT(fixture_name, test_name, timeout) \ |
| __TEST_F_IMPL(fixture_name, test_name, -1, timeout) |
| |
| #define __TEST_F_IMPL(fixture_name, test_name, signal, tmout) \ |
| static void fixture_name##_##test_name( \ |
| struct __test_metadata *_metadata, \ |
| FIXTURE_DATA(fixture_name) *self, \ |
| const FIXTURE_VARIANT(fixture_name) *variant); \ |
| static inline void wrapper_##fixture_name##_##test_name( \ |
| struct __test_metadata *_metadata, \ |
| struct __fixture_variant_metadata *variant) \ |
| { \ |
| /* fixture data is alloced, setup, and torn down per call. */ \ |
| FIXTURE_DATA(fixture_name) self; \ |
| memset(&self, 0, sizeof(FIXTURE_DATA(fixture_name))); \ |
| if (setjmp(_metadata->env) == 0) { \ |
| fixture_name##_setup(_metadata, &self, variant->data); \ |
| /* Let setup failure terminate early. */ \ |
| if (!_metadata->passed) \ |
| return; \ |
| _metadata->setup_completed = true; \ |
| fixture_name##_##test_name(_metadata, &self, variant->data); \ |
| } \ |
| if (_metadata->setup_completed) \ |
| fixture_name##_teardown(_metadata, &self, variant->data); \ |
| __test_check_assert(_metadata); \ |
| } \ |
| static struct __test_metadata \ |
| _##fixture_name##_##test_name##_object = { \ |
| .name = #test_name, \ |
| .fn = &wrapper_##fixture_name##_##test_name, \ |
| .fixture = &_##fixture_name##_fixture_object, \ |
| .termsig = signal, \ |
| .timeout = tmout, \ |
| }; \ |
| static void __attribute__((constructor)) \ |
| _register_##fixture_name##_##test_name(void) \ |
| { \ |
| __register_test(&_##fixture_name##_##test_name##_object); \ |
| } \ |
| static void fixture_name##_##test_name( \ |
| struct __test_metadata __attribute__((unused)) *_metadata, \ |
| FIXTURE_DATA(fixture_name) __attribute__((unused)) *self, \ |
| const FIXTURE_VARIANT(fixture_name) \ |
| __attribute__((unused)) *variant) |
| |
| /** |
| * TEST_HARNESS_MAIN - Simple wrapper to run the test harness |
| * |
| * .. code-block:: c |
| * |
| * TEST_HARNESS_MAIN |
| * |
| * Use once to append a main() to the test file. |
| */ |
| #define TEST_HARNESS_MAIN \ |
| static void __attribute__((constructor)) \ |
| __constructor_order_last(void) \ |
| { \ |
| if (!__constructor_order) \ |
| __constructor_order = _CONSTRUCTOR_ORDER_BACKWARD; \ |
| } \ |
| int main(int argc, char **argv) { \ |
| return test_harness_run(argc, argv); \ |
| } |
| |
| /** |
| * DOC: operators |
| * |
| * Operators for use in TEST() and TEST_F(). |
| * ASSERT_* calls will stop test execution immediately. |
| * EXPECT_* calls will emit a failure warning, note it, and continue. |
| */ |
| |
| /** |
| * ASSERT_EQ() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_EQ(expected, measured): expected == measured |
| */ |
| #define ASSERT_EQ(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, ==, 1) |
| |
| /** |
| * ASSERT_NE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_NE(expected, measured): expected != measured |
| */ |
| #define ASSERT_NE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, !=, 1) |
| |
| /** |
| * ASSERT_LT() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_LT(expected, measured): expected < measured |
| */ |
| #define ASSERT_LT(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, <, 1) |
| |
| /** |
| * ASSERT_LE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_LE(expected, measured): expected <= measured |
| */ |
| #define ASSERT_LE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, <=, 1) |
| |
| /** |
| * ASSERT_GT() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_GT(expected, measured): expected > measured |
| */ |
| #define ASSERT_GT(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, >, 1) |
| |
| /** |
| * ASSERT_GE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_GE(expected, measured): expected >= measured |
| */ |
| #define ASSERT_GE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, >=, 1) |
| |
| /** |
| * ASSERT_NULL() |
| * |
| * @seen: measured value |
| * |
| * ASSERT_NULL(measured): NULL == measured |
| */ |
| #define ASSERT_NULL(seen) \ |
| __EXPECT(NULL, "NULL", seen, #seen, ==, 1) |
| |
| /** |
| * ASSERT_TRUE() |
| * |
| * @seen: measured value |
| * |
| * ASSERT_TRUE(measured): measured != 0 |
| */ |
| #define ASSERT_TRUE(seen) \ |
| __EXPECT(0, "0", seen, #seen, !=, 1) |
| |
| /** |
| * ASSERT_FALSE() |
| * |
| * @seen: measured value |
| * |
| * ASSERT_FALSE(measured): measured == 0 |
| */ |
| #define ASSERT_FALSE(seen) \ |
| __EXPECT(0, "0", seen, #seen, ==, 1) |
| |
| /** |
| * ASSERT_STREQ() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_STREQ(expected, measured): !strcmp(expected, measured) |
| */ |
| #define ASSERT_STREQ(expected, seen) \ |
| __EXPECT_STR(expected, seen, ==, 1) |
| |
| /** |
| * ASSERT_STRNE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * ASSERT_STRNE(expected, measured): strcmp(expected, measured) |
| */ |
| #define ASSERT_STRNE(expected, seen) \ |
| __EXPECT_STR(expected, seen, !=, 1) |
| |
| /** |
| * EXPECT_EQ() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_EQ(expected, measured): expected == measured |
| */ |
| #define EXPECT_EQ(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, ==, 0) |
| |
| /** |
| * EXPECT_NE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_NE(expected, measured): expected != measured |
| */ |
| #define EXPECT_NE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, !=, 0) |
| |
| /** |
| * EXPECT_LT() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_LT(expected, measured): expected < measured |
| */ |
| #define EXPECT_LT(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, <, 0) |
| |
| /** |
| * EXPECT_LE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_LE(expected, measured): expected <= measured |
| */ |
| #define EXPECT_LE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, <=, 0) |
| |
| /** |
| * EXPECT_GT() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_GT(expected, measured): expected > measured |
| */ |
| #define EXPECT_GT(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, >, 0) |
| |
| /** |
| * EXPECT_GE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_GE(expected, measured): expected >= measured |
| */ |
| #define EXPECT_GE(expected, seen) \ |
| __EXPECT(expected, #expected, seen, #seen, >=, 0) |
| |
| /** |
| * EXPECT_NULL() |
| * |
| * @seen: measured value |
| * |
| * EXPECT_NULL(measured): NULL == measured |
| */ |
| #define EXPECT_NULL(seen) \ |
| __EXPECT(NULL, "NULL", seen, #seen, ==, 0) |
| |
| /** |
| * EXPECT_TRUE() |
| * |
| * @seen: measured value |
| * |
| * EXPECT_TRUE(measured): 0 != measured |
| */ |
| #define EXPECT_TRUE(seen) \ |
| __EXPECT(0, "0", seen, #seen, !=, 0) |
| |
| /** |
| * EXPECT_FALSE() |
| * |
| * @seen: measured value |
| * |
| * EXPECT_FALSE(measured): 0 == measured |
| */ |
| #define EXPECT_FALSE(seen) \ |
| __EXPECT(0, "0", seen, #seen, ==, 0) |
| |
| /** |
| * EXPECT_STREQ() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_STREQ(expected, measured): !strcmp(expected, measured) |
| */ |
| #define EXPECT_STREQ(expected, seen) \ |
| __EXPECT_STR(expected, seen, ==, 0) |
| |
| /** |
| * EXPECT_STRNE() |
| * |
| * @expected: expected value |
| * @seen: measured value |
| * |
| * EXPECT_STRNE(expected, measured): strcmp(expected, measured) |
| */ |
| #define EXPECT_STRNE(expected, seen) \ |
| __EXPECT_STR(expected, seen, !=, 0) |
| |
| #ifndef ARRAY_SIZE |
| #define ARRAY_SIZE(a) (sizeof(a) / sizeof(a[0])) |
| #endif |
| |
| /* Support an optional handler after and ASSERT_* or EXPECT_*. The approach is |
| * not thread-safe, but it should be fine in most sane test scenarios. |
| * |
| * Using __bail(), which optionally abort()s, is the easiest way to early |
| * return while still providing an optional block to the API consumer. |
| */ |
| #define OPTIONAL_HANDLER(_assert) \ |
| for (; _metadata->trigger; _metadata->trigger = \ |
| __bail(_assert, _metadata)) |
| |
| #define __INC_STEP(_metadata) \ |
| /* Keep "step" below 255 (which is used for "SKIP" reporting). */ \ |
| if (_metadata->passed && _metadata->step < 253) \ |
| _metadata->step++; |
| |
| #define is_signed_type(var) (!!(((__typeof__(var))(-1)) < (__typeof__(var))1)) |
| |
| #define __EXPECT(_expected, _expected_str, _seen, _seen_str, _t, _assert) do { \ |
| /* Avoid multiple evaluation of the cases */ \ |
| __typeof__(_expected) __exp = (_expected); \ |
| __typeof__(_seen) __seen = (_seen); \ |
| if (_assert) __INC_STEP(_metadata); \ |
| if (!(__exp _t __seen)) { \ |
| /* Report with actual signedness to avoid weird output. */ \ |
| switch (is_signed_type(__exp) * 2 + is_signed_type(__seen)) { \ |
| case 0: { \ |
| unsigned long long __exp_print = (uintptr_t)__exp; \ |
| unsigned long long __seen_print = (uintptr_t)__seen; \ |
| __TH_LOG("Expected %s (%llu) %s %s (%llu)", \ |
| _expected_str, __exp_print, #_t, \ |
| _seen_str, __seen_print); \ |
| break; \ |
| } \ |
| case 1: { \ |
| unsigned long long __exp_print = (uintptr_t)__exp; \ |
| long long __seen_print = (intptr_t)__seen; \ |
| __TH_LOG("Expected %s (%llu) %s %s (%lld)", \ |
| _expected_str, __exp_print, #_t, \ |
| _seen_str, __seen_print); \ |
| break; \ |
| } \ |
| case 2: { \ |
| long long __exp_print = (intptr_t)__exp; \ |
| unsigned long long __seen_print = (uintptr_t)__seen; \ |
| __TH_LOG("Expected %s (%lld) %s %s (%llu)", \ |
| _expected_str, __exp_print, #_t, \ |
| _seen_str, __seen_print); \ |
| break; \ |
| } \ |
| case 3: { \ |
| long long __exp_print = (intptr_t)__exp; \ |
| long long __seen_print = (intptr_t)__seen; \ |
| __TH_LOG("Expected %s (%lld) %s %s (%lld)", \ |
| _expected_str, __exp_print, #_t, \ |
| _seen_str, __seen_print); \ |
| break; \ |
| } \ |
| } \ |
| _metadata->passed = 0; \ |
| /* Ensure the optional handler is triggered */ \ |
| _metadata->trigger = 1; \ |
| } \ |
| } while (0); OPTIONAL_HANDLER(_assert) |
| |
| #define __EXPECT_STR(_expected, _seen, _t, _assert) do { \ |
| const char *__exp = (_expected); \ |
| const char *__seen = (_seen); \ |
| if (_assert) __INC_STEP(_metadata); \ |
| if (!(strcmp(__exp, __seen) _t 0)) { \ |
| __TH_LOG("Expected '%s' %s '%s'.", __exp, #_t, __seen); \ |
| _metadata->passed = 0; \ |
| _metadata->trigger = 1; \ |
| } \ |
| } while (0); OPTIONAL_HANDLER(_assert) |
| |
| /* List helpers */ |
| #define __LIST_APPEND(head, item) \ |
| { \ |
| /* Circular linked list where only prev is circular. */ \ |
| if (head == NULL) { \ |
| head = item; \ |
| item->next = NULL; \ |
| item->prev = item; \ |
| return; \ |
| } \ |
| if (__constructor_order == _CONSTRUCTOR_ORDER_FORWARD) { \ |
| item->next = NULL; \ |
| item->prev = head->prev; \ |
| item->prev->next = item; \ |
| head->prev = item; \ |
| } else { \ |
| item->next = head; \ |
| item->next->prev = item; \ |
| item->prev = item; \ |
| head = item; \ |
| } \ |
| } |
| |
| struct __test_results { |
| char reason[1024]; /* Reason for test result */ |
| }; |
| |
| struct __test_metadata; |
| struct __fixture_variant_metadata; |
| |
| /* Contains all the information about a fixture. */ |
| struct __fixture_metadata { |
| const char *name; |
| struct __test_metadata *tests; |
| struct __fixture_variant_metadata *variant; |
| struct __fixture_metadata *prev, *next; |
| } _fixture_global __attribute__((unused)) = { |
| .name = "global", |
| .prev = &_fixture_global, |
| }; |
| |
| static struct __fixture_metadata *__fixture_list = &_fixture_global; |
| static int __constructor_order; |
| |
| #define _CONSTRUCTOR_ORDER_FORWARD 1 |
| #define _CONSTRUCTOR_ORDER_BACKWARD -1 |
| |
| static inline void __register_fixture(struct __fixture_metadata *f) |
| { |
| __LIST_APPEND(__fixture_list, f); |
| } |
| |
| struct __fixture_variant_metadata { |
| const char *name; |
| const void *data; |
| struct __fixture_variant_metadata *prev, *next; |
| }; |
| |
| static inline void |
| __register_fixture_variant(struct __fixture_metadata *f, |
| struct __fixture_variant_metadata *variant) |
| { |
| __LIST_APPEND(f->variant, variant); |
| } |
| |
| /* Contains all the information for test execution and status checking. */ |
| struct __test_metadata { |
| const char *name; |
| void (*fn)(struct __test_metadata *, |
| struct __fixture_variant_metadata *); |
| pid_t pid; /* pid of test when being run */ |
| struct __fixture_metadata *fixture; |
| int termsig; |
| int passed; |
| int skip; /* did SKIP get used? */ |
| int trigger; /* extra handler after the evaluation */ |
| int timeout; /* seconds to wait for test timeout */ |
| bool timed_out; /* did this test timeout instead of exiting? */ |
| __u8 step; |
| bool no_print; /* manual trigger when TH_LOG_STREAM is not available */ |
| bool aborted; /* stopped test due to failed ASSERT */ |
| bool setup_completed; /* did setup finish? */ |
| jmp_buf env; /* for exiting out of test early */ |
| struct __test_results *results; |
| struct __test_metadata *prev, *next; |
| }; |
| |
| /* |
| * Since constructors are called in reverse order, reverse the test |
| * list so tests are run in source declaration order. |
| * https://gcc.gnu.org/onlinedocs/gccint/Initialization.html |
| * However, it seems not all toolchains do this correctly, so use |
| * __constructor_order to detect which direction is called first |
| * and adjust list building logic to get things running in the right |
| * direction. |
| */ |
| static inline void __register_test(struct __test_metadata *t) |
| { |
| __LIST_APPEND(t->fixture->tests, t); |
| } |
| |
| static inline int __bail(int for_realz, struct __test_metadata *t) |
| { |
| /* if this is ASSERT, return immediately. */ |
| if (for_realz) { |
| t->aborted = true; |
| longjmp(t->env, 1); |
| } |
| /* otherwise, end the for loop and continue. */ |
| return 0; |
| } |
| |
| static inline void __test_check_assert(struct __test_metadata *t) |
| { |
| if (t->aborted) { |
| if (t->no_print) |
| _exit(t->step); |
| abort(); |
| } |
| } |
| |
| struct __test_metadata *__active_test; |
| static void __timeout_handler(int sig, siginfo_t *info, void *ucontext) |
| { |
| struct __test_metadata *t = __active_test; |
| |
| /* Sanity check handler execution environment. */ |
| if (!t) { |
| fprintf(TH_LOG_STREAM, |
| "# no active test in SIGALRM handler!?\n"); |
| abort(); |
| } |
| if (sig != SIGALRM || sig != info->si_signo) { |
| fprintf(TH_LOG_STREAM, |
| "# %s: SIGALRM handler caught signal %d!?\n", |
| t->name, sig != SIGALRM ? sig : info->si_signo); |
| abort(); |
| } |
| |
| t->timed_out = true; |
| // signal process group |
| kill(-(t->pid), SIGKILL); |
| } |
| |
| void __wait_for_test(struct __test_metadata *t) |
| { |
| struct sigaction action = { |
| .sa_sigaction = __timeout_handler, |
| .sa_flags = SA_SIGINFO, |
| }; |
| struct sigaction saved_action; |
| int status; |
| |
| if (sigaction(SIGALRM, &action, &saved_action)) { |
| t->passed = 0; |
| fprintf(TH_LOG_STREAM, |
| "# %s: unable to install SIGALRM handler\n", |
| t->name); |
| return; |
| } |
| __active_test = t; |
| t->timed_out = false; |
| alarm(t->timeout); |
| waitpid(t->pid, &status, 0); |
| alarm(0); |
| if (sigaction(SIGALRM, &saved_action, NULL)) { |
| t->passed = 0; |
| fprintf(TH_LOG_STREAM, |
| "# %s: unable to uninstall SIGALRM handler\n", |
| t->name); |
| return; |
| } |
| __active_test = NULL; |
| |
| if (t->timed_out) { |
| t->passed = 0; |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test terminated by timeout\n", t->name); |
| } else if (WIFEXITED(status)) { |
| if (t->termsig != -1) { |
| t->passed = 0; |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test exited normally instead of by signal (code: %d)\n", |
| t->name, |
| WEXITSTATUS(status)); |
| } else { |
| switch (WEXITSTATUS(status)) { |
| /* Success */ |
| case 0: |
| t->passed = 1; |
| break; |
| /* SKIP */ |
| case 255: |
| t->passed = 1; |
| t->skip = 1; |
| break; |
| /* Other failure, assume step report. */ |
| default: |
| t->passed = 0; |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test failed at step #%d\n", |
| t->name, |
| WEXITSTATUS(status)); |
| } |
| } |
| } else if (WIFSIGNALED(status)) { |
| t->passed = 0; |
| if (WTERMSIG(status) == SIGABRT) { |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test terminated by assertion\n", |
| t->name); |
| } else if (WTERMSIG(status) == t->termsig) { |
| t->passed = 1; |
| } else { |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test terminated unexpectedly by signal %d\n", |
| t->name, |
| WTERMSIG(status)); |
| } |
| } else { |
| fprintf(TH_LOG_STREAM, |
| "# %s: Test ended in some other way [%u]\n", |
| t->name, |
| status); |
| } |
| } |
| |
| void __run_test(struct __fixture_metadata *f, |
| struct __fixture_variant_metadata *variant, |
| struct __test_metadata *t) |
| { |
| /* reset test struct */ |
| t->passed = 1; |
| t->skip = 0; |
| t->trigger = 0; |
| t->step = 1; |
| t->no_print = 0; |
| memset(t->results->reason, 0, sizeof(t->results->reason)); |
| |
| ksft_print_msg(" RUN %s%s%s.%s ...\n", |
| f->name, variant->name[0] ? "." : "", variant->name, t->name); |
| |
| /* Make sure output buffers are flushed before fork */ |
| fflush(stdout); |
| fflush(stderr); |
| |
| t->pid = fork(); |
| if (t->pid < 0) { |
| ksft_print_msg("ERROR SPAWNING TEST CHILD\n"); |
| t->passed = 0; |
| } else if (t->pid == 0) { |
| setpgrp(); |
| t->fn(t, variant); |
| if (t->skip) |
| _exit(255); |
| /* Pass is exit 0 */ |
| if (t->passed) |
| _exit(0); |
| /* Something else happened, report the step. */ |
| _exit(t->step); |
| } else { |
| __wait_for_test(t); |
| } |
| ksft_print_msg(" %4s %s%s%s.%s\n", t->passed ? "OK" : "FAIL", |
| f->name, variant->name[0] ? "." : "", variant->name, t->name); |
| |
| if (t->skip) |
| ksft_test_result_skip("%s\n", t->results->reason[0] ? |
| t->results->reason : "unknown"); |
| else |
| ksft_test_result(t->passed, "%s%s%s.%s\n", |
| f->name, variant->name[0] ? "." : "", variant->name, t->name); |
| } |
| |
| static int test_harness_run(int __attribute__((unused)) argc, |
| char __attribute__((unused)) **argv) |
| { |
| struct __fixture_variant_metadata no_variant = { .name = "", }; |
| struct __fixture_variant_metadata *v; |
| struct __fixture_metadata *f; |
| struct __test_results *results; |
| struct __test_metadata *t; |
| int ret = 0; |
| unsigned int case_count = 0, test_count = 0; |
| unsigned int count = 0; |
| unsigned int pass_count = 0; |
| |
| for (f = __fixture_list; f; f = f->next) { |
| for (v = f->variant ?: &no_variant; v; v = v->next) { |
| case_count++; |
| for (t = f->tests; t; t = t->next) |
| test_count++; |
| } |
| } |
| |
| results = mmap(NULL, sizeof(*results), PROT_READ | PROT_WRITE, |
| MAP_SHARED | MAP_ANONYMOUS, -1, 0); |
| |
| ksft_print_header(); |
| ksft_set_plan(test_count); |
| ksft_print_msg("Starting %u tests from %u test cases.\n", |
| test_count, case_count); |
| for (f = __fixture_list; f; f = f->next) { |
| for (v = f->variant ?: &no_variant; v; v = v->next) { |
| for (t = f->tests; t; t = t->next) { |
| count++; |
| t->results = results; |
| __run_test(f, v, t); |
| t->results = NULL; |
| if (t->passed) |
| pass_count++; |
| else |
| ret = 1; |
| } |
| } |
| } |
| munmap(results, sizeof(*results)); |
| |
| ksft_print_msg("%s: %u / %u tests passed.\n", ret ? "FAILED" : "PASSED", |
| pass_count, count); |
| ksft_exit(ret == 0); |
| |
| /* unreachable */ |
| return KSFT_FAIL; |
| } |
| |
| static void __attribute__((constructor)) __constructor_order_first(void) |
| { |
| if (!__constructor_order) |
| __constructor_order = _CONSTRUCTOR_ORDER_FORWARD; |
| } |
| |
| #endif /* __KSELFTEST_HARNESS_H */ |