6c8ae362bb
This commit is truly horrible. We want to run ASan with leak checking enabled in CI so we can catch memory leaks before they're introduced. This works well with Pixman, and with NIR-only drivers like iris or Panfrost. But when we run under llvmpipe - which we do under CI - we start failing because: - Mesa pulls in llvmpipe via dlopen - llvmpipe pulls in LLVM itself via DT_NEEDED - initialising LLVM's global type/etc systems performs thread-local allocations - llvmpipe can't free those allocations since the application might also be using LLVM - Weston stops using GL and destroys all GL objects, leading to Mesa unloading llvmpipe like it should - with everything disappearing from the process's vmap, ASan can no longer keep track of still-reachable pointers - tests fail because LLVM is 'leaking' Usually, an alternative is to LD_PRELOAD a shim which overrides dlclose() to be a no-op. This is not usable here, because when $LD_PRELOAD is not empty and ASan is not first in it, ASan immediately errors out. Prepending ASan doesn't work, because we run our tests through Meson (which also invokes Ninja), leading to LSan exploding over CPython and Ninja, which is not what we're interested in. It would be possible to inject _both_ ASan and a dlclose-does-nothing shim DSO into the LD_PRELOAD environment for every test, but that seems even worse, especially as Meson strongly discourages globbing for random files in the root. So, here we are, doing what we can: finding where swrast_dri.so (aka llvmpipe) lives, stashing that in an environment variable, and deliberately leaking a dlopen handle which we never close to ensure that neither llvmpipe nor LLVM leave our process's address space before we exit. Signed-off-by: Daniel Stone <daniels@collabora.com>
703 lines
18 KiB
C
703 lines
18 KiB
C
/*
|
|
* Copyright © 2012 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining
|
|
* a copy of this software and associated documentation files (the
|
|
* "Software"), to deal in the Software without restriction, including
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
* the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the
|
|
* next paragraph) shall be included in all copies or substantial
|
|
* portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
#include "config.h"
|
|
|
|
#include <unistd.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <sys/wait.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <signal.h>
|
|
#include <getopt.h>
|
|
#include <dlfcn.h>
|
|
|
|
#include "test-config.h"
|
|
#include "weston-test-runner.h"
|
|
#include "weston-testsuite-data.h"
|
|
#include "shared/string-helpers.h"
|
|
|
|
/* This is a glibc extension; if we can't use it then make it harmless. */
|
|
#ifndef RTLD_NODELETE
|
|
#define RTLD_NODELETE 0
|
|
#endif
|
|
|
|
/**
|
|
* \defgroup testharness Test harness
|
|
* \defgroup testharness_private Test harness private
|
|
*/
|
|
|
|
extern const struct weston_test_entry __start_test_section, __stop_test_section;
|
|
|
|
struct weston_test_run_info {
|
|
char name[512];
|
|
int fixture_nr;
|
|
};
|
|
|
|
static const struct weston_test_run_info *test_run_info_;
|
|
|
|
/** Get the test name string with counter
|
|
*
|
|
* \return The test name with fixture number \c -f%%d added. For an array
|
|
* driven test, e.g. defined with TEST_P(), the name has also a \c -e%%d
|
|
* suffix to indicate the array element number.
|
|
*
|
|
* This is only usable from code paths inside TEST(), TEST_P(), PLUGIN_TEST()
|
|
* etc. defined functions.
|
|
*
|
|
* \ingroup testharness
|
|
*/
|
|
const char *
|
|
get_test_name(void)
|
|
{
|
|
return test_run_info_->name;
|
|
}
|
|
|
|
/** Get the current fixture index
|
|
*
|
|
* Returns the current fixture index which can be used directly as an index
|
|
* into the array passed as an argument to DECLARE_FIXTURE_SETUP_WITH_ARG().
|
|
*
|
|
* This is only usable from code paths inside TEST(), TEST_P(), PLUGIN_TEST()
|
|
* etc. defined functions.
|
|
*
|
|
* \ingroup testharness
|
|
*/
|
|
int
|
|
get_test_fixture_index(void)
|
|
{
|
|
return test_run_info_->fixture_nr - 1;
|
|
}
|
|
|
|
/** Print into test log
|
|
*
|
|
* This is exactly like printf() except the output goes to the test log,
|
|
* which is at stderr.
|
|
*
|
|
* \param fmt printf format string
|
|
*
|
|
* \ingroup testharness
|
|
*/
|
|
void
|
|
testlog(const char *fmt, ...)
|
|
{
|
|
va_list argp;
|
|
|
|
va_start(argp, fmt);
|
|
vfprintf(stderr, fmt, argp);
|
|
va_end(argp);
|
|
}
|
|
|
|
static const void *
|
|
fixture_setup_array_get_arg(const struct fixture_setup_array *fsa, int findex)
|
|
{
|
|
const char *array_data = fsa->array;
|
|
|
|
if (array_data)
|
|
return array_data + findex * fsa->element_size;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const char *
|
|
fixture_setup_array_get_name(const struct fixture_setup_array *fsa, int findex)
|
|
{
|
|
const char *element_data = fixture_setup_array_get_arg(fsa, findex);
|
|
const struct fixture_metadata *meta;
|
|
|
|
if (!element_data)
|
|
return "";
|
|
|
|
meta = (const void *)(element_data + fsa->meta_offset);
|
|
|
|
return meta->name;
|
|
}
|
|
|
|
static const struct weston_test_entry *
|
|
find_test(const char *name)
|
|
{
|
|
const struct weston_test_entry *t;
|
|
|
|
for (t = &__start_test_section; t < &__stop_test_section; t++)
|
|
if (strcmp(t->name, name) == 0)
|
|
return t;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static enum test_result_code
|
|
run_test(int fixture_nr, const struct weston_test_entry *t, void *data,
|
|
int iteration)
|
|
{
|
|
struct weston_test_run_info info;
|
|
|
|
if (data) {
|
|
snprintf(info.name, sizeof(info.name), "%s-f%02d-e%02d",
|
|
t->name, fixture_nr, iteration);
|
|
} else {
|
|
snprintf(info.name, sizeof(info.name), "%s-f%02d",
|
|
t->name, fixture_nr);
|
|
}
|
|
info.fixture_nr = fixture_nr;
|
|
|
|
test_run_info_ = &info;
|
|
t->run(data);
|
|
test_run_info_ = NULL;
|
|
|
|
/*
|
|
* XXX: We should return t->run(data); but that requires changing
|
|
* the function signature and stop using assert() in tests.
|
|
* https://gitlab.freedesktop.org/wayland/weston/issues/311
|
|
*/
|
|
return RESULT_OK;
|
|
}
|
|
|
|
static void
|
|
list_tests(void)
|
|
{
|
|
const struct fixture_setup_array *fsa;
|
|
const struct weston_test_entry *t;
|
|
int i;
|
|
|
|
fsa = fixture_setup_array_get_();
|
|
|
|
if (fsa->n_elements > 1) {
|
|
printf("Fixture setups:\n");
|
|
for (i = 0; i < fsa->n_elements; i++) {
|
|
printf("%2d: %s\n", i + 1,
|
|
fixture_setup_array_get_name(fsa, i));
|
|
}
|
|
} else {
|
|
printf("One fixture setup.\n");
|
|
}
|
|
|
|
printf("Test names:\n");
|
|
for (t = &__start_test_section; t < &__stop_test_section; t++) {
|
|
printf(" %s\n", t->name);
|
|
if (t->n_elements > 1)
|
|
printf(" with array of %d cases\n", t->n_elements);
|
|
}
|
|
}
|
|
|
|
struct weston_test_harness {
|
|
int32_t fixt_ind;
|
|
char *chosen_testname;
|
|
int32_t case_ind;
|
|
|
|
struct wet_testsuite_data data;
|
|
};
|
|
|
|
typedef void (*weston_test_cb)(struct wet_testsuite_data *suite_data,
|
|
const struct weston_test_entry *t,
|
|
const void *test_data,
|
|
int iteration);
|
|
|
|
static void
|
|
for_each_test_case(struct wet_testsuite_data *data, weston_test_cb cb)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < data->tests_count; i++) {
|
|
const struct weston_test_entry *t = &data->tests[i];
|
|
const void *current_test_data = t->table_data;
|
|
int elem;
|
|
int elem_end;
|
|
|
|
if (data->case_index == -1) {
|
|
elem = 0;
|
|
elem_end = t->n_elements;
|
|
} else {
|
|
elem = data->case_index;
|
|
elem_end = elem + 1;
|
|
}
|
|
|
|
for (; elem < elem_end; elem++) {
|
|
current_test_data = (char *)t->table_data +
|
|
elem * t->element_size;
|
|
cb(data, t, current_test_data, elem);
|
|
}
|
|
}
|
|
}
|
|
|
|
static const char *
|
|
result_to_str(enum test_result_code ret)
|
|
{
|
|
static const char *names[] = {
|
|
[RESULT_FAIL] = "fail",
|
|
[RESULT_HARD_ERROR] = "hard error",
|
|
[RESULT_OK] = "ok",
|
|
#if WESTON_TEST_SKIP_IS_FAILURE
|
|
[RESULT_SKIP] = "skip error",
|
|
#else
|
|
[RESULT_SKIP] = "skip",
|
|
#endif
|
|
};
|
|
|
|
assert(ret >= 0 && ret < ARRAY_LENGTH(names));
|
|
return names[ret];
|
|
}
|
|
|
|
static void
|
|
run_case(struct wet_testsuite_data *suite_data,
|
|
const struct weston_test_entry *t,
|
|
const void *test_data,
|
|
int iteration)
|
|
{
|
|
enum test_result_code ret;
|
|
const char *fail = "";
|
|
const char *skip = "";
|
|
int fixture_nr = suite_data->fixture_iteration + 1;
|
|
int iteration_nr = iteration + 1;
|
|
|
|
testlog("*** Run %s %s/%d\n",
|
|
suite_data->fixture_name, t->name, iteration_nr);
|
|
|
|
if (suite_data->type == TEST_TYPE_PLUGIN) {
|
|
ret = run_test(fixture_nr, t, suite_data->compositor,
|
|
iteration);
|
|
} else {
|
|
ret = run_test(fixture_nr, t, (void *)test_data, iteration);
|
|
}
|
|
|
|
switch (ret) {
|
|
case RESULT_OK:
|
|
suite_data->passed++;
|
|
break;
|
|
case RESULT_FAIL:
|
|
case RESULT_HARD_ERROR:
|
|
suite_data->failed++;
|
|
fail = "not ";
|
|
break;
|
|
case RESULT_SKIP:
|
|
suite_data->skipped++;
|
|
skip = " # SKIP";
|
|
#if WESTON_TEST_SKIP_IS_FAILURE
|
|
fail = "not ";
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
testlog("*** Result %s %s/%d: %s\n",
|
|
suite_data->fixture_name, t->name, iteration_nr,
|
|
result_to_str(ret));
|
|
|
|
suite_data->counter++;
|
|
printf("%sok %d %s %s/%d%s\n", fail, suite_data->counter,
|
|
suite_data->fixture_name, t->name, iteration_nr, skip);
|
|
}
|
|
|
|
/* This function might run in a new thread */
|
|
static void
|
|
testsuite_run(struct wet_testsuite_data *data)
|
|
{
|
|
for_each_test_case(data, run_case);
|
|
}
|
|
|
|
static void
|
|
count_case(struct wet_testsuite_data *suite_data,
|
|
const struct weston_test_entry *t,
|
|
const void *test_data,
|
|
int iteration)
|
|
{
|
|
suite_data->total++;
|
|
}
|
|
|
|
static void
|
|
tap_plan(struct wet_testsuite_data *data, int count_fixtures)
|
|
{
|
|
data->total = 0;
|
|
for_each_test_case(data, count_case);
|
|
|
|
printf("1..%d\n", data->total * count_fixtures);
|
|
}
|
|
|
|
static void
|
|
skip_case(struct wet_testsuite_data *suite_data,
|
|
const struct weston_test_entry *t,
|
|
const void *test_data,
|
|
int iteration)
|
|
{
|
|
const char *skip_error = "";
|
|
int iteration_nr = iteration + 1;
|
|
|
|
#if WESTON_TEST_SKIP_IS_FAILURE
|
|
skip_error = "not ";
|
|
#endif
|
|
|
|
suite_data->counter++;
|
|
printf("%sok %d %s %s/%d # SKIP fixture\n",
|
|
skip_error, suite_data->counter,
|
|
suite_data->fixture_name, t->name, iteration_nr);
|
|
}
|
|
|
|
static void
|
|
tap_skip_fixture(struct wet_testsuite_data *data)
|
|
{
|
|
for_each_test_case(data, skip_case);
|
|
}
|
|
|
|
static void
|
|
help(const char *exe)
|
|
{
|
|
printf(
|
|
"Usage: %s [options] [testname [number]]\n"
|
|
"\n"
|
|
"This is a Weston test suite executable that runs some tests.\n"
|
|
"Options:\n"
|
|
" -f, --fixture N Run only fixture number N. 0 runs all (default).\n"
|
|
" -h, --help Print this help and exit with success.\n"
|
|
" -l, --list List all tests in this executable and exit with success.\n"
|
|
"testname: Optional; name of the test to execute instead of all tests.\n"
|
|
"number: Optional; for a multi-case test, run the given case only.\n"
|
|
"Both fixture and case numbering starts from 1.\n",
|
|
exe);
|
|
}
|
|
|
|
static void
|
|
parse_command_line(struct weston_test_harness *harness, int argc, char **argv)
|
|
{
|
|
int c;
|
|
static const struct option opts[] = {
|
|
{ "fixture", required_argument, NULL, 'f' },
|
|
{ "help", no_argument, NULL, 'h' },
|
|
{ "list", no_argument, NULL, 'l' },
|
|
{ 0, 0, NULL, 0 }
|
|
};
|
|
|
|
while ((c = getopt_long(argc, argv, "f:hl", opts, NULL)) != -1) {
|
|
switch (c) {
|
|
case 'f':
|
|
if (!safe_strtoint(optarg, &harness->fixt_ind)) {
|
|
fprintf(stderr,
|
|
"Error: '%s' does not look like a number (command line).\n",
|
|
optarg);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
harness->fixt_ind--; /* convert base-1 to base 0 */
|
|
break;
|
|
case 'h':
|
|
help(argv[0]);
|
|
exit(RESULT_OK);
|
|
case 'l':
|
|
list_tests();
|
|
exit(RESULT_OK);
|
|
case 0:
|
|
break;
|
|
default:
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
}
|
|
|
|
if (optind < argc)
|
|
harness->chosen_testname = argv[optind++];
|
|
|
|
if (optind < argc) {
|
|
if (!safe_strtoint(argv[optind], &harness->case_ind)) {
|
|
fprintf(stderr,
|
|
"Error: '%s' does not look like a number (command line).\n",
|
|
argv[optind]);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
harness->case_ind--; /* convert base-1 to base 0 */
|
|
optind++;
|
|
}
|
|
|
|
if (optind < argc) {
|
|
fprintf(stderr, "Unexpected extra arguments given (command line).\n\n");
|
|
help(argv[0]);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
}
|
|
|
|
static struct weston_test_harness *
|
|
weston_test_harness_create(int argc, char **argv)
|
|
{
|
|
const struct fixture_setup_array *fsa;
|
|
struct weston_test_harness *harness;
|
|
|
|
harness = zalloc(sizeof(*harness));
|
|
assert(harness);
|
|
|
|
harness->fixt_ind = -1;
|
|
harness->case_ind = -1;
|
|
parse_command_line(harness, argc, argv);
|
|
|
|
fsa = fixture_setup_array_get_();
|
|
if (harness->fixt_ind < -1 || harness->fixt_ind >= fsa->n_elements) {
|
|
fprintf(stderr,
|
|
"Error: fixture index %d (command line) is invalid for this program.\n",
|
|
harness->fixt_ind + 1);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
|
|
if (harness->chosen_testname) {
|
|
const struct weston_test_entry *t;
|
|
|
|
t = find_test(harness->chosen_testname);
|
|
if (!t) {
|
|
fprintf(stderr,
|
|
"Error: test '%s' not found (command line).\n",
|
|
harness->chosen_testname);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
|
|
if (harness->case_ind < -1 ||
|
|
harness->case_ind >= t->n_elements) {
|
|
fprintf(stderr,
|
|
"Error: case index %d (command line) is invalid for this test.\n",
|
|
harness->case_ind + 1);
|
|
exit(RESULT_HARD_ERROR);
|
|
}
|
|
|
|
harness->data.tests = t;
|
|
harness->data.tests_count = 1;
|
|
harness->data.case_index = harness->case_ind;
|
|
} else {
|
|
harness->data.tests = &__start_test_section;
|
|
harness->data.tests_count =
|
|
&__stop_test_section - &__start_test_section;
|
|
harness->data.case_index = -1;
|
|
}
|
|
|
|
harness->data.run = testsuite_run;
|
|
|
|
return harness;
|
|
}
|
|
|
|
static void
|
|
weston_test_harness_destroy(struct weston_test_harness *harness)
|
|
{
|
|
free(harness);
|
|
}
|
|
|
|
static enum test_result_code
|
|
counts_to_result(const struct wet_testsuite_data *data)
|
|
{
|
|
#if WESTON_TEST_SKIP_IS_FAILURE
|
|
if (data->skipped > 0)
|
|
return RESULT_FAIL;
|
|
#endif
|
|
|
|
/* RESULT_SKIP is reserved for fixture setup itself skipping everything */
|
|
if (data->total == data->passed + data->skipped)
|
|
return RESULT_OK;
|
|
return RESULT_FAIL;
|
|
}
|
|
|
|
/** Execute all tests as client tests
|
|
*
|
|
* \param harness The test harness context.
|
|
* \param setup The compositor configuration.
|
|
*
|
|
* Initializes the compositor with the given setup and executes the compositor.
|
|
* The compositor creates a new thread where all tests in the test program are
|
|
* serially executed. Once the thread finishes, the compositor returns from its
|
|
* event loop and cleans up.
|
|
*
|
|
* Returns RESULT_SKIP if the requested compositor features, e.g. GL-renderer,
|
|
* are not built.
|
|
*
|
|
* \sa DECLARE_FIXTURE_SETUP(), DECLARE_FIXTURE_SETUP_WITH_ARG()
|
|
* \ingroup testharness
|
|
*/
|
|
enum test_result_code
|
|
weston_test_harness_execute_as_client(struct weston_test_harness *harness,
|
|
const struct compositor_setup *setup)
|
|
{
|
|
struct wet_testsuite_data *data = &harness->data;
|
|
|
|
data->type = TEST_TYPE_CLIENT;
|
|
return execute_compositor(setup, data);
|
|
}
|
|
|
|
/** Execute all tests as plugin tests
|
|
*
|
|
* \param harness The test harness context.
|
|
* \param setup The compositor configuration.
|
|
*
|
|
* Initializes the compositor with the given setup and executes the compositor.
|
|
* The compositor executes all tests in the test program serially from an idle
|
|
* handler, then returns from its event loop and cleans up.
|
|
*
|
|
* Returns RESULT_SKIP if the requested compositor features, e.g. GL-renderer,
|
|
* are not built.
|
|
*
|
|
* \sa DECLARE_FIXTURE_SETUP(), DECLARE_FIXTURE_SETUP_WITH_ARG()
|
|
* \ingroup testharness
|
|
*/
|
|
enum test_result_code
|
|
weston_test_harness_execute_as_plugin(struct weston_test_harness *harness,
|
|
const struct compositor_setup *setup)
|
|
{
|
|
struct wet_testsuite_data *data = &harness->data;
|
|
|
|
data->type = TEST_TYPE_PLUGIN;
|
|
return execute_compositor(setup, data);
|
|
}
|
|
|
|
/** Execute all tests as standalone tests
|
|
*
|
|
* \param harness The test harness context.
|
|
*
|
|
* Executes all tests in the test program serially without any further setup,
|
|
* particularly without any compositor instance created.
|
|
*
|
|
* \sa DECLARE_FIXTURE_SETUP(), DECLARE_FIXTURE_SETUP_WITH_ARG()
|
|
* \ingroup testharness
|
|
*/
|
|
enum test_result_code
|
|
weston_test_harness_execute_standalone(struct weston_test_harness *harness)
|
|
{
|
|
struct wet_testsuite_data *data = &harness->data;
|
|
|
|
data->type = TEST_TYPE_STANDALONE;
|
|
data->run(data);
|
|
|
|
return RESULT_OK;
|
|
}
|
|
|
|
/** Fixture data array getter method
|
|
*
|
|
* DECLARE_FIXTURE_SETUP_WITH_ARG() overrides this in test programs.
|
|
* The default implementation has no data and makes the tests run once.
|
|
*
|
|
* \ingroup testharness
|
|
*/
|
|
__attribute__((weak)) const struct fixture_setup_array *
|
|
fixture_setup_array_get_(void)
|
|
{
|
|
/* A dummy fixture without a data array. */
|
|
static const struct fixture_setup_array default_fsa = {
|
|
.array = NULL,
|
|
.element_size = 0,
|
|
.n_elements = 1,
|
|
.meta_offset = 0,
|
|
};
|
|
|
|
return &default_fsa;
|
|
}
|
|
|
|
/** Fixture setup function
|
|
*
|
|
* DECLARE_FIXTURE_SETUP() and DECLARE_FIXTURE_SETUP_WITH_ARG() override
|
|
* this in test programs.
|
|
* The default implementation just calls
|
|
* weston_test_harness_execute_standalone().
|
|
*
|
|
* \ingroup testharness
|
|
*/
|
|
__attribute__((weak)) enum test_result_code
|
|
fixture_setup_run_(struct weston_test_harness *harness, const void *arg_)
|
|
{
|
|
return weston_test_harness_execute_standalone(harness);
|
|
}
|
|
|
|
static void
|
|
fixture_report(const struct wet_testsuite_data *d, enum test_result_code ret)
|
|
{
|
|
int fixture_nr = d->fixture_iteration + 1;
|
|
|
|
testlog("--- Fixture %d (%s) %s: passed %d, skipped %d, failed %d, total %d\n",
|
|
fixture_nr, d->fixture_name, result_to_str(ret),
|
|
d->passed, d->skipped, d->failed, d->total);
|
|
}
|
|
|
|
int
|
|
main(int argc, char *argv[])
|
|
{
|
|
struct weston_test_harness *harness;
|
|
enum test_result_code ret;
|
|
enum test_result_code result = RESULT_OK;
|
|
const struct fixture_setup_array *fsa;
|
|
const char *leak_dl_handle;
|
|
int fi;
|
|
int fi_end;
|
|
|
|
/* This is horrific, but it gives us working leak checking. If we
|
|
* actually unload llvmpipe, then we also unload LLVM, and some global
|
|
* setup it's done - which llvmpipe can't tear down because the actual
|
|
* client might be using LLVM instead.
|
|
*
|
|
* Turns out if llvmpipe is always live, then the pointers are always
|
|
* reachable, so LeakSanitizer just tells us about our own code rather
|
|
* than LLVM's, so ...
|
|
*/
|
|
leak_dl_handle = getenv("WESTON_CI_LEAK_DL_HANDLE");
|
|
if (leak_dl_handle)
|
|
(void) dlopen(leak_dl_handle, RTLD_LAZY | RTLD_GLOBAL | RTLD_NODELETE);
|
|
|
|
harness = weston_test_harness_create(argc, argv);
|
|
|
|
fsa = fixture_setup_array_get_();
|
|
|
|
if (harness->fixt_ind == -1) {
|
|
fi = 0;
|
|
fi_end = fsa->n_elements;
|
|
} else {
|
|
fi = harness->fixt_ind;
|
|
fi_end = fi + 1;
|
|
}
|
|
|
|
tap_plan(&harness->data, fi_end - fi);
|
|
testlog("Iterating through %d fixtures.\n", fi_end - fi);
|
|
|
|
for (; fi < fi_end; fi++) {
|
|
const void *arg = fixture_setup_array_get_arg(fsa, fi);
|
|
|
|
harness->data.fixture_iteration = fi;
|
|
harness->data.fixture_name = fixture_setup_array_get_name(fsa,
|
|
fi);
|
|
harness->data.passed = 0;
|
|
harness->data.skipped = 0;
|
|
harness->data.failed = 0;
|
|
|
|
testlog("--- Fixture %d (%s)...\n",
|
|
fi + 1, harness->data.fixture_name);
|
|
|
|
ret = fixture_setup_run_(harness, arg);
|
|
fixture_report(&harness->data, ret);
|
|
|
|
if (ret == RESULT_SKIP) {
|
|
tap_skip_fixture(&harness->data);
|
|
#if WESTON_TEST_SKIP_IS_FAILURE
|
|
ret = RESULT_FAIL;
|
|
#else
|
|
continue;
|
|
#endif
|
|
}
|
|
|
|
if (ret != RESULT_OK && result != RESULT_HARD_ERROR)
|
|
result = ret;
|
|
else if (counts_to_result(&harness->data) != RESULT_OK)
|
|
result = RESULT_FAIL;
|
|
}
|
|
|
|
weston_test_harness_destroy(harness);
|
|
|
|
return result;
|
|
}
|