[PATCH] tests: Refactor parallel test execution
Alexander Krutwig
alexander.krutwig at embedded-brains.de
Wed Mar 4 12:33:53 UTC 2015
---
cpukit/sapi/Makefile.am | 1 +
cpukit/sapi/include/rtems/test.h | 109 ++++++++++++++
cpukit/sapi/src/testparallel.c | 151 +++++++++++++++++++
testsuites/smptests/smpatomic01/init.c | 264 ++++++++++++++-------------------
4 files changed, 375 insertions(+), 150 deletions(-)
create mode 100644 cpukit/sapi/src/testparallel.c
diff --git a/cpukit/sapi/Makefile.am b/cpukit/sapi/Makefile.am
index a7eaad5..b58b3a2 100644
--- a/cpukit/sapi/Makefile.am
+++ b/cpukit/sapi/Makefile.am
@@ -42,6 +42,7 @@ libsapi_a_SOURCES += src/profilingiterate.c
libsapi_a_SOURCES += src/profilingreportxml.c
libsapi_a_SOURCES += src/testbeginend.c
libsapi_a_SOURCES += src/testextension.c
+libsapi_a_SOURCES += src/testparallel.c
libsapi_a_CPPFLAGS = $(AM_CPPFLAGS)
include $(srcdir)/preinstall.am
diff --git a/cpukit/sapi/include/rtems/test.h b/cpukit/sapi/include/rtems/test.h
index 48a33a0..096b9ad 100644
--- a/cpukit/sapi/include/rtems/test.h
+++ b/cpukit/sapi/include/rtems/test.h
@@ -17,6 +17,8 @@
#include <rtems.h>
#include <rtems/bspIo.h>
+#include <rtems/score/atomic.h>
+#include <rtems/score/smpbarrier.h>
#ifdef __cplusplus
extern "C" {
@@ -116,6 +118,113 @@ static inline int rtems_test_endk(void)
return rtems_test_end_with_plugin(printk_plugin, NULL);
}
+/**
+ * @brief Internal context for parallel job execution.
+ */
+typedef struct {
+ Atomic_Ulong stop;
+ SMP_barrier_Control barrier;
+ size_t worker_count;
+ rtems_id stop_worker_timer_id;
+ rtems_id master_id;
+} rtems_test_parallel_context;
+
+/**
+ * @brief Basic parallel job description.
+ */
+typedef struct {
+ /**
+ * @brief Initialization handler.
+ *
+ * This handler executes only in the context of the master worker before the
+ * job body handler.
+ *
+ * @param[in] ctx The parallel context.
+ * @param[in] arg The user specified argument.
+ */
+ void (*init)(
+ rtems_test_parallel_context *ctx,
+ void *arg
+ );
+
+ /**
+ * @brief Job body handler.
+ *
+ * @param[in] ctx The parallel context.
+ * @param[in] arg The user specified argument.
+ * @param[in] worker_index The worker index. It ranges from 0 to the
+ * processor count minus one.
+ */
+ void (*body)(
+ rtems_test_parallel_context *ctx,
+ void *arg,
+ size_t worker_index
+ );
+
+ /**
+ * @brief Finalization handler.
+ *
+ * This handler executes only in the context of the master worker after the
+ * job body handler.
+ *
+ * @param[in] ctx The parallel context.
+ * @param[in] arg The user specified argument.
+ */
+ void (*fini)(
+ rtems_test_parallel_context *ctx,
+ void *arg
+ );
+
+ void *arg;
+} rtems_test_parallel_job;
+
+/**
+ * @brief Indicates if a job body should stop its work loop.
+ *
+ * @param[in] ctx The parallel context.
+ *
+ * @retval true The job body should stop its work loop and return to the caller.
+ * @retval false Otherwise.
+ */
+static inline bool rtems_test_parallel_stop_job(
+ rtems_test_parallel_context *ctx
+)
+{
+ return _Atomic_Load_ulong(&ctx->stop, ATOMIC_ORDER_RELAXED) != 0;
+}
+
+/**
+ * @brief Indicates if a worker is the master worker.
+ *
+ * The master worker is the thread that called rtems_test_parallel().
+ *
+ * @param[in] worker_index The worker index.
+ *
+ * @retval true This is the master worker.
+ * @retval false Otherwise.
+ */
+static inline bool rtems_test_parallel_is_master_worker(size_t worker_index)
+{
+ return worker_index == 0;
+}
+
+/**
+ * @brief Runs a bunch of jobs in parallel on all processors of the system.
+ *
+ * There are SMP barriers before and after the job body.
+ *
+ * @param[in] ctx The parallel context
+ * @param[in] non_master_worker_priority Indicates the priority of the worker
+ * @param[in] jobs Indicates the pointer to the first job to be executed
+ * @param[in] job_count Indicates the total number of jobs to be executed
+ */
+void rtems_test_parallel(
+ rtems_test_parallel_context *ctx,
+ rtems_task_priority non_master_worker_priority,
+ const rtems_test_parallel_job *jobs,
+ size_t job_count
+);
+
/** @} */
#ifdef __cplusplus
diff --git a/cpukit/sapi/src/testparallel.c b/cpukit/sapi/src/testparallel.c
new file mode 100644
index 0000000..e47ebbd
--- /dev/null
+++ b/cpukit/sapi/src/testparallel.c
@@ -0,0 +1,151 @@
+/*
+ * Copyright (c) 2013-2015 embedded brains GmbH. All rights reserved.
+ *
+ * embedded brains GmbH
+ * Dornierstr. 4
+ * 82178 Puchheim
+ * Germany
+ * <rtems at embedded-brains.de>
+ *
+ * The license and distribution terms for this file may be
+ * found in the file LICENSE in this distribution or at
+ * http://www.rtems.org/license/LICENSE.
+ */
+
+
+#ifdef HAVE_CONFIG_H
+ #include "config.h"
+#endif
+
+#include <rtems/test.h>
+#include <rtems/score/assert.h>
+#include <rtems.h>
+
+static void stop_worker_timer(rtems_id timer_id, void *arg)
+{
+ rtems_test_parallel_context *ctx = arg;
+
+ _Atomic_Store_ulong(&ctx->stop, 1, ATOMIC_ORDER_RELAXED);
+}
+
+static void start_worker_stop_timer(rtems_test_parallel_context *ctx)
+{
+ rtems_status_code sc;
+
+ _Atomic_Store_ulong(&ctx->stop, 0, ATOMIC_ORDER_RELEASE);
+
+ sc = rtems_timer_fire_after(
+ ctx->stop_worker_timer_id,
+ rtems_clock_get_ticks_per_second(),
+ stop_worker_timer,
+ ctx
+ );
+ _Assert(sc == RTEMS_SUCCESSFUL);
+}
+
+static void run_tests(
+ rtems_test_parallel_context *ctx,
+ const rtems_test_parallel_job *jobs,
+ size_t job_count,
+ size_t worker_index
+)
+{
+ SMP_barrier_State bs = SMP_BARRIER_STATE_INITIALIZER;
+ size_t i;
+
+ for (i = 0; i < job_count; ++i) {
+ const rtems_test_parallel_job *job = &jobs[i];
+
+ if (rtems_test_parallel_is_master_worker(worker_index)) {
+ start_worker_stop_timer(ctx);
+ (*job->init)(ctx, job->arg);
+ }
+
+ _SMP_barrier_Wait(&ctx->barrier, &bs, ctx->worker_count);
+
+ (*job->body)(ctx, job->arg, worker_index);
+
+ _SMP_barrier_Wait(&ctx->barrier, &bs, ctx->worker_count);
+
+ if (rtems_test_parallel_is_master_worker(worker_index)) {
+ (*job->fini)(ctx, job->arg);
+ }
+ }
+}
+
+typedef struct {
+ rtems_test_parallel_context *ctx;
+ const rtems_test_parallel_job *jobs;
+ size_t job_count;
+ size_t worker_index;
+} worker_arg;
+
+static void worker_task(rtems_task_argument arg)
+{
+ worker_arg warg = *(worker_arg *) arg;
+ rtems_status_code sc;
+
+ sc = rtems_event_transient_send(warg.ctx->master_id);
+ _Assert(sc == RTEMS_SUCCESSFUL);
+
+ run_tests(warg.ctx, warg.jobs, warg.job_count, warg.worker_index);
+
+ rtems_task_delete(RTEMS_SELF);
+}
+
+void rtems_test_parallel(
+ rtems_test_parallel_context *ctx,
+ rtems_task_priority non_master_worker_priority,
+ const rtems_test_parallel_job *jobs,
+ size_t job_count
+)
+{
+ rtems_status_code sc;
+ size_t worker_index;
+
+ _Atomic_Init_ulong(&ctx->stop, 0);
+ _SMP_barrier_Control_initialize(&ctx->barrier);
+ ctx->worker_count = rtems_get_processor_count();
+ ctx->master_id = rtems_task_self();
+
+ sc = rtems_timer_create(
+ rtems_build_name('S', 'T', 'O', 'P'),
+ &ctx->stop_worker_timer_id
+ );
+ if (sc != RTEMS_SUCCESSFUL) {
+ rtems_fatal_error_occurred(sc);
+ }
+
+ for (worker_index = 1; worker_index < ctx->worker_count; ++worker_index) {
+ worker_arg warg = {
+ .ctx = ctx,
+ .jobs = jobs,
+ .job_count = job_count,
+ .worker_index = worker_index
+ };
+ rtems_id worker_id;
+
+ sc = rtems_task_create(
+ rtems_build_name('W', 'O', 'R', 'K'),
+ non_master_worker_priority,
+ RTEMS_MINIMUM_STACK_SIZE,
+ RTEMS_DEFAULT_MODES,
+ RTEMS_DEFAULT_ATTRIBUTES,
+ &worker_id
+ );
+ if (sc != RTEMS_SUCCESSFUL) {
+ rtems_fatal_error_occurred(sc);
+ }
+
+ sc = rtems_task_start(worker_id, worker_task, (rtems_task_argument) &warg);
+ _Assert(sc == RTEMS_SUCCESSFUL);
+
+ sc = rtems_event_transient_receive(RTEMS_WAIT, RTEMS_NO_TIMEOUT);
+ _Assert(sc == RTEMS_SUCCESSFUL);
+ }
+
+ run_tests(ctx, jobs, job_count, 0);
+
+ sc = rtems_timer_delete(ctx->stop_worker_timer_id);
+ _Assert(sc == RTEMS_SUCCESSFUL);
+}
diff --git a/testsuites/smptests/smpatomic01/init.c b/testsuites/smptests/smpatomic01/init.c
index cc28dee..55d5ffb 100644
--- a/testsuites/smptests/smpatomic01/init.c
+++ b/testsuites/smptests/smpatomic01/init.c
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2013-2014 embedded brains GmbH. All rights reserved.
+ * Copyright (c) 2013-2015 embedded brains GmbH. All rights reserved.
*
* embedded brains GmbH
* Dornierstr. 4
@@ -19,8 +19,8 @@
#endif
#include <rtems/score/atomic.h>
-#include <rtems/score/smpbarrier.h>
#include <rtems.h>
+#include <rtems/test.h>
#include <limits.h>
#include <string.h>
@@ -35,10 +35,7 @@ const char rtems_test_name[] = "SMPATOMIC 1";
#define CPU_COUNT 32
typedef struct {
- Atomic_Ulong stop;
- SMP_barrier_Control barrier;
- size_t worker_count;
- rtems_id stop_worker_timer_id;
+ rtems_test_parallel_context base;
Atomic_Uint atomic_int_value;
Atomic_Ulong atomic_value;
unsigned long per_worker_value[CPU_COUNT];
@@ -46,31 +43,12 @@ typedef struct {
char unused_space_for_cache_line_separation[128];
unsigned long second_value;
Atomic_Flag global_flag;
-} test_context;
+} smpatomic01_context;
-typedef struct {
- void (*init)(test_context *ctx);
- void (*body)(test_context *ctx, size_t worker_index);
- void (*fini)(test_context *ctx);
-} test_case;
-
-static test_context test_instance = {
- .stop = ATOMIC_INITIALIZER_ULONG(0),
- .barrier = SMP_BARRIER_CONTROL_INITIALIZER
-};
-
-static bool stop(test_context *ctx)
-{
- return _Atomic_Load_ulong(&ctx->stop, ATOMIC_ORDER_RELAXED) != 0;
-}
-
-static bool is_master_worker(size_t worker_index)
-{
- return worker_index == 0;
-}
+static smpatomic01_context test_instance;
static void test_fini(
- test_context *ctx,
+ smpatomic01_context *ctx,
const char *test,
bool atomic
)
@@ -81,7 +59,11 @@ static void test_fini(
printf("=== atomic %s test case ===\n", test);
- for (worker_index = 0; worker_index < ctx->worker_count; ++worker_index) {
+ for (
+ worker_index = 0;
+ worker_index < ctx->base.worker_count;
+ ++worker_index
+ ) {
unsigned long worker_value = ctx->per_worker_value[worker_index];
expected_value += worker_value;
@@ -108,16 +90,24 @@ static void test_fini(
rtems_test_assert(expected_value == actual_value);
}
-static void test_atomic_add_init(test_context *ctx)
+
+static void test_atomic_add_init(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
_Atomic_Init_ulong(&ctx->atomic_value, 0);
}
-static void test_atomic_add_body(test_context *ctx, size_t worker_index)
+static void test_atomic_add_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
unsigned long counter = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
++counter;
_Atomic_Fetch_add_ulong(&ctx->atomic_value, 1, ATOMIC_ORDER_RELAXED);
}
@@ -125,22 +115,31 @@ static void test_atomic_add_body(test_context *ctx, size_t worker_index)
ctx->per_worker_value[worker_index] = counter;
}
-static void test_atomic_add_fini(test_context *ctx)
+static void test_atomic_add_fini(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
test_fini(ctx, "add", true);
}
-static void test_atomic_flag_init(test_context *ctx)
+static void test_atomic_flag_init(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
_Atomic_Flag_clear(&ctx->global_flag, ATOMIC_ORDER_RELEASE);
ctx->normal_value = 0;
}
-static void test_atomic_flag_body(test_context *ctx, size_t worker_index)
+static void test_atomic_flag_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
unsigned long counter = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
while (_Atomic_Flag_test_and_set(&ctx->global_flag, ATOMIC_ORDER_ACQUIRE)) {
/* Wait */
}
@@ -154,21 +153,30 @@ static void test_atomic_flag_body(test_context *ctx, size_t worker_index)
ctx->per_worker_value[worker_index] = counter;
}
-static void test_atomic_flag_fini(test_context *ctx)
+static void test_atomic_flag_fini(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
test_fini(ctx, "flag", false);
}
-static void test_atomic_sub_init(test_context *ctx)
+static void test_atomic_sub_init(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
_Atomic_Init_ulong(&ctx->atomic_value, 0);
}
-static void test_atomic_sub_body(test_context *ctx, size_t worker_index)
+static void test_atomic_sub_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
unsigned long counter = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
--counter;
_Atomic_Fetch_sub_ulong(&ctx->atomic_value, 1, ATOMIC_ORDER_RELAXED);
}
@@ -176,22 +184,34 @@ static void test_atomic_sub_body(test_context *ctx, size_t worker_index)
ctx->per_worker_value[worker_index] = counter;
}
-static void test_atomic_sub_fini(test_context *ctx)
+static void test_atomic_sub_fini(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
test_fini(ctx, "sub", true);
}
-static void test_atomic_compare_exchange_init(test_context *ctx)
+static void test_atomic_compare_exchange_init(
+ rtems_test_parallel_context *base,
+ void *arg
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
_Atomic_Init_ulong(&ctx->atomic_value, 0);
ctx->normal_value = 0;
}
-static void test_atomic_compare_exchange_body(test_context *ctx, size_t worker_index)
+static void test_atomic_compare_exchange_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
unsigned long counter = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
bool success;
do {
@@ -215,22 +235,37 @@ static void test_atomic_compare_exchange_body(test_context *ctx, size_t worker_i
ctx->per_worker_value[worker_index] = counter;
}
-static void test_atomic_compare_exchange_fini(test_context *ctx)
+static void test_atomic_compare_exchange_fini(
+ rtems_test_parallel_context *base,
+ void *arg
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
test_fini(ctx, "compare exchange", false);
}
-static void test_atomic_or_and_init(test_context *ctx)
+static void test_atomic_or_and_init(
+ rtems_test_parallel_context *base,
+ void *arg
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
_Atomic_Init_ulong(&ctx->atomic_value, 0);
}
-static void test_atomic_or_and_body(test_context *ctx, size_t worker_index)
+static void test_atomic_or_and_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
unsigned long the_bit = 1UL << worker_index;
unsigned long current_bit = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
unsigned long previous;
if (current_bit != 0) {
@@ -255,31 +290,44 @@ static void test_atomic_or_and_body(test_context *ctx, size_t worker_index)
ctx->per_worker_value[worker_index] = current_bit;
}
-static void test_atomic_or_and_fini(test_context *ctx)
+static void test_atomic_or_and_fini(
+ rtems_test_parallel_context *base,
+ void *arg
+)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
test_fini(ctx, "or/and", true);
}
-static void test_atomic_fence_init(test_context *ctx)
+static void test_atomic_fence_init(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
ctx->normal_value = 0;
ctx->second_value = 0;
_Atomic_Fence(ATOMIC_ORDER_RELEASE);
}
-static void test_atomic_fence_body(test_context *ctx, size_t worker_index)
+static void test_atomic_fence_body(
+ rtems_test_parallel_context *base,
+ void *arg,
+ size_t worker_index
+)
{
- if (is_master_worker(worker_index)) {
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
+ if (rtems_test_parallel_is_master_worker(worker_index)) {
unsigned long counter = 0;
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
++counter;
ctx->normal_value = counter;
_Atomic_Fence(ATOMIC_ORDER_RELEASE);
ctx->second_value = counter;
}
} else {
- while (!stop(ctx)) {
+ while (!rtems_test_parallel_stop_job(&ctx->base)) {
unsigned long n;
unsigned long s;
@@ -292,8 +340,10 @@ static void test_atomic_fence_body(test_context *ctx, size_t worker_index)
}
}
-static void test_atomic_fence_fini(test_context *ctx)
+static void test_atomic_fence_fini(rtems_test_parallel_context *base, void *arg)
{
+ smpatomic01_context *ctx = (smpatomic01_context *) base;
+
printf(
"=== atomic fence test case ===\n"
"normal value = %lu, second value = %lu\n",
@@ -302,7 +352,7 @@ static void test_atomic_fence_fini(test_context *ctx)
);
}
-static const test_case test_cases[] = {
+static const rtems_test_parallel_job test_jobs[] = {
{
test_atomic_add_init,
test_atomic_add_body,
@@ -330,104 +380,18 @@ static const test_case test_cases[] = {
},
};
-#define TEST_COUNT RTEMS_ARRAY_SIZE(test_cases)
-
-static void stop_worker_timer(rtems_id timer_id, void *arg)
-{
- test_context *ctx = arg;
-
- _Atomic_Store_ulong(&ctx->stop, 1, ATOMIC_ORDER_RELAXED);
-}
-
-static void start_worker_stop_timer(test_context *ctx)
-{
- rtems_status_code sc;
-
- _Atomic_Store_ulong(&ctx->stop, 0, ATOMIC_ORDER_RELEASE);
-
- sc = rtems_timer_fire_after(
- ctx->stop_worker_timer_id,
- rtems_clock_get_ticks_per_second(),
- stop_worker_timer,
- ctx
- );
- rtems_test_assert(sc == RTEMS_SUCCESSFUL);
-}
-
-static void run_tests(test_context *ctx, size_t worker_index)
-{
- SMP_barrier_State bs = SMP_BARRIER_STATE_INITIALIZER;
- size_t test;
-
- for (test = 0; test < TEST_COUNT; ++test) {
- const test_case *tc = &test_cases[test];
-
- if (is_master_worker(worker_index)) {
- start_worker_stop_timer(ctx);
- (*tc->init)(ctx);
- }
-
- _SMP_barrier_Wait(&ctx->barrier, &bs, ctx->worker_count);
-
- (*tc->body)(ctx, worker_index);
-
- _SMP_barrier_Wait(&ctx->barrier, &bs, ctx->worker_count);
-
- if (is_master_worker(worker_index)) {
- (*tc->fini)(ctx);
- }
- }
-}
-
-static void worker_task(size_t worker_index)
-{
- test_context *ctx = &test_instance;
-
- run_tests(ctx, worker_index);
-
- (void) rtems_task_suspend(RTEMS_SELF);
- rtems_test_assert(0);
-}
-
-static void test(void)
-{
- test_context *ctx = &test_instance;
- rtems_status_code sc;
- size_t worker_index;
-
- ctx->worker_count = rtems_get_processor_count();
-
- sc = rtems_timer_create(
- rtems_build_name('S', 'T', 'O', 'P'),
- &ctx->stop_worker_timer_id
- );
- rtems_test_assert(sc == RTEMS_SUCCESSFUL);
-
- for (worker_index = 1; worker_index < ctx->worker_count; ++worker_index) {
- rtems_id worker_id;
-
- sc = rtems_task_create(
- rtems_build_name('W', 'O', 'R', 'K'),
- WORKER_PRIORITY,
- RTEMS_MINIMUM_STACK_SIZE,
- RTEMS_DEFAULT_MODES,
- RTEMS_DEFAULT_ATTRIBUTES,
- &worker_id
- );
- rtems_test_assert(sc == RTEMS_SUCCESSFUL);
-
- sc = rtems_task_start(worker_id, worker_task, worker_index);
- rtems_test_assert(sc == RTEMS_SUCCESSFUL);
- }
-
- run_tests(ctx, 0);
-}
-
static void Init(rtems_task_argument arg)
{
+ smpatomic01_context *ctx = &test_instance;
+
TEST_BEGIN();
- test();
+ rtems_test_parallel(
+ &ctx->base,
+ WORKER_PRIORITY,
+ &test_jobs[0],
+ RTEMS_ARRAY_SIZE(test_jobs)
+ );
TEST_END();
rtems_test_exit(0);
--
1.8.4.5
More information about the devel
mailing list