[rtems-libbsd commit] EPOCH(9): Add epoch_drain_callbacks()

Sebastian Huber sebh at rtems.org
Tue Jul 9 12:02:04 UTC 2019


Module:    rtems-libbsd
Branch:    master
Commit:    bc2ba9a9cdc7381c2a4f2ae6ee303be636f31368
Changeset: http://git.rtems.org/rtems-libbsd/commit/?id=bc2ba9a9cdc7381c2a4f2ae6ee303be636f31368

Author:    Sebastian Huber <sebastian.huber at embedded-brains.de>
Date:      Tue Jul  9 13:17:18 2019 +0200

EPOCH(9): Add epoch_drain_callbacks()

---

 rtemsbsd/include/sys/epoch.h        |  8 ++++
 rtemsbsd/rtems/rtems-kernel-epoch.c | 82 ++++++++++++++++++++++++++++++++++++-
 testsuite/epoch01/test_main.c       | 50 ++++++++++++++++++++++
 3 files changed, 138 insertions(+), 2 deletions(-)

diff --git a/rtemsbsd/include/sys/epoch.h b/rtemsbsd/include/sys/epoch.h
index 4babb4b..b7dda3d 100644
--- a/rtemsbsd/include/sys/epoch.h
+++ b/rtemsbsd/include/sys/epoch.h
@@ -34,6 +34,8 @@
 #include <sys/cdefs.h>
 #ifdef _KERNEL
 #include <sys/lock.h>
+#include <sys/_mutex.h>
+#include <sys/_sx.h>
 #include <sys/pcpu.h>
 #include <rtems/score/percpudata.h>
 #endif
@@ -66,6 +68,8 @@ struct epoch_record {
 	ck_epoch_record_t er_record;
 	struct epoch_tdlist er_tdlist;
 	uint32_t er_cpuid;
+	struct epoch_context er_drain_ctx;
+	struct epoch *er_parent;
 } __aligned(EPOCH_ALIGN);
 
 typedef struct epoch {
@@ -73,6 +77,9 @@ typedef struct epoch {
 	uintptr_t e_pcpu_record_offset;
 	int	e_flags;
 	SLIST_ENTRY(epoch) e_link;	/* List of all epochs */
+	struct sx e_drain_sx;
+	struct mtx e_drain_mtx;
+	volatile int e_drain_count;
 } *epoch_t;
 
 extern struct epoch _bsd_global_epoch;
@@ -110,6 +117,7 @@ void	epoch_wait_preempt(epoch_t epoch);
 
 void	epoch_call(epoch_t epoch, epoch_context_t ctx,
 	    void (*callback) (epoch_context_t));
+void	epoch_drain_callbacks(epoch_t epoch);
 
 int	_bsd_in_epoch(epoch_t epoch);
 #define	in_epoch(epoch) _bsd_in_epoch(epoch)
diff --git a/rtemsbsd/rtems/rtems-kernel-epoch.c b/rtemsbsd/rtems/rtems-kernel-epoch.c
index c7682a6..f4fff0e 100644
--- a/rtemsbsd/rtems/rtems-kernel-epoch.c
+++ b/rtemsbsd/rtems/rtems-kernel-epoch.c
@@ -33,10 +33,11 @@
 #include <sys/types.h>
 #include <sys/kernel.h>
 #include <sys/epoch.h>
-#ifdef INVARIANTS
+#include <sys/mutex.h>
+#include <sys/sx.h>
 #include <sys/systm.h>
-#endif
 
+#include <machine/atomic.h>
 #include <machine/cpu.h>
 
 #include <rtems.h>
@@ -75,6 +76,8 @@ _bsd_epoch_init(epoch_t epoch, uintptr_t pcpu_record_offset, int flags)
 	ck_epoch_init(&epoch->e_epoch);
 	epoch->e_flags = flags;
 	epoch->e_pcpu_record_offset = pcpu_record_offset;
+	sx_init(&epoch->e_drain_sx, "epoch-drain-sx");
+	mtx_init(&epoch->e_drain_mtx, "epoch-drain-mtx", NULL, MTX_DEF);
 
 	cpu_count = rtems_scheduler_get_processor_maximum();
 
@@ -89,6 +92,7 @@ _bsd_epoch_init(epoch_t epoch, uintptr_t pcpu_record_offset, int flags)
 		TAILQ_INIT(__DEVOLATILE(struct epoch_tdlist *,
 		    &er->er_tdlist));
 		er->er_cpuid = cpu_index;
+		er->er_parent = epoch;
 	}
 
 	SLIST_INSERT_HEAD(&epoch_list, epoch, e_link);
@@ -380,3 +384,77 @@ _bsd_in_epoch(epoch_t epoch)
 	return (in);
 }
 #endif
+
+static void
+epoch_drain_cb(struct epoch_context *ctx)
+{
+	struct epoch *epoch =
+	    __containerof(ctx, struct epoch_record, er_drain_ctx)->er_parent;
+
+	if (atomic_fetchadd_int(&epoch->e_drain_count, -1) == 1) {
+		mtx_lock(&epoch->e_drain_mtx);
+		wakeup(epoch);
+		mtx_unlock(&epoch->e_drain_mtx);
+	}
+}
+
+#ifdef RTEMS_SMP
+static void
+epoch_call_drain_cb(void *arg)
+{
+	epoch_t epoch;
+	Per_CPU_Control *cpu;
+	struct epoch_record *er;
+
+	epoch = arg;
+	cpu = _Per_CPU_Get();
+	er = EPOCH_GET_RECORD(cpu, epoch);
+	epoch_call(epoch, &er->er_drain_ctx, epoch_drain_cb);
+}
+#endif
+
+void
+epoch_drain_callbacks(epoch_t epoch)
+{
+#ifdef RTEMS_SMP
+	uint32_t cpu_index;
+	uint32_t cpu_max;
+	rtems_id id;
+	rtems_status_code sc;
+#else
+	struct epoch_record *er;
+#endif
+
+	sx_xlock(&epoch->e_drain_sx);
+	mtx_lock(&epoch->e_drain_mtx);
+
+#ifdef RTEMS_SMP
+	cpu_max = rtems_scheduler_get_processor_maximum();
+
+	for (cpu_index = 0; cpu_index <= cpu_max; ++cpu_index) {
+		sc = rtems_scheduler_ident_by_processor(cpu_index, &id);
+		if (sc == RTEMS_SUCCESSFUL) {
+			epoch->e_drain_count++;
+		}
+	}
+
+	for (cpu_index = 0; cpu_index <= cpu_max; ++cpu_index) {
+		sc = rtems_scheduler_ident_by_processor(cpu_index, &id);
+		if (sc == RTEMS_SUCCESSFUL) {
+			_SMP_Unicast_action(cpu_index, epoch_call_drain_cb,
+			    epoch);
+		}
+	}
+#else
+	epoch->e_drain_count = 1;
+	er = EPOCH_GET_RECORD(0, epoch);
+	epoch_call(epoch, &er->er_drain_ctx, epoch_drain_cb);
+#endif
+
+	while (epoch->e_drain_count != 0) {
+		msleep(epoch, &epoch->e_drain_mtx, PZERO, "EDRAIN", 0);
+	}
+
+	mtx_unlock(&epoch->e_drain_mtx);
+	sx_xunlock(&epoch->e_drain_sx);
+}
diff --git a/testsuite/epoch01/test_main.c b/testsuite/epoch01/test_main.c
index 7088a12..b72431d 100644
--- a/testsuite/epoch01/test_main.c
+++ b/testsuite/epoch01/test_main.c
@@ -357,6 +357,51 @@ test_enter_list_op_exit_preempt_fini(rtems_test_parallel_context *base,
 }
 
 static void
+test_enter_list_op_exit_drain_body(rtems_test_parallel_context *base,
+    void *arg, size_t active_workers, size_t worker_index)
+{
+	test_context *ctx;
+	epoch_t e;
+	uint32_t counter;
+	uint32_t removals;
+	uint32_t item_counter[CPU_COUNT];
+
+	ctx = (test_context *)base;
+	e = global_epoch;
+	counter = 0;
+	removals = 0;
+	memset(item_counter, 0, sizeof(item_counter));
+
+	while (!rtems_test_parallel_stop_job(&ctx->base)) {
+		test_item *rm;
+
+		epoch_enter(e);
+		++counter;
+		rm = test_remove_item(ctx, item_counter, &removals,
+		    worker_index);
+		epoch_exit(e);
+
+		if (rm != NULL) {
+			epoch_call(e, &rm->ec, test_list_callback);
+			epoch_drain_callbacks(e);
+		}
+	}
+
+	ctx->stats.counter[worker_index] = counter;
+	ctx->stats.removals[worker_index] = removals;
+	memcpy(ctx->stats.item_counter[worker_index], item_counter,
+	    sizeof(ctx->stats.item_counter[worker_index]));
+}
+
+static void
+test_enter_list_op_exit_drain_fini(rtems_test_parallel_context *base,
+    void *arg, size_t active_workers)
+{
+
+	test_fini(base, "EnterListOpExitDrain", active_workers);
+}
+
+static void
 test_thread_local_mutex_body(rtems_test_parallel_context *base, void *arg,
     size_t active_workers, size_t worker_index)
 {
@@ -443,6 +488,11 @@ static const rtems_test_parallel_job test_jobs[] = {
 		.fini = test_enter_list_op_exit_preempt_fini,
 		.cascade = true
 	}, {
+		.init = test_list_init,
+		.body = test_enter_list_op_exit_drain_body,
+		.fini = test_enter_list_op_exit_drain_fini,
+		.cascade = true
+	}, {
 		.init = test_init,
 		.body = test_thread_local_mutex_body,
 		.fini = test_thread_local_mutex_fini,




More information about the vc mailing list