[PATCH 4/4] score: Move per-CPU jobs support

Sebastian Huber sebastian.huber at embedded-brains.de
Wed Jul 28 11:18:04 UTC 2021


Add percpujobs.c to contain the per-CPU jobs implementation.
---
 cpukit/Makefile.am                    |   1 +
 cpukit/score/src/percpujobs.c         | 124 ++++++++++++++++++++++++++
 cpukit/score/src/smpmulticastaction.c |  89 +-----------------
 spec/build/cpukit/objsmp.yml          |   4 +-
 4 files changed, 130 insertions(+), 88 deletions(-)
 create mode 100644 cpukit/score/src/percpujobs.c

diff --git a/cpukit/Makefile.am b/cpukit/Makefile.am
index c83167668d..5673c4e8fb 100644
--- a/cpukit/Makefile.am
+++ b/cpukit/Makefile.am
@@ -1150,6 +1150,7 @@ endif
 
 if HAS_SMP
 
+librtemscpu_a_SOURCES += score/src/percpujobs.c
 librtemscpu_a_SOURCES += score/src/percpustatewait.c
 librtemscpu_a_SOURCES += score/src/profilingsmplock.c
 librtemscpu_a_SOURCES += score/src/schedulerdefaultpinunpin.c
diff --git a/cpukit/score/src/percpujobs.c b/cpukit/score/src/percpujobs.c
new file mode 100644
index 0000000000..4ce96dc738
--- /dev/null
+++ b/cpukit/score/src/percpujobs.c
@@ -0,0 +1,124 @@
+/* SPDX-License-Identifier: BSD-2-Clause */
+
+/**
+ * @file
+ *
+ * @ingroup RTEMSScorePerCPU
+ *
+ * @brief This source file contains the implementation of _Per_CPU_Add_job(),
+ *   _Per_CPU_Perform_jobs(), and _Per_CPU_Wait_for_job().
+ */
+
+/*
+ * Copyright (C) 2019 embedded brains GmbH
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ *    notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ *    notice, this list of conditions and the following disclaimer in the
+ *    documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#ifdef HAVE_CONFIG_H
+#include "config.h"
+#endif
+
+#include <rtems/score/smpimpl.h>
+#include <rtems/score/assert.h>
+
+#define _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, lock_context ) \
+  _ISR_lock_ISR_disable_and_acquire( &( cpu )->Jobs.Lock, lock_context )
+
+#define _Per_CPU_Jobs_release_and_ISR_enable( cpu, lock_context ) \
+  _ISR_lock_Release_and_ISR_enable( &( cpu )->Jobs.Lock, lock_context )
+
+void _Per_CPU_Perform_jobs( Per_CPU_Control *cpu )
+{
+  ISR_lock_Context  lock_context;
+  Per_CPU_Job      *job;
+
+  _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
+  job = cpu->Jobs.head;
+  cpu->Jobs.head = NULL;
+  _Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
+
+  while ( job != NULL ) {
+    const Per_CPU_Job_context *context;
+    Per_CPU_Job               *next;
+
+    context = job->context;
+    next = job->next;
+    ( *context->handler )( context->arg );
+    _Atomic_Store_ulong( &job->done, PER_CPU_JOB_DONE, ATOMIC_ORDER_RELEASE );
+
+    job = next;
+  }
+}
+
+void _Per_CPU_Add_job( Per_CPU_Control *cpu, Per_CPU_Job *job )
+{
+  ISR_lock_Context lock_context;
+
+  _Atomic_Store_ulong( &job->done, 0, ATOMIC_ORDER_RELAXED );
+  _Assert( job->next == NULL );
+
+  _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
+
+  if ( cpu->Jobs.head == NULL ) {
+    cpu->Jobs.head = job;
+  } else {
+    *cpu->Jobs.tail = job;
+  }
+
+  cpu->Jobs.tail = &job->next;
+
+  _Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
+}
+
+void _Per_CPU_Wait_for_job(
+  const Per_CPU_Control *cpu,
+  const Per_CPU_Job     *job
+)
+{
+  while (
+    _Atomic_Load_ulong( &job->done, ATOMIC_ORDER_ACQUIRE )
+      != PER_CPU_JOB_DONE
+  ) {
+    Per_CPU_Control *cpu_self;
+
+    switch ( _Per_CPU_Get_state( cpu ) ) {
+      case PER_CPU_STATE_INITIAL:
+      case PER_CPU_STATE_READY_TO_START_MULTITASKING:
+      case PER_CPU_STATE_UP:
+        /*
+         * Calling this function with the current processor is intentional.  We
+         * have to perform our own jobs here in case inter-processor interrupts
+         * are not working.
+         */
+        cpu_self = _Per_CPU_Get();
+        _SMP_Try_to_process_message(
+          cpu_self,
+          _Atomic_Load_ulong( &cpu_self->message, ATOMIC_ORDER_RELAXED )
+        );
+        break;
+      default:
+        _SMP_Fatal( SMP_FATAL_WRONG_CPU_STATE_TO_PERFORM_JOBS );
+        break;
+    }
+  }
+}
diff --git a/cpukit/score/src/smpmulticastaction.c b/cpukit/score/src/smpmulticastaction.c
index 8dbdef80c7..03d9e064fe 100644
--- a/cpukit/score/src/smpmulticastaction.c
+++ b/cpukit/score/src/smpmulticastaction.c
@@ -3,12 +3,11 @@
 /**
  * @file
  *
- * @ingroup RTEMSScorePerCPU
  * @ingroup RTEMSScoreSMP
  *
- * @brief This source file contains the implementation of _Per_CPU_Add_job(),
- *   _Per_CPU_Perform_jobs(), _Per_CPU_Wait_for_job(), _SMP_Broadcast_action(),
- *   _SMP_Multicast_action(), _SMP_Othercast_action(), and _SMP_Synchronize().
+ * @brief This source file contains the implementation of
+ *   _SMP_Broadcast_action(), _SMP_Multicast_action(), _SMP_Othercast_action(),
+ *   and _SMP_Synchronize().
  */
 
 /*
@@ -43,88 +42,6 @@
 #include <rtems/score/smpimpl.h>
 #include <rtems/score/assert.h>
 
-#define _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, lock_context ) \
-  _ISR_lock_ISR_disable_and_acquire( &( cpu )->Jobs.Lock, lock_context )
-
-#define _Per_CPU_Jobs_release_and_ISR_enable( cpu, lock_context ) \
-  _ISR_lock_Release_and_ISR_enable( &( cpu )->Jobs.Lock, lock_context )
-
-void _Per_CPU_Perform_jobs( Per_CPU_Control *cpu )
-{
-  ISR_lock_Context  lock_context;
-  Per_CPU_Job      *job;
-
-  _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
-  job = cpu->Jobs.head;
-  cpu->Jobs.head = NULL;
-  _Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
-
-  while ( job != NULL ) {
-    const Per_CPU_Job_context *context;
-    Per_CPU_Job               *next;
-
-    context = job->context;
-    next = job->next;
-    ( *context->handler )( context->arg );
-    _Atomic_Store_ulong( &job->done, PER_CPU_JOB_DONE, ATOMIC_ORDER_RELEASE );
-
-    job = next;
-  }
-}
-
-void _Per_CPU_Add_job( Per_CPU_Control *cpu, Per_CPU_Job *job )
-{
-  ISR_lock_Context lock_context;
-
-  _Atomic_Store_ulong( &job->done, 0, ATOMIC_ORDER_RELAXED );
-  _Assert( job->next == NULL );
-
-  _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
-
-  if ( cpu->Jobs.head == NULL ) {
-    cpu->Jobs.head = job;
-  } else {
-    *cpu->Jobs.tail = job;
-  }
-
-  cpu->Jobs.tail = &job->next;
-
-  _Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
-}
-
-void _Per_CPU_Wait_for_job(
-  const Per_CPU_Control *cpu,
-  const Per_CPU_Job     *job
-)
-{
-  while (
-    _Atomic_Load_ulong( &job->done, ATOMIC_ORDER_ACQUIRE )
-      != PER_CPU_JOB_DONE
-  ) {
-    Per_CPU_Control *cpu_self;
-
-    switch ( _Per_CPU_Get_state( cpu ) ) {
-      case PER_CPU_STATE_INITIAL:
-      case PER_CPU_STATE_READY_TO_START_MULTITASKING:
-      case PER_CPU_STATE_UP:
-        /*
-         * Calling this function with the current processor is intentional.  We
-         * have to perform our own jobs here in case inter-processor interrupts
-         * are not working.
-         */
-        cpu_self = _Per_CPU_Get();
-        _SMP_Try_to_process_message(
-          cpu_self,
-          _Atomic_Load_ulong( &cpu_self->message, ATOMIC_ORDER_RELAXED )
-        );
-        break;
-      default:
-        _SMP_Fatal( SMP_FATAL_WRONG_CPU_STATE_TO_PERFORM_JOBS );
-        break;
-    }
-  }
-}
-
 typedef struct {
   Per_CPU_Job_context Context;
   Per_CPU_Job         Jobs[ CPU_MAXIMUM_PROCESSORS ];
diff --git a/spec/build/cpukit/objsmp.yml b/spec/build/cpukit/objsmp.yml
index 1eb1faa3d6..992ed50666 100644
--- a/spec/build/cpukit/objsmp.yml
+++ b/spec/build/cpukit/objsmp.yml
@@ -5,12 +5,12 @@ copyrights:
 - Copyright (C) 2020 embedded brains GmbH (http://www.embedded-brains.de)
 cppflags: []
 cxxflags: []
-enabled-by:
-- RTEMS_SMP
+enabled-by: RTEMS_SMP
 includes: []
 install: []
 links: []
 source:
+- cpukit/score/src/percpujobs.c
 - cpukit/score/src/percpustatewait.c
 - cpukit/score/src/profilingsmplock.c
 - cpukit/score/src/schedulerdefaultaskforhelp.c
-- 
2.26.2



More information about the devel mailing list