[rtems commit] score: Add _Per_CPU_Add_job()
Sebastian Huber
sebh at rtems.org
Tue May 21 04:25:00 UTC 2019
Module: rtems
Branch: master
Commit: 85d6e845d2b9c47779cc96438e3aa6e894b4a71a
Changeset: http://git.rtems.org/rtems/commit/?id=85d6e845d2b9c47779cc96438e3aa6e894b4a71a
Author: Sebastian Huber <sebastian.huber at embedded-brains.de>
Date: Fri Apr 19 11:01:31 2019 +0200
score: Add _Per_CPU_Add_job()
---
cpukit/include/rtems/score/percpu.h | 15 ++++++++++-
cpukit/score/src/smpmulticastaction.c | 37 ++++++++++++++------------
testsuites/smptests/smpmulticast01/init.c | 43 +++++++++++++++++++++++++++++++
3 files changed, 78 insertions(+), 17 deletions(-)
diff --git a/cpukit/include/rtems/score/percpu.h b/cpukit/include/rtems/score/percpu.h
index eff71c4..85e10fb 100644
--- a/cpukit/include/rtems/score/percpu.h
+++ b/cpukit/include/rtems/score/percpu.h
@@ -792,12 +792,25 @@ bool _Per_CPU_State_wait_for_non_initial_state(
);
/**
- * @brief Performs the jobs of the specified processor.
+ * @brief Performs the jobs of the specified processor in FIFO order.
*
* @param[in, out] cpu The jobs of this processor will be performed.
*/
void _Per_CPU_Perform_jobs( Per_CPU_Control *cpu );
+/**
+ * @brief Adds the job to the tail of the processing list of the specified
+ * processor.
+ *
+ * This function does not send the SMP_MESSAGE_PERFORM_JOBS message the
+ * specified processor.
+ *
+ * @param[in, out] cpu The processor to add the job.
+ * @param[in, out] job The job. The Per_CPU_Job::context member must be
+ * initialized by the caller.
+ */
+void _Per_CPU_Add_job( Per_CPU_Control *cpu, Per_CPU_Job *job );
+
#endif /* defined( RTEMS_SMP ) */
/*
diff --git a/cpukit/score/src/smpmulticastaction.c b/cpukit/score/src/smpmulticastaction.c
index ad3b153..1207000 100644
--- a/cpukit/score/src/smpmulticastaction.c
+++ b/cpukit/score/src/smpmulticastaction.c
@@ -57,6 +57,24 @@ void _Per_CPU_Perform_jobs( Per_CPU_Control *cpu )
_Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
}
+}
+
+void _Per_CPU_Add_job( Per_CPU_Control *cpu, Per_CPU_Job *job )
+{
+ ISR_lock_Context lock_context;
+
+ _Atomic_Store_ulong( &job->done, 0, ATOMIC_ORDER_RELAXED );
+ _Assert( job->next == NULL );
+
+ _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
+
+ if ( cpu->Jobs.head == NULL ) {
+ cpu->Jobs.head = job;
+ } else {
+ *cpu->Jobs.tail = job;
+ }
+
+ cpu->Jobs.tail = &job->next;
_Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
}
@@ -97,27 +115,14 @@ static void _SMP_Issue_action_jobs(
for ( cpu_index = 0; cpu_index < cpu_max; ++cpu_index ) {
if ( _Processor_mask_Is_set( targets, cpu_index ) ) {
- ISR_lock_Context lock_context;
- Per_CPU_Job *job;
- Per_CPU_Control *cpu;
+ Per_CPU_Job *job;
+ Per_CPU_Control *cpu;
job = &jobs->Jobs[ cpu_index ];
- _Atomic_Store_ulong( &job->done, 0, ATOMIC_ORDER_RELAXED );
- _Assert( job->next == NULL );
job->context = &jobs->Context;
-
cpu = _Per_CPU_Get_by_index( cpu_index );
- _Per_CPU_Jobs_ISR_disable_and_acquire( cpu, &lock_context );
-
- if ( cpu->Jobs.head == NULL ) {
- cpu->Jobs.head = job;
- } else {
- *cpu->Jobs.tail = job;
- }
-
- cpu->Jobs.tail = &job->next;
- _Per_CPU_Jobs_release_and_ISR_enable( cpu, &lock_context );
+ _Per_CPU_Add_job( cpu, job );
_SMP_Send_message( cpu_index, SMP_MESSAGE_PERFORM_JOBS );
}
}
diff --git a/testsuites/smptests/smpmulticast01/init.c b/testsuites/smptests/smpmulticast01/init.c
index 5f10400..9e5d15f 100644
--- a/testsuites/smptests/smpmulticast01/init.c
+++ b/testsuites/smptests/smpmulticast01/init.c
@@ -351,6 +351,49 @@ static void test_wrong_cpu_state_to_perform_jobs(void)
rtems_fatal(RTEMS_FATAL_SOURCE_APPLICATION, 0);
}
+#define TEST_JOB_ORDER_JOBS 3
+
+static Per_CPU_Job job_order_jobs[TEST_JOB_ORDER_JOBS];
+
+static void job_order_handler_0(void *arg)
+{
+ T_step(1, "invalid job order");
+}
+
+static void job_order_handler_1(void *arg)
+{
+ T_step(2, "invalid job order");
+}
+
+static void job_order_handler_2(void *arg)
+{
+ T_step(3, "invalid job order");
+}
+
+static const Per_CPU_Job_context job_order_contexts[TEST_JOB_ORDER_JOBS] = {
+ { .handler = job_order_handler_0 },
+ { .handler = job_order_handler_1 },
+ { .handler = job_order_handler_2 }
+};
+
+T_TEST_CASE(JobOrder)
+{
+ Per_CPU_Control *cpu_self;
+ size_t i;
+
+ T_plan(4);
+ cpu_self = _Thread_Dispatch_disable();
+
+ for (i = 0; i < TEST_JOB_ORDER_JOBS; ++i) {
+ job_order_jobs[i].context = &job_order_contexts[i];
+ _Per_CPU_Add_job(cpu_self, &job_order_jobs[i]);
+ }
+
+ T_step(0, "wrong job processing time");
+ _SMP_Send_message(_Per_CPU_Get_index(cpu_self), SMP_MESSAGE_PERFORM_JOBS);
+ _Thread_Dispatch_enable(cpu_self);
+}
+
T_TEST_CASE(UnicastDuringMultitaskingIRQDisabled)
{
test_unicast(&test_instance, multicast_action_irq_disabled);
More information about the vc
mailing list