[RTEMS Project] #2963: Add a testsuite top level confguration file that is common to all tests.
RTEMS trac
trac at rtems.org
Fri Mar 31 06:32:12 UTC 2017
#2963: Add a testsuite top level confguration file that is common to all tests.
-------------------------+------------------------------
Reporter: Chris Johns | Owner: joel.sherrill@…
Type: defect | Status: assigned
Priority: normal | Milestone: 4.12.0
Component: testing | Version: 4.12
Severity: normal | Resolution:
Keywords: testing |
-------------------------+------------------------------
Comment (by Sebastian Huber):
Replying to [comment:2 Chris Johns]:
> I have never looked at the gcc testsuite and from what I have read about
it was not flattering.
My comment was not about the GCC testsuite in general.
>
> I do not think adding 500+ files to state `fileio` is a `user-input`
test and will never complete is not good. Maybe global is not a great
word, maybe common is better. We need accurate data to determine the
results of tests.
Why 500+ files, its just one:
{{{
diff --git a/testsuites/samples/fileio/init.c
b/testsuites/samples/fileio/init.c
index 07ec2c6..68942e8 100644
--- a/testsuites/samples/fileio/init.c
+++ b/testsuites/samples/fileio/init.c
@@ -34,6 +34,7 @@
#include <rtems/nvdisk-sram.h>
#include <rtems/shell.h>
+/* FANCY TEST COMMENT: user-input */
const char rtems_test_name[] = "FILE I/O";
#if FILEIO_BUILD
}}}
>
> It is similar to the work you have been doing to have a common linkercmd
file where ever possible. It is the same thing or are you saying we should
create a separate linker command file for every bsp as well? ;)
>
> Look at the results with a work in progress `rtems-test` for
`erc32-run`:
>
> {{{
> Passed: 546
> Failed: 1
> User Input: 4
> Expected Fail: 0
> Indeterminate: 0
> Timeout: 6
> Invalid: 1
> ------------------
> Total: 558
>
> Failures:
> spcontext01.exe
> User Input:
> fileio.exe
> top.exe
> termios.exe
> monitor.exe
> Timeouts:
> jffs2_fssymlink.exe
> mrfs_fserror.exe
> dhrystone.exe
> fsdosfsformat01.exe
> imfs_fsrdwr.exe
> whetstone.exe
> Invalid:
> minimum.exe
> Average test time: 0:00:00.481800
> Testing time : 0:04:28.844749
> }}}
>
> Note, the benchmark tests have broken parallel testing because of the
time they now take.
On my host these benchmark tests did run less than 3 minutes.
--
Ticket URL: <http://devel.rtems.org/ticket/2963#comment:3>
RTEMS Project <http://www.rtems.org/>
RTEMS Project
More information about the bugs
mailing list