[PATCH] tester: Make the SIS time limit user configurable
Sebastian Huber
sebastian.huber at embedded-brains.de
Thu Jul 7 06:41:59 UTC 2022
On 06/07/2022 23:55, Chris Johns wrote:
> On 6/7/2022 6:00 pm, Sebastian Huber wrote:
>> Yes, if tests go wrong the tester can kill a test execution after the specified
>> timeout.
>
> Killing should be taken as a sign something in the test equipment is broken.
A simulator which kills itself after an arbitrary amount of time is a
broken test equipment.
>
>> Why do we need this arbitrary SIS -tlim of 400 s?
>
> There was a few values and I select this one based on those. If you have a
> better value please say.
I started this discussion by the best value from my point of view: no
time limit for the simulator.
>
> I do not recommend removing the time limit option from testing.
Why don't you recommend removing the time limit option?
> If you want to
> operate that way create a user config with:
>
> [erc32-sis]
> sis_time_limit = >
>> Which problem does this solve?
>
> Repeatable test results across wide ranging hosts and host operating systems.
I don't see how an arbitrary simulator timeout helps here. Killing the
SIS is very reliably. I have never seen a zombie SIS process after an
rtems-test exit.
>
>> Why can't we let the tests run in SIS without a limit just like we do it for
>> Qemu?
>
> Qemu is painful to make work in a consistent and reliable way.
>
>> Normally, if a test is done, it terminates the SIS execution.
>
> The time outs are for the cases that end abnormally.
Yes, this timeout should be defined by the --timeout command line
option. The timeout depends on the tests you run. This is also selected
through the command line. Test executions stopped by the tester due to a
timeout are reported as a timed out test, however, test executions
stopped by the simulator due to a simulator internal timeout are
reported as "failed".
>
>> The new performance tests can be used to catch performance regressions.
>
> We need to consider the existing benchmarks.
>
>> In the
>> long run I think we need a more modular approach. For example, one component
>> which runs tests and reports the test output. Another component which analyses
>> the test output. The test outputs can be archived.
>
> The rtems-test command is required to report regressions and as simply as
> possible. A user wants to build, test then know what they have matches what we
> released.
Yes, the rtems-test command should do this, however, the machinery it
uses could be more modular. If you want to catch performance regressions
you need history data.
--
embedded brains GmbH
Herr Sebastian HUBER
Dornierstr. 4
82178 Puchheim
Germany
email: sebastian.huber at embedded-brains.de
phone: +49-89-18 94 741 - 16
fax: +49-89-18 94 741 - 08
Registergericht: Amtsgericht München
Registernummer: HRB 157899
Vertretungsberechtigte Geschäftsführer: Peter Rasmussen, Thomas Dörfler
Unsere Datenschutzerklärung finden Sie hier:
https://embedded-brains.de/datenschutzerklaerung/
More information about the devel
mailing list