On 26/9/19 3:25 am, Gedare Bloom wrote: > I just meant that the view of the running tests is not useful for comparison,
Yes. It lets you see if something major is wrong with a test run. The very original implementation listed the failing tests by name but this added little value and you could not track the state of the run so it was removed. > it is exactly this summary (the end result) that helps. Great, this is the important bit. > If we had regular testing, > parsing the results and producing a status matrix could help for understanding > the tiers. I'm not saying I know how this would be accomplished, and it seems > it > would require coordination among community members who test on different bsps. A score board? This would be really nice to have. I think something that hooks into procmail and monitors the emails posted to the bu...@rtems.org list would work. I would to encourage anyone and everyone to post results for BSPs they have. A score board can then be used to maintain the tiers. Also there is a ticket to have the tester take the console output and generate the results. If this also posted the results email it would help make the test results more widely available. The tester work so far has been unfunded or GSoC projects and I cannot see this changing soon however it is vital to our users, the community and a wider audience that we have quality current published results. Chris _______________________________________________ devel mailing list devel@rtems.org http://lists.rtems.org/mailman/listinfo/devel