Alright, so basically it seems like people are not ok with the not requiring stressnewtest approach. So I guess that means that we need to identify -1s willing to help resolve this problem…
Who would like to help? Thanks, Mark > On Feb 28, 2020, at 11:12 AM, Ernest Burghardt <eburgha...@pivotal.io> wrote: > > -1 to limiting any tests... if there are issues with the tests let's fix > that. we have too many commits coming in with little or no testing over > new/changed code, so I can't see how removing any existing test coverage as > a good idea > > Best, > EB > > On Fri, Feb 28, 2020 at 10:58 AM Mark Hanson <mhan...@pivotal.io> wrote: > >> Just to make sure we are clear, I am not suggesting that we disable >> stressnewtest, but that we make it not required. It would still run and >> provide feedback, but it would not give us an unwarranted green in my >> approach. >> >>> On Feb 28, 2020, at 10:49 AM, Ju@N <jujora...@gmail.com> wrote: >>> >>> +1 to what Owen said, I don't think disabling *StressNewTest* is a >>> good idea. >>> >>> On Fri, 28 Feb 2020 at 18:35, Owen Nichols <onich...@pivotal.io> wrote: >>> >>>> -1 to making StressNew not required >>>> >>>> +1 to eliminating the current loophole — StressNew should never give a >>>> free pass. >>>> >>>> Any time your PR is having trouble passing StressNew, please bring it up >>>> on the dev list. We can review on a case-by-case basis and decide >> whether >>>> to try increasing the timeout, changing the repeat count, refactoring >> the >>>> PR, or as an absolute last resort requesting authorization for an >> override >>>> (for example, a change to spotless rules might touch a huge number of >> files >>>> but carry no risk). >>>> >>>> One bug we should fix is that StressNew sometimes counts more files >>>> touched than really were, especially if you had many commits or merges >> or >>>> rebases on your PR branch. Possible workarounds there include squashing >>>> and/or creating a new PR and/or splitting into multiple PRs. I’ve spent >>>> some time trying to reproduce why files are mis-counted, with no >> success, >>>> but perhaps someone cleverer with git could provide a fix. >>>> >>>> Another issue is that StressNew is only in the PR pipeline, not the main >>>> develop pipeline. This feels like an asymmetry where PRs must pass a >>>> “higher” standard. We should consider adding some form of StressNew to >> the >>>> main pipeline as well (maybe compare to the previous SHA that passed?). >>>> >>>> The original motivation for the 25-file limit was an attempt to limit >> how >>>> long StressNew might run for. Since concourse already applies a >> timeout, >>>> that check is unnecessary. However, a compromise solution might be to >> use >>>> the number of files changed to dial back the number of repeats, e.g. >> stay >>>> with 50 repeats if fewer than 25 files changed, otherwise compute 1250 / >>>> <#-files-changed> and do only that many repeats (e.g. if 50 files >> changed, >>>> run all tests 25x instead of 50x). >>>> >>>> While StressNew is intended to catch new flaky tests, it can also catch >>>> poorly-designed tests that fail just by running twice in the same VM. >> This >>>> may be a sign that the test does not clean up properly and could be >>>> polluting other tests in unexpected ways? It might be useful to run a >>>> “StressNew” with repeats=2 over a much broader scope, maybe even all >> tests, >>>> at least once in a while? >>>> >>>> >>>>> On Feb 28, 2020, at 9:51 AM, Mark Hanson <mhan...@pivotal.io> wrote: >>>>> >>>>> Hi All, >>>>> >>>>> Proposal: Force StressNewTest to fail a change with 25 or more files >>>> rather than pass it without running it. >>>>> >>>>> Currently, the StressNewTest job in the pipeline will just pass a job >>>> that has more than 25 files changed. It will be marked as green with no >>>> work done. There are reasons, relating to run time being too long to be >>>> tracked by concourse, so we just let it through as a pass. I think this >> is >>>> a bad signal. I think that this should automatically be a failure in the >>>> short term. As a result, I also think it should not be required. It is a >>>> bad signal, and I think that by making it a fail, this will at least not >>>> give us a false sense of security. I understand that this opens the >> flood >>>> gates so to speak, but I don’t think as a community it is a big problem >>>> because we can still say that you should not merge if the StressNewTest >>>> fails because of your test. >>>>> >>>>> I personally find the false sense of security more troubling than >>>> anything. Hence the reason, I would like this to be >>>>> >>>>> Let me know what you think.. >>>>> >>>>> Thanks, >>>>> Mark >>>> >>>> >>> >>> -- >>> Ju@N >> >>