I've been promoting the idea that running fstests once is nice, but things get interesting if you try to run fstests multiple times until a failure is found. It turns out at least kdevops has found tests which fail with a failure rate of typically 1/2 to 1/30 average failure rate. That is 1/2 means a failure can happen 50% of the time, whereas 1/30 means it takes 30 runs to find the failure. I have tried my best to annotate failure rates when I know what they might be on the test expunge list, as an example: workflows/fstests/expunges/5.17.0-rc7/xfs/unassigned/xfs_reflink.txt:generic/530 # failure rate about 1/15 https://gist.github.com/mcgrof/4129074db592c170e6bf748aa11d783d The term "failure rate 1/15" is 16 characters long, so I'd like to propose to standardize a way to represent this. How about generic/530 # F:1/15 Then we could extend the definition. F being current estimate, and this can be just how long it took to find the first failure. A more valuable figure would be failure rate avarage, so running the test multiple times, say 10, to see what the failure rate is and then averaging the failure out. So this could be a more accurate representation. For this how about: generic/530 # FA:1/15 This would mean on average there failure rate has been found to be about 1/15, and this was determined based on 10 runs. We should also go extend check for fstests/blktests to run a test until a failure is found and report back the number of successes. Thoughts? Note: yes failure rates lower than 1/100 do exist but they are rare creatures. I love them though as my experience shows so far that they uncover hidden bones in the closet, and they they make take months and a lot of eyeballs to resolve. Luis