On Mon, Oct 11, 2021 at 10:14 AM Daniel Latypov <dlatypov@xxxxxxxxxx> wrote: > > On Fri, Oct 8, 2021 at 7:35 PM 'David Gow' via KUnit Development > <kunit-dev@xxxxxxxxxxxxxxxx> wrote: > > > > On Fri, Oct 8, 2021 at 5:03 AM Daniel Latypov <dlatypov@xxxxxxxxxx> wrote: > > > > > > From: Rae Moar <rmoar@xxxxxxxxxx> > > > > > > Update to kunit_parser to improve compatibility with KTAP > > > specification including arbitrarily nested tests. Patch accomplishes > > > three major changes: > > > > > > - Use a general Test object to represent all tests rather than TestCase > > > and TestSuite objects. This allows for easier implementation of arbitrary > > > levels of nested tests and promotes the idea that both test suites and test > > > cases are tests. > > > > > > - Print errors incrementally rather than all at once after the > > > parsing finishes to maximize information given to the user in the > > > case of the parser given invalid input and to increase the helpfulness > > > of the timestamps given during printing. Note that kunit.py parse does > > > not print incrementally yet. However, this fix brings us closer to > > > this feature. > > > > > > - Increase compatibility for different formats of input. Arbitrary levels > > > of nested tests supported. Also, test cases and test suites are now > > > supported to be present on the same level of testing. > > > > > > This patch now implements the draft KTAP specification here: > > > https://lore.kernel.org/linux-kselftest/CA+GJov6tdjvY9x12JsJT14qn6c7NViJxqaJk+r-K1YJzPggFDQ@xxxxxxxxxxxxxx/ > > > We'll update the parser as the spec evolves. > > > > > > This patch adjusts the kunit_tool_test.py file to check for > > > the correct outputs from the new parser and adds a new test to check > > > the parsing for a KTAP result log with correct format for multiple nested > > > subtests (test_is_test_passed-all_passed_nested.log). > > > > > > This patch also alters the kunit_json.py file to allow for arbitrarily > > > nested tests. > > > > > > Signed-off-by: Rae Moar <rmoar@xxxxxxxxxx> > > > Reviewed-by: Brendan Higgins <brendanhiggins@xxxxxxxxxx> > > > Signed-off-by: Daniel Latypov <dlatypov@xxxxxxxxxx> > > > Reviewed-by: David Gow <davidgow@xxxxxxxxxx> > > > --- > > > > Found a regression here with the KASAN tests and treating "BUG:" as a crash. > > > > Take this output, for example: > > [19:32:07] # Subtest: kasan > > [19:32:07] 1..48 > > [19:32:07] # kasan: pass:42 fail:0 skip:6 total:48 > > [19:32:07] # Totals: pass:42 fail:0 skip:6 total:48 > > [19:32:07] ok 1 - kasan > > [19:32:07] ===================== [CRASHED] kasan ====================== > > [19:32:07] ============================================================ > > [19:32:07] Testing complete. Passed: 4, Failed: 0, Crashed: 38, > > Skipped: 6, Errors: 0 > > > > The in-kernel totals are correctly: 42 passed, 7 skipped, 0 failed. In > > kunit_tool, only 4 tests are recorded as passed, and 38 are marked as > > crashed. > > > > > > > Change log from v6: > > > https://lore.kernel.org/linux-kselftest/20211006170049.106852-1-dlatypov@xxxxxxxxxx/ > > > - Rebase onto shuah/linux-kselftest/kunit > > > - fix one new unit test failure (s/suites/test.subtests) > > > > > > Change log from v5: > > > https://lore.kernel.org/linux-kselftest/20211006001447.20919-1-dlatypov@xxxxxxxxxx/ > > > - Tweak commit message to reflect the KTAP spec is a draft > > > - Add missing Signed-off-by > > > - Tweak docstrings > > > > > > Change log from v3,4: > > > https://lore.kernel.org/linux-kselftest/20210901190623.315736-1-rmoar@xxxxxxxxxx/ > > > - Move test_kselftest_nested from LinuxSourceTreeTest => KUnitParserTest. > > > - Resolve conflict with hermetic testing patches. > > > - max_status is no longer defined, so we need to use the TestCounts > > > type now. And to keep --raw_output working, we need to set this to > > > SUCCESS to avoid the default assumption that the kernel crashed. > > > > > > Ignore v4, was accidentally based on v2. > > > > > > Change log from v2: > > > https://lore.kernel.org/linux-kselftest/20210826195505.3066755-1-rmoar@xxxxxxxxxx/ > > > - Fixes bug of type disagreement in kunit_json.py for build_dir > > > - Removes raw_output() > > > - Changes docstrings in kunit_parser.py (class docstring, LineStream > > > docstrings, add_error(), total(), get_status(), all parsing methods) > > > - Fixes bug of not printing diagnostic log in the case of end of lines > > > - Sets default status of all tests to TEST_CRASHED > > > - Adds and prints empty tests with crashed status in case of missing > > > tests > > > - Prints 'subtest' in instance of 1 subtest instead of 'subtests' > > > - Includes checking for 'BUG:' message in search of crash messages in > > > log (note that parse_crash_in_log method could be removed but would > > > require deleting tests in kunit_tool_test.py that include the crash > > > message that is no longer used. If removed, parser would still print > > > log in cases of test crashed or failure, which would now include > > > missing subtests) > > > > So this causes problems with the KASAN tests, because they include > > KASAN errors in the log (which are expected), and these messages do > > contain "BUG:". > > Even though the KASAN integration code then marks the test as success, > > and it shows up as "ok" in the KTAP output, kunit_tool now marks it as > > crashed. > > > > There are two obvious solutions to this: > > - Update KASAN to not include "BUG" in the message for KASAN failures > > which are expected. > > - Alter this patch to not mark tests as crashed just because they have > > "BUG" in their logs. > > > > (There are also more complicated solutions, like having a "failure > > expected" message added to the log, and only ignoring "BUG:" if that > > exists in kunit_tool, but that seems needlessly complicated.) > > > > I feel that, in the short term, we should revert this change, and not > > treat "BUG:" specially. We can add it back in as a separate patch if > > we want to fix this issue differently. > > Will do. > > I also found another bug relating to test status counting. > > If there's NO_TESTS, add_status() will increment the counts.passed field. > This means if you get > > $ ./tools/testing/kunit/kunit.py exec 'nomatch' > [10:13:34] Starting KUnit Kernel (1/1)... > [10:13:34] ============================================================ > [10:13:37] [ERROR] Test main: 0 tests run! > [10:13:37] ============================================================ > [10:13:37] Testing complete. Passed: 1, Failed: 0, Crashed: 0, > Skipped: 0, Errors: 1 > <exit with status code 0> > I think something like this resolves those two issues. I'm just not sure what the intent behind the SUCCESS or NO_TESTS checks was. diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py index f01fd565f978..bd3e859bc4e5 100644 --- a/tools/testing/kunit/kunit_parser.py +++ b/tools/testing/kunit/kunit_parser.py @@ -156,17 +156,13 @@ class TestCounts: Parameters: status - status to be added to the TestCounts object """ - if status == TestStatus.SUCCESS or \ - status == TestStatus.NO_TESTS: - # if status is NO_TESTS the most appropriate - # attribute to increment is passed because - # the test did not fail, crash or get skipped. + if status == TestStatus.SUCCESS: self.passed += 1 elif status == TestStatus.FAILURE: self.failed += 1 elif status == TestStatus.SKIPPED: self.skipped += 1 - else: + elif status != TestStatus.NO_TESTS: self.crashed += 1 class LineStream: @@ -475,8 +471,7 @@ def parse_diagnostic(lines: LineStream) -> List[str]: log.append(lines.pop()) return log -DIAGNOSTIC_CRASH_MESSAGE = re.compile( - r'^(BUG:|# .*?: kunit test case crashed!$)') +DIAGNOSTIC_CRASH_MESSAGE = re.compile(r'^# .*?: kunit test case crashed!$') def parse_crash_in_log(test: Test) -> bool: """ @@ -642,8 +637,7 @@ def print_summary_line(test: Test) -> None: test - Test object representing current test being printed """ - if test.status == TestStatus.SUCCESS or \ - test.status == TestStatus.NO_TESTS: + if test.status == TestStatus.SUCCESS: color = green elif test.status == TestStatus.SKIPPED: color = yellow