[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAGS_qxoOj+sDOHY8VZv4fw7_XiXqDWcYEC1LAiSUHv38dXh_uw@mail.gmail.com>
Date: Thu, 28 Oct 2021 15:36:20 -0700
From: Daniel Latypov <dlatypov@...gle.com>
To: David Gow <davidgow@...gle.com>
Cc: Brendan Higgins <brendanhiggins@...gle.com>,
Rae Moar <rmr167@...il.com>,
Shuah Khan <skhan@...uxfoundation.org>,
kunit-dev@...glegroups.com, linux-kselftest@...r.kernel.org,
linux-kernel@...r.kernel.org
Subject: Re: [PATCH v3 2/4] kunit: tool: Report an error if any test has no subtests
On Wed, Oct 27, 2021 at 11:42 PM David Gow <davidgow@...gle.com> wrote:
>
> It's possible for a test to have a subtest header, but zero valid
> subtests. We used to error on this if the test plan had no subtests
> listed, but it's possible to have subtests without a test plan (indeed,
> this is how parameterised tests work).
>
> Tests with 0 subtests now have the result NO_TESTS, and will report an
> error (which does not halt test execution, but is printed in a scary red
> colour and is noted in the results summary).
Tested by tweaking ext4 tests (and running with patch 3)
[15:04:33] =============== ext4_inode_test (1 subtest) ================
[15:04:33] ============== inode_test_xtimestamp_decoding ==============
[15:04:33] [ERROR] Test inode_test_xtimestamp_decoding: 0 tests run!
[15:04:33] ====== [NO TESTS RUN] inode_test_xtimestamp_decoding =======
[15:04:33] ================ [SKIPPED] ext4_inode_test =================
[15:04:33] ============================================================
[15:04:33] Testing complete. Passed: 0, Failed: 0, Crashed: 0,
Skipped: 1, Errors: 1
[15:04:33] Elapsed time: 48.581s total, 0.000s configuring, 45.486s
building, 2.992s running
It's maybe a bit confusing to have ERROR, NO TESTS RUN, and SKIPPED
all printed for the same thing.
An alternative would be to drop the error, giving
[15:04:33] =============== ext4_inode_test (1 subtest) ================
[15:04:33] ============== inode_test_xtimestamp_decoding ==============
[15:04:33] ====== [NO TESTS RUN] inode_test_xtimestamp_decoding =======
[15:04:33] ================ [SKIPPED] ext4_inode_test =================
[15:04:33] ============================================================
But looking at it, I think I prefer the more explicit ERROR being there.
>
> Signed-off-by: David Gow <davidgow@...gle.com>
Reviewed-by: Daniel Latypov <dlatypov@...gle.com>
A few optional nits below.
> ---
>
> Changes since v2:
> https://lore.kernel.org/linux-kselftest/20211027013702.2039566-2-davidgow@google.com/
> - Report NO_TESTS as '[NO TESTS RUN]' in yellow, instead of '[FAILED]'
> in red, particularly since it doesn't get counted as a failure.
>
> tools/testing/kunit/kunit_parser.py | 16 +++++++++++-----
> tools/testing/kunit/kunit_tool_test.py | 9 +++++++++
> .../test_is_test_passed-no_tests_no_plan.log | 7 +++++++
> 3 files changed, 27 insertions(+), 5 deletions(-)
> create mode 100644 tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log
>
> diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py
> index 50ded55c168c..68c847e8ca58 100644
> --- a/tools/testing/kunit/kunit_parser.py
> +++ b/tools/testing/kunit/kunit_parser.py
> @@ -360,9 +360,6 @@ def parse_test_plan(lines: LineStream, test: Test) -> bool:
> test.log.append(lines.pop())
> expected_count = int(match.group(1))
> test.expected_count = expected_count
> - if expected_count == 0:
> - test.status = TestStatus.NO_TESTS
> - test.add_error('0 tests run!')
> return True
>
> TEST_RESULT = re.compile(r'^(ok|not ok) ([0-9]+) (- )?([^#]*)( # .*)?$')
> @@ -589,6 +586,8 @@ def format_test_result(test: Test) -> str:
> return (green('[PASSED] ') + test.name)
> elif test.status == TestStatus.SKIPPED:
> return (yellow('[SKIPPED] ') + test.name)
> + elif test.status == TestStatus.NO_TESTS:
> + return (yellow('[NO TESTS RUN] ') + test.name)
> elif test.status == TestStatus.TEST_CRASHED:
> print_log(test.log)
> return (red('[CRASHED] ') + test.name)
> @@ -731,6 +730,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test:
> # test plan
> test.name = "main"
> parse_test_plan(lines, test)
> + parent_test = True
> else:
> # If KTAP/TAP header is not found, test must be subtest
> # header or test result line so parse attempt to parser
> @@ -744,7 +744,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test:
> expected_count = test.expected_count
> subtests = []
> test_num = 1
> - while expected_count is None or test_num <= expected_count:
> + while parent_test and (expected_count is None or test_num <= expected_count):
> # Loop to parse any subtests.
> # Break after parsing expected number of tests or
> # if expected number of tests is unknown break when test
> @@ -779,9 +779,15 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test:
> parse_test_result(lines, test, expected_num)
> else:
> test.add_error('missing subtest result line!')
> +
> + # Check for there being no tests
> + if parent_test and len(subtests) == 0:
> + test.status = TestStatus.NO_TESTS
> + test.add_error('0 tests run!')
> +
> # Add statuses to TestCounts attribute in Test object
> bubble_up_test_results(test)
> - if parent_test:
> + if parent_test and not main:
> # If test has subtests and is not the main test object, print
> # footer.
> print_test_footer(test)
> diff --git a/tools/testing/kunit/kunit_tool_test.py b/tools/testing/kunit/kunit_tool_test.py
> index bc8793145713..c59fe0777387 100755
> --- a/tools/testing/kunit/kunit_tool_test.py
> +++ b/tools/testing/kunit/kunit_tool_test.py
> @@ -208,6 +208,15 @@ class KUnitParserTest(unittest.TestCase):
> self.assertEqual(
> kunit_parser.TestStatus.NO_TESTS,
> result.status)
I'd prefer we split these test cases out.
Perhaps:
def test_no_tests_empty_plan(self):
...
def test_no_tests_no_plan(self):
... # this new test
> + no_plan_log = test_data_path('test_is_test_passed-no_tests_no_plan.log')
> + with open(no_plan_log) as file:
> + result = kunit_parser.parse_run_tests(
> + kunit_parser.extract_tap_lines(file.readlines()))
> + self.assertEqual(0, len(result.test.subtests[0].subtests[0].subtests))
> + self.assertEqual(
> + kunit_parser.TestStatus.NO_TESTS,
> + result.test.subtests[0].subtests[0].status)
optional:
self.assertEqual(1, result.test.counts.errors)
> +
>
> def test_no_kunit_output(self):
> crash_log = test_data_path('test_insufficient_memory.log')
> diff --git a/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log
> new file mode 100644
> index 000000000000..dd873c981108
> --- /dev/null
> +++ b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log
> @@ -0,0 +1,7 @@
> +TAP version 14
> +1..1
> + # Subtest: suite
> + 1..1
> + # Subtest: case
> + ok 1 - case # SKIP
> +ok 1 - suite
> --
> 2.33.0.1079.g6e70778dc9-goog
>
Powered by blists - more mailing lists