mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-28 21:45:01 +08:00
acd8e8407b
When a number of tests fail, it can be useful to get higher-level statistics of how many tests are failing (or how many parameters are failing in parameterised tests), and in what cases or suites. This is already done by some non-KUnit tests, so add support for automatically generating these for KUnit tests. This change adds a 'kunit.stats_enabled' switch which has three values: - 0: No stats are printed (current behaviour) - 1: Stats are printed only for tests/suites with more than one subtest (new default) - 2: Always print test statistics For parameterised tests, the summary line looks as follows: " # inode_test_xtimestamp_decoding: pass:16 fail:0 skip:0 total:16" For test suites, there are two lines looking like this: "# ext4_inode_test: pass:1 fail:0 skip:0 total:1" "# Totals: pass:16 fail:0 skip:0 total:16" The first line gives the number of direct subtests, the second "Totals" line is the accumulated sum of all tests and test parameters. This format is based on the one used by kselftest[1]. [1]: https://elixir.bootlin.com/linux/latest/source/tools/testing/selftests/kselftest.h#L109 Signed-off-by: David Gow <davidgow@google.com> Reviewed-by: Brendan Higgins <brendanhiggins@google.com> Signed-off-by: Shuah Khan <skhan@linuxfoundation.org>
429 lines
12 KiB
Python
429 lines
12 KiB
Python
# SPDX-License-Identifier: GPL-2.0
|
|
#
|
|
# Parses test results from a kernel dmesg log.
|
|
#
|
|
# Copyright (C) 2019, Google LLC.
|
|
# Author: Felix Guo <felixguoxiuping@gmail.com>
|
|
# Author: Brendan Higgins <brendanhiggins@google.com>
|
|
|
|
import re
|
|
|
|
from collections import namedtuple
|
|
from datetime import datetime
|
|
from enum import Enum, auto
|
|
from functools import reduce
|
|
from typing import Iterable, Iterator, List, Optional, Tuple
|
|
|
|
TestResult = namedtuple('TestResult', ['status','suites','log'])
|
|
|
|
class TestSuite(object):
|
|
def __init__(self) -> None:
|
|
self.status = TestStatus.SUCCESS
|
|
self.name = ''
|
|
self.cases = [] # type: List[TestCase]
|
|
|
|
def __str__(self) -> str:
|
|
return 'TestSuite(' + str(self.status) + ',' + self.name + ',' + str(self.cases) + ')'
|
|
|
|
def __repr__(self) -> str:
|
|
return str(self)
|
|
|
|
class TestCase(object):
|
|
def __init__(self) -> None:
|
|
self.status = TestStatus.SUCCESS
|
|
self.name = ''
|
|
self.log = [] # type: List[str]
|
|
|
|
def __str__(self) -> str:
|
|
return 'TestCase(' + str(self.status) + ',' + self.name + ',' + str(self.log) + ')'
|
|
|
|
def __repr__(self) -> str:
|
|
return str(self)
|
|
|
|
class TestStatus(Enum):
|
|
SUCCESS = auto()
|
|
FAILURE = auto()
|
|
SKIPPED = auto()
|
|
TEST_CRASHED = auto()
|
|
NO_TESTS = auto()
|
|
FAILURE_TO_PARSE_TESTS = auto()
|
|
|
|
class LineStream:
|
|
"""Provides a peek()/pop() interface over an iterator of (line#, text)."""
|
|
_lines: Iterator[Tuple[int, str]]
|
|
_next: Tuple[int, str]
|
|
_done: bool
|
|
|
|
def __init__(self, lines: Iterator[Tuple[int, str]]):
|
|
self._lines = lines
|
|
self._done = False
|
|
self._next = (0, '')
|
|
self._get_next()
|
|
|
|
def _get_next(self) -> None:
|
|
try:
|
|
self._next = next(self._lines)
|
|
except StopIteration:
|
|
self._done = True
|
|
|
|
def peek(self) -> str:
|
|
return self._next[1]
|
|
|
|
def pop(self) -> str:
|
|
n = self._next
|
|
self._get_next()
|
|
return n[1]
|
|
|
|
def __bool__(self) -> bool:
|
|
return not self._done
|
|
|
|
# Only used by kunit_tool_test.py.
|
|
def __iter__(self) -> Iterator[str]:
|
|
while bool(self):
|
|
yield self.pop()
|
|
|
|
def line_number(self) -> int:
|
|
return self._next[0]
|
|
|
|
kunit_start_re = re.compile(r'TAP version [0-9]+$')
|
|
kunit_end_re = re.compile('(List of all partitions:|'
|
|
'Kernel panic - not syncing: VFS:|reboot: System halted)')
|
|
|
|
def extract_tap_lines(kernel_output: Iterable[str]) -> LineStream:
|
|
def isolate_kunit_output(kernel_output: Iterable[str]) -> Iterator[Tuple[int, str]]:
|
|
line_num = 0
|
|
started = False
|
|
for line in kernel_output:
|
|
line_num += 1
|
|
line = line.rstrip() # line always has a trailing \n
|
|
if kunit_start_re.search(line):
|
|
prefix_len = len(line.split('TAP version')[0])
|
|
started = True
|
|
yield line_num, line[prefix_len:]
|
|
elif kunit_end_re.search(line):
|
|
break
|
|
elif started:
|
|
yield line_num, line[prefix_len:]
|
|
return LineStream(lines=isolate_kunit_output(kernel_output))
|
|
|
|
DIVIDER = '=' * 60
|
|
|
|
RESET = '\033[0;0m'
|
|
|
|
def red(text) -> str:
|
|
return '\033[1;31m' + text + RESET
|
|
|
|
def yellow(text) -> str:
|
|
return '\033[1;33m' + text + RESET
|
|
|
|
def green(text) -> str:
|
|
return '\033[1;32m' + text + RESET
|
|
|
|
def print_with_timestamp(message) -> None:
|
|
print('[%s] %s' % (datetime.now().strftime('%H:%M:%S'), message))
|
|
|
|
def format_suite_divider(message) -> str:
|
|
return '======== ' + message + ' ========'
|
|
|
|
def print_suite_divider(message) -> None:
|
|
print_with_timestamp(DIVIDER)
|
|
print_with_timestamp(format_suite_divider(message))
|
|
|
|
def print_log(log) -> None:
|
|
for m in log:
|
|
print_with_timestamp(m)
|
|
|
|
TAP_ENTRIES = re.compile(r'^(TAP|[\s]*ok|[\s]*not ok|[\s]*[0-9]+\.\.[0-9]+|[\s]*# (Subtest:|.*: kunit test case crashed!)).*$')
|
|
|
|
def consume_non_diagnostic(lines: LineStream) -> None:
|
|
while lines and not TAP_ENTRIES.match(lines.peek()):
|
|
lines.pop()
|
|
|
|
def save_non_diagnostic(lines: LineStream, test_case: TestCase) -> None:
|
|
while lines and not TAP_ENTRIES.match(lines.peek()):
|
|
test_case.log.append(lines.peek())
|
|
lines.pop()
|
|
|
|
OkNotOkResult = namedtuple('OkNotOkResult', ['is_ok','description', 'text'])
|
|
|
|
OK_NOT_OK_SKIP = re.compile(r'^[\s]*(ok|not ok) [0-9]+ - (.*) # SKIP(.*)$')
|
|
|
|
OK_NOT_OK_SUBTEST = re.compile(r'^[\s]+(ok|not ok) [0-9]+ - (.*)$')
|
|
|
|
OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) ([0-9]+) - (.*)$')
|
|
|
|
def parse_ok_not_ok_test_case(lines: LineStream, test_case: TestCase) -> bool:
|
|
save_non_diagnostic(lines, test_case)
|
|
if not lines:
|
|
test_case.status = TestStatus.TEST_CRASHED
|
|
return True
|
|
line = lines.peek()
|
|
match = OK_NOT_OK_SUBTEST.match(line)
|
|
while not match and lines:
|
|
line = lines.pop()
|
|
match = OK_NOT_OK_SUBTEST.match(line)
|
|
if match:
|
|
test_case.log.append(lines.pop())
|
|
test_case.name = match.group(2)
|
|
skip_match = OK_NOT_OK_SKIP.match(line)
|
|
if skip_match:
|
|
test_case.status = TestStatus.SKIPPED
|
|
return True
|
|
if test_case.status == TestStatus.TEST_CRASHED:
|
|
return True
|
|
if match.group(1) == 'ok':
|
|
test_case.status = TestStatus.SUCCESS
|
|
else:
|
|
test_case.status = TestStatus.FAILURE
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
SUBTEST_DIAGNOSTIC = re.compile(r'^[\s]+# (.*)$')
|
|
DIAGNOSTIC_CRASH_MESSAGE = re.compile(r'^[\s]+# .*?: kunit test case crashed!$')
|
|
|
|
def parse_diagnostic(lines: LineStream, test_case: TestCase) -> bool:
|
|
save_non_diagnostic(lines, test_case)
|
|
if not lines:
|
|
return False
|
|
line = lines.peek()
|
|
match = SUBTEST_DIAGNOSTIC.match(line)
|
|
if match:
|
|
test_case.log.append(lines.pop())
|
|
crash_match = DIAGNOSTIC_CRASH_MESSAGE.match(line)
|
|
if crash_match:
|
|
test_case.status = TestStatus.TEST_CRASHED
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def parse_test_case(lines: LineStream) -> Optional[TestCase]:
|
|
test_case = TestCase()
|
|
save_non_diagnostic(lines, test_case)
|
|
while parse_diagnostic(lines, test_case):
|
|
pass
|
|
if parse_ok_not_ok_test_case(lines, test_case):
|
|
return test_case
|
|
else:
|
|
return None
|
|
|
|
SUBTEST_HEADER = re.compile(r'^[\s]+# Subtest: (.*)$')
|
|
|
|
def parse_subtest_header(lines: LineStream) -> Optional[str]:
|
|
consume_non_diagnostic(lines)
|
|
if not lines:
|
|
return None
|
|
match = SUBTEST_HEADER.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return match.group(1)
|
|
else:
|
|
return None
|
|
|
|
SUBTEST_PLAN = re.compile(r'[\s]+[0-9]+\.\.([0-9]+)')
|
|
|
|
def parse_subtest_plan(lines: LineStream) -> Optional[int]:
|
|
consume_non_diagnostic(lines)
|
|
match = SUBTEST_PLAN.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return int(match.group(1))
|
|
else:
|
|
return None
|
|
|
|
def max_status(left: TestStatus, right: TestStatus) -> TestStatus:
|
|
if left == right:
|
|
return left
|
|
elif left == TestStatus.TEST_CRASHED or right == TestStatus.TEST_CRASHED:
|
|
return TestStatus.TEST_CRASHED
|
|
elif left == TestStatus.FAILURE or right == TestStatus.FAILURE:
|
|
return TestStatus.FAILURE
|
|
elif left == TestStatus.SKIPPED:
|
|
return right
|
|
else:
|
|
return left
|
|
|
|
def parse_ok_not_ok_test_suite(lines: LineStream,
|
|
test_suite: TestSuite,
|
|
expected_suite_index: int) -> bool:
|
|
consume_non_diagnostic(lines)
|
|
if not lines:
|
|
test_suite.status = TestStatus.TEST_CRASHED
|
|
return False
|
|
line = lines.peek()
|
|
match = OK_NOT_OK_MODULE.match(line)
|
|
if match:
|
|
lines.pop()
|
|
if match.group(1) == 'ok':
|
|
test_suite.status = TestStatus.SUCCESS
|
|
else:
|
|
test_suite.status = TestStatus.FAILURE
|
|
skip_match = OK_NOT_OK_SKIP.match(line)
|
|
if skip_match:
|
|
test_suite.status = TestStatus.SKIPPED
|
|
suite_index = int(match.group(2))
|
|
if suite_index != expected_suite_index:
|
|
print_with_timestamp(
|
|
red('[ERROR] ') + 'expected_suite_index ' +
|
|
str(expected_suite_index) + ', but got ' +
|
|
str(suite_index))
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def bubble_up_errors(status_list: Iterable[TestStatus]) -> TestStatus:
|
|
return reduce(max_status, status_list, TestStatus.SKIPPED)
|
|
|
|
def bubble_up_test_case_errors(test_suite: TestSuite) -> TestStatus:
|
|
max_test_case_status = bubble_up_errors(x.status for x in test_suite.cases)
|
|
return max_status(max_test_case_status, test_suite.status)
|
|
|
|
def parse_test_suite(lines: LineStream, expected_suite_index: int) -> Optional[TestSuite]:
|
|
if not lines:
|
|
return None
|
|
consume_non_diagnostic(lines)
|
|
test_suite = TestSuite()
|
|
test_suite.status = TestStatus.SUCCESS
|
|
name = parse_subtest_header(lines)
|
|
if not name:
|
|
return None
|
|
test_suite.name = name
|
|
expected_test_case_num = parse_subtest_plan(lines)
|
|
if expected_test_case_num is None:
|
|
return None
|
|
while expected_test_case_num > 0:
|
|
test_case = parse_test_case(lines)
|
|
if not test_case:
|
|
break
|
|
test_suite.cases.append(test_case)
|
|
expected_test_case_num -= 1
|
|
if parse_ok_not_ok_test_suite(lines, test_suite, expected_suite_index):
|
|
test_suite.status = bubble_up_test_case_errors(test_suite)
|
|
return test_suite
|
|
elif not lines:
|
|
print_with_timestamp(red('[ERROR] ') + 'ran out of lines before end token')
|
|
return test_suite
|
|
else:
|
|
print(f'failed to parse end of suite "{name}", at line {lines.line_number()}: {lines.peek()}')
|
|
return None
|
|
|
|
TAP_HEADER = re.compile(r'^TAP version 14$')
|
|
|
|
def parse_tap_header(lines: LineStream) -> bool:
|
|
consume_non_diagnostic(lines)
|
|
if TAP_HEADER.match(lines.peek()):
|
|
lines.pop()
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
TEST_PLAN = re.compile(r'[0-9]+\.\.([0-9]+)')
|
|
|
|
def parse_test_plan(lines: LineStream) -> Optional[int]:
|
|
consume_non_diagnostic(lines)
|
|
match = TEST_PLAN.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return int(match.group(1))
|
|
else:
|
|
return None
|
|
|
|
def bubble_up_suite_errors(test_suites: Iterable[TestSuite]) -> TestStatus:
|
|
return bubble_up_errors(x.status for x in test_suites)
|
|
|
|
def parse_test_result(lines: LineStream) -> TestResult:
|
|
consume_non_diagnostic(lines)
|
|
if not lines or not parse_tap_header(lines):
|
|
return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines)
|
|
expected_test_suite_num = parse_test_plan(lines)
|
|
if expected_test_suite_num == 0:
|
|
return TestResult(TestStatus.NO_TESTS, [], lines)
|
|
elif expected_test_suite_num is None:
|
|
return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines)
|
|
test_suites = []
|
|
for i in range(1, expected_test_suite_num + 1):
|
|
test_suite = parse_test_suite(lines, i)
|
|
if test_suite:
|
|
test_suites.append(test_suite)
|
|
else:
|
|
print_with_timestamp(
|
|
red('[ERROR] ') + ' expected ' +
|
|
str(expected_test_suite_num) +
|
|
' test suites, but got ' + str(i - 2))
|
|
break
|
|
test_suite = parse_test_suite(lines, -1)
|
|
if test_suite:
|
|
print_with_timestamp(red('[ERROR] ') +
|
|
'got unexpected test suite: ' + test_suite.name)
|
|
if test_suites:
|
|
return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines)
|
|
else:
|
|
return TestResult(TestStatus.NO_TESTS, [], lines)
|
|
|
|
class TestCounts:
|
|
passed: int
|
|
failed: int
|
|
crashed: int
|
|
skipped: int
|
|
|
|
def __init__(self):
|
|
self.passed = 0
|
|
self.failed = 0
|
|
self.crashed = 0
|
|
self.skipped = 0
|
|
|
|
def total(self) -> int:
|
|
return self.passed + self.failed + self.crashed + self.skipped
|
|
|
|
def print_and_count_results(test_result: TestResult) -> TestCounts:
|
|
counts = TestCounts()
|
|
for test_suite in test_result.suites:
|
|
if test_suite.status == TestStatus.SUCCESS:
|
|
print_suite_divider(green('[PASSED] ') + test_suite.name)
|
|
elif test_suite.status == TestStatus.SKIPPED:
|
|
print_suite_divider(yellow('[SKIPPED] ') + test_suite.name)
|
|
elif test_suite.status == TestStatus.TEST_CRASHED:
|
|
print_suite_divider(red('[CRASHED] ' + test_suite.name))
|
|
else:
|
|
print_suite_divider(red('[FAILED] ') + test_suite.name)
|
|
for test_case in test_suite.cases:
|
|
if test_case.status == TestStatus.SUCCESS:
|
|
counts.passed += 1
|
|
print_with_timestamp(green('[PASSED] ') + test_case.name)
|
|
elif test_case.status == TestStatus.SKIPPED:
|
|
counts.skipped += 1
|
|
print_with_timestamp(yellow('[SKIPPED] ') + test_case.name)
|
|
elif test_case.status == TestStatus.TEST_CRASHED:
|
|
counts.crashed += 1
|
|
print_with_timestamp(red('[CRASHED] ' + test_case.name))
|
|
print_log(map(yellow, test_case.log))
|
|
print_with_timestamp('')
|
|
else:
|
|
counts.failed += 1
|
|
print_with_timestamp(red('[FAILED] ') + test_case.name)
|
|
print_log(map(yellow, test_case.log))
|
|
print_with_timestamp('')
|
|
return counts
|
|
|
|
def parse_run_tests(kernel_output: Iterable[str]) -> TestResult:
|
|
counts = TestCounts()
|
|
lines = extract_tap_lines(kernel_output)
|
|
test_result = parse_test_result(lines)
|
|
if test_result.status == TestStatus.NO_TESTS:
|
|
print(red('[ERROR] ') + yellow('no tests run!'))
|
|
elif test_result.status == TestStatus.FAILURE_TO_PARSE_TESTS:
|
|
print(red('[ERROR] ') + yellow('could not parse test results!'))
|
|
else:
|
|
counts = print_and_count_results(test_result)
|
|
print_with_timestamp(DIVIDER)
|
|
if test_result.status == TestStatus.SUCCESS:
|
|
fmt = green
|
|
elif test_result.status == TestStatus.SKIPPED:
|
|
fmt = yellow
|
|
else:
|
|
fmt =red
|
|
print_with_timestamp(
|
|
fmt('Testing complete. %d tests run. %d failed. %d crashed. %d skipped.' %
|
|
(counts.total(), counts.failed, counts.crashed, counts.skipped)))
|
|
return test_result
|