mirror of
git://git.yoctoproject.org/poky.git
synced 2025-07-19 21:09:03 +02:00

Enable reporting selected test case result given the user provided the selected test case id. If both test result id and test case id were provided, report the selected test case result from the selected test result id. (From OE-Core rev: 7161310ce32d6e0c397d0132808d556bdc80d183) Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com> Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
314 lines
16 KiB
Python
314 lines
16 KiB
Python
# test result tool - report text based test results
|
|
#
|
|
# Copyright (c) 2019, Intel Corporation.
|
|
# Copyright (c) 2019, Linux Foundation
|
|
#
|
|
# SPDX-License-Identifier: GPL-2.0-only
|
|
#
|
|
|
|
import os
|
|
import glob
|
|
import json
|
|
import resulttool.resultutils as resultutils
|
|
from oeqa.utils.git import GitRepo
|
|
import oeqa.utils.gitarchive as gitarchive
|
|
|
|
|
|
class ResultsTextReport(object):
|
|
def __init__(self):
|
|
self.ptests = {}
|
|
self.ltptests = {}
|
|
self.ltpposixtests = {}
|
|
self.result_types = {'passed': ['PASSED', 'passed', 'PASS', 'XFAIL'],
|
|
'failed': ['FAILED', 'failed', 'FAIL', 'ERROR', 'error', 'UNKNOWN', 'XPASS'],
|
|
'skipped': ['SKIPPED', 'skipped', 'UNSUPPORTED', 'UNTESTED', 'UNRESOLVED']}
|
|
|
|
|
|
def handle_ptest_result(self, k, status, result, machine):
|
|
if machine not in self.ptests:
|
|
self.ptests[machine] = {}
|
|
|
|
if k == 'ptestresult.sections':
|
|
# Ensure tests without any test results still show up on the report
|
|
for suite in result['ptestresult.sections']:
|
|
if suite not in self.ptests[machine]:
|
|
self.ptests[machine][suite] = {
|
|
'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-',
|
|
'failed_testcases': [], "testcases": set(),
|
|
}
|
|
if 'duration' in result['ptestresult.sections'][suite]:
|
|
self.ptests[machine][suite]['duration'] = result['ptestresult.sections'][suite]['duration']
|
|
if 'timeout' in result['ptestresult.sections'][suite]:
|
|
self.ptests[machine][suite]['duration'] += " T"
|
|
return True
|
|
|
|
# process test result
|
|
try:
|
|
_, suite, test = k.split(".", 2)
|
|
except ValueError:
|
|
return True
|
|
|
|
# Handle 'glib-2.0'
|
|
if 'ptestresult.sections' in result and suite not in result['ptestresult.sections']:
|
|
try:
|
|
_, suite, suite1, test = k.split(".", 3)
|
|
if suite + "." + suite1 in result['ptestresult.sections']:
|
|
suite = suite + "." + suite1
|
|
except ValueError:
|
|
pass
|
|
|
|
if suite not in self.ptests[machine]:
|
|
self.ptests[machine][suite] = {
|
|
'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-',
|
|
'failed_testcases': [], "testcases": set(),
|
|
}
|
|
|
|
# do not process duplicate results
|
|
if test in self.ptests[machine][suite]["testcases"]:
|
|
print("Warning duplicate ptest result '{}.{}' for {}".format(suite, test, machine))
|
|
return False
|
|
|
|
for tk in self.result_types:
|
|
if status in self.result_types[tk]:
|
|
self.ptests[machine][suite][tk] += 1
|
|
self.ptests[machine][suite]["testcases"].add(test)
|
|
return True
|
|
|
|
def handle_ltptest_result(self, k, status, result, machine):
|
|
if machine not in self.ltptests:
|
|
self.ltptests[machine] = {}
|
|
|
|
if k == 'ltpresult.sections':
|
|
# Ensure tests without any test results still show up on the report
|
|
for suite in result['ltpresult.sections']:
|
|
if suite not in self.ltptests[machine]:
|
|
self.ltptests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []}
|
|
if 'duration' in result['ltpresult.sections'][suite]:
|
|
self.ltptests[machine][suite]['duration'] = result['ltpresult.sections'][suite]['duration']
|
|
if 'timeout' in result['ltpresult.sections'][suite]:
|
|
self.ltptests[machine][suite]['duration'] += " T"
|
|
return
|
|
try:
|
|
_, suite, test = k.split(".", 2)
|
|
except ValueError:
|
|
return
|
|
# Handle 'glib-2.0'
|
|
if 'ltpresult.sections' in result and suite not in result['ltpresult.sections']:
|
|
try:
|
|
_, suite, suite1, test = k.split(".", 3)
|
|
print("split2: %s %s %s" % (suite, suite1, test))
|
|
if suite + "." + suite1 in result['ltpresult.sections']:
|
|
suite = suite + "." + suite1
|
|
except ValueError:
|
|
pass
|
|
if suite not in self.ltptests[machine]:
|
|
self.ltptests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []}
|
|
for tk in self.result_types:
|
|
if status in self.result_types[tk]:
|
|
self.ltptests[machine][suite][tk] += 1
|
|
|
|
def handle_ltpposixtest_result(self, k, status, result, machine):
|
|
if machine not in self.ltpposixtests:
|
|
self.ltpposixtests[machine] = {}
|
|
|
|
if k == 'ltpposixresult.sections':
|
|
# Ensure tests without any test results still show up on the report
|
|
for suite in result['ltpposixresult.sections']:
|
|
if suite not in self.ltpposixtests[machine]:
|
|
self.ltpposixtests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []}
|
|
if 'duration' in result['ltpposixresult.sections'][suite]:
|
|
self.ltpposixtests[machine][suite]['duration'] = result['ltpposixresult.sections'][suite]['duration']
|
|
return
|
|
try:
|
|
_, suite, test = k.split(".", 2)
|
|
except ValueError:
|
|
return
|
|
# Handle 'glib-2.0'
|
|
if 'ltpposixresult.sections' in result and suite not in result['ltpposixresult.sections']:
|
|
try:
|
|
_, suite, suite1, test = k.split(".", 3)
|
|
if suite + "." + suite1 in result['ltpposixresult.sections']:
|
|
suite = suite + "." + suite1
|
|
except ValueError:
|
|
pass
|
|
if suite not in self.ltpposixtests[machine]:
|
|
self.ltpposixtests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []}
|
|
for tk in self.result_types:
|
|
if status in self.result_types[tk]:
|
|
self.ltpposixtests[machine][suite][tk] += 1
|
|
|
|
def get_aggregated_test_result(self, logger, testresult, machine):
|
|
test_count_report = {'passed': 0, 'failed': 0, 'skipped': 0, 'failed_testcases': []}
|
|
result = testresult.get('result', [])
|
|
for k in result:
|
|
test_status = result[k].get('status', [])
|
|
if k.startswith("ptestresult."):
|
|
if not self.handle_ptest_result(k, test_status, result, machine):
|
|
continue
|
|
elif k.startswith("ltpresult."):
|
|
self.handle_ltptest_result(k, test_status, result, machine)
|
|
elif k.startswith("ltpposixresult."):
|
|
self.handle_ltpposixtest_result(k, test_status, result, machine)
|
|
|
|
# process result if it was not skipped by a handler
|
|
for tk in self.result_types:
|
|
if test_status in self.result_types[tk]:
|
|
test_count_report[tk] += 1
|
|
if test_status in self.result_types['failed']:
|
|
test_count_report['failed_testcases'].append(k)
|
|
return test_count_report
|
|
|
|
def print_test_report(self, template_file_name, test_count_reports):
|
|
from jinja2 import Environment, FileSystemLoader
|
|
script_path = os.path.dirname(os.path.realpath(__file__))
|
|
file_loader = FileSystemLoader(script_path + '/template')
|
|
env = Environment(loader=file_loader, trim_blocks=True)
|
|
template = env.get_template(template_file_name)
|
|
havefailed = False
|
|
reportvalues = []
|
|
machines = []
|
|
cols = ['passed', 'failed', 'skipped']
|
|
maxlen = {'passed' : 0, 'failed' : 0, 'skipped' : 0, 'result_id': 0, 'testseries' : 0, 'ptest' : 0 ,'ltptest': 0, 'ltpposixtest': 0}
|
|
for line in test_count_reports:
|
|
total_tested = line['passed'] + line['failed'] + line['skipped']
|
|
vals = {}
|
|
vals['result_id'] = line['result_id']
|
|
vals['testseries'] = line['testseries']
|
|
vals['sort'] = line['testseries'] + "_" + line['result_id']
|
|
vals['failed_testcases'] = line['failed_testcases']
|
|
for k in cols:
|
|
vals[k] = "%d (%s%%)" % (line[k], format(line[k] / total_tested * 100, '.0f'))
|
|
for k in maxlen:
|
|
if k in vals and len(vals[k]) > maxlen[k]:
|
|
maxlen[k] = len(vals[k])
|
|
reportvalues.append(vals)
|
|
if line['failed_testcases']:
|
|
havefailed = True
|
|
if line['machine'] not in machines:
|
|
machines.append(line['machine'])
|
|
reporttotalvalues = {}
|
|
for k in cols:
|
|
reporttotalvalues[k] = '%s' % sum([line[k] for line in test_count_reports])
|
|
reporttotalvalues['count'] = '%s' % len(test_count_reports)
|
|
for (machine, report) in self.ptests.items():
|
|
for ptest in self.ptests[machine]:
|
|
if len(ptest) > maxlen['ptest']:
|
|
maxlen['ptest'] = len(ptest)
|
|
for (machine, report) in self.ltptests.items():
|
|
for ltptest in self.ltptests[machine]:
|
|
if len(ltptest) > maxlen['ltptest']:
|
|
maxlen['ltptest'] = len(ltptest)
|
|
for (machine, report) in self.ltpposixtests.items():
|
|
for ltpposixtest in self.ltpposixtests[machine]:
|
|
if len(ltpposixtest) > maxlen['ltpposixtest']:
|
|
maxlen['ltpposixtest'] = len(ltpposixtest)
|
|
output = template.render(reportvalues=reportvalues,
|
|
reporttotalvalues=reporttotalvalues,
|
|
havefailed=havefailed,
|
|
machines=machines,
|
|
ptests=self.ptests,
|
|
ltptests=self.ltptests,
|
|
ltpposixtests=self.ltpposixtests,
|
|
maxlen=maxlen)
|
|
print(output)
|
|
|
|
def view_test_report(self, logger, source_dir, branch, commit, tag, use_regression_map, raw_test, selected_test_case_only):
|
|
def print_selected_testcase_result(testresults, selected_test_case_only):
|
|
for testsuite in testresults:
|
|
for resultid in testresults[testsuite]:
|
|
result = testresults[testsuite][resultid]['result']
|
|
test_case_result = result.get(selected_test_case_only, {})
|
|
if test_case_result.get('status'):
|
|
print('Found selected test case result for %s from %s' % (selected_test_case_only,
|
|
resultid))
|
|
print(test_case_result['status'])
|
|
else:
|
|
print('Could not find selected test case result for %s from %s' % (selected_test_case_only,
|
|
resultid))
|
|
if test_case_result.get('log'):
|
|
print(test_case_result['log'])
|
|
test_count_reports = []
|
|
configmap = resultutils.store_map
|
|
if use_regression_map:
|
|
configmap = resultutils.regression_map
|
|
if commit:
|
|
if tag:
|
|
logger.warning("Ignoring --tag as --commit was specified")
|
|
tag_name = "{branch}/{commit_number}-g{commit}/{tag_number}"
|
|
repo = GitRepo(source_dir)
|
|
revs = gitarchive.get_test_revs(logger, repo, tag_name, branch=branch)
|
|
rev_index = gitarchive.rev_find(revs, 'commit', commit)
|
|
testresults = resultutils.git_get_result(repo, revs[rev_index][2], configmap=configmap)
|
|
elif tag:
|
|
repo = GitRepo(source_dir)
|
|
testresults = resultutils.git_get_result(repo, [tag], configmap=configmap)
|
|
else:
|
|
testresults = resultutils.load_resultsdata(source_dir, configmap=configmap)
|
|
if raw_test:
|
|
raw_results = {}
|
|
for testsuite in testresults:
|
|
result = testresults[testsuite].get(raw_test, {})
|
|
if result:
|
|
raw_results[testsuite] = {raw_test: result}
|
|
if raw_results:
|
|
if selected_test_case_only:
|
|
print_selected_testcase_result(raw_results, selected_test_case_only)
|
|
else:
|
|
print(json.dumps(raw_results, sort_keys=True, indent=4))
|
|
else:
|
|
print('Could not find raw test result for %s' % raw_test)
|
|
return 0
|
|
if selected_test_case_only:
|
|
print_selected_testcase_result(testresults, selected_test_case_only)
|
|
return 0
|
|
for testsuite in testresults:
|
|
for resultid in testresults[testsuite]:
|
|
skip = False
|
|
result = testresults[testsuite][resultid]
|
|
machine = result['configuration']['MACHINE']
|
|
|
|
# Check to see if there is already results for these kinds of tests for the machine
|
|
for key in result['result'].keys():
|
|
testtype = str(key).split('.')[0]
|
|
if ((machine in self.ltptests and testtype == "ltpiresult" and self.ltptests[machine]) or
|
|
(machine in self.ltpposixtests and testtype == "ltpposixresult" and self.ltpposixtests[machine])):
|
|
print("Already have test results for %s on %s, skipping %s" %(str(key).split('.')[0], machine, resultid))
|
|
skip = True
|
|
break
|
|
if skip:
|
|
break
|
|
|
|
test_count_report = self.get_aggregated_test_result(logger, result, machine)
|
|
test_count_report['machine'] = machine
|
|
test_count_report['testseries'] = result['configuration']['TESTSERIES']
|
|
test_count_report['result_id'] = resultid
|
|
test_count_reports.append(test_count_report)
|
|
self.print_test_report('test_report_full_text.txt', test_count_reports)
|
|
|
|
def report(args, logger):
|
|
report = ResultsTextReport()
|
|
report.view_test_report(logger, args.source_dir, args.branch, args.commit, args.tag, args.use_regression_map,
|
|
args.raw_test_only, args.selected_test_case_only)
|
|
return 0
|
|
|
|
def register_commands(subparsers):
|
|
"""Register subcommands from this plugin"""
|
|
parser_build = subparsers.add_parser('report', help='summarise test results',
|
|
description='print a text-based summary of the test results',
|
|
group='analysis')
|
|
parser_build.set_defaults(func=report)
|
|
parser_build.add_argument('source_dir',
|
|
help='source file/directory/URL that contain the test result files to summarise')
|
|
parser_build.add_argument('--branch', '-B', default='master', help="Branch to find commit in")
|
|
parser_build.add_argument('--commit', help="Revision to report")
|
|
parser_build.add_argument('-t', '--tag', default='',
|
|
help='source_dir is a git repository, report on the tag specified from that repository')
|
|
parser_build.add_argument('-m', '--use_regression_map', action='store_true',
|
|
help='instead of the default "store_map", use the "regression_map" for report')
|
|
parser_build.add_argument('-r', '--raw_test_only', default='',
|
|
help='output raw test result only for the user provided test result id')
|
|
parser_build.add_argument('-s', '--selected_test_case_only', default='',
|
|
help='output selected test case result for the user provided test case id, if both test '
|
|
'result id and test case id are provided then output the selected test case result '
|
|
'from the provided test result id')
|