mirror of
git://git.yoctoproject.org/poky.git
synced 2025-07-19 21:09:03 +02:00

This adds SPDX license headers in place of the wide assortment of things currently in our script headers. We default to GPL-2.0-only except for the oeqa code where it was clearly submitted and marked as MIT on the most part or some scripts which had the "or later" GPL versioning. The patch also drops other obsolete bits of file headers where they were encoountered such as editor modelines, obsolete maintainer information or the phrase "All rights reserved" which is now obsolete and not required in copyright headers (in this case its actually confusing for licensing as all rights were not reserved). More work is needed for OE-Core but this takes care of the bulk of the scripts and meta/lib directories. The top level LICENSE files are tweaked to match the new structure and the SPDX naming. (From OE-Core rev: f8c9c511b5f1b7dbd45b77f345cb6c048ae6763e) Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
206 lines
10 KiB
Python
Executable File
206 lines
10 KiB
Python
Executable File
# test case management tool - manual execution from testopia test cases
|
|
#
|
|
# Copyright (c) 2018, Intel Corporation.
|
|
#
|
|
# SPDX-License-Identifier: GPL-2.0-only
|
|
#
|
|
|
|
import argparse
|
|
import json
|
|
import os
|
|
import sys
|
|
import datetime
|
|
import re
|
|
import copy
|
|
from oeqa.core.runner import OETestResultJSONHelper
|
|
|
|
|
|
def load_json_file(f):
|
|
with open(f, "r") as filedata:
|
|
return json.load(filedata)
|
|
|
|
def write_json_file(f, json_data):
|
|
os.makedirs(os.path.dirname(f), exist_ok=True)
|
|
with open(f, 'w') as filedata:
|
|
filedata.write(json.dumps(json_data, sort_keys=True, indent=4))
|
|
|
|
class ManualTestRunner(object):
|
|
|
|
def _get_test_module(self, case_file):
|
|
return os.path.basename(case_file).split('.')[0]
|
|
|
|
def _get_input(self, config):
|
|
while True:
|
|
output = input('{} = '.format(config))
|
|
if re.match('^[a-z0-9-.]+$', output):
|
|
break
|
|
print('Only lowercase alphanumeric, hyphen and dot are allowed. Please try again')
|
|
return output
|
|
|
|
def _get_available_config_options(self, config_options, test_module, target_config):
|
|
avail_config_options = None
|
|
if test_module in config_options:
|
|
avail_config_options = config_options[test_module].get(target_config)
|
|
return avail_config_options
|
|
|
|
def _choose_config_option(self, options):
|
|
while True:
|
|
output = input('{} = '.format('Option index number'))
|
|
if output in options:
|
|
break
|
|
print('Only integer index inputs from above available configuration options are allowed. Please try again.')
|
|
return options[output]
|
|
|
|
def _get_config(self, config_options, test_module):
|
|
from oeqa.utils.metadata import get_layers
|
|
from oeqa.utils.commands import get_bb_var
|
|
from resulttool.resultutils import store_map
|
|
|
|
layers = get_layers(get_bb_var('BBLAYERS'))
|
|
configurations = {}
|
|
configurations['LAYERS'] = layers
|
|
configurations['STARTTIME'] = datetime.datetime.now().strftime('%Y%m%d%H%M%S')
|
|
configurations['TEST_TYPE'] = 'manual'
|
|
configurations['TEST_MODULE'] = test_module
|
|
|
|
extra_config = set(store_map['manual']) - set(configurations)
|
|
for config in sorted(extra_config):
|
|
avail_config_options = self._get_available_config_options(config_options, test_module, config)
|
|
if avail_config_options:
|
|
print('---------------------------------------------')
|
|
print('These are available configuration #%s options:' % config)
|
|
print('---------------------------------------------')
|
|
for option, _ in sorted(avail_config_options.items(), key=lambda x: int(x[0])):
|
|
print('%s: %s' % (option, avail_config_options[option]))
|
|
print('Please select configuration option, enter the integer index number.')
|
|
value_conf = self._choose_config_option(avail_config_options)
|
|
print('---------------------------------------------\n')
|
|
else:
|
|
print('---------------------------------------------')
|
|
print('This is configuration #%s. Please provide configuration value(use "None" if not applicable).' % config)
|
|
print('---------------------------------------------')
|
|
value_conf = self._get_input('Configuration Value')
|
|
print('---------------------------------------------\n')
|
|
configurations[config] = value_conf
|
|
return configurations
|
|
|
|
def _execute_test_steps(self, case):
|
|
test_result = {}
|
|
print('------------------------------------------------------------------------')
|
|
print('Executing test case: %s' % case['test']['@alias'])
|
|
print('------------------------------------------------------------------------')
|
|
print('You have total %s test steps to be executed.' % len(case['test']['execution']))
|
|
print('------------------------------------------------------------------------\n')
|
|
for step, _ in sorted(case['test']['execution'].items(), key=lambda x: int(x[0])):
|
|
print('Step %s: %s' % (step, case['test']['execution'][step]['action']))
|
|
expected_output = case['test']['execution'][step]['expected_results']
|
|
if expected_output:
|
|
print('Expected output: %s' % expected_output)
|
|
while True:
|
|
done = input('\nPlease provide test results: (P)assed/(F)ailed/(B)locked/(S)kipped? \n').lower()
|
|
result_types = {'p':'PASSED',
|
|
'f':'FAILED',
|
|
'b':'BLOCKED',
|
|
's':'SKIPPED'}
|
|
if done in result_types:
|
|
for r in result_types:
|
|
if done == r:
|
|
res = result_types[r]
|
|
if res == 'FAILED':
|
|
log_input = input('\nPlease enter the error and the description of the log: (Ex:log:211 Error Bitbake)\n')
|
|
test_result.update({case['test']['@alias']: {'status': '%s' % res, 'log': '%s' % log_input}})
|
|
else:
|
|
test_result.update({case['test']['@alias']: {'status': '%s' % res}})
|
|
break
|
|
print('Invalid input!')
|
|
return test_result
|
|
|
|
def _get_write_dir(self):
|
|
return os.environ['BUILDDIR'] + '/tmp/log/manual/'
|
|
|
|
def run_test(self, case_file, config_options_file, testcase_config_file):
|
|
test_module = self._get_test_module(case_file)
|
|
cases = load_json_file(case_file)
|
|
config_options = {}
|
|
if config_options_file:
|
|
config_options = load_json_file(config_options_file)
|
|
configurations = self._get_config(config_options, test_module)
|
|
result_id = 'manual_%s_%s' % (test_module, configurations['STARTTIME'])
|
|
test_results = {}
|
|
if testcase_config_file:
|
|
test_case_config = load_json_file(testcase_config_file)
|
|
test_case_to_execute = test_case_config['testcases']
|
|
for case in copy.deepcopy(cases) :
|
|
if case['test']['@alias'] not in test_case_to_execute:
|
|
cases.remove(case)
|
|
|
|
print('\nTotal number of test cases in this test suite: %s\n' % len(cases))
|
|
for c in cases:
|
|
test_result = self._execute_test_steps(c)
|
|
test_results.update(test_result)
|
|
return configurations, result_id, self._get_write_dir(), test_results
|
|
|
|
def _get_true_false_input(self, input_message):
|
|
yes_list = ['Y', 'YES']
|
|
no_list = ['N', 'NO']
|
|
while True:
|
|
more_config_option = input(input_message).upper()
|
|
if more_config_option in yes_list or more_config_option in no_list:
|
|
break
|
|
print('Invalid input!')
|
|
if more_config_option in no_list:
|
|
return False
|
|
return True
|
|
|
|
def make_config_option_file(self, logger, case_file, config_options_file):
|
|
config_options = {}
|
|
if config_options_file:
|
|
config_options = load_json_file(config_options_file)
|
|
new_test_module = self._get_test_module(case_file)
|
|
print('Creating configuration options file for test module: %s' % new_test_module)
|
|
new_config_options = {}
|
|
|
|
while True:
|
|
config_name = input('\nPlease provide test configuration to create:\n').upper()
|
|
new_config_options[config_name] = {}
|
|
while True:
|
|
config_value = self._get_input('Configuration possible option value')
|
|
config_option_index = len(new_config_options[config_name]) + 1
|
|
new_config_options[config_name][config_option_index] = config_value
|
|
more_config_option = self._get_true_false_input('\nIs there more configuration option input: (Y)es/(N)o\n')
|
|
if not more_config_option:
|
|
break
|
|
more_config = self._get_true_false_input('\nIs there more configuration to create: (Y)es/(N)o\n')
|
|
if not more_config:
|
|
break
|
|
|
|
if new_config_options:
|
|
config_options[new_test_module] = new_config_options
|
|
if not config_options_file:
|
|
config_options_file = os.path.join(self._get_write_dir(), 'manual_config_options.json')
|
|
write_json_file(config_options_file, config_options)
|
|
logger.info('Configuration option file created at %s' % config_options_file)
|
|
|
|
def manualexecution(args, logger):
|
|
testrunner = ManualTestRunner()
|
|
if args.make_config_options_file:
|
|
testrunner.make_config_option_file(logger, args.file, args.config_options_file)
|
|
return 0
|
|
configurations, result_id, write_dir, test_results = testrunner.run_test(args.file, args.config_options_file, args.testcase_config_file)
|
|
resultjsonhelper = OETestResultJSONHelper()
|
|
resultjsonhelper.dump_testresult_file(write_dir, configurations, result_id, test_results)
|
|
return 0
|
|
|
|
def register_commands(subparsers):
|
|
"""Register subcommands from this plugin"""
|
|
parser_build = subparsers.add_parser('manualexecution', help='helper script for results populating during manual test execution.',
|
|
description='helper script for results populating during manual test execution. You can find manual test case JSON file in meta/lib/oeqa/manual/',
|
|
group='manualexecution')
|
|
parser_build.set_defaults(func=manualexecution)
|
|
parser_build.add_argument('file', help='specify path to manual test case JSON file.Note: Please use \"\" to encapsulate the file path.')
|
|
parser_build.add_argument('-c', '--config-options-file', default='',
|
|
help='the config options file to import and used as available configuration option selection or make config option file')
|
|
parser_build.add_argument('-m', '--make-config-options-file', action='store_true',
|
|
help='make the configuration options file based on provided inputs')
|
|
parser_build.add_argument('-t', '--testcase-config-file', default='',
|
|
help='the testcase configuration file to enable user to run a selected set of test case') |