summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--etc/vnftest/vnftest.yaml6
-rw-r--r--tools/vnftest_test_config.yaml6
-rw-r--r--vnftest/cmd/commands/task.py6
-rw-r--r--vnftest/common/constants.py6
-rw-r--r--vnftest/common/html_template.py13
-rw-r--r--vnftest/core/report.py7
-rw-r--r--vnftest/core/task.py403
-rw-r--r--vnftest/core/testcase.py10
-rw-r--r--vnftest/dispatcher/base.py2
-rw-r--r--vnftest/dispatcher/file.py10
-rw-r--r--vnftest/dispatcher/http.py21
-rw-r--r--vnftest/tests/unit/core/test_task.py15
-rw-r--r--vnftest/tests/unit/onap/test_rest_call.py5
13 files changed, 266 insertions, 244 deletions
diff --git a/etc/vnftest/vnftest.yaml b/etc/vnftest/vnftest.yaml
index cd5b961..f1b7750 100644
--- a/etc/vnftest/vnftest.yaml
+++ b/etc/vnftest/vnftest.yaml
@@ -16,8 +16,4 @@ dir:
conf: /etc/nvftest
repos: /home/onap/repos/vnftest
log: /var/log/vnftest
-
-file:
- output_file: /tmp/vnftest.out
- html_file: /tmp/vnftest.htm
- reporting_file: /tmp/report.html
+ report: /tmp \ No newline at end of file
diff --git a/tools/vnftest_test_config.yaml b/tools/vnftest_test_config.yaml
index cb2aa4b..ca72694 100644
--- a/tools/vnftest_test_config.yaml
+++ b/tools/vnftest_test_config.yaml
@@ -16,8 +16,4 @@ dir:
conf: ./
repos: ./
log: ./
-
-file:
- output_file: /tmp/vnftest.out
- html_file: /tmp/vnftest.htm
- reporting_file: /tmp/report.html
+ report: /tmp \ No newline at end of file
diff --git a/vnftest/cmd/commands/task.py b/vnftest/cmd/commands/task.py
index e10ac30..b839c8a 100644
--- a/vnftest/cmd/commands/task.py
+++ b/vnftest/cmd/commands/task.py
@@ -45,19 +45,19 @@ class TaskCommands(object): # pragma: no cover
output_file_default, default=output_file_default)
@cliargs("--suite", help="process test suite file instead of a task file",
action="store_true")
- def do_start(self, args, **kwargs):
+ def do_start(self, args):
param = change_osloobj_to_paras(args)
self.output_file = param.output_file
result = {}
LOG.info('Task START')
try:
- result = Task().start(param, **kwargs)
+ result = Task(param).start()
except Exception as e:
self._write_error_data(e)
LOG.exception("")
- if result.get('result', {}).get('criteria') == 'PASS':
+ if result.get('criteria') == 'PASS':
LOG.info('Task SUCCESS')
else:
LOG.info('Task FAILED')
diff --git a/vnftest/common/constants.py b/vnftest/common/constants.py
index 9996ca6..9634708 100644
--- a/vnftest/common/constants.py
+++ b/vnftest/common/constants.py
@@ -88,6 +88,7 @@ PACKAGE_DIR = get_param('dir.packages', join(VNFTEST_ROOT_PATH, '../../packages/
CONF_FILE = join(CONF_DIR, 'vnftest.conf')
REPOS_DIR = get_param('dir.repos', join(VNFTEST_ROOT_PATH, '..'))
LOG_DIR = get_param('dir.log', '/var/log/vnftest/')
+REPORT_DIR = get_param('dir.report', '/tmp/')
TASK_LOG_DIR = get_param('dir.tasklog', '/var/log/vnftest/')
CONF_SAMPLE_DIR = join(REPOS_DIR, 'etc/vnftest/')
@@ -95,11 +96,6 @@ SAMPLE_CASE_DIR = join(REPOS_DIR, 'samples')
TESTCASE_DIR = join(VNFTEST_ROOT_PATH, 'vnftest/test_config/onap/test_cases/')
TESTSUITE_DIR = join(VNFTEST_ROOT_PATH, 'vnftest/test_config/onap/test_suites/')
-# file
-DEFAULT_OUTPUT_FILE = get_param('file.output_file', '/tmp/vnftest.out')
-DEFAULT_HTML_FILE = get_param('file.html_file', '/tmp/vnftest.htm')
-REPORTING_FILE = get_param('file.reporting_file', '/tmp/report.html')
-
# api
API_PORT = 5000
DOCKER_URL = 'unix://var/run/docker.sock'
diff --git a/vnftest/common/html_template.py b/vnftest/common/html_template.py
index 7a91781..0196ba3 100644
--- a/vnftest/common/html_template.py
+++ b/vnftest/common/html_template.py
@@ -146,9 +146,9 @@ report_template = """
<hr/>
<div>
- <div>Task ID : {{result.task_id}} </div>
+ <div>Task ID : {{task_id}} </div>
<div style="margin-top:5px;">Criteria :
- <font> {{result.criteria}}</font>
+ <font> {{criteria}}</font>
</div>
<hr/>
@@ -160,7 +160,7 @@ report_template = """
<th>value</th>
</tr>
<tbody>
- {% for key, value in result.info.items() %}
+ {% for key, value in info.items() %}
<tr>
<td>{{ loop.index }}</td>
<td>{{key}}</td>
@@ -179,12 +179,11 @@ report_template = """
<th>value</th>
</tr>
<tbody>
- {% for key, value in result.testcases.items() %}
+ {% for testcase in testcases %}
<tr>
<td>{{ loop.index }}</td>
- <td>{{key}}</td>
- <td>{{value.criteria}}</td>
- <td>{{value.output}}</td>
+ <td>{{testcase.name}}</td>
+ <td>{{testcase.criteria}}</td>
</tr>
{% endfor %}
</tbody>
diff --git a/vnftest/core/report.py b/vnftest/core/report.py
index b8f8bb7..c648848 100644
--- a/vnftest/core/report.py
+++ b/vnftest/core/report.py
@@ -19,6 +19,7 @@ from __future__ import absolute_import
from __future__ import print_function
# import ast
+import os
import re
import uuid
@@ -120,7 +121,9 @@ class Report(object):
"Timestamp": self.Timestamp,
"task_id": self.task_id,
"table": table_vals})
- with open(consts.DEFAULT_HTML_FILE, "w") as file_open:
+ file_name = 'vnftest_' + str(self.task_id) + '.htm'
+ report_file = os.path.join(consts.REPORT_DIR, file_name)
+ with open(report_file, "w") as file_open:
file_open.write(Template_html.render(Context_html))
- print("Report generated. View /tmp/vnftest.htm")
+ print("Report generated. View " + report_file)
diff --git a/vnftest/core/task.py b/vnftest/core/task.py
index 41756af..a919b1f 100644
--- a/vnftest/core/task.py
+++ b/vnftest/core/task.py
@@ -26,7 +26,6 @@ import logging
import sys
import time
import uuid
-from collections import OrderedDict
import ipaddress
import os
@@ -61,9 +60,13 @@ class Task(object): # pragma: no cover
Set of commands to manage benchmark tasks.
"""
- def __init__(self):
+ def __init__(self, args):
self.context = None
self.outputs = {}
+ self.args = args or {}
+ task_id = getattr(args, 'task_id', None)
+ self.task_id = task_id if task_id is not None else str(uuid.uuid4())
+ self.task_info = TaskInfo(task_id)
def _set_dispatchers(self, output_config):
dispatchers = output_config.get('DEFAULT', {}).get('dispatcher',
@@ -71,14 +74,9 @@ class Task(object): # pragma: no cover
out_types = [s.strip() for s in dispatchers.split(',')]
output_config['DEFAULT']['dispatcher'] = out_types
- def start(self, args, **kwargs):
+ def start(self):
atexit.register(self.atexit_handler)
-
- task_id = getattr(args, 'task_id')
- self.task_id = task_id if task_id else str(uuid.uuid4())
-
self._set_log()
-
try:
output_config = utils.parse_ini_file(CONF_FILE)
except Exception:
@@ -86,7 +84,7 @@ class Task(object): # pragma: no cover
output_config = {}
self._init_output_config(output_config)
- self._set_output_config(output_config, args.output_file)
+ self._set_output_config(output_config, self.args.output_file)
LOG.debug('Output configuration is: %s', output_config)
self._set_dispatchers(output_config)
@@ -94,79 +92,77 @@ class Task(object): # pragma: no cover
# update dispatcher list
if 'file' in output_config['DEFAULT']['dispatcher']:
result = {'status': 0, 'result': {}}
- utils.write_json_to_file(args.output_file, result)
+ utils.write_json_to_file(self.args.output_file, result)
total_start_time = time.time()
- parser = TaskParser(args.inputfile)
+ parser = TaskParser(self.args.inputfile)
- if args.suite:
+ if self.args.suite:
# 1.parse suite, return suite_params info
task_files, task_args_list, task_args_fnames = \
parser.parse_suite()
else:
task_files = [parser.path]
- task_args_list = [args.task_args]
- task_args_fnames = [args.task_args_file]
+ task_args_list = [self.args.task_args]
+ task_args_fnames = [self.args.task_args_file]
LOG.debug("task_files:%s, task_args_list:%s, task_args_fnames:%s",
task_files, task_args_list, task_args_fnames)
- if args.parse_only:
+ if self.args.parse_only:
sys.exit(0)
- testcases = {}
- # parse task_files
- for i in range(0, len(task_files)):
- one_task_start_time = time.time()
- # the output of the previous task is the input of the new task
- inputs = copy.deepcopy(self.outputs)
- task_args_file = task_args_fnames[i]
- task_args = task_args_list[i]
- try:
- if task_args_file:
- with utils.load_resource(task_args_file) as f:
- inputs.update(parse_task_args("task_args_file", f.read()))
- inputs.update(parse_task_args("task_args", task_args))
- except TypeError:
- raise TypeError()
- parser.path = task_files[i]
- steps, run_in_parallel, meet_precondition, ret_context = \
- parser.parse_task(self.task_id, inputs)
-
- self.context = ret_context
-
- if not meet_precondition:
- LOG.info("meet_precondition is %s, please check envrionment",
- meet_precondition)
- continue
-
- case_name = os.path.splitext(os.path.basename(task_files[i]))[0]
- try:
- data = self._run(steps, run_in_parallel, args.output_file, inputs)
- except KeyboardInterrupt:
- raise
- except Exception:
- LOG.error('Testcase: "%s" FAILED!!!', case_name, exc_info=True)
- testcases[case_name] = {'criteria': 'FAIL', 'tc_data': []}
- else:
- criteria = self.evaluate_task_criteria(data)
- testcases[case_name] = {'criteria': criteria, 'tc_data': data, 'output': self.outputs}
-
- if args.keep_deploy:
- # keep deployment, forget about stack
- # (hide it for exit handler)
- self.context = None
- else:
- self.context.undeploy()
- self.context = None
- one_task_end_time = time.time()
- LOG.info("Task %s finished in %d secs", task_files[i],
- one_task_end_time - one_task_start_time)
-
- result = self._get_format_result(testcases)
+ try:
+ for i in range(0, len(task_files)):
+ one_task_start_time = time.time()
+ # the output of the previous task is the input of the new task
+ inputs = copy.deepcopy(self.outputs)
+ task_args_file = task_args_fnames[i]
+ task_args = task_args_list[i]
+ try:
+ if task_args_file:
+ with utils.load_resource(task_args_file) as f:
+ inputs.update(parse_task_args("task_args_file", f.read()))
+ inputs.update(parse_task_args("task_args", task_args))
+ except TypeError:
+ raise TypeError()
+ parser.path = task_files[i]
+ steps, run_in_parallel, meet_precondition, ret_context = \
+ parser.parse_task(self.task_id, inputs)
+
+ self.context = ret_context
+
+ if not meet_precondition:
+ LOG.info("meet_precondition is %s, please check envrionment",
+ meet_precondition)
+ continue
- self._do_output(output_config, result)
- self._generate_reporting(result)
+ case_name = os.path.splitext(os.path.basename(task_files[i]))[0]
+ try:
+ self._run(steps, case_name, run_in_parallel, self.args.output_file, inputs)
+ except KeyboardInterrupt:
+ raise
+ except Exception:
+ LOG.error('Testcase: "%s" FAILED!!!', case_name, exc_info=True)
+
+ if self.args.keep_deploy:
+ # keep deployment, forget about stack
+ # (hide it for exit handler)
+ self.context = None
+ else:
+ self.context.undeploy()
+ self.context = None
+ one_task_end_time = time.time()
+ LOG.info("Task %s finished in %d secs", task_files[i],
+ one_task_end_time - one_task_start_time)
+ except Exception as e:
+ LOG.error("Task fatal error", e)
+ self.task_info.task_fatal()
+ finally:
+ self.task_info.task_end()
+
+ self._do_output(output_config)
+ self._generate_reporting()
total_end_time = time.time()
LOG.info("Total finished in %d secs",
@@ -176,14 +172,16 @@ class Task(object): # pragma: no cover
LOG.info("To generate report, execute command "
"'vnftest report generate %(task_id)s %(tc)s'", step)
LOG.info("Task ALL DONE, exiting")
- return result
+ return self.task_info.result()
- def _generate_reporting(self, result):
+ def _generate_reporting(self):
env = Environment()
- with open(constants.REPORTING_FILE, 'w') as f:
- f.write(env.from_string(report_template).render(result))
+ file_name = 'report_' + str(self.task_id) + '.html'
+ report_file = os.path.join(constants.REPORT_DIR, file_name)
+ with open(report_file, 'w') as f:
+ f.write(env.from_string(report_template).render(self.task_info.result()))
- LOG.info("Report can be found in '%s'", constants.REPORTING_FILE)
+ LOG.info("Report can be found in '%s'", report_file)
def _set_log(self):
log_format = '%(asctime)s %(name)s %(filename)s:%(lineno)d %(levelname)s %(message)s'
@@ -221,104 +219,61 @@ class Task(object): # pragma: no cover
k = 'dispatcher_{}'.format(output_config['DEFAULT']['dispatcher'])
output_config[k]['target'] = target
- def _get_format_result(self, testcases):
- criteria = self._get_task_criteria(testcases)
-
- info = {
- 'deploy_step': os.environ.get('DEPLOY_STEP', 'unknown'),
- 'installer': os.environ.get('INSTALLER_TYPE', 'unknown'),
- 'pod_name': os.environ.get('NODE_NAME', 'unknown'),
- 'version': os.environ.get('VNFTEST_BRANCH', 'unknown')
- }
-
- result = {
- 'status': 1,
- 'result': {
- 'criteria': criteria,
- 'task_id': self.task_id,
- 'info': info,
- 'testcases': testcases
- }
- }
-
- return result
-
- def _get_task_criteria(self, testcases):
- criteria = any(t.get('criteria') != 'PASS' for t in testcases.values())
- if criteria:
- return 'FAIL'
- else:
- return 'PASS'
-
- def evaluate_task_criteria(self, steps_result_list):
- for step_result in steps_result_list:
- errors_list = step_result['errors']
- if errors_list is not None and len(errors_list) > 0:
- return 'FAIL'
- return 'PASS'
-
- def _do_output(self, output_config, result):
+ def _do_output(self, output_config):
dispatchers = DispatcherBase.get(output_config)
for dispatcher in dispatchers:
- dispatcher.flush_result_data(result)
+ dispatcher.flush_result_data(self.task_id, self.task_info.result())
- def _run(self, steps, run_in_parallel, output_file, inputs):
+ def _run(self, steps, case_name, run_in_parallel, output_file, inputs):
"""Deploys context and calls runners"""
if self.context:
self.context.deploy()
- background_runners = []
-
- result = []
- # Start all background steps
- for step in filter(_is_background_step, steps):
- step["runner"] = dict(type="Duration", duration=1000000000)
- runner = self.run_one_step(step, output_file, inputs)
- background_runners.append(runner)
-
- runners = []
- if run_in_parallel:
- for step in steps:
- if not _is_background_step(step):
- runner = self.run_one_step(step, output_file, inputs)
- runners.append(runner)
-
- # Wait for runners to finish
- for runner in runners:
- status = runner_join(runner, background_runners, self.outputs, result)
- if status != 0:
- raise RuntimeError(
- "{0} runner status {1}".format(runner.__execution_type__, status))
- LOG.info("Runner ended, output in %s", output_file)
- else:
- # run serially
+ try:
+ self.task_info.testcase_start(case_name)
for step in steps:
- if not _is_background_step(step):
- runner = self.run_one_step(step, output_file, inputs)
- status = runner_join(runner, background_runners, self.outputs, result)
- if status != 0:
- LOG.error('Step NO.%s: "%s" ERROR!',
- steps.index(step) + 1,
- step.get('type'))
- raise RuntimeError(
- "{0} runner status {1}".format(runner.__execution_type__, status))
- LOG.info("Runner ended, output in %s", output_file)
-
- # Abort background runners
- for runner in background_runners:
- runner.abort()
-
- # Wait for background runners to finish
- for runner in background_runners:
- status = runner.join(self.outputs, result)
- if status is None:
- # Nuke if it did not stop nicely
- base_runner.Runner.terminate(runner)
- runner.join(self.outputs, result)
- base_runner.Runner.release(runner)
-
- print("Background task ended")
- return result
+ step_unique_id = self.task_info.step_add(case_name, step['name'])
+ step['step_unique_id'] = step_unique_id
+
+ background_runners = []
+ result = []
+ # Start all background steps
+ for step in filter(_is_background_step, steps):
+ step["runner"] = dict(type="Duration", duration=1000000000)
+ self.task_info.step_start(step['step_unique_id'])
+ runner = self.run_one_step(step, output_file, inputs)
+ background_runners.append([step, runner])
+
+ runners = []
+ if run_in_parallel:
+ for step in steps:
+ if not _is_background_step(step):
+ self.task_info.step_start(step['step_unique_id'])
+ runner = self.run_one_step(step, output_file, inputs)
+ runners.append([step, runner])
+
+ # Wait for runners to finish
+ for runner_item in runners:
+ self.finalize_step(runner_item[0], runner_item[1], result)
+ else:
+ # run serially
+ for step in steps:
+ if not _is_background_step(step):
+ self.task_info.step_start(step['step_unique_id'])
+ runner = self.run_one_step(step, output_file, inputs)
+ self.finalize_step(step, runner, result)
+
+ # Abort background runners
+ for runner_item in background_runners:
+ runner_item[1].abort()
+
+ # Wait for background runners to finish
+ for runner_item in background_runners:
+ runner = runner_item[1]
+ self.finalize_step(step, runner, result)
+ return result
+ finally:
+ self.task_info.testcase_end(case_name)
def atexit_handler(self):
"""handler for process termination"""
@@ -355,6 +310,16 @@ class Task(object): # pragma: no cover
runner.run(step_cfg, self.context, inputs)
return runner
+ def finalize_step(self, step, runner, result):
+ step_result_list = []
+ status = runner_join(runner, self.outputs, step_result_list)
+ self.task_info.step_end(step['step_unique_id'], step_result_list)
+ if status != 0:
+ raise RuntimeError(
+ "{0} runner status {1}".format(runner.__execution_type__, status))
+ LOG.info("Runner ended")
+ result.extend(step_result_list)
+
class TaskParser(object): # pragma: no cover
"""Parser for task config files in yaml format"""
@@ -492,26 +457,17 @@ class TaskParser(object): # pragma: no cover
if "precondition" in cfg:
precondition = cfg["precondition"]
installer_type = precondition.get("installer_type", None)
- deploy_steps = precondition.get("deploy_steps", None)
tc_fit_pods = precondition.get("pod_name", None)
installer_type_env = os.environ.get('INSTALL_TYPE', None)
- deploy_step_env = os.environ.get('DEPLOY_STEP', None)
pod_name_env = os.environ.get('NODE_NAME', None)
LOG.info("installer_type: %s, installer_type_env: %s",
installer_type, installer_type_env)
- LOG.info("deploy_steps: %s, deploy_step_env: %s",
- deploy_steps, deploy_step_env)
LOG.info("tc_fit_pods: %s, pod_name_env: %s",
tc_fit_pods, pod_name_env)
if installer_type and installer_type_env:
if installer_type_env not in installer_type:
return False
- if deploy_steps and deploy_step_env:
- deploy_steps_list = deploy_steps.split(',')
- for deploy_step in deploy_steps_list:
- if deploy_step_env.startswith(deploy_step):
- return True
return False
if tc_fit_pods and pod_name_env:
if pod_name_env not in tc_fit_pods:
@@ -519,6 +475,97 @@ class TaskParser(object): # pragma: no cover
return True
+class TaskInfo(object):
+ info_dict = {}
+ info_dict['status'] = 'IN_PROGRESS'
+ info_dict['criteria'] = 'N/A'
+ result_info_dict = {}
+ info_dict['info'] = result_info_dict
+ test_cases_list = []
+ info_dict['testcases'] = test_cases_list
+ helper_test_cases_dict = {}
+ helper_test_steps_dict = {}
+ step_id_helper = 0
+
+ def __init__(self, task_id):
+ self.info_dict['task_id'] = task_id
+
+ def task_end(self):
+ if self.info_dict['criteria'] == 'N/A':
+ criteria = 'PASS'
+ for testcase in self.info_dict['testcases']:
+ if testcase['criteria'] == 'FAIL':
+ criteria = 'FAIL'
+ break
+ self.info_dict['criteria'] = criteria
+ self.info_dict['status'] = 'FINISHED'
+
+ def task_fatal(self):
+ self.info_dict['criteria'] = 'FATAL'
+ self.info_dict['status'] = 'FINISHED'
+
+ def testcase_start(self, testcase_name):
+ testcase_dict = {}
+ testcase_dict['name'] = testcase_name
+ testcase_dict['criteria'] = 'N/A'
+ testcase_dict['status'] = 'IN_PROGRESS'
+ testcase_dict['steps'] = []
+ self.test_cases_list.append(testcase_dict)
+ self.helper_test_cases_dict[testcase_name] = testcase_dict
+
+ def testcase_end(self, testcase_name):
+ testcase_dict = self.helper_test_cases_dict[testcase_name]
+ criteria = 'PASS'
+ for step in testcase_dict['steps']:
+ if step['criteria'] == 'FAIL':
+ criteria = 'FAIL'
+ break
+ testcase_dict['criteria'] = criteria
+ testcase_dict['status'] = 'FINISHED'
+
+ def step_add(self, testcase_name, step_name):
+ step_dict = {}
+ step_dict['name'] = step_name
+ step_dict['criteria'] = 'N/A'
+ step_dict['status'] = 'NOT_STARTED'
+ step_dict['results'] = []
+ testcase_dict = self.helper_test_cases_dict[testcase_name]
+ testcase_dict['steps'].append(step_dict)
+ self.step_id_helper += 1
+ step_unique_id = step_name + '_' + str(self.step_id_helper)
+ self.helper_test_steps_dict[step_unique_id] = step_dict
+ return step_unique_id
+
+ def step_start(self, step_unique_id):
+ step_dict = self.helper_test_steps_dict[step_unique_id]
+ step_dict['status'] = 'IN_PROGRESS'
+
+ def step_end(self, step_unique_id, result_list):
+ step_dict = self.helper_test_steps_dict[step_unique_id]
+ errors_count = 0
+ for result in result_list:
+ result_item = {
+ 'timestamp': result['timestamp'],
+ 'sequence': result['sequence'],
+ 'data': [],
+ 'errors': []
+ }
+ for k, v in result['data'].items():
+ result_item['data'].append({'type': 'String', 'key': k, 'value': str(v)})
+
+ for error in result['errors']:
+ result_item['errors'].append({'type': 'String', 'key': 'error', 'value': str(error)})
+ errors_count += 1
+ step_dict['results'].append(result_item)
+ if errors_count > 0:
+ step_dict['criteria'] = 'FAIL'
+ else:
+ step_dict['criteria'] = 'PASS'
+ step_dict['status'] = 'FINISHED'
+
+ def result(self):
+ return self.info_dict
+
def is_ip_addr(addr):
"""check if string addr is an IP address"""
try:
@@ -541,22 +588,10 @@ def _is_background_step(step):
return False
-def runner_join(runner, background_runners, outputs, result):
- """join (wait for) a runner, exit process at runner failure
- :param background_runners:
- :type background_runners:
- :param outputs:
- :type outputs: dict
- :param result:
- :type result: list
- """
+def runner_join(runner, outputs, result):
while runner.poll() is None:
outputs.update(runner.get_output())
result.extend(runner.get_result())
- # drain all the background runner queues
- for background in background_runners:
- outputs.update(background.get_output())
- result.extend(background.get_result())
status = runner.join(outputs, result)
base_runner.Runner.release(runner)
return status
diff --git a/vnftest/core/testcase.py b/vnftest/core/testcase.py
index 2c5c4b4..df86dd6 100644
--- a/vnftest/core/testcase.py
+++ b/vnftest/core/testcase.py
@@ -63,14 +63,13 @@ class Testcase(object):
LOG.exception('Failed to load test case:\n%s\n', testcase_file)
raise
- description, installer, deploy_steps = self._parse_testcase(
+ description, installer = self._parse_testcase(
testcase_info)
record = {
'Name': testcase_file.split(".")[0],
'Description': description,
- 'installer': installer,
- 'deploy_steps': deploy_steps
+ 'installer': installer
}
return record
@@ -82,11 +81,8 @@ class Testcase(object):
test_precondition = testcase_cfg.get('precondition', {})
installer_type = test_precondition.get('installer_type', 'all')
- deploy_steps = test_precondition.get('deploy_steps', 'all')
-
description = self._get_description(testcase_cfg)
-
- return description, installer_type, deploy_steps
+ return description, installer_type
def _get_description(self, testcase_cfg):
try:
diff --git a/vnftest/dispatcher/base.py b/vnftest/dispatcher/base.py
index e53dd96..932b8ef 100644
--- a/vnftest/dispatcher/base.py
+++ b/vnftest/dispatcher/base.py
@@ -46,5 +46,5 @@ class Base(object):
return list_dispatcher
@abc.abstractmethod
- def flush_result_data(self, data):
+ def flush_result_data(self, id, data):
"""Flush result data into permanent storage media interface."""
diff --git a/vnftest/dispatcher/file.py b/vnftest/dispatcher/file.py
index 83d0fee..2504457 100644
--- a/vnftest/dispatcher/file.py
+++ b/vnftest/dispatcher/file.py
@@ -16,6 +16,8 @@
from __future__ import absolute_import
+import os
+
from vnftest.dispatcher.base import Base as DispatchBase
from vnftest.common import constants as consts
from vnftest.common import utils
@@ -29,8 +31,8 @@ class FileDispatcher(DispatchBase):
def __init__(self, conf):
super(FileDispatcher, self).__init__(conf)
- self.target = conf['dispatcher_file'].get('file_path',
- consts.DEFAULT_OUTPUT_FILE)
- def flush_result_data(self, data):
- utils.write_json_to_file(self.target, data)
+ def flush_result_data(self, id, data):
+ file_name = 'vnftest_' + str(id) + '.out'
+ target = self.conf['dispatcher_file'].get('file_path', os.path.join(consts.REPORT_DIR, file_name))
+ utils.write_json_to_file(target, data)
diff --git a/vnftest/dispatcher/http.py b/vnftest/dispatcher/http.py
index da66c90..900792a 100644
--- a/vnftest/dispatcher/http.py
+++ b/vnftest/dispatcher/http.py
@@ -41,7 +41,7 @@ class HttpDispatcher(DispatchBase):
self.timeout = int(http_conf.get('timeout', 5))
self.target = http_conf.get('target', 'http://127.0.0.1:8000/results')
- def flush_result_data(self, data):
+ def flush_result_data(self, id, data):
if self.target == '':
# if the target was not set, do not do anything
LOG.error('Dispatcher target was not set, no data will'
@@ -54,28 +54,25 @@ class HttpDispatcher(DispatchBase):
self.criteria = result['criteria']
testcases = result['testcases']
- for case, data in testcases.items():
- self._upload_case_result(case, data)
+ for testcase in testcases:
+ self._upload_case_result(testcase)
- def _upload_case_result(self, case, data):
+ def _upload_case_result(self, testcase):
try:
- step_data = data.get('tc_data', [])[0]
+ step_data = testcase.get('steps', [])[0]
+ step_result = step_data['results'][0]
except IndexError:
current_time = datetime.now()
else:
- timestamp = float(step_data.get('timestamp', 0.0))
+ timestamp = float(step_result.get('timestamp', 0.0))
current_time = datetime.fromtimestamp(timestamp)
result = {
"project_name": "vnftest",
- "case_name": case,
+ "case_name": testcase['name'],
"description": "vnftest ci step status",
- "step": self.info.get('deploy_step'),
- "version": self.info.get('version'),
- "pod_name": self.info.get('pod_name'),
- "installer": self.info.get('installer'),
"build_tag": os.environ.get('BUILD_TAG'),
- "criteria": data.get('criteria'),
+ "criteria": testcase.get('criteria'),
"start_date": current_time.strftime('%Y-%m-%d %H:%M:%S'),
"stop_date": current_time.strftime('%Y-%m-%d %H:%M:%S'),
"trust_indicator": "",
diff --git a/vnftest/tests/unit/core/test_task.py b/vnftest/tests/unit/core/test_task.py
index d554951..7f26d61 100644
--- a/vnftest/tests/unit/core/test_task.py
+++ b/vnftest/tests/unit/core/test_task.py
@@ -26,14 +26,14 @@ from vnftest.core import task
class TaskTestCase(unittest.TestCase):
def test_set_dispatchers(self):
- t = task.Task()
+ t = task.Task({})
output_config = {"DEFAULT": {"dispatcher": "file, http"}}
t._set_dispatchers(output_config)
self.assertEqual(output_config, output_config)
@mock.patch.object(task, 'DispatcherBase')
def test__do_output(self, mock_dispatcher):
- t = task.Task()
+ t = task.Task({})
output_config = {"DEFAULT": {"dispatcher": "file, http"}}
dispatcher1 = mock.MagicMock()
@@ -44,7 +44,7 @@ class TaskTestCase(unittest.TestCase):
mock_dispatcher.get = mock.MagicMock(return_value=[dispatcher1,
dispatcher2])
- self.assertIsNone(t._do_output(output_config, {}))
+ self.assertIsNone(t._do_output(output_config))
@mock.patch.object(task, 'Context')
@mock.patch.object(task, 'base_runner')
@@ -55,16 +55,17 @@ class TaskTestCase(unittest.TestCase):
'interval': 1,
'type': 'Duration'
},
- 'type': 'Dummy'
+ 'type': 'Dummy',
+ 'name': 'Dummy Step'
}
- t = task.Task()
+ t = task.Task({})
runner = mock.Mock()
runner.join.return_value = 0
runner.get_output.return_value = {}
runner.get_result.return_value = []
mock_base_runner.Runner.get.return_value = runner
- t._run([step], False, "vnftest.out", {})
+ t._run([step], 'dummy_case', False, "vnftest.out", {})
self.assertTrue(runner.run.called)
def test_parse_suite_no_constraint_no_args(self):
@@ -121,7 +122,7 @@ class TaskTestCase(unittest.TestCase):
@mock.patch.object(task, 'utils')
@mock.patch('logging.root')
def test_set_log(self, mock_logging_root, *args):
- task_obj = task.Task()
+ task_obj = task.Task({})
task_obj.task_id = 'task_id'
task_obj._set_log()
mock_logging_root.addHandler.assert_called()
diff --git a/vnftest/tests/unit/onap/test_rest_call.py b/vnftest/tests/unit/onap/test_rest_call.py
index 13aced7..b100ecb 100644
--- a/vnftest/tests/unit/onap/test_rest_call.py
+++ b/vnftest/tests/unit/onap/test_rest_call.py
@@ -21,6 +21,7 @@ from vnftest.core import task
class RestCallTestCase(testtools.TestCase):
step = {
+ 'name': 'DummyName',
'type': 'RestCall',
'options': {
'file': "dummy.yaml",
@@ -49,10 +50,10 @@ class RestCallTestCase(testtools.TestCase):
@mock.patch('vnftest.steps.rest_call.RestCall.execute_operation_impl')
def test_run(self, mock_execute_operation):
mock_execute_operation.return_value = {'body': {'value': 'output1'}}
- t = task.Task()
+ t = task.Task({})
context_cfg = {}
context = Context.get("CSAR")
context.init(context_cfg)
t.context = context
- output = t._run([RestCallTestCase.step], False, "vnftest.out", {})
+ output = t._run([RestCallTestCase.step], 'dummy_case', False, "vnftest.out", {})
self.assertEquals(output[0]['data']['output1'], 'output1')