| #!/usr/bin/env python3 |
| |
| from __future__ import print_function |
| import gc |
| import sys |
| import os |
| import select |
| import signal |
| import unittest |
| import tempfile |
| import time |
| import faulthandler |
| import random |
| import copy |
| import psutil |
| import platform |
| from collections import deque |
| from threading import Thread, Event |
| from inspect import getdoc, isclass |
| from traceback import format_exception |
| from logging import FileHandler, DEBUG, Formatter |
| |
| import scapy.compat |
| from scapy.packet import Raw |
| import hook as hookmodule |
| from vpp_pg_interface import VppPGInterface |
| from vpp_sub_interface import VppSubInterface |
| from vpp_lo_interface import VppLoInterface |
| from vpp_bvi_interface import VppBviInterface |
| from vpp_papi_provider import VppPapiProvider |
| import vpp_papi |
| from vpp_papi.vpp_stats import VPPStats |
| from vpp_papi.vpp_transport_shmem import VppTransportShmemIOError |
| from log import RED, GREEN, YELLOW, double_line_delim, single_line_delim, \ |
| get_logger, colorize |
| from vpp_object import VppObjectRegistry |
| from util import ppp, is_core_present |
| from scapy.layers.inet import IPerror, TCPerror, UDPerror, ICMPerror |
| from scapy.layers.inet6 import ICMPv6DestUnreach, ICMPv6EchoRequest |
| from scapy.layers.inet6 import ICMPv6EchoReply |
| |
| if os.name == 'posix' and sys.version_info[0] < 3: |
| # using subprocess32 is recommended by python official documentation |
| # @ https://docs.python.org/2/library/subprocess.html |
| import subprocess32 as subprocess |
| else: |
| import subprocess |
| |
| # Python2/3 compatible |
| try: |
| input = raw_input |
| except NameError: |
| pass |
| |
| PASS = 0 |
| FAIL = 1 |
| ERROR = 2 |
| SKIP = 3 |
| TEST_RUN = 4 |
| |
| |
| class BoolEnvironmentVariable(object): |
| |
| def __init__(self, env_var_name, default='n', true_values=None): |
| self.name = env_var_name |
| self.default = default |
| self.true_values = true_values if true_values is not None else \ |
| ("y", "yes", "1") |
| |
| def __bool__(self): |
| return os.getenv(self.name, self.default).lower() in self.true_values |
| |
| if sys.version_info[0] == 2: |
| __nonzero__ = __bool__ |
| |
| def __repr__(self): |
| return 'BoolEnvironmentVariable(%r, default=%r, true_values=%r)' % \ |
| (self.name, self.default, self.true_values) |
| |
| |
| debug_framework = BoolEnvironmentVariable('TEST_DEBUG') |
| if debug_framework: |
| import debug_internal |
| |
| """ |
| Test framework module. |
| |
| The module provides a set of tools for constructing and running tests and |
| representing the results. |
| """ |
| |
| |
| class VppDiedError(Exception): |
| """ exception for reporting that the subprocess has died.""" |
| |
| signals_by_value = {v: k for k, v in signal.__dict__.items() if |
| k.startswith('SIG') and not k.startswith('SIG_')} |
| |
| def __init__(self, rv=None, testcase=None, method_name=None): |
| self.rv = rv |
| self.signal_name = None |
| self.testcase = testcase |
| self.method_name = method_name |
| |
| try: |
| self.signal_name = VppDiedError.signals_by_value[-rv] |
| except (KeyError, TypeError): |
| pass |
| |
| if testcase is None and method_name is None: |
| in_msg = '' |
| else: |
| in_msg = 'running %s.%s ' % (testcase, method_name) |
| |
| msg = "VPP subprocess died %sunexpectedly with return code: %d%s." % ( |
| in_msg, |
| self.rv, |
| ' [%s]' % (self.signal_name if |
| self.signal_name is not None else '')) |
| super(VppDiedError, self).__init__(msg) |
| |
| |
| class _PacketInfo(object): |
| """Private class to create packet info object. |
| |
| Help process information about the next packet. |
| Set variables to default values. |
| """ |
| #: Store the index of the packet. |
| index = -1 |
| #: Store the index of the source packet generator interface of the packet. |
| src = -1 |
| #: Store the index of the destination packet generator interface |
| #: of the packet. |
| dst = -1 |
| #: Store expected ip version |
| ip = -1 |
| #: Store expected upper protocol |
| proto = -1 |
| #: Store the copy of the former packet. |
| data = None |
| |
| def __eq__(self, other): |
| index = self.index == other.index |
| src = self.src == other.src |
| dst = self.dst == other.dst |
| data = self.data == other.data |
| return index and src and dst and data |
| |
| |
| def pump_output(testclass): |
| """ pump output from vpp stdout/stderr to proper queues """ |
| stdout_fragment = "" |
| stderr_fragment = "" |
| while not testclass.pump_thread_stop_flag.is_set(): |
| readable = select.select([testclass.vpp.stdout.fileno(), |
| testclass.vpp.stderr.fileno(), |
| testclass.pump_thread_wakeup_pipe[0]], |
| [], [])[0] |
| if testclass.vpp.stdout.fileno() in readable: |
| read = os.read(testclass.vpp.stdout.fileno(), 102400) |
| if len(read) > 0: |
| split = read.decode('ascii', |
| errors='backslashreplace').splitlines(True) |
| if len(stdout_fragment) > 0: |
| split[0] = "%s%s" % (stdout_fragment, split[0]) |
| if len(split) > 0 and split[-1].endswith("\n"): |
| limit = None |
| else: |
| limit = -1 |
| stdout_fragment = split[-1] |
| testclass.vpp_stdout_deque.extend(split[:limit]) |
| if not testclass.cache_vpp_output: |
| for line in split[:limit]: |
| testclass.logger.info( |
| "VPP STDOUT: %s" % line.rstrip("\n")) |
| if testclass.vpp.stderr.fileno() in readable: |
| read = os.read(testclass.vpp.stderr.fileno(), 102400) |
| if len(read) > 0: |
| split = read.decode('ascii', |
| errors='backslashreplace').splitlines(True) |
| if len(stderr_fragment) > 0: |
| split[0] = "%s%s" % (stderr_fragment, split[0]) |
| if len(split) > 0 and split[-1].endswith("\n"): |
| limit = None |
| else: |
| limit = -1 |
| stderr_fragment = split[-1] |
| |
| testclass.vpp_stderr_deque.extend(split[:limit]) |
| if not testclass.cache_vpp_output: |
| for line in split[:limit]: |
| testclass.logger.error( |
| "VPP STDERR: %s" % line.rstrip("\n")) |
| # ignoring the dummy pipe here intentionally - the |
| # flag will take care of properly terminating the loop |
| |
| |
| def _is_skip_aarch64_set(): |
| return BoolEnvironmentVariable('SKIP_AARCH64') |
| |
| |
| is_skip_aarch64_set = _is_skip_aarch64_set() |
| |
| |
| def _is_platform_aarch64(): |
| return platform.machine() == 'aarch64' |
| |
| |
| is_platform_aarch64 = _is_platform_aarch64() |
| |
| |
| def _running_extended_tests(): |
| return BoolEnvironmentVariable("EXTENDED_TESTS") |
| |
| |
| running_extended_tests = _running_extended_tests() |
| |
| |
| def _running_gcov_tests(): |
| return BoolEnvironmentVariable("GCOV_TESTS") |
| |
| |
| running_gcov_tests = _running_gcov_tests() |
| |
| |
| def _running_on_centos(): |
| os_id = os.getenv("OS_ID", "") |
| return True if "centos" in os_id.lower() else False |
| |
| |
| running_on_centos = _running_on_centos() |
| |
| |
| class KeepAliveReporter(object): |
| """ |
| Singleton object which reports test start to parent process |
| """ |
| _shared_state = {} |
| |
| def __init__(self): |
| self.__dict__ = self._shared_state |
| self._pipe = None |
| |
| @property |
| def pipe(self): |
| return self._pipe |
| |
| @pipe.setter |
| def pipe(self, pipe): |
| if self._pipe is not None: |
| raise Exception("Internal error - pipe should only be set once.") |
| self._pipe = pipe |
| |
| def send_keep_alive(self, test, desc=None): |
| """ |
| Write current test tmpdir & desc to keep-alive pipe to signal liveness |
| """ |
| if self.pipe is None: |
| # if not running forked.. |
| return |
| |
| if isclass(test): |
| desc = '%s (%s)' % (desc, unittest.util.strclass(test)) |
| else: |
| desc = test.id() |
| |
| self.pipe.send((desc, test.vpp_bin, test.tempdir, test.vpp.pid)) |
| |
| |
| class VppTestCase(unittest.TestCase): |
| """This subclass is a base class for VPP test cases that are implemented as |
| classes. It provides methods to create and run test case. |
| """ |
| |
| extra_vpp_punt_config = [] |
| extra_vpp_plugin_config = [] |
| vapi_response_timeout = 5 |
| |
| @property |
| def packet_infos(self): |
| """List of packet infos""" |
| return self._packet_infos |
| |
| @classmethod |
| def get_packet_count_for_if_idx(cls, dst_if_index): |
| """Get the number of packet info for specified destination if index""" |
| if dst_if_index in cls._packet_count_for_dst_if_idx: |
| return cls._packet_count_for_dst_if_idx[dst_if_index] |
| else: |
| return 0 |
| |
| @classmethod |
| def instance(cls): |
| """Return the instance of this testcase""" |
| return cls.test_instance |
| |
| @classmethod |
| def set_debug_flags(cls, d): |
| cls.gdbserver_port = 7777 |
| cls.debug_core = False |
| cls.debug_gdb = False |
| cls.debug_gdbserver = False |
| cls.debug_all = False |
| if d is None: |
| return |
| dl = d.lower() |
| if dl == "core": |
| cls.debug_core = True |
| elif dl == "gdb" or dl == "gdb-all": |
| cls.debug_gdb = True |
| elif dl == "gdbserver" or dl == "gdbserver-all": |
| cls.debug_gdbserver = True |
| else: |
| raise Exception("Unrecognized DEBUG option: '%s'" % d) |
| if dl == "gdb-all" or dl == "gdbserver-all": |
| cls.debug_all = True |
| |
| @staticmethod |
| def get_least_used_cpu(): |
| cpu_usage_list = [set(range(psutil.cpu_count()))] |
| vpp_processes = [p for p in psutil.process_iter(attrs=['pid', 'name']) |
| if 'vpp_main' == p.info['name']] |
| for vpp_process in vpp_processes: |
| for cpu_usage_set in cpu_usage_list: |
| try: |
| cpu_num = vpp_process.cpu_num() |
| if cpu_num in cpu_usage_set: |
| cpu_usage_set_index = cpu_usage_list.index( |
| cpu_usage_set) |
| if cpu_usage_set_index == len(cpu_usage_list) - 1: |
| cpu_usage_list.append({cpu_num}) |
| else: |
| cpu_usage_list[cpu_usage_set_index + 1].add( |
| cpu_num) |
| cpu_usage_set.remove(cpu_num) |
| break |
| except psutil.NoSuchProcess: |
| pass |
| |
| for cpu_usage_set in cpu_usage_list: |
| if len(cpu_usage_set) > 0: |
| min_usage_set = cpu_usage_set |
| break |
| |
| return random.choice(tuple(min_usage_set)) |
| |
| @classmethod |
| def setUpConstants(cls): |
| """ Set-up the test case class based on environment variables """ |
| cls.step = BoolEnvironmentVariable('STEP') |
| d = os.getenv("DEBUG", None) |
| # inverted case to handle '' == True |
| c = os.getenv("CACHE_OUTPUT", "1") |
| cls.cache_vpp_output = False if c.lower() in ("n", "no", "0") else True |
| cls.set_debug_flags(d) |
| cls.vpp_bin = os.getenv('VPP_BIN', "vpp") |
| cls.plugin_path = os.getenv('VPP_PLUGIN_PATH') |
| cls.test_plugin_path = os.getenv('VPP_TEST_PLUGIN_PATH') |
| cls.extern_plugin_path = os.getenv('EXTERN_PLUGINS') |
| plugin_path = None |
| if cls.plugin_path is not None: |
| if cls.extern_plugin_path is not None: |
| plugin_path = "%s:%s" % ( |
| cls.plugin_path, cls.extern_plugin_path) |
| else: |
| plugin_path = cls.plugin_path |
| elif cls.extern_plugin_path is not None: |
| plugin_path = cls.extern_plugin_path |
| debug_cli = "" |
| if cls.step or cls.debug_gdb or cls.debug_gdbserver: |
| debug_cli = "cli-listen localhost:5002" |
| coredump_size = None |
| size = os.getenv("COREDUMP_SIZE") |
| if size is not None: |
| coredump_size = "coredump-size %s" % size |
| if coredump_size is None: |
| coredump_size = "coredump-size unlimited" |
| |
| cpu_core_number = cls.get_least_used_cpu() |
| if not hasattr(cls, "worker_config"): |
| cls.worker_config = "" |
| |
| default_variant = os.getenv("VARIANT") |
| if default_variant is not None: |
| default_variant = "defaults { %s 100 }" % default_variant |
| else: |
| default_variant = "" |
| |
| cls.vpp_cmdline = [cls.vpp_bin, "unix", |
| "{", "nodaemon", debug_cli, "full-coredump", |
| coredump_size, "runtime-dir", cls.tempdir, "}", |
| "api-trace", "{", "on", "}", "api-segment", "{", |
| "prefix", cls.shm_prefix, "}", "cpu", "{", |
| "main-core", str(cpu_core_number), |
| cls.worker_config, "}", |
| "physmem", "{", "max-size", "32m", "}", |
| "statseg", "{", "socket-name", cls.stats_sock, "}", |
| "socksvr", "{", "socket-name", cls.api_sock, "}", |
| "node { ", default_variant, "}", |
| "plugins", |
| "{", "plugin", "dpdk_plugin.so", "{", "disable", |
| "}", "plugin", "rdma_plugin.so", "{", "disable", |
| "}", "plugin", "unittest_plugin.so", "{", "enable", |
| "}"] + cls.extra_vpp_plugin_config + ["}", ] |
| |
| if cls.extra_vpp_punt_config is not None: |
| cls.vpp_cmdline.extend(cls.extra_vpp_punt_config) |
| if plugin_path is not None: |
| cls.vpp_cmdline.extend(["plugin_path", plugin_path]) |
| if cls.test_plugin_path is not None: |
| cls.vpp_cmdline.extend(["test_plugin_path", cls.test_plugin_path]) |
| |
| cls.logger.info("vpp_cmdline args: %s" % cls.vpp_cmdline) |
| cls.logger.info("vpp_cmdline: %s" % " ".join(cls.vpp_cmdline)) |
| |
| @classmethod |
| def wait_for_enter(cls): |
| if cls.debug_gdbserver: |
| print(double_line_delim) |
| print("Spawned GDB server with PID: %d" % cls.vpp.pid) |
| elif cls.debug_gdb: |
| print(double_line_delim) |
| print("Spawned VPP with PID: %d" % cls.vpp.pid) |
| else: |
| cls.logger.debug("Spawned VPP with PID: %d" % cls.vpp.pid) |
| return |
| print(single_line_delim) |
| print("You can debug VPP using:") |
| if cls.debug_gdbserver: |
| print("sudo gdb " + cls.vpp_bin + |
| " -ex 'target remote localhost:{port}'" |
| .format(port=cls.gdbserver_port)) |
| print("Now is the time to attach gdb by running the above " |
| "command, set up breakpoints etc., then resume VPP from " |
| "within gdb by issuing the 'continue' command") |
| cls.gdbserver_port += 1 |
| elif cls.debug_gdb: |
| print("sudo gdb " + cls.vpp_bin + " -ex 'attach %s'" % cls.vpp.pid) |
| print("Now is the time to attach gdb by running the above " |
| "command and set up breakpoints etc., then resume VPP from" |
| " within gdb by issuing the 'continue' command") |
| print(single_line_delim) |
| input("Press ENTER to continue running the testcase...") |
| |
| @classmethod |
| def run_vpp(cls): |
| cmdline = cls.vpp_cmdline |
| |
| if cls.debug_gdbserver: |
| gdbserver = '/usr/bin/gdbserver' |
| if not os.path.isfile(gdbserver) or \ |
| not os.access(gdbserver, os.X_OK): |
| raise Exception("gdbserver binary '%s' does not exist or is " |
| "not executable" % gdbserver) |
| |
| cmdline = [gdbserver, 'localhost:{port}' |
| .format(port=cls.gdbserver_port)] + cls.vpp_cmdline |
| cls.logger.info("Gdbserver cmdline is %s", " ".join(cmdline)) |
| |
| try: |
| cls.vpp = subprocess.Popen(cmdline, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE) |
| except subprocess.CalledProcessError as e: |
| cls.logger.critical("Subprocess returned with non-0 return code: (" |
| "%s)", e.returncode) |
| raise |
| except OSError as e: |
| cls.logger.critical("Subprocess returned with OS error: " |
| "(%s) %s", e.errno, e.strerror) |
| raise |
| except Exception as e: |
| cls.logger.exception("Subprocess returned unexpected from " |
| "%s:", cmdline) |
| raise |
| |
| cls.wait_for_enter() |
| |
| @classmethod |
| def wait_for_coredump(cls): |
| corefile = cls.tempdir + "/core" |
| if os.path.isfile(corefile): |
| cls.logger.error("Waiting for coredump to complete: %s", corefile) |
| curr_size = os.path.getsize(corefile) |
| deadline = time.time() + 60 |
| ok = False |
| while time.time() < deadline: |
| cls.sleep(1) |
| size = curr_size |
| curr_size = os.path.getsize(corefile) |
| if size == curr_size: |
| ok = True |
| break |
| if not ok: |
| cls.logger.error("Timed out waiting for coredump to complete:" |
| " %s", corefile) |
| else: |
| cls.logger.error("Coredump complete: %s, size %d", |
| corefile, curr_size) |
| |
| @classmethod |
| def setUpClass(cls): |
| """ |
| Perform class setup before running the testcase |
| Remove shared memory files, start vpp and connect the vpp-api |
| """ |
| super(VppTestCase, cls).setUpClass() |
| gc.collect() # run garbage collection first |
| cls.logger = get_logger(cls.__name__) |
| seed = os.environ["RND_SEED"] |
| random.seed(seed) |
| if hasattr(cls, 'parallel_handler'): |
| cls.logger.addHandler(cls.parallel_handler) |
| cls.logger.propagate = False |
| |
| cls.tempdir = tempfile.mkdtemp( |
| prefix='vpp-unittest-%s-' % cls.__name__) |
| cls.stats_sock = "%s/stats.sock" % cls.tempdir |
| cls.api_sock = "%s/api.sock" % cls.tempdir |
| cls.file_handler = FileHandler("%s/log.txt" % cls.tempdir) |
| cls.file_handler.setFormatter( |
| Formatter(fmt='%(asctime)s,%(msecs)03d %(message)s', |
| datefmt="%H:%M:%S")) |
| cls.file_handler.setLevel(DEBUG) |
| cls.logger.addHandler(cls.file_handler) |
| cls.logger.debug("--- setUpClass() for %s called ---" % |
| cls.__name__) |
| cls.shm_prefix = os.path.basename(cls.tempdir) |
| os.chdir(cls.tempdir) |
| cls.logger.info("Temporary dir is %s, shm prefix is %s", |
| cls.tempdir, cls.shm_prefix) |
| cls.logger.debug("Random seed is %s" % seed) |
| cls.setUpConstants() |
| cls.reset_packet_infos() |
| cls._captures = [] |
| cls.verbose = 0 |
| cls.vpp_dead = False |
| cls.registry = VppObjectRegistry() |
| cls.vpp_startup_failed = False |
| cls.reporter = KeepAliveReporter() |
| # need to catch exceptions here because if we raise, then the cleanup |
| # doesn't get called and we might end with a zombie vpp |
| try: |
| cls.run_vpp() |
| cls.reporter.send_keep_alive(cls, 'setUpClass') |
| VppTestResult.current_test_case_info = TestCaseInfo( |
| cls.logger, cls.tempdir, cls.vpp.pid, cls.vpp_bin) |
| cls.vpp_stdout_deque = deque() |
| cls.vpp_stderr_deque = deque() |
| cls.pump_thread_stop_flag = Event() |
| cls.pump_thread_wakeup_pipe = os.pipe() |
| cls.pump_thread = Thread(target=pump_output, args=(cls,)) |
| cls.pump_thread.daemon = True |
| cls.pump_thread.start() |
| if cls.debug_gdb or cls.debug_gdbserver: |
| cls.vapi_response_timeout = 0 |
| cls.vapi = VppPapiProvider(cls.shm_prefix, cls.shm_prefix, cls, |
| cls.vapi_response_timeout) |
| if cls.step: |
| hook = hookmodule.StepHook(cls) |
| else: |
| hook = hookmodule.PollHook(cls) |
| cls.vapi.register_hook(hook) |
| cls.statistics = VPPStats(socketname=cls.stats_sock) |
| try: |
| hook.poll_vpp() |
| except VppDiedError: |
| cls.vpp_startup_failed = True |
| cls.logger.critical( |
| "VPP died shortly after startup, check the" |
| " output to standard error for possible cause") |
| raise |
| try: |
| cls.vapi.connect() |
| except vpp_papi.VPPIOError as e: |
| cls.logger.debug("Exception connecting to vapi: %s" % e) |
| cls.vapi.disconnect() |
| |
| if cls.debug_gdbserver: |
| print(colorize("You're running VPP inside gdbserver but " |
| "VPP-API connection failed, did you forget " |
| "to 'continue' VPP from within gdb?", RED)) |
| raise |
| except vpp_papi.VPPRuntimeError as e: |
| cls.logger.debug("%s" % e) |
| cls.quit() |
| raise |
| except Exception as e: |
| cls.logger.debug("Exception connecting to VPP: %s" % e) |
| cls.quit() |
| raise |
| |
| @classmethod |
| def _debug_quit(cls): |
| if (cls.debug_gdbserver or cls.debug_gdb): |
| try: |
| cls.vpp.poll() |
| |
| if cls.vpp.returncode is None: |
| print() |
| print(double_line_delim) |
| print("VPP or GDB server is still running") |
| print(single_line_delim) |
| input("When done debugging, press ENTER to kill the " |
| "process and finish running the testcase...") |
| except AttributeError: |
| pass |
| |
| @classmethod |
| def quit(cls): |
| """ |
| Disconnect vpp-api, kill vpp and cleanup shared memory files |
| """ |
| cls._debug_quit() |
| |
| # first signal that we want to stop the pump thread, then wake it up |
| if hasattr(cls, 'pump_thread_stop_flag'): |
| cls.pump_thread_stop_flag.set() |
| if hasattr(cls, 'pump_thread_wakeup_pipe'): |
| os.write(cls.pump_thread_wakeup_pipe[1], b'ding dong wake up') |
| if hasattr(cls, 'pump_thread'): |
| cls.logger.debug("Waiting for pump thread to stop") |
| cls.pump_thread.join() |
| if hasattr(cls, 'vpp_stderr_reader_thread'): |
| cls.logger.debug("Waiting for stderr pump to stop") |
| cls.vpp_stderr_reader_thread.join() |
| |
| if hasattr(cls, 'vpp'): |
| if hasattr(cls, 'vapi'): |
| cls.logger.debug(cls.vapi.vpp.get_stats()) |
| cls.logger.debug("Disconnecting class vapi client on %s", |
| cls.__name__) |
| cls.vapi.disconnect() |
| cls.logger.debug("Deleting class vapi attribute on %s", |
| cls.__name__) |
| del cls.vapi |
| cls.vpp.poll() |
| if cls.vpp.returncode is None: |
| cls.wait_for_coredump() |
| cls.logger.debug("Sending TERM to vpp") |
| cls.vpp.terminate() |
| cls.logger.debug("Waiting for vpp to die") |
| cls.vpp.communicate() |
| cls.logger.debug("Deleting class vpp attribute on %s", |
| cls.__name__) |
| del cls.vpp |
| |
| if cls.vpp_startup_failed: |
| stdout_log = cls.logger.info |
| stderr_log = cls.logger.critical |
| else: |
| stdout_log = cls.logger.info |
| stderr_log = cls.logger.info |
| |
| if hasattr(cls, 'vpp_stdout_deque'): |
| stdout_log(single_line_delim) |
| stdout_log('VPP output to stdout while running %s:', cls.__name__) |
| stdout_log(single_line_delim) |
| vpp_output = "".join(cls.vpp_stdout_deque) |
| with open(cls.tempdir + '/vpp_stdout.txt', 'w') as f: |
| f.write(vpp_output) |
| stdout_log('\n%s', vpp_output) |
| stdout_log(single_line_delim) |
| |
| if hasattr(cls, 'vpp_stderr_deque'): |
| stderr_log(single_line_delim) |
| stderr_log('VPP output to stderr while running %s:', cls.__name__) |
| stderr_log(single_line_delim) |
| vpp_output = "".join(cls.vpp_stderr_deque) |
| with open(cls.tempdir + '/vpp_stderr.txt', 'w') as f: |
| f.write(vpp_output) |
| stderr_log('\n%s', vpp_output) |
| stderr_log(single_line_delim) |
| |
| @classmethod |
| def tearDownClass(cls): |
| """ Perform final cleanup after running all tests in this test-case """ |
| cls.logger.debug("--- tearDownClass() for %s called ---" % |
| cls.__name__) |
| cls.reporter.send_keep_alive(cls, 'tearDownClass') |
| cls.quit() |
| cls.file_handler.close() |
| cls.reset_packet_infos() |
| if debug_framework: |
| debug_internal.on_tear_down_class(cls) |
| |
| def show_commands_at_teardown(self): |
| """ Allow subclass specific teardown logging additions.""" |
| self.logger.info("--- No test specific show commands provided. ---") |
| |
| def tearDown(self): |
| """ Show various debug prints after each test """ |
| self.logger.debug("--- tearDown() for %s.%s(%s) called ---" % |
| (self.__class__.__name__, self._testMethodName, |
| self._testMethodDoc)) |
| |
| try: |
| if not self.vpp_dead: |
| self.logger.debug(self.vapi.cli("show trace max 1000")) |
| self.logger.info(self.vapi.ppcli("show interface")) |
| self.logger.info(self.vapi.ppcli("show hardware")) |
| self.logger.info(self.statistics.set_errors_str()) |
| self.logger.info(self.vapi.ppcli("show run")) |
| self.logger.info(self.vapi.ppcli("show log")) |
| self.logger.info(self.vapi.ppcli("show bihash")) |
| self.logger.info("Logging testcase specific show commands.") |
| self.show_commands_at_teardown() |
| self.registry.remove_vpp_config(self.logger) |
| # Save/Dump VPP api trace log |
| m = self._testMethodName |
| api_trace = "vpp_api_trace.%s.%d.log" % (m, self.vpp.pid) |
| tmp_api_trace = "/tmp/%s" % api_trace |
| vpp_api_trace_log = "%s/%s" % (self.tempdir, api_trace) |
| self.logger.info(self.vapi.ppcli("api trace save %s" % api_trace)) |
| self.logger.info("Moving %s to %s\n" % (tmp_api_trace, |
| vpp_api_trace_log)) |
| os.rename(tmp_api_trace, vpp_api_trace_log) |
| self.logger.info(self.vapi.ppcli("api trace custom-dump %s" % |
| vpp_api_trace_log)) |
| except VppTransportShmemIOError: |
| self.logger.debug("VppTransportShmemIOError: Vpp dead. " |
| "Cannot log show commands.") |
| self.vpp_dead = True |
| else: |
| self.registry.unregister_all(self.logger) |
| |
| def setUp(self): |
| """ Clear trace before running each test""" |
| super(VppTestCase, self).setUp() |
| self.reporter.send_keep_alive(self) |
| if self.vpp_dead: |
| |
| raise VppDiedError(rv=None, testcase=self.__class__.__name__, |
| method_name=self._testMethodName) |
| self.sleep(.1, "during setUp") |
| self.vpp_stdout_deque.append( |
| "--- test setUp() for %s.%s(%s) starts here ---\n" % |
| (self.__class__.__name__, self._testMethodName, |
| self._testMethodDoc)) |
| self.vpp_stderr_deque.append( |
| "--- test setUp() for %s.%s(%s) starts here ---\n" % |
| (self.__class__.__name__, self._testMethodName, |
| self._testMethodDoc)) |
| self.vapi.cli("clear trace") |
| # store the test instance inside the test class - so that objects |
| # holding the class can access instance methods (like assertEqual) |
| type(self).test_instance = self |
| |
| @classmethod |
| def pg_enable_capture(cls, interfaces=None): |
| """ |
| Enable capture on packet-generator interfaces |
| |
| :param interfaces: iterable interface indexes (if None, |
| use self.pg_interfaces) |
| |
| """ |
| if interfaces is None: |
| interfaces = cls.pg_interfaces |
| for i in interfaces: |
| i.enable_capture() |
| |
| @classmethod |
| def register_capture(cls, cap_name): |
| """ Register a capture in the testclass """ |
| # add to the list of captures with current timestamp |
| cls._captures.append((time.time(), cap_name)) |
| |
| @classmethod |
| def get_vpp_time(cls): |
| # processes e.g. "Time now 2.190522, Wed, 11 Mar 2020 17:29:54 GMT" |
| # returns float("2.190522") |
| timestr = cls.vapi.cli('show clock') |
| head, sep, tail = timestr.partition(',') |
| head, sep, tail = head.partition('Time now') |
| return float(tail) |
| |
| @classmethod |
| def sleep_on_vpp_time(cls, sec): |
| """ Sleep according to time in VPP world """ |
| # On a busy system with many processes |
| # we might end up with VPP time being slower than real world |
| # So take that into account when waiting for VPP to do something |
| start_time = cls.get_vpp_time() |
| while cls.get_vpp_time() - start_time < sec: |
| cls.sleep(0.1) |
| |
| @classmethod |
| def pg_start(cls): |
| """ Enable the PG, wait till it is done, then clean up """ |
| cls.vapi.cli("trace add pg-input 1000") |
| cls.vapi.cli('packet-generator enable') |
| # PG, when starts, runs to completion - |
| # so let's avoid a race condition, |
| # and wait a little till it's done. |
| # Then clean it up - and then be gone. |
| deadline = time.time() + 300 |
| while cls.vapi.cli('show packet-generator').find("Yes") != -1: |
| cls.sleep(0.01) # yield |
| if time.time() > deadline: |
| cls.logger.error("Timeout waiting for pg to stop") |
| break |
| for stamp, cap_name in cls._captures: |
| cls.vapi.cli('packet-generator delete %s' % cap_name) |
| cls._captures = [] |
| |
| @classmethod |
| def create_pg_interfaces(cls, interfaces, gso=0, gso_size=0): |
| """ |
| Create packet-generator interfaces. |
| |
| :param interfaces: iterable indexes of the interfaces. |
| :returns: List of created interfaces. |
| |
| """ |
| result = [] |
| for i in interfaces: |
| intf = VppPGInterface(cls, i, gso, gso_size) |
| setattr(cls, intf.name, intf) |
| result.append(intf) |
| cls.pg_interfaces = result |
| return result |
| |
| @classmethod |
| def create_loopback_interfaces(cls, count): |
| """ |
| Create loopback interfaces. |
| |
| :param count: number of interfaces created. |
| :returns: List of created interfaces. |
| """ |
| result = [VppLoInterface(cls) for i in range(count)] |
| for intf in result: |
| setattr(cls, intf.name, intf) |
| cls.lo_interfaces = result |
| return result |
| |
| @classmethod |
| def create_bvi_interfaces(cls, count): |
| """ |
| Create BVI interfaces. |
| |
| :param count: number of interfaces created. |
| :returns: List of created interfaces. |
| """ |
| result = [VppBviInterface(cls) for i in range(count)] |
| for intf in result: |
| setattr(cls, intf.name, intf) |
| cls.bvi_interfaces = result |
| return result |
| |
| @staticmethod |
| def extend_packet(packet, size, padding=' '): |
| """ |
| Extend packet to given size by padding with spaces or custom padding |
| NOTE: Currently works only when Raw layer is present. |
| |
| :param packet: packet |
| :param size: target size |
| :param padding: padding used to extend the payload |
| |
| """ |
| packet_len = len(packet) + 4 |
| extend = size - packet_len |
| if extend > 0: |
| num = (extend // len(padding)) + 1 |
| packet[Raw].load += (padding * num)[:extend].encode("ascii") |
| |
| @classmethod |
| def reset_packet_infos(cls): |
| """ Reset the list of packet info objects and packet counts to zero """ |
| cls._packet_infos = {} |
| cls._packet_count_for_dst_if_idx = {} |
| |
| @classmethod |
| def create_packet_info(cls, src_if, dst_if): |
| """ |
| Create packet info object containing the source and destination indexes |
| and add it to the testcase's packet info list |
| |
| :param VppInterface src_if: source interface |
| :param VppInterface dst_if: destination interface |
| |
| :returns: _PacketInfo object |
| |
| """ |
| info = _PacketInfo() |
| info.index = len(cls._packet_infos) |
| info.src = src_if.sw_if_index |
| info.dst = dst_if.sw_if_index |
| if isinstance(dst_if, VppSubInterface): |
| dst_idx = dst_if.parent.sw_if_index |
| else: |
| dst_idx = dst_if.sw_if_index |
| if dst_idx in cls._packet_count_for_dst_if_idx: |
| cls._packet_count_for_dst_if_idx[dst_idx] += 1 |
| else: |
| cls._packet_count_for_dst_if_idx[dst_idx] = 1 |
| cls._packet_infos[info.index] = info |
| return info |
| |
| @staticmethod |
| def info_to_payload(info): |
| """ |
| Convert _PacketInfo object to packet payload |
| |
| :param info: _PacketInfo object |
| |
| :returns: string containing serialized data from packet info |
| """ |
| return "%d %d %d %d %d" % (info.index, info.src, info.dst, |
| info.ip, info.proto) |
| |
| @staticmethod |
| def payload_to_info(payload, payload_field='load'): |
| """ |
| Convert packet payload to _PacketInfo object |
| |
| :param payload: packet payload |
| :type payload: <class 'scapy.packet.Raw'> |
| :param payload_field: packet fieldname of payload "load" for |
| <class 'scapy.packet.Raw'> |
| :type payload_field: str |
| :returns: _PacketInfo object containing de-serialized data from payload |
| |
| """ |
| numbers = getattr(payload, payload_field).split() |
| info = _PacketInfo() |
| info.index = int(numbers[0]) |
| info.src = int(numbers[1]) |
| info.dst = int(numbers[2]) |
| info.ip = int(numbers[3]) |
| info.proto = int(numbers[4]) |
| return info |
| |
| def get_next_packet_info(self, info): |
| """ |
| Iterate over the packet info list stored in the testcase |
| Start iteration with first element if info is None |
| Continue based on index in info if info is specified |
| |
| :param info: info or None |
| :returns: next info in list or None if no more infos |
| """ |
| if info is None: |
| next_index = 0 |
| else: |
| next_index = info.index + 1 |
| if next_index == len(self._packet_infos): |
| return None |
| else: |
| return self._packet_infos[next_index] |
| |
| def get_next_packet_info_for_interface(self, src_index, info): |
| """ |
| Search the packet info list for the next packet info with same source |
| interface index |
| |
| :param src_index: source interface index to search for |
| :param info: packet info - where to start the search |
| :returns: packet info or None |
| |
| """ |
| while True: |
| info = self.get_next_packet_info(info) |
| if info is None: |
| return None |
| if info.src == src_index: |
| return info |
| |
| def get_next_packet_info_for_interface2(self, src_index, dst_index, info): |
| """ |
| Search the packet info list for the next packet info with same source |
| and destination interface indexes |
| |
| :param src_index: source interface index to search for |
| :param dst_index: destination interface index to search for |
| :param info: packet info - where to start the search |
| :returns: packet info or None |
| |
| """ |
| while True: |
| info = self.get_next_packet_info_for_interface(src_index, info) |
| if info is None: |
| return None |
| if info.dst == dst_index: |
| return info |
| |
| def assert_equal(self, real_value, expected_value, name_or_class=None): |
| if name_or_class is None: |
| self.assertEqual(real_value, expected_value) |
| return |
| try: |
| msg = "Invalid %s: %d('%s') does not match expected value %d('%s')" |
| msg = msg % (getdoc(name_or_class).strip(), |
| real_value, str(name_or_class(real_value)), |
| expected_value, str(name_or_class(expected_value))) |
| except Exception: |
| msg = "Invalid %s: %s does not match expected value %s" % ( |
| name_or_class, real_value, expected_value) |
| |
| self.assertEqual(real_value, expected_value, msg) |
| |
| def assert_in_range(self, |
| real_value, |
| expected_min, |
| expected_max, |
| name=None): |
| if name is None: |
| msg = None |
| else: |
| msg = "Invalid %s: %s out of range <%s,%s>" % ( |
| name, real_value, expected_min, expected_max) |
| self.assertTrue(expected_min <= real_value <= expected_max, msg) |
| |
| def assert_packet_checksums_valid(self, packet, |
| ignore_zero_udp_checksums=True): |
| received = packet.__class__(scapy.compat.raw(packet)) |
| udp_layers = ['UDP', 'UDPerror'] |
| checksum_fields = ['cksum', 'chksum'] |
| checksums = [] |
| counter = 0 |
| temp = received.__class__(scapy.compat.raw(received)) |
| while True: |
| layer = temp.getlayer(counter) |
| if layer: |
| layer = layer.copy() |
| layer.remove_payload() |
| for cf in checksum_fields: |
| if hasattr(layer, cf): |
| if ignore_zero_udp_checksums and \ |
| 0 == getattr(layer, cf) and \ |
| layer.name in udp_layers: |
| continue |
| delattr(temp.getlayer(counter), cf) |
| checksums.append((counter, cf)) |
| else: |
| break |
| counter = counter + 1 |
| if 0 == len(checksums): |
| return |
| temp = temp.__class__(scapy.compat.raw(temp)) |
| for layer, cf in checksums: |
| calc_sum = getattr(temp[layer], cf) |
| self.assert_equal( |
| getattr(received[layer], cf), calc_sum, |
| "packet checksum on layer #%d: %s" % (layer, temp[layer].name)) |
| self.logger.debug( |
| "Checksum field `%s` on `%s` layer has correct value `%s`" % |
| (cf, temp[layer].name, calc_sum)) |
| |
| def assert_checksum_valid(self, received_packet, layer, |
| field_name='chksum', |
| ignore_zero_checksum=False): |
| """ Check checksum of received packet on given layer """ |
| received_packet_checksum = getattr(received_packet[layer], field_name) |
| if ignore_zero_checksum and 0 == received_packet_checksum: |
| return |
| recalculated = received_packet.__class__( |
| scapy.compat.raw(received_packet)) |
| delattr(recalculated[layer], field_name) |
| recalculated = recalculated.__class__(scapy.compat.raw(recalculated)) |
| self.assert_equal(received_packet_checksum, |
| getattr(recalculated[layer], field_name), |
| "packet checksum on layer: %s" % layer) |
| |
| def assert_ip_checksum_valid(self, received_packet, |
| ignore_zero_checksum=False): |
| self.assert_checksum_valid(received_packet, 'IP', |
| ignore_zero_checksum=ignore_zero_checksum) |
| |
| def assert_tcp_checksum_valid(self, received_packet, |
| ignore_zero_checksum=False): |
| self.assert_checksum_valid(received_packet, 'TCP', |
| ignore_zero_checksum=ignore_zero_checksum) |
| |
| def assert_udp_checksum_valid(self, received_packet, |
| ignore_zero_checksum=True): |
| self.assert_checksum_valid(received_packet, 'UDP', |
| ignore_zero_checksum=ignore_zero_checksum) |
| |
| def assert_embedded_icmp_checksum_valid(self, received_packet): |
| if received_packet.haslayer(IPerror): |
| self.assert_checksum_valid(received_packet, 'IPerror') |
| if received_packet.haslayer(TCPerror): |
| self.assert_checksum_valid(received_packet, 'TCPerror') |
| if received_packet.haslayer(UDPerror): |
| self.assert_checksum_valid(received_packet, 'UDPerror', |
| ignore_zero_checksum=True) |
| if received_packet.haslayer(ICMPerror): |
| self.assert_checksum_valid(received_packet, 'ICMPerror') |
| |
| def assert_icmp_checksum_valid(self, received_packet): |
| self.assert_checksum_valid(received_packet, 'ICMP') |
| self.assert_embedded_icmp_checksum_valid(received_packet) |
| |
| def assert_icmpv6_checksum_valid(self, pkt): |
| if pkt.haslayer(ICMPv6DestUnreach): |
| self.assert_checksum_valid(pkt, 'ICMPv6DestUnreach', 'cksum') |
| self.assert_embedded_icmp_checksum_valid(pkt) |
| if pkt.haslayer(ICMPv6EchoRequest): |
| self.assert_checksum_valid(pkt, 'ICMPv6EchoRequest', 'cksum') |
| if pkt.haslayer(ICMPv6EchoReply): |
| self.assert_checksum_valid(pkt, 'ICMPv6EchoReply', 'cksum') |
| |
| def get_packet_counter(self, counter): |
| if counter.startswith("/"): |
| counter_value = self.statistics.get_counter(counter) |
| else: |
| counters = self.vapi.cli("sh errors").split('\n') |
| counter_value = 0 |
| for i in range(1, len(counters) - 1): |
| results = counters[i].split() |
| if results[1] == counter: |
| counter_value = int(results[0]) |
| break |
| return counter_value |
| |
| def assert_packet_counter_equal(self, counter, expected_value): |
| counter_value = self.get_packet_counter(counter) |
| self.assert_equal(counter_value, expected_value, |
| "packet counter `%s'" % counter) |
| |
| def assert_error_counter_equal(self, counter, expected_value): |
| counter_value = self.statistics.get_err_counter(counter) |
| self.assert_equal(counter_value, expected_value, |
| "error counter `%s'" % counter) |
| |
| @classmethod |
| def sleep(cls, timeout, remark=None): |
| |
| # /* Allow sleep(0) to maintain win32 semantics, and as decreed |
| # * by Guido, only the main thread can be interrupted. |
| # */ |
| # https://github.com/python/cpython/blob/6673decfa0fb078f60587f5cb5e98460eea137c2/Modules/timemodule.c#L1892 # noqa |
| if timeout == 0: |
| # yield quantum |
| if hasattr(os, 'sched_yield'): |
| os.sched_yield() |
| else: |
| time.sleep(0) |
| return |
| |
| if hasattr(cls, 'logger'): |
| cls.logger.debug("Starting sleep for %es (%s)", timeout, remark) |
| before = time.time() |
| time.sleep(timeout) |
| after = time.time() |
| if hasattr(cls, 'logger') and after - before > 2 * timeout: |
| cls.logger.error("unexpected self.sleep() result - " |
| "slept for %es instead of ~%es!", |
| after - before, timeout) |
| if hasattr(cls, 'logger'): |
| cls.logger.debug( |
| "Finished sleep (%s) - slept %es (wanted %es)", |
| remark, after - before, timeout) |
| |
| def pg_send(self, intf, pkts, worker=None): |
| self.vapi.cli("clear trace") |
| intf.add_stream(pkts, worker=worker) |
| self.pg_enable_capture(self.pg_interfaces) |
| self.pg_start() |
| |
| def send_and_assert_no_replies(self, intf, pkts, remark="", timeout=None): |
| self.pg_send(intf, pkts) |
| if not timeout: |
| timeout = 1 |
| for i in self.pg_interfaces: |
| i.get_capture(0, timeout=timeout) |
| i.assert_nothing_captured(remark=remark) |
| timeout = 0.1 |
| |
| def send_and_expect(self, intf, pkts, output, n_rx=None, worker=None): |
| if not n_rx: |
| n_rx = len(pkts) |
| self.pg_send(intf, pkts, worker=worker) |
| rx = output.get_capture(n_rx) |
| return rx |
| |
| def send_and_expect_only(self, intf, pkts, output, timeout=None): |
| self.pg_send(intf, pkts) |
| rx = output.get_capture(len(pkts)) |
| outputs = [output] |
| if not timeout: |
| timeout = 1 |
| for i in self.pg_interfaces: |
| if i not in outputs: |
| i.get_capture(0, timeout=timeout) |
| i.assert_nothing_captured() |
| timeout = 0.1 |
| |
| return rx |
| |
| |
| def get_testcase_doc_name(test): |
| return getdoc(test.__class__).splitlines()[0] |
| |
| |
| def get_test_description(descriptions, test): |
| short_description = test.shortDescription() |
| if descriptions and short_description: |
| return short_description |
| else: |
| return str(test) |
| |
| |
| class TestCaseInfo(object): |
| def __init__(self, logger, tempdir, vpp_pid, vpp_bin_path): |
| self.logger = logger |
| self.tempdir = tempdir |
| self.vpp_pid = vpp_pid |
| self.vpp_bin_path = vpp_bin_path |
| self.core_crash_test = None |
| |
| |
| class VppTestResult(unittest.TestResult): |
| """ |
| @property result_string |
| String variable to store the test case result string. |
| @property errors |
| List variable containing 2-tuples of TestCase instances and strings |
| holding formatted tracebacks. Each tuple represents a test which |
| raised an unexpected exception. |
| @property failures |
| List variable containing 2-tuples of TestCase instances and strings |
| holding formatted tracebacks. Each tuple represents a test where |
| a failure was explicitly signalled using the TestCase.assert*() |
| methods. |
| """ |
| |
| failed_test_cases_info = set() |
| core_crash_test_cases_info = set() |
| current_test_case_info = None |
| |
| def __init__(self, stream=None, descriptions=None, verbosity=None, |
| runner=None): |
| """ |
| :param stream File descriptor to store where to report test results. |
| Set to the standard error stream by default. |
| :param descriptions Boolean variable to store information if to use |
| test case descriptions. |
| :param verbosity Integer variable to store required verbosity level. |
| """ |
| super(VppTestResult, self).__init__(stream, descriptions, verbosity) |
| self.stream = stream |
| self.descriptions = descriptions |
| self.verbosity = verbosity |
| self.result_string = None |
| self.runner = runner |
| |
| def addSuccess(self, test): |
| """ |
| Record a test succeeded result |
| |
| :param test: |
| |
| """ |
| if self.current_test_case_info: |
| self.current_test_case_info.logger.debug( |
| "--- addSuccess() %s.%s(%s) called" % (test.__class__.__name__, |
| test._testMethodName, |
| test._testMethodDoc)) |
| unittest.TestResult.addSuccess(self, test) |
| self.result_string = colorize("OK", GREEN) |
| |
| self.send_result_through_pipe(test, PASS) |
| |
| def addSkip(self, test, reason): |
| """ |
| Record a test skipped. |
| |
| :param test: |
| :param reason: |
| |
| """ |
| if self.current_test_case_info: |
| self.current_test_case_info.logger.debug( |
| "--- addSkip() %s.%s(%s) called, reason is %s" % |
| (test.__class__.__name__, test._testMethodName, |
| test._testMethodDoc, reason)) |
| unittest.TestResult.addSkip(self, test, reason) |
| self.result_string = colorize("SKIP", YELLOW) |
| |
| self.send_result_through_pipe(test, SKIP) |
| |
| def symlink_failed(self): |
| if self.current_test_case_info: |
| try: |
| failed_dir = os.getenv('FAILED_DIR') |
| link_path = os.path.join( |
| failed_dir, |
| '%s-FAILED' % |
| os.path.basename(self.current_test_case_info.tempdir)) |
| if self.current_test_case_info.logger: |
| self.current_test_case_info.logger.debug( |
| "creating a link to the failed test") |
| self.current_test_case_info.logger.debug( |
| "os.symlink(%s, %s)" % |
| (self.current_test_case_info.tempdir, link_path)) |
| if os.path.exists(link_path): |
| if self.current_test_case_info.logger: |
| self.current_test_case_info.logger.debug( |
| 'symlink already exists') |
| else: |
| os.symlink(self.current_test_case_info.tempdir, link_path) |
| |
| except Exception as e: |
| if self.current_test_case_info.logger: |
| self.current_test_case_info.logger.error(e) |
| |
| def send_result_through_pipe(self, test, result): |
| if hasattr(self, 'test_framework_result_pipe'): |
| pipe = self.test_framework_result_pipe |
| if pipe: |
| pipe.send((test.id(), result)) |
| |
| def log_error(self, test, err, fn_name): |
| if self.current_test_case_info: |
| if isinstance(test, unittest.suite._ErrorHolder): |
| test_name = test.description |
| else: |
| test_name = '%s.%s(%s)' % (test.__class__.__name__, |
| test._testMethodName, |
| test._testMethodDoc) |
| self.current_test_case_info.logger.debug( |
| "--- %s() %s called, err is %s" % |
| (fn_name, test_name, err)) |
| self.current_test_case_info.logger.debug( |
| "formatted exception is:\n%s" % |
| "".join(format_exception(*err))) |
| |
| def add_error(self, test, err, unittest_fn, error_type): |
| if error_type == FAIL: |
| self.log_error(test, err, 'addFailure') |
| error_type_str = colorize("FAIL", RED) |
| elif error_type == ERROR: |
| self.log_error(test, err, 'addError') |
| error_type_str = colorize("ERROR", RED) |
| else: |
| raise Exception('Error type %s cannot be used to record an ' |
| 'error or a failure' % error_type) |
| |
| unittest_fn(self, test, err) |
| if self.current_test_case_info: |
| self.result_string = "%s [ temp dir used by test case: %s ]" % \ |
| (error_type_str, |
| self.current_test_case_info.tempdir) |
| self.symlink_failed() |
| self.failed_test_cases_info.add(self.current_test_case_info) |
| if is_core_present(self.current_test_case_info.tempdir): |
| if not self.current_test_case_info.core_crash_test: |
| if isinstance(test, unittest.suite._ErrorHolder): |
| test_name = str(test) |
| else: |
| test_name = "'{!s}' ({!s})".format( |
| get_testcase_doc_name(test), test.id()) |
| self.current_test_case_info.core_crash_test = test_name |
| self.core_crash_test_cases_info.add( |
| self.current_test_case_info) |
| else: |
| self.result_string = '%s [no temp dir]' % error_type_str |
| |
| self.send_result_through_pipe(test, error_type) |
| |
| def addFailure(self, test, err): |
| """ |
| Record a test failed result |
| |
| :param test: |
| :param err: error message |
| |
| """ |
| self.add_error(test, err, unittest.TestResult.addFailure, FAIL) |
| |
| def addError(self, test, err): |
| """ |
| Record a test error result |
| |
| :param test: |
| :param err: error message |
| |
| """ |
| self.add_error(test, err, unittest.TestResult.addError, ERROR) |
| |
| def getDescription(self, test): |
| """ |
| Get test description |
| |
| :param test: |
| :returns: test description |
| |
| """ |
| return get_test_description(self.descriptions, test) |
| |
| def startTest(self, test): |
| """ |
| Start a test |
| |
| :param test: |
| |
| """ |
| |
| def print_header(test): |
| if not hasattr(test.__class__, '_header_printed'): |
| print(double_line_delim) |
| print(colorize(getdoc(test).splitlines()[0], GREEN)) |
| print(double_line_delim) |
| test.__class__._header_printed = True |
| |
| print_header(test) |
| self.start_test = time.time() |
| unittest.TestResult.startTest(self, test) |
| if self.verbosity > 0: |
| self.stream.writeln( |
| "Starting " + self.getDescription(test) + " ...") |
| self.stream.writeln(single_line_delim) |
| |
| def stopTest(self, test): |
| """ |
| Called when the given test has been run |
| |
| :param test: |
| |
| """ |
| unittest.TestResult.stopTest(self, test) |
| |
| if self.verbosity > 0: |
| self.stream.writeln(single_line_delim) |
| self.stream.writeln("%-73s%s" % (self.getDescription(test), |
| self.result_string)) |
| self.stream.writeln(single_line_delim) |
| else: |
| self.stream.writeln("%-68s %4.2f %s" % |
| (self.getDescription(test), |
| time.time() - self.start_test, |
| self.result_string)) |
| |
| self.send_result_through_pipe(test, TEST_RUN) |
| |
| def printErrors(self): |
| """ |
| Print errors from running the test case |
| """ |
| if len(self.errors) > 0 or len(self.failures) > 0: |
| self.stream.writeln() |
| self.printErrorList('ERROR', self.errors) |
| self.printErrorList('FAIL', self.failures) |
| |
| # ^^ that is the last output from unittest before summary |
| if not self.runner.print_summary: |
| devnull = unittest.runner._WritelnDecorator(open(os.devnull, 'w')) |
| self.stream = devnull |
| self.runner.stream = devnull |
| |
| def printErrorList(self, flavour, errors): |
| """ |
| Print error list to the output stream together with error type |
| and test case description. |
| |
| :param flavour: error type |
| :param errors: iterable errors |
| |
| """ |
| for test, err in errors: |
| self.stream.writeln(double_line_delim) |
| self.stream.writeln("%s: %s" % |
| (flavour, self.getDescription(test))) |
| self.stream.writeln(single_line_delim) |
| self.stream.writeln("%s" % err) |
| |
| |
| class VppTestRunner(unittest.TextTestRunner): |
| """ |
| A basic test runner implementation which prints results to standard error. |
| """ |
| |
| @property |
| def resultclass(self): |
| """Class maintaining the results of the tests""" |
| return VppTestResult |
| |
| def __init__(self, keep_alive_pipe=None, descriptions=True, verbosity=1, |
| result_pipe=None, failfast=False, buffer=False, |
| resultclass=None, print_summary=True, **kwargs): |
| # ignore stream setting here, use hard-coded stdout to be in sync |
| # with prints from VppTestCase methods ... |
| super(VppTestRunner, self).__init__(sys.stdout, descriptions, |
| verbosity, failfast, buffer, |
| resultclass, **kwargs) |
| KeepAliveReporter.pipe = keep_alive_pipe |
| |
| self.orig_stream = self.stream |
| self.resultclass.test_framework_result_pipe = result_pipe |
| |
| self.print_summary = print_summary |
| |
| def _makeResult(self): |
| return self.resultclass(self.stream, |
| self.descriptions, |
| self.verbosity, |
| self) |
| |
| def run(self, test): |
| """ |
| Run the tests |
| |
| :param test: |
| |
| """ |
| faulthandler.enable() # emit stack trace to stderr if killed by signal |
| |
| result = super(VppTestRunner, self).run(test) |
| if not self.print_summary: |
| self.stream = self.orig_stream |
| result.stream = self.orig_stream |
| return result |
| |
| |
| class Worker(Thread): |
| def __init__(self, executable_args, logger, env=None, *args, **kwargs): |
| super(Worker, self).__init__(*args, **kwargs) |
| self.logger = logger |
| self.args = executable_args |
| if hasattr(self, 'testcase') and self.testcase.debug_all: |
| if self.testcase.debug_gdbserver: |
| self.args = ['/usr/bin/gdbserver', 'localhost:{port}' |
| .format(port=self.testcase.gdbserver_port)] + args |
| elif self.testcase.debug_gdb and hasattr(self, 'wait_for_gdb'): |
| self.args.append(self.wait_for_gdb) |
| self.app_bin = executable_args[0] |
| self.app_name = os.path.basename(self.app_bin) |
| if hasattr(self, 'role'): |
| self.app_name += ' {role}'.format(role=self.role) |
| self.process = None |
| self.result = None |
| env = {} if env is None else env |
| self.env = copy.deepcopy(env) |
| |
| def wait_for_enter(self): |
| if not hasattr(self, 'testcase'): |
| return |
| if self.testcase.debug_all and self.testcase.debug_gdbserver: |
| print() |
| print(double_line_delim) |
| print("Spawned GDB Server for '{app}' with PID: {pid}" |
| .format(app=self.app_name, pid=self.process.pid)) |
| elif self.testcase.debug_all and self.testcase.debug_gdb: |
| print() |
| print(double_line_delim) |
| print("Spawned '{app}' with PID: {pid}" |
| .format(app=self.app_name, pid=self.process.pid)) |
| else: |
| return |
| print(single_line_delim) |
| print("You can debug '{app}' using:".format(app=self.app_name)) |
| if self.testcase.debug_gdbserver: |
| print("sudo gdb " + self.app_bin + |
| " -ex 'target remote localhost:{port}'" |
| .format(port=self.testcase.gdbserver_port)) |
| print("Now is the time to attach gdb by running the above " |
| "command, set up breakpoints etc., then resume from " |
| "within gdb by issuing the 'continue' command") |
| self.testcase.gdbserver_port += 1 |
| elif self.testcase.debug_gdb: |
| print("sudo gdb " + self.app_bin + |
| " -ex 'attach {pid}'".format(pid=self.process.pid)) |
| print("Now is the time to attach gdb by running the above " |
| "command and set up breakpoints etc., then resume from" |
| " within gdb by issuing the 'continue' command") |
| print(single_line_delim) |
| input("Press ENTER to continue running the testcase...") |
| |
| def run(self): |
| executable = self.args[0] |
| if not os.path.exists(executable) or not os.access( |
| executable, os.F_OK | os.X_OK): |
| # Exit code that means some system file did not exist, |
| # could not be opened, or had some other kind of error. |
| self.result = os.EX_OSFILE |
| raise EnvironmentError( |
| "executable '%s' is not found or executable." % executable) |
| self.logger.debug("Running executable: '{app}'" |
| .format(app=' '.join(self.args))) |
| env = os.environ.copy() |
| env.update(self.env) |
| env["CK_LOG_FILE_NAME"] = "-" |
| self.process = subprocess.Popen( |
| self.args, shell=False, env=env, preexec_fn=os.setpgrp, |
| stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| self.wait_for_enter() |
| out, err = self.process.communicate() |
| self.logger.debug("Finished running `{app}'".format(app=self.app_name)) |
| self.logger.info("Return code is `%s'" % self.process.returncode) |
| self.logger.info(single_line_delim) |
| self.logger.info("Executable `{app}' wrote to stdout:" |
| .format(app=self.app_name)) |
| self.logger.info(single_line_delim) |
| self.logger.info(out.decode('utf-8')) |
| self.logger.info(single_line_delim) |
| self.logger.info("Executable `{app}' wrote to stderr:" |
| .format(app=self.app_name)) |
| self.logger.info(single_line_delim) |
| self.logger.info(err.decode('utf-8')) |
| self.logger.info(single_line_delim) |
| self.result = self.process.returncode |
| |
| |
| if __name__ == '__main__': |
| pass |