| """ |
| A simple testing framework for lldb using python's unit testing framework. |
| |
| Tests for lldb are written as python scripts which take advantage of the script |
| bridging provided by LLDB.framework to interact with lldb core. |
| |
| A specific naming pattern is followed by the .py script to be recognized as |
| a module which implements a test scenario, namely, Test*.py. |
| |
| To specify the directories where "Test*.py" python test scripts are located, |
| you need to pass in a list of directory names. By default, the current |
| working directory is searched if nothing is specified on the command line. |
| |
| Type: |
| |
| ./dotest.py -h |
| |
| for available options. |
| """ |
| |
| from __future__ import absolute_import |
| from __future__ import print_function |
| |
| # System modules |
| import sys |
| import atexit |
| import importlib |
| import os |
| import errno |
| import platform |
| import progress |
| import signal |
| import socket |
| import subprocess |
| import sys |
| import inspect |
| |
| # Third-party modules |
| import six |
| import unittest2 |
| |
| # LLDB Modules |
| import lldbsuite |
| from . import dotest_args |
| from . import lldbtest_config |
| from . import test_categories |
| from . import test_results |
| from .test_results import EventBuilder |
| from ..support import seven |
| |
| def is_exe(fpath): |
| """Returns true if fpath is an executable.""" |
| return os.path.isfile(fpath) and os.access(fpath, os.X_OK) |
| |
| def which(program): |
| """Returns the full path to a program; None otherwise.""" |
| fpath, fname = os.path.split(program) |
| if fpath: |
| if is_exe(program): |
| return program |
| else: |
| for path in os.environ["PATH"].split(os.pathsep): |
| exe_file = os.path.join(path, program) |
| if is_exe(exe_file): |
| return exe_file |
| return None |
| |
| class _WritelnDecorator(object): |
| """Used to decorate file-like objects with a handy 'writeln' method""" |
| def __init__(self,stream): |
| self.stream = stream |
| |
| def __getattr__(self, attr): |
| if attr in ('stream', '__getstate__'): |
| raise AttributeError(attr) |
| return getattr(self.stream,attr) |
| |
| def writeln(self, arg=None): |
| if arg: |
| self.write(arg) |
| self.write('\n') # text-mode streams translate to \r\n if needed |
| |
| # |
| # Global variables: |
| # |
| |
| # The test suite. |
| suite = unittest2.TestSuite() |
| |
| # By default, benchmarks tests are not run. |
| just_do_benchmarks_test = False |
| |
| dont_do_dsym_test = False |
| dont_do_dwarf_test = False |
| dont_do_dwo_test = False |
| |
| # The blacklist is optional (-b blacklistFile) and allows a central place to skip |
| # testclass's and/or testclass.testmethod's. |
| blacklist = None |
| |
| # The dictionary as a result of sourcing blacklistFile. |
| blacklistConfig = {} |
| |
| # The list of categories we said we care about |
| categoriesList = None |
| # set to true if we are going to use categories for cherry-picking test cases |
| useCategories = False |
| # Categories we want to skip |
| skipCategories = [] |
| # use this to track per-category failures |
| failuresPerCategory = {} |
| |
| # The path to LLDB.framework is optional. |
| lldbFrameworkPath = None |
| |
| # The config file is optional. |
| configFile = None |
| |
| # Test suite repeat count. Can be overwritten with '-# count'. |
| count = 1 |
| |
| # The dictionary as a result of sourcing configFile. |
| config = {} |
| # The pre_flight and post_flight functions come from reading a config file. |
| pre_flight = None |
| post_flight = None |
| # So do the lldbtest_remote_sandbox and lldbtest_remote_shell_template variables. |
| lldbtest_remote_sandbox = None |
| lldbtest_remote_shell_template = None |
| |
| # The 'archs' and 'compilers' can be specified via either command line or configFile, |
| # with the command line overriding the configFile. The corresponding options can be |
| # specified more than once. For example, "-A x86_64 -A i386" => archs=['x86_64', 'i386'] |
| # and "-C gcc -C clang" => compilers=['gcc', 'clang']. |
| archs = None # Must be initialized after option parsing |
| compilers = None # Must be initialized after option parsing |
| |
| # The arch might dictate some specific CFLAGS to be passed to the toolchain to build |
| # the inferior programs. The global variable cflags_extras provides a hook to do |
| # just that. |
| cflags_extras = '' |
| |
| # Dump the Python sys.path variable. Use '-D' to dump sys.path. |
| dumpSysPath = False |
| |
| # Full path of the benchmark executable, as specified by the '-e' option. |
| bmExecutable = None |
| # The breakpoint specification of bmExecutable, as specified by the '-x' option. |
| bmBreakpointSpec = None |
| # The benchmark iteration count, as specified by the '-y' option. |
| bmIterationCount = -1 |
| |
| # By default, don't exclude any directories. Use '-X' to add one excluded directory. |
| excluded = set(['.svn', '.git']) |
| |
| # By default, failfast is False. Use '-F' to overwrite it. |
| failfast = False |
| |
| # The filters (testclass.testmethod) used to admit tests into our test suite. |
| filters = [] |
| |
| # The runhooks is a list of lldb commands specifically for the debugger. |
| # Use '-k' to specify a runhook. |
| runHooks = [] |
| |
| # If '-g' is specified, the filterspec is not exclusive. If a test module does |
| # not contain testclass.testmethod which matches the filterspec, the whole test |
| # module is still admitted into our test suite. fs4all flag defaults to True. |
| fs4all = True |
| |
| # Ignore the build search path relative to this script to locate the lldb.py module. |
| ignore = False |
| |
| # By default, we do not skip build and cleanup. Use '-S' option to override. |
| skip_build_and_cleanup = False |
| |
| # By default, we skip long running test case. Use '-l' option to override. |
| skip_long_running_test = True |
| |
| # By default, we print the build dir, lldb version, and svn info. Use '-n' option to |
| # turn it off. |
| noHeaders = False |
| |
| # Parsable mode silences headers, and any other output this script might generate, and instead |
| # prints machine-readable output similar to what clang tests produce. |
| parsable = False |
| |
| # The regular expression pattern to match against eligible filenames as our test cases. |
| regexp = None |
| |
| # By default, tests are executed in place and cleanups are performed afterwards. |
| # Use '-r dir' option to relocate the tests and their intermediate files to a |
| # different directory and to forgo any cleanups. The directory specified must |
| # not exist yet. |
| rdir = None |
| |
| # By default, recorded session info for errored/failed test are dumped into its |
| # own file under a session directory named after the timestamp of the test suite |
| # run. Use '-s session-dir-name' to specify a specific dir name. |
| sdir_name = None |
| |
| # Set this flag if there is any session info dumped during the test run. |
| sdir_has_content = False |
| |
| # svn_info stores the output from 'svn info lldb.base.dir'. |
| svn_info = '' |
| |
| # svn_silent means do not try to obtain svn status |
| svn_silent = True |
| |
| # Default verbosity is 0. |
| verbose = 1 |
| |
| # Set to True only if verbose is 0 and LLDB trace mode is off. |
| progress_bar = False |
| |
| # By default, search from the script directory. |
| # We can't use sys.path[0] to determine the script directory |
| # because it doesn't work under a debugger |
| testdirs = [ os.path.dirname(os.path.realpath(__file__)) ] |
| |
| # Separator string. |
| separator = '-' * 70 |
| |
| failed = False |
| |
| # LLDB Remote platform setting |
| lldb_platform_name = None |
| lldb_platform_url = None |
| lldb_platform_working_dir = None |
| |
| # Parallel execution settings |
| is_inferior_test_runner = False |
| multiprocess_test_subdir = None |
| num_threads = None |
| output_on_success = False |
| no_multiprocess_test_runner = False |
| test_runner_name = None |
| |
| # Test results handling globals |
| results_filename = None |
| results_port = None |
| results_file_object = None |
| results_formatter_name = None |
| results_formatter_object = None |
| results_formatter_options = None |
| |
| # The names of all tests. Used to assert we don't have two tests with the same base name. |
| all_tests = set() |
| |
| def usage(parser): |
| parser.print_help() |
| if verbose > 0: |
| print(""" |
| Examples: |
| |
| This is an example of using the -f option to pinpoint to a specific test class |
| and test method to be run: |
| |
| $ ./dotest.py -f ClassTypesTestCase.test_with_dsym_and_run_command |
| ---------------------------------------------------------------------- |
| Collected 1 test |
| |
| test_with_dsym_and_run_command (TestClassTypes.ClassTypesTestCase) |
| Test 'frame variable this' when stopped on a class constructor. ... ok |
| |
| ---------------------------------------------------------------------- |
| Ran 1 test in 1.396s |
| |
| OK |
| |
| And this is an example of using the -p option to run a single file (the filename |
| matches the pattern 'ObjC' and it happens to be 'TestObjCMethods.py'): |
| |
| $ ./dotest.py -v -p ObjC |
| ---------------------------------------------------------------------- |
| Collected 4 tests |
| |
| test_break_with_dsym (TestObjCMethods.FoundationTestCase) |
| Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok |
| test_break_with_dwarf (TestObjCMethods.FoundationTestCase) |
| Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok |
| test_data_type_and_expr_with_dsym (TestObjCMethods.FoundationTestCase) |
| Lookup objective-c data types and evaluate expressions. ... ok |
| test_data_type_and_expr_with_dwarf (TestObjCMethods.FoundationTestCase) |
| Lookup objective-c data types and evaluate expressions. ... ok |
| |
| ---------------------------------------------------------------------- |
| Ran 4 tests in 16.661s |
| |
| OK |
| |
| Running of this script also sets up the LLDB_TEST environment variable so that |
| individual test cases can locate their supporting files correctly. The script |
| tries to set up Python's search paths for modules by looking at the build tree |
| relative to this script. See also the '-i' option in the following example. |
| |
| Finally, this is an example of using the lldb.py module distributed/installed by |
| Xcode4 to run against the tests under the 'forward' directory, and with the '-w' |
| option to add some delay between two tests. It uses ARCH=x86_64 to specify that |
| as the architecture and CC=clang to specify the compiler used for the test run: |
| |
| $ PYTHONPATH=/Xcode4/Library/PrivateFrameworks/LLDB.framework/Versions/A/Resources/Python ARCH=x86_64 CC=clang ./dotest.py -v -w -i forward |
| |
| Session logs for test failures/errors will go into directory '2010-11-11-13_56_16' |
| ---------------------------------------------------------------------- |
| Collected 2 tests |
| |
| test_with_dsym_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase) |
| Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok |
| test_with_dwarf_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase) |
| Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok |
| |
| ---------------------------------------------------------------------- |
| Ran 2 tests in 5.659s |
| |
| OK |
| |
| The 'Session ...' verbiage is recently introduced (see also the '-s' option) to |
| notify the directory containing the session logs for test failures or errors. |
| In case there is any test failure/error, a similar message is appended at the |
| end of the stderr output for your convenience. |
| |
| ENABLING LOGS FROM TESTS |
| |
| Option 1: |
| |
| Writing logs into different files per test case:: |
| |
| This option is particularly useful when multiple dotest instances are created |
| by dosep.py |
| |
| $ ./dotest.py --channel "lldb all" |
| |
| $ ./dotest.py --channel "lldb all" --channel "gdb-remote packets" |
| |
| These log files are written to: |
| |
| <session-dir>/<test-id>-host.log (logs from lldb host process) |
| <session-dir>/<test-id>-server.log (logs from debugserver/lldb-server) |
| <session-dir>/<test-id>-<test-result>.log (console logs) |
| |
| By default, logs from successful runs are deleted. Use the --log-success flag |
| to create reference logs for debugging. |
| |
| $ ./dotest.py --log-success |
| |
| Option 2: (DEPRECATED) |
| |
| The following options can only enable logs from the host lldb process. |
| Only categories from the "lldb" or "gdb-remote" channels can be enabled |
| They also do not automatically enable logs in locally running debug servers. |
| Also, logs from all test case are written into each log file |
| |
| o LLDB_LOG: if defined, specifies the log file pathname for the 'lldb' subsystem |
| with a default option of 'event process' if LLDB_LOG_OPTION is not defined. |
| |
| o GDB_REMOTE_LOG: if defined, specifies the log file pathname for the |
| 'process.gdb-remote' subsystem with a default option of 'packets' if |
| GDB_REMOTE_LOG_OPTION is not defined. |
| |
| """) |
| sys.exit(0) |
| |
| |
| def setCrashInfoHook_Mac(text): |
| from . import crashinfo |
| crashinfo.setCrashReporterDescription(text) |
| |
| # implement this in some suitable way for your platform, and then bind it |
| # to setCrashInfoHook |
| def setCrashInfoHook_NonMac(text): |
| pass |
| |
| setCrashInfoHook = None |
| |
| def deleteCrashInfoDylib(dylib_path): |
| try: |
| # Need to modify this to handle multiple tests running at the same time. If we move this |
| # to the test's real dir, all should be we run sequentially within a test directory. |
| # os.remove(dylib_path) |
| None |
| finally: |
| pass |
| |
| def setupCrashInfoHook(): |
| global setCrashInfoHook |
| setCrashInfoHook = setCrashInfoHook_NonMac # safe default |
| if platform.system() == "Darwin": |
| from . import lock |
| test_dir = os.environ['LLDB_TEST'] |
| if not test_dir or not os.path.exists(test_dir): |
| return |
| dylib_lock = os.path.join(test_dir,"crashinfo.lock") |
| dylib_src = os.path.join(test_dir,"crashinfo.c") |
| dylib_dst = os.path.join(test_dir,"crashinfo.so") |
| try: |
| compile_lock = lock.Lock(dylib_lock) |
| compile_lock.acquire() |
| if not os.path.isfile(dylib_dst) or os.path.getmtime(dylib_dst) < os.path.getmtime(dylib_src): |
| # we need to compile |
| cmd = "SDKROOT= xcrun clang %s -o %s -framework Python -Xlinker -dylib -iframework /System/Library/Frameworks/ -Xlinker -F /System/Library/Frameworks/" % (dylib_src,dylib_dst) |
| if subprocess.call(cmd,shell=True) != 0 or not os.path.isfile(dylib_dst): |
| raise Exception('command failed: "{}"'.format(cmd)) |
| finally: |
| compile_lock.release() |
| del compile_lock |
| |
| setCrashInfoHook = setCrashInfoHook_Mac |
| |
| else: |
| pass |
| |
| def shouldSkipBecauseOfCategories(test_categories): |
| global useCategories, categoriesList, skipCategories |
| |
| if useCategories: |
| if len(test_categories) == 0 or len(categoriesList & set(test_categories)) == 0: |
| return True |
| |
| for category in skipCategories: |
| if category in test_categories: |
| return True |
| |
| return False |
| |
| def parseOptionsAndInitTestdirs(): |
| """Initialize the list of directories containing our unittest scripts. |
| |
| '-h/--help as the first option prints out usage info and exit the program. |
| """ |
| |
| global just_do_benchmarks_test |
| global dont_do_dsym_test |
| global dont_do_dwarf_test |
| global dont_do_dwo_test |
| global blacklist |
| global blacklistConfig |
| global categoriesList |
| global validCategories |
| global useCategories |
| global skipCategories |
| global lldbFrameworkPath |
| global configFile |
| global archs |
| global compilers |
| global count |
| global dumpSysPath |
| global bmExecutable |
| global bmBreakpointSpec |
| global bmIterationCount |
| global failfast |
| global filters |
| global fs4all |
| global ignore |
| global progress_bar |
| global runHooks |
| global skip_build_and_cleanup |
| global skip_long_running_test |
| global noHeaders |
| global parsable |
| global regexp |
| global rdir |
| global sdir_name |
| global svn_silent |
| global verbose |
| global testdirs |
| global lldb_platform_name |
| global lldb_platform_url |
| global lldb_platform_working_dir |
| global setCrashInfoHook |
| global is_inferior_test_runner |
| global multiprocess_test_subdir |
| global num_threads |
| global output_on_success |
| global no_multiprocess_test_runner |
| global test_runner_name |
| global results_filename |
| global results_formatter_name |
| global results_formatter_options |
| global results_port |
| |
| do_help = False |
| |
| platform_system = platform.system() |
| platform_machine = platform.machine() |
| |
| parser = dotest_args.create_parser() |
| args = dotest_args.parse_args(parser, sys.argv[1:]) |
| |
| if args.unset_env_varnames: |
| for env_var in args.unset_env_varnames: |
| if env_var in os.environ: |
| # From Python Doc: When unsetenv() is supported, deletion of items in os.environ |
| # is automatically translated into a corresponding call to unsetenv(). |
| del os.environ[env_var] |
| #os.unsetenv(env_var) |
| |
| if args.set_env_vars: |
| for env_var in args.set_env_vars: |
| parts = env_var.split('=', 1) |
| if len(parts) == 1: |
| os.environ[parts[0]] = "" |
| else: |
| os.environ[parts[0]] = parts[1] |
| |
| # only print the args if being verbose (and parsable is off) |
| if args.v and not args.q: |
| print(sys.argv) |
| |
| if args.h: |
| do_help = True |
| |
| if args.compilers: |
| compilers = args.compilers |
| else: |
| # Use a compiler appropriate appropriate for the Apple SDK if one was specified |
| if platform_system == 'Darwin' and args.apple_sdk: |
| compilers = [seven.get_command_output('xcrun -sdk "%s" -find clang 2> /dev/null' % (args.apple_sdk))] |
| else: |
| # 'clang' on ubuntu 14.04 is 3.4 so we try clang-3.5 first |
| candidateCompilers = ['clang-3.5', 'clang', 'gcc'] |
| for candidate in candidateCompilers: |
| if which(candidate): |
| compilers = [candidate] |
| break |
| |
| if args.channels: |
| lldbtest_config.channels = args.channels |
| |
| if args.log_success: |
| lldbtest_config.log_success = args.log_success |
| |
| # Set SDKROOT if we are using an Apple SDK |
| if platform_system == 'Darwin' and args.apple_sdk: |
| os.environ['SDKROOT'] = seven.get_command_output('xcrun --sdk "%s" --show-sdk-path 2> /dev/null' % (args.apple_sdk)) |
| |
| if args.archs: |
| archs = args.archs |
| for arch in archs: |
| if arch.startswith('arm') and platform_system == 'Darwin' and not args.apple_sdk: |
| os.environ['SDKROOT'] = seven.get_command_output('xcrun --sdk iphoneos.internal --show-sdk-path 2> /dev/null') |
| if not os.path.exists(os.environ['SDKROOT']): |
| os.environ['SDKROOT'] = seven.get_command_output('xcrun --sdk iphoneos --show-sdk-path 2> /dev/null') |
| else: |
| archs = [platform_machine] |
| |
| if args.categoriesList: |
| categoriesList = set(test_categories.validate(args.categoriesList, False)) |
| useCategories = True |
| else: |
| categoriesList = [] |
| |
| if args.skipCategories: |
| skipCategories = test_categories.validate(args.skipCategories, False) |
| |
| if args.D: |
| dumpSysPath = True |
| |
| if args.E: |
| cflags_extras = args.E |
| os.environ['CFLAGS_EXTRAS'] = cflags_extras |
| |
| # argparse makes sure we have correct options |
| if args.N == 'dwarf': |
| dont_do_dwarf_test = True |
| elif args.N == 'dwo': |
| dont_do_dwo_test = True |
| elif args.N == 'dsym': |
| dont_do_dsym_test = True |
| |
| if args.a or args.plus_a: |
| print("Options '-a' and '+a' have been deprecated. Please use the test category\n" |
| "functionality (-G pyapi, --skip-category pyapi) instead.") |
| sys.exit(1) |
| |
| if args.m or args.plus_m: |
| print("Options '-m' and '+m' have been deprecated. Please use the test category\n" |
| "functionality (-G lldb-mi, --skip-category lldb-mi) instead.") |
| sys.exit(1) |
| |
| if args.plus_b: |
| just_do_benchmarks_test = True |
| |
| if args.b: |
| if args.b.startswith('-'): |
| usage(parser) |
| blacklistFile = args.b |
| if not os.path.isfile(blacklistFile): |
| print('Blacklist file:', blacklistFile, 'does not exist!') |
| usage(parser) |
| # Now read the blacklist contents and assign it to blacklist. |
| execfile(blacklistFile, globals(), blacklistConfig) |
| blacklist = blacklistConfig.get('blacklist') |
| |
| if args.c: |
| if args.c.startswith('-'): |
| usage(parser) |
| configFile = args.c |
| if not os.path.isfile(configFile): |
| print('Config file:', configFile, 'does not exist!') |
| usage(parser) |
| |
| if args.d: |
| sys.stdout.write("Suspending the process %d to wait for debugger to attach...\n" % os.getpid()) |
| sys.stdout.flush() |
| os.kill(os.getpid(), signal.SIGSTOP) |
| |
| if args.e: |
| if args.e.startswith('-'): |
| usage(parser) |
| bmExecutable = args.e |
| if not is_exe(bmExecutable): |
| usage(parser) |
| |
| if args.F: |
| failfast = True |
| |
| if args.f: |
| if any([x.startswith('-') for x in args.f]): |
| usage(parser) |
| filters.extend(args.f) |
| # Shut off multiprocessing mode when additional filters are specified. |
| # The rational is that the user is probably going after a very specific |
| # test and doesn't need a bunch of parallel test runners all looking for |
| # it in a frenzy. Also, '-v' now spits out all test run output even |
| # on success, so the standard recipe for redoing a failing test (with -v |
| # and a -f to filter to the specific test) now causes all test scanning |
| # (in parallel) to print results for do-nothing runs in a very distracting |
| # manner. If we really need filtered parallel runs in the future, consider |
| # adding a --no-output-on-success that prevents -v from setting |
| # output-on-success. |
| no_multiprocess_test_runner = True |
| |
| if args.g: |
| fs4all = False |
| |
| if args.i: |
| ignore = True |
| |
| if args.k: |
| runHooks.extend(args.k) |
| |
| if args.l: |
| skip_long_running_test = False |
| |
| if args.framework: |
| lldbFrameworkPath = args.framework |
| |
| if args.executable: |
| lldbtest_config.lldbExec = args.executable |
| |
| if args.libcxx: |
| os.environ["LIBCXX_PATH"] = args.libcxx |
| |
| if args.n: |
| noHeaders = True |
| |
| if args.p: |
| if args.p.startswith('-'): |
| usage(parser) |
| regexp = args.p |
| |
| if args.q: |
| noHeaders = True |
| parsable = True |
| |
| if args.P and not args.v: |
| progress_bar = True |
| verbose = 0 |
| |
| if args.R: |
| if args.R.startswith('-'): |
| usage(parser) |
| rdir = os.path.abspath(args.R) |
| if os.path.exists(rdir): |
| import shutil |
| print('Removing tree:', rdir) |
| shutil.rmtree(rdir) |
| |
| if args.r: |
| if args.r.startswith('-'): |
| usage(parser) |
| rdir = os.path.abspath(args.r) |
| if os.path.exists(rdir): |
| print('Relocated directory:', rdir, 'must not exist!') |
| usage(parser) |
| |
| if args.S: |
| skip_build_and_cleanup = True |
| |
| if args.s: |
| if args.s.startswith('-'): |
| usage(parser) |
| sdir_name = args.s |
| |
| if args.t: |
| os.environ['LLDB_COMMAND_TRACE'] = 'YES' |
| |
| if args.T: |
| svn_silent = False |
| |
| if args.v: |
| verbose = 2 |
| |
| if args.w: |
| os.environ['LLDB_WAIT_BETWEEN_TEST_CASES'] = 'YES' |
| |
| if args.X: |
| if args.X.startswith('-'): |
| usage(parser) |
| excluded.add(args.X) |
| |
| if args.x: |
| if args.x.startswith('-'): |
| usage(parser) |
| bmBreakpointSpec = args.x |
| |
| # argparse makes sure we have a number |
| if args.y: |
| bmIterationCount = args.y |
| |
| # argparse makes sure we have a number |
| if args.sharp: |
| count = args.sharp |
| |
| if sys.platform.startswith('win32'): |
| os.environ['LLDB_DISABLE_CRASH_DIALOG'] = str(args.disable_crash_dialog) |
| os.environ['LLDB_LAUNCH_INFERIORS_WITHOUT_CONSOLE'] = str(args.hide_inferior_console) |
| |
| if do_help == True: |
| usage(parser) |
| |
| if args.no_multiprocess: |
| no_multiprocess_test_runner = True |
| |
| if args.inferior: |
| is_inferior_test_runner = True |
| |
| # Turn on output_on_sucess if either explicitly added or -v specified. |
| if args.output_on_success or args.v: |
| output_on_success = True |
| |
| if args.num_threads: |
| num_threads = args.num_threads |
| |
| if args.test_subdir: |
| multiprocess_test_subdir = args.test_subdir |
| |
| if args.test_runner_name: |
| test_runner_name = args.test_runner_name |
| |
| # Capture test results-related args. |
| if args.results_file: |
| results_filename = args.results_file |
| |
| if args.results_port: |
| results_port = args.results_port |
| |
| if args.results_file and args.results_port: |
| sys.stderr.write( |
| "only one of --results-file and --results-port should " |
| "be specified\n") |
| usage(args) |
| |
| if args.results_formatter: |
| results_formatter_name = args.results_formatter |
| if args.results_formatter_options: |
| results_formatter_options = args.results_formatter_options |
| |
| if args.lldb_platform_name: |
| lldb_platform_name = args.lldb_platform_name |
| if args.lldb_platform_url: |
| lldb_platform_url = args.lldb_platform_url |
| if args.lldb_platform_working_dir: |
| lldb_platform_working_dir = args.lldb_platform_working_dir |
| |
| if args.event_add_entries and len(args.event_add_entries) > 0: |
| entries = {} |
| # Parse out key=val pairs, separated by comma |
| for keyval in args.event_add_entries.split(","): |
| key_val_entry = keyval.split("=") |
| if len(key_val_entry) == 2: |
| (key, val) = key_val_entry |
| val_parts = val.split(':') |
| if len(val_parts) > 1: |
| (val, val_type) = val_parts |
| if val_type == 'int': |
| val = int(val) |
| entries[key] = val |
| # Tell the event builder to create all events with these |
| # key/val pairs in them. |
| if len(entries) > 0: |
| test_results.EventBuilder.add_entries_to_all_events(entries) |
| |
| # Gather all the dirs passed on the command line. |
| if len(args.args) > 0: |
| testdirs = list(map(os.path.abspath, args.args)) |
| # Shut off multiprocessing mode when test directories are specified. |
| no_multiprocess_test_runner = True |
| |
| # If '-r dir' is specified, the tests should be run under the relocated |
| # directory. Let's copy the testdirs over. |
| if rdir: |
| from shutil import copytree, ignore_patterns |
| |
| tmpdirs = [] |
| orig_testdirs = testdirs[:] |
| for srcdir in testdirs: |
| # For example, /Volumes/data/lldb/svn/ToT/test/functionalities/watchpoint/hello_watchpoint |
| # shall be split into ['/Volumes/data/lldb/svn/ToT/', 'functionalities/watchpoint/hello_watchpoint']. |
| # Utilize the relative path to the 'test' directory to make our destination dir path. |
| if ("test" + os.sep) in srcdir: |
| to_split_on = "test" + os.sep |
| else: |
| to_split_on = "test" |
| dstdir = os.path.join(rdir, srcdir.split(to_split_on)[1]) |
| dstdir = dstdir.rstrip(os.sep) |
| # Don't copy the *.pyc and .svn stuffs. |
| copytree(srcdir, dstdir, ignore=ignore_patterns('*.pyc', '.svn')) |
| tmpdirs.append(dstdir) |
| |
| # This will be our modified testdirs. |
| testdirs = tmpdirs |
| |
| # With '-r dir' specified, there's no cleanup of intermediate test files. |
| os.environ["LLDB_DO_CLEANUP"] = 'NO' |
| |
| # If the original testdirs is ['test'], the make directory has already been copied |
| # recursively and is contained within the rdir/test dir. For anything |
| # else, we would need to copy over the make directory and its contents, |
| # so that, os.listdir(rdir) looks like, for example: |
| # |
| # array_types conditional_break make |
| # |
| # where the make directory contains the Makefile.rules file. |
| if len(testdirs) != 1 or os.path.basename(orig_testdirs[0]) != 'test': |
| scriptdir = os.path.dirname(__file__) |
| # Don't copy the .svn stuffs. |
| copytree(os.path.join(scriptdir, 'make'), os.path.join(rdir, 'make'), |
| ignore=ignore_patterns('.svn')) |
| |
| #print("testdirs:", testdirs) |
| |
| # Source the configFile if specified. |
| # The side effect, if any, will be felt from this point on. An example |
| # config file may be these simple two lines: |
| # |
| # sys.stderr = open("/tmp/lldbtest-stderr", "w") |
| # sys.stdout = open("/tmp/lldbtest-stdout", "w") |
| # |
| # which will reassign the two file objects to sys.stderr and sys.stdout, |
| # respectively. |
| # |
| # See also lldb-trunk/examples/test/usage-config. |
| global config, pre_flight, post_flight, lldbtest_remote_sandbox, lldbtest_remote_shell_template |
| if configFile: |
| # Pass config (a dictionary) as the locals namespace for side-effect. |
| execfile(configFile, globals(), config) |
| #print("config:", config) |
| if "pre_flight" in config: |
| pre_flight = config["pre_flight"] |
| if not six.callable(pre_flight): |
| print("fatal error: pre_flight is not callable, exiting.") |
| sys.exit(1) |
| if "post_flight" in config: |
| post_flight = config["post_flight"] |
| if not six.callable(post_flight): |
| print("fatal error: post_flight is not callable, exiting.") |
| sys.exit(1) |
| if "lldbtest_remote_sandbox" in config: |
| lldbtest_remote_sandbox = config["lldbtest_remote_sandbox"] |
| if "lldbtest_remote_shell_template" in config: |
| lldbtest_remote_shell_template = config["lldbtest_remote_shell_template"] |
| #print("sys.stderr:", sys.stderr) |
| #print("sys.stdout:", sys.stdout) |
| |
| def getXcodeOutputPaths(lldbRootDirectory): |
| result = [] |
| |
| # These are for xcode build directories. |
| xcode3_build_dir = ['build'] |
| xcode4_build_dir = ['build', 'lldb', 'Build', 'Products'] |
| |
| configurations = [['Debug'], ['DebugClang'], ['Release'], ['BuildAndIntegration']] |
| xcode_build_dirs = [xcode3_build_dir, xcode4_build_dir] |
| for configuration in configurations: |
| for xcode_build_dir in xcode_build_dirs: |
| outputPath = os.path.join(lldbRootDirectory, *(xcode_build_dir+configuration) ) |
| result.append(outputPath) |
| |
| return result |
| |
| |
| def createSocketToLocalPort(port): |
| def socket_closer(s): |
| """Close down an opened socket properly.""" |
| s.shutdown(socket.SHUT_RDWR) |
| s.close() |
| |
| sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) |
| sock.connect(("localhost", port)) |
| return (sock, lambda: socket_closer(sock)) |
| |
| |
| def setupTestResults(): |
| """Sets up test results-related objects based on arg settings.""" |
| global results_filename |
| global results_file_object |
| global results_formatter_name |
| global results_formatter_object |
| global results_formatter_options |
| global results_port |
| |
| default_formatter_name = None |
| cleanup_func = None |
| |
| if results_filename: |
| # Open the results file for writing. |
| if results_filename == 'stdout': |
| results_file_object = sys.stdout |
| cleanup_func = None |
| elif results_filename == 'stderr': |
| results_file_object = sys.stderr |
| cleanup_func = None |
| else: |
| results_file_object = open(results_filename, "w") |
| cleanup_func = results_file_object.close |
| default_formatter_name = "lldbsuite.test.test_results.XunitFormatter" |
| elif results_port: |
| # Connect to the specified localhost port. |
| results_file_object, cleanup_func = createSocketToLocalPort( |
| results_port) |
| default_formatter_name = "lldbsuite.test.test_results.RawPickledFormatter" |
| |
| if results_file_object: |
| # We care about the formatter. Choose user-specified or, if |
| # none specified, use the default for the output type. |
| if results_formatter_name: |
| formatter_name = results_formatter_name |
| else: |
| formatter_name = default_formatter_name |
| |
| # Create an instance of the class. First figure out the package/module. |
| components = formatter_name.split(".") |
| module = importlib.import_module(".".join(components[:-1])) |
| |
| # Create the class name we need to load. |
| clazz = getattr(module, components[-1]) |
| |
| # Handle formatter options for the results formatter class. |
| formatter_arg_parser = clazz.arg_parser() |
| if results_formatter_options and len(results_formatter_options) > 0: |
| command_line_options = results_formatter_options |
| else: |
| command_line_options = [] |
| |
| formatter_options = formatter_arg_parser.parse_args( |
| command_line_options) |
| |
| # Create the TestResultsFormatter given the processed options. |
| results_formatter_object = clazz(results_file_object, formatter_options) |
| |
| # Start the results formatter session - we'll only have one |
| # during a given dotest process invocation. |
| initialize_event = EventBuilder.bare_event("initialize") |
| if isMultiprocessTestRunner(): |
| if test_runner_name is not None and test_runner_name == "serial": |
| # Only one worker queue here. |
| worker_count = 1 |
| else: |
| # Workers will be the number of threads specified. |
| worker_count = num_threads |
| else: |
| worker_count = 1 |
| initialize_event["worker_count"] = worker_count |
| |
| results_formatter_object.handle_event(initialize_event) |
| |
| def shutdown_formatter(): |
| # Tell the formatter to write out anything it may have |
| # been saving until the very end (e.g. xUnit results |
| # can't complete its output until this point). |
| results_formatter_object.send_terminate_as_needed() |
| |
| # And now close out the output file-like object. |
| if cleanup_func is not None: |
| cleanup_func() |
| |
| atexit.register(shutdown_formatter) |
| |
| |
| def getOutputPaths(lldbRootDirectory): |
| """ |
| Returns typical build output paths for the lldb executable |
| |
| lldbDirectory - path to the root of the lldb svn/git repo |
| """ |
| result = [] |
| |
| if sys.platform == 'darwin': |
| result.extend(getXcodeOutputPaths(lldbRootDirectory)) |
| |
| # cmake builds? look for build or build/host folder next to llvm directory |
| # lldb is located in llvm/tools/lldb so we need to go up three levels |
| llvmParentDir = os.path.abspath(os.path.join(lldbRootDirectory, os.pardir, os.pardir, os.pardir)) |
| result.append(os.path.join(llvmParentDir, 'build', 'bin')) |
| result.append(os.path.join(llvmParentDir, 'build', 'host', 'bin')) |
| |
| # some cmake developers keep their build directory beside their lldb directory |
| lldbParentDir = os.path.abspath(os.path.join(lldbRootDirectory, os.pardir)) |
| result.append(os.path.join(lldbParentDir, 'build', 'bin')) |
| result.append(os.path.join(lldbParentDir, 'build', 'host', 'bin')) |
| |
| return result |
| |
| def setupSysPath(): |
| """ |
| Add LLDB.framework/Resources/Python to the search paths for modules. |
| As a side effect, we also discover the 'lldb' executable and export it here. |
| """ |
| |
| global rdir |
| global testdirs |
| global dumpSysPath |
| global noHeaders |
| global svn_info |
| global svn_silent |
| global lldbFrameworkPath |
| |
| # Get the directory containing the current script. |
| if "DOTEST_PROFILE" in os.environ and "DOTEST_SCRIPT_DIR" in os.environ: |
| scriptPath = os.environ["DOTEST_SCRIPT_DIR"] |
| else: |
| scriptPath = os.path.dirname(os.path.realpath(__file__)) |
| if not scriptPath.endswith('test'): |
| print("This script expects to reside in lldb's test directory.") |
| sys.exit(-1) |
| |
| if rdir: |
| # Set up the LLDB_TEST environment variable appropriately, so that the |
| # individual tests can be located relatively. |
| # |
| # See also lldbtest.TestBase.setUpClass(cls). |
| if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test': |
| os.environ["LLDB_TEST"] = os.path.join(rdir, 'test') |
| else: |
| os.environ["LLDB_TEST"] = rdir |
| else: |
| os.environ["LLDB_TEST"] = scriptPath |
| |
| # Set up the LLDB_SRC environment variable, so that the tests can locate |
| # the LLDB source code. |
| os.environ["LLDB_SRC"] = lldbsuite.lldb_root |
| |
| pluginPath = os.path.join(scriptPath, 'plugins') |
| toolsLLDBMIPath = os.path.join(scriptPath, 'tools', 'lldb-mi') |
| toolsLLDBServerPath = os.path.join(scriptPath, 'tools', 'lldb-server') |
| |
| # Insert script dir, plugin dir, lldb-mi dir and lldb-server dir to the sys.path. |
| sys.path.insert(0, pluginPath) |
| sys.path.insert(0, toolsLLDBMIPath) # Adding test/tools/lldb-mi to the path makes it easy |
| # to "import lldbmi_testcase" from the MI tests |
| sys.path.insert(0, toolsLLDBServerPath) # Adding test/tools/lldb-server to the path makes it easy |
| # to "import lldbgdbserverutils" from the lldb-server tests |
| |
| # This is the root of the lldb git/svn checkout |
| # When this changes over to a package instead of a standalone script, this |
| # will be `lldbsuite.lldb_root` |
| lldbRootDirectory = lldbsuite.lldb_root |
| |
| # Some of the tests can invoke the 'lldb' command directly. |
| # We'll try to locate the appropriate executable right here. |
| |
| # The lldb executable can be set from the command line |
| # if it's not set, we try to find it now |
| # first, we try the environment |
| if not lldbtest_config.lldbExec: |
| # First, you can define an environment variable LLDB_EXEC specifying the |
| # full pathname of the lldb executable. |
| if "LLDB_EXEC" in os.environ: |
| lldbtest_config.lldbExec = os.environ["LLDB_EXEC"] |
| |
| if not lldbtest_config.lldbExec: |
| outputPaths = getOutputPaths(lldbRootDirectory) |
| for outputPath in outputPaths: |
| candidatePath = os.path.join(outputPath, 'lldb') |
| if is_exe(candidatePath): |
| lldbtest_config.lldbExec = candidatePath |
| break |
| |
| if not lldbtest_config.lldbExec: |
| # Last, check the path |
| lldbtest_config.lldbExec = which('lldb') |
| |
| if lldbtest_config.lldbExec and not is_exe(lldbtest_config.lldbExec): |
| print("'{}' is not a path to a valid executable".format(lldbtest_config.lldbExec)) |
| lldbtest_config.lldbExec = None |
| |
| if not lldbtest_config.lldbExec: |
| print("The 'lldb' executable cannot be located. Some of the tests may not be run as a result.") |
| sys.exit(-1) |
| |
| lldbLibDir = os.path.dirname(lldbtest_config.lldbExec) # confusingly, this is the "bin" directory |
| os.environ["LLDB_LIB_DIR"] = lldbLibDir |
| lldbImpLibDir = os.path.join(lldbLibDir, '..', 'lib') if sys.platform.startswith('win32') else lldbLibDir |
| os.environ["LLDB_IMPLIB_DIR"] = lldbImpLibDir |
| if not noHeaders: |
| print("LLDB library dir:", os.environ["LLDB_LIB_DIR"]) |
| print("LLDB import library dir:", os.environ["LLDB_IMPLIB_DIR"]) |
| os.system('%s -v' % lldbtest_config.lldbExec) |
| |
| # Assume lldb-mi is in same place as lldb |
| # If not found, disable the lldb-mi tests |
| lldbMiExec = None |
| if lldbtest_config.lldbExec and is_exe(lldbtest_config.lldbExec + "-mi"): |
| lldbMiExec = lldbtest_config.lldbExec + "-mi" |
| if not lldbMiExec: |
| if not shouldSkipBecauseOfCategories(["lldb-mi"]): |
| print("The 'lldb-mi' executable cannot be located. The lldb-mi tests can not be run as a result.") |
| skipCategories.append("lldb-mi") |
| else: |
| os.environ["LLDBMI_EXEC"] = lldbMiExec |
| |
| # Skip printing svn/git information when running in parsable (lit-test compatibility) mode |
| if not svn_silent and not parsable: |
| if os.path.isdir(os.path.join(lldbRootDirectory, '.svn')) and which("svn") is not None: |
| pipe = subprocess.Popen([which("svn"), "info", lldbRootDirectory], stdout = subprocess.PIPE) |
| svn_info = pipe.stdout.read() |
| elif os.path.isdir(os.path.join(lldbRootDirectory, '.git')) and which("git") is not None: |
| pipe = subprocess.Popen([which("git"), "svn", "info", lldbRootDirectory], stdout = subprocess.PIPE) |
| svn_info = pipe.stdout.read() |
| if not noHeaders: |
| print(svn_info) |
| |
| global ignore |
| |
| lldbPythonDir = None # The directory that contains 'lldb/__init__.py' |
| if lldbFrameworkPath: |
| candidatePath = os.path.join(lldbFrameworkPath, 'Resources', 'Python') |
| if os.path.isfile(os.path.join(candidatePath, 'lldb/__init__.py')): |
| lldbPythonDir = candidatePath |
| if not lldbPythonDir: |
| print('Resources/Python/lldb/__init__.py was not found in ' + lldbFrameworkPath) |
| sys.exit(-1) |
| else: |
| # The '-i' option is used to skip looking for lldb.py in the build tree. |
| if ignore: |
| return |
| |
| # If our lldb supports the -P option, use it to find the python path: |
| init_in_python_dir = os.path.join('lldb', '__init__.py') |
| |
| lldb_dash_p_result = subprocess.check_output([lldbtest_config.lldbExec, "-P"], stderr=subprocess.STDOUT, universal_newlines=True) |
| |
| if lldb_dash_p_result and not lldb_dash_p_result.startswith(("<", "lldb: invalid option:")) \ |
| and not lldb_dash_p_result.startswith("Traceback"): |
| lines = lldb_dash_p_result.splitlines() |
| |
| # Workaround for readline vs libedit issue on FreeBSD. If stdout |
| # is not a terminal Python executes |
| # rl_variable_bind ("enable-meta-key", "off"); |
| # This produces a warning with FreeBSD's libedit because the |
| # enable-meta-key variable is unknown. Not an issue on Apple |
| # because cpython commit f0ab6f9f0603 added a #ifndef __APPLE__ |
| # around the call. See http://bugs.python.org/issue19884 for more |
| # information. For now we just discard the warning output. |
| if len(lines) >= 1 and lines[0].startswith("bind: Invalid command"): |
| lines.pop(0) |
| |
| # Taking the last line because lldb outputs |
| # 'Cannot read termcap database;\nusing dumb terminal settings.\n' |
| # before the path |
| if len(lines) >= 1 and os.path.isfile(os.path.join(lines[-1], init_in_python_dir)): |
| lldbPythonDir = lines[-1] |
| if "freebsd" in sys.platform or "linux" in sys.platform: |
| os.environ['LLDB_LIB_DIR'] = os.path.join(lldbPythonDir, '..', '..') |
| |
| if not lldbPythonDir: |
| if platform.system() == "Darwin": |
| python_resource_dir = ['LLDB.framework', 'Resources', 'Python'] |
| outputPaths = getXcodeOutputPaths() |
| for outputPath in outputPaths: |
| candidatePath = os.path.join(outputPath, python_resource_dir) |
| if os.path.isfile(os.path.join(candidatePath, init_in_python_dir)): |
| lldbPythonDir = candidatePath |
| break |
| |
| if not lldbPythonDir: |
| print('This script requires lldb.py to be in either ' + dbgPath + ',', end=' ') |
| print(relPath + ', or ' + baiPath + '. Some tests might fail.') |
| else: |
| print("Unable to load lldb extension module. Possible reasons for this include:") |
| print(" 1) LLDB was built with LLDB_DISABLE_PYTHON=1") |
| print(" 2) PYTHONPATH and PYTHONHOME are not set correctly. PYTHONHOME should refer to") |
| print(" the version of Python that LLDB built and linked against, and PYTHONPATH") |
| print(" should contain the Lib directory for the same python distro, as well as the") |
| print(" location of LLDB\'s site-packages folder.") |
| print(" 3) A different version of Python than that which was built against is exported in") |
| print(" the system\'s PATH environment variable, causing conflicts.") |
| print(" 4) The executable '%s' could not be found. Please check " % lldbExecutable) |
| print(" that it exists and is executable.") |
| |
| if lldbPythonDir: |
| lldbPythonDir = os.path.normpath(lldbPythonDir) |
| # Some of the code that uses this path assumes it hasn't resolved the Versions... link. |
| # If the path we've constructed looks like that, then we'll strip out the Versions/A part. |
| (before, frameWithVersion, after) = lldbPythonDir.rpartition("LLDB.framework/Versions/A") |
| if frameWithVersion != "" : |
| lldbPythonDir = before + "LLDB.framework" + after |
| |
| lldbPythonDir = os.path.abspath(lldbPythonDir) |
| |
| # If tests need to find LLDB_FRAMEWORK, now they can do it |
| os.environ["LLDB_FRAMEWORK"] = os.path.dirname(os.path.dirname(lldbPythonDir)) |
| |
| # This is to locate the lldb.py module. Insert it right after sys.path[0]. |
| sys.path[1:1] = [lldbPythonDir] |
| if dumpSysPath: |
| print("sys.path:", sys.path) |
| |
| def visit(prefix, dir, names): |
| """Visitor function for os.path.walk(path, visit, arg).""" |
| |
| global suite |
| global regexp |
| global filters |
| global fs4all |
| global excluded |
| global all_tests |
| |
| if set(dir.split(os.sep)).intersection(excluded): |
| #print("Detected an excluded dir component: %s" % dir) |
| return |
| |
| for name in names: |
| if '.py' == os.path.splitext(name)[1] and name.startswith(prefix): |
| |
| if name in all_tests: |
| raise Exception("Found multiple tests with the name %s" % name) |
| all_tests.add(name) |
| |
| # Try to match the regexp pattern, if specified. |
| if regexp: |
| import re |
| if re.search(regexp, name): |
| #print("Filename: '%s' matches pattern: '%s'" % (name, regexp)) |
| pass |
| else: |
| #print("Filename: '%s' does not match pattern: '%s'" % (name, regexp)) |
| continue |
| |
| # We found a match for our test. Add it to the suite. |
| |
| # Update the sys.path first. |
| if not sys.path.count(dir): |
| sys.path.insert(0, dir) |
| base = os.path.splitext(name)[0] |
| |
| # Thoroughly check the filterspec against the base module and admit |
| # the (base, filterspec) combination only when it makes sense. |
| filterspec = None |
| for filterspec in filters: |
| # Optimistically set the flag to True. |
| filtered = True |
| module = __import__(base) |
| parts = filterspec.split('.') |
| obj = module |
| for part in parts: |
| try: |
| parent, obj = obj, getattr(obj, part) |
| except AttributeError: |
| # The filterspec has failed. |
| filtered = False |
| break |
| |
| # If filtered, we have a good filterspec. Add it. |
| if filtered: |
| #print("adding filter spec %s to module %s" % (filterspec, module)) |
| suite.addTests( |
| unittest2.defaultTestLoader.loadTestsFromName(filterspec, module)) |
| continue |
| |
| # Forgo this module if the (base, filterspec) combo is invalid |
| # and no '-g' option is specified |
| if filters and fs4all and not filtered: |
| continue |
| |
| # Add either the filtered test case(s) (which is done before) or the entire test class. |
| if not filterspec or not filtered: |
| # A simple case of just the module name. Also the failover case |
| # from the filterspec branch when the (base, filterspec) combo |
| # doesn't make sense. |
| suite.addTests(unittest2.defaultTestLoader.loadTestsFromName(base)) |
| |
| |
| def disabledynamics(): |
| import lldb |
| ci = lldb.DBG.GetCommandInterpreter() |
| res = lldb.SBCommandReturnObject() |
| ci.HandleCommand("setting set target.prefer-dynamic-value no-dynamic-values", res, False) |
| if not res.Succeeded(): |
| raise Exception('disabling dynamic type support failed') |
| |
| def lldbLoggings(): |
| import lldb |
| """Check and do lldb loggings if necessary.""" |
| |
| # Turn on logging for debugging purposes if ${LLDB_LOG} environment variable is |
| # defined. Use ${LLDB_LOG} to specify the log file. |
| ci = lldb.DBG.GetCommandInterpreter() |
| res = lldb.SBCommandReturnObject() |
| if ("LLDB_LOG" in os.environ): |
| open(os.environ["LLDB_LOG"], 'w').close() |
| if ("LLDB_LOG_OPTION" in os.environ): |
| lldb_log_option = os.environ["LLDB_LOG_OPTION"] |
| else: |
| lldb_log_option = "event process expr state api" |
| ci.HandleCommand( |
| "log enable -n -f " + os.environ["LLDB_LOG"] + " lldb " + lldb_log_option, |
| res) |
| if not res.Succeeded(): |
| raise Exception('log enable failed (check LLDB_LOG env variable)') |
| |
| if ("LLDB_LINUX_LOG" in os.environ): |
| open(os.environ["LLDB_LINUX_LOG"], 'w').close() |
| if ("LLDB_LINUX_LOG_OPTION" in os.environ): |
| lldb_log_option = os.environ["LLDB_LINUX_LOG_OPTION"] |
| else: |
| lldb_log_option = "event process expr state api" |
| ci.HandleCommand( |
| "log enable -n -f " + os.environ["LLDB_LINUX_LOG"] + " linux " + lldb_log_option, |
| res) |
| if not res.Succeeded(): |
| raise Exception('log enable failed (check LLDB_LINUX_LOG env variable)') |
| |
| # Ditto for gdb-remote logging if ${GDB_REMOTE_LOG} environment variable is defined. |
| # Use ${GDB_REMOTE_LOG} to specify the log file. |
| if ("GDB_REMOTE_LOG" in os.environ): |
| if ("GDB_REMOTE_LOG_OPTION" in os.environ): |
| gdb_remote_log_option = os.environ["GDB_REMOTE_LOG_OPTION"] |
| else: |
| gdb_remote_log_option = "packets process" |
| ci.HandleCommand( |
| "log enable -n -f " + os.environ["GDB_REMOTE_LOG"] + " gdb-remote " |
| + gdb_remote_log_option, |
| res) |
| if not res.Succeeded(): |
| raise Exception('log enable failed (check GDB_REMOTE_LOG env variable)') |
| |
| def getMyCommandLine(): |
| return ' '.join(sys.argv) |
| |
| # ======================================== # |
| # # |
| # Execution of the test driver starts here # |
| # # |
| # ======================================== # |
| |
| def checkDsymForUUIDIsNotOn(): |
| cmd = ["defaults", "read", "com.apple.DebugSymbols"] |
| pipe = subprocess.Popen(cmd, stdout = subprocess.PIPE, stderr = subprocess.STDOUT) |
| cmd_output = pipe.stdout.read() |
| if cmd_output and "DBGFileMappedPaths = " in cmd_output: |
| print("%s =>" % ' '.join(cmd)) |
| print(cmd_output) |
| print("Disable automatic lookup and caching of dSYMs before running the test suite!") |
| print("Exiting...") |
| sys.exit(0) |
| |
| def exitTestSuite(exitCode = None): |
| import lldb |
| lldb.SBDebugger.Terminate() |
| if exitCode: |
| sys.exit(exitCode) |
| |
| |
| def isMultiprocessTestRunner(): |
| # We're not multiprocess when we're either explicitly |
| # the inferior (as specified by the multiprocess test |
| # runner) OR we've been told to skip using the multiprocess |
| # test runner |
| return not (is_inferior_test_runner or no_multiprocess_test_runner) |
| |
| def getVersionForSDK(sdk): |
| sdk = str.lower(sdk) |
| full_path = seven.get_command_output('xcrun -sdk %s --show-sdk-path' % sdk) |
| basename = os.path.basename(full_path) |
| basename = os.path.splitext(basename)[0] |
| basename = str.lower(basename) |
| ver = basename.replace(sdk, '') |
| return ver |
| |
| def getPathForSDK(sdk): |
| sdk = str.lower(sdk) |
| full_path = seven.get_command_output('xcrun -sdk %s --show-sdk-path' % sdk) |
| if os.path.exists(full_path): return full_path |
| return None |
| |
| def setDefaultTripleForPlatform(): |
| if lldb_platform_name == 'ios-simulator': |
| triple_str = 'x86_64-apple-ios%s' % (getVersionForSDK('iphonesimulator')) |
| os.environ['TRIPLE'] = triple_str |
| return {'TRIPLE':triple_str} |
| return {} |
| |
| def run_suite(): |
| global just_do_benchmarks_test |
| global dont_do_dsym_test |
| global dont_do_dwarf_test |
| global dont_do_dwo_test |
| global blacklist |
| global blacklistConfig |
| global categoriesList |
| global validCategories |
| global useCategories |
| global skipCategories |
| global lldbFrameworkPath |
| global configFile |
| global archs |
| global compilers |
| global count |
| global dumpSysPath |
| global bmExecutable |
| global bmBreakpointSpec |
| global bmIterationCount |
| global failed |
| global failfast |
| global filters |
| global fs4all |
| global ignore |
| global progress_bar |
| global runHooks |
| global skip_build_and_cleanup |
| global skip_long_running_test |
| global noHeaders |
| global parsable |
| global regexp |
| global rdir |
| global sdir_name |
| global svn_silent |
| global verbose |
| global testdirs |
| global lldb_platform_name |
| global lldb_platform_url |
| global lldb_platform_working_dir |
| global setCrashInfoHook |
| global is_inferior_test_runner |
| global multiprocess_test_subdir |
| global num_threads |
| global output_on_success |
| global no_multiprocess_test_runner |
| global test_runner_name |
| global results_filename |
| global results_formatter_name |
| global results_formatter_options |
| global results_port |
| |
| # On MacOS X, check to make sure that domain for com.apple.DebugSymbols defaults |
| # does not exist before proceeding to running the test suite. |
| if sys.platform.startswith("darwin"): |
| checkDsymForUUIDIsNotOn() |
| |
| # |
| # Start the actions by first parsing the options while setting up the test |
| # directories, followed by setting up the search paths for lldb utilities; |
| # then, we walk the directory trees and collect the tests into our test suite. |
| # |
| parseOptionsAndInitTestdirs() |
| |
| # Setup test results (test results formatter and output handling). |
| setupTestResults() |
| |
| # If we are running as the multiprocess test runner, kick off the |
| # multiprocess test runner here. |
| if isMultiprocessTestRunner(): |
| from . import dosep |
| dosep.main(output_on_success, num_threads, multiprocess_test_subdir, |
| test_runner_name, results_formatter_object) |
| raise Exception("should never get here") |
| elif is_inferior_test_runner: |
| # Shut off Ctrl-C processing in inferiors. The parallel |
| # test runner handles this more holistically. |
| signal.signal(signal.SIGINT, signal.SIG_IGN) |
| |
| setupSysPath() |
| setupCrashInfoHook() |
| |
| # |
| # If '-l' is specified, do not skip the long running tests. |
| if not skip_long_running_test: |
| os.environ["LLDB_SKIP_LONG_RUNNING_TEST"] = "NO" |
| |
| # For the time being, let's bracket the test runner within the |
| # lldb.SBDebugger.Initialize()/Terminate() pair. |
| import lldb |
| |
| # Create a singleton SBDebugger in the lldb namespace. |
| lldb.DBG = lldb.SBDebugger.Create() |
| |
| if lldb_platform_name: |
| print("Setting up remote platform '%s'" % (lldb_platform_name)) |
| lldb.remote_platform = lldb.SBPlatform(lldb_platform_name) |
| lldb.remote_platform_name = lldb_platform_name |
| if not lldb.remote_platform.IsValid(): |
| print("error: unable to create the LLDB platform named '%s'." % (lldb_platform_name)) |
| exitTestSuite(1) |
| if lldb_platform_url: |
| # We must connect to a remote platform if a LLDB platform URL was specified |
| print("Connecting to remote platform '%s' at '%s'..." % (lldb_platform_name, lldb_platform_url)) |
| lldb.platform_url = lldb_platform_url |
| platform_connect_options = lldb.SBPlatformConnectOptions(lldb_platform_url) |
| err = lldb.remote_platform.ConnectRemote(platform_connect_options) |
| if err.Success(): |
| print("Connected.") |
| else: |
| print("error: failed to connect to remote platform using URL '%s': %s" % (lldb_platform_url, err)) |
| exitTestSuite(1) |
| else: |
| lldb.platform_url = None |
| |
| platform_changes = setDefaultTripleForPlatform() |
| first = True |
| for key in platform_changes: |
| if first: |
| print("Environment variables setup for platform support:") |
| first = False |
| print("%s = %s" % (key,platform_changes[key])) |
| |
| if lldb_platform_working_dir: |
| print("Setting remote platform working directory to '%s'..." % (lldb_platform_working_dir)) |
| lldb.remote_platform.SetWorkingDirectory(lldb_platform_working_dir) |
| lldb.remote_platform_working_dir = lldb_platform_working_dir |
| lldb.DBG.SetSelectedPlatform(lldb.remote_platform) |
| else: |
| lldb.remote_platform = None |
| lldb.remote_platform_working_dir = None |
| lldb.platform_url = None |
| |
| target_platform = lldb.DBG.GetSelectedPlatform().GetTriple().split('-')[2] |
| |
| # By default, both dsym, dwarf and dwo tests are performed. |
| # Use @dsym_test, @dwarf_test or @dwo_test decorators, defined in lldbtest.py, to mark a test as |
| # a dsym, dwarf or dwo test. Use '-N dsym', '-N dwarf' or '-N dwo' to exclude dsym, dwarf or |
| # dwo tests from running. |
| dont_do_dsym_test = dont_do_dsym_test or any(platform in target_platform for platform in ["linux", "freebsd", "windows"]) |
| dont_do_dwo_test = dont_do_dwo_test or any(platform in target_platform for platform in ["darwin", "macosx", "ios"]) |
| |
| # Don't do debugserver tests on everything except OS X. |
| dont_do_debugserver_test = "linux" in target_platform or "freebsd" in target_platform or "windows" in target_platform |
| |
| # Don't do lldb-server (llgs) tests on anything except Linux. |
| dont_do_llgs_test = not ("linux" in target_platform) |
| |
| # |
| # Walk through the testdirs while collecting tests. |
| # |
| for testdir in testdirs: |
| for (dirpath, dirnames, filenames) in os.walk(testdir): |
| visit('Test', dirpath, filenames) |
| |
| # |
| # Now that we have loaded all the test cases, run the whole test suite. |
| # |
| |
| # Put the blacklist in the lldb namespace, to be used by lldb.TestBase. |
| lldb.blacklist = blacklist |
| |
| # The pre_flight and post_flight come from reading a config file. |
| lldb.pre_flight = pre_flight |
| lldb.post_flight = post_flight |
| def getsource_if_available(obj): |
| """ |
| Return the text of the source code for an object if available. Otherwise, |
| a print representation is returned. |
| """ |
| import inspect |
| try: |
| return inspect.getsource(obj) |
| except: |
| return repr(obj) |
| |
| if not noHeaders: |
| print("lldb.pre_flight:", getsource_if_available(lldb.pre_flight)) |
| print("lldb.post_flight:", getsource_if_available(lldb.post_flight)) |
| |
| # If either pre_flight or post_flight is defined, set lldb.test_remote to True. |
| if lldb.pre_flight or lldb.post_flight: |
| lldb.test_remote = True |
| else: |
| lldb.test_remote = False |
| |
| # So do the lldbtest_remote_sandbox and lldbtest_remote_shell_template variables. |
| lldb.lldbtest_remote_sandbox = lldbtest_remote_sandbox |
| lldb.lldbtest_remote_sandboxed_executable = None |
| lldb.lldbtest_remote_shell_template = lldbtest_remote_shell_template |
| |
| # Put all these test decorators in the lldb namespace. |
| lldb.just_do_benchmarks_test = just_do_benchmarks_test |
| lldb.dont_do_dsym_test = dont_do_dsym_test |
| lldb.dont_do_dwarf_test = dont_do_dwarf_test |
| lldb.dont_do_dwo_test = dont_do_dwo_test |
| lldb.dont_do_debugserver_test = dont_do_debugserver_test |
| lldb.dont_do_llgs_test = dont_do_llgs_test |
| |
| # Do we need to skip build and cleanup? |
| lldb.skip_build_and_cleanup = skip_build_and_cleanup |
| |
| # Put bmExecutable, bmBreakpointSpec, and bmIterationCount into the lldb namespace, too. |
| lldb.bmExecutable = bmExecutable |
| lldb.bmBreakpointSpec = bmBreakpointSpec |
| lldb.bmIterationCount = bmIterationCount |
| |
| # And don't forget the runHooks! |
| lldb.runHooks = runHooks |
| |
| # Turn on lldb loggings if necessary. |
| lldbLoggings() |
| |
| # Disable default dynamic types for testing purposes |
| disabledynamics() |
| |
| # Install the control-c handler. |
| unittest2.signals.installHandler() |
| |
| # If sdir_name is not specified through the '-s sdir_name' option, get a |
| # timestamp string and export it as LLDB_SESSION_DIR environment var. This will |
| # be used when/if we want to dump the session info of individual test cases |
| # later on. |
| # |
| # See also TestBase.dumpSessionInfo() in lldbtest.py. |
| import datetime |
| # The windows platforms don't like ':' in the pathname. |
| timestamp_started = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S") |
| if not sdir_name: |
| sdir_name = timestamp_started |
| os.environ["LLDB_SESSION_DIRNAME"] = os.path.join(os.getcwd(), sdir_name) |
| |
| if not noHeaders: |
| sys.stderr.write("\nSession logs for test failures/errors/unexpected successes" |
| " will go into directory '%s'\n" % sdir_name) |
| sys.stderr.write("Command invoked: %s\n" % getMyCommandLine()) |
| |
| if not os.path.isdir(sdir_name): |
| try: |
| os.mkdir(sdir_name) |
| except OSError as exception: |
| if exception.errno != errno.EEXIST: |
| raise |
| where_to_save_session = os.getcwd() |
| fname = os.path.join(sdir_name, "TestStarted-%d" % os.getpid()) |
| with open(fname, "w") as f: |
| print("Test started at: %s\n" % timestamp_started, file=f) |
| print(svn_info, file=f) |
| print("Command invoked: %s\n" % getMyCommandLine(), file=f) |
| |
| # |
| # Invoke the default TextTestRunner to run the test suite, possibly iterating |
| # over different configurations. |
| # |
| |
| iterArchs = False |
| iterCompilers = False |
| |
| if not archs and "archs" in config: |
| archs = config["archs"] |
| |
| if isinstance(archs, list) and len(archs) >= 1: |
| iterArchs = True |
| |
| if not compilers and "compilers" in config: |
| compilers = config["compilers"] |
| |
| # |
| # Add some intervention here to sanity check that the compilers requested are sane. |
| # If found not to be an executable program, the invalid one is dropped from the list. |
| for i in range(len(compilers)): |
| c = compilers[i] |
| if which(c): |
| continue |
| else: |
| if sys.platform.startswith("darwin"): |
| pipe = subprocess.Popen(['xcrun', '-find', c], stdout = subprocess.PIPE, stderr = subprocess.STDOUT) |
| cmd_output = pipe.stdout.read() |
| if cmd_output: |
| if "not found" in cmd_output: |
| print("dropping %s from the compilers used" % c) |
| compilers.remove(i) |
| else: |
| compilers[i] = cmd_output.split('\n')[0] |
| print("'xcrun -find %s' returning %s" % (c, compilers[i])) |
| |
| if not parsable: |
| print("compilers=%s" % str(compilers)) |
| |
| if not compilers or len(compilers) == 0: |
| print("No eligible compiler found, exiting.") |
| exitTestSuite(1) |
| |
| if isinstance(compilers, list) and len(compilers) >= 1: |
| iterCompilers = True |
| |
| # Make a shallow copy of sys.path, we need to manipulate the search paths later. |
| # This is only necessary if we are relocated and with different configurations. |
| if rdir: |
| old_sys_path = sys.path[:] |
| # If we iterate on archs or compilers, there is a chance we want to split stderr/stdout. |
| if iterArchs or iterCompilers: |
| old_stderr = sys.stderr |
| old_stdout = sys.stdout |
| new_stderr = None |
| new_stdout = None |
| |
| # Iterating over all possible architecture and compiler combinations. |
| for ia in range(len(archs) if iterArchs else 1): |
| archConfig = "" |
| if iterArchs: |
| os.environ["ARCH"] = archs[ia] |
| archConfig = "arch=%s" % archs[ia] |
| for ic in range(len(compilers) if iterCompilers else 1): |
| if iterCompilers: |
| os.environ["CC"] = compilers[ic] |
| configString = "%s compiler=%s" % (archConfig, compilers[ic]) |
| else: |
| configString = archConfig |
| |
| if iterArchs or iterCompilers: |
| # Translate ' ' to '-' for pathname component. |
| if six.PY2: |
| tbl = string.maketrans(' ', '-') |
| else: |
| tbl = str.maketrans(' ', '-') |
| configPostfix = configString.translate(tbl) |
| |
| # Check whether we need to split stderr/stdout into configuration |
| # specific files. |
| if old_stderr.name != '<stderr>' and config.get('split_stderr'): |
| if new_stderr: |
| new_stderr.close() |
| new_stderr = open("%s.%s" % (old_stderr.name, configPostfix), "w") |
| sys.stderr = new_stderr |
| if old_stdout.name != '<stdout>' and config.get('split_stdout'): |
| if new_stdout: |
| new_stdout.close() |
| new_stdout = open("%s.%s" % (old_stdout.name, configPostfix), "w") |
| sys.stdout = new_stdout |
| |
| # If we specified a relocated directory to run the test suite, do |
| # the extra housekeeping to copy the testdirs to a configStringified |
| # directory and to update sys.path before invoking the test runner. |
| # The purpose is to separate the configuration-specific directories |
| # from each other. |
| if rdir: |
| from shutil import copytree, rmtree, ignore_patterns |
| |
| newrdir = "%s.%s" % (rdir, configPostfix) |
| |
| # Copy the tree to a new directory with postfix name configPostfix. |
| if os.path.exists(newrdir): |
| rmtree(newrdir) |
| copytree(rdir, newrdir, ignore=ignore_patterns('*.pyc', '*.o', '*.d')) |
| |
| # Update the LLDB_TEST environment variable to reflect new top |
| # level test directory. |
| # |
| # See also lldbtest.TestBase.setUpClass(cls). |
| if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test': |
| os.environ["LLDB_TEST"] = os.path.join(newrdir, 'test') |
| else: |
| os.environ["LLDB_TEST"] = newrdir |
| |
| # And update the Python search paths for modules. |
| sys.path = [x.replace(rdir, newrdir, 1) for x in old_sys_path] |
| |
| # Output the configuration. |
| if not parsable: |
| sys.stderr.write("\nConfiguration: " + configString + "\n") |
| |
| #print("sys.stderr name is", sys.stderr.name) |
| #print("sys.stdout name is", sys.stdout.name) |
| |
| # First, write out the number of collected test cases. |
| if not parsable: |
| sys.stderr.write(separator + "\n") |
| sys.stderr.write("Collected %d test%s\n\n" |
| % (suite.countTestCases(), |
| suite.countTestCases() != 1 and "s" or "")) |
| |
| class LLDBTestResult(unittest2.TextTestResult): |
| """ |
| Enforce a singleton pattern to allow introspection of test progress. |
| |
| Overwrite addError(), addFailure(), and addExpectedFailure() methods |
| to enable each test instance to track its failure/error status. It |
| is used in the LLDB test framework to emit detailed trace messages |
| to a log file for easier human inspection of test failures/errors. |
| """ |
| __singleton__ = None |
| __ignore_singleton__ = False |
| |
| @staticmethod |
| def getTerminalSize(): |
| import os |
| env = os.environ |
| def ioctl_GWINSZ(fd): |
| try: |
| import fcntl, termios, struct, os |
| cr = struct.unpack('hh', fcntl.ioctl(fd, termios.TIOCGWINSZ, |
| '1234')) |
| except: |
| return |
| return cr |
| cr = ioctl_GWINSZ(0) or ioctl_GWINSZ(1) or ioctl_GWINSZ(2) |
| if not cr: |
| try: |
| fd = os.open(os.ctermid(), os.O_RDONLY) |
| cr = ioctl_GWINSZ(fd) |
| os.close(fd) |
| except: |
| pass |
| if not cr: |
| cr = (env.get('LINES', 25), env.get('COLUMNS', 80)) |
| return int(cr[1]), int(cr[0]) |
| |
| def __init__(self, *args): |
| if not LLDBTestResult.__ignore_singleton__ and LLDBTestResult.__singleton__: |
| raise Exception("LLDBTestResult instantiated more than once") |
| super(LLDBTestResult, self).__init__(*args) |
| LLDBTestResult.__singleton__ = self |
| # Now put this singleton into the lldb module namespace. |
| lldb.test_result = self |
| # Computes the format string for displaying the counter. |
| global suite |
| counterWidth = len(str(suite.countTestCases())) |
| self.fmt = "%" + str(counterWidth) + "d: " |
| self.indentation = ' ' * (counterWidth + 2) |
| # This counts from 1 .. suite.countTestCases(). |
| self.counter = 0 |
| (width, height) = LLDBTestResult.getTerminalSize() |
| self.progressbar = None |
| global progress_bar |
| if width > 10 and not parsable and progress_bar: |
| try: |
| self.progressbar = progress.ProgressWithEvents(stdout=self.stream,start=0,end=suite.countTestCases(),width=width-10) |
| except: |
| self.progressbar = None |
| self.results_formatter = results_formatter_object |
| |
| def _config_string(self, test): |
| compiler = getattr(test, "getCompiler", None) |
| arch = getattr(test, "getArchitecture", None) |
| return "%s-%s" % (compiler() if compiler else "", arch() if arch else "") |
| |
| def _exc_info_to_string(self, err, test): |
| """Overrides superclass TestResult's method in order to append |
| our test config info string to the exception info string.""" |
| if hasattr(test, "getArchitecture") and hasattr(test, "getCompiler"): |
| return '%sConfig=%s-%s' % (super(LLDBTestResult, self)._exc_info_to_string(err, test), |
| test.getArchitecture(), |
| test.getCompiler()) |
| else: |
| return super(LLDBTestResult, self)._exc_info_to_string(err, test) |
| |
| def getDescription(self, test): |
| doc_first_line = test.shortDescription() |
| if self.descriptions and doc_first_line: |
| return '\n'.join((str(test), self.indentation + doc_first_line)) |
| else: |
| return str(test) |
| |
| def getCategoriesForTest(self,test): |
| if hasattr(test,"_testMethodName"): |
| test_method = getattr(test,"_testMethodName") |
| test_method = getattr(test,test_method) |
| else: |
| test_method = None |
| if test_method != None and hasattr(test_method,"getCategories"): |
| test_categories = test_method.getCategories(test) |
| elif hasattr(test,"getCategories"): |
| test_categories = test.getCategories() |
| elif inspect.ismethod(test) and test.__self__ != None and hasattr(test.__self__,"getCategories"): |
| test_categories = test.__self__.getCategories() |
| else: |
| test_categories = [] |
| if test_categories == None: |
| test_categories = [] |
| return test_categories |
| |
| def hardMarkAsSkipped(self,test): |
| getattr(test, test._testMethodName).__func__.__unittest_skip__ = True |
| getattr(test, test._testMethodName).__func__.__unittest_skip_why__ = "test case does not fall in any category of interest for this run" |
| test.__class__.__unittest_skip__ = True |
| test.__class__.__unittest_skip_why__ = "test case does not fall in any category of interest for this run" |
| |
| def startTest(self, test): |
| if shouldSkipBecauseOfCategories(self.getCategoriesForTest(test)): |
| self.hardMarkAsSkipped(test) |
| global setCrashInfoHook |
| setCrashInfoHook("%s at %s" % (str(test),inspect.getfile(test.__class__))) |
| self.counter += 1 |
| #if self.counter == 4: |
| # import crashinfo |
| # crashinfo.testCrashReporterDescription(None) |
| test.test_number = self.counter |
| if self.showAll: |
| self.stream.write(self.fmt % self.counter) |
| super(LLDBTestResult, self).startTest(test) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_start(test)) |
| |
| def addSuccess(self, test): |
| global parsable |
| super(LLDBTestResult, self).addSuccess(test) |
| if parsable: |
| self.stream.write("PASS: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_success(test)) |
| |
| def addError(self, test, err): |
| global sdir_has_content |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addError(test, err) |
| method = getattr(test, "markError", None) |
| if method: |
| method() |
| if parsable: |
| self.stream.write("FAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_error(test, err)) |
| |
| def addCleanupError(self, test, err): |
| global sdir_has_content |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addCleanupError(test, err) |
| method = getattr(test, "markCleanupError", None) |
| if method: |
| method() |
| if parsable: |
| self.stream.write("CLEANUP ERROR: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_cleanup_error( |
| test, err)) |
| |
| def addFailure(self, test, err): |
| global sdir_has_content |
| global failuresPerCategory |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addFailure(test, err) |
| method = getattr(test, "markFailure", None) |
| if method: |
| method() |
| if parsable: |
| self.stream.write("FAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if useCategories: |
| test_categories = self.getCategoriesForTest(test) |
| for category in test_categories: |
| if category in failuresPerCategory: |
| failuresPerCategory[category] = failuresPerCategory[category] + 1 |
| else: |
| failuresPerCategory[category] = 1 |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_failure(test, err)) |
| |
| |
| def addExpectedFailure(self, test, err, bugnumber): |
| global sdir_has_content |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addExpectedFailure(test, err, bugnumber) |
| method = getattr(test, "markExpectedFailure", None) |
| if method: |
| method(err, bugnumber) |
| if parsable: |
| self.stream.write("XFAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_expected_failure( |
| test, err, bugnumber)) |
| |
| def addSkip(self, test, reason): |
| global sdir_has_content |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addSkip(test, reason) |
| method = getattr(test, "markSkippedTest", None) |
| if method: |
| method() |
| if parsable: |
| self.stream.write("UNSUPPORTED: LLDB (%s) :: %s (%s) \n" % (self._config_string(test), str(test), reason)) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_skip(test, reason)) |
| |
| def addUnexpectedSuccess(self, test, bugnumber): |
| global sdir_has_content |
| global parsable |
| sdir_has_content = True |
| super(LLDBTestResult, self).addUnexpectedSuccess(test, bugnumber) |
| method = getattr(test, "markUnexpectedSuccess", None) |
| if method: |
| method(bugnumber) |
| if parsable: |
| self.stream.write("XPASS: LLDB (%s) :: %s\n" % (self._config_string(test), str(test))) |
| if self.results_formatter: |
| self.results_formatter.handle_event( |
| EventBuilder.event_for_unexpected_success( |
| test, bugnumber)) |
| |
| |
| if parsable: |
| v = 0 |
| elif progress_bar: |
| v = 1 |
| else: |
| v = verbose |
| |
| # Invoke the test runner. |
| if count == 1: |
| result = unittest2.TextTestRunner(stream=sys.stderr, |
| verbosity=v, |
| failfast=failfast, |
| resultclass=LLDBTestResult).run(suite) |
| else: |
| # We are invoking the same test suite more than once. In this case, |
| # mark __ignore_singleton__ flag as True so the signleton pattern is |
| # not enforced. |
| LLDBTestResult.__ignore_singleton__ = True |
| for i in range(count): |
| |
| result = unittest2.TextTestRunner(stream=sys.stderr, |
| verbosity=v, |
| failfast=failfast, |
| resultclass=LLDBTestResult).run(suite) |
| |
| failed = failed or not result.wasSuccessful() |
| |
| if sdir_has_content and not parsable: |
| sys.stderr.write("Session logs for test failures/errors/unexpected successes" |
| " can be found in directory '%s'\n" % sdir_name) |
| |
| if useCategories and len(failuresPerCategory) > 0: |
| sys.stderr.write("Failures per category:\n") |
| for category in failuresPerCategory: |
| sys.stderr.write("%s - %d\n" % (category,failuresPerCategory[category])) |
| |
| os.chdir(where_to_save_session) |
| fname = os.path.join(sdir_name, "TestFinished-%d" % os.getpid()) |
| with open(fname, "w") as f: |
| print("Test finished at: %s\n" % datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S"), file=f) |
| |
| # Terminate the test suite if ${LLDB_TESTSUITE_FORCE_FINISH} is defined. |
| # This should not be necessary now. |
| if ("LLDB_TESTSUITE_FORCE_FINISH" in os.environ): |
| print("Terminating Test suite...") |
| subprocess.Popen(["/bin/sh", "-c", "kill %s; exit 0" % (os.getpid())]) |
| |
| # Exiting. |
| exitTestSuite(failed) |
| |
| if __name__ == "__main__": |
| print(__file__ + " is for use as a module only. It should not be run as a standalone script.") |
| sys.exit(-1) |