| #!/usr/bin/env python |
| |
| """ |
| Run the test suite using a separate process for each test file. |
| |
| Each test will run with a time limit of 10 minutes by default. |
| |
| Override the default time limit of 10 minutes by setting |
| the environment variable LLDB_TEST_TIMEOUT. |
| |
| E.g., export LLDB_TEST_TIMEOUT=10m |
| |
| Override the time limit for individual tests by setting |
| the environment variable LLDB_[TEST NAME]_TIMEOUT. |
| |
| E.g., export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=2m |
| |
| Set to "0" to run without time limit. |
| |
| E.g., export LLDB_TEST_TIMEOUT=0 |
| or export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=0 |
| |
| To collect core files for timed out tests, do the following before running dosep.py |
| |
| OSX |
| ulimit -c unlimited |
| sudo sysctl -w kern.corefile=core.%P |
| |
| Linux: |
| ulimit -c unlimited |
| echo core.%p | sudo tee /proc/sys/kernel/core_pattern |
| """ |
| |
| import multiprocessing |
| import os |
| import fnmatch |
| import platform |
| import re |
| import dotest_args |
| import shlex |
| import subprocess |
| import sys |
| |
| from optparse import OptionParser |
| |
| def get_timeout_command(): |
| """Search for a suitable timeout command.""" |
| if sys.platform.startswith("win32"): |
| return None |
| try: |
| subprocess.call("timeout", stderr=subprocess.PIPE) |
| return "timeout" |
| except OSError: |
| pass |
| try: |
| subprocess.call("gtimeout", stderr=subprocess.PIPE) |
| return "gtimeout" |
| except OSError: |
| pass |
| return None |
| |
| timeout_command = get_timeout_command() |
| |
| default_timeout = os.getenv("LLDB_TEST_TIMEOUT") or "10m" |
| |
| # Status codes for running command with timeout. |
| eTimedOut, ePassed, eFailed = 124, 0, 1 |
| |
| def parse_test_results(output): |
| passes = 0 |
| failures = 0 |
| for result in output: |
| pass_count = re.search("^RESULT:.*([0-9]+) passes", result, re.MULTILINE) |
| fail_count = re.search("^RESULT:.*([0-9]+) failures", result, re.MULTILINE) |
| error_count = re.search("^RESULT:.*([0-9]+) errors", result, re.MULTILINE) |
| this_fail_count = 0 |
| this_error_count = 0 |
| if pass_count != None: |
| passes = passes + int(pass_count.group(1)) |
| if fail_count != None: |
| failures = failures + int(fail_count.group(1)) |
| if error_count != None: |
| failures = failures + int(error_count.group(1)) |
| pass |
| return passes, failures |
| |
| def call_with_timeout(command, timeout): |
| """Run command with a timeout if possible.""" |
| """-s QUIT will create a coredump if they are enabled on your system""" |
| process = None |
| if timeout_command and timeout != "0": |
| command = [timeout_command, '-s', 'QUIT', timeout] + command |
| # Specifying a value for close_fds is unsupported on Windows when using subprocess.PIPE |
| if os.name != "nt": |
| process = subprocess.Popen(command, stdin=subprocess.PIPE, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE, |
| close_fds=True) |
| else: |
| process = subprocess.Popen(command, stdin=subprocess.PIPE, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE) |
| output = process.communicate() |
| exit_status = process.returncode |
| passes, failures = parse_test_results(output) |
| return exit_status, passes, failures |
| |
| def process_dir(root, files, test_root, dotest_argv): |
| """Examine a directory for tests, and invoke any found within it.""" |
| timed_out = [] |
| failed = [] |
| passed = [] |
| pass_sub_count = 0 |
| fail_sub_count = 0 |
| for name in files: |
| path = os.path.join(root, name) |
| |
| # We're only interested in the test file with the "Test*.py" naming pattern. |
| if not name.startswith("Test") or not name.endswith(".py"): |
| continue |
| |
| # Neither a symbolically linked file. |
| if os.path.islink(path): |
| continue |
| |
| script_file = os.path.join(test_root, "dotest.py") |
| command = ([sys.executable, script_file] + |
| dotest_argv + |
| ["-p", name, root]) |
| |
| timeout_name = os.path.basename(os.path.splitext(name)[0]).upper() |
| |
| timeout = os.getenv("LLDB_%s_TIMEOUT" % timeout_name) or default_timeout |
| |
| exit_status, pass_count, fail_count = call_with_timeout(command, timeout) |
| |
| pass_sub_count = pass_sub_count + pass_count |
| fail_sub_count = fail_sub_count + fail_count |
| |
| if exit_status == ePassed: |
| passed.append(name) |
| else: |
| if eTimedOut == exit_status: |
| timed_out.append(name) |
| failed.append(name) |
| return (timed_out, failed, passed, fail_sub_count, pass_sub_count) |
| |
| in_q = None |
| out_q = None |
| |
| def process_dir_worker(arg_tuple): |
| """Worker thread main loop when in multithreaded mode. |
| Takes one directory specification at a time and works on it.""" |
| (root, files, test_root, dotest_argv) = arg_tuple |
| return process_dir(root, files, test_root, dotest_argv) |
| |
| def walk_and_invoke(test_directory, test_subdir, dotest_argv, num_threads): |
| """Look for matched files and invoke test driver on each one. |
| In single-threaded mode, each test driver is invoked directly. |
| In multi-threaded mode, submit each test driver to a worker |
| queue, and then wait for all to complete. |
| |
| test_directory - lldb/test/ directory |
| test_subdir - lldb/test/ or a subfolder with the tests we're interested in running |
| """ |
| |
| # Collect the test files that we'll run. |
| test_work_items = [] |
| for root, dirs, files in os.walk(test_subdir, topdown=False): |
| test_work_items.append((root, files, test_directory, dotest_argv)) |
| |
| # Run the items, either in a pool (for multicore speedup) or |
| # calling each individually. |
| if num_threads > 1: |
| pool = multiprocessing.Pool(num_threads) |
| test_results = pool.map(process_dir_worker, test_work_items) |
| else: |
| test_results = [] |
| for work_item in test_work_items: |
| test_results.append(process_dir_worker(work_item)) |
| |
| timed_out = [] |
| failed = [] |
| passed = [] |
| fail_sub_count = 0 |
| pass_sub_count = 0 |
| |
| for test_result in test_results: |
| (dir_timed_out, dir_failed, dir_passed, dir_fail_sub_count, dir_pass_sub_count) = test_result |
| timed_out += dir_timed_out |
| failed += dir_failed |
| passed += dir_passed |
| fail_sub_count = fail_sub_count + dir_fail_sub_count |
| pass_sub_count = pass_sub_count + dir_pass_sub_count |
| |
| return (timed_out, failed, passed, fail_sub_count, pass_sub_count) |
| |
| def getExpectedTimeouts(platform_name): |
| # returns a set of test filenames that might timeout |
| # are we running against a remote target? |
| if platform_name is None: |
| target = sys.platform |
| remote = False |
| else: |
| m = re.search('remote-(\w+)', platform_name) |
| target = m.group(1) |
| remote = True |
| |
| expected_timeout = set() |
| |
| if target.startswith("linux"): |
| expected_timeout |= { |
| "TestAttachDenied.py", |
| "TestAttachResume.py", |
| "TestConnectRemote.py", |
| "TestCreateAfterAttach.py", |
| "TestEvents.py", |
| "TestExitDuringStep.py", |
| "TestThreadStepOut.py", |
| } |
| elif target.startswith("android"): |
| expected_timeout |= { |
| "TestExitDuringStep.py", |
| "TestHelloWorld.py", |
| } |
| elif target.startswith("freebsd"): |
| expected_timeout |= { |
| "TestBreakpointConditions.py", |
| "TestChangeProcessGroup.py", |
| "TestValueObjectRecursion.py", |
| "TestWatchpointConditionAPI.py", |
| } |
| elif target.startswith("darwin"): |
| expected_timeout |= { |
| "TestThreadSpecificBreakpoint.py", # times out on MBP Retina, Mid 2012 |
| } |
| return expected_timeout |
| |
| def touch(fname, times=None): |
| with open(fname, 'a'): |
| os.utime(fname, times) |
| |
| def find(pattern, path): |
| result = [] |
| for root, dirs, files in os.walk(path): |
| for name in files: |
| if fnmatch.fnmatch(name, pattern): |
| result.append(os.path.join(root, name)) |
| return result |
| |
| def main(): |
| # We can't use sys.path[0] to determine the script directory |
| # because it doesn't work under a debugger |
| test_directory = os.path.dirname(os.path.realpath(__file__)) |
| parser = OptionParser(usage="""\ |
| Run lldb test suite using a separate process for each test file. |
| |
| Each test will run with a time limit of 10 minutes by default. |
| |
| Override the default time limit of 10 minutes by setting |
| the environment variable LLDB_TEST_TIMEOUT. |
| |
| E.g., export LLDB_TEST_TIMEOUT=10m |
| |
| Override the time limit for individual tests by setting |
| the environment variable LLDB_[TEST NAME]_TIMEOUT. |
| |
| E.g., export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=2m |
| |
| Set to "0" to run without time limit. |
| |
| E.g., export LLDB_TEST_TIMEOUT=0 |
| or export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=0 |
| """) |
| parser.add_option('-o', '--options', |
| type='string', action='store', |
| dest='dotest_options', |
| help="""The options passed to 'dotest.py' if specified.""") |
| |
| parser.add_option('-t', '--threads', |
| type='int', |
| dest='num_threads', |
| help="""The number of threads to use when running tests separately.""") |
| |
| opts, args = parser.parse_args() |
| dotest_option_string = opts.dotest_options |
| |
| is_posix = (os.name == "posix") |
| dotest_argv = shlex.split(dotest_option_string, posix=is_posix) if dotest_option_string else [] |
| |
| parser = dotest_args.create_parser() |
| dotest_options = dotest_args.parse_args(parser, dotest_argv) |
| |
| if not dotest_options.s: |
| # no session log directory, we need to add this to prevent |
| # every dotest invocation from creating its own directory |
| import datetime |
| # The windows platforms don't like ':' in the pathname. |
| timestamp_started = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S") |
| dotest_argv.append('-s') |
| dotest_argv.append(timestamp_started) |
| dotest_options.s = timestamp_started |
| |
| session_dir = os.path.join(os.getcwd(), dotest_options.s) |
| |
| # The root directory was specified on the command line |
| if len(args) == 0: |
| test_subdir = test_directory |
| else: |
| test_subdir = os.path.join(test_directory, args[0]) |
| |
| # clean core files in test tree from previous runs (Linux) |
| cores = find('core.*', test_subdir) |
| for core in cores: |
| os.unlink(core) |
| |
| if opts.num_threads: |
| num_threads = opts.num_threads |
| else: |
| num_threads_str = os.environ.get("LLDB_TEST_THREADS") |
| if num_threads_str: |
| num_threads = int(num_threads_str) |
| else: |
| num_threads = multiprocessing.cpu_count() |
| if num_threads < 1: |
| num_threads = 1 |
| |
| system_info = " ".join(platform.uname()) |
| (timed_out, failed, passed, all_fails, all_passes) = walk_and_invoke(test_directory, test_subdir, dotest_argv, num_threads) |
| |
| timed_out = set(timed_out) |
| num_test_files = len(failed) + len(passed) |
| num_tests = all_fails + all_passes |
| |
| # move core files into session dir |
| cores = find('core.*', test_subdir) |
| for core in cores: |
| dst = core.replace(test_directory, "")[1:] |
| dst = dst.replace(os.path.sep, "-") |
| os.rename(core, os.path.join(session_dir, dst)) |
| |
| # remove expected timeouts from failures |
| expected_timeout = getExpectedTimeouts(dotest_options.lldb_platform_name) |
| for xtime in expected_timeout: |
| if xtime in timed_out: |
| timed_out.remove(xtime) |
| failed.remove(xtime) |
| result = "ExpectedTimeout" |
| elif xtime in passed: |
| result = "UnexpectedCompletion" |
| else: |
| result = None # failed |
| |
| if result: |
| test_name = os.path.splitext(xtime)[0] |
| touch(os.path.join(session_dir, "{}-{}".format(result, test_name))) |
| |
| print "Ran %d test suites (%d failed) (%f%%)" % (num_test_files, len(failed), 100.0*len(failed)/num_test_files) |
| print "Ran %d test cases (%d failed) (%f%%)" % (num_tests, all_fails, 100.0*all_fails/num_tests) |
| if len(failed) > 0: |
| failed.sort() |
| print "Failing Tests (%d)" % len(failed) |
| for f in failed: |
| print "%s: LLDB (suite) :: %s (%s)" % ( |
| "TIMEOUT" if f in timed_out else "FAIL", f, system_info |
| ) |
| sys.exit(1) |
| sys.exit(0) |
| |
| if __name__ == '__main__': |
| main() |