| #!/usr/bin/env python |
| |
| """ |
| MultiTestRunner - Harness for running multiple tests in the simple clang style. |
| |
| TODO |
| -- |
| - Use configuration file for clang specific stuff |
| - Use a timeout / ulimit |
| - Detect signaled failures (abort) |
| - Better support for finding tests |
| |
| - Support "disabling" tests? The advantage of making this distinct from XFAIL |
| is it makes it more obvious that it is a temporary measure (and MTR can put |
| in a separate category). |
| """ |
| |
| # TOD |
| import os, sys, re, random, time |
| import threading |
| import ProgressBar |
| import TestRunner |
| from TestRunner import TestStatus |
| from Queue import Queue |
| |
| kTestFileExtensions = set(['.mi','.i','.c','.cpp','.m','.mm','.ll']) |
| |
| def getTests(inputs): |
| for path in inputs: |
| if not os.path.exists(path): |
| print >>sys.stderr,"WARNING: Invalid test \"%s\""%(path,) |
| continue |
| |
| if not os.path.isdir(path): |
| yield path |
| |
| for dirpath,dirnames,filenames in os.walk(path): |
| # FIXME: This doesn't belong here |
| if 'Output' in dirnames: |
| dirnames.remove('Output') |
| for f in filenames: |
| base,ext = os.path.splitext(f) |
| if ext in kTestFileExtensions: |
| yield os.path.join(dirpath,f) |
| |
| class TestingProgressDisplay: |
| def __init__(self, opts, numTests, progressBar=None): |
| self.opts = opts |
| self.numTests = numTests |
| self.digits = len(str(self.numTests)) |
| self.current = None |
| self.lock = threading.Lock() |
| self.progressBar = progressBar |
| self.progress = 0. |
| |
| def update(self, index, tr): |
| # Avoid locking overhead in quiet mode |
| if self.opts.quiet and not tr.failed(): |
| return |
| |
| # Output lock |
| self.lock.acquire() |
| try: |
| self.handleUpdate(index, tr) |
| finally: |
| self.lock.release() |
| |
| def finish(self): |
| if self.progressBar: |
| self.progressBar.clear() |
| elif self.opts.succinct: |
| sys.stdout.write('\n') |
| |
| def handleUpdate(self, index, tr): |
| if self.progressBar: |
| if tr.failed(): |
| self.progressBar.clear() |
| else: |
| # Force monotonicity |
| self.progress = max(self.progress, float(index)/self.numTests) |
| self.progressBar.update(self.progress, tr.path) |
| return |
| elif self.opts.succinct: |
| if not tr.failed(): |
| sys.stdout.write('.') |
| sys.stdout.flush() |
| return |
| else: |
| sys.stdout.write('\n') |
| |
| status = TestStatus.getName(tr.code).upper() |
| print '%s: %s (%*d of %*d)' % (status, tr.path, |
| self.digits, index+1, |
| self.digits, self.numTests) |
| |
| if tr.failed() and self.opts.showOutput: |
| print "%s TEST '%s' FAILED %s" % ('*'*20, tr.path, '*'*20) |
| print tr.output |
| print "*" * 20 |
| |
| class TestResult: |
| def __init__(self, path, code, output, elapsed): |
| self.path = path |
| self.code = code |
| self.output = output |
| self.elapsed = elapsed |
| |
| def failed(self): |
| return self.code in (TestStatus.Fail,TestStatus.XPass) |
| |
| class TestProvider: |
| def __init__(self, opts, tests, display): |
| self.opts = opts |
| self.tests = tests |
| self.index = 0 |
| self.lock = threading.Lock() |
| self.results = [None]*len(self.tests) |
| self.startTime = time.time() |
| self.progress = display |
| |
| def get(self): |
| self.lock.acquire() |
| try: |
| if self.opts.maxTime is not None: |
| if time.time() - self.startTime > self.opts.maxTime: |
| return None |
| if self.index >= len(self.tests): |
| return None |
| item = self.tests[self.index],self.index |
| self.index += 1 |
| return item |
| finally: |
| self.lock.release() |
| |
| def setResult(self, index, result): |
| self.results[index] = result |
| self.progress.update(index, result) |
| |
| class Tester(threading.Thread): |
| def __init__(self, provider): |
| threading.Thread.__init__(self) |
| self.provider = provider |
| |
| def run(self): |
| while 1: |
| item = self.provider.get() |
| if item is None: |
| break |
| self.runTest(item) |
| |
| def runTest(self, (path, index)): |
| base = TestRunner.getTestOutputBase('Output', path) |
| numTests = len(self.provider.tests) |
| digits = len(str(numTests)) |
| code = None |
| elapsed = None |
| try: |
| opts = self.provider.opts |
| if opts.debugDoNotTest: |
| code = None |
| else: |
| startTime = time.time() |
| code, output = TestRunner.runOneTest(path, base, |
| opts.clang, opts.clangcc, |
| opts.useValgrind) |
| elapsed = time.time() - startTime |
| except KeyboardInterrupt: |
| # This is a sad hack. Unfortunately subprocess goes |
| # bonkers with ctrl-c and we start forking merrily. |
| print 'Ctrl-C detected, goodbye.' |
| os.kill(0,9) |
| |
| self.provider.setResult(index, TestResult(path, code, output, elapsed)) |
| |
| def detectCPUs(): |
| """ |
| Detects the number of CPUs on a system. Cribbed from pp. |
| """ |
| # Linux, Unix and MacOS: |
| if hasattr(os, "sysconf"): |
| if os.sysconf_names.has_key("SC_NPROCESSORS_ONLN"): |
| # Linux & Unix: |
| ncpus = os.sysconf("SC_NPROCESSORS_ONLN") |
| if isinstance(ncpus, int) and ncpus > 0: |
| return ncpus |
| else: # OSX: |
| return int(os.popen2("sysctl -n hw.ncpu")[1].read()) |
| # Windows: |
| if os.environ.has_key("NUMBER_OF_PROCESSORS"): |
| ncpus = int(os.environ["NUMBER_OF_PROCESSORS"]); |
| if ncpus > 0: |
| return ncpus |
| return 1 # Default |
| |
| def main(): |
| global options |
| from optparse import OptionParser |
| parser = OptionParser("usage: %prog [options] {inputs}") |
| parser.add_option("-j", "--threads", dest="numThreads", |
| help="Number of testing threads", |
| type=int, action="store", |
| default=detectCPUs()) |
| parser.add_option("", "--clang", dest="clang", |
| help="Program to use as \"clang\"", |
| action="store", default=None) |
| parser.add_option("", "--clang-cc", dest="clangcc", |
| help="Program to use as \"clang-cc\"", |
| action="store", default=None) |
| parser.add_option("", "--vg", dest="useValgrind", |
| help="Run tests under valgrind", |
| action="store_true", default=False) |
| parser.add_option("-v", "--verbose", dest="showOutput", |
| help="Show all test output", |
| action="store_true", default=False) |
| parser.add_option("-q", "--quiet", dest="quiet", |
| help="Suppress no error output", |
| action="store_true", default=False) |
| parser.add_option("-s", "--succinct", dest="succinct", |
| help="Reduce amount of output", |
| action="store_true", default=False) |
| parser.add_option("", "--max-tests", dest="maxTests", |
| help="Maximum number of tests to run", |
| action="store", type=int, default=None) |
| parser.add_option("", "--max-time", dest="maxTime", |
| help="Maximum time to spend testing (in seconds)", |
| action="store", type=float, default=None) |
| parser.add_option("", "--shuffle", dest="shuffle", |
| help="Run tests in random order", |
| action="store_true", default=False) |
| parser.add_option("", "--seed", dest="seed", |
| help="Seed for random number generator (default: random)", |
| action="store", default=None) |
| parser.add_option("", "--no-progress-bar", dest="useProgressBar", |
| help="Do not use curses based progress bar", |
| action="store_false", default=True) |
| parser.add_option("", "--debug-do-not-test", dest="debugDoNotTest", |
| help="DEBUG: Skip running actual test script", |
| action="store_true", default=False) |
| parser.add_option("", "--time-tests", dest="timeTests", |
| help="Track elapsed wall time for each test", |
| action="store_true", default=False) |
| parser.add_option("", "--path", dest="path", |
| help="Additional paths to add to testing environment", |
| action="append", type=str, default=[]) |
| |
| (opts, args) = parser.parse_args() |
| |
| if not args: |
| parser.error('No inputs specified') |
| |
| # FIXME: Move into configuration object. |
| TestRunner.kChildEnv["PATH"] = os.pathsep.join(opts.path + |
| [TestRunner.kChildEnv['PATH']]) |
| |
| if opts.clang is None: |
| opts.clang = TestRunner.inferClang() |
| if opts.clangcc is None: |
| opts.clangcc = TestRunner.inferClangCC(opts.clang) |
| |
| # FIXME: It could be worth loading these in parallel with testing. |
| allTests = list(getTests(args)) |
| allTests.sort() |
| |
| tests = allTests |
| if opts.seed is not None: |
| try: |
| seed = int(opts.seed) |
| except: |
| parser.error('--seed argument should be an integer') |
| random.seed(seed) |
| if opts.shuffle: |
| random.shuffle(tests) |
| if opts.maxTests is not None: |
| tests = tests[:opts.maxTests] |
| |
| extra = '' |
| if len(tests) != len(allTests): |
| extra = ' of %d'%(len(allTests),) |
| header = '-- Testing: %d%s tests, %d threads --'%(len(tests),extra, |
| opts.numThreads) |
| |
| progressBar = None |
| if not opts.quiet: |
| if opts.useProgressBar: |
| try: |
| tc = ProgressBar.TerminalController() |
| progressBar = ProgressBar.ProgressBar(tc, header) |
| except ValueError: |
| pass |
| |
| if not progressBar: |
| print header |
| |
| display = TestingProgressDisplay(opts, len(tests), progressBar) |
| provider = TestProvider(opts, tests, display) |
| |
| testers = [Tester(provider) for i in range(opts.numThreads)] |
| startTime = time.time() |
| for t in testers: |
| t.start() |
| try: |
| for t in testers: |
| t.join() |
| except KeyboardInterrupt: |
| sys.exit(1) |
| |
| display.finish() |
| |
| if not opts.quiet: |
| print 'Testing Time: %.2fs'%(time.time() - startTime) |
| |
| # List test results organized organized by kind. |
| byCode = {} |
| for t in provider.results: |
| if t: |
| if t.code not in byCode: |
| byCode[t.code] = [] |
| byCode[t.code].append(t) |
| for title,code in (('Unexpected Passing Tests', TestStatus.XPass), |
| ('Failing Tests', TestStatus.Fail)): |
| elts = byCode.get(code) |
| if not elts: |
| continue |
| print '*'*20 |
| print '%s (%d):' % (title, len(elts)) |
| for tr in elts: |
| print '\t%s'%(tr.path,) |
| |
| numFailures = len(byCode.get(TestStatus.Fail,[])) |
| if numFailures: |
| print '\nFailures: %d' % (numFailures,) |
| sys.exit(1) |
| |
| if opts.timeTests: |
| print '\nTest Times:' |
| provider.results.sort(key=lambda t: t and t.elapsed) |
| for tr in provider.results: |
| if tr: |
| print '%.2fs: %s' % (tr.elapsed, tr.path) |
| |
| if __name__=='__main__': |
| main() |