blob: e1c5ce5abaea4625c59697dce391f5d3abb2ca0d [file] [log] [blame]
#!/usr/bin/env python
"""
An LTP [execution and] parsing wrapper.
Used as a second layer for ease-of-use with users as many developers
complain about complexity involved with trying to use LTP in my
organization -_-.
Copyright (C) 2009, Garrett Cooper
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License along
with this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
"""
from optparse import OptionGroup, OptionParser
import os, re, sys
class ResultsParseException(Exception):
""" Extended class for parsing LTP results. """
def parse_ltp_results(exec_log, output_log, verbose=0):
""" Function for parsing LTP results.
1. The exec log is the log with the results in summary form.
And now a note from our sponsors about exec logs...
startup='Thu Oct 1 06:42:07 2009'
tag=abort01 stime=1254379327 dur=2 exit=exited stat=0 core=no cu=0 cs=16
tag=accept01 stime=1254379329 dur=0 exit=exited stat=0 core=no cu=1 cs=0
tag=access01 stime=1254379329 dur=0 exit=exited stat=0 core=no cu=0 cs=0
tag=access02 stime=1254379329 dur=0 exit=exited stat=0 core=no cu=0 cs=0
tag=access03 stime=1254379329 dur=1 exit=exited stat=0 core=no cu=0 cs=1
[...]
a. tag is the test tag name.
b. stime is the system time at the start of the exec.
c. dur is the total duration of the test.
d. exit tells you what the result was. Valid values are:
- exited
- signaled
- stopped
- unknown
See run_child in pan.c.
e. stat is the exit status.
f. core answers the question: `did I dump core?'.
g. cu is the cutime (cumulative user time).
h. cs is the cstime (cumulative system time).
2. The output log is the log with all of the terse results.
3. verbose tells us whether or not we need to include the passed results.
"""
if not os.access(exec_log, os.R_OK):
raise ResultsParseException("Exec log - %s - specified doesn't exist"
% exec_log)
elif 1 < verbose and not os.access(output_log, os.R_OK):
# Need the output log for context to the end user.
raise ResultsParseException("Output log - %s - specified doesn't exist"
% output_log )
context = None
failed = [ ]
passed = 0
if 2 <= verbose:
passed = [ ]
target_vals = ( 'exited', '0', 'no' )
fd = open(exec_log, 'r')
try:
content = fd.read()
matches = re.finditer('tag=(?P<tag>\w+).+exit=(?P<exit>\w+) '
'stat=(?P<stat>\d+) core=(?P<core>\w+)', content)
finally:
fd.close()
if not matches:
raise ResultsParseException("No parseable results were found in the "
"exec log - `%s'."% exec_log)
for match in matches:
if ((match.group('exit'), match.group('stat'), match.group('core')) !=
target_vals):
failed.append(match.group('tag'))
elif 2 <= verbose:
passed.append(match.group('tag'))
else:
passed += 1
# Save memory on large files because lists can eat up a fair amount of
# memory.
matches = None
if 1 <= verbose:
context = { }
search_tags = failed[:]
if 2 <= verbose:
search_tags += passed
search_tags.sort()
fd = open(output_log, 'r')
try:
try:
lines = fd.readlines()
finally:
fd.close()
fd.close()
end_output = '<<<execution_status>>>'
output_start = '<<<test_output>>>'
tag_re = re.compile('tag=(\w+)')
grab_output = False
i = 0
local_context = ''
line_len = len(lines)
search_tag = None
while i < line_len:
if lines[i].startswith(end_output):
if search_tag:
context[search_tag] = local_context
grab_output = False
local_context = ''
search_tag = None
if not search_tag:
while i < len(lines):
match = tag_re.match(lines[i])
if match and match.group(1) in search_tags:
search_tag = match.group(1)
break
i += 1
elif lines[i].startswith(output_start):
grab_output = True
elif grab_output:
local_context += lines[i]
i += 1
for k in context.keys():
if k not in search_tags:
raise ResultsParseException('Leftover token in search '
'keys: %s' % k)
except Exception, exc:
# XXX (garrcoop): change from Exception to soft error and print
# out warning with logging module.
raise ResultsParseException('Encountered exception reading output '
'for context: %s' % str(exc))
return failed, passed, context
def determine_context(output_log, testsuite, test_set, context):
""" Return a set of context values mapping test_set -> context. """
test_set_context = {}
for test in test_set:
if test in context:
test_context = context[test]
del context[test]
else:
test_context = ('Could not determine context for %s; please see '
'output log - %s' % (test, output_log))
test_set_context['%s : %s' % (testsuite, test)] = test_context
return test_set_context
def print_context(output_dest, header, testsuite_context):
""" Print out testsuite_context to output_dest, heading it up with
header.
"""
output_dest.write('\n'.join(['', '=' * 40, header, '-' * 40, '']))
for test, context in testsuite_context.items():
output_dest.write('<output test="%s">\n%s\n</output>\n' %
(test, context.strip()))
def main():
""" main. """
parser = OptionParser(prog=os.path.basename(sys.argv[0]),
usage='usage: %prog [options] test ...',
version='0.0.1')
ltpdir = os.getenv('LTPROOT', '@prefix@')
parser.add_option('-l', '--ltp-dir', dest='ltp_dir',
default=ltpdir, help='LTP directory [default: %default]')
parser.add_option('-L', '--log-dir', dest='log_dir',
default=None,
help=('directory for [storing and] retrieving logs '
'[default: %s/output]' % ltpdir),
metavar='DIR')
parser.add_option('-p', '--postprocess-only', dest='postprocess_only',
default=False, action='store_true',
help=("Don't execute runltp; just postprocess logs "
"[default: %default]."))
parser.add_option('-o', '--output-file', dest='output_file',
default=None,
help='File to output results')
parser.add_option('-r', '--runltp-opts', dest='runltp_opts',
default='',
help=('options to pass directly to runltp (will '
'suppress -q).'))
group = OptionGroup(parser, 'Logging',
'If --summary-mode is 0, then the summary output is '
'suppressed. '
'If --summary-mode is 1 [the default], then summary '
'output will be displayed for test execution'
'If --summary-mode is 2, then summary output will be '
'provided on a per-test suite basis. If only '
'one test suite is specified, this has the same net '
"effect as `--summary-mode 1'"
'If --verbose is specified once, prints out failed '
'test information with additional context. '
'If --verbose is specified twice, prints out the '
'failed and passed test context, as well as the '
'summary.')
parser.add_option('-s', '--summary-mode', dest='summary_mode', default=1,
type='int',
help='See Logging.')
parser.add_option('-v', '--verbose', dest='verbose', default=0,
action='count',
help=('Increases context verbosity from tests. See '
'Verbosity for more details.'))
parser.add_option_group(group)
group = OptionGroup(parser, 'Copyright',
'%(prog)s version %(version)s, Copyright (C) 2009, '
'Garrett Cooper %(prog)s comes with ABSOLUTELY NO '
'WARRANTY; '
'This is free software, and you are welcome to '
'redistribute it under certain conditions (See the '
'license tort in %(file)s for more details).'
% { 'file' : os.path.abspath(__file__),
'prog' : parser.prog,
'version' : parser.version })
parser.add_option_group(group)
opts, args = parser.parse_args()
# Remove -q from the opts string, as long as it's a standalone option.
runltp_opts = re.sub('^((?<!\S)+\-q\s+|\-q|\s+\-q(?!\S))$', '',
opts.runltp_opts)
if not opts.log_dir:
opts.log_dir = os.path.join(opts.ltp_dir, 'output')
if not opts.summary_mode and not opts.verbose:
parser.error('You cannot suppress summary output and disable '
'verbosity.')
elif opts.summary_mode not in range(3):
parser.error('--summary-mode must be a value between 0 and 2.')
if len(args) == 0:
# This matches the default test suite list in runltp when -f isn't
# specified. Look for `SCENFILES'.
args = [ 'syscalls', 'fs', 'fsx', 'dio', 'io', 'mm', 'ipc', 'sched',
'math', 'nptl', 'pty', 'containers', 'fs_bind', 'controllers',
'filecaps', 'cap_bounds', 'fcntl-locktests', 'connectors',
'admin_tools', 'timers', 'power_management_tests', 'numa',
'hugetlb', 'commands', 'hyperthreading' ]
if opts.output_file:
output_dir = os.path.dirname(opts.output_file)
if output_dir:
# Not cwd; let's check to make sure that the directory does or
# does not exist.
if not os.path.exists(output_dir):
# We need to make the directory.
os.makedirs(os.path.dirname(opts.output_file))
elif not os.path.isdir(os.path.abspath(output_dir)):
# Path exists, but isn't a file. Oops!
parser.error('Dirname for path specified - %s - is not valid'
% output_dir)
else:
# Current path (cwd)
opts.output_file = os.path.join(os.getcwd(), opts.output_file)
output_dest = open(opts.output_file, 'w')
else:
output_dest = sys.stdout
try:
failed_context = {}
passed_context = {}
failed_count = 0
passed_count = 0
if opts.summary_mode == 2 and len(args) == 1:
opts.summary_mode = 1
for testsuite in args:
# Iterate over the provided test list
context = {}
exec_log = os.path.join(opts.log_dir, '%s-exec.log' % testsuite)
output_log = os.path.join(opts.log_dir, ('%s-output.log'
% testsuite))
failed_subset = {}
runtest_file = os.path.join(opts.ltp_dir, 'runtest', testsuite)
if not opts.postprocess_only:
for log in [ exec_log, output_log ]:
if os.path.isfile(log):
os.remove(log)
if not os.access(runtest_file, os.R_OK):
output_dest.write("%s doesn't exist; skipping "
"test\n" % runtest_file)
continue
os.system(' '.join([ os.path.join(opts.ltp_dir, 'runltp'),
runltp_opts, '-f', testsuite,
'-l', exec_log, '-o', output_log ]))
try:
failed_subset, passed_css, context = \
parse_ltp_results(exec_log, output_log,
verbose=opts.verbose)
except ResultsParseException, rpe:
output_dest.write('Error encountered when parsing results for '
'test - %s: %s\n' % (testsuite, str(rpe)))
continue
failed_count += len(failed_subset)
failed_subset_context = {}
passed_subset_context = {}
if opts.verbose:
failed_subset_context = determine_context(output_log,
testsuite,
failed_subset,
context)
if type(passed_css) == list:
passed_count += len(passed_css)
if opts.verbose == 2:
passed_subset_context = determine_context(output_log,
testsuite,
passed_css,
context)
else:
passed_count += passed_css
if opts.summary_mode == 1:
failed_context.update(failed_subset_context)
passed_context.update(passed_subset_context)
else:
if 1 <= opts.verbose:
# Print out failed testcases.
print_context(output_dest,
'FAILED TESTCASES for %s' % testsuite,
failed_subset_context)
if opts.verbose == 2:
# Print out passed testcases with context.
print_context(output_dest,
'PASSED TESTCASES for %s' % testsuite,
passed_subset_context)
if opts.summary_mode == 2:
output_dest.write("""
========================================
SUMMARY for: %s
----------------------------------------
PASS - %d
FAIL - %d
----------------------------------------
""" % (testsuite, passed_count, len(failed_subset)))
if opts.summary_mode == 1:
# Print out overall results.
if 1 <= opts.verbose:
# Print out failed testcases with context.
print_context(output_dest, "FAILED TESTCASES", failed_context)
if opts.verbose == 2:
# Print out passed testcases with context.
print_context(output_dest, "PASSED TESTCASES", passed_context)
output_dest.write("""
========================================
SUMMARY for tests:
%s
----------------------------------------
PASS - %d
FAIL - %d
----------------------------------------
""" % (' '.join(args), passed_count, failed_count))
finally:
if output_dest != sys.stdout:
output_dest.close()
if __name__ == '__main__':
main()