blob: 9daf188a5ea681235f1e208a145a2237aa25cd4d [file] [log] [blame]
# Copyright 2007 Google Inc. Released under the GPL v2
import re, os, sys, traceback, subprocess, tempfile, time, pickle, glob
from autotest_lib.server import installable_object, utils
from autotest_lib.client.common_lib import log, error, debug
from autotest_lib.client.common_lib import global_config, packages
AUTOTEST_SVN = 'svn://test.kernel.org/autotest/trunk/client'
AUTOTEST_HTTP = 'http://test.kernel.org/svn/autotest/trunk/client'
# Timeouts for powering down and up respectively
HALT_TIME = 300
BOOT_TIME = 1800
CRASH_RECOVERY_TIME = 9000
class BaseAutotest(installable_object.InstallableObject):
"""
This class represents the Autotest program.
Autotest is used to run tests automatically and collect the results.
It also supports profilers.
Implementation details:
This is a leaf class in an abstract class hierarchy, it must
implement the unimplemented methods in parent classes.
"""
def __init__(self, host = None):
self.host = host
self.got = False
self.installed = False
self.serverdir = utils.get_server_dir()
super(BaseAutotest, self).__init__()
self.logger = debug.get_logger(module='server')
install_in_tmpdir = False
@classmethod
def set_install_in_tmpdir(cls, flag):
""" Sets a flag that controls whether or not Autotest should by
default be installed in a "standard" directory (e.g.
/home/autotest, /usr/local/autotest) or a temporary directory. """
cls.install_in_tmpdir = flag
def _get_install_dir(self, host):
""" Determines the location where autotest should be installed on
host. If self.install_in_tmpdir is set, it will return a unique
temporary directory that autotest can be installed in. """
try:
autodir = _get_autodir(host)
except error.AutotestRunError:
autodir = '/usr/local/autotest'
if self.install_in_tmpdir:
autodir = host.get_tmp_dir(parent=autodir)
return autodir
@log.record
def install(self, host=None, autodir=None):
self._install(host=host, autodir=autodir)
def install_base(self, host=None, autodir=None):
""" Performs a lightweight autotest install. Useful for when you
want to run some client-side code but don't want to pay the cost
of a full installation. """
self._install(host=host, autodir=autodir, lightweight=True)
def _install(self, host=None, autodir=None, lightweight=False):
"""
Install autotest. If get() was not called previously, an
attempt will be made to install from the autotest svn
repository.
Args:
host: a Host instance on which autotest will be installed
autodir: location on the remote host to install to
lightweight: exclude tests, deps and profilers, if possible
Raises:
AutoservError: if a tarball was not specified and
the target host does not have svn installed in its path"""
if not host:
host = self.host
if not self.got:
self.get()
host.wait_up(timeout=30)
host.setup()
print "Installing autotest on %s" % host.hostname
# set up the autotest directory on the remote machine
if not autodir:
autodir = self._get_install_dir(host)
host.set_autodir(autodir)
host.run('mkdir -p "%s"' % utils.sh_escape(autodir))
# Fetch the autotest client from the nearest repository
try:
c = global_config.global_config
repos = c.get_config_value("PACKAGES", 'fetch_location', type=list)
pkgmgr = packages.PackageManager(autodir, hostname=host.hostname,
repo_urls=repos,
do_locking=False,
run_function=host.run,
run_function_dargs=dict(timeout=600))
# The packages dir is used to store all the packages that
# are fetched on that client. (for the tests,deps etc.
# too apart from the client)
pkg_dir = os.path.join(autodir, 'packages')
# clean up the autodir except for the packages directory
host.run('cd %s && ls | grep -v "^packages$"'
' | xargs rm -rf && rm -rf .[^.]*' % autodir)
pkgmgr.install_pkg('autotest', 'client', pkg_dir, autodir,
preserve_install_dir=True)
self.installed = True
return
except global_config.ConfigError, e:
print ("Could not install autotest using the"
" packaging system %s" % e)
except (packages.PackageInstallError, error.AutoservRunError), e:
print "Could not install autotest from %s : %s " % (repos, e)
# try to install from file or directory
if self.source_material:
if os.path.isdir(self.source_material):
# Copy autotest recursively
if lightweight:
dirs_to_exclude = set(["tests", "site_tests", "deps",
"tools", "profilers"])
light_files = [os.path.join(self.source_material, f)
for f in os.listdir(self.source_material)
if f not in dirs_to_exclude]
host.send_file(light_files, autodir, delete_dest=True)
# create empty dirs for all the stuff we excluded
commands = []
for path in dirs_to_exclude:
abs_path = os.path.join(autodir, path)
abs_path = utils.sh_escape(abs_path)
commands.append("mkdir -p '%s'" % abs_path)
host.run(';'.join(commands))
else:
host.send_file(self.source_material, autodir,
delete_dest=True)
else:
# Copy autotest via tarball
e_msg = 'Installation method not yet implemented!'
raise NotImplementedError(e_msg)
print "Installation of autotest completed"
self.installed = True
return
# if that fails try to install using svn
if utils.run('which svn').exit_status:
raise error.AutoservError('svn not found on target machine: %s'
% host.name)
try:
host.run('svn checkout %s %s' % (AUTOTEST_SVN, autodir))
except error.AutoservRunError, e:
host.run('svn checkout %s %s' % (AUTOTEST_HTTP, autodir))
print "Installation of autotest completed"
self.installed = True
def get(self, location = None):
if not location:
location = os.path.join(self.serverdir, '../client')
location = os.path.abspath(location)
# If there's stuff run on our client directory already, it
# can cause problems. Try giving it a quick clean first.
cwd = os.getcwd()
os.chdir(location)
os.system('tools/make_clean')
os.chdir(cwd)
super(BaseAutotest, self).get(location)
self.got = True
def run(self, control_file, results_dir = '.', host = None,
timeout=None, tag=None, parallel_flag=False, background=False):
"""
Run an autotest job on the remote machine.
Args:
control_file: an open file-like-obj of the control file
results_dir: a str path where the results should be stored
on the local filesystem
host: a Host instance on which the control file should
be run
tag: tag name for the client side instance of autotest
parallel_flag: flag set when multiple jobs are run at the
same time
background: indicates that the client should be launched as
a background job; the code calling run will be
responsible for monitoring the client and
collecting the results
Raises:
AutotestRunError: if there is a problem executing
the control file
"""
host = self._get_host_and_setup(host)
results_dir = os.path.abspath(results_dir)
if tag:
results_dir = os.path.join(results_dir, tag)
atrun = _Run(host, results_dir, tag, parallel_flag, background)
self._do_run(control_file, results_dir, host, atrun, timeout)
def _get_host_and_setup(self, host):
if not host:
host = self.host
if not self.installed:
self.install(host)
host.wait_up(timeout=30)
return host
def _do_run(self, control_file, results_dir, host, atrun, timeout):
try:
atrun.verify_machine()
except:
print "Verify failed on %s. Reinstalling autotest" % host.hostname
self.install(host)
atrun.verify_machine()
debug = os.path.join(results_dir, 'debug')
try:
os.makedirs(debug)
except Exception:
pass
delete_file_list = [atrun.remote_control_file,
atrun.remote_control_file + '.state',
atrun.manual_control_file,
atrun.manual_control_file + '.state']
cmd = ';'.join('rm -f ' + control for control in delete_file_list)
host.run(cmd, ignore_status=True)
tmppath = utils.get(control_file)
cfile = "job.default_boot_tag(%r)\n" % host.job.last_boot_tag
cfile += "job.default_test_cleanup(%r)\n" % host.job.run_test_cleanup
# If the packaging system is being used, add the repository list.
try:
c = global_config.global_config
repos = c.get_config_value("PACKAGES", 'fetch_location', type=list)
pkgmgr = packages.PackageManager('autotest', hostname=host.hostname,
repo_urls=repos)
cfile += 'job.add_repository(%s)\n' % pkgmgr.repo_urls
except global_config.ConfigError, e:
pass
cfile += open(tmppath).read()
open(tmppath, "w").write(cfile)
# Create and copy state file to remote_control_file + '.state'
sysinfo_state = {"__sysinfo": host.job.sysinfo.serialize()}
state_file = self._create_state_file(host.job, sysinfo_state)
host.send_file(state_file, atrun.remote_control_file + '.state')
os.remove(state_file)
# Copy control_file to remote_control_file on the host
host.send_file(tmppath, atrun.remote_control_file)
if os.path.abspath(tmppath) != os.path.abspath(control_file):
os.remove(tmppath)
try:
atrun.execute_control(timeout=timeout)
finally:
if not atrun.background:
collector = log_collector(host, atrun.tag, results_dir)
collector.collect_client_job_results()
self._process_client_state_file(host, atrun, results_dir)
def _create_state_file(self, job, state_dict):
""" Create a state file from a dictionary. Returns the path of the
state file. """
fd, path = tempfile.mkstemp(dir=job.tmpdir)
state_file = os.fdopen(fd, "w")
pickle.dump(state_dict, state_file)
state_file.close()
return path
def _process_client_state_file(self, host, atrun, results_dir):
state_file = os.path.basename(atrun.remote_control_file) + ".state"
state_path = os.path.join(results_dir, state_file)
try:
state_dict = pickle.load(open(state_path))
except Exception, e:
msg = "Ignoring error while loading client job state file: %s" % e
self.logger.warning(msg)
state_dict = {}
# clear out the state file
# TODO: stash the file away somewhere useful instead
try:
os.remove(state_path)
except Exception:
pass
msg = "Persistent state variables pulled back from %s: %s"
msg %= (host.hostname, state_dict)
print msg
if "__run_test_cleanup" in state_dict:
if state_dict["__run_test_cleanup"]:
host.job.enable_test_cleanup()
else:
host.job.disable_test_cleanup()
if "__last_boot_tag" in state_dict:
host.job.last_boot_tag = state_dict["__last_boot_tag"]
if "__sysinfo" in state_dict:
host.job.sysinfo.deserialize(state_dict["__sysinfo"])
def run_timed_test(self, test_name, results_dir='.', host=None,
timeout=None, tag=None, *args, **dargs):
"""
Assemble a tiny little control file to just run one test,
and run it as an autotest client-side test
"""
if not host:
host = self.host
if not self.installed:
self.install(host)
opts = ["%s=%s" % (o[0], repr(o[1])) for o in dargs.items()]
cmd = ", ".join([repr(test_name)] + map(repr, args) + opts)
control = "job.run_test(%s)\n" % cmd
self.run(control, results_dir, host, timeout=timeout, tag=tag)
def run_test(self, test_name, results_dir='.', host=None, tag=None,
*args, **dargs):
self.run_timed_test(test_name, results_dir, host, timeout=None,
tag=tag, *args, **dargs)
class _Run(object):
"""
Represents a run of autotest control file. This class maintains
all the state necessary as an autotest control file is executed.
It is not intended to be used directly, rather control files
should be run using the run method in Autotest.
"""
def __init__(self, host, results_dir, tag, parallel_flag, background):
self.host = host
self.results_dir = results_dir
self.env = host.env
self.tag = tag
self.parallel_flag = parallel_flag
self.background = background
self.autodir = _get_autodir(self.host)
control = os.path.join(self.autodir, 'control')
if tag:
control += '.' + tag
self.manual_control_file = control
self.remote_control_file = control + '.autoserv'
self.logger = debug.get_logger(module='server')
def verify_machine(self):
binary = os.path.join(self.autodir, 'bin/autotest')
try:
self.host.run('ls %s > /dev/null 2>&1' % binary)
except:
raise "Autotest does not appear to be installed"
if not self.parallel_flag:
tmpdir = os.path.join(self.autodir, 'tmp')
download = os.path.join(self.autodir, 'tests/download')
self.host.run('umount %s' % tmpdir, ignore_status=True)
self.host.run('umount %s' % download, ignore_status=True)
def get_full_cmd(self, section):
# build up the full command we want to run over the host
cmd = [os.path.join(self.autodir, 'bin/autotest_client')]
if not self.background:
cmd.append('-H autoserv')
if section > 0:
cmd.append('-c')
if self.tag:
cmd.append('-t %s' % self.tag)
if self.host.job.use_external_logging():
cmd.append('-l')
cmd.append(self.remote_control_file)
if self.background:
cmd = ['nohup'] + cmd + ['>/dev/null 2>/dev/null &']
return ' '.join(cmd)
def get_client_log(self, section):
""" Find what the "next" client.log.* file should be and open it. """
debug_dir = os.path.join(self.results_dir, "debug")
client_logs = glob.glob(os.path.join(debug_dir, "client.log.*"))
next_log = os.path.join(debug_dir, "client.log.%d" % len(client_logs))
return open(next_log, "w", 0)
def execute_section(self, section, timeout, stderr_redirector):
print "Executing %s/bin/autotest %s/control phase %d" % \
(self.autodir, self.autodir, section)
full_cmd = self.get_full_cmd(section)
client_log = self.get_client_log(section)
try:
old_resultdir = self.host.job.resultdir
self.host.job.resultdir = self.results_dir
result = self.host.run(full_cmd, ignore_status=True,
timeout=timeout,
stdout_tee=client_log,
stderr_tee=stderr_redirector)
finally:
self.host.job.resultdir = old_resultdir
if result.exit_status == 1:
raise error.AutotestRunError("client job was aborted")
if not self.background and not result.stderr:
raise error.AutotestRunError(
"execute_section: %s failed to return anything\n"
"stdout:%s\n" % (full_cmd, result.stdout))
return stderr_redirector.last_line
def _wait_for_reboot(self):
self.logger.info("Client is rebooting")
self.logger.info("Waiting for client to halt")
if not self.host.wait_down(HALT_TIME):
err = "%s failed to shutdown after %d"
err %= (self.host.hostname, HALT_TIME)
raise error.AutotestRunError(err)
self.logger.info("Client down, waiting for restart")
if not self.host.wait_up(BOOT_TIME):
# since reboot failed
# hardreset the machine once if possible
# before failing this control file
warning = "%s did not come back up, hard resetting"
warning %= self.host.hostname
self.logger.warning(warning)
try:
self.host.hardreset(wait=False)
except (AttributeError, error.AutoservUnsupportedError):
warning = "Hard reset unsupported on %s"
warning %= self.host.hostname
self.logger.warning(warning)
raise error.AutotestRunError("%s failed to boot after %ds" %
(self.host.hostname, BOOT_TIME))
self.host.reboot_followup()
def execute_control(self, timeout=None):
section = 0
start_time = time.time()
logger = client_logger(self.host, self.tag, self.results_dir)
try:
while not timeout or time.time() < start_time + timeout:
if timeout:
section_timeout = start_time + timeout - time.time()
else:
section_timeout = None
last = self.execute_section(section, section_timeout,
logger)
if self.background:
return
section += 1
if re.match(r'^END .*\t----\t----\t.*$', last):
print "Client complete"
return
elif re.match('^\t*GOOD\t----\treboot\.start.*$', last):
try:
self._wait_for_reboot()
except error.AutotestRunError, e:
self.host.job.record("ABORT", None, "reboot", str(e))
self.host.job.record("END ABORT", None, None, str(e))
raise
continue
# if we reach here, something unexpected happened
msg = "Autotest client terminated unexpectedly"
self.host.job.record("END ABORT", None, None, msg)
# give the client machine a chance to recover from a crash
self.host.wait_up(CRASH_RECOVERY_TIME)
msg = ("Aborting - unexpected final status message from "
"client: %s\n") % last
raise error.AutotestRunError(msg)
finally:
logger.close()
# should only get here if we timed out
assert timeout
raise error.AutotestTimeoutError()
def _get_autodir(host):
autodir = host.get_autodir()
if autodir:
return autodir
try:
# There's no clean way to do this. readlink may not exist
cmd = "python -c 'import os,sys; print os.readlink(sys.argv[1])' /etc/autotest.conf 2> /dev/null"
autodir = os.path.dirname(host.run(cmd).stdout)
if autodir:
return autodir
except error.AutoservRunError:
pass
for path in ['/usr/local/autotest', '/home/autotest']:
try:
host.run('ls %s > /dev/null 2>&1' %
os.path.join(path, 'bin/autotest'))
return path
except error.AutoservRunError:
pass
raise error.AutotestRunError("Cannot figure out autotest directory")
class log_collector(object):
def __init__(self, host, client_tag, results_dir):
self.host = host
if not client_tag:
client_tag = "default"
self.client_results_dir = os.path.join(host.get_autodir(), "results",
client_tag)
self.server_results_dir = results_dir
def collect_client_job_results(self):
""" A method that collects all the current results of a running
client job into the results dir. By default does nothing as no
client job is running, but when running a client job you can override
this with something that will actually do something. """
# make an effort to wait for the machine to come up
try:
self.host.wait_up(timeout=30)
except error.AutoservError:
# don't worry about any errors, we'll try and
# get the results anyway
pass
# Copy all dirs in default to results_dir
try:
keyval_path = self._prepare_for_copying_logs()
self.host.get_file(self.client_results_dir + '/',
self.server_results_dir)
self._process_copied_logs(keyval_path)
self._postprocess_copied_logs()
except Exception:
# well, don't stop running just because we couldn't get logs
print "Unexpected error copying test result logs, continuing ..."
traceback.print_exc(file=sys.stdout)
def _prepare_for_copying_logs(self):
server_keyval = os.path.join(self.server_results_dir, 'keyval')
if not os.path.exists(server_keyval):
# Client-side keyval file can be copied directly
return
# Copy client-side keyval to temporary location
suffix = '.keyval_%s' % self.host.hostname
fd, keyval_path = tempfile.mkstemp(suffix)
os.close(fd)
try:
client_keyval = os.path.join(self.client_results_dir, 'keyval')
try:
self.host.get_file(client_keyval, keyval_path)
finally:
# We will squirrel away the client side keyval
# away and move it back when we are done
remote_temp_dir = self.host.get_tmp_dir()
self.temp_keyval_path = os.path.join(remote_temp_dir, "keyval")
self.host.run('mv %s %s' % (client_keyval,
self.temp_keyval_path))
except (error.AutoservRunError, error.AutoservSSHTimeout):
print "Prepare for copying logs failed"
return keyval_path
def _process_copied_logs(self, keyval_path):
if not keyval_path:
# Client-side keyval file was copied directly
return
# Append contents of keyval_<host> file to keyval file
try:
# Read in new and old keyval files
new_keyval = utils.read_keyval(keyval_path)
old_keyval = utils.read_keyval(self.server_results_dir)
# 'Delete' from new keyval entries that are in both
tmp_keyval = {}
for key, val in new_keyval.iteritems():
if key not in old_keyval:
tmp_keyval[key] = val
# Append new info to keyval file
utils.write_keyval(self.server_results_dir, tmp_keyval)
# Delete keyval_<host> file
os.remove(keyval_path)
except IOError:
print "Process copied logs failed"
def _postprocess_copied_logs(self):
# we can now put our keyval file back
client_keyval = os.path.join(self.client_results_dir, 'keyval')
try:
self.host.run('mv %s %s' % (self.temp_keyval_path, client_keyval))
except Exception:
pass
# a file-like object for catching stderr from an autotest client and
# extracting status logs from it
class client_logger(object):
"""Partial file object to write to both stdout and
the status log file. We only implement those methods
utils.run() actually calls.
Note that this class is fairly closely coupled with server_job, as it
uses special job._ methods to actually carry out the loggging.
"""
status_parser = re.compile(r"^AUTOTEST_STATUS:([^:]*):(.*)$")
test_complete_parser = re.compile(r"^AUTOTEST_TEST_COMPLETE:(.*)$")
extract_indent = re.compile(r"^(\t*).*$")
def __init__(self, host, tag, server_results_dir):
self.host = host
self.job = host.job
self.log_collector = log_collector(host, tag, server_results_dir)
self.leftover = ""
self.last_line = ""
self.logs = {}
def _process_log_dict(self, log_dict):
log_list = log_dict.pop("logs", [])
for key in sorted(log_dict.iterkeys()):
log_list += self._process_log_dict(log_dict.pop(key))
return log_list
def _process_logs(self):
"""Go through the accumulated logs in self.log and print them
out to stdout and the status log. Note that this processes
logs in an ordering where:
1) logs to different tags are never interleaved
2) logs to x.y come before logs to x.y.z for all z
3) logs to x.y come before x.z whenever y < z
Note that this will in general not be the same as the
chronological ordering of the logs. However, if a chronological
ordering is desired that one can be reconstructed from the
status log by looking at timestamp lines."""
log_list = self._process_log_dict(self.logs)
for line in log_list:
self.job._record_prerendered(line + '\n')
if log_list:
self.last_line = log_list[-1]
def _process_quoted_line(self, tag, line):
"""Process a line quoted with an AUTOTEST_STATUS flag. If the
tag is blank then we want to push out all the data we've been
building up in self.logs, and then the newest line. If the
tag is not blank, then push the line into the logs for handling
later."""
print line
if tag == "":
self._process_logs()
self.job._record_prerendered(line + '\n')
self.last_line = line
else:
tag_parts = [int(x) for x in tag.split(".")]
log_dict = self.logs
for part in tag_parts:
log_dict = log_dict.setdefault(part, {})
log_list = log_dict.setdefault("logs", [])
log_list.append(line)
def _process_line(self, line):
"""Write out a line of data to the appropriate stream. Status
lines sent by autotest will be prepended with
"AUTOTEST_STATUS", and all other lines are ssh error
messages."""
status_match = self.status_parser.search(line)
test_complete_match = self.test_complete_parser.search(line)
if status_match:
tag, line = status_match.groups()
self._process_quoted_line(tag, line)
elif test_complete_match:
fifo_path, = test_complete_match.groups()
self.log_collector.collect_client_job_results()
self.host.run("echo A > %s" % fifo_path)
else:
print line
def _format_warnings(self, last_line, warnings):
# use the indentation of whatever the last log line was
indent = self.extract_indent.match(last_line).group(1)
# if the last line starts a new group, add an extra indent
if last_line.lstrip('\t').startswith("START\t"):
indent += '\t'
return [self.job._render_record("WARN", None, None, msg,
timestamp, indent).rstrip('\n')
for timestamp, msg in warnings]
def _process_warnings(self, last_line, log_dict, warnings):
if log_dict.keys() in ([], ["logs"]):
# there are no sub-jobs, just append the warnings here
warnings = self._format_warnings(last_line, warnings)
log_list = log_dict.setdefault("logs", [])
log_list += warnings
for warning in warnings:
sys.stdout.write(warning + '\n')
else:
# there are sub-jobs, so put the warnings in there
log_list = log_dict.get("logs", [])
if log_list:
last_line = log_list[-1]
for key in sorted(log_dict.iterkeys()):
if key != "logs":
self._process_warnings(last_line,
log_dict[key],
warnings)
def write(self, data):
# first check for any new console warnings
warnings = self.job._read_warnings()
self._process_warnings(self.last_line, self.logs, warnings)
# now process the newest data written out
data = self.leftover + data
lines = data.split("\n")
# process every line but the last one
for line in lines[:-1]:
self._process_line(line)
# save the last line for later processing
# since we may not have the whole line yet
self.leftover = lines[-1]
def flush(self):
sys.stdout.flush()
def close(self):
if self.leftover:
self._process_line(self.leftover)
self._process_logs()
self.flush()
# site_autotest.py may be non-existant or empty, make sure that an appropriate
# SiteAutotest class is created nevertheless
try:
from site_autotest import SiteAutotest
except ImportError:
class SiteAutotest(BaseAutotest):
pass
class Autotest(SiteAutotest):
pass