| import os, time, pickle, logging |
| |
| from autotest_lib.server import utils, profiler |
| |
| |
| # import any site hooks for the crashdump and crashinfo collection |
| get_site_crashdumps = utils.import_site_function( |
| __file__, "autotest_lib.server.site_crashcollect", "get_site_crashdumps", |
| lambda host, test_start_time: None) |
| get_site_crashinfo = utils.import_site_function( |
| __file__, "autotest_lib.server.site_crashcollect", "get_site_crashinfo", |
| lambda host, test_start_time: None) |
| |
| |
| def get_crashdumps(host, test_start_time): |
| get_site_crashdumps(host, test_start_time) |
| |
| |
| def get_crashinfo(host, test_start_time): |
| logging.info("Collecting crash information...") |
| |
| # include crashdumps as part of the general crashinfo |
| get_crashdumps(host, test_start_time) |
| |
| if wait_for_machine_to_recover(host): |
| # run any site-specific collection |
| get_site_crashinfo(host, test_start_time) |
| |
| crashinfo_dir = get_crashinfo_dir(host) |
| collect_log_file(host, "/var/log/messages", crashinfo_dir) |
| collect_log_file(host, "/var/log/monitor-ssh-reboots", crashinfo_dir) |
| collect_command(host, "dmesg", os.path.join(crashinfo_dir, "dmesg")) |
| collect_profiler_data(host, crashinfo_dir) |
| collect_uncollected_logs(host) |
| |
| |
| def wait_for_machine_to_recover(host, hours_to_wait=4.0): |
| """Wait for a machine (possibly down) to become accessible again. |
| |
| @param host: A RemoteHost instance to wait on |
| @param hours_to_wait: Number of hours to wait before giving up |
| |
| @returns: True if the machine comes back up, False otherwise |
| """ |
| current_time = time.strftime("%b %d %H:%M:%S", time.localtime()) |
| logging.info("Waiting four hours for %s to come up (%s)", |
| host.hostname, current_time) |
| if not host.wait_up(timeout=hours_to_wait * 3600): |
| logging.warning("%s down, unable to collect crash info", |
| host.hostname) |
| return False |
| else: |
| logging.info("%s is back up, collecting crash info", host.hostname) |
| return True |
| |
| |
| def get_crashinfo_dir(host): |
| """Find and if necessary create a directory to store crashinfo in. |
| |
| @param host: The RemoteHost object that crashinfo will be collected from |
| |
| @returns: The path to an existing directory for writing crashinfo into |
| """ |
| host_resultdir = getattr(getattr(host, "job", None), "resultdir", None) |
| if host_resultdir: |
| infodir = host_resultdir |
| else: |
| infodir = os.path.abspath(os.getcwd()) |
| infodir = os.path.join(infodir, "crashinfo.%s" % host.hostname) |
| if not os.path.exists(infodir): |
| os.mkdir(infodir) |
| return infodir |
| |
| |
| def collect_log_file(host, log_path, dest_path): |
| """Collects a log file from the remote machine. |
| |
| Log files are collected from the remote machine and written into the |
| destination path. If dest_path is a directory, the log file will be named |
| using the basename of the remote log path. |
| |
| @param host: The RemoteHost to collect logs from |
| @param log_path: The remote path to collect the log file from |
| @param dest_path: A path (file or directory) to write the copies logs into |
| """ |
| logging.info("Collecting %s...", log_path) |
| try: |
| host.get_file(log_path, dest_path, preserve_perm=False) |
| except Exception: |
| logging.warning("Collection of %s failed", log_path) |
| |
| |
| |
| def collect_command(host, command, dest_path): |
| """Collects the result of a command on the remote machine. |
| |
| The standard output of the command will be collected and written into the |
| desitionation path. The destination path is assumed to be filename and |
| not a directory. |
| |
| @param host: The RemoteHost to collect from |
| @param command: A shell command to run on the remote machine and capture |
| the output from. |
| @param dest_path: A file path to write the results of the log into |
| """ |
| logging.info("Collecting '%s' ...", command) |
| devnull = open("/dev/null", "w") |
| try: |
| try: |
| result = host.run(command, stdout_tee=devnull).stdout |
| utils.open_write_close(dest_path, result) |
| except Exception, e: |
| logging.warning("Collection of '%s' failed:\n%s", command, e) |
| finally: |
| devnull.close() |
| |
| |
| def collect_profiler_data(host, dest_path): |
| """Collects any leftover profiler data that can be found. |
| |
| Any profiler data found will be written into a subdirectory of the |
| crashinfo path called "profiler.$REMOTEDIR" where $REMOTEDIR is the |
| basename of the remote profiler data path. |
| |
| @param host: The RemoteHost to collect from |
| @param dest_path: A directory to copy the profiler results into |
| """ |
| logging.info("Collecting any server-side profiler data lying around...") |
| try: |
| cmd = "ls %s" % profiler.PROFILER_TMPDIR |
| profiler_dirs = [path for path in host.run(cmd).stdout.split() |
| if path.startswith("autoserv-")] |
| for profiler_dir in profiler_dirs: |
| remote_path = profiler.get_profiler_results_dir(profiler_dir) |
| remote_exists = host.run("ls %s" % remote_path, |
| ignore_status=True).exit_status == 0 |
| if not remote_exists: |
| continue |
| local_path = os.path.join(dest_path, "profiler." + profiler_dir) |
| os.mkdir(local_path) |
| host.get_file(remote_path + "/", local_path) |
| except Exception, e: |
| logging.warning("Collection of profiler data failed with:\n%s", e) |
| |
| |
| def collect_uncollected_logs(host): |
| """Collects any leftover uncollected logs from the client. |
| |
| @param host: The RemoteHost to collect from |
| """ |
| if not host.job.uncollected_log_file: |
| host.job.uncollected_log_file = '' |
| |
| if host.job and os.path.exists(host.job.uncollected_log_file): |
| try: |
| logs = pickle.load(open(host.job.uncollected_log_file)) |
| for hostname, remote_path, local_path in logs: |
| if hostname == host.hostname: |
| logging.info("Retrieving logs from %s:%s into %s", |
| hostname, remote_path, local_path) |
| host.get_file(remote_path + "/", local_path + "/") |
| except Exception, e: |
| logging.warning("Error while trying to collect stranded " |
| "Autotest client logs: %s", e) |