| #!/usr/bin/python |
| """ |
| Usage: ./cron_scripts/log_distiller.py job_id path_to_logfile |
| If the job_id is a suite it will find all subjobs. |
| You need to change the location of the log it will parse. |
| The job_id needs to be in the afe database. |
| """ |
| import abc |
| import datetime |
| import os |
| import re |
| import pprint |
| import subprocess |
| import sys |
| import time |
| |
| import common |
| from autotest_lib.server import frontend |
| |
| |
| LOGFIE = './logs/scheduler.log.2014-04-17-16.51.47' |
| # logfile name format: scheduler.log.2014-02-14-18.10.56 |
| time_format = '%Y-%m-%d-%H.%M.%S' |
| logfile_regex = r'scheduler.log.([0-9,.,-]+)' |
| logdir = os.path.join('/usr/local/autotest', 'logs') |
| |
| class StateMachineViolation(Exception): |
| pass |
| |
| |
| class LogLineException(Exception): |
| pass |
| |
| |
| def should_process_log(time_str, time_format, cutoff_days=7): |
| """Returns true if the logs was created after cutoff days. |
| |
| @param time_str: A string representing the time. |
| eg: 2014-02-14-18.10.56 |
| @param time_format: A string representing the format of the time string. |
| ref: http://docs.python.org/2/library/datetime.html#strftime-strptime-behavior |
| @param cutoff_days: Int representind the cutoff in days. |
| |
| @return: Returns True if time_str has aged more than cutoff_days. |
| """ |
| log_time = datetime.datetime.strptime(time_str, time_format) |
| now = datetime.datetime.strptime(time.strftime(time_format), time_format) |
| cutoff = now - datetime.timedelta(days=cutoff_days) |
| return log_time < cutoff |
| |
| |
| def apply_regex(regex, line): |
| """Simple regex applicator. |
| |
| @param regex: Regex to apply. |
| @param line: The line to apply regex on. |
| |
| @return: A tuple with the matching groups, if there was a match. |
| """ |
| log_match = re.match(regex, line) |
| if log_match: |
| return log_match.groups() |
| |
| |
| class StateMachineParser(object): |
| """Abstract class that enforces state transition ordering. |
| |
| Classes inheriting from StateMachineParser need to define an |
| expected_transitions dictionary. The SMP will pop 'to' states |
| from the dictionary as they occur, so you cannot same state transitions |
| unless you specify 2 of them. |
| """ |
| __metaclass__ = abc.ABCMeta |
| |
| |
| @abc.abstractmethod |
| def __init__(self): |
| self.visited_states = [] |
| self.expected_transitions = {} |
| |
| |
| def advance_state(self, from_state, to_state): |
| """Checks that a transition is valid. |
| |
| @param from_state: A string representind the state the host is leaving. |
| @param to_state: The state The host is going to, represented as a string. |
| |
| @raises LogLineException: If an invalid state transition was |
| detected. |
| """ |
| # TODO: Updating to the same state is a waste of bw. |
| if from_state and from_state == to_state: |
| return ('Updating to the same state is a waste of BW: %s->%s' % |
| (from_state, to_state)) |
| return |
| |
| if (from_state in self.expected_transitions and |
| to_state in self.expected_transitions[from_state]): |
| self.expected_transitions[from_state].remove(to_state) |
| self.visited_states.append(to_state) |
| else: |
| return (from_state, to_state) |
| |
| |
| class SingleJobHostSMP(StateMachineParser): |
| def __init__(self): |
| self.visited_states = [] |
| self.expected_transitions = { |
| 'Ready': ['Resetting', 'Verifying', 'Pending', 'Provisioning'], |
| 'Resetting': ['Ready', 'Provisioning'], |
| 'Pending': ['Running'], |
| 'Provisioning': ['Repairing'], |
| 'Running': ['Ready'] |
| } |
| |
| |
| def check_transitions(self, hostline): |
| if hostline.line_info['field'] == 'status': |
| self.advance_state(hostline.line_info['state'], |
| hostline.line_info['value']) |
| |
| |
| class SingleJobHqeSMP(StateMachineParser): |
| def __init__(self): |
| self.visited_states = [] |
| self.expected_transitions = { |
| 'Queued': ['Starting', 'Resetting', 'Aborted'], |
| 'Resetting': ['Pending', 'Provisioning'], |
| 'Provisioning': ['Pending', 'Queued', 'Repairing'], |
| 'Pending': ['Starting'], |
| 'Starting': ['Running'], |
| 'Running': ['Gathering', 'Parsing'], |
| 'Gathering': ['Parsing'], |
| 'Parsing': ['Completed', 'Aborted'] |
| } |
| |
| |
| def check_transitions(self, hqeline): |
| invalid_states = self.advance_state( |
| hqeline.line_info['from_state'], hqeline.line_info['to_state']) |
| if not invalid_states: |
| return |
| |
| # Deal with repair. |
| if (invalid_states[0] == 'Queued' and |
| 'Running' in self.visited_states): |
| raise StateMachineViolation('Unrecognized state transition ' |
| '%s->%s, expected transitions are %s' % |
| (invalid_states[0], invalid_states[1], |
| self.expected_transitions)) |
| |
| |
| class LogLine(object): |
| """Line objects. |
| |
| All classes inheriting from LogLine represent a line of some sort. |
| A line is responsible for parsing itself, and invoking an SMP to |
| validate state transitions. A line can be part of several state machines. |
| """ |
| line_format = '%s' |
| |
| |
| def __init__(self, state_machine_parsers): |
| """ |
| @param state_machine_parsers: A list of smp objects to use to validate |
| state changes on these types of lines.. |
| """ |
| self.smps = state_machine_parsers |
| |
| # Because, this is easier to flush. |
| self.line_info = {} |
| |
| |
| def parse_line(self, line): |
| """Apply a line regex and save any information the parsed line contains. |
| |
| @param line: A string representing a line. |
| """ |
| # Regex for all the things. |
| line_rgx = '(.*)' |
| parsed_line = apply_regex(line_rgx, line) |
| if parsed_line: |
| self.line_info['line'] = parsed_line[0] |
| |
| |
| def flush(self): |
| """Call any state machine parsers, persist line info if needed. |
| """ |
| for smp in self.smps: |
| smp.check_transitions(self) |
| # TODO: persist this? |
| self.line_info={} |
| |
| |
| def format_line(self): |
| try: |
| return self.line_format % self.line_info |
| except KeyError: |
| return self.line_info['line'] |
| |
| |
| class TimeLine(LogLine): |
| """Filters timestamps for scheduler logs. |
| """ |
| |
| def parse_line(self, line): |
| super(TimeLine, self).parse_line(line) |
| |
| # Regex for isolating the date and time from scheduler logs, eg: |
| # 02/16 16:04:36.573 INFO |scheduler_:0574|... |
| line_rgx = '([0-9,/,:,., ]+)(.*)' |
| parsed_line = apply_regex(line_rgx, self.line_info['line']) |
| if parsed_line: |
| self.line_info['time'] = parsed_line[0] |
| self.line_info['line'] = parsed_line[1] |
| |
| |
| class HostLine(TimeLine): |
| """Manages hosts line parsing. |
| """ |
| line_format = (' \t\t %(time)s %(host)s, currently in %(state)s, ' |
| 'updated %(field)s->%(value)s') |
| |
| |
| def record_state_transition(self, line): |
| """Apply the state_transition_rgx to a line and record state changes. |
| |
| @param line: The line we're expecting to contain a state transition. |
| """ |
| state_transition_rgx = ".* ([a-zA-Z]+) updating {'([a-zA-Z]+)': ('[a-zA-Z]+'|[0-9])}.*" |
| match = apply_regex(state_transition_rgx, line) |
| if match: |
| self.line_info['state'] = match[0] |
| self.line_info['field'] = match[1] |
| self.line_info['value'] = match[2].replace("'", "") |
| |
| |
| def parse_line(self, line): |
| super(HostLine, self).parse_line(line) |
| |
| # Regex for getting host status. Eg: |
| # 172.22.4 in Running updating {'status': 'Running'} |
| line_rgx = '.*Host (([0-9,.,a-z,-]+).*)' |
| parsed_line = apply_regex(line_rgx, self.line_info['line']) |
| if parsed_line: |
| self.line_info['line'] = parsed_line[0] |
| self.line_info['host'] = parsed_line[1] |
| self.record_state_transition(self.line_info['line']) |
| return self.format_line() |
| |
| |
| class HQELine(TimeLine): |
| """Manages HQE line parsing. |
| """ |
| line_format = ('%(time)s %(hqe)s, currently in %(from_state)s, ' |
| 'updated to %(to_state)s. Flags: %(flags)s') |
| |
| |
| def record_state_transition(self, line): |
| """Apply the state_transition_rgx to a line and record state changes. |
| |
| @param line: The line we're expecting to contain a state transition. |
| """ |
| # Regex for getting hqe status. Eg: |
| # status:Running [active] -> Gathering |
| state_transition_rgx = ".*status:([a-zA-Z]+)( \[[a-z\,]+\])? -> ([a-zA-Z]+)" |
| match = apply_regex(state_transition_rgx, line) |
| if match: |
| self.line_info['from_state'] = match[0] |
| self.line_info['flags'] = match[1] |
| self.line_info['to_state'] = match[2] |
| |
| |
| def parse_line(self, line): |
| super(HQELine, self).parse_line(line) |
| line_rgx = r'.*\| HQE: (([0-9]+).*)' |
| parsed_line = apply_regex(line_rgx, self.line_info['line']) |
| if parsed_line: |
| self.line_info['line'] = parsed_line[0] |
| self.line_info['hqe'] = parsed_line[1] |
| self.record_state_transition(self.line_info['line']) |
| return self.format_line() |
| |
| |
| class LogCrawler(object): |
| """Crawl logs. |
| |
| Log crawlers are meant to apply some basic preprocessing to a log, and crawl |
| the output validating state changes. They manage line and state machine |
| creation. The initial filtering applied to the log needs to be grab all lines |
| that match an action, such as the running of a job. |
| """ |
| |
| def __init__(self, log_name): |
| self.log = log_name |
| self.filter_command = 'cat %s' % log_name |
| |
| |
| def preprocess_log(self): |
| """Apply some basic filtering to the log. |
| """ |
| proc = subprocess.Popen(self.filter_command, |
| shell=True, stdout=subprocess.PIPE) |
| out, err = proc.communicate() |
| return out |
| |
| |
| class SchedulerLogCrawler(LogCrawler): |
| """A log crawler for the scheduler logs. |
| |
| This crawler is only capable of processing information about a single job. |
| """ |
| |
| def __init__(self, log_name, **kwargs): |
| super(SchedulerLogCrawler, self).__init__(log_name) |
| self.job_id = kwargs['job_id'] |
| self.line_processors = [HostLine([SingleJobHostSMP()]), |
| HQELine([SingleJobHqeSMP()])] |
| self.filter_command = ('%s | grep "for job: %s"' % |
| (self.filter_command, self.job_id)) |
| |
| |
| def parse_log(self): |
| """Parse each line of the preprocessed log output. |
| |
| Pass each line through each possible line_processor. The one that matches |
| will populate itself, call flush, this will walk the state machine of that |
| line to the next step. |
| """ |
| out = self.preprocess_log() |
| response = [] |
| for job_line in out.split('\n'): |
| parsed_line = None |
| for processor in self.line_processors: |
| line = processor.parse_line(job_line) |
| if line and parsed_line: |
| raise LogLineException('Multiple Parsers claiming the line %s: ' |
| 'previous parsing: %s, current parsing: %s ' % |
| (job_line, parsed_line, line)) |
| elif line: |
| parsed_line = line |
| try: |
| processor.flush() |
| except StateMachineViolation as e: |
| response.append(str(e)) |
| raise StateMachineViolation(response) |
| response.append(parsed_line if parsed_line else job_line) |
| return response |
| |
| |
| def process_logs(): |
| if len(sys.argv) < 2: |
| print ('Usage: ./cron_scripts/log_distiller.py 0 8415620 ' |
| 'You need to change the location of the log it will parse.' |
| 'The job_id needs to be in the afe database.') |
| sys.exit(1) |
| |
| job_id = int(sys.argv[1]) |
| rpc = frontend.AFE() |
| suite_jobs = rpc.run('get_jobs', id=job_id) |
| if not suite_jobs[0]['parent_job']: |
| suite_jobs = rpc.run('get_jobs', parent_job=job_id) |
| try: |
| logfile = sys.argv[2] |
| except Exception: |
| logfile = LOGFILE |
| |
| for job in suite_jobs: |
| log_crawler = SchedulerLogCrawler(logfile, job_id=job['id']) |
| for line in log_crawler.parse_log(): |
| print line |
| return |
| |
| |
| if __name__ == '__main__': |
| process_logs() |