| # Copyright 2004-2005 Elemental Security, Inc. All Rights Reserved. |
| # Licensed to PSF under a Contributor Agreement. |
| |
| """Convert graminit.[ch] spit out by pgen to Python code. |
| |
| Pgen is the Python parser generator. It is useful to quickly create a |
| parser from a grammar file in Python's grammar notation. But I don't |
| want my parsers to be written in C (yet), so I'm translating the |
| parsing tables to Python data structures and writing a Python parse |
| engine. |
| |
| Note that the token numbers are constants determined by the standard |
| Python tokenizer. The standard token module defines these numbers and |
| their names (the names are not used much). The token numbers are |
| hardcoded into the Python tokenizer and into pgen. A Python |
| implementation of the Python tokenizer is also available, in the |
| standard tokenize module. |
| |
| On the other hand, symbol numbers (representing the grammar's |
| non-terminals) are assigned by pgen based on the actual grammar |
| input. |
| |
| Note: this module is pretty much obsolete; the pgen module generates |
| equivalent grammar tables directly from the Grammar.txt input file |
| without having to invoke the Python pgen C program. |
| |
| """ |
| |
| # Python imports |
| import re |
| |
| # Local imports |
| from pgen2 import grammar, token |
| |
| |
| class Converter(grammar.Grammar): |
| """Grammar subclass that reads classic pgen output files. |
| |
| The run() method reads the tables as produced by the pgen parser |
| generator, typically contained in two C files, graminit.h and |
| graminit.c. The other methods are for internal use only. |
| |
| See the base class for more documentation. |
| |
| """ |
| |
| def run(self, graminit_h, graminit_c): |
| """Load the grammar tables from the text files written by pgen.""" |
| self.parse_graminit_h(graminit_h) |
| self.parse_graminit_c(graminit_c) |
| self.finish_off() |
| |
| def parse_graminit_h(self, filename): |
| """Parse the .h file written by pgen. (Internal) |
| |
| This file is a sequence of #define statements defining the |
| nonterminals of the grammar as numbers. We build two tables |
| mapping the numbers to names and back. |
| |
| """ |
| try: |
| f = open(filename) |
| except IOError as err: |
| print("Can't open %s: %s" % (filename, err)) |
| return False |
| self.symbol2number = {} |
| self.number2symbol = {} |
| lineno = 0 |
| for line in f: |
| lineno += 1 |
| mo = re.match(r"^#define\s+(\w+)\s+(\d+)$", line) |
| if not mo and line.strip(): |
| print("%s(%s): can't parse %s" % (filename, lineno, |
| line.strip())) |
| else: |
| symbol, number = mo.groups() |
| number = int(number) |
| assert symbol not in self.symbol2number |
| assert number not in self.number2symbol |
| self.symbol2number[symbol] = number |
| self.number2symbol[number] = symbol |
| return True |
| |
| def parse_graminit_c(self, filename): |
| """Parse the .c file written by pgen. (Internal) |
| |
| The file looks as follows. The first two lines are always this: |
| |
| #include "pgenheaders.h" |
| #include "grammar.h" |
| |
| After that come four blocks: |
| |
| 1) one or more state definitions |
| 2) a table defining dfas |
| 3) a table defining labels |
| 4) a struct defining the grammar |
| |
| A state definition has the following form: |
| - one or more arc arrays, each of the form: |
| static arc arcs_<n>_<m>[<k>] = { |
| {<i>, <j>}, |
| ... |
| }; |
| - followed by a state array, of the form: |
| static state states_<s>[<t>] = { |
| {<k>, arcs_<n>_<m>}, |
| ... |
| }; |
| |
| """ |
| try: |
| f = open(filename) |
| except IOError as err: |
| print("Can't open %s: %s" % (filename, err)) |
| return False |
| # The code below essentially uses f's iterator-ness! |
| lineno = 0 |
| |
| # Expect the two #include lines |
| lineno, line = lineno+1, next(f) |
| assert line == '#include "pgenheaders.h"\n', (lineno, line) |
| lineno, line = lineno+1, next(f) |
| assert line == '#include "grammar.h"\n', (lineno, line) |
| |
| # Parse the state definitions |
| lineno, line = lineno+1, next(f) |
| allarcs = {} |
| states = [] |
| while line.startswith("static arc "): |
| while line.startswith("static arc "): |
| mo = re.match(r"static arc arcs_(\d+)_(\d+)\[(\d+)\] = {$", |
| line) |
| assert mo, (lineno, line) |
| n, m, k = list(map(int, mo.groups())) |
| arcs = [] |
| for _ in range(k): |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"\s+{(\d+), (\d+)},$", line) |
| assert mo, (lineno, line) |
| i, j = list(map(int, mo.groups())) |
| arcs.append((i, j)) |
| lineno, line = lineno+1, next(f) |
| assert line == "};\n", (lineno, line) |
| allarcs[(n, m)] = arcs |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"static state states_(\d+)\[(\d+)\] = {$", line) |
| assert mo, (lineno, line) |
| s, t = list(map(int, mo.groups())) |
| assert s == len(states), (lineno, line) |
| state = [] |
| for _ in range(t): |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"\s+{(\d+), arcs_(\d+)_(\d+)},$", line) |
| assert mo, (lineno, line) |
| k, n, m = list(map(int, mo.groups())) |
| arcs = allarcs[n, m] |
| assert k == len(arcs), (lineno, line) |
| state.append(arcs) |
| states.append(state) |
| lineno, line = lineno+1, next(f) |
| assert line == "};\n", (lineno, line) |
| lineno, line = lineno+1, next(f) |
| self.states = states |
| |
| # Parse the dfas |
| dfas = {} |
| mo = re.match(r"static dfa dfas\[(\d+)\] = {$", line) |
| assert mo, (lineno, line) |
| ndfas = int(mo.group(1)) |
| for i in range(ndfas): |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r'\s+{(\d+), "(\w+)", (\d+), (\d+), states_(\d+),$', |
| line) |
| assert mo, (lineno, line) |
| symbol = mo.group(2) |
| number, x, y, z = list(map(int, mo.group(1, 3, 4, 5))) |
| assert self.symbol2number[symbol] == number, (lineno, line) |
| assert self.number2symbol[number] == symbol, (lineno, line) |
| assert x == 0, (lineno, line) |
| state = states[z] |
| assert y == len(state), (lineno, line) |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r'\s+("(?:\\\d\d\d)*")},$', line) |
| assert mo, (lineno, line) |
| first = {} |
| rawbitset = eval(mo.group(1)) |
| for i, c in enumerate(rawbitset): |
| byte = ord(c) |
| for j in range(8): |
| if byte & (1<<j): |
| first[i*8 + j] = 1 |
| dfas[number] = (state, first) |
| lineno, line = lineno+1, next(f) |
| assert line == "};\n", (lineno, line) |
| self.dfas = dfas |
| |
| # Parse the labels |
| labels = [] |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"static label labels\[(\d+)\] = {$", line) |
| assert mo, (lineno, line) |
| nlabels = int(mo.group(1)) |
| for i in range(nlabels): |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r'\s+{(\d+), (0|"\w+")},$', line) |
| assert mo, (lineno, line) |
| x, y = mo.groups() |
| x = int(x) |
| if y == "0": |
| y = None |
| else: |
| y = eval(y) |
| labels.append((x, y)) |
| lineno, line = lineno+1, next(f) |
| assert line == "};\n", (lineno, line) |
| self.labels = labels |
| |
| # Parse the grammar struct |
| lineno, line = lineno+1, next(f) |
| assert line == "grammar _PyParser_Grammar = {\n", (lineno, line) |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"\s+(\d+),$", line) |
| assert mo, (lineno, line) |
| ndfas = int(mo.group(1)) |
| assert ndfas == len(self.dfas) |
| lineno, line = lineno+1, next(f) |
| assert line == "\tdfas,\n", (lineno, line) |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"\s+{(\d+), labels},$", line) |
| assert mo, (lineno, line) |
| nlabels = int(mo.group(1)) |
| assert nlabels == len(self.labels), (lineno, line) |
| lineno, line = lineno+1, next(f) |
| mo = re.match(r"\s+(\d+)$", line) |
| assert mo, (lineno, line) |
| start = int(mo.group(1)) |
| assert start in self.number2symbol, (lineno, line) |
| self.start = start |
| lineno, line = lineno+1, next(f) |
| assert line == "};\n", (lineno, line) |
| try: |
| lineno, line = lineno+1, next(f) |
| except StopIteration: |
| pass |
| else: |
| assert 0, (lineno, line) |
| |
| def finish_off(self): |
| """Create additional useful structures. (Internal).""" |
| self.keywords = {} # map from keyword strings to arc labels |
| self.tokens = {} # map from numeric token values to arc labels |
| for ilabel, (type, value) in enumerate(self.labels): |
| if type == token.NAME and value is not None: |
| self.keywords[value] = ilabel |
| elif value is None: |
| self.tokens[type] = ilabel |