blob: b7ae7d31b0fb7b2c5a706edc395af60907621f28 [file] [log] [blame]
Guido van Rossumb51eaa11997-03-07 00:21:55 +00001"""Tokenization help for Python programs.
Guido van Rossum4d8e8591992-01-01 19:34:47 +00002
Florent Xicluna43e4ea12010-09-03 19:54:02 +00003tokenize(readline) is a generator that breaks a stream of bytes into
4Python tokens. It decodes the bytes according to PEP-0263 for
5determining source file encoding.
Trent Nelson428de652008-03-18 22:41:35 +00006
Florent Xicluna43e4ea12010-09-03 19:54:02 +00007It accepts a readline-like method which is called repeatedly to get the
8next line of input (or b"" for EOF). It generates 5-tuples with these
9members:
Tim Peters4efb6e92001-06-29 23:51:08 +000010
11 the token type (see token.py)
12 the token (a string)
13 the starting (row, column) indices of the token (a 2-tuple of ints)
14 the ending (row, column) indices of the token (a 2-tuple of ints)
15 the original line (string)
16
17It is designed to match the working of the Python tokenizer exactly, except
18that it produces COMMENT tokens for comments and gives type OP for all
Florent Xicluna43e4ea12010-09-03 19:54:02 +000019operators. Additionally, all token lists start with an ENCODING token
20which tells you which encoding was used to decode the bytes stream.
21"""
Guido van Rossumb51eaa11997-03-07 00:21:55 +000022
Ka-Ping Yee244c5932001-03-01 13:56:40 +000023__author__ = 'Ka-Ping Yee <ping@lfw.org>'
Trent Nelson428de652008-03-18 22:41:35 +000024__credits__ = ('GvR, ESR, Tim Peters, Thomas Wouters, Fred Drake, '
25 'Skip Montanaro, Raymond Hettinger, Trent Nelson, '
26 'Michael Foord')
Brett Cannonf3042782011-02-22 03:25:12 +000027import builtins
Florent Xicluna43e4ea12010-09-03 19:54:02 +000028import re
29import sys
Guido van Rossumfc6f5331997-03-07 00:21:12 +000030from token import *
Benjamin Peterson433f32c2008-12-12 01:25:05 +000031from codecs import lookup, BOM_UTF8
Raymond Hettinger3fb79c72010-09-09 07:15:18 +000032import collections
Victor Stinner58c07522010-11-09 01:08:59 +000033from io import TextIOWrapper
Trent Nelson428de652008-03-18 22:41:35 +000034cookie_re = re.compile("coding[:=]\s*([-\w.]+)")
Guido van Rossum4d8e8591992-01-01 19:34:47 +000035
Skip Montanaro40fc1602001-03-01 04:27:19 +000036import token
Alexander Belopolskyb9d10d02010-11-11 14:07:41 +000037__all__ = token.__all__ + ["COMMENT", "tokenize", "detect_encoding",
38 "NL", "untokenize", "ENCODING", "TokenInfo"]
Skip Montanaro40fc1602001-03-01 04:27:19 +000039del token
40
Guido van Rossum1aec3231997-04-08 14:24:39 +000041COMMENT = N_TOKENS
42tok_name[COMMENT] = 'COMMENT'
Guido van Rossuma90c78b1998-04-03 16:05:38 +000043NL = N_TOKENS + 1
44tok_name[NL] = 'NL'
Trent Nelson428de652008-03-18 22:41:35 +000045ENCODING = N_TOKENS + 2
46tok_name[ENCODING] = 'ENCODING'
47N_TOKENS += 3
Meador Inge00c7f852012-01-19 00:44:45 -060048EXACT_TOKEN_TYPES = {
49 '(': LPAR,
50 ')': RPAR,
51 '[': LSQB,
52 ']': RSQB,
53 ':': COLON,
54 ',': COMMA,
55 ';': SEMI,
56 '+': PLUS,
57 '-': MINUS,
58 '*': STAR,
59 '/': SLASH,
60 '|': VBAR,
61 '&': AMPER,
62 '<': LESS,
63 '>': GREATER,
64 '=': EQUAL,
65 '.': DOT,
66 '%': PERCENT,
67 '{': LBRACE,
68 '}': RBRACE,
69 '==': EQEQUAL,
70 '!=': NOTEQUAL,
71 '<=': LESSEQUAL,
72 '>=': GREATEREQUAL,
73 '~': TILDE,
74 '^': CIRCUMFLEX,
75 '<<': LEFTSHIFT,
76 '>>': RIGHTSHIFT,
77 '**': DOUBLESTAR,
78 '+=': PLUSEQUAL,
79 '-=': MINEQUAL,
80 '*=': STAREQUAL,
81 '/=': SLASHEQUAL,
82 '%=': PERCENTEQUAL,
83 '&=': AMPEREQUAL,
84 '|=': VBAREQUAL,
85 '^=': CIRCUMFLEXEQUAL,
86 '<<=': LEFTSHIFTEQUAL,
87 '>>=': RIGHTSHIFTEQUAL,
88 '**=': DOUBLESTAREQUAL,
89 '//': DOUBLESLASH,
90 '//=': DOUBLESLASHEQUAL,
91 '@': AT
92}
Guido van Rossum1aec3231997-04-08 14:24:39 +000093
Raymond Hettinger3fb79c72010-09-09 07:15:18 +000094class TokenInfo(collections.namedtuple('TokenInfo', 'type string start end line')):
Raymond Hettingeraa17a7f2009-04-29 14:21:25 +000095 def __repr__(self):
Raymond Hettingera0e79402010-09-09 08:29:05 +000096 annotated_type = '%d (%s)' % (self.type, tok_name[self.type])
97 return ('TokenInfo(type=%s, string=%r, start=%r, end=%r, line=%r)' %
98 self._replace(type=annotated_type))
Raymond Hettingeraa17a7f2009-04-29 14:21:25 +000099
Meador Inge00c7f852012-01-19 00:44:45 -0600100 @property
101 def exact_type(self):
102 if self.type == OP and self.string in EXACT_TOKEN_TYPES:
103 return EXACT_TOKEN_TYPES[self.string]
104 else:
105 return self.type
106
Eric S. Raymondb08b2d32001-02-09 11:10:16 +0000107def group(*choices): return '(' + '|'.join(choices) + ')'
Guido van Rossum68468eb2003-02-27 20:14:51 +0000108def any(*choices): return group(*choices) + '*'
109def maybe(*choices): return group(*choices) + '?'
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000110
Antoine Pitroufd036452008-08-19 17:56:33 +0000111# Note: we use unicode matching for names ("\w") but ascii matching for
112# number literals.
Guido van Rossum3b631771997-10-27 20:44:15 +0000113Whitespace = r'[ \f\t]*'
114Comment = r'#[^\r\n]*'
115Ignore = Whitespace + any(r'\\\r?\n' + Whitespace) + maybe(Comment)
Benjamin Peterson33856de2010-08-30 14:41:20 +0000116Name = r'\w+'
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000117
Antoine Pitroufd036452008-08-19 17:56:33 +0000118Hexnumber = r'0[xX][0-9a-fA-F]+'
Georg Brandlfceab5a2008-01-19 20:08:23 +0000119Binnumber = r'0[bB][01]+'
120Octnumber = r'0[oO][0-7]+'
Antoine Pitroufd036452008-08-19 17:56:33 +0000121Decnumber = r'(?:0+|[1-9][0-9]*)'
Guido van Rossumcd16bf62007-06-13 18:07:49 +0000122Intnumber = group(Hexnumber, Binnumber, Octnumber, Decnumber)
Antoine Pitroufd036452008-08-19 17:56:33 +0000123Exponent = r'[eE][-+]?[0-9]+'
124Pointfloat = group(r'[0-9]+\.[0-9]*', r'\.[0-9]+') + maybe(Exponent)
125Expfloat = r'[0-9]+' + Exponent
Guido van Rossum1aec3231997-04-08 14:24:39 +0000126Floatnumber = group(Pointfloat, Expfloat)
Antoine Pitroufd036452008-08-19 17:56:33 +0000127Imagnumber = group(r'[0-9]+[jJ]', Floatnumber + r'[jJ]')
Guido van Rossum1aec3231997-04-08 14:24:39 +0000128Number = group(Imagnumber, Floatnumber, Intnumber)
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000129
Tim Petersde495832000-10-07 05:09:39 +0000130# Tail end of ' string.
131Single = r"[^'\\]*(?:\\.[^'\\]*)*'"
132# Tail end of " string.
133Double = r'[^"\\]*(?:\\.[^"\\]*)*"'
134# Tail end of ''' string.
135Single3 = r"[^'\\]*(?:(?:\\.|'(?!''))[^'\\]*)*'''"
136# Tail end of """ string.
137Double3 = r'[^"\\]*(?:(?:\\.|"(?!""))[^"\\]*)*"""'
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000138Triple = group("[bBuU]?[rR]?'''", '[bBuU]?[rR]?"""')
Tim Petersde495832000-10-07 05:09:39 +0000139# Single-line ' or " string.
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000140String = group(r"[bBuU]?[rR]?'[^\n'\\]*(?:\\.[^\n'\\]*)*'",
141 r'[bBuU]?[rR]?"[^\n"\\]*(?:\\.[^\n"\\]*)*"')
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000142
Tim Petersde495832000-10-07 05:09:39 +0000143# Because of leftmost-then-longest match semantics, be sure to put the
144# longest operators first (e.g., if = came before ==, == would get
145# recognized as two instances of =).
Guido van Rossumb053cd82006-08-24 03:53:23 +0000146Operator = group(r"\*\*=?", r">>=?", r"<<=?", r"!=",
Neal Norwitzc1505362006-12-28 06:47:50 +0000147 r"//=?", r"->",
Tim Petersde495832000-10-07 05:09:39 +0000148 r"[+\-*/%&|^=<>]=?",
149 r"~")
Thomas Wouterse1519a12000-08-24 21:44:52 +0000150
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000151Bracket = '[][(){}]'
Georg Brandldde00282007-03-18 19:01:53 +0000152Special = group(r'\r?\n', r'\.\.\.', r'[:;.,@]')
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000153Funny = group(Operator, Bracket, Special)
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000154
Guido van Rossum3b631771997-10-27 20:44:15 +0000155PlainToken = group(Number, Funny, String, Name)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000156Token = Ignore + PlainToken
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000157
Tim Petersde495832000-10-07 05:09:39 +0000158# First (or only) line of ' or " string.
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000159ContStr = group(r"[bBuU]?[rR]?'[^\n'\\]*(?:\\.[^\n'\\]*)*" +
Ka-Ping Yee1ff08b12001-01-15 22:04:30 +0000160 group("'", r'\\\r?\n'),
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000161 r'[bBuU]?[rR]?"[^\n"\\]*(?:\\.[^\n"\\]*)*' +
Ka-Ping Yee1ff08b12001-01-15 22:04:30 +0000162 group('"', r'\\\r?\n'))
Guido van Rossum3b631771997-10-27 20:44:15 +0000163PseudoExtras = group(r'\\\r?\n', Comment, Triple)
164PseudoToken = Whitespace + group(PseudoExtras, Number, Funny, ContStr, Name)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000165
Benjamin Peterson33856de2010-08-30 14:41:20 +0000166def _compile(expr):
167 return re.compile(expr, re.UNICODE)
168
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200169endpats = {"'": Single, '"': Double,
170 "'''": Single3, '"""': Double3,
171 "r'''": Single3, 'r"""': Double3,
172 "b'''": Single3, 'b"""': Double3,
173 "br'''": Single3, 'br"""': Double3,
174 "R'''": Single3, 'R"""': Double3,
175 "B'''": Single3, 'B"""': Double3,
176 "bR'''": Single3, 'bR"""': Double3,
177 "Br'''": Single3, 'Br"""': Double3,
178 "BR'''": Single3, 'BR"""': Double3,
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000179 "u'''": Single3, 'u"""': Double3,
180 "ur'''": Single3, 'ur"""': Double3,
181 "R'''": Single3, 'R"""': Double3,
182 "U'''": Single3, 'U"""': Double3,
183 "uR'''": Single3, 'uR"""': Double3,
184 "Ur'''": Single3, 'Ur"""': Double3,
185 "UR'''": Single3, 'UR"""': Double3,
186 'r': None, 'R': None, 'b': None, 'B': None,
187 'u': None, 'U': None}
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000188
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000189triple_quoted = {}
190for t in ("'''", '"""',
191 "r'''", 'r"""', "R'''", 'R"""',
Guido van Rossum4fe72f92007-11-12 17:40:10 +0000192 "b'''", 'b"""', "B'''", 'B"""',
193 "br'''", 'br"""', "Br'''", 'Br"""',
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000194 "bR'''", 'bR"""', "BR'''", 'BR"""',
195 "u'''", 'u"""', "U'''", 'U"""',
196 "ur'''", 'ur"""', "Ur'''", 'Ur"""',
197 "uR'''", 'uR"""', "UR'''", 'UR"""'):
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000198 triple_quoted[t] = t
199single_quoted = {}
200for t in ("'", '"',
201 "r'", 'r"', "R'", 'R"',
Guido van Rossum4fe72f92007-11-12 17:40:10 +0000202 "b'", 'b"', "B'", 'B"',
203 "br'", 'br"', "Br'", 'Br"',
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000204 "bR'", 'bR"', "BR'", 'BR"' ,
205 "u'", 'u"', "U'", 'U"',
206 "ur'", 'ur"', "Ur'", 'Ur"',
207 "uR'", 'uR"', "UR'", 'UR"' ):
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000208 single_quoted[t] = t
209
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000210tabsize = 8
Fred Drake9b8d8012000-08-17 04:45:13 +0000211
Ka-Ping Yee28c62bb2001-03-23 05:22:49 +0000212class TokenError(Exception): pass
213
214class StopTokenizing(Exception): pass
Fred Drake9b8d8012000-08-17 04:45:13 +0000215
Tim Peters5ca576e2001-06-18 22:08:13 +0000216
Thomas Wouters89f507f2006-12-13 04:49:30 +0000217class Untokenizer:
218
219 def __init__(self):
220 self.tokens = []
221 self.prev_row = 1
222 self.prev_col = 0
Trent Nelson428de652008-03-18 22:41:35 +0000223 self.encoding = None
Thomas Wouters89f507f2006-12-13 04:49:30 +0000224
225 def add_whitespace(self, start):
226 row, col = start
227 assert row <= self.prev_row
228 col_offset = col - self.prev_col
229 if col_offset:
230 self.tokens.append(" " * col_offset)
231
232 def untokenize(self, iterable):
233 for t in iterable:
234 if len(t) == 2:
235 self.compat(t, iterable)
236 break
237 tok_type, token, start, end, line = t
Trent Nelson428de652008-03-18 22:41:35 +0000238 if tok_type == ENCODING:
239 self.encoding = token
240 continue
Thomas Wouters89f507f2006-12-13 04:49:30 +0000241 self.add_whitespace(start)
242 self.tokens.append(token)
243 self.prev_row, self.prev_col = end
244 if tok_type in (NEWLINE, NL):
245 self.prev_row += 1
246 self.prev_col = 0
247 return "".join(self.tokens)
248
249 def compat(self, token, iterable):
250 startline = False
251 indents = []
252 toks_append = self.tokens.append
253 toknum, tokval = token
Trent Nelson428de652008-03-18 22:41:35 +0000254
Thomas Wouters89f507f2006-12-13 04:49:30 +0000255 if toknum in (NAME, NUMBER):
256 tokval += ' '
257 if toknum in (NEWLINE, NL):
258 startline = True
Christian Heimesba4af492008-03-28 00:55:15 +0000259 prevstring = False
Thomas Wouters89f507f2006-12-13 04:49:30 +0000260 for tok in iterable:
261 toknum, tokval = tok[:2]
Trent Nelson428de652008-03-18 22:41:35 +0000262 if toknum == ENCODING:
263 self.encoding = tokval
264 continue
Thomas Wouters89f507f2006-12-13 04:49:30 +0000265
266 if toknum in (NAME, NUMBER):
267 tokval += ' '
268
Christian Heimesba4af492008-03-28 00:55:15 +0000269 # Insert a space between two consecutive strings
270 if toknum == STRING:
271 if prevstring:
272 tokval = ' ' + tokval
273 prevstring = True
274 else:
275 prevstring = False
276
Thomas Wouters89f507f2006-12-13 04:49:30 +0000277 if toknum == INDENT:
278 indents.append(tokval)
279 continue
280 elif toknum == DEDENT:
281 indents.pop()
282 continue
283 elif toknum in (NEWLINE, NL):
284 startline = True
285 elif startline and indents:
286 toks_append(indents[-1])
287 startline = False
288 toks_append(tokval)
Raymond Hettinger68c04532005-06-10 11:05:19 +0000289
Trent Nelson428de652008-03-18 22:41:35 +0000290
Raymond Hettinger68c04532005-06-10 11:05:19 +0000291def untokenize(iterable):
292 """Transform tokens back into Python source code.
Trent Nelson428de652008-03-18 22:41:35 +0000293 It returns a bytes object, encoded using the ENCODING
294 token, which is the first token sequence output by tokenize.
Raymond Hettinger68c04532005-06-10 11:05:19 +0000295
296 Each element returned by the iterable must be a token sequence
Thomas Wouters89f507f2006-12-13 04:49:30 +0000297 with at least two elements, a token number and token value. If
298 only two tokens are passed, the resulting output is poor.
Raymond Hettinger68c04532005-06-10 11:05:19 +0000299
Thomas Wouters89f507f2006-12-13 04:49:30 +0000300 Round-trip invariant for full input:
301 Untokenized source will match input source exactly
302
303 Round-trip invariant for limited intput:
Trent Nelson428de652008-03-18 22:41:35 +0000304 # Output bytes will tokenize the back to the input
305 t1 = [tok[:2] for tok in tokenize(f.readline)]
Raymond Hettinger68c04532005-06-10 11:05:19 +0000306 newcode = untokenize(t1)
Trent Nelson428de652008-03-18 22:41:35 +0000307 readline = BytesIO(newcode).readline
308 t2 = [tok[:2] for tok in tokenize(readline)]
Raymond Hettinger68c04532005-06-10 11:05:19 +0000309 assert t1 == t2
310 """
Thomas Wouters89f507f2006-12-13 04:49:30 +0000311 ut = Untokenizer()
Trent Nelson428de652008-03-18 22:41:35 +0000312 out = ut.untokenize(iterable)
313 if ut.encoding is not None:
314 out = out.encode(ut.encoding)
315 return out
Raymond Hettinger68c04532005-06-10 11:05:19 +0000316
Trent Nelson428de652008-03-18 22:41:35 +0000317
Benjamin Petersond3afada2009-10-09 21:43:09 +0000318def _get_normal_name(orig_enc):
319 """Imitates get_normal_name in tokenizer.c."""
320 # Only care about the first 12 characters.
321 enc = orig_enc[:12].lower().replace("_", "-")
322 if enc == "utf-8" or enc.startswith("utf-8-"):
323 return "utf-8"
324 if enc in ("latin-1", "iso-8859-1", "iso-latin-1") or \
325 enc.startswith(("latin-1-", "iso-8859-1-", "iso-latin-1-")):
326 return "iso-8859-1"
327 return orig_enc
328
Trent Nelson428de652008-03-18 22:41:35 +0000329def detect_encoding(readline):
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000330 """
Trent Nelson428de652008-03-18 22:41:35 +0000331 The detect_encoding() function is used to detect the encoding that should
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000332 be used to decode a Python source file. It requires one argment, readline,
Trent Nelson428de652008-03-18 22:41:35 +0000333 in the same way as the tokenize() generator.
334
335 It will call readline a maximum of twice, and return the encoding used
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000336 (as a string) and a list of any lines (left as bytes) it has read in.
Trent Nelson428de652008-03-18 22:41:35 +0000337
338 It detects the encoding from the presence of a utf-8 bom or an encoding
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000339 cookie as specified in pep-0263. If both a bom and a cookie are present,
340 but disagree, a SyntaxError will be raised. If the encoding cookie is an
341 invalid charset, raise a SyntaxError. Note that if a utf-8 bom is found,
Benjamin Peterson689a5582010-03-18 22:29:52 +0000342 'utf-8-sig' is returned.
Trent Nelson428de652008-03-18 22:41:35 +0000343
344 If no encoding is specified, then the default of 'utf-8' will be returned.
345 """
Trent Nelson428de652008-03-18 22:41:35 +0000346 bom_found = False
347 encoding = None
Benjamin Peterson689a5582010-03-18 22:29:52 +0000348 default = 'utf-8'
Trent Nelson428de652008-03-18 22:41:35 +0000349 def read_or_stop():
350 try:
351 return readline()
352 except StopIteration:
353 return b''
354
355 def find_cookie(line):
356 try:
357 line_string = line.decode('ascii')
358 except UnicodeDecodeError:
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000359 return None
360
361 matches = cookie_re.findall(line_string)
362 if not matches:
363 return None
Benjamin Petersond3afada2009-10-09 21:43:09 +0000364 encoding = _get_normal_name(matches[0])
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000365 try:
366 codec = lookup(encoding)
367 except LookupError:
368 # This behaviour mimics the Python interpreter
369 raise SyntaxError("unknown encoding: " + encoding)
370
Benjamin Peterson1613ed82010-03-18 22:34:15 +0000371 if bom_found:
372 if codec.name != 'utf-8':
373 # This behaviour mimics the Python interpreter
374 raise SyntaxError('encoding problem: utf-8')
375 encoding += '-sig'
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000376 return encoding
Trent Nelson428de652008-03-18 22:41:35 +0000377
378 first = read_or_stop()
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000379 if first.startswith(BOM_UTF8):
Trent Nelson428de652008-03-18 22:41:35 +0000380 bom_found = True
381 first = first[3:]
Benjamin Peterson689a5582010-03-18 22:29:52 +0000382 default = 'utf-8-sig'
Trent Nelson428de652008-03-18 22:41:35 +0000383 if not first:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000384 return default, []
Trent Nelson428de652008-03-18 22:41:35 +0000385
386 encoding = find_cookie(first)
387 if encoding:
388 return encoding, [first]
389
390 second = read_or_stop()
391 if not second:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000392 return default, [first]
Trent Nelson428de652008-03-18 22:41:35 +0000393
394 encoding = find_cookie(second)
395 if encoding:
396 return encoding, [first, second]
397
Benjamin Peterson689a5582010-03-18 22:29:52 +0000398 return default, [first, second]
Trent Nelson428de652008-03-18 22:41:35 +0000399
400
Victor Stinner58c07522010-11-09 01:08:59 +0000401def open(filename):
402 """Open a file in read only mode using the encoding detected by
403 detect_encoding().
404 """
Brett Cannonf3042782011-02-22 03:25:12 +0000405 buffer = builtins.open(filename, 'rb')
Victor Stinner58c07522010-11-09 01:08:59 +0000406 encoding, lines = detect_encoding(buffer.readline)
407 buffer.seek(0)
408 text = TextIOWrapper(buffer, encoding, line_buffering=True)
409 text.mode = 'r'
410 return text
411
412
Trent Nelson428de652008-03-18 22:41:35 +0000413def tokenize(readline):
414 """
415 The tokenize() generator requires one argment, readline, which
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000416 must be a callable object which provides the same interface as the
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000417 readline() method of built-in file objects. Each call to the function
Trent Nelson428de652008-03-18 22:41:35 +0000418 should return one line of input as bytes. Alternately, readline
Raymond Hettinger68c04532005-06-10 11:05:19 +0000419 can be a callable function terminating with StopIteration:
Trent Nelson428de652008-03-18 22:41:35 +0000420 readline = open(myfile, 'rb').__next__ # Example of alternate readline
Tim Peters8ac14952002-05-23 15:15:30 +0000421
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000422 The generator produces 5-tuples with these members: the token type; the
423 token string; a 2-tuple (srow, scol) of ints specifying the row and
424 column where the token begins in the source; a 2-tuple (erow, ecol) of
425 ints specifying the row and column where the token ends in the source;
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000426 and the line on which the token was found. The line passed is the
Tim Peters8ac14952002-05-23 15:15:30 +0000427 logical line; continuation lines are included.
Trent Nelson428de652008-03-18 22:41:35 +0000428
429 The first token sequence will always be an ENCODING token
430 which tells you which encoding was used to decode the bytes stream.
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000431 """
Benjamin Peterson21db77e2009-11-14 16:27:26 +0000432 # This import is here to avoid problems when the itertools module is not
433 # built yet and tokenize is imported.
Benjamin Peterson81dd8b92009-11-14 18:09:17 +0000434 from itertools import chain, repeat
Trent Nelson428de652008-03-18 22:41:35 +0000435 encoding, consumed = detect_encoding(readline)
Benjamin Peterson81dd8b92009-11-14 18:09:17 +0000436 rl_gen = iter(readline, b"")
437 empty = repeat(b"")
438 return _tokenize(chain(consumed, rl_gen, empty).__next__, encoding)
Trent Nelson428de652008-03-18 22:41:35 +0000439
440
441def _tokenize(readline, encoding):
Guido van Rossum1aec3231997-04-08 14:24:39 +0000442 lnum = parenlev = continued = 0
Benjamin Peterson33856de2010-08-30 14:41:20 +0000443 numchars = '0123456789'
Guido van Rossumde655271997-04-09 17:15:54 +0000444 contstr, needcont = '', 0
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000445 contline = None
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000446 indents = [0]
Guido van Rossum1aec3231997-04-08 14:24:39 +0000447
Trent Nelson428de652008-03-18 22:41:35 +0000448 if encoding is not None:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000449 if encoding == "utf-8-sig":
450 # BOM will already have been stripped.
451 encoding = "utf-8"
Raymond Hettingera48db392009-04-29 00:34:27 +0000452 yield TokenInfo(ENCODING, encoding, (0, 0), (0, 0), '')
Benjamin Peterson0fe14382008-06-05 23:07:42 +0000453 while True: # loop over lines in stream
Raymond Hettinger68c04532005-06-10 11:05:19 +0000454 try:
455 line = readline()
456 except StopIteration:
Trent Nelson428de652008-03-18 22:41:35 +0000457 line = b''
458
459 if encoding is not None:
460 line = line.decode(encoding)
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000461 lnum += 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000462 pos, max = 0, len(line)
463
464 if contstr: # continued string
Guido van Rossumde655271997-04-09 17:15:54 +0000465 if not line:
Collin Winterce36ad82007-08-30 01:19:48 +0000466 raise TokenError("EOF in multi-line string", strstart)
Guido van Rossum3b631771997-10-27 20:44:15 +0000467 endmatch = endprog.match(line)
468 if endmatch:
469 pos = end = endmatch.end(0)
Raymond Hettingera48db392009-04-29 00:34:27 +0000470 yield TokenInfo(STRING, contstr + line[:end],
Thomas Wouters89f507f2006-12-13 04:49:30 +0000471 strstart, (lnum, end), contline + line)
Guido van Rossumde655271997-04-09 17:15:54 +0000472 contstr, needcont = '', 0
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000473 contline = None
Guido van Rossumde655271997-04-09 17:15:54 +0000474 elif needcont and line[-2:] != '\\\n' and line[-3:] != '\\\r\n':
Raymond Hettingera48db392009-04-29 00:34:27 +0000475 yield TokenInfo(ERRORTOKEN, contstr + line,
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000476 strstart, (lnum, len(line)), contline)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000477 contstr = ''
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000478 contline = None
Guido van Rossumde655271997-04-09 17:15:54 +0000479 continue
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000480 else:
481 contstr = contstr + line
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000482 contline = contline + line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000483 continue
484
Guido van Rossum1aec3231997-04-08 14:24:39 +0000485 elif parenlev == 0 and not continued: # new statement
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000486 if not line: break
487 column = 0
Guido van Rossum1aec3231997-04-08 14:24:39 +0000488 while pos < max: # measure leading whitespace
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000489 if line[pos] == ' ':
490 column += 1
491 elif line[pos] == '\t':
492 column = (column//tabsize + 1)*tabsize
493 elif line[pos] == '\f':
494 column = 0
495 else:
496 break
497 pos += 1
498 if pos == max:
499 break
Guido van Rossum1aec3231997-04-08 14:24:39 +0000500
501 if line[pos] in '#\r\n': # skip comments or blank lines
Thomas Wouters89f507f2006-12-13 04:49:30 +0000502 if line[pos] == '#':
503 comment_token = line[pos:].rstrip('\r\n')
504 nl_pos = pos + len(comment_token)
Raymond Hettingera48db392009-04-29 00:34:27 +0000505 yield TokenInfo(COMMENT, comment_token,
Thomas Wouters89f507f2006-12-13 04:49:30 +0000506 (lnum, pos), (lnum, pos + len(comment_token)), line)
Raymond Hettingera48db392009-04-29 00:34:27 +0000507 yield TokenInfo(NL, line[nl_pos:],
Thomas Wouters89f507f2006-12-13 04:49:30 +0000508 (lnum, nl_pos), (lnum, len(line)), line)
509 else:
Raymond Hettingera48db392009-04-29 00:34:27 +0000510 yield TokenInfo((NL, COMMENT)[line[pos] == '#'], line[pos:],
Guido van Rossum1aec3231997-04-08 14:24:39 +0000511 (lnum, pos), (lnum, len(line)), line)
512 continue
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000513
514 if column > indents[-1]: # count indents or dedents
515 indents.append(column)
Raymond Hettingera48db392009-04-29 00:34:27 +0000516 yield TokenInfo(INDENT, line[:pos], (lnum, 0), (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000517 while column < indents[-1]:
Raymond Hettingerda99d1c2005-06-21 07:43:58 +0000518 if column not in indents:
519 raise IndentationError(
Thomas Wouters00ee7ba2006-08-21 19:07:27 +0000520 "unindent does not match any outer indentation level",
521 ("<tokenize>", lnum, pos, line))
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000522 indents = indents[:-1]
Raymond Hettingera48db392009-04-29 00:34:27 +0000523 yield TokenInfo(DEDENT, '', (lnum, pos), (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000524
525 else: # continued statement
Guido van Rossumde655271997-04-09 17:15:54 +0000526 if not line:
Collin Winterce36ad82007-08-30 01:19:48 +0000527 raise TokenError("EOF in multi-line statement", (lnum, 0))
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000528 continued = 0
529
530 while pos < max:
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200531 pseudomatch = _compile(PseudoToken).match(line, pos)
Guido van Rossum3b631771997-10-27 20:44:15 +0000532 if pseudomatch: # scan for tokens
533 start, end = pseudomatch.span(1)
Guido van Rossumde655271997-04-09 17:15:54 +0000534 spos, epos, pos = (lnum, start), (lnum, end), end
Guido van Rossum1aec3231997-04-08 14:24:39 +0000535 token, initial = line[start:end], line[start]
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000536
Georg Brandldde00282007-03-18 19:01:53 +0000537 if (initial in numchars or # ordinary number
538 (initial == '.' and token != '.' and token != '...')):
Raymond Hettingera48db392009-04-29 00:34:27 +0000539 yield TokenInfo(NUMBER, token, spos, epos, line)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000540 elif initial in '\r\n':
Raymond Hettingera48db392009-04-29 00:34:27 +0000541 yield TokenInfo(NL if parenlev > 0 else NEWLINE,
Thomas Wouters89f507f2006-12-13 04:49:30 +0000542 token, spos, epos, line)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000543 elif initial == '#':
Thomas Wouters89f507f2006-12-13 04:49:30 +0000544 assert not token.endswith("\n")
Raymond Hettingera48db392009-04-29 00:34:27 +0000545 yield TokenInfo(COMMENT, token, spos, epos, line)
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000546 elif token in triple_quoted:
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200547 endprog = _compile(endpats[token])
Guido van Rossum3b631771997-10-27 20:44:15 +0000548 endmatch = endprog.match(line, pos)
549 if endmatch: # all on one line
550 pos = endmatch.end(0)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000551 token = line[start:pos]
Raymond Hettingera48db392009-04-29 00:34:27 +0000552 yield TokenInfo(STRING, token, spos, (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000553 else:
Guido van Rossum1aec3231997-04-08 14:24:39 +0000554 strstart = (lnum, start) # multiple lines
555 contstr = line[start:]
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000556 contline = line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000557 break
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000558 elif initial in single_quoted or \
559 token[:2] in single_quoted or \
560 token[:3] in single_quoted:
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000561 if token[-1] == '\n': # continued string
Guido van Rossum1aec3231997-04-08 14:24:39 +0000562 strstart = (lnum, start)
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200563 endprog = _compile(endpats[initial] or
564 endpats[token[1]] or
565 endpats[token[2]])
Guido van Rossumde655271997-04-09 17:15:54 +0000566 contstr, needcont = line[start:], 1
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000567 contline = line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000568 break
569 else: # ordinary string
Raymond Hettingera48db392009-04-29 00:34:27 +0000570 yield TokenInfo(STRING, token, spos, epos, line)
Benjamin Peterson33856de2010-08-30 14:41:20 +0000571 elif initial.isidentifier(): # ordinary name
Raymond Hettingera48db392009-04-29 00:34:27 +0000572 yield TokenInfo(NAME, token, spos, epos, line)
Guido van Rossum3b631771997-10-27 20:44:15 +0000573 elif initial == '\\': # continued stmt
574 continued = 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000575 else:
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000576 if initial in '([{':
577 parenlev += 1
578 elif initial in ')]}':
579 parenlev -= 1
Raymond Hettingera48db392009-04-29 00:34:27 +0000580 yield TokenInfo(OP, token, spos, epos, line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000581 else:
Raymond Hettingera48db392009-04-29 00:34:27 +0000582 yield TokenInfo(ERRORTOKEN, line[pos],
Guido van Rossumde655271997-04-09 17:15:54 +0000583 (lnum, pos), (lnum, pos+1), line)
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000584 pos += 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000585
586 for indent in indents[1:]: # pop remaining indent levels
Raymond Hettingera48db392009-04-29 00:34:27 +0000587 yield TokenInfo(DEDENT, '', (lnum, 0), (lnum, 0), '')
588 yield TokenInfo(ENDMARKER, '', (lnum, 0), (lnum, 0), '')
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000589
Trent Nelson428de652008-03-18 22:41:35 +0000590
591# An undocumented, backwards compatible, API for all the places in the standard
592# library that expect to be able to use tokenize with strings
593def generate_tokens(readline):
594 return _tokenize(readline, None)
Raymond Hettinger6c60d092010-09-09 04:32:39 +0000595
Meador Inge14c0f032011-10-07 08:53:38 -0500596def main():
597 import argparse
598
599 # Helper error handling routines
600 def perror(message):
601 print(message, file=sys.stderr)
602
603 def error(message, filename=None, location=None):
604 if location:
605 args = (filename,) + location + (message,)
606 perror("%s:%d:%d: error: %s" % args)
607 elif filename:
608 perror("%s: error: %s" % (filename, message))
609 else:
610 perror("error: %s" % message)
611 sys.exit(1)
612
613 # Parse the arguments and options
614 parser = argparse.ArgumentParser(prog='python -m tokenize')
615 parser.add_argument(dest='filename', nargs='?',
616 metavar='filename.py',
617 help='the file to tokenize; defaults to stdin')
Meador Inge00c7f852012-01-19 00:44:45 -0600618 parser.add_argument('-e', '--exact', dest='exact', action='store_true',
619 help='display token names using the exact type')
Meador Inge14c0f032011-10-07 08:53:38 -0500620 args = parser.parse_args()
621
622 try:
623 # Tokenize the input
624 if args.filename:
625 filename = args.filename
626 with builtins.open(filename, 'rb') as f:
627 tokens = list(tokenize(f.readline))
628 else:
629 filename = "<stdin>"
630 tokens = _tokenize(sys.stdin.readline, None)
631
632 # Output the tokenization
633 for token in tokens:
Meador Inge00c7f852012-01-19 00:44:45 -0600634 token_type = token.type
635 if args.exact:
636 token_type = token.exact_type
Meador Inge14c0f032011-10-07 08:53:38 -0500637 token_range = "%d,%d-%d,%d:" % (token.start + token.end)
638 print("%-20s%-15s%-15r" %
Meador Inge00c7f852012-01-19 00:44:45 -0600639 (token_range, tok_name[token_type], token.string))
Meador Inge14c0f032011-10-07 08:53:38 -0500640 except IndentationError as err:
641 line, column = err.args[1][1:3]
642 error(err.args[0], filename, (line, column))
643 except TokenError as err:
644 line, column = err.args[1]
645 error(err.args[0], filename, (line, column))
646 except SyntaxError as err:
647 error(err, filename)
648 except IOError as err:
649 error(err)
650 except KeyboardInterrupt:
651 print("interrupted\n")
652 except Exception as err:
653 perror("unexpected error: %s" % err)
654 raise
655
Raymond Hettinger6c60d092010-09-09 04:32:39 +0000656if __name__ == "__main__":
Meador Inge14c0f032011-10-07 08:53:38 -0500657 main()