blob: c05f764dd86e787c5d2b57064594e61a7d6a577d [file] [log] [blame]
Guido van Rossumb51eaa11997-03-07 00:21:55 +00001"""Tokenization help for Python programs.
Guido van Rossum4d8e8591992-01-01 19:34:47 +00002
Florent Xicluna43e4ea12010-09-03 19:54:02 +00003tokenize(readline) is a generator that breaks a stream of bytes into
4Python tokens. It decodes the bytes according to PEP-0263 for
5determining source file encoding.
Trent Nelson428de652008-03-18 22:41:35 +00006
Florent Xicluna43e4ea12010-09-03 19:54:02 +00007It accepts a readline-like method which is called repeatedly to get the
8next line of input (or b"" for EOF). It generates 5-tuples with these
9members:
Tim Peters4efb6e92001-06-29 23:51:08 +000010
11 the token type (see token.py)
12 the token (a string)
13 the starting (row, column) indices of the token (a 2-tuple of ints)
14 the ending (row, column) indices of the token (a 2-tuple of ints)
15 the original line (string)
16
17It is designed to match the working of the Python tokenizer exactly, except
18that it produces COMMENT tokens for comments and gives type OP for all
Florent Xicluna43e4ea12010-09-03 19:54:02 +000019operators. Additionally, all token lists start with an ENCODING token
20which tells you which encoding was used to decode the bytes stream.
21"""
Guido van Rossumb51eaa11997-03-07 00:21:55 +000022
Ka-Ping Yee244c5932001-03-01 13:56:40 +000023__author__ = 'Ka-Ping Yee <ping@lfw.org>'
Trent Nelson428de652008-03-18 22:41:35 +000024__credits__ = ('GvR, ESR, Tim Peters, Thomas Wouters, Fred Drake, '
25 'Skip Montanaro, Raymond Hettinger, Trent Nelson, '
26 'Michael Foord')
Brett Cannonf3042782011-02-22 03:25:12 +000027import builtins
Florent Xicluna43e4ea12010-09-03 19:54:02 +000028import re
29import sys
Guido van Rossumfc6f5331997-03-07 00:21:12 +000030from token import *
Benjamin Peterson433f32c2008-12-12 01:25:05 +000031from codecs import lookup, BOM_UTF8
Raymond Hettinger3fb79c72010-09-09 07:15:18 +000032import collections
Victor Stinner58c07522010-11-09 01:08:59 +000033from io import TextIOWrapper
Trent Nelson428de652008-03-18 22:41:35 +000034cookie_re = re.compile("coding[:=]\s*([-\w.]+)")
Guido van Rossum4d8e8591992-01-01 19:34:47 +000035
Skip Montanaro40fc1602001-03-01 04:27:19 +000036import token
Alexander Belopolskyb9d10d02010-11-11 14:07:41 +000037__all__ = token.__all__ + ["COMMENT", "tokenize", "detect_encoding",
38 "NL", "untokenize", "ENCODING", "TokenInfo"]
Skip Montanaro40fc1602001-03-01 04:27:19 +000039del token
40
Guido van Rossum1aec3231997-04-08 14:24:39 +000041COMMENT = N_TOKENS
42tok_name[COMMENT] = 'COMMENT'
Guido van Rossuma90c78b1998-04-03 16:05:38 +000043NL = N_TOKENS + 1
44tok_name[NL] = 'NL'
Trent Nelson428de652008-03-18 22:41:35 +000045ENCODING = N_TOKENS + 2
46tok_name[ENCODING] = 'ENCODING'
47N_TOKENS += 3
Meador Inge00c7f852012-01-19 00:44:45 -060048EXACT_TOKEN_TYPES = {
49 '(': LPAR,
50 ')': RPAR,
51 '[': LSQB,
52 ']': RSQB,
53 ':': COLON,
54 ',': COMMA,
55 ';': SEMI,
56 '+': PLUS,
57 '-': MINUS,
58 '*': STAR,
59 '/': SLASH,
60 '|': VBAR,
61 '&': AMPER,
62 '<': LESS,
63 '>': GREATER,
64 '=': EQUAL,
65 '.': DOT,
66 '%': PERCENT,
67 '{': LBRACE,
68 '}': RBRACE,
69 '==': EQEQUAL,
70 '!=': NOTEQUAL,
71 '<=': LESSEQUAL,
72 '>=': GREATEREQUAL,
73 '~': TILDE,
74 '^': CIRCUMFLEX,
75 '<<': LEFTSHIFT,
76 '>>': RIGHTSHIFT,
77 '**': DOUBLESTAR,
78 '+=': PLUSEQUAL,
79 '-=': MINEQUAL,
80 '*=': STAREQUAL,
81 '/=': SLASHEQUAL,
82 '%=': PERCENTEQUAL,
83 '&=': AMPEREQUAL,
84 '|=': VBAREQUAL,
85 '^=': CIRCUMFLEXEQUAL,
86 '<<=': LEFTSHIFTEQUAL,
87 '>>=': RIGHTSHIFTEQUAL,
88 '**=': DOUBLESTAREQUAL,
89 '//': DOUBLESLASH,
90 '//=': DOUBLESLASHEQUAL,
91 '@': AT
92}
Guido van Rossum1aec3231997-04-08 14:24:39 +000093
Raymond Hettinger3fb79c72010-09-09 07:15:18 +000094class TokenInfo(collections.namedtuple('TokenInfo', 'type string start end line')):
Raymond Hettingeraa17a7f2009-04-29 14:21:25 +000095 def __repr__(self):
Raymond Hettingera0e79402010-09-09 08:29:05 +000096 annotated_type = '%d (%s)' % (self.type, tok_name[self.type])
97 return ('TokenInfo(type=%s, string=%r, start=%r, end=%r, line=%r)' %
98 self._replace(type=annotated_type))
Raymond Hettingeraa17a7f2009-04-29 14:21:25 +000099
Meador Inge00c7f852012-01-19 00:44:45 -0600100 @property
101 def exact_type(self):
102 if self.type == OP and self.string in EXACT_TOKEN_TYPES:
103 return EXACT_TOKEN_TYPES[self.string]
104 else:
105 return self.type
106
Eric S. Raymondb08b2d32001-02-09 11:10:16 +0000107def group(*choices): return '(' + '|'.join(choices) + ')'
Guido van Rossum68468eb2003-02-27 20:14:51 +0000108def any(*choices): return group(*choices) + '*'
109def maybe(*choices): return group(*choices) + '?'
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000110
Antoine Pitroufd036452008-08-19 17:56:33 +0000111# Note: we use unicode matching for names ("\w") but ascii matching for
112# number literals.
Guido van Rossum3b631771997-10-27 20:44:15 +0000113Whitespace = r'[ \f\t]*'
114Comment = r'#[^\r\n]*'
115Ignore = Whitespace + any(r'\\\r?\n' + Whitespace) + maybe(Comment)
Benjamin Peterson33856de2010-08-30 14:41:20 +0000116Name = r'\w+'
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000117
Antoine Pitroufd036452008-08-19 17:56:33 +0000118Hexnumber = r'0[xX][0-9a-fA-F]+'
Georg Brandlfceab5a2008-01-19 20:08:23 +0000119Binnumber = r'0[bB][01]+'
120Octnumber = r'0[oO][0-7]+'
Antoine Pitroufd036452008-08-19 17:56:33 +0000121Decnumber = r'(?:0+|[1-9][0-9]*)'
Guido van Rossumcd16bf62007-06-13 18:07:49 +0000122Intnumber = group(Hexnumber, Binnumber, Octnumber, Decnumber)
Antoine Pitroufd036452008-08-19 17:56:33 +0000123Exponent = r'[eE][-+]?[0-9]+'
124Pointfloat = group(r'[0-9]+\.[0-9]*', r'\.[0-9]+') + maybe(Exponent)
125Expfloat = r'[0-9]+' + Exponent
Guido van Rossum1aec3231997-04-08 14:24:39 +0000126Floatnumber = group(Pointfloat, Expfloat)
Antoine Pitroufd036452008-08-19 17:56:33 +0000127Imagnumber = group(r'[0-9]+[jJ]', Floatnumber + r'[jJ]')
Guido van Rossum1aec3231997-04-08 14:24:39 +0000128Number = group(Imagnumber, Floatnumber, Intnumber)
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000129
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000130StringPrefix = r'(?:[uU][rR]?|[bB][rR]|[rR][bB]|[rR]|[uU])?'
131
Tim Petersde495832000-10-07 05:09:39 +0000132# Tail end of ' string.
133Single = r"[^'\\]*(?:\\.[^'\\]*)*'"
134# Tail end of " string.
135Double = r'[^"\\]*(?:\\.[^"\\]*)*"'
136# Tail end of ''' string.
137Single3 = r"[^'\\]*(?:(?:\\.|'(?!''))[^'\\]*)*'''"
138# Tail end of """ string.
139Double3 = r'[^"\\]*(?:(?:\\.|"(?!""))[^"\\]*)*"""'
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000140Triple = group(StringPrefix + "'''", StringPrefix + '"""')
Tim Petersde495832000-10-07 05:09:39 +0000141# Single-line ' or " string.
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000142String = group(StringPrefix + r"'[^\n'\\]*(?:\\.[^\n'\\]*)*'",
143 StringPrefix + r'"[^\n"\\]*(?:\\.[^\n"\\]*)*"')
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000144
Tim Petersde495832000-10-07 05:09:39 +0000145# Because of leftmost-then-longest match semantics, be sure to put the
146# longest operators first (e.g., if = came before ==, == would get
147# recognized as two instances of =).
Guido van Rossumb053cd82006-08-24 03:53:23 +0000148Operator = group(r"\*\*=?", r">>=?", r"<<=?", r"!=",
Neal Norwitzc1505362006-12-28 06:47:50 +0000149 r"//=?", r"->",
Tim Petersde495832000-10-07 05:09:39 +0000150 r"[+\-*/%&|^=<>]=?",
151 r"~")
Thomas Wouterse1519a12000-08-24 21:44:52 +0000152
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000153Bracket = '[][(){}]'
Georg Brandldde00282007-03-18 19:01:53 +0000154Special = group(r'\r?\n', r'\.\.\.', r'[:;.,@]')
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000155Funny = group(Operator, Bracket, Special)
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000156
Guido van Rossum3b631771997-10-27 20:44:15 +0000157PlainToken = group(Number, Funny, String, Name)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000158Token = Ignore + PlainToken
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000159
Tim Petersde495832000-10-07 05:09:39 +0000160# First (or only) line of ' or " string.
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000161ContStr = group(StringPrefix + r"'[^\n'\\]*(?:\\.[^\n'\\]*)*" +
Ka-Ping Yee1ff08b12001-01-15 22:04:30 +0000162 group("'", r'\\\r?\n'),
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000163 StringPrefix + r'"[^\n"\\]*(?:\\.[^\n"\\]*)*' +
Ka-Ping Yee1ff08b12001-01-15 22:04:30 +0000164 group('"', r'\\\r?\n'))
Guido van Rossum3b631771997-10-27 20:44:15 +0000165PseudoExtras = group(r'\\\r?\n', Comment, Triple)
166PseudoToken = Whitespace + group(PseudoExtras, Number, Funny, ContStr, Name)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000167
Benjamin Peterson33856de2010-08-30 14:41:20 +0000168def _compile(expr):
169 return re.compile(expr, re.UNICODE)
170
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200171endpats = {"'": Single, '"': Double,
172 "'''": Single3, '"""': Double3,
173 "r'''": Single3, 'r"""': Double3,
174 "b'''": Single3, 'b"""': Double3,
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200175 "R'''": Single3, 'R"""': Double3,
176 "B'''": Single3, 'B"""': Double3,
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000177 "br'''": Single3, 'br"""': Double3,
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200178 "bR'''": Single3, 'bR"""': Double3,
179 "Br'''": Single3, 'Br"""': Double3,
180 "BR'''": Single3, 'BR"""': Double3,
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000181 "rb'''": Single3, 'rb"""': Double3,
182 "Rb'''": Single3, 'Rb"""': Double3,
183 "rB'''": Single3, 'rB"""': Double3,
184 "RB'''": Single3, 'RB"""': Double3,
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000185 "u'''": Single3, 'u"""': Double3,
186 "ur'''": Single3, 'ur"""': Double3,
187 "R'''": Single3, 'R"""': Double3,
188 "U'''": Single3, 'U"""': Double3,
189 "uR'''": Single3, 'uR"""': Double3,
190 "Ur'''": Single3, 'Ur"""': Double3,
191 "UR'''": Single3, 'UR"""': Double3,
192 'r': None, 'R': None, 'b': None, 'B': None,
193 'u': None, 'U': None}
Guido van Rossum4d8e8591992-01-01 19:34:47 +0000194
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000195triple_quoted = {}
196for t in ("'''", '"""',
197 "r'''", 'r"""', "R'''", 'R"""',
Guido van Rossum4fe72f92007-11-12 17:40:10 +0000198 "b'''", 'b"""', "B'''", 'B"""',
199 "br'''", 'br"""', "Br'''", 'Br"""',
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000200 "bR'''", 'bR"""', "BR'''", 'BR"""',
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000201 "rb'''", 'rb"""', "rB'''", 'rB"""',
202 "Rb'''", 'Rb"""', "RB'''", 'RB"""',
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000203 "u'''", 'u"""', "U'''", 'U"""',
204 "ur'''", 'ur"""', "Ur'''", 'Ur"""',
205 "uR'''", 'uR"""', "UR'''", 'UR"""'):
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000206 triple_quoted[t] = t
207single_quoted = {}
208for t in ("'", '"',
209 "r'", 'r"', "R'", 'R"',
Guido van Rossum4fe72f92007-11-12 17:40:10 +0000210 "b'", 'b"', "B'", 'B"',
211 "br'", 'br"', "Br'", 'Br"',
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000212 "bR'", 'bR"', "BR'", 'BR"' ,
Armin Ronacherc0eaeca2012-03-04 13:07:57 +0000213 "rb'", 'rb"', "rB'", 'rB"',
214 "Rb'", 'Rb"', "RB'", 'RB"' ,
Armin Ronacher6ecf77b2012-03-04 12:04:06 +0000215 "u'", 'u"', "U'", 'U"',
216 "ur'", 'ur"', "Ur'", 'Ur"',
217 "uR'", 'uR"', "UR'", 'UR"' ):
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000218 single_quoted[t] = t
219
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000220tabsize = 8
Fred Drake9b8d8012000-08-17 04:45:13 +0000221
Ka-Ping Yee28c62bb2001-03-23 05:22:49 +0000222class TokenError(Exception): pass
223
224class StopTokenizing(Exception): pass
Fred Drake9b8d8012000-08-17 04:45:13 +0000225
Tim Peters5ca576e2001-06-18 22:08:13 +0000226
Thomas Wouters89f507f2006-12-13 04:49:30 +0000227class Untokenizer:
228
229 def __init__(self):
230 self.tokens = []
231 self.prev_row = 1
232 self.prev_col = 0
Trent Nelson428de652008-03-18 22:41:35 +0000233 self.encoding = None
Thomas Wouters89f507f2006-12-13 04:49:30 +0000234
235 def add_whitespace(self, start):
236 row, col = start
237 assert row <= self.prev_row
238 col_offset = col - self.prev_col
239 if col_offset:
240 self.tokens.append(" " * col_offset)
241
242 def untokenize(self, iterable):
243 for t in iterable:
244 if len(t) == 2:
245 self.compat(t, iterable)
246 break
247 tok_type, token, start, end, line = t
Trent Nelson428de652008-03-18 22:41:35 +0000248 if tok_type == ENCODING:
249 self.encoding = token
250 continue
Thomas Wouters89f507f2006-12-13 04:49:30 +0000251 self.add_whitespace(start)
252 self.tokens.append(token)
253 self.prev_row, self.prev_col = end
254 if tok_type in (NEWLINE, NL):
255 self.prev_row += 1
256 self.prev_col = 0
257 return "".join(self.tokens)
258
259 def compat(self, token, iterable):
260 startline = False
261 indents = []
262 toks_append = self.tokens.append
263 toknum, tokval = token
Trent Nelson428de652008-03-18 22:41:35 +0000264
Thomas Wouters89f507f2006-12-13 04:49:30 +0000265 if toknum in (NAME, NUMBER):
266 tokval += ' '
267 if toknum in (NEWLINE, NL):
268 startline = True
Christian Heimesba4af492008-03-28 00:55:15 +0000269 prevstring = False
Thomas Wouters89f507f2006-12-13 04:49:30 +0000270 for tok in iterable:
271 toknum, tokval = tok[:2]
Trent Nelson428de652008-03-18 22:41:35 +0000272 if toknum == ENCODING:
273 self.encoding = tokval
274 continue
Thomas Wouters89f507f2006-12-13 04:49:30 +0000275
276 if toknum in (NAME, NUMBER):
277 tokval += ' '
278
Christian Heimesba4af492008-03-28 00:55:15 +0000279 # Insert a space between two consecutive strings
280 if toknum == STRING:
281 if prevstring:
282 tokval = ' ' + tokval
283 prevstring = True
284 else:
285 prevstring = False
286
Thomas Wouters89f507f2006-12-13 04:49:30 +0000287 if toknum == INDENT:
288 indents.append(tokval)
289 continue
290 elif toknum == DEDENT:
291 indents.pop()
292 continue
293 elif toknum in (NEWLINE, NL):
294 startline = True
295 elif startline and indents:
296 toks_append(indents[-1])
297 startline = False
298 toks_append(tokval)
Raymond Hettinger68c04532005-06-10 11:05:19 +0000299
Trent Nelson428de652008-03-18 22:41:35 +0000300
Raymond Hettinger68c04532005-06-10 11:05:19 +0000301def untokenize(iterable):
302 """Transform tokens back into Python source code.
Trent Nelson428de652008-03-18 22:41:35 +0000303 It returns a bytes object, encoded using the ENCODING
304 token, which is the first token sequence output by tokenize.
Raymond Hettinger68c04532005-06-10 11:05:19 +0000305
306 Each element returned by the iterable must be a token sequence
Thomas Wouters89f507f2006-12-13 04:49:30 +0000307 with at least two elements, a token number and token value. If
308 only two tokens are passed, the resulting output is poor.
Raymond Hettinger68c04532005-06-10 11:05:19 +0000309
Thomas Wouters89f507f2006-12-13 04:49:30 +0000310 Round-trip invariant for full input:
311 Untokenized source will match input source exactly
312
313 Round-trip invariant for limited intput:
Trent Nelson428de652008-03-18 22:41:35 +0000314 # Output bytes will tokenize the back to the input
315 t1 = [tok[:2] for tok in tokenize(f.readline)]
Raymond Hettinger68c04532005-06-10 11:05:19 +0000316 newcode = untokenize(t1)
Trent Nelson428de652008-03-18 22:41:35 +0000317 readline = BytesIO(newcode).readline
318 t2 = [tok[:2] for tok in tokenize(readline)]
Raymond Hettinger68c04532005-06-10 11:05:19 +0000319 assert t1 == t2
320 """
Thomas Wouters89f507f2006-12-13 04:49:30 +0000321 ut = Untokenizer()
Trent Nelson428de652008-03-18 22:41:35 +0000322 out = ut.untokenize(iterable)
323 if ut.encoding is not None:
324 out = out.encode(ut.encoding)
325 return out
Raymond Hettinger68c04532005-06-10 11:05:19 +0000326
Trent Nelson428de652008-03-18 22:41:35 +0000327
Benjamin Petersond3afada2009-10-09 21:43:09 +0000328def _get_normal_name(orig_enc):
329 """Imitates get_normal_name in tokenizer.c."""
330 # Only care about the first 12 characters.
331 enc = orig_enc[:12].lower().replace("_", "-")
332 if enc == "utf-8" or enc.startswith("utf-8-"):
333 return "utf-8"
334 if enc in ("latin-1", "iso-8859-1", "iso-latin-1") or \
335 enc.startswith(("latin-1-", "iso-8859-1-", "iso-latin-1-")):
336 return "iso-8859-1"
337 return orig_enc
338
Trent Nelson428de652008-03-18 22:41:35 +0000339def detect_encoding(readline):
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000340 """
Trent Nelson428de652008-03-18 22:41:35 +0000341 The detect_encoding() function is used to detect the encoding that should
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000342 be used to decode a Python source file. It requires one argment, readline,
Trent Nelson428de652008-03-18 22:41:35 +0000343 in the same way as the tokenize() generator.
344
345 It will call readline a maximum of twice, and return the encoding used
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000346 (as a string) and a list of any lines (left as bytes) it has read in.
Trent Nelson428de652008-03-18 22:41:35 +0000347
348 It detects the encoding from the presence of a utf-8 bom or an encoding
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000349 cookie as specified in pep-0263. If both a bom and a cookie are present,
350 but disagree, a SyntaxError will be raised. If the encoding cookie is an
351 invalid charset, raise a SyntaxError. Note that if a utf-8 bom is found,
Benjamin Peterson689a5582010-03-18 22:29:52 +0000352 'utf-8-sig' is returned.
Trent Nelson428de652008-03-18 22:41:35 +0000353
354 If no encoding is specified, then the default of 'utf-8' will be returned.
355 """
Trent Nelson428de652008-03-18 22:41:35 +0000356 bom_found = False
357 encoding = None
Benjamin Peterson689a5582010-03-18 22:29:52 +0000358 default = 'utf-8'
Trent Nelson428de652008-03-18 22:41:35 +0000359 def read_or_stop():
360 try:
361 return readline()
362 except StopIteration:
363 return b''
364
365 def find_cookie(line):
366 try:
Martin v. Löwis63674f42012-04-20 14:36:47 +0200367 # Decode as UTF-8. Either the line is an encoding declaration,
368 # in which case it should be pure ASCII, or it must be UTF-8
369 # per default encoding.
370 line_string = line.decode('utf-8')
Trent Nelson428de652008-03-18 22:41:35 +0000371 except UnicodeDecodeError:
Martin v. Löwis63674f42012-04-20 14:36:47 +0200372 raise SyntaxError("invalid or missing encoding declaration")
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000373
374 matches = cookie_re.findall(line_string)
375 if not matches:
376 return None
Benjamin Petersond3afada2009-10-09 21:43:09 +0000377 encoding = _get_normal_name(matches[0])
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000378 try:
379 codec = lookup(encoding)
380 except LookupError:
381 # This behaviour mimics the Python interpreter
382 raise SyntaxError("unknown encoding: " + encoding)
383
Benjamin Peterson1613ed82010-03-18 22:34:15 +0000384 if bom_found:
385 if codec.name != 'utf-8':
386 # This behaviour mimics the Python interpreter
387 raise SyntaxError('encoding problem: utf-8')
388 encoding += '-sig'
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000389 return encoding
Trent Nelson428de652008-03-18 22:41:35 +0000390
391 first = read_or_stop()
Benjamin Peterson433f32c2008-12-12 01:25:05 +0000392 if first.startswith(BOM_UTF8):
Trent Nelson428de652008-03-18 22:41:35 +0000393 bom_found = True
394 first = first[3:]
Benjamin Peterson689a5582010-03-18 22:29:52 +0000395 default = 'utf-8-sig'
Trent Nelson428de652008-03-18 22:41:35 +0000396 if not first:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000397 return default, []
Trent Nelson428de652008-03-18 22:41:35 +0000398
399 encoding = find_cookie(first)
400 if encoding:
401 return encoding, [first]
402
403 second = read_or_stop()
404 if not second:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000405 return default, [first]
Trent Nelson428de652008-03-18 22:41:35 +0000406
407 encoding = find_cookie(second)
408 if encoding:
409 return encoding, [first, second]
410
Benjamin Peterson689a5582010-03-18 22:29:52 +0000411 return default, [first, second]
Trent Nelson428de652008-03-18 22:41:35 +0000412
413
Victor Stinner58c07522010-11-09 01:08:59 +0000414def open(filename):
415 """Open a file in read only mode using the encoding detected by
416 detect_encoding().
417 """
Brett Cannonf3042782011-02-22 03:25:12 +0000418 buffer = builtins.open(filename, 'rb')
Victor Stinner58c07522010-11-09 01:08:59 +0000419 encoding, lines = detect_encoding(buffer.readline)
420 buffer.seek(0)
421 text = TextIOWrapper(buffer, encoding, line_buffering=True)
422 text.mode = 'r'
423 return text
424
425
Trent Nelson428de652008-03-18 22:41:35 +0000426def tokenize(readline):
427 """
428 The tokenize() generator requires one argment, readline, which
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000429 must be a callable object which provides the same interface as the
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000430 readline() method of built-in file objects. Each call to the function
Trent Nelson428de652008-03-18 22:41:35 +0000431 should return one line of input as bytes. Alternately, readline
Raymond Hettinger68c04532005-06-10 11:05:19 +0000432 can be a callable function terminating with StopIteration:
Trent Nelson428de652008-03-18 22:41:35 +0000433 readline = open(myfile, 'rb').__next__ # Example of alternate readline
Tim Peters8ac14952002-05-23 15:15:30 +0000434
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000435 The generator produces 5-tuples with these members: the token type; the
436 token string; a 2-tuple (srow, scol) of ints specifying the row and
437 column where the token begins in the source; a 2-tuple (erow, ecol) of
438 ints specifying the row and column where the token ends in the source;
Florent Xicluna43e4ea12010-09-03 19:54:02 +0000439 and the line on which the token was found. The line passed is the
Tim Peters8ac14952002-05-23 15:15:30 +0000440 logical line; continuation lines are included.
Trent Nelson428de652008-03-18 22:41:35 +0000441
442 The first token sequence will always be an ENCODING token
443 which tells you which encoding was used to decode the bytes stream.
Raymond Hettingerd1fa3db2002-05-15 02:56:03 +0000444 """
Benjamin Peterson21db77e2009-11-14 16:27:26 +0000445 # This import is here to avoid problems when the itertools module is not
446 # built yet and tokenize is imported.
Benjamin Peterson81dd8b92009-11-14 18:09:17 +0000447 from itertools import chain, repeat
Trent Nelson428de652008-03-18 22:41:35 +0000448 encoding, consumed = detect_encoding(readline)
Benjamin Peterson81dd8b92009-11-14 18:09:17 +0000449 rl_gen = iter(readline, b"")
450 empty = repeat(b"")
451 return _tokenize(chain(consumed, rl_gen, empty).__next__, encoding)
Trent Nelson428de652008-03-18 22:41:35 +0000452
453
454def _tokenize(readline, encoding):
Guido van Rossum1aec3231997-04-08 14:24:39 +0000455 lnum = parenlev = continued = 0
Benjamin Peterson33856de2010-08-30 14:41:20 +0000456 numchars = '0123456789'
Guido van Rossumde655271997-04-09 17:15:54 +0000457 contstr, needcont = '', 0
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000458 contline = None
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000459 indents = [0]
Guido van Rossum1aec3231997-04-08 14:24:39 +0000460
Trent Nelson428de652008-03-18 22:41:35 +0000461 if encoding is not None:
Benjamin Peterson689a5582010-03-18 22:29:52 +0000462 if encoding == "utf-8-sig":
463 # BOM will already have been stripped.
464 encoding = "utf-8"
Raymond Hettingera48db392009-04-29 00:34:27 +0000465 yield TokenInfo(ENCODING, encoding, (0, 0), (0, 0), '')
Benjamin Peterson0fe14382008-06-05 23:07:42 +0000466 while True: # loop over lines in stream
Raymond Hettinger68c04532005-06-10 11:05:19 +0000467 try:
468 line = readline()
469 except StopIteration:
Trent Nelson428de652008-03-18 22:41:35 +0000470 line = b''
471
472 if encoding is not None:
473 line = line.decode(encoding)
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000474 lnum += 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000475 pos, max = 0, len(line)
476
477 if contstr: # continued string
Guido van Rossumde655271997-04-09 17:15:54 +0000478 if not line:
Collin Winterce36ad82007-08-30 01:19:48 +0000479 raise TokenError("EOF in multi-line string", strstart)
Guido van Rossum3b631771997-10-27 20:44:15 +0000480 endmatch = endprog.match(line)
481 if endmatch:
482 pos = end = endmatch.end(0)
Raymond Hettingera48db392009-04-29 00:34:27 +0000483 yield TokenInfo(STRING, contstr + line[:end],
Thomas Wouters89f507f2006-12-13 04:49:30 +0000484 strstart, (lnum, end), contline + line)
Guido van Rossumde655271997-04-09 17:15:54 +0000485 contstr, needcont = '', 0
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000486 contline = None
Guido van Rossumde655271997-04-09 17:15:54 +0000487 elif needcont and line[-2:] != '\\\n' and line[-3:] != '\\\r\n':
Raymond Hettingera48db392009-04-29 00:34:27 +0000488 yield TokenInfo(ERRORTOKEN, contstr + line,
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000489 strstart, (lnum, len(line)), contline)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000490 contstr = ''
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000491 contline = None
Guido van Rossumde655271997-04-09 17:15:54 +0000492 continue
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000493 else:
494 contstr = contstr + line
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000495 contline = contline + line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000496 continue
497
Guido van Rossum1aec3231997-04-08 14:24:39 +0000498 elif parenlev == 0 and not continued: # new statement
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000499 if not line: break
500 column = 0
Guido van Rossum1aec3231997-04-08 14:24:39 +0000501 while pos < max: # measure leading whitespace
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000502 if line[pos] == ' ':
503 column += 1
504 elif line[pos] == '\t':
505 column = (column//tabsize + 1)*tabsize
506 elif line[pos] == '\f':
507 column = 0
508 else:
509 break
510 pos += 1
511 if pos == max:
512 break
Guido van Rossum1aec3231997-04-08 14:24:39 +0000513
514 if line[pos] in '#\r\n': # skip comments or blank lines
Thomas Wouters89f507f2006-12-13 04:49:30 +0000515 if line[pos] == '#':
516 comment_token = line[pos:].rstrip('\r\n')
517 nl_pos = pos + len(comment_token)
Raymond Hettingera48db392009-04-29 00:34:27 +0000518 yield TokenInfo(COMMENT, comment_token,
Thomas Wouters89f507f2006-12-13 04:49:30 +0000519 (lnum, pos), (lnum, pos + len(comment_token)), line)
Raymond Hettingera48db392009-04-29 00:34:27 +0000520 yield TokenInfo(NL, line[nl_pos:],
Thomas Wouters89f507f2006-12-13 04:49:30 +0000521 (lnum, nl_pos), (lnum, len(line)), line)
522 else:
Raymond Hettingera48db392009-04-29 00:34:27 +0000523 yield TokenInfo((NL, COMMENT)[line[pos] == '#'], line[pos:],
Guido van Rossum1aec3231997-04-08 14:24:39 +0000524 (lnum, pos), (lnum, len(line)), line)
525 continue
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000526
527 if column > indents[-1]: # count indents or dedents
528 indents.append(column)
Raymond Hettingera48db392009-04-29 00:34:27 +0000529 yield TokenInfo(INDENT, line[:pos], (lnum, 0), (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000530 while column < indents[-1]:
Raymond Hettingerda99d1c2005-06-21 07:43:58 +0000531 if column not in indents:
532 raise IndentationError(
Thomas Wouters00ee7ba2006-08-21 19:07:27 +0000533 "unindent does not match any outer indentation level",
534 ("<tokenize>", lnum, pos, line))
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000535 indents = indents[:-1]
Raymond Hettingera48db392009-04-29 00:34:27 +0000536 yield TokenInfo(DEDENT, '', (lnum, pos), (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000537
538 else: # continued statement
Guido van Rossumde655271997-04-09 17:15:54 +0000539 if not line:
Collin Winterce36ad82007-08-30 01:19:48 +0000540 raise TokenError("EOF in multi-line statement", (lnum, 0))
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000541 continued = 0
542
543 while pos < max:
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200544 pseudomatch = _compile(PseudoToken).match(line, pos)
Guido van Rossum3b631771997-10-27 20:44:15 +0000545 if pseudomatch: # scan for tokens
546 start, end = pseudomatch.span(1)
Guido van Rossumde655271997-04-09 17:15:54 +0000547 spos, epos, pos = (lnum, start), (lnum, end), end
Guido van Rossum1aec3231997-04-08 14:24:39 +0000548 token, initial = line[start:end], line[start]
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000549
Georg Brandldde00282007-03-18 19:01:53 +0000550 if (initial in numchars or # ordinary number
551 (initial == '.' and token != '.' and token != '...')):
Raymond Hettingera48db392009-04-29 00:34:27 +0000552 yield TokenInfo(NUMBER, token, spos, epos, line)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000553 elif initial in '\r\n':
Raymond Hettingera48db392009-04-29 00:34:27 +0000554 yield TokenInfo(NL if parenlev > 0 else NEWLINE,
Thomas Wouters89f507f2006-12-13 04:49:30 +0000555 token, spos, epos, line)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000556 elif initial == '#':
Thomas Wouters89f507f2006-12-13 04:49:30 +0000557 assert not token.endswith("\n")
Raymond Hettingera48db392009-04-29 00:34:27 +0000558 yield TokenInfo(COMMENT, token, spos, epos, line)
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000559 elif token in triple_quoted:
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200560 endprog = _compile(endpats[token])
Guido van Rossum3b631771997-10-27 20:44:15 +0000561 endmatch = endprog.match(line, pos)
562 if endmatch: # all on one line
563 pos = endmatch.end(0)
Guido van Rossum1aec3231997-04-08 14:24:39 +0000564 token = line[start:pos]
Raymond Hettingera48db392009-04-29 00:34:27 +0000565 yield TokenInfo(STRING, token, spos, (lnum, pos), line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000566 else:
Guido van Rossum1aec3231997-04-08 14:24:39 +0000567 strstart = (lnum, start) # multiple lines
568 contstr = line[start:]
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000569 contline = line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000570 break
Guido van Rossum9d6897a2002-08-24 06:54:19 +0000571 elif initial in single_quoted or \
572 token[:2] in single_quoted or \
573 token[:3] in single_quoted:
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000574 if token[-1] == '\n': # continued string
Guido van Rossum1aec3231997-04-08 14:24:39 +0000575 strstart = (lnum, start)
Antoine Pitrou10a99b02011-10-11 15:45:56 +0200576 endprog = _compile(endpats[initial] or
577 endpats[token[1]] or
578 endpats[token[2]])
Guido van Rossumde655271997-04-09 17:15:54 +0000579 contstr, needcont = line[start:], 1
Guido van Rossuma90c78b1998-04-03 16:05:38 +0000580 contline = line
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000581 break
582 else: # ordinary string
Raymond Hettingera48db392009-04-29 00:34:27 +0000583 yield TokenInfo(STRING, token, spos, epos, line)
Benjamin Peterson33856de2010-08-30 14:41:20 +0000584 elif initial.isidentifier(): # ordinary name
Raymond Hettingera48db392009-04-29 00:34:27 +0000585 yield TokenInfo(NAME, token, spos, epos, line)
Guido van Rossum3b631771997-10-27 20:44:15 +0000586 elif initial == '\\': # continued stmt
587 continued = 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000588 else:
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000589 if initial in '([{':
590 parenlev += 1
591 elif initial in ')]}':
592 parenlev -= 1
Raymond Hettingera48db392009-04-29 00:34:27 +0000593 yield TokenInfo(OP, token, spos, epos, line)
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000594 else:
Raymond Hettingera48db392009-04-29 00:34:27 +0000595 yield TokenInfo(ERRORTOKEN, line[pos],
Guido van Rossumde655271997-04-09 17:15:54 +0000596 (lnum, pos), (lnum, pos+1), line)
Benjamin Petersona0dfa822009-11-13 02:25:08 +0000597 pos += 1
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000598
599 for indent in indents[1:]: # pop remaining indent levels
Raymond Hettingera48db392009-04-29 00:34:27 +0000600 yield TokenInfo(DEDENT, '', (lnum, 0), (lnum, 0), '')
601 yield TokenInfo(ENDMARKER, '', (lnum, 0), (lnum, 0), '')
Guido van Rossumfc6f5331997-03-07 00:21:12 +0000602
Trent Nelson428de652008-03-18 22:41:35 +0000603
604# An undocumented, backwards compatible, API for all the places in the standard
605# library that expect to be able to use tokenize with strings
606def generate_tokens(readline):
607 return _tokenize(readline, None)
Raymond Hettinger6c60d092010-09-09 04:32:39 +0000608
Meador Inge14c0f032011-10-07 08:53:38 -0500609def main():
610 import argparse
611
612 # Helper error handling routines
613 def perror(message):
614 print(message, file=sys.stderr)
615
616 def error(message, filename=None, location=None):
617 if location:
618 args = (filename,) + location + (message,)
619 perror("%s:%d:%d: error: %s" % args)
620 elif filename:
621 perror("%s: error: %s" % (filename, message))
622 else:
623 perror("error: %s" % message)
624 sys.exit(1)
625
626 # Parse the arguments and options
627 parser = argparse.ArgumentParser(prog='python -m tokenize')
628 parser.add_argument(dest='filename', nargs='?',
629 metavar='filename.py',
630 help='the file to tokenize; defaults to stdin')
Meador Inge00c7f852012-01-19 00:44:45 -0600631 parser.add_argument('-e', '--exact', dest='exact', action='store_true',
632 help='display token names using the exact type')
Meador Inge14c0f032011-10-07 08:53:38 -0500633 args = parser.parse_args()
634
635 try:
636 # Tokenize the input
637 if args.filename:
638 filename = args.filename
639 with builtins.open(filename, 'rb') as f:
640 tokens = list(tokenize(f.readline))
641 else:
642 filename = "<stdin>"
643 tokens = _tokenize(sys.stdin.readline, None)
644
645 # Output the tokenization
646 for token in tokens:
Meador Inge00c7f852012-01-19 00:44:45 -0600647 token_type = token.type
648 if args.exact:
649 token_type = token.exact_type
Meador Inge14c0f032011-10-07 08:53:38 -0500650 token_range = "%d,%d-%d,%d:" % (token.start + token.end)
651 print("%-20s%-15s%-15r" %
Meador Inge00c7f852012-01-19 00:44:45 -0600652 (token_range, tok_name[token_type], token.string))
Meador Inge14c0f032011-10-07 08:53:38 -0500653 except IndentationError as err:
654 line, column = err.args[1][1:3]
655 error(err.args[0], filename, (line, column))
656 except TokenError as err:
657 line, column = err.args[1]
658 error(err.args[0], filename, (line, column))
659 except SyntaxError as err:
660 error(err, filename)
661 except IOError as err:
662 error(err)
663 except KeyboardInterrupt:
664 print("interrupted\n")
665 except Exception as err:
666 perror("unexpected error: %s" % err)
667 raise
668
Raymond Hettinger6c60d092010-09-09 04:32:39 +0000669if __name__ == "__main__":
Meador Inge14c0f032011-10-07 08:53:38 -0500670 main()