blob: 197b574e67b5df2c1bf8d1a502ca0ee7fffb4406 [file] [log] [blame]
Georg Brandl116aa622007-08-15 14:28:22 +00001:mod:`tokenize` --- Tokenizer for Python source
2===============================================
3
4.. module:: tokenize
5 :synopsis: Lexical scanner for Python source code.
6.. moduleauthor:: Ka Ping Yee
7.. sectionauthor:: Fred L. Drake, Jr. <fdrake@acm.org>
8
9
10The :mod:`tokenize` module provides a lexical scanner for Python source code,
Trent Nelson428de652008-03-18 22:41:35 +000011implemented in Python. The scanner in this module returns comments as tokens
12as well, making it useful for implementing "pretty-printers," including
13colorizers for on-screen displays.
Georg Brandl116aa622007-08-15 14:28:22 +000014
Georg Brandl9afde1c2007-11-01 20:32:30 +000015The primary entry point is a :term:`generator`:
Georg Brandl116aa622007-08-15 14:28:22 +000016
Trent Nelson428de652008-03-18 22:41:35 +000017.. function:: tokenize(readline)
Georg Brandl116aa622007-08-15 14:28:22 +000018
Trent Nelson428de652008-03-18 22:41:35 +000019 The :func:`tokenize` generator requires one argument, *readline*, which
Georg Brandl116aa622007-08-15 14:28:22 +000020 must be a callable object which provides the same interface as the
21 :meth:`readline` method of built-in file objects (see section
Georg Brandl48310cd2009-01-03 21:18:54 +000022 :ref:`bltin-file-objects`). Each call to the function should return one
Trent Nelson428de652008-03-18 22:41:35 +000023 line of input as bytes.
Georg Brandl116aa622007-08-15 14:28:22 +000024
Georg Brandl48310cd2009-01-03 21:18:54 +000025 The generator produces 5-tuples with these members: the token type; the
26 token string; a 2-tuple ``(srow, scol)`` of ints specifying the row and
27 column where the token begins in the source; a 2-tuple ``(erow, ecol)`` of
28 ints specifying the row and column where the token ends in the source; and
Georg Brandlc28e1fa2008-06-10 19:20:26 +000029 the line on which the token was found. The line passed (the last tuple item)
30 is the *logical* line; continuation lines are included.
Georg Brandl48310cd2009-01-03 21:18:54 +000031
Georg Brandlc28e1fa2008-06-10 19:20:26 +000032 :func:`tokenize` determines the source encoding of the file by looking for a
33 UTF-8 BOM or encoding cookie, according to :pep:`263`.
Georg Brandl116aa622007-08-15 14:28:22 +000034
Georg Brandl55ac8f02007-09-01 13:51:09 +000035
Georg Brandl116aa622007-08-15 14:28:22 +000036All constants from the :mod:`token` module are also exported from
Trent Nelson428de652008-03-18 22:41:35 +000037:mod:`tokenize`, as are three additional token type values:
Georg Brandl116aa622007-08-15 14:28:22 +000038
Georg Brandl116aa622007-08-15 14:28:22 +000039.. data:: COMMENT
40
41 Token value used to indicate a comment.
42
43
44.. data:: NL
45
46 Token value used to indicate a non-terminating newline. The NEWLINE token
Georg Brandl48310cd2009-01-03 21:18:54 +000047 indicates the end of a logical line of Python code; NL tokens are generated
Trent Nelson428de652008-03-18 22:41:35 +000048 when a logical line of code is continued over multiple physical lines.
Georg Brandl116aa622007-08-15 14:28:22 +000049
Trent Nelson428de652008-03-18 22:41:35 +000050
51.. data:: ENCODING
52
Georg Brandl48310cd2009-01-03 21:18:54 +000053 Token value that indicates the encoding used to decode the source bytes
54 into text. The first token returned by :func:`tokenize` will always be an
Trent Nelson428de652008-03-18 22:41:35 +000055 ENCODING token.
56
57
Georg Brandl48310cd2009-01-03 21:18:54 +000058Another function is provided to reverse the tokenization process. This is
59useful for creating tools that tokenize a script, modify the token stream, and
Trent Nelson428de652008-03-18 22:41:35 +000060write back the modified script.
Georg Brandl116aa622007-08-15 14:28:22 +000061
62
63.. function:: untokenize(iterable)
64
Trent Nelson428de652008-03-18 22:41:35 +000065 Converts tokens back into Python source code. The *iterable* must return
Georg Brandl48310cd2009-01-03 21:18:54 +000066 sequences with at least two elements, the token type and the token string.
Trent Nelson428de652008-03-18 22:41:35 +000067 Any additional sequence elements are ignored.
Georg Brandl48310cd2009-01-03 21:18:54 +000068
Trent Nelson428de652008-03-18 22:41:35 +000069 The reconstructed script is returned as a single string. The result is
70 guaranteed to tokenize back to match the input so that the conversion is
Georg Brandl48310cd2009-01-03 21:18:54 +000071 lossless and round-trips are assured. The guarantee applies only to the
72 token type and token string as the spacing between tokens (column
Trent Nelson428de652008-03-18 22:41:35 +000073 positions) may change.
Georg Brandl48310cd2009-01-03 21:18:54 +000074
75 It returns bytes, encoded using the ENCODING token, which is the first
Trent Nelson428de652008-03-18 22:41:35 +000076 token sequence output by :func:`tokenize`.
Georg Brandl116aa622007-08-15 14:28:22 +000077
Georg Brandl116aa622007-08-15 14:28:22 +000078
Trent Nelson428de652008-03-18 22:41:35 +000079:func:`tokenize` needs to detect the encoding of source files it tokenizes. The
80function it uses to do this is available:
81
82.. function:: detect_encoding(readline)
83
Georg Brandl48310cd2009-01-03 21:18:54 +000084 The :func:`detect_encoding` function is used to detect the encoding that
85 should be used to decode a Python source file. It requires one argment,
Trent Nelson428de652008-03-18 22:41:35 +000086 readline, in the same way as the :func:`tokenize` generator.
Georg Brandl48310cd2009-01-03 21:18:54 +000087
Trent Nelson428de652008-03-18 22:41:35 +000088 It will call readline a maximum of twice, and return the encoding used
89 (as a string) and a list of any lines (not decoded from bytes) it has read
90 in.
Georg Brandl48310cd2009-01-03 21:18:54 +000091
Trent Nelson428de652008-03-18 22:41:35 +000092 It detects the encoding from the presence of a utf-8 bom or an encoding
93 cookie as specified in pep-0263. If both a bom and a cookie are present,
94 but disagree, a SyntaxError will be raised.
Trent Nelson428de652008-03-18 22:41:35 +000095
Georg Brandl48310cd2009-01-03 21:18:54 +000096 If no encoding is specified, then the default of 'utf-8' will be returned.
97
98
Georg Brandl116aa622007-08-15 14:28:22 +000099Example of a script re-writer that transforms float literals into Decimal
100objects::
101
Trent Nelson428de652008-03-18 22:41:35 +0000102 def decistmt(s):
103 """Substitute Decimals for floats in a string of statements.
Georg Brandl48310cd2009-01-03 21:18:54 +0000104
Trent Nelson428de652008-03-18 22:41:35 +0000105 >>> from decimal import Decimal
106 >>> s = 'print(+21.3e-5*-.1234/81.7)'
107 >>> decistmt(s)
108 "print (+Decimal ('21.3e-5')*-Decimal ('.1234')/Decimal ('81.7'))"
Georg Brandl48310cd2009-01-03 21:18:54 +0000109
Trent Nelson428de652008-03-18 22:41:35 +0000110 The format of the exponent is inherited from the platform C library.
111 Known cases are "e-007" (Windows) and "e-07" (not Windows). Since
112 we're only showing 12 digits, and the 13th isn't close to 5, the
113 rest of the output should be platform-independent.
Georg Brandl48310cd2009-01-03 21:18:54 +0000114
Trent Nelson428de652008-03-18 22:41:35 +0000115 >>> exec(s) #doctest: +ELLIPSIS
116 -3.21716034272e-0...7
Georg Brandl48310cd2009-01-03 21:18:54 +0000117
Trent Nelson428de652008-03-18 22:41:35 +0000118 Output from calculations with Decimal should be identical across all
119 platforms.
Georg Brandl48310cd2009-01-03 21:18:54 +0000120
Trent Nelson428de652008-03-18 22:41:35 +0000121 >>> exec(decistmt(s))
122 -3.217160342717258261933904529E-7
123 """
124 result = []
125 g = tokenize(BytesIO(s.encode('utf-8')).readline) # tokenize the string
126 for toknum, tokval, _, _, _ in g:
127 if toknum == NUMBER and '.' in tokval: # replace NUMBER tokens
128 result.extend([
129 (NAME, 'Decimal'),
130 (OP, '('),
131 (STRING, repr(tokval)),
132 (OP, ')')
133 ])
134 else:
135 result.append((toknum, tokval))
136 return untokenize(result).decode('utf-8')
Georg Brandl116aa622007-08-15 14:28:22 +0000137
Georg Brandl116aa622007-08-15 14:28:22 +0000138