blob: 70919ca998c4807ca5751fee857585d54389369d [file] [log] [blame]
Georg Brandl116aa622007-08-15 14:28:22 +00001:mod:`tokenize` --- Tokenizer for Python source
2===============================================
3
4.. module:: tokenize
5 :synopsis: Lexical scanner for Python source code.
6.. moduleauthor:: Ka Ping Yee
7.. sectionauthor:: Fred L. Drake, Jr. <fdrake@acm.org>
8
Raymond Hettinger10480942011-01-10 03:26:08 +00009**Source code:** :source:`Lib/tokenize.py`
Georg Brandl116aa622007-08-15 14:28:22 +000010
Raymond Hettinger4f707fd2011-01-10 19:54:11 +000011--------------
12
Georg Brandl116aa622007-08-15 14:28:22 +000013The :mod:`tokenize` module provides a lexical scanner for Python source code,
Trent Nelson428de652008-03-18 22:41:35 +000014implemented in Python. The scanner in this module returns comments as tokens
15as well, making it useful for implementing "pretty-printers," including
16colorizers for on-screen displays.
Georg Brandl116aa622007-08-15 14:28:22 +000017
Meador Inge972cfb92012-01-19 00:22:22 -060018To simplify token stream handling, all :ref:`operators` and :ref:`delimiters`
19tokens are returned using the generic :data:`token.OP` token type. The exact
20type can be determined by checking the token ``string`` field on the
21:term:`named tuple` returned from :func:`tokenize.tokenize` for the character
22sequence that identifies a specific operator token.
23
Georg Brandl9afde1c2007-11-01 20:32:30 +000024The primary entry point is a :term:`generator`:
Georg Brandl116aa622007-08-15 14:28:22 +000025
Trent Nelson428de652008-03-18 22:41:35 +000026.. function:: tokenize(readline)
Georg Brandl116aa622007-08-15 14:28:22 +000027
Trent Nelson428de652008-03-18 22:41:35 +000028 The :func:`tokenize` generator requires one argument, *readline*, which
Georg Brandl116aa622007-08-15 14:28:22 +000029 must be a callable object which provides the same interface as the
Antoine Pitrou4adb2882010-01-04 18:50:53 +000030 :meth:`io.IOBase.readline` method of file objects. Each call to the
31 function should return one line of input as bytes.
Georg Brandl116aa622007-08-15 14:28:22 +000032
Georg Brandl48310cd2009-01-03 21:18:54 +000033 The generator produces 5-tuples with these members: the token type; the
34 token string; a 2-tuple ``(srow, scol)`` of ints specifying the row and
35 column where the token begins in the source; a 2-tuple ``(erow, ecol)`` of
36 ints specifying the row and column where the token ends in the source; and
Georg Brandlc28e1fa2008-06-10 19:20:26 +000037 the line on which the token was found. The line passed (the last tuple item)
Raymond Hettingera48db392009-04-29 00:34:27 +000038 is the *logical* line; continuation lines are included. The 5 tuple is
39 returned as a :term:`named tuple` with the field names:
40 ``type string start end line``.
41
42 .. versionchanged:: 3.1
43 Added support for named tuples.
Georg Brandl48310cd2009-01-03 21:18:54 +000044
Georg Brandlc28e1fa2008-06-10 19:20:26 +000045 :func:`tokenize` determines the source encoding of the file by looking for a
46 UTF-8 BOM or encoding cookie, according to :pep:`263`.
Georg Brandl116aa622007-08-15 14:28:22 +000047
Georg Brandl55ac8f02007-09-01 13:51:09 +000048
Georg Brandl116aa622007-08-15 14:28:22 +000049All constants from the :mod:`token` module are also exported from
Trent Nelson428de652008-03-18 22:41:35 +000050:mod:`tokenize`, as are three additional token type values:
Georg Brandl116aa622007-08-15 14:28:22 +000051
Georg Brandl116aa622007-08-15 14:28:22 +000052.. data:: COMMENT
53
54 Token value used to indicate a comment.
55
56
57.. data:: NL
58
59 Token value used to indicate a non-terminating newline. The NEWLINE token
Georg Brandl48310cd2009-01-03 21:18:54 +000060 indicates the end of a logical line of Python code; NL tokens are generated
Trent Nelson428de652008-03-18 22:41:35 +000061 when a logical line of code is continued over multiple physical lines.
Georg Brandl116aa622007-08-15 14:28:22 +000062
Trent Nelson428de652008-03-18 22:41:35 +000063
64.. data:: ENCODING
65
Georg Brandl48310cd2009-01-03 21:18:54 +000066 Token value that indicates the encoding used to decode the source bytes
67 into text. The first token returned by :func:`tokenize` will always be an
Trent Nelson428de652008-03-18 22:41:35 +000068 ENCODING token.
69
70
Georg Brandl48310cd2009-01-03 21:18:54 +000071Another function is provided to reverse the tokenization process. This is
72useful for creating tools that tokenize a script, modify the token stream, and
Trent Nelson428de652008-03-18 22:41:35 +000073write back the modified script.
Georg Brandl116aa622007-08-15 14:28:22 +000074
75
76.. function:: untokenize(iterable)
77
Trent Nelson428de652008-03-18 22:41:35 +000078 Converts tokens back into Python source code. The *iterable* must return
Georg Brandl48310cd2009-01-03 21:18:54 +000079 sequences with at least two elements, the token type and the token string.
Trent Nelson428de652008-03-18 22:41:35 +000080 Any additional sequence elements are ignored.
Georg Brandl48310cd2009-01-03 21:18:54 +000081
Trent Nelson428de652008-03-18 22:41:35 +000082 The reconstructed script is returned as a single string. The result is
83 guaranteed to tokenize back to match the input so that the conversion is
Georg Brandl48310cd2009-01-03 21:18:54 +000084 lossless and round-trips are assured. The guarantee applies only to the
85 token type and token string as the spacing between tokens (column
Trent Nelson428de652008-03-18 22:41:35 +000086 positions) may change.
Georg Brandl48310cd2009-01-03 21:18:54 +000087
88 It returns bytes, encoded using the ENCODING token, which is the first
Trent Nelson428de652008-03-18 22:41:35 +000089 token sequence output by :func:`tokenize`.
Georg Brandl116aa622007-08-15 14:28:22 +000090
Georg Brandl116aa622007-08-15 14:28:22 +000091
Trent Nelson428de652008-03-18 22:41:35 +000092:func:`tokenize` needs to detect the encoding of source files it tokenizes. The
93function it uses to do this is available:
94
95.. function:: detect_encoding(readline)
96
Georg Brandl48310cd2009-01-03 21:18:54 +000097 The :func:`detect_encoding` function is used to detect the encoding that
Georg Brandlae2dbe22009-03-13 19:04:40 +000098 should be used to decode a Python source file. It requires one argument,
Trent Nelson428de652008-03-18 22:41:35 +000099 readline, in the same way as the :func:`tokenize` generator.
Georg Brandl48310cd2009-01-03 21:18:54 +0000100
Trent Nelson428de652008-03-18 22:41:35 +0000101 It will call readline a maximum of twice, and return the encoding used
102 (as a string) and a list of any lines (not decoded from bytes) it has read
103 in.
Georg Brandl48310cd2009-01-03 21:18:54 +0000104
Ezio Melottia8f6f1e2009-12-20 12:24:57 +0000105 It detects the encoding from the presence of a UTF-8 BOM or an encoding
106 cookie as specified in :pep:`263`. If both a BOM and a cookie are present,
Benjamin Peterson689a5582010-03-18 22:29:52 +0000107 but disagree, a SyntaxError will be raised. Note that if the BOM is found,
108 ``'utf-8-sig'`` will be returned as an encoding.
Trent Nelson428de652008-03-18 22:41:35 +0000109
Benjamin Petersonb3a48292010-03-18 22:43:41 +0000110 If no encoding is specified, then the default of ``'utf-8'`` will be
111 returned.
112
Victor Stinner58c07522010-11-09 01:08:59 +0000113 Use :func:`open` to open Python source files: it uses
114 :func:`detect_encoding` to detect the file encoding.
Benjamin Petersonb3a48292010-03-18 22:43:41 +0000115
Victor Stinner58c07522010-11-09 01:08:59 +0000116
117.. function:: open(filename)
118
119 Open a file in read only mode using the encoding detected by
120 :func:`detect_encoding`.
121
122 .. versionadded:: 3.2
Georg Brandl48310cd2009-01-03 21:18:54 +0000123
124
Raymond Hettinger6c60d092010-09-09 04:32:39 +0000125Example of a script rewriter that transforms float literals into Decimal
Georg Brandl116aa622007-08-15 14:28:22 +0000126objects::
127
Ezio Melottia8f6f1e2009-12-20 12:24:57 +0000128 from tokenize import tokenize, untokenize, NUMBER, STRING, NAME, OP
129 from io import BytesIO
130
Trent Nelson428de652008-03-18 22:41:35 +0000131 def decistmt(s):
132 """Substitute Decimals for floats in a string of statements.
Georg Brandl48310cd2009-01-03 21:18:54 +0000133
Trent Nelson428de652008-03-18 22:41:35 +0000134 >>> from decimal import Decimal
135 >>> s = 'print(+21.3e-5*-.1234/81.7)'
136 >>> decistmt(s)
137 "print (+Decimal ('21.3e-5')*-Decimal ('.1234')/Decimal ('81.7'))"
Georg Brandl48310cd2009-01-03 21:18:54 +0000138
Trent Nelson428de652008-03-18 22:41:35 +0000139 The format of the exponent is inherited from the platform C library.
140 Known cases are "e-007" (Windows) and "e-07" (not Windows). Since
141 we're only showing 12 digits, and the 13th isn't close to 5, the
142 rest of the output should be platform-independent.
Georg Brandl48310cd2009-01-03 21:18:54 +0000143
Trent Nelson428de652008-03-18 22:41:35 +0000144 >>> exec(s) #doctest: +ELLIPSIS
145 -3.21716034272e-0...7
Georg Brandl48310cd2009-01-03 21:18:54 +0000146
Trent Nelson428de652008-03-18 22:41:35 +0000147 Output from calculations with Decimal should be identical across all
148 platforms.
Georg Brandl48310cd2009-01-03 21:18:54 +0000149
Trent Nelson428de652008-03-18 22:41:35 +0000150 >>> exec(decistmt(s))
151 -3.217160342717258261933904529E-7
152 """
153 result = []
154 g = tokenize(BytesIO(s.encode('utf-8')).readline) # tokenize the string
155 for toknum, tokval, _, _, _ in g:
156 if toknum == NUMBER and '.' in tokval: # replace NUMBER tokens
157 result.extend([
158 (NAME, 'Decimal'),
159 (OP, '('),
160 (STRING, repr(tokval)),
161 (OP, ')')
162 ])
163 else:
164 result.append((toknum, tokval))
165 return untokenize(result).decode('utf-8')
Georg Brandl116aa622007-08-15 14:28:22 +0000166