Issue #25977: Fix typos in Lib/tokenize.py
Patch by John Walker.
diff --git a/Lib/tokenize.py b/Lib/tokenize.py
index 65d06e5..9fd676c 100644
--- a/Lib/tokenize.py
+++ b/Lib/tokenize.py
@@ -328,8 +328,8 @@
Round-trip invariant for full input:
Untokenized source will match input source exactly
- Round-trip invariant for limited intput:
- # Output bytes will tokenize the back to the input
+ Round-trip invariant for limited input:
+ # Output bytes will tokenize back to the input
t1 = [tok[:2] for tok in tokenize(f.readline)]
newcode = untokenize(t1)
readline = BytesIO(newcode).readline
@@ -465,10 +465,10 @@
def tokenize(readline):
"""
- The tokenize() generator requires one argment, readline, which
+ The tokenize() generator requires one argument, readline, which
must be a callable object which provides the same interface as the
readline() method of built-in file objects. Each call to the function
- should return one line of input as bytes. Alternately, readline
+ should return one line of input as bytes. Alternatively, readline
can be a callable function terminating with StopIteration:
readline = open(myfile, 'rb').__next__ # Example of alternate readline