Make test_tokenize really pass -- don't add extra output.
diff --git a/Lib/test/test_tokenize.py b/Lib/test/test_tokenize.py
index 788a04b..9ef6563 100644
--- a/Lib/test/test_tokenize.py
+++ b/Lib/test/test_tokenize.py
@@ -98,7 +98,7 @@
 # and tokenized again from the latter.  The test fails if the second
 # tokenization doesn't match the first.
 def test_roundtrip(f):
-    ## print 'Testing:', f
+    ## print('Testing:', f)
     # Get the encoding first
     fobj = open(f, encoding="latin-1")
     first2lines = fobj.readline() + fobj.readline()
@@ -106,7 +106,7 @@
     m = re.search(r"coding:\s*(\S+)", first2lines)
     if m:
         encoding = m.group(1)
-        print("    coding:", encoding)
+        ## print("    coding:", encoding)
     else:
         encoding = "utf-8"
     fobj = open(f, encoding=encoding)