Reduced memory burden by iterating over the normalization test input
file directly (instead of sucking it all into a list of lines first).
diff --git a/Lib/test/regrtest.py b/Lib/test/regrtest.py
index d0650e6..f870527 100755
--- a/Lib/test/regrtest.py
+++ b/Lib/test/regrtest.py
@@ -506,7 +506,7 @@
 #     test_normalization
 #         Whether a skip is expected here depends on whether a large test
 #         input file has been downloaded.  test_normalization.skip_expected
-#         controls that
+#         controls that.
 
 _expectations = {
     'win32':
diff --git a/Lib/test/test_normalization.py b/Lib/test/test_normalization.py
index b673739..7e18c97 100644
--- a/Lib/test/test_normalization.py
+++ b/Lib/test/test_normalization.py
@@ -33,10 +33,8 @@
         raise TestSkipped(TESTDATAFILE + " not found, download from " +
                     "http://www.unicode.org/Public/UNIDATA/" + TESTDATAFILE)
 
-    data = open(TESTDATAFILE).readlines()
-
     part1_data = {}
-    for line in data:
+    for line in open(TESTDATAFILE):
         if '#' in line:
             line = line.split('#')[0]
         line = line.strip()