Reduced memory burden by iterating over the normalization test input
file directly (instead of sucking it all into a list of lines first).
diff --git a/Lib/test/regrtest.py b/Lib/test/regrtest.py
index d0650e6..f870527 100755
--- a/Lib/test/regrtest.py
+++ b/Lib/test/regrtest.py
@@ -506,7 +506,7 @@
# test_normalization
# Whether a skip is expected here depends on whether a large test
# input file has been downloaded. test_normalization.skip_expected
-# controls that
+# controls that.
_expectations = {
'win32':
diff --git a/Lib/test/test_normalization.py b/Lib/test/test_normalization.py
index b673739..7e18c97 100644
--- a/Lib/test/test_normalization.py
+++ b/Lib/test/test_normalization.py
@@ -33,10 +33,8 @@
raise TestSkipped(TESTDATAFILE + " not found, download from " +
"http://www.unicode.org/Public/UNIDATA/" + TESTDATAFILE)
- data = open(TESTDATAFILE).readlines()
-
part1_data = {}
- for line in data:
+ for line in open(TESTDATAFILE):
if '#' in line:
line = line.split('#')[0]
line = line.strip()