| """ Test script for the Unicode implementation. |
| |
| Written by Marc-Andre Lemburg (mal@lemburg.com). |
| |
| (c) Copyright CNRI, All Rights Reserved. NO WARRANTY. |
| |
| """ |
| import _string |
| import codecs |
| import itertools |
| import operator |
| import struct |
| import string |
| import sys |
| import unittest |
| import warnings |
| from test import support, string_tests |
| |
| # Error handling (bad decoder return) |
| def search_function(encoding): |
| def decode1(input, errors="strict"): |
| return 42 # not a tuple |
| def encode1(input, errors="strict"): |
| return 42 # not a tuple |
| def encode2(input, errors="strict"): |
| return (42, 42) # no unicode |
| def decode2(input, errors="strict"): |
| return (42, 42) # no unicode |
| if encoding=="test.unicode1": |
| return (encode1, decode1, None, None) |
| elif encoding=="test.unicode2": |
| return (encode2, decode2, None, None) |
| else: |
| return None |
| codecs.register(search_function) |
| |
| def duplicate_string(text): |
| """ |
| Try to get a fresh clone of the specified text: |
| new object with a reference count of 1. |
| |
| This is a best-effort: latin1 single letters and the empty |
| string ('') are singletons and cannot be cloned. |
| """ |
| return text.encode().decode() |
| |
| class StrSubclass(str): |
| pass |
| |
| class UnicodeTest(string_tests.CommonTest, |
| string_tests.MixinStrUnicodeUserStringTest, |
| string_tests.MixinStrUnicodeTest, |
| unittest.TestCase): |
| |
| type2test = str |
| |
| def checkequalnofix(self, result, object, methodname, *args): |
| method = getattr(object, methodname) |
| realresult = method(*args) |
| self.assertEqual(realresult, result) |
| self.assertTrue(type(realresult) is type(result)) |
| |
| # if the original is returned make sure that |
| # this doesn't happen with subclasses |
| if realresult is object: |
| class usub(str): |
| def __repr__(self): |
| return 'usub(%r)' % str.__repr__(self) |
| object = usub(object) |
| method = getattr(object, methodname) |
| realresult = method(*args) |
| self.assertEqual(realresult, result) |
| self.assertTrue(object is not realresult) |
| |
| def test_literals(self): |
| self.assertEqual('\xff', '\u00ff') |
| self.assertEqual('\uffff', '\U0000ffff') |
| self.assertRaises(SyntaxError, eval, '\'\\Ufffffffe\'') |
| self.assertRaises(SyntaxError, eval, '\'\\Uffffffff\'') |
| self.assertRaises(SyntaxError, eval, '\'\\U%08x\'' % 0x110000) |
| # raw strings should not have unicode escapes |
| self.assertNotEqual(r"\u0020", " ") |
| |
| def test_ascii(self): |
| if not sys.platform.startswith('java'): |
| # Test basic sanity of repr() |
| self.assertEqual(ascii('abc'), "'abc'") |
| self.assertEqual(ascii('ab\\c'), "'ab\\\\c'") |
| self.assertEqual(ascii('ab\\'), "'ab\\\\'") |
| self.assertEqual(ascii('\\c'), "'\\\\c'") |
| self.assertEqual(ascii('\\'), "'\\\\'") |
| self.assertEqual(ascii('\n'), "'\\n'") |
| self.assertEqual(ascii('\r'), "'\\r'") |
| self.assertEqual(ascii('\t'), "'\\t'") |
| self.assertEqual(ascii('\b'), "'\\x08'") |
| self.assertEqual(ascii("'\""), """'\\'"'""") |
| self.assertEqual(ascii("'\""), """'\\'"'""") |
| self.assertEqual(ascii("'"), '''"'"''') |
| self.assertEqual(ascii('"'), """'"'""") |
| latin1repr = ( |
| "'\\x00\\x01\\x02\\x03\\x04\\x05\\x06\\x07\\x08\\t\\n\\x0b\\x0c\\r" |
| "\\x0e\\x0f\\x10\\x11\\x12\\x13\\x14\\x15\\x16\\x17\\x18\\x19\\x1a" |
| "\\x1b\\x1c\\x1d\\x1e\\x1f !\"#$%&\\'()*+,-./0123456789:;<=>?@ABCDEFGHI" |
| "JKLMNOPQRSTUVWXYZ[\\\\]^_`abcdefghijklmnopqrstuvwxyz{|}~\\x7f" |
| "\\x80\\x81\\x82\\x83\\x84\\x85\\x86\\x87\\x88\\x89\\x8a\\x8b\\x8c\\x8d" |
| "\\x8e\\x8f\\x90\\x91\\x92\\x93\\x94\\x95\\x96\\x97\\x98\\x99\\x9a\\x9b" |
| "\\x9c\\x9d\\x9e\\x9f\\xa0\\xa1\\xa2\\xa3\\xa4\\xa5\\xa6\\xa7\\xa8\\xa9" |
| "\\xaa\\xab\\xac\\xad\\xae\\xaf\\xb0\\xb1\\xb2\\xb3\\xb4\\xb5\\xb6\\xb7" |
| "\\xb8\\xb9\\xba\\xbb\\xbc\\xbd\\xbe\\xbf\\xc0\\xc1\\xc2\\xc3\\xc4\\xc5" |
| "\\xc6\\xc7\\xc8\\xc9\\xca\\xcb\\xcc\\xcd\\xce\\xcf\\xd0\\xd1\\xd2\\xd3" |
| "\\xd4\\xd5\\xd6\\xd7\\xd8\\xd9\\xda\\xdb\\xdc\\xdd\\xde\\xdf\\xe0\\xe1" |
| "\\xe2\\xe3\\xe4\\xe5\\xe6\\xe7\\xe8\\xe9\\xea\\xeb\\xec\\xed\\xee\\xef" |
| "\\xf0\\xf1\\xf2\\xf3\\xf4\\xf5\\xf6\\xf7\\xf8\\xf9\\xfa\\xfb\\xfc\\xfd" |
| "\\xfe\\xff'") |
| testrepr = ascii(''.join(map(chr, range(256)))) |
| self.assertEqual(testrepr, latin1repr) |
| # Test ascii works on wide unicode escapes without overflow. |
| self.assertEqual(ascii("\U00010000" * 39 + "\uffff" * 4096), |
| ascii("\U00010000" * 39 + "\uffff" * 4096)) |
| |
| class WrongRepr: |
| def __repr__(self): |
| return b'byte-repr' |
| self.assertRaises(TypeError, ascii, WrongRepr()) |
| |
| def test_repr(self): |
| if not sys.platform.startswith('java'): |
| # Test basic sanity of repr() |
| self.assertEqual(repr('abc'), "'abc'") |
| self.assertEqual(repr('ab\\c'), "'ab\\\\c'") |
| self.assertEqual(repr('ab\\'), "'ab\\\\'") |
| self.assertEqual(repr('\\c'), "'\\\\c'") |
| self.assertEqual(repr('\\'), "'\\\\'") |
| self.assertEqual(repr('\n'), "'\\n'") |
| self.assertEqual(repr('\r'), "'\\r'") |
| self.assertEqual(repr('\t'), "'\\t'") |
| self.assertEqual(repr('\b'), "'\\x08'") |
| self.assertEqual(repr("'\""), """'\\'"'""") |
| self.assertEqual(repr("'\""), """'\\'"'""") |
| self.assertEqual(repr("'"), '''"'"''') |
| self.assertEqual(repr('"'), """'"'""") |
| latin1repr = ( |
| "'\\x00\\x01\\x02\\x03\\x04\\x05\\x06\\x07\\x08\\t\\n\\x0b\\x0c\\r" |
| "\\x0e\\x0f\\x10\\x11\\x12\\x13\\x14\\x15\\x16\\x17\\x18\\x19\\x1a" |
| "\\x1b\\x1c\\x1d\\x1e\\x1f !\"#$%&\\'()*+,-./0123456789:;<=>?@ABCDEFGHI" |
| "JKLMNOPQRSTUVWXYZ[\\\\]^_`abcdefghijklmnopqrstuvwxyz{|}~\\x7f" |
| "\\x80\\x81\\x82\\x83\\x84\\x85\\x86\\x87\\x88\\x89\\x8a\\x8b\\x8c\\x8d" |
| "\\x8e\\x8f\\x90\\x91\\x92\\x93\\x94\\x95\\x96\\x97\\x98\\x99\\x9a\\x9b" |
| "\\x9c\\x9d\\x9e\\x9f\\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9" |
| "\xaa\xab\xac\\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7" |
| "\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5" |
| "\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3" |
| "\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1" |
| "\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef" |
| "\xf0\xf1\xf2\xf3\xf4\xf5\xf6\xf7\xf8\xf9\xfa\xfb\xfc\xfd" |
| "\xfe\xff'") |
| testrepr = repr(''.join(map(chr, range(256)))) |
| self.assertEqual(testrepr, latin1repr) |
| # Test repr works on wide unicode escapes without overflow. |
| self.assertEqual(repr("\U00010000" * 39 + "\uffff" * 4096), |
| repr("\U00010000" * 39 + "\uffff" * 4096)) |
| |
| class WrongRepr: |
| def __repr__(self): |
| return b'byte-repr' |
| self.assertRaises(TypeError, repr, WrongRepr()) |
| |
| def test_iterators(self): |
| # Make sure unicode objects have an __iter__ method |
| it = "\u1111\u2222\u3333".__iter__() |
| self.assertEqual(next(it), "\u1111") |
| self.assertEqual(next(it), "\u2222") |
| self.assertEqual(next(it), "\u3333") |
| self.assertRaises(StopIteration, next, it) |
| |
| def test_count(self): |
| string_tests.CommonTest.test_count(self) |
| # check mixed argument types |
| self.checkequalnofix(3, 'aaa', 'count', 'a') |
| self.checkequalnofix(0, 'aaa', 'count', 'b') |
| self.checkequalnofix(3, 'aaa', 'count', 'a') |
| self.checkequalnofix(0, 'aaa', 'count', 'b') |
| self.checkequalnofix(0, 'aaa', 'count', 'b') |
| self.checkequalnofix(1, 'aaa', 'count', 'a', -1) |
| self.checkequalnofix(3, 'aaa', 'count', 'a', -10) |
| self.checkequalnofix(2, 'aaa', 'count', 'a', 0, -1) |
| self.checkequalnofix(0, 'aaa', 'count', 'a', 0, -10) |
| # test mixed kinds |
| self.checkequal(10, '\u0102' + 'a' * 10, 'count', 'a') |
| self.checkequal(10, '\U00100304' + 'a' * 10, 'count', 'a') |
| self.checkequal(10, '\U00100304' + '\u0102' * 10, 'count', '\u0102') |
| self.checkequal(0, 'a' * 10, 'count', '\u0102') |
| self.checkequal(0, 'a' * 10, 'count', '\U00100304') |
| self.checkequal(0, '\u0102' * 10, 'count', '\U00100304') |
| self.checkequal(10, '\u0102' + 'a_' * 10, 'count', 'a_') |
| self.checkequal(10, '\U00100304' + 'a_' * 10, 'count', 'a_') |
| self.checkequal(10, '\U00100304' + '\u0102_' * 10, 'count', '\u0102_') |
| self.checkequal(0, 'a' * 10, 'count', 'a\u0102') |
| self.checkequal(0, 'a' * 10, 'count', 'a\U00100304') |
| self.checkequal(0, '\u0102' * 10, 'count', '\u0102\U00100304') |
| |
| def test_find(self): |
| string_tests.CommonTest.test_find(self) |
| # test implementation details of the memchr fast path |
| self.checkequal(100, 'a' * 100 + '\u0102', 'find', '\u0102') |
| self.checkequal(-1, 'a' * 100 + '\u0102', 'find', '\u0201') |
| self.checkequal(-1, 'a' * 100 + '\u0102', 'find', '\u0120') |
| self.checkequal(-1, 'a' * 100 + '\u0102', 'find', '\u0220') |
| self.checkequal(100, 'a' * 100 + '\U00100304', 'find', '\U00100304') |
| self.checkequal(-1, 'a' * 100 + '\U00100304', 'find', '\U00100204') |
| self.checkequal(-1, 'a' * 100 + '\U00100304', 'find', '\U00102004') |
| # check mixed argument types |
| self.checkequalnofix(0, 'abcdefghiabc', 'find', 'abc') |
| self.checkequalnofix(9, 'abcdefghiabc', 'find', 'abc', 1) |
| self.checkequalnofix(-1, 'abcdefghiabc', 'find', 'def', 4) |
| |
| self.assertRaises(TypeError, 'hello'.find) |
| self.assertRaises(TypeError, 'hello'.find, 42) |
| # test mixed kinds |
| self.checkequal(100, '\u0102' * 100 + 'a', 'find', 'a') |
| self.checkequal(100, '\U00100304' * 100 + 'a', 'find', 'a') |
| self.checkequal(100, '\U00100304' * 100 + '\u0102', 'find', '\u0102') |
| self.checkequal(-1, 'a' * 100, 'find', '\u0102') |
| self.checkequal(-1, 'a' * 100, 'find', '\U00100304') |
| self.checkequal(-1, '\u0102' * 100, 'find', '\U00100304') |
| self.checkequal(100, '\u0102' * 100 + 'a_', 'find', 'a_') |
| self.checkequal(100, '\U00100304' * 100 + 'a_', 'find', 'a_') |
| self.checkequal(100, '\U00100304' * 100 + '\u0102_', 'find', '\u0102_') |
| self.checkequal(-1, 'a' * 100, 'find', 'a\u0102') |
| self.checkequal(-1, 'a' * 100, 'find', 'a\U00100304') |
| self.checkequal(-1, '\u0102' * 100, 'find', '\u0102\U00100304') |
| |
| def test_rfind(self): |
| string_tests.CommonTest.test_rfind(self) |
| # test implementation details of the memrchr fast path |
| self.checkequal(0, '\u0102' + 'a' * 100 , 'rfind', '\u0102') |
| self.checkequal(-1, '\u0102' + 'a' * 100 , 'rfind', '\u0201') |
| self.checkequal(-1, '\u0102' + 'a' * 100 , 'rfind', '\u0120') |
| self.checkequal(-1, '\u0102' + 'a' * 100 , 'rfind', '\u0220') |
| self.checkequal(0, '\U00100304' + 'a' * 100, 'rfind', '\U00100304') |
| self.checkequal(-1, '\U00100304' + 'a' * 100, 'rfind', '\U00100204') |
| self.checkequal(-1, '\U00100304' + 'a' * 100, 'rfind', '\U00102004') |
| # check mixed argument types |
| self.checkequalnofix(9, 'abcdefghiabc', 'rfind', 'abc') |
| self.checkequalnofix(12, 'abcdefghiabc', 'rfind', '') |
| self.checkequalnofix(12, 'abcdefghiabc', 'rfind', '') |
| # test mixed kinds |
| self.checkequal(0, 'a' + '\u0102' * 100, 'rfind', 'a') |
| self.checkequal(0, 'a' + '\U00100304' * 100, 'rfind', 'a') |
| self.checkequal(0, '\u0102' + '\U00100304' * 100, 'rfind', '\u0102') |
| self.checkequal(-1, 'a' * 100, 'rfind', '\u0102') |
| self.checkequal(-1, 'a' * 100, 'rfind', '\U00100304') |
| self.checkequal(-1, '\u0102' * 100, 'rfind', '\U00100304') |
| self.checkequal(0, '_a' + '\u0102' * 100, 'rfind', '_a') |
| self.checkequal(0, '_a' + '\U00100304' * 100, 'rfind', '_a') |
| self.checkequal(0, '_\u0102' + '\U00100304' * 100, 'rfind', '_\u0102') |
| self.checkequal(-1, 'a' * 100, 'rfind', '\u0102a') |
| self.checkequal(-1, 'a' * 100, 'rfind', '\U00100304a') |
| self.checkequal(-1, '\u0102' * 100, 'rfind', '\U00100304\u0102') |
| |
| def test_index(self): |
| string_tests.CommonTest.test_index(self) |
| self.checkequalnofix(0, 'abcdefghiabc', 'index', '') |
| self.checkequalnofix(3, 'abcdefghiabc', 'index', 'def') |
| self.checkequalnofix(0, 'abcdefghiabc', 'index', 'abc') |
| self.checkequalnofix(9, 'abcdefghiabc', 'index', 'abc', 1) |
| self.assertRaises(ValueError, 'abcdefghiabc'.index, 'hib') |
| self.assertRaises(ValueError, 'abcdefghiab'.index, 'abc', 1) |
| self.assertRaises(ValueError, 'abcdefghi'.index, 'ghi', 8) |
| self.assertRaises(ValueError, 'abcdefghi'.index, 'ghi', -1) |
| # test mixed kinds |
| self.checkequal(100, '\u0102' * 100 + 'a', 'index', 'a') |
| self.checkequal(100, '\U00100304' * 100 + 'a', 'index', 'a') |
| self.checkequal(100, '\U00100304' * 100 + '\u0102', 'index', '\u0102') |
| self.assertRaises(ValueError, ('a' * 100).index, '\u0102') |
| self.assertRaises(ValueError, ('a' * 100).index, '\U00100304') |
| self.assertRaises(ValueError, ('\u0102' * 100).index, '\U00100304') |
| self.checkequal(100, '\u0102' * 100 + 'a_', 'index', 'a_') |
| self.checkequal(100, '\U00100304' * 100 + 'a_', 'index', 'a_') |
| self.checkequal(100, '\U00100304' * 100 + '\u0102_', 'index', '\u0102_') |
| self.assertRaises(ValueError, ('a' * 100).index, 'a\u0102') |
| self.assertRaises(ValueError, ('a' * 100).index, 'a\U00100304') |
| self.assertRaises(ValueError, ('\u0102' * 100).index, '\u0102\U00100304') |
| |
| def test_rindex(self): |
| string_tests.CommonTest.test_rindex(self) |
| self.checkequalnofix(12, 'abcdefghiabc', 'rindex', '') |
| self.checkequalnofix(3, 'abcdefghiabc', 'rindex', 'def') |
| self.checkequalnofix(9, 'abcdefghiabc', 'rindex', 'abc') |
| self.checkequalnofix(0, 'abcdefghiabc', 'rindex', 'abc', 0, -1) |
| |
| self.assertRaises(ValueError, 'abcdefghiabc'.rindex, 'hib') |
| self.assertRaises(ValueError, 'defghiabc'.rindex, 'def', 1) |
| self.assertRaises(ValueError, 'defghiabc'.rindex, 'abc', 0, -1) |
| self.assertRaises(ValueError, 'abcdefghi'.rindex, 'ghi', 0, 8) |
| self.assertRaises(ValueError, 'abcdefghi'.rindex, 'ghi', 0, -1) |
| # test mixed kinds |
| self.checkequal(0, 'a' + '\u0102' * 100, 'rindex', 'a') |
| self.checkequal(0, 'a' + '\U00100304' * 100, 'rindex', 'a') |
| self.checkequal(0, '\u0102' + '\U00100304' * 100, 'rindex', '\u0102') |
| self.assertRaises(ValueError, ('a' * 100).rindex, '\u0102') |
| self.assertRaises(ValueError, ('a' * 100).rindex, '\U00100304') |
| self.assertRaises(ValueError, ('\u0102' * 100).rindex, '\U00100304') |
| self.checkequal(0, '_a' + '\u0102' * 100, 'rindex', '_a') |
| self.checkequal(0, '_a' + '\U00100304' * 100, 'rindex', '_a') |
| self.checkequal(0, '_\u0102' + '\U00100304' * 100, 'rindex', '_\u0102') |
| self.assertRaises(ValueError, ('a' * 100).rindex, '\u0102a') |
| self.assertRaises(ValueError, ('a' * 100).rindex, '\U00100304a') |
| self.assertRaises(ValueError, ('\u0102' * 100).rindex, '\U00100304\u0102') |
| |
| def test_maketrans_translate(self): |
| # these work with plain translate() |
| self.checkequalnofix('bbbc', 'abababc', 'translate', |
| {ord('a'): None}) |
| self.checkequalnofix('iiic', 'abababc', 'translate', |
| {ord('a'): None, ord('b'): ord('i')}) |
| self.checkequalnofix('iiix', 'abababc', 'translate', |
| {ord('a'): None, ord('b'): ord('i'), ord('c'): 'x'}) |
| self.checkequalnofix('c', 'abababc', 'translate', |
| {ord('a'): None, ord('b'): ''}) |
| self.checkequalnofix('xyyx', 'xzx', 'translate', |
| {ord('z'): 'yy'}) |
| |
| # this needs maketrans() |
| self.checkequalnofix('abababc', 'abababc', 'translate', |
| {'b': '<i>'}) |
| tbl = self.type2test.maketrans({'a': None, 'b': '<i>'}) |
| self.checkequalnofix('<i><i><i>c', 'abababc', 'translate', tbl) |
| # test alternative way of calling maketrans() |
| tbl = self.type2test.maketrans('abc', 'xyz', 'd') |
| self.checkequalnofix('xyzzy', 'abdcdcbdddd', 'translate', tbl) |
| |
| # various tests switching from ASCII to latin1 or the opposite; |
| # same length, remove a letter, or replace with a longer string. |
| self.assertEqual("[a]".translate(str.maketrans('a', 'X')), |
| "[X]") |
| self.assertEqual("[a]".translate(str.maketrans({'a': 'X'})), |
| "[X]") |
| self.assertEqual("[a]".translate(str.maketrans({'a': None})), |
| "[]") |
| self.assertEqual("[a]".translate(str.maketrans({'a': 'XXX'})), |
| "[XXX]") |
| self.assertEqual("[a]".translate(str.maketrans({'a': '\xe9'})), |
| "[\xe9]") |
| self.assertEqual('axb'.translate(str.maketrans({'a': None, 'b': '123'})), |
| "x123") |
| self.assertEqual('axb'.translate(str.maketrans({'a': None, 'b': '\xe9'})), |
| "x\xe9") |
| |
| # test non-ASCII (don't take the fast-path) |
| self.assertEqual("[a]".translate(str.maketrans({'a': '<\xe9>'})), |
| "[<\xe9>]") |
| self.assertEqual("[\xe9]".translate(str.maketrans({'\xe9': 'a'})), |
| "[a]") |
| self.assertEqual("[\xe9]".translate(str.maketrans({'\xe9': None})), |
| "[]") |
| self.assertEqual("[\xe9]".translate(str.maketrans({'\xe9': '123'})), |
| "[123]") |
| self.assertEqual("[a\xe9]".translate(str.maketrans({'a': '<\u20ac>'})), |
| "[<\u20ac>\xe9]") |
| |
| # invalid Unicode characters |
| invalid_char = 0x10ffff+1 |
| for before in "a\xe9\u20ac\U0010ffff": |
| mapping = str.maketrans({before: invalid_char}) |
| text = "[%s]" % before |
| self.assertRaises(ValueError, text.translate, mapping) |
| |
| # errors |
| self.assertRaises(TypeError, self.type2test.maketrans) |
| self.assertRaises(ValueError, self.type2test.maketrans, 'abc', 'defg') |
| self.assertRaises(TypeError, self.type2test.maketrans, 2, 'def') |
| self.assertRaises(TypeError, self.type2test.maketrans, 'abc', 2) |
| self.assertRaises(TypeError, self.type2test.maketrans, 'abc', 'def', 2) |
| self.assertRaises(ValueError, self.type2test.maketrans, {'xy': 2}) |
| self.assertRaises(TypeError, self.type2test.maketrans, {(1,): 2}) |
| |
| self.assertRaises(TypeError, 'hello'.translate) |
| self.assertRaises(TypeError, 'abababc'.translate, 'abc', 'xyz') |
| |
| def test_split(self): |
| string_tests.CommonTest.test_split(self) |
| |
| # test mixed kinds |
| for left, right in ('ba', '\u0101\u0100', '\U00010301\U00010300'): |
| left *= 9 |
| right *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| self.checkequal([left + right], |
| left + right, 'split', delim) |
| self.checkequal([left, right], |
| left + delim + right, 'split', delim) |
| self.checkequal([left + right], |
| left + right, 'split', delim * 2) |
| self.checkequal([left, right], |
| left + delim * 2 + right, 'split', delim *2) |
| |
| def test_rsplit(self): |
| string_tests.CommonTest.test_rsplit(self) |
| # test mixed kinds |
| for left, right in ('ba', '\u0101\u0100', '\U00010301\U00010300'): |
| left *= 9 |
| right *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| self.checkequal([left + right], |
| left + right, 'rsplit', delim) |
| self.checkequal([left, right], |
| left + delim + right, 'rsplit', delim) |
| self.checkequal([left + right], |
| left + right, 'rsplit', delim * 2) |
| self.checkequal([left, right], |
| left + delim * 2 + right, 'rsplit', delim *2) |
| |
| def test_partition(self): |
| string_tests.MixinStrUnicodeUserStringTest.test_partition(self) |
| # test mixed kinds |
| self.checkequal(('ABCDEFGH', '', ''), 'ABCDEFGH', 'partition', '\u4200') |
| for left, right in ('ba', '\u0101\u0100', '\U00010301\U00010300'): |
| left *= 9 |
| right *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| self.checkequal((left + right, '', ''), |
| left + right, 'partition', delim) |
| self.checkequal((left, delim, right), |
| left + delim + right, 'partition', delim) |
| self.checkequal((left + right, '', ''), |
| left + right, 'partition', delim * 2) |
| self.checkequal((left, delim * 2, right), |
| left + delim * 2 + right, 'partition', delim * 2) |
| |
| def test_rpartition(self): |
| string_tests.MixinStrUnicodeUserStringTest.test_rpartition(self) |
| # test mixed kinds |
| self.checkequal(('', '', 'ABCDEFGH'), 'ABCDEFGH', 'rpartition', '\u4200') |
| for left, right in ('ba', '\u0101\u0100', '\U00010301\U00010300'): |
| left *= 9 |
| right *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| self.checkequal(('', '', left + right), |
| left + right, 'rpartition', delim) |
| self.checkequal((left, delim, right), |
| left + delim + right, 'rpartition', delim) |
| self.checkequal(('', '', left + right), |
| left + right, 'rpartition', delim * 2) |
| self.checkequal((left, delim * 2, right), |
| left + delim * 2 + right, 'rpartition', delim * 2) |
| |
| def test_join(self): |
| string_tests.MixinStrUnicodeUserStringTest.test_join(self) |
| |
| class MyWrapper: |
| def __init__(self, sval): self.sval = sval |
| def __str__(self): return self.sval |
| |
| # mixed arguments |
| self.checkequalnofix('a b c d', ' ', 'join', ['a', 'b', 'c', 'd']) |
| self.checkequalnofix('abcd', '', 'join', ('a', 'b', 'c', 'd')) |
| self.checkequalnofix('w x y z', ' ', 'join', string_tests.Sequence('wxyz')) |
| self.checkequalnofix('a b c d', ' ', 'join', ['a', 'b', 'c', 'd']) |
| self.checkequalnofix('a b c d', ' ', 'join', ['a', 'b', 'c', 'd']) |
| self.checkequalnofix('abcd', '', 'join', ('a', 'b', 'c', 'd')) |
| self.checkequalnofix('w x y z', ' ', 'join', string_tests.Sequence('wxyz')) |
| self.checkraises(TypeError, ' ', 'join', ['1', '2', MyWrapper('foo')]) |
| self.checkraises(TypeError, ' ', 'join', ['1', '2', '3', bytes()]) |
| self.checkraises(TypeError, ' ', 'join', [1, 2, 3]) |
| self.checkraises(TypeError, ' ', 'join', ['1', '2', 3]) |
| |
| @unittest.skipIf(sys.maxsize > 2**32, |
| 'needs too much memory on a 64-bit platform') |
| def test_join_overflow(self): |
| size = int(sys.maxsize**0.5) + 1 |
| seq = ('A' * size,) * size |
| self.assertRaises(OverflowError, ''.join, seq) |
| |
| def test_replace(self): |
| string_tests.CommonTest.test_replace(self) |
| |
| # method call forwarded from str implementation because of unicode argument |
| self.checkequalnofix('one@two!three!', 'one!two!three!', 'replace', '!', '@', 1) |
| self.assertRaises(TypeError, 'replace'.replace, "r", 42) |
| # test mixed kinds |
| for left, right in ('ba', '\u0101\u0100', '\U00010301\U00010300'): |
| left *= 9 |
| right *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| for repl in ('d', '\u0103', '\U00010303'): |
| self.checkequal(left + right, |
| left + right, 'replace', delim, repl) |
| self.checkequal(left + repl + right, |
| left + delim + right, |
| 'replace', delim, repl) |
| self.checkequal(left + right, |
| left + right, 'replace', delim * 2, repl) |
| self.checkequal(left + repl + right, |
| left + delim * 2 + right, |
| 'replace', delim * 2, repl) |
| |
| @support.cpython_only |
| def test_replace_id(self): |
| pattern = 'abc' |
| text = 'abc def' |
| self.assertIs(text.replace(pattern, pattern), text) |
| |
| def test_bytes_comparison(self): |
| with support.check_warnings(): |
| warnings.simplefilter('ignore', BytesWarning) |
| self.assertEqual('abc' == b'abc', False) |
| self.assertEqual('abc' != b'abc', True) |
| self.assertEqual('abc' == bytearray(b'abc'), False) |
| self.assertEqual('abc' != bytearray(b'abc'), True) |
| |
| def test_comparison(self): |
| # Comparisons: |
| self.assertEqual('abc', 'abc') |
| self.assertTrue('abcd' > 'abc') |
| self.assertTrue('abc' < 'abcd') |
| |
| if 0: |
| # Move these tests to a Unicode collation module test... |
| # Testing UTF-16 code point order comparisons... |
| |
| # No surrogates, no fixup required. |
| self.assertTrue('\u0061' < '\u20ac') |
| # Non surrogate below surrogate value, no fixup required |
| self.assertTrue('\u0061' < '\ud800\udc02') |
| |
| # Non surrogate above surrogate value, fixup required |
| def test_lecmp(s, s2): |
| self.assertTrue(s < s2) |
| |
| def test_fixup(s): |
| s2 = '\ud800\udc01' |
| test_lecmp(s, s2) |
| s2 = '\ud900\udc01' |
| test_lecmp(s, s2) |
| s2 = '\uda00\udc01' |
| test_lecmp(s, s2) |
| s2 = '\udb00\udc01' |
| test_lecmp(s, s2) |
| s2 = '\ud800\udd01' |
| test_lecmp(s, s2) |
| s2 = '\ud900\udd01' |
| test_lecmp(s, s2) |
| s2 = '\uda00\udd01' |
| test_lecmp(s, s2) |
| s2 = '\udb00\udd01' |
| test_lecmp(s, s2) |
| s2 = '\ud800\ude01' |
| test_lecmp(s, s2) |
| s2 = '\ud900\ude01' |
| test_lecmp(s, s2) |
| s2 = '\uda00\ude01' |
| test_lecmp(s, s2) |
| s2 = '\udb00\ude01' |
| test_lecmp(s, s2) |
| s2 = '\ud800\udfff' |
| test_lecmp(s, s2) |
| s2 = '\ud900\udfff' |
| test_lecmp(s, s2) |
| s2 = '\uda00\udfff' |
| test_lecmp(s, s2) |
| s2 = '\udb00\udfff' |
| test_lecmp(s, s2) |
| |
| test_fixup('\ue000') |
| test_fixup('\uff61') |
| |
| # Surrogates on both sides, no fixup required |
| self.assertTrue('\ud800\udc02' < '\ud84d\udc56') |
| |
| def test_islower(self): |
| super().test_islower() |
| self.checkequalnofix(False, '\u1FFc', 'islower') |
| self.assertFalse('\u2167'.islower()) |
| self.assertTrue('\u2177'.islower()) |
| # non-BMP, uppercase |
| self.assertFalse('\U00010401'.islower()) |
| self.assertFalse('\U00010427'.islower()) |
| # non-BMP, lowercase |
| self.assertTrue('\U00010429'.islower()) |
| self.assertTrue('\U0001044E'.islower()) |
| # non-BMP, non-cased |
| self.assertFalse('\U0001F40D'.islower()) |
| self.assertFalse('\U0001F46F'.islower()) |
| |
| def test_isupper(self): |
| super().test_isupper() |
| if not sys.platform.startswith('java'): |
| self.checkequalnofix(False, '\u1FFc', 'isupper') |
| self.assertTrue('\u2167'.isupper()) |
| self.assertFalse('\u2177'.isupper()) |
| # non-BMP, uppercase |
| self.assertTrue('\U00010401'.isupper()) |
| self.assertTrue('\U00010427'.isupper()) |
| # non-BMP, lowercase |
| self.assertFalse('\U00010429'.isupper()) |
| self.assertFalse('\U0001044E'.isupper()) |
| # non-BMP, non-cased |
| self.assertFalse('\U0001F40D'.isupper()) |
| self.assertFalse('\U0001F46F'.isupper()) |
| |
| def test_istitle(self): |
| super().test_istitle() |
| self.checkequalnofix(True, '\u1FFc', 'istitle') |
| self.checkequalnofix(True, 'Greek \u1FFcitlecases ...', 'istitle') |
| |
| # non-BMP, uppercase + lowercase |
| self.assertTrue('\U00010401\U00010429'.istitle()) |
| self.assertTrue('\U00010427\U0001044E'.istitle()) |
| # apparently there are no titlecased (Lt) non-BMP chars in Unicode 6 |
| for ch in ['\U00010429', '\U0001044E', '\U0001F40D', '\U0001F46F']: |
| self.assertFalse(ch.istitle(), '{!a} is not title'.format(ch)) |
| |
| def test_isspace(self): |
| super().test_isspace() |
| self.checkequalnofix(True, '\u2000', 'isspace') |
| self.checkequalnofix(True, '\u200a', 'isspace') |
| self.checkequalnofix(False, '\u2014', 'isspace') |
| # apparently there are no non-BMP spaces chars in Unicode 6 |
| for ch in ['\U00010401', '\U00010427', '\U00010429', '\U0001044E', |
| '\U0001F40D', '\U0001F46F']: |
| self.assertFalse(ch.isspace(), '{!a} is not space.'.format(ch)) |
| |
| def test_isalnum(self): |
| super().test_isalnum() |
| for ch in ['\U00010401', '\U00010427', '\U00010429', '\U0001044E', |
| '\U0001D7F6', '\U00011066', '\U000104A0', '\U0001F107']: |
| self.assertTrue(ch.isalnum(), '{!a} is alnum.'.format(ch)) |
| |
| def test_isalpha(self): |
| super().test_isalpha() |
| self.checkequalnofix(True, '\u1FFc', 'isalpha') |
| # non-BMP, cased |
| self.assertTrue('\U00010401'.isalpha()) |
| self.assertTrue('\U00010427'.isalpha()) |
| self.assertTrue('\U00010429'.isalpha()) |
| self.assertTrue('\U0001044E'.isalpha()) |
| # non-BMP, non-cased |
| self.assertFalse('\U0001F40D'.isalpha()) |
| self.assertFalse('\U0001F46F'.isalpha()) |
| |
| def test_isdecimal(self): |
| self.checkequalnofix(False, '', 'isdecimal') |
| self.checkequalnofix(False, 'a', 'isdecimal') |
| self.checkequalnofix(True, '0', 'isdecimal') |
| self.checkequalnofix(False, '\u2460', 'isdecimal') # CIRCLED DIGIT ONE |
| self.checkequalnofix(False, '\xbc', 'isdecimal') # VULGAR FRACTION ONE QUARTER |
| self.checkequalnofix(True, '\u0660', 'isdecimal') # ARABIC-INDIC DIGIT ZERO |
| self.checkequalnofix(True, '0123456789', 'isdecimal') |
| self.checkequalnofix(False, '0123456789a', 'isdecimal') |
| |
| self.checkraises(TypeError, 'abc', 'isdecimal', 42) |
| |
| for ch in ['\U00010401', '\U00010427', '\U00010429', '\U0001044E', |
| '\U0001F40D', '\U0001F46F', '\U00011065', '\U0001F107']: |
| self.assertFalse(ch.isdecimal(), '{!a} is not decimal.'.format(ch)) |
| for ch in ['\U0001D7F6', '\U00011066', '\U000104A0']: |
| self.assertTrue(ch.isdecimal(), '{!a} is decimal.'.format(ch)) |
| |
| def test_isdigit(self): |
| super().test_isdigit() |
| self.checkequalnofix(True, '\u2460', 'isdigit') |
| self.checkequalnofix(False, '\xbc', 'isdigit') |
| self.checkequalnofix(True, '\u0660', 'isdigit') |
| |
| for ch in ['\U00010401', '\U00010427', '\U00010429', '\U0001044E', |
| '\U0001F40D', '\U0001F46F', '\U00011065']: |
| self.assertFalse(ch.isdigit(), '{!a} is not a digit.'.format(ch)) |
| for ch in ['\U0001D7F6', '\U00011066', '\U000104A0', '\U0001F107']: |
| self.assertTrue(ch.isdigit(), '{!a} is a digit.'.format(ch)) |
| |
| def test_isnumeric(self): |
| self.checkequalnofix(False, '', 'isnumeric') |
| self.checkequalnofix(False, 'a', 'isnumeric') |
| self.checkequalnofix(True, '0', 'isnumeric') |
| self.checkequalnofix(True, '\u2460', 'isnumeric') |
| self.checkequalnofix(True, '\xbc', 'isnumeric') |
| self.checkequalnofix(True, '\u0660', 'isnumeric') |
| self.checkequalnofix(True, '0123456789', 'isnumeric') |
| self.checkequalnofix(False, '0123456789a', 'isnumeric') |
| |
| self.assertRaises(TypeError, "abc".isnumeric, 42) |
| |
| for ch in ['\U00010401', '\U00010427', '\U00010429', '\U0001044E', |
| '\U0001F40D', '\U0001F46F']: |
| self.assertFalse(ch.isnumeric(), '{!a} is not numeric.'.format(ch)) |
| for ch in ['\U00011065', '\U0001D7F6', '\U00011066', |
| '\U000104A0', '\U0001F107']: |
| self.assertTrue(ch.isnumeric(), '{!a} is numeric.'.format(ch)) |
| |
| def test_isidentifier(self): |
| self.assertTrue("a".isidentifier()) |
| self.assertTrue("Z".isidentifier()) |
| self.assertTrue("_".isidentifier()) |
| self.assertTrue("b0".isidentifier()) |
| self.assertTrue("bc".isidentifier()) |
| self.assertTrue("b_".isidentifier()) |
| self.assertTrue("µ".isidentifier()) |
| self.assertTrue("𝔘𝔫𝔦𝔠𝔬𝔡𝔢".isidentifier()) |
| |
| self.assertFalse(" ".isidentifier()) |
| self.assertFalse("[".isidentifier()) |
| self.assertFalse("©".isidentifier()) |
| self.assertFalse("0".isidentifier()) |
| |
| def test_isprintable(self): |
| self.assertTrue("".isprintable()) |
| self.assertTrue(" ".isprintable()) |
| self.assertTrue("abcdefg".isprintable()) |
| self.assertFalse("abcdefg\n".isprintable()) |
| # some defined Unicode character |
| self.assertTrue("\u0374".isprintable()) |
| # undefined character |
| self.assertFalse("\u0378".isprintable()) |
| # single surrogate character |
| self.assertFalse("\ud800".isprintable()) |
| |
| self.assertTrue('\U0001F46F'.isprintable()) |
| self.assertFalse('\U000E0020'.isprintable()) |
| |
| def test_surrogates(self): |
| for s in ('a\uD800b\uDFFF', 'a\uDFFFb\uD800', |
| 'a\uD800b\uDFFFa', 'a\uDFFFb\uD800a'): |
| self.assertTrue(s.islower()) |
| self.assertFalse(s.isupper()) |
| self.assertFalse(s.istitle()) |
| for s in ('A\uD800B\uDFFF', 'A\uDFFFB\uD800', |
| 'A\uD800B\uDFFFA', 'A\uDFFFB\uD800A'): |
| self.assertFalse(s.islower()) |
| self.assertTrue(s.isupper()) |
| self.assertTrue(s.istitle()) |
| |
| for meth_name in ('islower', 'isupper', 'istitle'): |
| meth = getattr(str, meth_name) |
| for s in ('\uD800', '\uDFFF', '\uD800\uD800', '\uDFFF\uDFFF'): |
| self.assertFalse(meth(s), '%a.%s() is False' % (s, meth_name)) |
| |
| for meth_name in ('isalpha', 'isalnum', 'isdigit', 'isspace', |
| 'isdecimal', 'isnumeric', |
| 'isidentifier', 'isprintable'): |
| meth = getattr(str, meth_name) |
| for s in ('\uD800', '\uDFFF', '\uD800\uD800', '\uDFFF\uDFFF', |
| 'a\uD800b\uDFFF', 'a\uDFFFb\uD800', |
| 'a\uD800b\uDFFFa', 'a\uDFFFb\uD800a'): |
| self.assertFalse(meth(s), '%a.%s() is False' % (s, meth_name)) |
| |
| |
| def test_lower(self): |
| string_tests.CommonTest.test_lower(self) |
| self.assertEqual('\U00010427'.lower(), '\U0001044F') |
| self.assertEqual('\U00010427\U00010427'.lower(), |
| '\U0001044F\U0001044F') |
| self.assertEqual('\U00010427\U0001044F'.lower(), |
| '\U0001044F\U0001044F') |
| self.assertEqual('X\U00010427x\U0001044F'.lower(), |
| 'x\U0001044Fx\U0001044F') |
| self.assertEqual('fi'.lower(), 'fi') |
| self.assertEqual('\u0130'.lower(), '\u0069\u0307') |
| # Special case for GREEK CAPITAL LETTER SIGMA U+03A3 |
| self.assertEqual('\u03a3'.lower(), '\u03c3') |
| self.assertEqual('\u0345\u03a3'.lower(), '\u0345\u03c3') |
| self.assertEqual('A\u0345\u03a3'.lower(), 'a\u0345\u03c2') |
| self.assertEqual('A\u0345\u03a3a'.lower(), 'a\u0345\u03c3a') |
| self.assertEqual('A\u0345\u03a3'.lower(), 'a\u0345\u03c2') |
| self.assertEqual('A\u03a3\u0345'.lower(), 'a\u03c2\u0345') |
| self.assertEqual('\u03a3\u0345 '.lower(), '\u03c3\u0345 ') |
| self.assertEqual('\U0008fffe'.lower(), '\U0008fffe') |
| self.assertEqual('\u2177'.lower(), '\u2177') |
| |
| def test_casefold(self): |
| self.assertEqual('hello'.casefold(), 'hello') |
| self.assertEqual('hELlo'.casefold(), 'hello') |
| self.assertEqual('ß'.casefold(), 'ss') |
| self.assertEqual('fi'.casefold(), 'fi') |
| self.assertEqual('\u03a3'.casefold(), '\u03c3') |
| self.assertEqual('A\u0345\u03a3'.casefold(), 'a\u03b9\u03c3') |
| self.assertEqual('\u00b5'.casefold(), '\u03bc') |
| |
| def test_upper(self): |
| string_tests.CommonTest.test_upper(self) |
| self.assertEqual('\U0001044F'.upper(), '\U00010427') |
| self.assertEqual('\U0001044F\U0001044F'.upper(), |
| '\U00010427\U00010427') |
| self.assertEqual('\U00010427\U0001044F'.upper(), |
| '\U00010427\U00010427') |
| self.assertEqual('X\U00010427x\U0001044F'.upper(), |
| 'X\U00010427X\U00010427') |
| self.assertEqual('fi'.upper(), 'FI') |
| self.assertEqual('\u0130'.upper(), '\u0130') |
| self.assertEqual('\u03a3'.upper(), '\u03a3') |
| self.assertEqual('ß'.upper(), 'SS') |
| self.assertEqual('\u1fd2'.upper(), '\u0399\u0308\u0300') |
| self.assertEqual('\U0008fffe'.upper(), '\U0008fffe') |
| self.assertEqual('\u2177'.upper(), '\u2167') |
| |
| def test_capitalize(self): |
| string_tests.CommonTest.test_capitalize(self) |
| self.assertEqual('\U0001044F'.capitalize(), '\U00010427') |
| self.assertEqual('\U0001044F\U0001044F'.capitalize(), |
| '\U00010427\U0001044F') |
| self.assertEqual('\U00010427\U0001044F'.capitalize(), |
| '\U00010427\U0001044F') |
| self.assertEqual('\U0001044F\U00010427'.capitalize(), |
| '\U00010427\U0001044F') |
| self.assertEqual('X\U00010427x\U0001044F'.capitalize(), |
| 'X\U0001044Fx\U0001044F') |
| self.assertEqual('h\u0130'.capitalize(), 'H\u0069\u0307') |
| exp = '\u0399\u0308\u0300\u0069\u0307' |
| self.assertEqual('\u1fd2\u0130'.capitalize(), exp) |
| self.assertEqual('finnish'.capitalize(), 'FInnish') |
| self.assertEqual('A\u0345\u03a3'.capitalize(), 'A\u0345\u03c2') |
| |
| def test_title(self): |
| super().test_title() |
| self.assertEqual('\U0001044F'.title(), '\U00010427') |
| self.assertEqual('\U0001044F\U0001044F'.title(), |
| '\U00010427\U0001044F') |
| self.assertEqual('\U0001044F\U0001044F \U0001044F\U0001044F'.title(), |
| '\U00010427\U0001044F \U00010427\U0001044F') |
| self.assertEqual('\U00010427\U0001044F \U00010427\U0001044F'.title(), |
| '\U00010427\U0001044F \U00010427\U0001044F') |
| self.assertEqual('\U0001044F\U00010427 \U0001044F\U00010427'.title(), |
| '\U00010427\U0001044F \U00010427\U0001044F') |
| self.assertEqual('X\U00010427x\U0001044F X\U00010427x\U0001044F'.title(), |
| 'X\U0001044Fx\U0001044F X\U0001044Fx\U0001044F') |
| self.assertEqual('fiNNISH'.title(), 'Finnish') |
| self.assertEqual('A\u03a3 \u1fa1xy'.title(), 'A\u03c2 \u1fa9xy') |
| self.assertEqual('A\u03a3A'.title(), 'A\u03c3a') |
| |
| def test_swapcase(self): |
| string_tests.CommonTest.test_swapcase(self) |
| self.assertEqual('\U0001044F'.swapcase(), '\U00010427') |
| self.assertEqual('\U00010427'.swapcase(), '\U0001044F') |
| self.assertEqual('\U0001044F\U0001044F'.swapcase(), |
| '\U00010427\U00010427') |
| self.assertEqual('\U00010427\U0001044F'.swapcase(), |
| '\U0001044F\U00010427') |
| self.assertEqual('\U0001044F\U00010427'.swapcase(), |
| '\U00010427\U0001044F') |
| self.assertEqual('X\U00010427x\U0001044F'.swapcase(), |
| 'x\U0001044FX\U00010427') |
| self.assertEqual('fi'.swapcase(), 'FI') |
| self.assertEqual('\u0130'.swapcase(), '\u0069\u0307') |
| # Special case for GREEK CAPITAL LETTER SIGMA U+03A3 |
| self.assertEqual('\u03a3'.swapcase(), '\u03c3') |
| self.assertEqual('\u0345\u03a3'.swapcase(), '\u0399\u03c3') |
| self.assertEqual('A\u0345\u03a3'.swapcase(), 'a\u0399\u03c2') |
| self.assertEqual('A\u0345\u03a3a'.swapcase(), 'a\u0399\u03c3A') |
| self.assertEqual('A\u0345\u03a3'.swapcase(), 'a\u0399\u03c2') |
| self.assertEqual('A\u03a3\u0345'.swapcase(), 'a\u03c2\u0399') |
| self.assertEqual('\u03a3\u0345 '.swapcase(), '\u03c3\u0399 ') |
| self.assertEqual('\u03a3'.swapcase(), '\u03c3') |
| self.assertEqual('ß'.swapcase(), 'SS') |
| self.assertEqual('\u1fd2'.swapcase(), '\u0399\u0308\u0300') |
| |
| def test_center(self): |
| string_tests.CommonTest.test_center(self) |
| self.assertEqual('x'.center(2, '\U0010FFFF'), |
| 'x\U0010FFFF') |
| self.assertEqual('x'.center(3, '\U0010FFFF'), |
| '\U0010FFFFx\U0010FFFF') |
| self.assertEqual('x'.center(4, '\U0010FFFF'), |
| '\U0010FFFFx\U0010FFFF\U0010FFFF') |
| |
| @unittest.skipUnless(sys.maxsize == 2**31 - 1, "requires 32-bit system") |
| @support.cpython_only |
| def test_case_operation_overflow(self): |
| # Issue #22643 |
| size = 2**32//12 + 1 |
| try: |
| s = "ü" * size |
| except MemoryError: |
| self.skipTest('no enough memory (%.0f MiB required)' % (size / 2**20)) |
| try: |
| self.assertRaises(OverflowError, s.upper) |
| finally: |
| del s |
| |
| def test_contains(self): |
| # Testing Unicode contains method |
| self.assertIn('a', 'abdb') |
| self.assertIn('a', 'bdab') |
| self.assertIn('a', 'bdaba') |
| self.assertIn('a', 'bdba') |
| self.assertNotIn('a', 'bdb') |
| self.assertIn('a', 'bdba') |
| self.assertIn('a', ('a',1,None)) |
| self.assertIn('a', (1,None,'a')) |
| self.assertIn('a', ('a',1,None)) |
| self.assertIn('a', (1,None,'a')) |
| self.assertNotIn('a', ('x',1,'y')) |
| self.assertNotIn('a', ('x',1,None)) |
| self.assertNotIn('abcd', 'abcxxxx') |
| self.assertIn('ab', 'abcd') |
| self.assertIn('ab', 'abc') |
| self.assertIn('ab', (1,None,'ab')) |
| self.assertIn('', 'abc') |
| self.assertIn('', '') |
| self.assertIn('', 'abc') |
| self.assertNotIn('\0', 'abc') |
| self.assertIn('\0', '\0abc') |
| self.assertIn('\0', 'abc\0') |
| self.assertIn('a', '\0abc') |
| self.assertIn('asdf', 'asdf') |
| self.assertNotIn('asdf', 'asd') |
| self.assertNotIn('asdf', '') |
| |
| self.assertRaises(TypeError, "abc".__contains__) |
| # test mixed kinds |
| for fill in ('a', '\u0100', '\U00010300'): |
| fill *= 9 |
| for delim in ('c', '\u0102', '\U00010302'): |
| self.assertNotIn(delim, fill) |
| self.assertIn(delim, fill + delim) |
| self.assertNotIn(delim * 2, fill) |
| self.assertIn(delim * 2, fill + delim * 2) |
| |
| def test_issue18183(self): |
| '\U00010000\U00100000'.lower() |
| '\U00010000\U00100000'.casefold() |
| '\U00010000\U00100000'.upper() |
| '\U00010000\U00100000'.capitalize() |
| '\U00010000\U00100000'.title() |
| '\U00010000\U00100000'.swapcase() |
| '\U00100000'.center(3, '\U00010000') |
| '\U00100000'.ljust(3, '\U00010000') |
| '\U00100000'.rjust(3, '\U00010000') |
| |
| def test_format(self): |
| self.assertEqual(''.format(), '') |
| self.assertEqual('a'.format(), 'a') |
| self.assertEqual('ab'.format(), 'ab') |
| self.assertEqual('a{{'.format(), 'a{') |
| self.assertEqual('a}}'.format(), 'a}') |
| self.assertEqual('{{b'.format(), '{b') |
| self.assertEqual('}}b'.format(), '}b') |
| self.assertEqual('a{{b'.format(), 'a{b') |
| |
| # examples from the PEP: |
| import datetime |
| self.assertEqual("My name is {0}".format('Fred'), "My name is Fred") |
| self.assertEqual("My name is {0[name]}".format(dict(name='Fred')), |
| "My name is Fred") |
| self.assertEqual("My name is {0} :-{{}}".format('Fred'), |
| "My name is Fred :-{}") |
| |
| d = datetime.date(2007, 8, 18) |
| self.assertEqual("The year is {0.year}".format(d), |
| "The year is 2007") |
| |
| # classes we'll use for testing |
| class C: |
| def __init__(self, x=100): |
| self._x = x |
| def __format__(self, spec): |
| return spec |
| |
| class D: |
| def __init__(self, x): |
| self.x = x |
| def __format__(self, spec): |
| return str(self.x) |
| |
| # class with __str__, but no __format__ |
| class E: |
| def __init__(self, x): |
| self.x = x |
| def __str__(self): |
| return 'E(' + self.x + ')' |
| |
| # class with __repr__, but no __format__ or __str__ |
| class F: |
| def __init__(self, x): |
| self.x = x |
| def __repr__(self): |
| return 'F(' + self.x + ')' |
| |
| # class with __format__ that forwards to string, for some format_spec's |
| class G: |
| def __init__(self, x): |
| self.x = x |
| def __str__(self): |
| return "string is " + self.x |
| def __format__(self, format_spec): |
| if format_spec == 'd': |
| return 'G(' + self.x + ')' |
| return object.__format__(self, format_spec) |
| |
| class I(datetime.date): |
| def __format__(self, format_spec): |
| return self.strftime(format_spec) |
| |
| class J(int): |
| def __format__(self, format_spec): |
| return int.__format__(self * 2, format_spec) |
| |
| class M: |
| def __init__(self, x): |
| self.x = x |
| def __repr__(self): |
| return 'M(' + self.x + ')' |
| __str__ = None |
| |
| class N: |
| def __init__(self, x): |
| self.x = x |
| def __repr__(self): |
| return 'N(' + self.x + ')' |
| __format__ = None |
| |
| self.assertEqual(''.format(), '') |
| self.assertEqual('abc'.format(), 'abc') |
| self.assertEqual('{0}'.format('abc'), 'abc') |
| self.assertEqual('{0:}'.format('abc'), 'abc') |
| # self.assertEqual('{ 0 }'.format('abc'), 'abc') |
| self.assertEqual('X{0}'.format('abc'), 'Xabc') |
| self.assertEqual('{0}X'.format('abc'), 'abcX') |
| self.assertEqual('X{0}Y'.format('abc'), 'XabcY') |
| self.assertEqual('{1}'.format(1, 'abc'), 'abc') |
| self.assertEqual('X{1}'.format(1, 'abc'), 'Xabc') |
| self.assertEqual('{1}X'.format(1, 'abc'), 'abcX') |
| self.assertEqual('X{1}Y'.format(1, 'abc'), 'XabcY') |
| self.assertEqual('{0}'.format(-15), '-15') |
| self.assertEqual('{0}{1}'.format(-15, 'abc'), '-15abc') |
| self.assertEqual('{0}X{1}'.format(-15, 'abc'), '-15Xabc') |
| self.assertEqual('{{'.format(), '{') |
| self.assertEqual('}}'.format(), '}') |
| self.assertEqual('{{}}'.format(), '{}') |
| self.assertEqual('{{x}}'.format(), '{x}') |
| self.assertEqual('{{{0}}}'.format(123), '{123}') |
| self.assertEqual('{{{{0}}}}'.format(), '{{0}}') |
| self.assertEqual('}}{{'.format(), '}{') |
| self.assertEqual('}}x{{'.format(), '}x{') |
| |
| # weird field names |
| self.assertEqual("{0[foo-bar]}".format({'foo-bar':'baz'}), 'baz') |
| self.assertEqual("{0[foo bar]}".format({'foo bar':'baz'}), 'baz') |
| self.assertEqual("{0[ ]}".format({' ':3}), '3') |
| |
| self.assertEqual('{foo._x}'.format(foo=C(20)), '20') |
| self.assertEqual('{1}{0}'.format(D(10), D(20)), '2010') |
| self.assertEqual('{0._x.x}'.format(C(D('abc'))), 'abc') |
| self.assertEqual('{0[0]}'.format(['abc', 'def']), 'abc') |
| self.assertEqual('{0[1]}'.format(['abc', 'def']), 'def') |
| self.assertEqual('{0[1][0]}'.format(['abc', ['def']]), 'def') |
| self.assertEqual('{0[1][0].x}'.format(['abc', [D('def')]]), 'def') |
| |
| # strings |
| self.assertEqual('{0:.3s}'.format('abc'), 'abc') |
| self.assertEqual('{0:.3s}'.format('ab'), 'ab') |
| self.assertEqual('{0:.3s}'.format('abcdef'), 'abc') |
| self.assertEqual('{0:.0s}'.format('abcdef'), '') |
| self.assertEqual('{0:3.3s}'.format('abc'), 'abc') |
| self.assertEqual('{0:2.3s}'.format('abc'), 'abc') |
| self.assertEqual('{0:2.2s}'.format('abc'), 'ab') |
| self.assertEqual('{0:3.2s}'.format('abc'), 'ab ') |
| self.assertEqual('{0:x<0s}'.format('result'), 'result') |
| self.assertEqual('{0:x<5s}'.format('result'), 'result') |
| self.assertEqual('{0:x<6s}'.format('result'), 'result') |
| self.assertEqual('{0:x<7s}'.format('result'), 'resultx') |
| self.assertEqual('{0:x<8s}'.format('result'), 'resultxx') |
| self.assertEqual('{0: <7s}'.format('result'), 'result ') |
| self.assertEqual('{0:<7s}'.format('result'), 'result ') |
| self.assertEqual('{0:>7s}'.format('result'), ' result') |
| self.assertEqual('{0:>8s}'.format('result'), ' result') |
| self.assertEqual('{0:^8s}'.format('result'), ' result ') |
| self.assertEqual('{0:^9s}'.format('result'), ' result ') |
| self.assertEqual('{0:^10s}'.format('result'), ' result ') |
| self.assertEqual('{0:10000}'.format('a'), 'a' + ' ' * 9999) |
| self.assertEqual('{0:10000}'.format(''), ' ' * 10000) |
| self.assertEqual('{0:10000000}'.format(''), ' ' * 10000000) |
| |
| # issue 12546: use \x00 as a fill character |
| self.assertEqual('{0:\x00<6s}'.format('foo'), 'foo\x00\x00\x00') |
| self.assertEqual('{0:\x01<6s}'.format('foo'), 'foo\x01\x01\x01') |
| self.assertEqual('{0:\x00^6s}'.format('foo'), '\x00foo\x00\x00') |
| self.assertEqual('{0:^6s}'.format('foo'), ' foo ') |
| |
| self.assertEqual('{0:\x00<6}'.format(3), '3\x00\x00\x00\x00\x00') |
| self.assertEqual('{0:\x01<6}'.format(3), '3\x01\x01\x01\x01\x01') |
| self.assertEqual('{0:\x00^6}'.format(3), '\x00\x003\x00\x00\x00') |
| self.assertEqual('{0:<6}'.format(3), '3 ') |
| |
| self.assertEqual('{0:\x00<6}'.format(3.14), '3.14\x00\x00') |
| self.assertEqual('{0:\x01<6}'.format(3.14), '3.14\x01\x01') |
| self.assertEqual('{0:\x00^6}'.format(3.14), '\x003.14\x00') |
| self.assertEqual('{0:^6}'.format(3.14), ' 3.14 ') |
| |
| self.assertEqual('{0:\x00<12}'.format(3+2.0j), '(3+2j)\x00\x00\x00\x00\x00\x00') |
| self.assertEqual('{0:\x01<12}'.format(3+2.0j), '(3+2j)\x01\x01\x01\x01\x01\x01') |
| self.assertEqual('{0:\x00^12}'.format(3+2.0j), '\x00\x00\x00(3+2j)\x00\x00\x00') |
| self.assertEqual('{0:^12}'.format(3+2.0j), ' (3+2j) ') |
| |
| # format specifiers for user defined type |
| self.assertEqual('{0:abc}'.format(C()), 'abc') |
| |
| # !r, !s and !a coercions |
| self.assertEqual('{0!s}'.format('Hello'), 'Hello') |
| self.assertEqual('{0!s:}'.format('Hello'), 'Hello') |
| self.assertEqual('{0!s:15}'.format('Hello'), 'Hello ') |
| self.assertEqual('{0!s:15s}'.format('Hello'), 'Hello ') |
| self.assertEqual('{0!r}'.format('Hello'), "'Hello'") |
| self.assertEqual('{0!r:}'.format('Hello'), "'Hello'") |
| self.assertEqual('{0!r}'.format(F('Hello')), 'F(Hello)') |
| self.assertEqual('{0!r}'.format('\u0378'), "'\\u0378'") # nonprintable |
| self.assertEqual('{0!r}'.format('\u0374'), "'\u0374'") # printable |
| self.assertEqual('{0!r}'.format(F('\u0374')), 'F(\u0374)') |
| self.assertEqual('{0!a}'.format('Hello'), "'Hello'") |
| self.assertEqual('{0!a}'.format('\u0378'), "'\\u0378'") # nonprintable |
| self.assertEqual('{0!a}'.format('\u0374'), "'\\u0374'") # printable |
| self.assertEqual('{0!a:}'.format('Hello'), "'Hello'") |
| self.assertEqual('{0!a}'.format(F('Hello')), 'F(Hello)') |
| self.assertEqual('{0!a}'.format(F('\u0374')), 'F(\\u0374)') |
| |
| # test fallback to object.__format__ |
| self.assertEqual('{0}'.format({}), '{}') |
| self.assertEqual('{0}'.format([]), '[]') |
| self.assertEqual('{0}'.format([1]), '[1]') |
| |
| self.assertEqual('{0:d}'.format(G('data')), 'G(data)') |
| self.assertEqual('{0!s}'.format(G('data')), 'string is data') |
| |
| self.assertRaises(TypeError, '{0:^10}'.format, E('data')) |
| self.assertRaises(TypeError, '{0:^10s}'.format, E('data')) |
| self.assertRaises(TypeError, '{0:>15s}'.format, G('data')) |
| |
| self.assertEqual("{0:date: %Y-%m-%d}".format(I(year=2007, |
| month=8, |
| day=27)), |
| "date: 2007-08-27") |
| |
| # test deriving from a builtin type and overriding __format__ |
| self.assertEqual("{0}".format(J(10)), "20") |
| |
| |
| # string format specifiers |
| self.assertEqual('{0:}'.format('a'), 'a') |
| |
| # computed format specifiers |
| self.assertEqual("{0:.{1}}".format('hello world', 5), 'hello') |
| self.assertEqual("{0:.{1}s}".format('hello world', 5), 'hello') |
| self.assertEqual("{0:.{precision}s}".format('hello world', precision=5), 'hello') |
| self.assertEqual("{0:{width}.{precision}s}".format('hello world', width=10, precision=5), 'hello ') |
| self.assertEqual("{0:{width}.{precision}s}".format('hello world', width='10', precision='5'), 'hello ') |
| |
| # test various errors |
| self.assertRaises(ValueError, '{'.format) |
| self.assertRaises(ValueError, '}'.format) |
| self.assertRaises(ValueError, 'a{'.format) |
| self.assertRaises(ValueError, 'a}'.format) |
| self.assertRaises(ValueError, '{a'.format) |
| self.assertRaises(ValueError, '}a'.format) |
| self.assertRaises(IndexError, '{0}'.format) |
| self.assertRaises(IndexError, '{1}'.format, 'abc') |
| self.assertRaises(KeyError, '{x}'.format) |
| self.assertRaises(ValueError, "}{".format) |
| self.assertRaises(ValueError, "abc{0:{}".format) |
| self.assertRaises(ValueError, "{0".format) |
| self.assertRaises(IndexError, "{0.}".format) |
| self.assertRaises(ValueError, "{0.}".format, 0) |
| self.assertRaises(ValueError, "{0[}".format) |
| self.assertRaises(ValueError, "{0[}".format, []) |
| self.assertRaises(KeyError, "{0]}".format) |
| self.assertRaises(ValueError, "{0.[]}".format, 0) |
| self.assertRaises(ValueError, "{0..foo}".format, 0) |
| self.assertRaises(ValueError, "{0[0}".format, 0) |
| self.assertRaises(ValueError, "{0[0:foo}".format, 0) |
| self.assertRaises(KeyError, "{c]}".format) |
| self.assertRaises(ValueError, "{{ {{{0}}".format, 0) |
| self.assertRaises(ValueError, "{0}}".format, 0) |
| self.assertRaises(KeyError, "{foo}".format, bar=3) |
| self.assertRaises(ValueError, "{0!x}".format, 3) |
| self.assertRaises(ValueError, "{0!}".format, 0) |
| self.assertRaises(ValueError, "{0!rs}".format, 0) |
| self.assertRaises(ValueError, "{!}".format) |
| self.assertRaises(IndexError, "{:}".format) |
| self.assertRaises(IndexError, "{:s}".format) |
| self.assertRaises(IndexError, "{}".format) |
| big = "23098475029384702983476098230754973209482573" |
| self.assertRaises(ValueError, ("{" + big + "}").format) |
| self.assertRaises(ValueError, ("{[" + big + "]}").format, [0]) |
| |
| # issue 6089 |
| self.assertRaises(ValueError, "{0[0]x}".format, [None]) |
| self.assertRaises(ValueError, "{0[0](10)}".format, [None]) |
| |
| # can't have a replacement on the field name portion |
| self.assertRaises(TypeError, '{0[{1}]}'.format, 'abcdefg', 4) |
| |
| # exceed maximum recursion depth |
| self.assertRaises(ValueError, "{0:{1:{2}}}".format, 'abc', 's', '') |
| self.assertRaises(ValueError, "{0:{1:{2:{3:{4:{5:{6}}}}}}}".format, |
| 0, 1, 2, 3, 4, 5, 6, 7) |
| |
| # string format spec errors |
| self.assertRaises(ValueError, "{0:-s}".format, '') |
| self.assertRaises(ValueError, format, "", "-") |
| self.assertRaises(ValueError, "{0:=s}".format, '') |
| |
| # Alternate formatting is not supported |
| self.assertRaises(ValueError, format, '', '#') |
| self.assertRaises(ValueError, format, '', '#20') |
| |
| # Non-ASCII |
| self.assertEqual("{0:s}{1:s}".format("ABC", "\u0410\u0411\u0412"), |
| 'ABC\u0410\u0411\u0412') |
| self.assertEqual("{0:.3s}".format("ABC\u0410\u0411\u0412"), |
| 'ABC') |
| self.assertEqual("{0:.0s}".format("ABC\u0410\u0411\u0412"), |
| '') |
| |
| self.assertEqual("{[{}]}".format({"{}": 5}), "5") |
| self.assertEqual("{[{}]}".format({"{}" : "a"}), "a") |
| self.assertEqual("{[{]}".format({"{" : "a"}), "a") |
| self.assertEqual("{[}]}".format({"}" : "a"}), "a") |
| self.assertEqual("{[[]}".format({"[" : "a"}), "a") |
| self.assertEqual("{[!]}".format({"!" : "a"}), "a") |
| self.assertRaises(ValueError, "{a{}b}".format, 42) |
| self.assertRaises(ValueError, "{a{b}".format, 42) |
| self.assertRaises(ValueError, "{[}".format, 42) |
| |
| self.assertEqual("0x{:0{:d}X}".format(0x0,16), "0x0000000000000000") |
| |
| # Blocking fallback |
| m = M('data') |
| self.assertEqual("{!r}".format(m), 'M(data)') |
| self.assertRaises(TypeError, "{!s}".format, m) |
| self.assertRaises(TypeError, "{}".format, m) |
| n = N('data') |
| self.assertEqual("{!r}".format(n), 'N(data)') |
| self.assertEqual("{!s}".format(n), 'N(data)') |
| self.assertRaises(TypeError, "{}".format, n) |
| |
| def test_format_map(self): |
| self.assertEqual(''.format_map({}), '') |
| self.assertEqual('a'.format_map({}), 'a') |
| self.assertEqual('ab'.format_map({}), 'ab') |
| self.assertEqual('a{{'.format_map({}), 'a{') |
| self.assertEqual('a}}'.format_map({}), 'a}') |
| self.assertEqual('{{b'.format_map({}), '{b') |
| self.assertEqual('}}b'.format_map({}), '}b') |
| self.assertEqual('a{{b'.format_map({}), 'a{b') |
| |
| # using mappings |
| class Mapping(dict): |
| def __missing__(self, key): |
| return key |
| self.assertEqual('{hello}'.format_map(Mapping()), 'hello') |
| self.assertEqual('{a} {world}'.format_map(Mapping(a='hello')), 'hello world') |
| |
| class InternalMapping: |
| def __init__(self): |
| self.mapping = {'a': 'hello'} |
| def __getitem__(self, key): |
| return self.mapping[key] |
| self.assertEqual('{a}'.format_map(InternalMapping()), 'hello') |
| |
| |
| class C: |
| def __init__(self, x=100): |
| self._x = x |
| def __format__(self, spec): |
| return spec |
| self.assertEqual('{foo._x}'.format_map({'foo': C(20)}), '20') |
| |
| # test various errors |
| self.assertRaises(TypeError, ''.format_map) |
| self.assertRaises(TypeError, 'a'.format_map) |
| |
| self.assertRaises(ValueError, '{'.format_map, {}) |
| self.assertRaises(ValueError, '}'.format_map, {}) |
| self.assertRaises(ValueError, 'a{'.format_map, {}) |
| self.assertRaises(ValueError, 'a}'.format_map, {}) |
| self.assertRaises(ValueError, '{a'.format_map, {}) |
| self.assertRaises(ValueError, '}a'.format_map, {}) |
| |
| # issue #12579: can't supply positional params to format_map |
| self.assertRaises(ValueError, '{}'.format_map, {'a' : 2}) |
| self.assertRaises(ValueError, '{}'.format_map, 'a') |
| self.assertRaises(ValueError, '{a} {}'.format_map, {"a" : 2, "b" : 1}) |
| |
| def test_format_huge_precision(self): |
| format_string = ".{}f".format(sys.maxsize + 1) |
| with self.assertRaises(ValueError): |
| result = format(2.34, format_string) |
| |
| def test_format_huge_width(self): |
| format_string = "{}f".format(sys.maxsize + 1) |
| with self.assertRaises(ValueError): |
| result = format(2.34, format_string) |
| |
| def test_format_huge_item_number(self): |
| format_string = "{{{}:.6f}}".format(sys.maxsize + 1) |
| with self.assertRaises(ValueError): |
| result = format_string.format(2.34) |
| |
| def test_format_auto_numbering(self): |
| class C: |
| def __init__(self, x=100): |
| self._x = x |
| def __format__(self, spec): |
| return spec |
| |
| self.assertEqual('{}'.format(10), '10') |
| self.assertEqual('{:5}'.format('s'), 's ') |
| self.assertEqual('{!r}'.format('s'), "'s'") |
| self.assertEqual('{._x}'.format(C(10)), '10') |
| self.assertEqual('{[1]}'.format([1, 2]), '2') |
| self.assertEqual('{[a]}'.format({'a':4, 'b':2}), '4') |
| self.assertEqual('a{}b{}c'.format(0, 1), 'a0b1c') |
| |
| self.assertEqual('a{:{}}b'.format('x', '^10'), 'a x b') |
| self.assertEqual('a{:{}x}b'.format(20, '#'), 'a0x14b') |
| |
| # can't mix and match numbering and auto-numbering |
| self.assertRaises(ValueError, '{}{1}'.format, 1, 2) |
| self.assertRaises(ValueError, '{1}{}'.format, 1, 2) |
| self.assertRaises(ValueError, '{:{1}}'.format, 1, 2) |
| self.assertRaises(ValueError, '{0:{}}'.format, 1, 2) |
| |
| # can mix and match auto-numbering and named |
| self.assertEqual('{f}{}'.format(4, f='test'), 'test4') |
| self.assertEqual('{}{f}'.format(4, f='test'), '4test') |
| self.assertEqual('{:{f}}{g}{}'.format(1, 3, g='g', f=2), ' 1g3') |
| self.assertEqual('{f:{}}{}{g}'.format(2, 4, f=1, g='g'), ' 14g') |
| |
| def test_formatting(self): |
| string_tests.MixinStrUnicodeUserStringTest.test_formatting(self) |
| # Testing Unicode formatting strings... |
| self.assertEqual("%s, %s" % ("abc", "abc"), 'abc, abc') |
| self.assertEqual("%s, %s, %i, %f, %5.2f" % ("abc", "abc", 1, 2, 3), 'abc, abc, 1, 2.000000, 3.00') |
| self.assertEqual("%s, %s, %i, %f, %5.2f" % ("abc", "abc", 1, -2, 3), 'abc, abc, 1, -2.000000, 3.00') |
| self.assertEqual("%s, %s, %i, %f, %5.2f" % ("abc", "abc", -1, -2, 3.5), 'abc, abc, -1, -2.000000, 3.50') |
| self.assertEqual("%s, %s, %i, %f, %5.2f" % ("abc", "abc", -1, -2, 3.57), 'abc, abc, -1, -2.000000, 3.57') |
| self.assertEqual("%s, %s, %i, %f, %5.2f" % ("abc", "abc", -1, -2, 1003.57), 'abc, abc, -1, -2.000000, 1003.57') |
| if not sys.platform.startswith('java'): |
| self.assertEqual("%r, %r" % (b"abc", "abc"), "b'abc', 'abc'") |
| self.assertEqual("%r" % ("\u1234",), "'\u1234'") |
| self.assertEqual("%a" % ("\u1234",), "'\\u1234'") |
| self.assertEqual("%(x)s, %(y)s" % {'x':"abc", 'y':"def"}, 'abc, def') |
| self.assertEqual("%(x)s, %(\xfc)s" % {'x':"abc", '\xfc':"def"}, 'abc, def') |
| |
| self.assertEqual('%c' % 0x1234, '\u1234') |
| self.assertEqual('%c' % 0x21483, '\U00021483') |
| self.assertRaises(OverflowError, "%c".__mod__, (0x110000,)) |
| self.assertEqual('%c' % '\U00021483', '\U00021483') |
| self.assertRaises(TypeError, "%c".__mod__, "aa") |
| self.assertRaises(ValueError, "%.1\u1032f".__mod__, (1.0/3)) |
| self.assertRaises(TypeError, "%i".__mod__, "aa") |
| |
| # formatting jobs delegated from the string implementation: |
| self.assertEqual('...%(foo)s...' % {'foo':"abc"}, '...abc...') |
| self.assertEqual('...%(foo)s...' % {'foo':"abc"}, '...abc...') |
| self.assertEqual('...%(foo)s...' % {'foo':"abc"}, '...abc...') |
| self.assertEqual('...%(foo)s...' % {'foo':"abc"}, '...abc...') |
| self.assertEqual('...%(foo)s...' % {'foo':"abc",'def':123}, '...abc...') |
| self.assertEqual('...%(foo)s...' % {'foo':"abc",'def':123}, '...abc...') |
| self.assertEqual('...%s...%s...%s...%s...' % (1,2,3,"abc"), '...1...2...3...abc...') |
| self.assertEqual('...%%...%%s...%s...%s...%s...%s...' % (1,2,3,"abc"), '...%...%s...1...2...3...abc...') |
| self.assertEqual('...%s...' % "abc", '...abc...') |
| self.assertEqual('%*s' % (5,'abc',), ' abc') |
| self.assertEqual('%*s' % (-5,'abc',), 'abc ') |
| self.assertEqual('%*.*s' % (5,2,'abc',), ' ab') |
| self.assertEqual('%*.*s' % (5,3,'abc',), ' abc') |
| self.assertEqual('%i %*.*s' % (10, 5,3,'abc',), '10 abc') |
| self.assertEqual('%i%s %*.*s' % (10, 3, 5, 3, 'abc',), '103 abc') |
| self.assertEqual('%c' % 'a', 'a') |
| class Wrapper: |
| def __str__(self): |
| return '\u1234' |
| self.assertEqual('%s' % Wrapper(), '\u1234') |
| |
| # issue 3382 |
| NAN = float('nan') |
| INF = float('inf') |
| self.assertEqual('%f' % NAN, 'nan') |
| self.assertEqual('%F' % NAN, 'NAN') |
| self.assertEqual('%f' % INF, 'inf') |
| self.assertEqual('%F' % INF, 'INF') |
| |
| # PEP 393 |
| self.assertEqual('%.1s' % "a\xe9\u20ac", 'a') |
| self.assertEqual('%.2s' % "a\xe9\u20ac", 'a\xe9') |
| |
| #issue 19995 |
| class PseudoInt: |
| def __init__(self, value): |
| self.value = int(value) |
| def __int__(self): |
| return self.value |
| def __index__(self): |
| return self.value |
| class PseudoFloat: |
| def __init__(self, value): |
| self.value = float(value) |
| def __int__(self): |
| return int(self.value) |
| pi = PseudoFloat(3.1415) |
| letter_m = PseudoInt(109) |
| self.assertEqual('%x' % 42, '2a') |
| self.assertEqual('%X' % 15, 'F') |
| self.assertEqual('%o' % 9, '11') |
| self.assertEqual('%c' % 109, 'm') |
| self.assertEqual('%x' % letter_m, '6d') |
| self.assertEqual('%X' % letter_m, '6D') |
| self.assertEqual('%o' % letter_m, '155') |
| self.assertEqual('%c' % letter_m, 'm') |
| self.assertRaisesRegex(TypeError, '%x format: an integer is required, not float', operator.mod, '%x', 3.14), |
| self.assertRaisesRegex(TypeError, '%X format: an integer is required, not float', operator.mod, '%X', 2.11), |
| self.assertRaisesRegex(TypeError, '%o format: an integer is required, not float', operator.mod, '%o', 1.79), |
| self.assertRaisesRegex(TypeError, '%x format: an integer is required, not PseudoFloat', operator.mod, '%x', pi), |
| self.assertRaises(TypeError, operator.mod, '%c', pi), |
| |
| def test_formatting_with_enum(self): |
| # issue18780 |
| import enum |
| class Float(float, enum.Enum): |
| PI = 3.1415926 |
| class Int(enum.IntEnum): |
| IDES = 15 |
| class Str(str, enum.Enum): |
| ABC = 'abc' |
| # Testing Unicode formatting strings... |
| self.assertEqual("%s, %s" % (Str.ABC, Str.ABC), |
| 'Str.ABC, Str.ABC') |
| self.assertEqual("%s, %s, %d, %i, %u, %f, %5.2f" % |
| (Str.ABC, Str.ABC, |
| Int.IDES, Int.IDES, Int.IDES, |
| Float.PI, Float.PI), |
| 'Str.ABC, Str.ABC, 15, 15, 15, 3.141593, 3.14') |
| |
| # formatting jobs delegated from the string implementation: |
| self.assertEqual('...%(foo)s...' % {'foo':Str.ABC}, |
| '...Str.ABC...') |
| self.assertEqual('...%(foo)s...' % {'foo':Int.IDES}, |
| '...Int.IDES...') |
| self.assertEqual('...%(foo)i...' % {'foo':Int.IDES}, |
| '...15...') |
| self.assertEqual('...%(foo)d...' % {'foo':Int.IDES}, |
| '...15...') |
| self.assertEqual('...%(foo)u...' % {'foo':Int.IDES, 'def':Float.PI}, |
| '...15...') |
| self.assertEqual('...%(foo)f...' % {'foo':Float.PI,'def':123}, |
| '...3.141593...') |
| |
| def test_formatting_huge_precision(self): |
| format_string = "%.{}f".format(sys.maxsize + 1) |
| with self.assertRaises(ValueError): |
| result = format_string % 2.34 |
| |
| @support.cpython_only |
| def test_formatting_huge_precision_c_limits(self): |
| from _testcapi import INT_MAX |
| format_string = "%.{}f".format(INT_MAX + 1) |
| with self.assertRaises(ValueError): |
| result = format_string % 2.34 |
| |
| def test_formatting_huge_width(self): |
| format_string = "%{}f".format(sys.maxsize + 1) |
| with self.assertRaises(ValueError): |
| result = format_string % 2.34 |
| |
| def test_startswith_endswith_errors(self): |
| for meth in ('foo'.startswith, 'foo'.endswith): |
| with self.assertRaises(TypeError) as cm: |
| meth(['f']) |
| exc = str(cm.exception) |
| self.assertIn('str', exc) |
| self.assertIn('tuple', exc) |
| |
| @support.run_with_locale('LC_ALL', 'de_DE', 'fr_FR') |
| def test_format_float(self): |
| # should not format with a comma, but always with C locale |
| self.assertEqual('1.0', '%.1f' % 1.0) |
| |
| def test_constructor(self): |
| # unicode(obj) tests (this maps to PyObject_Unicode() at C level) |
| |
| self.assertEqual( |
| str('unicode remains unicode'), |
| 'unicode remains unicode' |
| ) |
| |
| for text in ('ascii', '\xe9', '\u20ac', '\U0010FFFF'): |
| subclass = StrSubclass(text) |
| self.assertEqual(str(subclass), text) |
| self.assertEqual(len(subclass), len(text)) |
| if text == 'ascii': |
| self.assertEqual(subclass.encode('ascii'), b'ascii') |
| self.assertEqual(subclass.encode('utf-8'), b'ascii') |
| |
| self.assertEqual( |
| str('strings are converted to unicode'), |
| 'strings are converted to unicode' |
| ) |
| |
| class StringCompat: |
| def __init__(self, x): |
| self.x = x |
| def __str__(self): |
| return self.x |
| |
| self.assertEqual( |
| str(StringCompat('__str__ compatible objects are recognized')), |
| '__str__ compatible objects are recognized' |
| ) |
| |
| # unicode(obj) is compatible to str(): |
| |
| o = StringCompat('unicode(obj) is compatible to str()') |
| self.assertEqual(str(o), 'unicode(obj) is compatible to str()') |
| self.assertEqual(str(o), 'unicode(obj) is compatible to str()') |
| |
| for obj in (123, 123.45, 123): |
| self.assertEqual(str(obj), str(str(obj))) |
| |
| # unicode(obj, encoding, error) tests (this maps to |
| # PyUnicode_FromEncodedObject() at C level) |
| |
| if not sys.platform.startswith('java'): |
| self.assertRaises( |
| TypeError, |
| str, |
| 'decoding unicode is not supported', |
| 'utf-8', |
| 'strict' |
| ) |
| |
| self.assertEqual( |
| str(b'strings are decoded to unicode', 'utf-8', 'strict'), |
| 'strings are decoded to unicode' |
| ) |
| |
| if not sys.platform.startswith('java'): |
| self.assertEqual( |
| str( |
| memoryview(b'character buffers are decoded to unicode'), |
| 'utf-8', |
| 'strict' |
| ), |
| 'character buffers are decoded to unicode' |
| ) |
| |
| self.assertRaises(TypeError, str, 42, 42, 42) |
| |
| def test_constructor_keyword_args(self): |
| """Pass various keyword argument combinations to the constructor.""" |
| # The object argument can be passed as a keyword. |
| self.assertEqual(str(object='foo'), 'foo') |
| self.assertEqual(str(object=b'foo', encoding='utf-8'), 'foo') |
| # The errors argument without encoding triggers "decode" mode. |
| self.assertEqual(str(b'foo', errors='strict'), 'foo') # not "b'foo'" |
| self.assertEqual(str(object=b'foo', errors='strict'), 'foo') |
| |
| def test_constructor_defaults(self): |
| """Check the constructor argument defaults.""" |
| # The object argument defaults to '' or b''. |
| self.assertEqual(str(), '') |
| self.assertEqual(str(errors='strict'), '') |
| utf8_cent = '¢'.encode('utf-8') |
| # The encoding argument defaults to utf-8. |
| self.assertEqual(str(utf8_cent, errors='strict'), '¢') |
| # The errors argument defaults to strict. |
| self.assertRaises(UnicodeDecodeError, str, utf8_cent, encoding='ascii') |
| |
| def test_codecs_utf7(self): |
| utfTests = [ |
| ('A\u2262\u0391.', b'A+ImIDkQ.'), # RFC2152 example |
| ('Hi Mom -\u263a-!', b'Hi Mom -+Jjo--!'), # RFC2152 example |
| ('\u65E5\u672C\u8A9E', b'+ZeVnLIqe-'), # RFC2152 example |
| ('Item 3 is \u00a31.', b'Item 3 is +AKM-1.'), # RFC2152 example |
| ('+', b'+-'), |
| ('+-', b'+--'), |
| ('+?', b'+-?'), |
| (r'\?', b'+AFw?'), |
| ('+?', b'+-?'), |
| (r'\\?', b'+AFwAXA?'), |
| (r'\\\?', b'+AFwAXABc?'), |
| (r'++--', b'+-+---'), |
| ('\U000abcde', b'+2m/c3g-'), # surrogate pairs |
| ('/', b'/'), |
| ] |
| |
| for (x, y) in utfTests: |
| self.assertEqual(x.encode('utf-7'), y) |
| |
| # Unpaired surrogates are passed through |
| self.assertEqual('\uD801'.encode('utf-7'), b'+2AE-') |
| self.assertEqual('\uD801x'.encode('utf-7'), b'+2AE-x') |
| self.assertEqual('\uDC01'.encode('utf-7'), b'+3AE-') |
| self.assertEqual('\uDC01x'.encode('utf-7'), b'+3AE-x') |
| self.assertEqual(b'+2AE-'.decode('utf-7'), '\uD801') |
| self.assertEqual(b'+2AE-x'.decode('utf-7'), '\uD801x') |
| self.assertEqual(b'+3AE-'.decode('utf-7'), '\uDC01') |
| self.assertEqual(b'+3AE-x'.decode('utf-7'), '\uDC01x') |
| |
| self.assertEqual('\uD801\U000abcde'.encode('utf-7'), b'+2AHab9ze-') |
| self.assertEqual(b'+2AHab9ze-'.decode('utf-7'), '\uD801\U000abcde') |
| |
| # Issue #2242: crash on some Windows/MSVC versions |
| self.assertEqual(b'+\xc1'.decode('utf-7', 'ignore'), '') |
| |
| # Direct encoded characters |
| set_d = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789'(),-./:?" |
| # Optional direct characters |
| set_o = '!"#$%&*;<=>@[]^_`{|}' |
| for c in set_d: |
| self.assertEqual(c.encode('utf7'), c.encode('ascii')) |
| self.assertEqual(c.encode('ascii').decode('utf7'), c) |
| for c in set_o: |
| self.assertEqual(c.encode('ascii').decode('utf7'), c) |
| |
| def test_codecs_utf8(self): |
| self.assertEqual(''.encode('utf-8'), b'') |
| self.assertEqual('\u20ac'.encode('utf-8'), b'\xe2\x82\xac') |
| self.assertEqual('\U00010002'.encode('utf-8'), b'\xf0\x90\x80\x82') |
| self.assertEqual('\U00023456'.encode('utf-8'), b'\xf0\xa3\x91\x96') |
| self.assertEqual('\ud800'.encode('utf-8', 'surrogatepass'), b'\xed\xa0\x80') |
| self.assertEqual('\udc00'.encode('utf-8', 'surrogatepass'), b'\xed\xb0\x80') |
| self.assertEqual(('\U00010002'*10).encode('utf-8'), |
| b'\xf0\x90\x80\x82'*10) |
| self.assertEqual( |
| '\u6b63\u78ba\u306b\u8a00\u3046\u3068\u7ffb\u8a33\u306f' |
| '\u3055\u308c\u3066\u3044\u307e\u305b\u3093\u3002\u4e00' |
| '\u90e8\u306f\u30c9\u30a4\u30c4\u8a9e\u3067\u3059\u304c' |
| '\u3001\u3042\u3068\u306f\u3067\u305f\u3089\u3081\u3067' |
| '\u3059\u3002\u5b9f\u969b\u306b\u306f\u300cWenn ist das' |
| ' Nunstuck git und'.encode('utf-8'), |
| b'\xe6\xad\xa3\xe7\xa2\xba\xe3\x81\xab\xe8\xa8\x80\xe3\x81' |
| b'\x86\xe3\x81\xa8\xe7\xbf\xbb\xe8\xa8\xb3\xe3\x81\xaf\xe3' |
| b'\x81\x95\xe3\x82\x8c\xe3\x81\xa6\xe3\x81\x84\xe3\x81\xbe' |
| b'\xe3\x81\x9b\xe3\x82\x93\xe3\x80\x82\xe4\xb8\x80\xe9\x83' |
| b'\xa8\xe3\x81\xaf\xe3\x83\x89\xe3\x82\xa4\xe3\x83\x84\xe8' |
| b'\xaa\x9e\xe3\x81\xa7\xe3\x81\x99\xe3\x81\x8c\xe3\x80\x81' |
| b'\xe3\x81\x82\xe3\x81\xa8\xe3\x81\xaf\xe3\x81\xa7\xe3\x81' |
| b'\x9f\xe3\x82\x89\xe3\x82\x81\xe3\x81\xa7\xe3\x81\x99\xe3' |
| b'\x80\x82\xe5\xae\x9f\xe9\x9a\x9b\xe3\x81\xab\xe3\x81\xaf' |
| b'\xe3\x80\x8cWenn ist das Nunstuck git und' |
| ) |
| |
| # UTF-8 specific decoding tests |
| self.assertEqual(str(b'\xf0\xa3\x91\x96', 'utf-8'), '\U00023456' ) |
| self.assertEqual(str(b'\xf0\x90\x80\x82', 'utf-8'), '\U00010002' ) |
| self.assertEqual(str(b'\xe2\x82\xac', 'utf-8'), '\u20ac' ) |
| |
| # Other possible utf-8 test cases: |
| # * strict decoding testing for all of the |
| # UTF8_ERROR cases in PyUnicode_DecodeUTF8 |
| |
| def test_utf8_decode_valid_sequences(self): |
| sequences = [ |
| # single byte |
| (b'\x00', '\x00'), (b'a', 'a'), (b'\x7f', '\x7f'), |
| # 2 bytes |
| (b'\xc2\x80', '\x80'), (b'\xdf\xbf', '\u07ff'), |
| # 3 bytes |
| (b'\xe0\xa0\x80', '\u0800'), (b'\xed\x9f\xbf', '\ud7ff'), |
| (b'\xee\x80\x80', '\uE000'), (b'\xef\xbf\xbf', '\uffff'), |
| # 4 bytes |
| (b'\xF0\x90\x80\x80', '\U00010000'), |
| (b'\xf4\x8f\xbf\xbf', '\U0010FFFF') |
| ] |
| for seq, res in sequences: |
| self.assertEqual(seq.decode('utf-8'), res) |
| |
| |
| def test_utf8_decode_invalid_sequences(self): |
| # continuation bytes in a sequence of 2, 3, or 4 bytes |
| continuation_bytes = [bytes([x]) for x in range(0x80, 0xC0)] |
| # start bytes of a 2-byte sequence equivalent to code points < 0x7F |
| invalid_2B_seq_start_bytes = [bytes([x]) for x in range(0xC0, 0xC2)] |
| # start bytes of a 4-byte sequence equivalent to code points > 0x10FFFF |
| invalid_4B_seq_start_bytes = [bytes([x]) for x in range(0xF5, 0xF8)] |
| invalid_start_bytes = ( |
| continuation_bytes + invalid_2B_seq_start_bytes + |
| invalid_4B_seq_start_bytes + [bytes([x]) for x in range(0xF7, 0x100)] |
| ) |
| |
| for byte in invalid_start_bytes: |
| self.assertRaises(UnicodeDecodeError, byte.decode, 'utf-8') |
| |
| for sb in invalid_2B_seq_start_bytes: |
| for cb in continuation_bytes: |
| self.assertRaises(UnicodeDecodeError, (sb+cb).decode, 'utf-8') |
| |
| for sb in invalid_4B_seq_start_bytes: |
| for cb1 in continuation_bytes[:3]: |
| for cb3 in continuation_bytes[:3]: |
| self.assertRaises(UnicodeDecodeError, |
| (sb+cb1+b'\x80'+cb3).decode, 'utf-8') |
| |
| for cb in [bytes([x]) for x in range(0x80, 0xA0)]: |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xE0'+cb+b'\x80').decode, 'utf-8') |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xE0'+cb+b'\xBF').decode, 'utf-8') |
| # surrogates |
| for cb in [bytes([x]) for x in range(0xA0, 0xC0)]: |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xED'+cb+b'\x80').decode, 'utf-8') |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xED'+cb+b'\xBF').decode, 'utf-8') |
| for cb in [bytes([x]) for x in range(0x80, 0x90)]: |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xF0'+cb+b'\x80\x80').decode, 'utf-8') |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xF0'+cb+b'\xBF\xBF').decode, 'utf-8') |
| for cb in [bytes([x]) for x in range(0x90, 0xC0)]: |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xF4'+cb+b'\x80\x80').decode, 'utf-8') |
| self.assertRaises(UnicodeDecodeError, |
| (b'\xF4'+cb+b'\xBF\xBF').decode, 'utf-8') |
| |
| def test_issue8271(self): |
| # Issue #8271: during the decoding of an invalid UTF-8 byte sequence, |
| # only the start byte and the continuation byte(s) are now considered |
| # invalid, instead of the number of bytes specified by the start byte. |
| # See http://www.unicode.org/versions/Unicode5.2.0/ch03.pdf (page 95, |
| # table 3-8, Row 2) for more information about the algorithm used. |
| FFFD = '\ufffd' |
| sequences = [ |
| # invalid start bytes |
| (b'\x80', FFFD), # continuation byte |
| (b'\x80\x80', FFFD*2), # 2 continuation bytes |
| (b'\xc0', FFFD), |
| (b'\xc0\xc0', FFFD*2), |
| (b'\xc1', FFFD), |
| (b'\xc1\xc0', FFFD*2), |
| (b'\xc0\xc1', FFFD*2), |
| # with start byte of a 2-byte sequence |
| (b'\xc2', FFFD), # only the start byte |
| (b'\xc2\xc2', FFFD*2), # 2 start bytes |
| (b'\xc2\xc2\xc2', FFFD*3), # 3 start bytes |
| (b'\xc2\x41', FFFD+'A'), # invalid continuation byte |
| # with start byte of a 3-byte sequence |
| (b'\xe1', FFFD), # only the start byte |
| (b'\xe1\xe1', FFFD*2), # 2 start bytes |
| (b'\xe1\xe1\xe1', FFFD*3), # 3 start bytes |
| (b'\xe1\xe1\xe1\xe1', FFFD*4), # 4 start bytes |
| (b'\xe1\x80', FFFD), # only 1 continuation byte |
| (b'\xe1\x41', FFFD+'A'), # invalid continuation byte |
| (b'\xe1\x41\x80', FFFD+'A'+FFFD), # invalid cb followed by valid cb |
| (b'\xe1\x41\x41', FFFD+'AA'), # 2 invalid continuation bytes |
| (b'\xe1\x80\x41', FFFD+'A'), # only 1 valid continuation byte |
| (b'\xe1\x80\xe1\x41', FFFD*2+'A'), # 1 valid and the other invalid |
| (b'\xe1\x41\xe1\x80', FFFD+'A'+FFFD), # 1 invalid and the other valid |
| # with start byte of a 4-byte sequence |
| (b'\xf1', FFFD), # only the start byte |
| (b'\xf1\xf1', FFFD*2), # 2 start bytes |
| (b'\xf1\xf1\xf1', FFFD*3), # 3 start bytes |
| (b'\xf1\xf1\xf1\xf1', FFFD*4), # 4 start bytes |
| (b'\xf1\xf1\xf1\xf1\xf1', FFFD*5), # 5 start bytes |
| (b'\xf1\x80', FFFD), # only 1 continuation bytes |
| (b'\xf1\x80\x80', FFFD), # only 2 continuation bytes |
| (b'\xf1\x80\x41', FFFD+'A'), # 1 valid cb and 1 invalid |
| (b'\xf1\x80\x41\x41', FFFD+'AA'), # 1 valid cb and 1 invalid |
| (b'\xf1\x80\x80\x41', FFFD+'A'), # 2 valid cb and 1 invalid |
| (b'\xf1\x41\x80', FFFD+'A'+FFFD), # 1 invalid cv and 1 valid |
| (b'\xf1\x41\x80\x80', FFFD+'A'+FFFD*2), # 1 invalid cb and 2 invalid |
| (b'\xf1\x41\x80\x41', FFFD+'A'+FFFD+'A'), # 2 invalid cb and 1 invalid |
| (b'\xf1\x41\x41\x80', FFFD+'AA'+FFFD), # 1 valid cb and 1 invalid |
| (b'\xf1\x41\xf1\x80', FFFD+'A'+FFFD), |
| (b'\xf1\x41\x80\xf1', FFFD+'A'+FFFD*2), |
| (b'\xf1\xf1\x80\x41', FFFD*2+'A'), |
| (b'\xf1\x41\xf1\xf1', FFFD+'A'+FFFD*2), |
| # with invalid start byte of a 4-byte sequence (rfc2279) |
| (b'\xf5', FFFD), # only the start byte |
| (b'\xf5\xf5', FFFD*2), # 2 start bytes |
| (b'\xf5\x80', FFFD*2), # only 1 continuation byte |
| (b'\xf5\x80\x80', FFFD*3), # only 2 continuation byte |
| (b'\xf5\x80\x80\x80', FFFD*4), # 3 continuation bytes |
| (b'\xf5\x80\x41', FFFD*2+'A'), # 1 valid cb and 1 invalid |
| (b'\xf5\x80\x41\xf5', FFFD*2+'A'+FFFD), |
| (b'\xf5\x41\x80\x80\x41', FFFD+'A'+FFFD*2+'A'), |
| # with invalid start byte of a 5-byte sequence (rfc2279) |
| (b'\xf8', FFFD), # only the start byte |
| (b'\xf8\xf8', FFFD*2), # 2 start bytes |
| (b'\xf8\x80', FFFD*2), # only one continuation byte |
| (b'\xf8\x80\x41', FFFD*2 + 'A'), # 1 valid cb and 1 invalid |
| (b'\xf8\x80\x80\x80\x80', FFFD*5), # invalid 5 bytes seq with 5 bytes |
| # with invalid start byte of a 6-byte sequence (rfc2279) |
| (b'\xfc', FFFD), # only the start byte |
| (b'\xfc\xfc', FFFD*2), # 2 start bytes |
| (b'\xfc\x80\x80', FFFD*3), # only 2 continuation bytes |
| (b'\xfc\x80\x80\x80\x80\x80', FFFD*6), # 6 continuation bytes |
| # invalid start byte |
| (b'\xfe', FFFD), |
| (b'\xfe\x80\x80', FFFD*3), |
| # other sequences |
| (b'\xf1\x80\x41\x42\x43', '\ufffd\x41\x42\x43'), |
| (b'\xf1\x80\xff\x42\x43', '\ufffd\ufffd\x42\x43'), |
| (b'\xf1\x80\xc2\x81\x43', '\ufffd\x81\x43'), |
| (b'\x61\xF1\x80\x80\xE1\x80\xC2\x62\x80\x63\x80\xBF\x64', |
| '\x61\uFFFD\uFFFD\uFFFD\x62\uFFFD\x63\uFFFD\uFFFD\x64'), |
| ] |
| for n, (seq, res) in enumerate(sequences): |
| self.assertRaises(UnicodeDecodeError, seq.decode, 'utf-8', 'strict') |
| self.assertEqual(seq.decode('utf-8', 'replace'), res) |
| self.assertEqual((seq+b'b').decode('utf-8', 'replace'), res+'b') |
| self.assertEqual(seq.decode('utf-8', 'ignore'), |
| res.replace('\uFFFD', '')) |
| |
| def to_bytestring(self, seq): |
| return bytes(int(c, 16) for c in seq.split()) |
| |
| def assertCorrectUTF8Decoding(self, seq, res, err): |
| """ |
| Check that an invalid UTF-8 sequence raises a UnicodeDecodeError when |
| 'strict' is used, returns res when 'replace' is used, and that doesn't |
| return anything when 'ignore' is used. |
| """ |
| with self.assertRaises(UnicodeDecodeError) as cm: |
| seq.decode('utf-8') |
| exc = cm.exception |
| |
| self.assertIn(err, str(exc)) |
| self.assertEqual(seq.decode('utf-8', 'replace'), res) |
| self.assertEqual((b'aaaa' + seq + b'bbbb').decode('utf-8', 'replace'), |
| 'aaaa' + res + 'bbbb') |
| res = res.replace('\ufffd', '') |
| self.assertEqual(seq.decode('utf-8', 'ignore'), res) |
| self.assertEqual((b'aaaa' + seq + b'bbbb').decode('utf-8', 'ignore'), |
| 'aaaa' + res + 'bbbb') |
| |
| def test_invalid_start_byte(self): |
| """ |
| Test that an 'invalid start byte' error is raised when the first byte |
| is not in the ASCII range or is not a valid start byte of a 2-, 3-, or |
| 4-bytes sequence. The invalid start byte is replaced with a single |
| U+FFFD when errors='replace'. |
| E.g. <80> is a continuation byte and can appear only after a start byte. |
| """ |
| FFFD = '\ufffd' |
| for byte in b'\x80\xA0\x9F\xBF\xC0\xC1\xF5\xFF': |
| self.assertCorrectUTF8Decoding(bytes([byte]), '\ufffd', |
| 'invalid start byte') |
| |
| def test_unexpected_end_of_data(self): |
| """ |
| Test that an 'unexpected end of data' error is raised when the string |
| ends after a start byte of a 2-, 3-, or 4-bytes sequence without having |
| enough continuation bytes. The incomplete sequence is replaced with a |
| single U+FFFD when errors='replace'. |
| E.g. in the sequence <F3 80 80>, F3 is the start byte of a 4-bytes |
| sequence, but it's followed by only 2 valid continuation bytes and the |
| last continuation bytes is missing. |
| Note: the continuation bytes must be all valid, if one of them is |
| invalid another error will be raised. |
| """ |
| sequences = [ |
| 'C2', 'DF', |
| 'E0 A0', 'E0 BF', 'E1 80', 'E1 BF', 'EC 80', 'EC BF', |
| 'ED 80', 'ED 9F', 'EE 80', 'EE BF', 'EF 80', 'EF BF', |
| 'F0 90', 'F0 BF', 'F0 90 80', 'F0 90 BF', 'F0 BF 80', 'F0 BF BF', |
| 'F1 80', 'F1 BF', 'F1 80 80', 'F1 80 BF', 'F1 BF 80', 'F1 BF BF', |
| 'F3 80', 'F3 BF', 'F3 80 80', 'F3 80 BF', 'F3 BF 80', 'F3 BF BF', |
| 'F4 80', 'F4 8F', 'F4 80 80', 'F4 80 BF', 'F4 8F 80', 'F4 8F BF' |
| ] |
| FFFD = '\ufffd' |
| for seq in sequences: |
| self.assertCorrectUTF8Decoding(self.to_bytestring(seq), '\ufffd', |
| 'unexpected end of data') |
| |
| def test_invalid_cb_for_2bytes_seq(self): |
| """ |
| Test that an 'invalid continuation byte' error is raised when the |
| continuation byte of a 2-bytes sequence is invalid. The start byte |
| is replaced by a single U+FFFD and the second byte is handled |
| separately when errors='replace'. |
| E.g. in the sequence <C2 41>, C2 is the start byte of a 2-bytes |
| sequence, but 41 is not a valid continuation byte because it's the |
| ASCII letter 'A'. |
| """ |
| FFFD = '\ufffd' |
| FFFDx2 = FFFD * 2 |
| sequences = [ |
| ('C2 00', FFFD+'\x00'), ('C2 7F', FFFD+'\x7f'), |
| ('C2 C0', FFFDx2), ('C2 FF', FFFDx2), |
| ('DF 00', FFFD+'\x00'), ('DF 7F', FFFD+'\x7f'), |
| ('DF C0', FFFDx2), ('DF FF', FFFDx2), |
| ] |
| for seq, res in sequences: |
| self.assertCorrectUTF8Decoding(self.to_bytestring(seq), res, |
| 'invalid continuation byte') |
| |
| def test_invalid_cb_for_3bytes_seq(self): |
| """ |
| Test that an 'invalid continuation byte' error is raised when the |
| continuation byte(s) of a 3-bytes sequence are invalid. When |
| errors='replace', if the first continuation byte is valid, the first |
| two bytes (start byte + 1st cb) are replaced by a single U+FFFD and the |
| third byte is handled separately, otherwise only the start byte is |
| replaced with a U+FFFD and the other continuation bytes are handled |
| separately. |
| E.g. in the sequence <E1 80 41>, E1 is the start byte of a 3-bytes |
| sequence, 80 is a valid continuation byte, but 41 is not a valid cb |
| because it's the ASCII letter 'A'. |
| Note: when the start byte is E0 or ED, the valid ranges for the first |
| continuation byte are limited to A0..BF and 80..9F respectively. |
| Python 2 used to consider all the bytes in range 80..BF valid when the |
| start byte was ED. This is fixed in Python 3. |
| """ |
| FFFD = '\ufffd' |
| FFFDx2 = FFFD * 2 |
| sequences = [ |
| ('E0 00', FFFD+'\x00'), ('E0 7F', FFFD+'\x7f'), ('E0 80', FFFDx2), |
| ('E0 9F', FFFDx2), ('E0 C0', FFFDx2), ('E0 FF', FFFDx2), |
| ('E0 A0 00', FFFD+'\x00'), ('E0 A0 7F', FFFD+'\x7f'), |
| ('E0 A0 C0', FFFDx2), ('E0 A0 FF', FFFDx2), |
| ('E0 BF 00', FFFD+'\x00'), ('E0 BF 7F', FFFD+'\x7f'), |
| ('E0 BF C0', FFFDx2), ('E0 BF FF', FFFDx2), ('E1 00', FFFD+'\x00'), |
| ('E1 7F', FFFD+'\x7f'), ('E1 C0', FFFDx2), ('E1 FF', FFFDx2), |
| ('E1 80 00', FFFD+'\x00'), ('E1 80 7F', FFFD+'\x7f'), |
| ('E1 80 C0', FFFDx2), ('E1 80 FF', FFFDx2), |
| ('E1 BF 00', FFFD+'\x00'), ('E1 BF 7F', FFFD+'\x7f'), |
| ('E1 BF C0', FFFDx2), ('E1 BF FF', FFFDx2), ('EC 00', FFFD+'\x00'), |
| ('EC 7F', FFFD+'\x7f'), ('EC C0', FFFDx2), ('EC FF', FFFDx2), |
| ('EC 80 00', FFFD+'\x00'), ('EC 80 7F', FFFD+'\x7f'), |
| ('EC 80 C0', FFFDx2), ('EC 80 FF', FFFDx2), |
| ('EC BF 00', FFFD+'\x00'), ('EC BF 7F', FFFD+'\x7f'), |
| ('EC BF C0', FFFDx2), ('EC BF FF', FFFDx2), ('ED 00', FFFD+'\x00'), |
| ('ED 7F', FFFD+'\x7f'), |
| ('ED A0', FFFDx2), ('ED BF', FFFDx2), # see note ^ |
| ('ED C0', FFFDx2), ('ED FF', FFFDx2), ('ED 80 00', FFFD+'\x00'), |
| ('ED 80 7F', FFFD+'\x7f'), ('ED 80 C0', FFFDx2), |
| ('ED 80 FF', FFFDx2), ('ED 9F 00', FFFD+'\x00'), |
| ('ED 9F 7F', FFFD+'\x7f'), ('ED 9F C0', FFFDx2), |
| ('ED 9F FF', FFFDx2), ('EE 00', FFFD+'\x00'), |
| ('EE 7F', FFFD+'\x7f'), ('EE C0', FFFDx2), ('EE FF', FFFDx2), |
| ('EE 80 00', FFFD+'\x00'), ('EE 80 7F', FFFD+'\x7f'), |
| ('EE 80 C0', FFFDx2), ('EE 80 FF', FFFDx2), |
| ('EE BF 00', FFFD+'\x00'), ('EE BF 7F', FFFD+'\x7f'), |
| ('EE BF C0', FFFDx2), ('EE BF FF', FFFDx2), ('EF 00', FFFD+'\x00'), |
| ('EF 7F', FFFD+'\x7f'), ('EF C0', FFFDx2), ('EF FF', FFFDx2), |
| ('EF 80 00', FFFD+'\x00'), ('EF 80 7F', FFFD+'\x7f'), |
| ('EF 80 C0', FFFDx2), ('EF 80 FF', FFFDx2), |
| ('EF BF 00', FFFD+'\x00'), ('EF BF 7F', FFFD+'\x7f'), |
| ('EF BF C0', FFFDx2), ('EF BF FF', FFFDx2), |
| ] |
| for seq, res in sequences: |
| self.assertCorrectUTF8Decoding(self.to_bytestring(seq), res, |
| 'invalid continuation byte') |
| |
| def test_invalid_cb_for_4bytes_seq(self): |
| """ |
| Test that an 'invalid continuation byte' error is raised when the |
| continuation byte(s) of a 4-bytes sequence are invalid. When |
| errors='replace',the start byte and all the following valid |
| continuation bytes are replaced with a single U+FFFD, and all the bytes |
| starting from the first invalid continuation bytes (included) are |
| handled separately. |
| E.g. in the sequence <E1 80 41>, E1 is the start byte of a 3-bytes |
| sequence, 80 is a valid continuation byte, but 41 is not a valid cb |
| because it's the ASCII letter 'A'. |
| Note: when the start byte is E0 or ED, the valid ranges for the first |
| continuation byte are limited to A0..BF and 80..9F respectively. |
| However, when the start byte is ED, Python 2 considers all the bytes |
| in range 80..BF valid. This is fixed in Python 3. |
| """ |
| FFFD = '\ufffd' |
| FFFDx2 = FFFD * 2 |
| sequences = [ |
| ('F0 00', FFFD+'\x00'), ('F0 7F', FFFD+'\x7f'), ('F0 80', FFFDx2), |
| ('F0 8F', FFFDx2), ('F0 C0', FFFDx2), ('F0 FF', FFFDx2), |
| ('F0 90 00', FFFD+'\x00'), ('F0 90 7F', FFFD+'\x7f'), |
| ('F0 90 C0', FFFDx2), ('F0 90 FF', FFFDx2), |
| ('F0 BF 00', FFFD+'\x00'), ('F0 BF 7F', FFFD+'\x7f'), |
| ('F0 BF C0', FFFDx2), ('F0 BF FF', FFFDx2), |
| ('F0 90 80 00', FFFD+'\x00'), ('F0 90 80 7F', FFFD+'\x7f'), |
| ('F0 90 80 C0', FFFDx2), ('F0 90 80 FF', FFFDx2), |
| ('F0 90 BF 00', FFFD+'\x00'), ('F0 90 BF 7F', FFFD+'\x7f'), |
| ('F0 90 BF C0', FFFDx2), ('F0 90 BF FF', FFFDx2), |
| ('F0 BF 80 00', FFFD+'\x00'), ('F0 BF 80 7F', FFFD+'\x7f'), |
| ('F0 BF 80 C0', FFFDx2), ('F0 BF 80 FF', FFFDx2), |
| ('F0 BF BF 00', FFFD+'\x00'), ('F0 BF BF 7F', FFFD+'\x7f'), |
| ('F0 BF BF C0', FFFDx2), ('F0 BF BF FF', FFFDx2), |
| ('F1 00', FFFD+'\x00'), ('F1 7F', FFFD+'\x7f'), ('F1 C0', FFFDx2), |
| ('F1 FF', FFFDx2), ('F1 80 00', FFFD+'\x00'), |
| ('F1 80 7F', FFFD+'\x7f'), ('F1 80 C0', FFFDx2), |
| ('F1 80 FF', FFFDx2), ('F1 BF 00', FFFD+'\x00'), |
| ('F1 BF 7F', FFFD+'\x7f'), ('F1 BF C0', FFFDx2), |
| ('F1 BF FF', FFFDx2), ('F1 80 80 00', FFFD+'\x00'), |
| ('F1 80 80 7F', FFFD+'\x7f'), ('F1 80 80 C0', FFFDx2), |
| ('F1 80 80 FF', FFFDx2), ('F1 80 BF 00', FFFD+'\x00'), |
| ('F1 80 BF 7F', FFFD+'\x7f'), ('F1 80 BF C0', FFFDx2), |
| ('F1 80 BF FF', FFFDx2), ('F1 BF 80 00', FFFD+'\x00'), |
| ('F1 BF 80 7F', FFFD+'\x7f'), ('F1 BF 80 C0', FFFDx2), |
| ('F1 BF 80 FF', FFFDx2), ('F1 BF BF 00', FFFD+'\x00'), |
| ('F1 BF BF 7F', FFFD+'\x7f'), ('F1 BF BF C0', FFFDx2), |
| ('F1 BF BF FF', FFFDx2), ('F3 00', FFFD+'\x00'), |
| ('F3 7F', FFFD+'\x7f'), ('F3 C0', FFFDx2), ('F3 FF', FFFDx2), |
| ('F3 80 00', FFFD+'\x00'), ('F3 80 7F', FFFD+'\x7f'), |
| ('F3 80 C0', FFFDx2), ('F3 80 FF', FFFDx2), |
| ('F3 BF 00', FFFD+'\x00'), ('F3 BF 7F', FFFD+'\x7f'), |
| ('F3 BF C0', FFFDx2), ('F3 BF FF', FFFDx2), |
| ('F3 80 80 00', FFFD+'\x00'), ('F3 80 80 7F', FFFD+'\x7f'), |
| ('F3 80 80 C0', FFFDx2), ('F3 80 80 FF', FFFDx2), |
| ('F3 80 BF 00', FFFD+'\x00'), ('F3 80 BF 7F', FFFD+'\x7f'), |
| ('F3 80 BF C0', FFFDx2), ('F3 80 BF FF', FFFDx2), |
| ('F3 BF 80 00', FFFD+'\x00'), ('F3 BF 80 7F', FFFD+'\x7f'), |
| ('F3 BF 80 C0', FFFDx2), ('F3 BF 80 FF', FFFDx2), |
| ('F3 BF BF 00', FFFD+'\x00'), ('F3 BF BF 7F', FFFD+'\x7f'), |
| ('F3 BF BF C0', FFFDx2), ('F3 BF BF FF', FFFDx2), |
| ('F4 00', FFFD+'\x00'), ('F4 7F', FFFD+'\x7f'), ('F4 90', FFFDx2), |
| ('F4 BF', FFFDx2), ('F4 C0', FFFDx2), ('F4 FF', FFFDx2), |
| ('F4 80 00', FFFD+'\x00'), ('F4 80 7F', FFFD+'\x7f'), |
| ('F4 80 C0', FFFDx2), ('F4 80 FF', FFFDx2), |
| ('F4 8F 00', FFFD+'\x00'), ('F4 8F 7F', FFFD+'\x7f'), |
| ('F4 8F C0', FFFDx2), ('F4 8F FF', FFFDx2), |
| ('F4 80 80 00', FFFD+'\x00'), ('F4 80 80 7F', FFFD+'\x7f'), |
| ('F4 80 80 C0', FFFDx2), ('F4 80 80 FF', FFFDx2), |
| ('F4 80 BF 00', FFFD+'\x00'), ('F4 80 BF 7F', FFFD+'\x7f'), |
| ('F4 80 BF C0', FFFDx2), ('F4 80 BF FF', FFFDx2), |
| ('F4 8F 80 00', FFFD+'\x00'), ('F4 8F 80 7F', FFFD+'\x7f'), |
| ('F4 8F 80 C0', FFFDx2), ('F4 8F 80 FF', FFFDx2), |
| ('F4 8F BF 00', FFFD+'\x00'), ('F4 8F BF 7F', FFFD+'\x7f'), |
| ('F4 8F BF C0', FFFDx2), ('F4 8F BF FF', FFFDx2) |
| ] |
| for seq, res in sequences: |
| self.assertCorrectUTF8Decoding(self.to_bytestring(seq), res, |
| 'invalid continuation byte') |
| |
| def test_codecs_idna(self): |
| # Test whether trailing dot is preserved |
| self.assertEqual("www.python.org.".encode("idna"), b"www.python.org.") |
| |
| def test_codecs_errors(self): |
| # Error handling (encoding) |
| self.assertRaises(UnicodeError, 'Andr\202 x'.encode, 'ascii') |
| self.assertRaises(UnicodeError, 'Andr\202 x'.encode, 'ascii','strict') |
| self.assertEqual('Andr\202 x'.encode('ascii','ignore'), b"Andr x") |
| self.assertEqual('Andr\202 x'.encode('ascii','replace'), b"Andr? x") |
| self.assertEqual('Andr\202 x'.encode('ascii', 'replace'), |
| 'Andr\202 x'.encode('ascii', errors='replace')) |
| self.assertEqual('Andr\202 x'.encode('ascii', 'ignore'), |
| 'Andr\202 x'.encode(encoding='ascii', errors='ignore')) |
| |
| # Error handling (decoding) |
| self.assertRaises(UnicodeError, str, b'Andr\202 x', 'ascii') |
| self.assertRaises(UnicodeError, str, b'Andr\202 x', 'ascii', 'strict') |
| self.assertEqual(str(b'Andr\202 x', 'ascii', 'ignore'), "Andr x") |
| self.assertEqual(str(b'Andr\202 x', 'ascii', 'replace'), 'Andr\uFFFD x') |
| self.assertEqual(str(b'\202 x', 'ascii', 'replace'), '\uFFFD x') |
| |
| # Error handling (unknown character names) |
| self.assertEqual(b"\\N{foo}xx".decode("unicode-escape", "ignore"), "xx") |
| |
| # Error handling (truncated escape sequence) |
| self.assertRaises(UnicodeError, b"\\".decode, "unicode-escape") |
| |
| self.assertRaises(TypeError, b"hello".decode, "test.unicode1") |
| self.assertRaises(TypeError, str, b"hello", "test.unicode2") |
| self.assertRaises(TypeError, "hello".encode, "test.unicode1") |
| self.assertRaises(TypeError, "hello".encode, "test.unicode2") |
| |
| # Error handling (wrong arguments) |
| self.assertRaises(TypeError, "hello".encode, 42, 42, 42) |
| |
| # Error handling (lone surrogate in PyUnicode_TransformDecimalToASCII()) |
| self.assertRaises(UnicodeError, float, "\ud800") |
| self.assertRaises(UnicodeError, float, "\udf00") |
| self.assertRaises(UnicodeError, complex, "\ud800") |
| self.assertRaises(UnicodeError, complex, "\udf00") |
| |
| def test_codecs(self): |
| # Encoding |
| self.assertEqual('hello'.encode('ascii'), b'hello') |
| self.assertEqual('hello'.encode('utf-7'), b'hello') |
| self.assertEqual('hello'.encode('utf-8'), b'hello') |
| self.assertEqual('hello'.encode('utf-8'), b'hello') |
| self.assertEqual('hello'.encode('utf-16-le'), b'h\000e\000l\000l\000o\000') |
| self.assertEqual('hello'.encode('utf-16-be'), b'\000h\000e\000l\000l\000o') |
| self.assertEqual('hello'.encode('latin-1'), b'hello') |
| |
| # Default encoding is utf-8 |
| self.assertEqual('\u2603'.encode(), b'\xe2\x98\x83') |
| |
| # Roundtrip safety for BMP (just the first 1024 chars) |
| for c in range(1024): |
| u = chr(c) |
| for encoding in ('utf-7', 'utf-8', 'utf-16', 'utf-16-le', |
| 'utf-16-be', 'raw_unicode_escape', |
| 'unicode_escape', 'unicode_internal'): |
| with warnings.catch_warnings(): |
| # unicode-internal has been deprecated |
| warnings.simplefilter("ignore", DeprecationWarning) |
| |
| self.assertEqual(str(u.encode(encoding),encoding), u) |
| |
| # Roundtrip safety for BMP (just the first 256 chars) |
| for c in range(256): |
| u = chr(c) |
| for encoding in ('latin-1',): |
| self.assertEqual(str(u.encode(encoding),encoding), u) |
| |
| # Roundtrip safety for BMP (just the first 128 chars) |
| for c in range(128): |
| u = chr(c) |
| for encoding in ('ascii',): |
| self.assertEqual(str(u.encode(encoding),encoding), u) |
| |
| # Roundtrip safety for non-BMP (just a few chars) |
| with warnings.catch_warnings(): |
| # unicode-internal has been deprecated |
| warnings.simplefilter("ignore", DeprecationWarning) |
| |
| u = '\U00010001\U00020002\U00030003\U00040004\U00050005' |
| for encoding in ('utf-8', 'utf-16', 'utf-16-le', 'utf-16-be', |
| 'raw_unicode_escape', |
| 'unicode_escape', 'unicode_internal'): |
| self.assertEqual(str(u.encode(encoding),encoding), u) |
| |
| # UTF-8 must be roundtrip safe for all code points |
| # (except surrogates, which are forbidden). |
| u = ''.join(map(chr, list(range(0, 0xd800)) + |
| list(range(0xe000, 0x110000)))) |
| for encoding in ('utf-8',): |
| self.assertEqual(str(u.encode(encoding),encoding), u) |
| |
| def test_codecs_charmap(self): |
| # 0-127 |
| s = bytes(range(128)) |
| for encoding in ( |
| 'cp037', 'cp1026', 'cp273', |
| 'cp437', 'cp500', 'cp720', 'cp737', 'cp775', 'cp850', |
| 'cp852', 'cp855', 'cp858', 'cp860', 'cp861', 'cp862', |
| 'cp863', 'cp865', 'cp866', 'cp1125', |
| 'iso8859_10', 'iso8859_13', 'iso8859_14', 'iso8859_15', |
| 'iso8859_2', 'iso8859_3', 'iso8859_4', 'iso8859_5', 'iso8859_6', |
| 'iso8859_7', 'iso8859_9', |
| 'koi8_r', 'koi8_t', 'koi8_u', 'kz1048', 'latin_1', |
| 'mac_cyrillic', 'mac_latin2', |
| |
| 'cp1250', 'cp1251', 'cp1252', 'cp1253', 'cp1254', 'cp1255', |
| 'cp1256', 'cp1257', 'cp1258', |
| 'cp856', 'cp857', 'cp864', 'cp869', 'cp874', |
| |
| 'mac_greek', 'mac_iceland','mac_roman', 'mac_turkish', |
| 'cp1006', 'iso8859_8', |
| |
| ### These have undefined mappings: |
| #'cp424', |
| |
| ### These fail the round-trip: |
| #'cp875' |
| |
| ): |
| self.assertEqual(str(s, encoding).encode(encoding), s) |
| |
| # 128-255 |
| s = bytes(range(128, 256)) |
| for encoding in ( |
| 'cp037', 'cp1026', 'cp273', |
| 'cp437', 'cp500', 'cp720', 'cp737', 'cp775', 'cp850', |
| 'cp852', 'cp855', 'cp858', 'cp860', 'cp861', 'cp862', |
| 'cp863', 'cp865', 'cp866', 'cp1125', |
| 'iso8859_10', 'iso8859_13', 'iso8859_14', 'iso8859_15', |
| 'iso8859_2', 'iso8859_4', 'iso8859_5', |
| 'iso8859_9', 'koi8_r', 'koi8_u', 'latin_1', |
| 'mac_cyrillic', 'mac_latin2', |
| |
| ### These have undefined mappings: |
| #'cp1250', 'cp1251', 'cp1252', 'cp1253', 'cp1254', 'cp1255', |
| #'cp1256', 'cp1257', 'cp1258', |
| #'cp424', 'cp856', 'cp857', 'cp864', 'cp869', 'cp874', |
| #'iso8859_3', 'iso8859_6', 'iso8859_7', 'koi8_t', 'kz1048', |
| #'mac_greek', 'mac_iceland','mac_roman', 'mac_turkish', |
| |
| ### These fail the round-trip: |
| #'cp1006', 'cp875', 'iso8859_8', |
| |
| ): |
| self.assertEqual(str(s, encoding).encode(encoding), s) |
| |
| def test_concatenation(self): |
| self.assertEqual(("abc" "def"), "abcdef") |
| self.assertEqual(("abc" "def"), "abcdef") |
| self.assertEqual(("abc" "def"), "abcdef") |
| self.assertEqual(("abc" "def" "ghi"), "abcdefghi") |
| self.assertEqual(("abc" "def" "ghi"), "abcdefghi") |
| |
| def test_printing(self): |
| class BitBucket: |
| def write(self, text): |
| pass |
| |
| out = BitBucket() |
| print('abc', file=out) |
| print('abc', 'def', file=out) |
| print('abc', 'def', file=out) |
| print('abc', 'def', file=out) |
| print('abc\n', file=out) |
| print('abc\n', end=' ', file=out) |
| print('abc\n', end=' ', file=out) |
| print('def\n', file=out) |
| print('def\n', file=out) |
| |
| def test_ucs4(self): |
| x = '\U00100000' |
| y = x.encode("raw-unicode-escape").decode("raw-unicode-escape") |
| self.assertEqual(x, y) |
| |
| y = br'\U00100000' |
| x = y.decode("raw-unicode-escape").encode("raw-unicode-escape") |
| self.assertEqual(x, y) |
| y = br'\U00010000' |
| x = y.decode("raw-unicode-escape").encode("raw-unicode-escape") |
| self.assertEqual(x, y) |
| |
| try: |
| br'\U11111111'.decode("raw-unicode-escape") |
| except UnicodeDecodeError as e: |
| self.assertEqual(e.start, 0) |
| self.assertEqual(e.end, 10) |
| else: |
| self.fail("Should have raised UnicodeDecodeError") |
| |
| def test_conversion(self): |
| # Make sure __str__() works properly |
| class ObjectToStr: |
| def __str__(self): |
| return "foo" |
| |
| class StrSubclassToStr(str): |
| def __str__(self): |
| return "foo" |
| |
| class StrSubclassToStrSubclass(str): |
| def __new__(cls, content=""): |
| return str.__new__(cls, 2*content) |
| def __str__(self): |
| return self |
| |
| self.assertEqual(str(ObjectToStr()), "foo") |
| self.assertEqual(str(StrSubclassToStr("bar")), "foo") |
| s = str(StrSubclassToStrSubclass("foo")) |
| self.assertEqual(s, "foofoo") |
| self.assertIs(type(s), StrSubclassToStrSubclass) |
| s = StrSubclass(StrSubclassToStrSubclass("foo")) |
| self.assertEqual(s, "foofoo") |
| self.assertIs(type(s), StrSubclass) |
| |
| def test_unicode_repr(self): |
| class s1: |
| def __repr__(self): |
| return '\\n' |
| |
| class s2: |
| def __repr__(self): |
| return '\\n' |
| |
| self.assertEqual(repr(s1()), '\\n') |
| self.assertEqual(repr(s2()), '\\n') |
| |
| def test_printable_repr(self): |
| self.assertEqual(repr('\U00010000'), "'%c'" % (0x10000,)) # printable |
| self.assertEqual(repr('\U00014000'), "'\\U00014000'") # nonprintable |
| |
| # This test only affects 32-bit platforms because expandtabs can only take |
| # an int as the max value, not a 64-bit C long. If expandtabs is changed |
| # to take a 64-bit long, this test should apply to all platforms. |
| @unittest.skipIf(sys.maxsize > (1 << 32) or struct.calcsize('P') != 4, |
| 'only applies to 32-bit platforms') |
| def test_expandtabs_overflows_gracefully(self): |
| self.assertRaises(OverflowError, 't\tt\t'.expandtabs, sys.maxsize) |
| |
| @support.cpython_only |
| def test_expandtabs_optimization(self): |
| s = 'abc' |
| self.assertIs(s.expandtabs(), s) |
| |
| def test_raiseMemError(self): |
| if struct.calcsize('P') == 8: |
| # 64 bits pointers |
| ascii_struct_size = 48 |
| compact_struct_size = 72 |
| else: |
| # 32 bits pointers |
| ascii_struct_size = 24 |
| compact_struct_size = 36 |
| |
| for char in ('a', '\xe9', '\u20ac', '\U0010ffff'): |
| code = ord(char) |
| if code < 0x100: |
| char_size = 1 # sizeof(Py_UCS1) |
| struct_size = ascii_struct_size |
| elif code < 0x10000: |
| char_size = 2 # sizeof(Py_UCS2) |
| struct_size = compact_struct_size |
| else: |
| char_size = 4 # sizeof(Py_UCS4) |
| struct_size = compact_struct_size |
| # Note: sys.maxsize is half of the actual max allocation because of |
| # the signedness of Py_ssize_t. Strings of maxlen-1 should in principle |
| # be allocatable, given enough memory. |
| maxlen = ((sys.maxsize - struct_size) // char_size) |
| alloc = lambda: char * maxlen |
| self.assertRaises(MemoryError, alloc) |
| self.assertRaises(MemoryError, alloc) |
| |
| def test_format_subclass(self): |
| class S(str): |
| def __str__(self): |
| return '__str__ overridden' |
| s = S('xxx') |
| self.assertEqual("%s" % s, '__str__ overridden') |
| self.assertEqual("{}".format(s), '__str__ overridden') |
| |
| def test_subclass_add(self): |
| class S(str): |
| def __add__(self, o): |
| return "3" |
| self.assertEqual(S("4") + S("5"), "3") |
| class S(str): |
| def __iadd__(self, o): |
| return "3" |
| s = S("1") |
| s += "4" |
| self.assertEqual(s, "3") |
| |
| def test_getnewargs(self): |
| text = 'abc' |
| args = text.__getnewargs__() |
| self.assertIsNot(args[0], text) |
| self.assertEqual(args[0], text) |
| self.assertEqual(len(args), 1) |
| |
| def test_resize(self): |
| for length in range(1, 100, 7): |
| # generate a fresh string (refcount=1) |
| text = 'a' * length + 'b' |
| |
| with support.check_warnings(('unicode_internal codec has been ' |
| 'deprecated', DeprecationWarning)): |
| # fill wstr internal field |
| abc = text.encode('unicode_internal') |
| self.assertEqual(abc.decode('unicode_internal'), text) |
| |
| # resize text: wstr field must be cleared and then recomputed |
| text += 'c' |
| abcdef = text.encode('unicode_internal') |
| self.assertNotEqual(abc, abcdef) |
| self.assertEqual(abcdef.decode('unicode_internal'), text) |
| |
| def test_compare(self): |
| # Issue #17615 |
| N = 10 |
| ascii = 'a' * N |
| ascii2 = 'z' * N |
| latin = '\x80' * N |
| latin2 = '\xff' * N |
| bmp = '\u0100' * N |
| bmp2 = '\uffff' * N |
| astral = '\U00100000' * N |
| astral2 = '\U0010ffff' * N |
| strings = ( |
| ascii, ascii2, |
| latin, latin2, |
| bmp, bmp2, |
| astral, astral2) |
| for text1, text2 in itertools.combinations(strings, 2): |
| equal = (text1 is text2) |
| self.assertEqual(text1 == text2, equal) |
| self.assertEqual(text1 != text2, not equal) |
| |
| if equal: |
| self.assertTrue(text1 <= text2) |
| self.assertTrue(text1 >= text2) |
| |
| # text1 is text2: duplicate strings to skip the "str1 == str2" |
| # optimization in unicode_compare_eq() and really compare |
| # character per character |
| copy1 = duplicate_string(text1) |
| copy2 = duplicate_string(text2) |
| self.assertIsNot(copy1, copy2) |
| |
| self.assertTrue(copy1 == copy2) |
| self.assertFalse(copy1 != copy2) |
| |
| self.assertTrue(copy1 <= copy2) |
| self.assertTrue(copy2 >= copy2) |
| |
| self.assertTrue(ascii < ascii2) |
| self.assertTrue(ascii < latin) |
| self.assertTrue(ascii < bmp) |
| self.assertTrue(ascii < astral) |
| self.assertFalse(ascii >= ascii2) |
| self.assertFalse(ascii >= latin) |
| self.assertFalse(ascii >= bmp) |
| self.assertFalse(ascii >= astral) |
| |
| self.assertFalse(latin < ascii) |
| self.assertTrue(latin < latin2) |
| self.assertTrue(latin < bmp) |
| self.assertTrue(latin < astral) |
| self.assertTrue(latin >= ascii) |
| self.assertFalse(latin >= latin2) |
| self.assertFalse(latin >= bmp) |
| self.assertFalse(latin >= astral) |
| |
| self.assertFalse(bmp < ascii) |
| self.assertFalse(bmp < latin) |
| self.assertTrue(bmp < bmp2) |
| self.assertTrue(bmp < astral) |
| self.assertTrue(bmp >= ascii) |
| self.assertTrue(bmp >= latin) |
| self.assertFalse(bmp >= bmp2) |
| self.assertFalse(bmp >= astral) |
| |
| self.assertFalse(astral < ascii) |
| self.assertFalse(astral < latin) |
| self.assertFalse(astral < bmp2) |
| self.assertTrue(astral < astral2) |
| self.assertTrue(astral >= ascii) |
| self.assertTrue(astral >= latin) |
| self.assertTrue(astral >= bmp2) |
| self.assertFalse(astral >= astral2) |
| |
| def test_free_after_iterating(self): |
| support.check_free_after_iterating(self, iter, str) |
| support.check_free_after_iterating(self, reversed, str) |
| |
| |
| class CAPITest(unittest.TestCase): |
| |
| # Test PyUnicode_FromFormat() |
| def test_from_format(self): |
| support.import_module('ctypes') |
| from ctypes import ( |
| pythonapi, py_object, sizeof, |
| c_int, c_long, c_longlong, c_ssize_t, |
| c_uint, c_ulong, c_ulonglong, c_size_t, c_void_p) |
| name = "PyUnicode_FromFormat" |
| _PyUnicode_FromFormat = getattr(pythonapi, name) |
| _PyUnicode_FromFormat.restype = py_object |
| |
| def PyUnicode_FromFormat(format, *args): |
| cargs = tuple( |
| py_object(arg) if isinstance(arg, str) else arg |
| for arg in args) |
| return _PyUnicode_FromFormat(format, *cargs) |
| |
| def check_format(expected, format, *args): |
| text = PyUnicode_FromFormat(format, *args) |
| self.assertEqual(expected, text) |
| |
| # ascii format, non-ascii argument |
| check_format('ascii\x7f=unicode\xe9', |
| b'ascii\x7f=%U', 'unicode\xe9') |
| |
| # non-ascii format, ascii argument: ensure that PyUnicode_FromFormatV() |
| # raises an error |
| self.assertRaisesRegex(ValueError, |
| r'^PyUnicode_FromFormatV\(\) expects an ASCII-encoded format ' |
| 'string, got a non-ASCII byte: 0xe9$', |
| PyUnicode_FromFormat, b'unicode\xe9=%s', 'ascii') |
| |
| # test "%c" |
| check_format('\uabcd', |
| b'%c', c_int(0xabcd)) |
| check_format('\U0010ffff', |
| b'%c', c_int(0x10ffff)) |
| with self.assertRaises(OverflowError): |
| PyUnicode_FromFormat(b'%c', c_int(0x110000)) |
| # Issue #18183 |
| check_format('\U00010000\U00100000', |
| b'%c%c', c_int(0x10000), c_int(0x100000)) |
| |
| # test "%" |
| check_format('%', |
| b'%') |
| check_format('%', |
| b'%%') |
| check_format('%s', |
| b'%%s') |
| check_format('[%]', |
| b'[%%]') |
| check_format('%abc', |
| b'%%%s', b'abc') |
| |
| # truncated string |
| check_format('abc', |
| b'%.3s', b'abcdef') |
| check_format('abc[\ufffd', |
| b'%.5s', 'abc[\u20ac]'.encode('utf8')) |
| check_format("'\\u20acABC'", |
| b'%A', '\u20acABC') |
| check_format("'\\u20", |
| b'%.5A', '\u20acABCDEF') |
| check_format("'\u20acABC'", |
| b'%R', '\u20acABC') |
| check_format("'\u20acA", |
| b'%.3R', '\u20acABCDEF') |
| check_format('\u20acAB', |
| b'%.3S', '\u20acABCDEF') |
| check_format('\u20acAB', |
| b'%.3U', '\u20acABCDEF') |
| check_format('\u20acAB', |
| b'%.3V', '\u20acABCDEF', None) |
| check_format('abc[\ufffd', |
| b'%.5V', None, 'abc[\u20ac]'.encode('utf8')) |
| |
| # following tests comes from #7330 |
| # test width modifier and precision modifier with %S |
| check_format("repr= abc", |
| b'repr=%5S', 'abc') |
| check_format("repr=ab", |
| b'repr=%.2S', 'abc') |
| check_format("repr= ab", |
| b'repr=%5.2S', 'abc') |
| |
| # test width modifier and precision modifier with %R |
| check_format("repr= 'abc'", |
| b'repr=%8R', 'abc') |
| check_format("repr='ab", |
| b'repr=%.3R', 'abc') |
| check_format("repr= 'ab", |
| b'repr=%5.3R', 'abc') |
| |
| # test width modifier and precision modifier with %A |
| check_format("repr= 'abc'", |
| b'repr=%8A', 'abc') |
| check_format("repr='ab", |
| b'repr=%.3A', 'abc') |
| check_format("repr= 'ab", |
| b'repr=%5.3A', 'abc') |
| |
| # test width modifier and precision modifier with %s |
| check_format("repr= abc", |
| b'repr=%5s', b'abc') |
| check_format("repr=ab", |
| b'repr=%.2s', b'abc') |
| check_format("repr= ab", |
| b'repr=%5.2s', b'abc') |
| |
| # test width modifier and precision modifier with %U |
| check_format("repr= abc", |
| b'repr=%5U', 'abc') |
| check_format("repr=ab", |
| b'repr=%.2U', 'abc') |
| check_format("repr= ab", |
| b'repr=%5.2U', 'abc') |
| |
| # test width modifier and precision modifier with %V |
| check_format("repr= abc", |
| b'repr=%5V', 'abc', b'123') |
| check_format("repr=ab", |
| b'repr=%.2V', 'abc', b'123') |
| check_format("repr= ab", |
| b'repr=%5.2V', 'abc', b'123') |
| check_format("repr= 123", |
| b'repr=%5V', None, b'123') |
| check_format("repr=12", |
| b'repr=%.2V', None, b'123') |
| check_format("repr= 12", |
| b'repr=%5.2V', None, b'123') |
| |
| # test integer formats (%i, %d, %u) |
| check_format('010', |
| b'%03i', c_int(10)) |
| check_format('0010', |
| b'%0.4i', c_int(10)) |
| check_format('-123', |
| b'%i', c_int(-123)) |
| check_format('-123', |
| b'%li', c_long(-123)) |
| check_format('-123', |
| b'%lli', c_longlong(-123)) |
| check_format('-123', |
| b'%zi', c_ssize_t(-123)) |
| |
| check_format('-123', |
| b'%d', c_int(-123)) |
| check_format('-123', |
| b'%ld', c_long(-123)) |
| check_format('-123', |
| b'%lld', c_longlong(-123)) |
| check_format('-123', |
| b'%zd', c_ssize_t(-123)) |
| |
| check_format('123', |
| b'%u', c_uint(123)) |
| check_format('123', |
| b'%lu', c_ulong(123)) |
| check_format('123', |
| b'%llu', c_ulonglong(123)) |
| check_format('123', |
| b'%zu', c_size_t(123)) |
| |
| # test long output |
| min_longlong = -(2 ** (8 * sizeof(c_longlong) - 1)) |
| max_longlong = -min_longlong - 1 |
| check_format(str(min_longlong), |
| b'%lld', c_longlong(min_longlong)) |
| check_format(str(max_longlong), |
| b'%lld', c_longlong(max_longlong)) |
| max_ulonglong = 2 ** (8 * sizeof(c_ulonglong)) - 1 |
| check_format(str(max_ulonglong), |
| b'%llu', c_ulonglong(max_ulonglong)) |
| PyUnicode_FromFormat(b'%p', c_void_p(-1)) |
| |
| # test padding (width and/or precision) |
| check_format('123'.rjust(10, '0'), |
| b'%010i', c_int(123)) |
| check_format('123'.rjust(100), |
| b'%100i', c_int(123)) |
| check_format('123'.rjust(100, '0'), |
| b'%.100i', c_int(123)) |
| check_format('123'.rjust(80, '0').rjust(100), |
| b'%100.80i', c_int(123)) |
| |
| check_format('123'.rjust(10, '0'), |
| b'%010u', c_uint(123)) |
| check_format('123'.rjust(100), |
| b'%100u', c_uint(123)) |
| check_format('123'.rjust(100, '0'), |
| b'%.100u', c_uint(123)) |
| check_format('123'.rjust(80, '0').rjust(100), |
| b'%100.80u', c_uint(123)) |
| |
| check_format('123'.rjust(10, '0'), |
| b'%010x', c_int(0x123)) |
| check_format('123'.rjust(100), |
| b'%100x', c_int(0x123)) |
| check_format('123'.rjust(100, '0'), |
| b'%.100x', c_int(0x123)) |
| check_format('123'.rjust(80, '0').rjust(100), |
| b'%100.80x', c_int(0x123)) |
| |
| # test %A |
| check_format(r"%A:'abc\xe9\uabcd\U0010ffff'", |
| b'%%A:%A', 'abc\xe9\uabcd\U0010ffff') |
| |
| # test %V |
| check_format('repr=abc', |
| b'repr=%V', 'abc', b'xyz') |
| |
| # Test string decode from parameter of %s using utf-8. |
| # b'\xe4\xba\xba\xe6\xb0\x91' is utf-8 encoded byte sequence of |
| # '\u4eba\u6c11' |
| check_format('repr=\u4eba\u6c11', |
| b'repr=%V', None, b'\xe4\xba\xba\xe6\xb0\x91') |
| |
| #Test replace error handler. |
| check_format('repr=abc\ufffd', |
| b'repr=%V', None, b'abc\xff') |
| |
| # not supported: copy the raw format string. these tests are just here |
| # to check for crashes and should not be considered as specifications |
| check_format('%s', |
| b'%1%s', b'abc') |
| check_format('%1abc', |
| b'%1abc') |
| check_format('%+i', |
| b'%+i', c_int(10)) |
| check_format('%.%s', |
| b'%.%s', b'abc') |
| |
| # Test PyUnicode_AsWideChar() |
| @support.cpython_only |
| def test_aswidechar(self): |
| from _testcapi import unicode_aswidechar |
| support.import_module('ctypes') |
| from ctypes import c_wchar, sizeof |
| |
| wchar, size = unicode_aswidechar('abcdef', 2) |
| self.assertEqual(size, 2) |
| self.assertEqual(wchar, 'ab') |
| |
| wchar, size = unicode_aswidechar('abc', 3) |
| self.assertEqual(size, 3) |
| self.assertEqual(wchar, 'abc') |
| |
| wchar, size = unicode_aswidechar('abc', 4) |
| self.assertEqual(size, 3) |
| self.assertEqual(wchar, 'abc\0') |
| |
| wchar, size = unicode_aswidechar('abc', 10) |
| self.assertEqual(size, 3) |
| self.assertEqual(wchar, 'abc\0') |
| |
| wchar, size = unicode_aswidechar('abc\0def', 20) |
| self.assertEqual(size, 7) |
| self.assertEqual(wchar, 'abc\0def\0') |
| |
| nonbmp = chr(0x10ffff) |
| if sizeof(c_wchar) == 2: |
| buflen = 3 |
| nchar = 2 |
| else: # sizeof(c_wchar) == 4 |
| buflen = 2 |
| nchar = 1 |
| wchar, size = unicode_aswidechar(nonbmp, buflen) |
| self.assertEqual(size, nchar) |
| self.assertEqual(wchar, nonbmp + '\0') |
| |
| # Test PyUnicode_AsWideCharString() |
| @support.cpython_only |
| def test_aswidecharstring(self): |
| from _testcapi import unicode_aswidecharstring |
| support.import_module('ctypes') |
| from ctypes import c_wchar, sizeof |
| |
| wchar, size = unicode_aswidecharstring('abc') |
| self.assertEqual(size, 3) |
| self.assertEqual(wchar, 'abc\0') |
| |
| wchar, size = unicode_aswidecharstring('abc\0def') |
| self.assertEqual(size, 7) |
| self.assertEqual(wchar, 'abc\0def\0') |
| |
| nonbmp = chr(0x10ffff) |
| if sizeof(c_wchar) == 2: |
| nchar = 2 |
| else: # sizeof(c_wchar) == 4 |
| nchar = 1 |
| wchar, size = unicode_aswidecharstring(nonbmp) |
| self.assertEqual(size, nchar) |
| self.assertEqual(wchar, nonbmp + '\0') |
| |
| # Test PyUnicode_AsUCS4() |
| @support.cpython_only |
| def test_asucs4(self): |
| from _testcapi import unicode_asucs4 |
| for s in ['abc', '\xa1\xa2', '\u4f60\u597d', 'a\U0001f600', |
| 'a\ud800b\udfffc', '\ud834\udd1e']: |
| l = len(s) |
| self.assertEqual(unicode_asucs4(s, l, 1), s+'\0') |
| self.assertEqual(unicode_asucs4(s, l, 0), s+'\uffff') |
| self.assertEqual(unicode_asucs4(s, l+1, 1), s+'\0\uffff') |
| self.assertEqual(unicode_asucs4(s, l+1, 0), s+'\0\uffff') |
| self.assertRaises(SystemError, unicode_asucs4, s, l-1, 1) |
| self.assertRaises(SystemError, unicode_asucs4, s, l-2, 0) |
| s = '\0'.join([s, s]) |
| self.assertEqual(unicode_asucs4(s, len(s), 1), s+'\0') |
| self.assertEqual(unicode_asucs4(s, len(s), 0), s+'\uffff') |
| |
| # Test PyUnicode_CopyCharacters() |
| @support.cpython_only |
| def test_copycharacters(self): |
| from _testcapi import unicode_copycharacters |
| |
| strings = [ |
| 'abcde', '\xa1\xa2\xa3\xa4\xa5', |
| '\u4f60\u597d\u4e16\u754c\uff01', |
| '\U0001f600\U0001f601\U0001f602\U0001f603\U0001f604' |
| ] |
| |
| for idx, from_ in enumerate(strings): |
| # wide -> narrow: exceed maxchar limitation |
| for to in strings[:idx]: |
| self.assertRaises( |
| SystemError, |
| unicode_copycharacters, to, 0, from_, 0, 5 |
| ) |
| # same kind |
| for from_start in range(5): |
| self.assertEqual( |
| unicode_copycharacters(from_, 0, from_, from_start, 5), |
| (from_[from_start:from_start+5].ljust(5, '\0'), |
| 5-from_start) |
| ) |
| for to_start in range(5): |
| self.assertEqual( |
| unicode_copycharacters(from_, to_start, from_, to_start, 5), |
| (from_[to_start:to_start+5].rjust(5, '\0'), |
| 5-to_start) |
| ) |
| # narrow -> wide |
| # Tests omitted since this creates invalid strings. |
| |
| s = strings[0] |
| self.assertRaises(IndexError, unicode_copycharacters, s, 6, s, 0, 5) |
| self.assertRaises(IndexError, unicode_copycharacters, s, -1, s, 0, 5) |
| self.assertRaises(IndexError, unicode_copycharacters, s, 0, s, 6, 5) |
| self.assertRaises(IndexError, unicode_copycharacters, s, 0, s, -1, 5) |
| self.assertRaises(SystemError, unicode_copycharacters, s, 1, s, 0, 5) |
| self.assertRaises(SystemError, unicode_copycharacters, s, 0, s, 0, -1) |
| self.assertRaises(SystemError, unicode_copycharacters, s, 0, b'', 0, 0) |
| |
| @support.cpython_only |
| def test_encode_decimal(self): |
| from _testcapi import unicode_encodedecimal |
| self.assertEqual(unicode_encodedecimal('123'), |
| b'123') |
| self.assertEqual(unicode_encodedecimal('\u0663.\u0661\u0664'), |
| b'3.14') |
| self.assertEqual(unicode_encodedecimal("\N{EM SPACE}3.14\N{EN SPACE}"), |
| b' 3.14 ') |
| self.assertRaises(UnicodeEncodeError, |
| unicode_encodedecimal, "123\u20ac", "strict") |
| self.assertRaisesRegex( |
| ValueError, |
| "^'decimal' codec can't encode character", |
| unicode_encodedecimal, "123\u20ac", "replace") |
| |
| @support.cpython_only |
| def test_transform_decimal(self): |
| from _testcapi import unicode_transformdecimaltoascii as transform_decimal |
| self.assertEqual(transform_decimal('123'), |
| '123') |
| self.assertEqual(transform_decimal('\u0663.\u0661\u0664'), |
| '3.14') |
| self.assertEqual(transform_decimal("\N{EM SPACE}3.14\N{EN SPACE}"), |
| "\N{EM SPACE}3.14\N{EN SPACE}") |
| self.assertEqual(transform_decimal('123\u20ac'), |
| '123\u20ac') |
| |
| @support.cpython_only |
| def test_pep393_utf8_caching_bug(self): |
| # Issue #25709: Problem with string concatenation and utf-8 cache |
| from _testcapi import getargs_s_hash |
| for k in 0x24, 0xa4, 0x20ac, 0x1f40d: |
| s = '' |
| for i in range(5): |
| # Due to CPython specific optimization the 's' string can be |
| # resized in-place. |
| s += chr(k) |
| # Parsing with the "s#" format code calls indirectly |
| # PyUnicode_AsUTF8AndSize() which creates the UTF-8 |
| # encoded string cached in the Unicode object. |
| self.assertEqual(getargs_s_hash(s), chr(k).encode() * (i + 1)) |
| # Check that the second call returns the same result |
| self.assertEqual(getargs_s_hash(s), chr(k).encode() * (i + 1)) |
| |
| class StringModuleTest(unittest.TestCase): |
| def test_formatter_parser(self): |
| def parse(format): |
| return list(_string.formatter_parser(format)) |
| |
| formatter = parse("prefix {2!s}xxx{0:^+10.3f}{obj.attr!s} {z[0]!s:10}") |
| self.assertEqual(formatter, [ |
| ('prefix ', '2', '', 's'), |
| ('xxx', '0', '^+10.3f', None), |
| ('', 'obj.attr', '', 's'), |
| (' ', 'z[0]', '10', 's'), |
| ]) |
| |
| formatter = parse("prefix {} suffix") |
| self.assertEqual(formatter, [ |
| ('prefix ', '', '', None), |
| (' suffix', None, None, None), |
| ]) |
| |
| formatter = parse("str") |
| self.assertEqual(formatter, [ |
| ('str', None, None, None), |
| ]) |
| |
| formatter = parse("") |
| self.assertEqual(formatter, []) |
| |
| formatter = parse("{0}") |
| self.assertEqual(formatter, [ |
| ('', '0', '', None), |
| ]) |
| |
| self.assertRaises(TypeError, _string.formatter_parser, 1) |
| |
| def test_formatter_field_name_split(self): |
| def split(name): |
| items = list(_string.formatter_field_name_split(name)) |
| items[1] = list(items[1]) |
| return items |
| self.assertEqual(split("obj"), ["obj", []]) |
| self.assertEqual(split("obj.arg"), ["obj", [(True, 'arg')]]) |
| self.assertEqual(split("obj[key]"), ["obj", [(False, 'key')]]) |
| self.assertEqual(split("obj.arg[key1][key2]"), [ |
| "obj", |
| [(True, 'arg'), |
| (False, 'key1'), |
| (False, 'key2'), |
| ]]) |
| self.assertRaises(TypeError, _string.formatter_field_name_split, 1) |
| |
| |
| if __name__ == "__main__": |
| unittest.main() |