| Guido van Rossum | 715287f | 2008-12-02 22:34:15 +0000 | [diff] [blame] | 1 | .. _unicode-howto: | 
 | 2 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 3 | ***************** | 
 | 4 |   Unicode HOWTO | 
 | 5 | ***************** | 
 | 6 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 7 | :Release: 1.12 | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 8 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 9 | This HOWTO discusses Python support for Unicode, and explains | 
| Benjamin Peterson | d7c3ed5 | 2010-06-27 22:32:30 +0000 | [diff] [blame] | 10 | various problems that people commonly encounter when trying to work | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 11 | with Unicode. | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 12 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 13 | Introduction to Unicode | 
 | 14 | ======================= | 
 | 15 |  | 
 | 16 | History of Character Codes | 
 | 17 | -------------------------- | 
 | 18 |  | 
 | 19 | In 1968, the American Standard Code for Information Interchange, better known by | 
 | 20 | its acronym ASCII, was standardized.  ASCII defined numeric codes for various | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 21 | characters, with the numeric values running from 0 to 127.  For example, the | 
 | 22 | lowercase letter 'a' is assigned 97 as its code value. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 23 |  | 
 | 24 | ASCII was an American-developed standard, so it only defined unaccented | 
 | 25 | characters.  There was an 'e', but no 'é' or 'Í'.  This meant that languages | 
 | 26 | which required accented characters couldn't be faithfully represented in ASCII. | 
 | 27 | (Actually the missing accents matter for English, too, which contains words such | 
 | 28 | as 'naïve' and 'café', and some publications have house styles which require | 
 | 29 | spellings such as 'coöperate'.) | 
 | 30 |  | 
 | 31 | For a while people just wrote programs that didn't display accents.  I remember | 
 | 32 | looking at Apple ][ BASIC programs, published in French-language publications in | 
 | 33 | the mid-1980s, that had lines like these:: | 
 | 34 |  | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 35 |    PRINT "FICHIER EST COMPLETE." | 
 | 36 |    PRINT "CARACTERE NON ACCEPTE." | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 37 |  | 
 | 38 | Those messages should contain accents, and they just look wrong to someone who | 
 | 39 | can read French. | 
 | 40 |  | 
 | 41 | In the 1980s, almost all personal computers were 8-bit, meaning that bytes could | 
 | 42 | hold values ranging from 0 to 255.  ASCII codes only went up to 127, so some | 
 | 43 | machines assigned values between 128 and 255 to accented characters.  Different | 
 | 44 | machines had different codes, however, which led to problems exchanging files. | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 45 | Eventually various commonly used sets of values for the 128--255 range emerged. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 46 | Some were true standards, defined by the International Standards Organization, | 
 | 47 | and some were **de facto** conventions that were invented by one company or | 
 | 48 | another and managed to catch on. | 
 | 49 |  | 
 | 50 | 255 characters aren't very many.  For example, you can't fit both the accented | 
 | 51 | characters used in Western Europe and the Cyrillic alphabet used for Russian | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 52 | into the 128--255 range because there are more than 127 such characters. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 53 |  | 
 | 54 | You could write files using different codes (all your Russian files in a coding | 
 | 55 | system called KOI8, all your French files in a different coding system called | 
 | 56 | Latin1), but what if you wanted to write a French document that quotes some | 
 | 57 | Russian text?  In the 1980s people began to want to solve this problem, and the | 
 | 58 | Unicode standardization effort began. | 
 | 59 |  | 
 | 60 | Unicode started out using 16-bit characters instead of 8-bit characters.  16 | 
 | 61 | bits means you have 2^16 = 65,536 distinct values available, making it possible | 
 | 62 | to represent many different characters from many different alphabets; an initial | 
 | 63 | goal was to have Unicode contain the alphabets for every single human language. | 
 | 64 | It turns out that even 16 bits isn't enough to meet that goal, and the modern | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 65 | Unicode specification uses a wider range of codes, 0 through 1,114,111 (0x10ffff | 
 | 66 | in base 16). | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 67 |  | 
 | 68 | There's a related ISO standard, ISO 10646.  Unicode and ISO 10646 were | 
 | 69 | originally separate efforts, but the specifications were merged with the 1.1 | 
 | 70 | revision of Unicode. | 
 | 71 |  | 
 | 72 | (This discussion of Unicode's history is highly simplified.  I don't think the | 
 | 73 | average Python programmer needs to worry about the historical details; consult | 
 | 74 | the Unicode consortium site listed in the References for more information.) | 
 | 75 |  | 
 | 76 |  | 
 | 77 | Definitions | 
 | 78 | ----------- | 
 | 79 |  | 
 | 80 | A **character** is the smallest possible component of a text.  'A', 'B', 'C', | 
 | 81 | etc., are all different characters.  So are 'È' and 'Í'.  Characters are | 
 | 82 | abstractions, and vary depending on the language or context you're talking | 
 | 83 | about.  For example, the symbol for ohms (Ω) is usually drawn much like the | 
 | 84 | capital letter omega (Ω) in the Greek alphabet (they may even be the same in | 
 | 85 | some fonts), but these are two different characters that have different | 
 | 86 | meanings. | 
 | 87 |  | 
 | 88 | The Unicode standard describes how characters are represented by **code | 
 | 89 | points**.  A code point is an integer value, usually denoted in base 16.  In the | 
 | 90 | standard, a code point is written using the notation U+12ca to mean the | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 91 | character with value 0x12ca (4,810 decimal).  The Unicode standard contains a lot | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 92 | of tables listing characters and their corresponding code points:: | 
 | 93 |  | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 94 |    0061    'a'; LATIN SMALL LETTER A | 
 | 95 |    0062    'b'; LATIN SMALL LETTER B | 
 | 96 |    0063    'c'; LATIN SMALL LETTER C | 
 | 97 |    ... | 
 | 98 |    007B    '{'; LEFT CURLY BRACKET | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 99 |  | 
 | 100 | Strictly, these definitions imply that it's meaningless to say 'this is | 
 | 101 | character U+12ca'.  U+12ca is a code point, which represents some particular | 
 | 102 | character; in this case, it represents the character 'ETHIOPIC SYLLABLE WI'.  In | 
 | 103 | informal contexts, this distinction between code points and characters will | 
 | 104 | sometimes be forgotten. | 
 | 105 |  | 
 | 106 | A character is represented on a screen or on paper by a set of graphical | 
 | 107 | elements that's called a **glyph**.  The glyph for an uppercase A, for example, | 
 | 108 | is two diagonal strokes and a horizontal stroke, though the exact details will | 
 | 109 | depend on the font being used.  Most Python code doesn't need to worry about | 
 | 110 | glyphs; figuring out the correct glyph to display is generally the job of a GUI | 
 | 111 | toolkit or a terminal's font renderer. | 
 | 112 |  | 
 | 113 |  | 
 | 114 | Encodings | 
 | 115 | --------- | 
 | 116 |  | 
 | 117 | To summarize the previous section: a Unicode string is a sequence of code | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 118 | points, which are numbers from 0 through 0x10ffff (1,114,111 decimal).  This | 
 | 119 | sequence needs to be represented as a set of bytes (meaning, values | 
 | 120 | from 0 through 255) in memory.  The rules for translating a Unicode string | 
 | 121 | into a sequence of bytes are called an **encoding**. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 122 |  | 
 | 123 | The first encoding you might think of is an array of 32-bit integers.  In this | 
 | 124 | representation, the string "Python" would look like this:: | 
 | 125 |  | 
 | 126 |        P           y           t           h           o           n | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 127 |     0x50 00 00 00 79 00 00 00 74 00 00 00 68 00 00 00 6f 00 00 00 6e 00 00 00 | 
 | 128 |        0  1  2  3  4  5  6  7  8  9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 129 |  | 
 | 130 | This representation is straightforward but using it presents a number of | 
 | 131 | problems. | 
 | 132 |  | 
 | 133 | 1. It's not portable; different processors order the bytes differently. | 
 | 134 |  | 
 | 135 | 2. It's very wasteful of space.  In most texts, the majority of the code points | 
 | 136 |    are less than 127, or less than 255, so a lot of space is occupied by zero | 
 | 137 |    bytes.  The above string takes 24 bytes compared to the 6 bytes needed for an | 
 | 138 |    ASCII representation.  Increased RAM usage doesn't matter too much (desktop | 
 | 139 |    computers have megabytes of RAM, and strings aren't usually that large), but | 
 | 140 |    expanding our usage of disk and network bandwidth by a factor of 4 is | 
 | 141 |    intolerable. | 
 | 142 |  | 
 | 143 | 3. It's not compatible with existing C functions such as ``strlen()``, so a new | 
 | 144 |    family of wide string functions would need to be used. | 
 | 145 |  | 
 | 146 | 4. Many Internet standards are defined in terms of textual data, and can't | 
 | 147 |    handle content with embedded zero bytes. | 
 | 148 |  | 
| Benjamin Peterson | d7c3ed5 | 2010-06-27 22:32:30 +0000 | [diff] [blame] | 149 | Generally people don't use this encoding, instead choosing other | 
 | 150 | encodings that are more efficient and convenient.  UTF-8 is probably | 
 | 151 | the most commonly supported encoding; it will be discussed below. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 152 |  | 
 | 153 | Encodings don't have to handle every possible Unicode character, and most | 
| Benjamin Peterson | 1f31697 | 2009-09-11 20:42:29 +0000 | [diff] [blame] | 154 | encodings don't.  The rules for converting a Unicode string into the ASCII | 
 | 155 | encoding, for example, are simple; for each code point: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 156 |  | 
 | 157 | 1. If the code point is < 128, each byte is the same as the value of the code | 
 | 158 |    point. | 
 | 159 |  | 
 | 160 | 2. If the code point is 128 or greater, the Unicode string can't be represented | 
 | 161 |    in this encoding.  (Python raises a :exc:`UnicodeEncodeError` exception in this | 
 | 162 |    case.) | 
 | 163 |  | 
 | 164 | Latin-1, also known as ISO-8859-1, is a similar encoding.  Unicode code points | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 165 | 0--255 are identical to the Latin-1 values, so converting to this encoding simply | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 166 | requires converting code points to byte values; if a code point larger than 255 | 
 | 167 | is encountered, the string can't be encoded into Latin-1. | 
 | 168 |  | 
 | 169 | Encodings don't have to be simple one-to-one mappings like Latin-1.  Consider | 
 | 170 | IBM's EBCDIC, which was used on IBM mainframes.  Letter values weren't in one | 
 | 171 | block: 'a' through 'i' had values from 129 to 137, but 'j' through 'r' were 145 | 
 | 172 | through 153.  If you wanted to use EBCDIC as an encoding, you'd probably use | 
 | 173 | some sort of lookup table to perform the conversion, but this is largely an | 
 | 174 | internal detail. | 
 | 175 |  | 
 | 176 | UTF-8 is one of the most commonly used encodings.  UTF stands for "Unicode | 
 | 177 | Transformation Format", and the '8' means that 8-bit numbers are used in the | 
 | 178 | encoding.  (There's also a UTF-16 encoding, but it's less frequently used than | 
 | 179 | UTF-8.)  UTF-8 uses the following rules: | 
 | 180 |  | 
 | 181 | 1. If the code point is <128, it's represented by the corresponding byte value. | 
 | 182 | 2. If the code point is between 128 and 0x7ff, it's turned into two byte values | 
 | 183 |    between 128 and 255. | 
 | 184 | 3. Code points >0x7ff are turned into three- or four-byte sequences, where each | 
 | 185 |    byte of the sequence is between 128 and 255. | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 186 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 187 | UTF-8 has several convenient properties: | 
 | 188 |  | 
 | 189 | 1. It can handle any Unicode code point. | 
 | 190 | 2. A Unicode string is turned into a string of bytes containing no embedded zero | 
 | 191 |    bytes.  This avoids byte-ordering issues, and means UTF-8 strings can be | 
 | 192 |    processed by C functions such as ``strcpy()`` and sent through protocols that | 
 | 193 |    can't handle zero bytes. | 
 | 194 | 3. A string of ASCII text is also valid UTF-8 text. | 
 | 195 | 4. UTF-8 is fairly compact; the majority of code points are turned into two | 
 | 196 |    bytes, and values less than 128 occupy only a single byte. | 
 | 197 | 5. If bytes are corrupted or lost, it's possible to determine the start of the | 
 | 198 |    next UTF-8-encoded code point and resynchronize.  It's also unlikely that | 
 | 199 |    random 8-bit data will look like valid UTF-8. | 
 | 200 |  | 
 | 201 |  | 
 | 202 |  | 
 | 203 | References | 
 | 204 | ---------- | 
 | 205 |  | 
 | 206 | The Unicode Consortium site at <http://www.unicode.org> has character charts, a | 
 | 207 | glossary, and PDF versions of the Unicode specification.  Be prepared for some | 
 | 208 | difficult reading.  <http://www.unicode.org/history/> is a chronology of the | 
 | 209 | origin and development of Unicode. | 
 | 210 |  | 
 | 211 | To help understand the standard, Jukka Korpela has written an introductory guide | 
 | 212 | to reading the Unicode character tables, available at | 
 | 213 | <http://www.cs.tut.fi/~jkorpela/unicode/guide.html>. | 
 | 214 |  | 
| Georg Brandl | ee8783d | 2009-09-16 16:00:31 +0000 | [diff] [blame] | 215 | Another good introductory article was written by Joel Spolsky | 
 | 216 | <http://www.joelonsoftware.com/articles/Unicode.html>. | 
 | 217 | If this introduction didn't make things clear to you, you should try reading this | 
 | 218 | alternate article before continuing. | 
 | 219 |  | 
 | 220 | .. Jason Orendorff XXX http://www.jorendorff.com/articles/unicode/ is broken | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 221 |  | 
 | 222 | Wikipedia entries are often helpful; see the entries for "character encoding" | 
 | 223 | <http://en.wikipedia.org/wiki/Character_encoding> and UTF-8 | 
 | 224 | <http://en.wikipedia.org/wiki/UTF-8>, for example. | 
 | 225 |  | 
 | 226 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 227 | Python's Unicode Support | 
 | 228 | ======================== | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 229 |  | 
 | 230 | Now that you've learned the rudiments of Unicode, we can look at Python's | 
 | 231 | Unicode features. | 
 | 232 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 233 | The String Type | 
 | 234 | --------------- | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 235 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 236 | Since Python 3.0, the language features a ``str`` type that contain Unicode | 
 | 237 | characters, meaning any string created using ``"unicode rocks!"``, ``'unicode | 
| Georg Brandl | 4f5f98d | 2009-05-04 21:01:20 +0000 | [diff] [blame] | 238 | rocks!'``, or the triple-quoted string syntax is stored as Unicode. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 239 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 240 | To insert a Unicode character that is not part ASCII, e.g., any letters with | 
 | 241 | accents, one can use escape sequences in their string literals as such:: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 242 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 243 |    >>> "\N{GREEK CAPITAL LETTER DELTA}"  # Using the character name | 
 | 244 |    '\u0394' | 
 | 245 |    >>> "\u0394"                          # Using a 16-bit hex value | 
 | 246 |    '\u0394' | 
 | 247 |    >>> "\U00000394"                      # Using a 32-bit hex value | 
 | 248 |    '\u0394' | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 249 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 250 | In addition, one can create a string using the :func:`decode` method of | 
 | 251 | :class:`bytes`.  This method takes an encoding, such as UTF-8, and, optionally, | 
 | 252 | an *errors* argument. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 253 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 254 | The *errors* argument specifies the response when the input string can't be | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 255 | converted according to the encoding's rules.  Legal values for this argument are | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 256 | 'strict' (raise a :exc:`UnicodeDecodeError` exception), 'replace' (use U+FFFD, | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 257 | 'REPLACEMENT CHARACTER'), or 'ignore' (just leave the character out of the | 
 | 258 | Unicode result).  The following examples show the differences:: | 
 | 259 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 260 |     >>> b'\x80abc'.decode("utf-8", "strict") | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 261 |     Traceback (most recent call last): | 
 | 262 |       File "<stdin>", line 1, in ? | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 263 |     UnicodeDecodeError: 'utf8' codec can't decode byte 0x80 in position 0: | 
 | 264 |                         unexpected code byte | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 265 |     >>> b'\x80abc'.decode("utf-8", "replace") | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 266 |     '�abc' | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 267 |     >>> b'\x80abc'.decode("utf-8", "ignore") | 
 | 268 |     'abc' | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 269 |  | 
| Benjamin Peterson | d7c3ed5 | 2010-06-27 22:32:30 +0000 | [diff] [blame] | 270 | Encodings are specified as strings containing the encoding's name.  Python 3.2 | 
 | 271 | comes with roughly 100 different encodings; see the Python Library Reference at | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 272 | :ref:`standard-encodings` for a list.  Some encodings have multiple names; for | 
 | 273 | example, 'latin-1', 'iso_8859_1' and '8859' are all synonyms for the same | 
 | 274 | encoding. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 275 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 276 | One-character Unicode strings can also be created with the :func:`chr` | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 277 | built-in function, which takes integers and returns a Unicode string of length 1 | 
 | 278 | that contains the corresponding code point.  The reverse operation is the | 
 | 279 | built-in :func:`ord` function that takes a one-character Unicode string and | 
 | 280 | returns the code point value:: | 
 | 281 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 282 |     >>> chr(57344) | 
 | 283 |     '\ue000' | 
 | 284 |     >>> ord('\ue000') | 
 | 285 |     57344 | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 286 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 287 | Converting to Bytes | 
 | 288 | ------------------- | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 289 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 290 | Another important str method is ``.encode([encoding], [errors='strict'])``, | 
 | 291 | which returns a ``bytes`` representation of the Unicode string, encoded in the | 
 | 292 | requested encoding.  The ``errors`` parameter is the same as the parameter of | 
 | 293 | the :meth:`decode` method, with one additional possibility; as well as 'strict', | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 294 | 'ignore', and 'replace' (which in this case inserts a question mark instead of | 
 | 295 | the unencodable character), you can also pass 'xmlcharrefreplace' which uses | 
 | 296 | XML's character references.  The following example shows the different results:: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 297 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 298 |     >>> u = chr(40960) + 'abcd' + chr(1972) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 299 |     >>> u.encode('utf-8') | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 300 |     b'\xea\x80\x80abcd\xde\xb4' | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 301 |     >>> u.encode('ascii') | 
 | 302 |     Traceback (most recent call last): | 
 | 303 |       File "<stdin>", line 1, in ? | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 304 |     UnicodeEncodeError: 'ascii' codec can't encode character '\ua000' in | 
 | 305 |                         position 0: ordinal not in range(128) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 306 |     >>> u.encode('ascii', 'ignore') | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 307 |     b'abcd' | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 308 |     >>> u.encode('ascii', 'replace') | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 309 |     b'?abcd?' | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 310 |     >>> u.encode('ascii', 'xmlcharrefreplace') | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 311 |     b'ꀀabcd޴' | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 312 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 313 | The low-level routines for registering and accessing the available encodings are | 
 | 314 | found in the :mod:`codecs` module.  However, the encoding and decoding functions | 
 | 315 | returned by this module are usually more low-level than is comfortable, so I'm | 
 | 316 | not going to describe the :mod:`codecs` module here.  If you need to implement a | 
 | 317 | completely new encoding, you'll need to learn about the :mod:`codecs` module | 
 | 318 | interfaces, but implementing encodings is a specialized task that also won't be | 
 | 319 | covered here.  Consult the Python documentation to learn more about this module. | 
 | 320 |  | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 321 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 322 | Unicode Literals in Python Source Code | 
 | 323 | -------------------------------------- | 
 | 324 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 325 | In Python source code, specific Unicode code points can be written using the | 
 | 326 | ``\u`` escape sequence, which is followed by four hex digits giving the code | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 327 | point.  The ``\U`` escape sequence is similar, but expects eight hex digits, | 
 | 328 | not four:: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 329 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 330 |     >>> s = "a\xac\u1234\u20ac\U00008000" | 
 | 331 |               ^^^^ two-digit hex escape | 
 | 332 |                    ^^^^^ four-digit Unicode escape | 
 | 333 |                               ^^^^^^^^^^ eight-digit Unicode escape | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 334 |     >>> for c in s:  print(ord(c), end=" ") | 
 | 335 |     ... | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 336 |     97 172 4660 8364 32768 | 
 | 337 |  | 
 | 338 | Using escape sequences for code points greater than 127 is fine in small doses, | 
 | 339 | but becomes an annoyance if you're using many accented characters, as you would | 
 | 340 | in a program with messages in French or some other accent-using language.  You | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 341 | can also assemble strings using the :func:`chr` built-in function, but this is | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 342 | even more tedious. | 
 | 343 |  | 
 | 344 | Ideally, you'd want to be able to write literals in your language's natural | 
 | 345 | encoding.  You could then edit Python source code with your favorite editor | 
 | 346 | which would display the accented characters naturally, and have the right | 
 | 347 | characters used at runtime. | 
 | 348 |  | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 349 | Python supports writing source code in UTF-8 by default, but you can use almost | 
 | 350 | any encoding if you declare the encoding being used.  This is done by including | 
 | 351 | a special comment as either the first or second line of the source file:: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 352 |  | 
 | 353 |     #!/usr/bin/env python | 
 | 354 |     # -*- coding: latin-1 -*- | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 355 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 356 |     u = 'abcdé' | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 357 |     print(ord(u[-1])) | 
 | 358 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 359 | The syntax is inspired by Emacs's notation for specifying variables local to a | 
 | 360 | file.  Emacs supports many different variables, but Python only supports | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 361 | 'coding'.  The ``-*-`` symbols indicate to Emacs that the comment is special; | 
 | 362 | they have no significance to Python but are a convention.  Python looks for | 
 | 363 | ``coding: name`` or ``coding=name`` in the comment. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 364 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 365 | If you don't include such a comment, the default encoding used will be UTF-8 as | 
 | 366 | already mentioned. | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 367 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 368 |  | 
 | 369 | Unicode Properties | 
 | 370 | ------------------ | 
 | 371 |  | 
 | 372 | The Unicode specification includes a database of information about code points. | 
 | 373 | For each code point that's defined, the information includes the character's | 
 | 374 | name, its category, the numeric value if applicable (Unicode has characters | 
 | 375 | representing the Roman numerals and fractions such as one-third and | 
 | 376 | four-fifths).  There are also properties related to the code point's use in | 
 | 377 | bidirectional text and other display-related properties. | 
 | 378 |  | 
 | 379 | The following program displays some information about several characters, and | 
 | 380 | prints the numeric value of one particular character:: | 
 | 381 |  | 
 | 382 |     import unicodedata | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 383 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 384 |     u = chr(233) + chr(0x0bf2) + chr(3972) + chr(6000) + chr(13231) | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 385 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 386 |     for i, c in enumerate(u): | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 387 |         print(i, '%04x' % ord(c), unicodedata.category(c), end=" ") | 
 | 388 |         print(unicodedata.name(c)) | 
 | 389 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 390 |     # Get numeric value of second character | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 391 |     print(unicodedata.numeric(u[1])) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 392 |  | 
 | 393 | When run, this prints:: | 
 | 394 |  | 
 | 395 |     0 00e9 Ll LATIN SMALL LETTER E WITH ACUTE | 
 | 396 |     1 0bf2 No TAMIL NUMBER ONE THOUSAND | 
 | 397 |     2 0f84 Mn TIBETAN MARK HALANTA | 
 | 398 |     3 1770 Lo TAGBANWA LETTER SA | 
 | 399 |     4 33af So SQUARE RAD OVER S SQUARED | 
 | 400 |     1000.0 | 
 | 401 |  | 
 | 402 | The category codes are abbreviations describing the nature of the character. | 
 | 403 | These are grouped into categories such as "Letter", "Number", "Punctuation", or | 
 | 404 | "Symbol", which in turn are broken up into subcategories.  To take the codes | 
 | 405 | from the above output, ``'Ll'`` means 'Letter, lowercase', ``'No'`` means | 
 | 406 | "Number, other", ``'Mn'`` is "Mark, nonspacing", and ``'So'`` is "Symbol, | 
 | 407 | other".  See | 
| Ezio Melotti | 4c5475d | 2010-03-22 23:16:42 +0000 | [diff] [blame] | 408 | <http://www.unicode.org/reports/tr44/#General_Category_Values> for a | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 409 | list of category codes. | 
 | 410 |  | 
 | 411 | References | 
 | 412 | ---------- | 
 | 413 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 414 | The ``str`` type is described in the Python library reference at | 
 | 415 | :ref:`typesseq`. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 416 |  | 
 | 417 | The documentation for the :mod:`unicodedata` module. | 
 | 418 |  | 
 | 419 | The documentation for the :mod:`codecs` module. | 
 | 420 |  | 
 | 421 | Marc-André Lemburg gave a presentation at EuroPython 2002 titled "Python and | 
 | 422 | Unicode".  A PDF version of his slides is available at | 
| Christian Heimes | dd15f6c | 2008-03-16 00:07:10 +0000 | [diff] [blame] | 423 | <http://downloads.egenix.com/python/Unicode-EPC2002-Talk.pdf>, and is an | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 424 | excellent overview of the design of Python's Unicode features (based on Python | 
 | 425 | 2, where the Unicode string type is called ``unicode`` and literals start with | 
 | 426 | ``u``). | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 427 |  | 
 | 428 |  | 
 | 429 | Reading and Writing Unicode Data | 
 | 430 | ================================ | 
 | 431 |  | 
 | 432 | Once you've written some code that works with Unicode data, the next problem is | 
 | 433 | input/output.  How do you get Unicode strings into your program, and how do you | 
 | 434 | convert Unicode into a form suitable for storage or transmission? | 
 | 435 |  | 
 | 436 | It's possible that you may not need to do anything depending on your input | 
 | 437 | sources and output destinations; you should check whether the libraries used in | 
 | 438 | your application support Unicode natively.  XML parsers often return Unicode | 
 | 439 | data, for example.  Many relational databases also support Unicode-valued | 
 | 440 | columns and can return Unicode values from an SQL query. | 
 | 441 |  | 
 | 442 | Unicode data is usually converted to a particular encoding before it gets | 
 | 443 | written to disk or sent over a socket.  It's possible to do all the work | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 444 | yourself: open a file, read an 8-bit byte string from it, and convert the string | 
 | 445 | with ``str(bytes, encoding)``.  However, the manual approach is not recommended. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 446 |  | 
 | 447 | One problem is the multi-byte nature of encodings; one Unicode character can be | 
 | 448 | represented by several bytes.  If you want to read the file in arbitrary-sized | 
 | 449 | chunks (say, 1K or 4K), you need to write error-handling code to catch the case | 
 | 450 | where only part of the bytes encoding a single Unicode character are read at the | 
 | 451 | end of a chunk.  One solution would be to read the entire file into memory and | 
 | 452 | then perform the decoding, but that prevents you from working with files that | 
 | 453 | are extremely large; if you need to read a 2Gb file, you need 2Gb of RAM. | 
 | 454 | (More, really, since for at least a moment you'd need to have both the encoded | 
 | 455 | string and its Unicode version in memory.) | 
 | 456 |  | 
 | 457 | The solution would be to use the low-level decoding interface to catch the case | 
 | 458 | of partial coding sequences.  The work of implementing this has already been | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 459 | done for you: the built-in :func:`open` function can return a file-like object | 
 | 460 | that assumes the file's contents are in a specified encoding and accepts Unicode | 
 | 461 | parameters for methods such as ``.read()`` and ``.write()``.  This works through | 
 | 462 | :func:`open`\'s *encoding* and *errors* parameters which are interpreted just | 
 | 463 | like those in string objects' :meth:`encode` and :meth:`decode` methods. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 464 |  | 
 | 465 | Reading Unicode from a file is therefore simple:: | 
 | 466 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 467 |     with open('unicode.rst', encoding='utf-8') as f: | 
 | 468 |         for line in f: | 
 | 469 |             print(repr(line)) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 470 |  | 
 | 471 | It's also possible to open files in update mode, allowing both reading and | 
 | 472 | writing:: | 
 | 473 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 474 |     with open('test', encoding='utf-8', mode='w+') as f: | 
 | 475 |         f.write('\u4500 blah blah blah\n') | 
 | 476 |         f.seek(0) | 
 | 477 |         print(repr(f.readline()[:1])) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 478 |  | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 479 | The Unicode character U+FEFF is used as a byte-order mark (BOM), and is often | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 480 | written as the first character of a file in order to assist with autodetection | 
 | 481 | of the file's byte ordering.  Some encodings, such as UTF-16, expect a BOM to be | 
 | 482 | present at the start of a file; when such an encoding is used, the BOM will be | 
 | 483 | automatically written as the first character and will be silently dropped when | 
 | 484 | the file is read.  There are variants of these encodings, such as 'utf-16-le' | 
 | 485 | and 'utf-16-be' for little-endian and big-endian encodings, that specify one | 
 | 486 | particular byte ordering and don't skip the BOM. | 
 | 487 |  | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 488 | In some areas, it is also convention to use a "BOM" at the start of UTF-8 | 
 | 489 | encoded files; the name is misleading since UTF-8 is not byte-order dependent. | 
 | 490 | The mark simply announces that the file is encoded in UTF-8.  Use the | 
 | 491 | 'utf-8-sig' codec to automatically skip the mark if present for reading such | 
 | 492 | files. | 
 | 493 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 494 |  | 
 | 495 | Unicode filenames | 
 | 496 | ----------------- | 
 | 497 |  | 
 | 498 | Most of the operating systems in common use today support filenames that contain | 
 | 499 | arbitrary Unicode characters.  Usually this is implemented by converting the | 
 | 500 | Unicode string into some encoding that varies depending on the system.  For | 
| Georg Brandl | c575c90 | 2008-09-13 17:46:05 +0000 | [diff] [blame] | 501 | example, Mac OS X uses UTF-8 while Windows uses a configurable encoding; on | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 502 | Windows, Python uses the name "mbcs" to refer to whatever the currently | 
 | 503 | configured encoding is.  On Unix systems, there will only be a filesystem | 
 | 504 | encoding if you've set the ``LANG`` or ``LC_CTYPE`` environment variables; if | 
 | 505 | you haven't, the default encoding is ASCII. | 
 | 506 |  | 
 | 507 | The :func:`sys.getfilesystemencoding` function returns the encoding to use on | 
 | 508 | your current system, in case you want to do the encoding manually, but there's | 
 | 509 | not much reason to bother.  When opening a file for reading or writing, you can | 
 | 510 | usually just provide the Unicode string as the filename, and it will be | 
 | 511 | automatically converted to the right encoding for you:: | 
 | 512 |  | 
| Georg Brandl | f694518 | 2008-02-01 11:56:49 +0000 | [diff] [blame] | 513 |     filename = 'filename\u4500abc' | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 514 |     with open(filename, 'w') as f: | 
 | 515 |         f.write('blah\n') | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 516 |  | 
 | 517 | Functions in the :mod:`os` module such as :func:`os.stat` will also accept Unicode | 
 | 518 | filenames. | 
 | 519 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 520 | Function :func:`os.listdir`, which returns filenames, raises an issue: should it return | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 521 | the Unicode version of filenames, or should it return byte strings containing | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 522 | the encoded versions?  :func:`os.listdir` will do both, depending on whether you | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 523 | provided the directory path as a byte string or a Unicode string.  If you pass a | 
 | 524 | Unicode string as the path, filenames will be decoded using the filesystem's | 
 | 525 | encoding and a list of Unicode strings will be returned, while passing a byte | 
 | 526 | path will return the byte string versions of the filenames.  For example, | 
 | 527 | assuming the default filesystem encoding is UTF-8, running the following | 
 | 528 | program:: | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 529 |  | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 530 |    fn = 'filename\u4500abc' | 
 | 531 |    f = open(fn, 'w') | 
 | 532 |    f.close() | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 533 |  | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 534 |    import os | 
 | 535 |    print(os.listdir(b'.')) | 
 | 536 |    print(os.listdir('.')) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 537 |  | 
 | 538 | will produce the following output:: | 
 | 539 |  | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 540 |    amk:~$ python t.py | 
 | 541 |    [b'.svn', b'filename\xe4\x94\x80abc', ...] | 
 | 542 |    ['.svn', 'filename\u4500abc', ...] | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 543 |  | 
 | 544 | The first list contains UTF-8-encoded filenames, and the second list contains | 
 | 545 | the Unicode versions. | 
 | 546 |  | 
| R. David Murray | 01054d7 | 2009-09-12 03:09:02 +0000 | [diff] [blame] | 547 | Note that in most occasions, the Unicode APIs should be used.  The bytes APIs | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 548 | should only be used on systems where undecodable file names can be present, | 
 | 549 | i.e. Unix systems. | 
 | 550 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 551 |  | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 552 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 553 | Tips for Writing Unicode-aware Programs | 
 | 554 | --------------------------------------- | 
 | 555 |  | 
 | 556 | This section provides some suggestions on writing software that deals with | 
 | 557 | Unicode. | 
 | 558 |  | 
 | 559 | The most important tip is: | 
 | 560 |  | 
 | 561 |     Software should only work with Unicode strings internally, converting to a | 
 | 562 |     particular encoding on output. | 
 | 563 |  | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 564 | If you attempt to write processing functions that accept both Unicode and byte | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 565 | strings, you will find your program vulnerable to bugs wherever you combine the | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 566 | two different kinds of strings.  There is no automatic encoding or decoding if | 
 | 567 | you do e.g. ``str + bytes``, a :exc:`TypeError` is raised for this expression. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 568 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 569 | When using data coming from a web browser or some other untrusted source, a | 
 | 570 | common technique is to check for illegal characters in a string before using the | 
 | 571 | string in a generated command line or storing it in a database.  If you're doing | 
 | 572 | this, be careful to check the string once it's in the form that will be used or | 
 | 573 | stored; it's possible for encodings to be used to disguise characters.  This is | 
 | 574 | especially true if the input data also specifies the encoding; many encodings | 
 | 575 | leave the commonly checked-for characters alone, but Python includes some | 
 | 576 | encodings such as ``'base64'`` that modify every single character. | 
 | 577 |  | 
 | 578 | For example, let's say you have a content management system that takes a Unicode | 
 | 579 | filename, and you want to disallow paths with a '/' character.  You might write | 
 | 580 | this code:: | 
 | 581 |  | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 582 |     def read_file(filename, encoding): | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 583 |         if '/' in filename: | 
 | 584 |             raise ValueError("'/' not allowed in filenames") | 
 | 585 |         unicode_name = filename.decode(encoding) | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 586 |         with open(unicode_name, 'r') as f: | 
 | 587 |             # ... return contents of file ... | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 588 |  | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 589 | However, if an attacker could specify the ``'base64'`` encoding, they could pass | 
 | 590 | ``'L2V0Yy9wYXNzd2Q='``, which is the base-64 encoded form of the string | 
 | 591 | ``'/etc/passwd'``, to read a system file.  The above code looks for ``'/'`` | 
 | 592 | characters in the encoded form and misses the dangerous character in the | 
 | 593 | resulting decoded form. | 
 | 594 |  | 
 | 595 | References | 
 | 596 | ---------- | 
 | 597 |  | 
 | 598 | The PDF slides for Marc-André Lemburg's presentation "Writing Unicode-aware | 
 | 599 | Applications in Python" are available at | 
| Christian Heimes | dd15f6c | 2008-03-16 00:07:10 +0000 | [diff] [blame] | 600 | <http://downloads.egenix.com/python/LSM2005-Developing-Unicode-aware-applications-in-Python.pdf> | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 601 | and discuss questions of character encodings as well as how to internationalize | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 602 | and localize an application.  These slides cover Python 2.x only. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 603 |  | 
 | 604 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 605 | Acknowledgements | 
 | 606 | ================ | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 607 |  | 
 | 608 | Thanks to the following people who have noted errors or offered suggestions on | 
 | 609 | this article: Nicholas Bastin, Marius Gedminas, Kent Johnson, Ken Krugler, | 
 | 610 | Marc-André Lemburg, Martin von Löwis, Chad Whitacre. | 
 | 611 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 612 | .. comment | 
 | 613 |    Revision History | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 614 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 615 |    Version 1.0: posted August 5 2005. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 616 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 617 |    Version 1.01: posted August 7 2005.  Corrects factual and markup errors; adds | 
 | 618 |    several links. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 619 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 620 |    Version 1.02: posted August 16 2005.  Corrects factual errors. | 
| Georg Brandl | 0c07422 | 2008-11-22 10:26:59 +0000 | [diff] [blame] | 621 |  | 
| Alexander Belopolsky | 93a6b13 | 2010-11-19 16:09:58 +0000 | [diff] [blame^] | 622 |    Version 1.1: Feb-Nov 2008.  Updates the document with respect to Python 3 changes. | 
 | 623 |  | 
 | 624 |    Version 1.11: posted June 20 2010.  Notes that Python 3.x is not covered, | 
 | 625 |    and that the HOWTO only covers 2.x. | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 626 |  | 
| Benjamin Peterson | d7c3ed5 | 2010-06-27 22:32:30 +0000 | [diff] [blame] | 627 | .. comment Describe Python 3.x support (new section? new document?) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 628 | .. comment Additional topic: building Python w/ UCS2 or UCS4 support | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 629 | .. comment Describe use of codecs.StreamRecoder and StreamReaderWriter | 
 | 630 |  | 
| Georg Brandl | 6911e3c | 2007-09-04 07:15:32 +0000 | [diff] [blame] | 631 | .. comment | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 632 |    Original outline: | 
 | 633 |  | 
 | 634 |    - [ ] Unicode introduction | 
 | 635 |        - [ ] ASCII | 
 | 636 |        - [ ] Terms | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 637 |            - [ ] Character | 
 | 638 |            - [ ] Code point | 
 | 639 |          - [ ] Encodings | 
 | 640 |             - [ ] Common encodings: ASCII, Latin-1, UTF-8 | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 641 |        - [ ] Unicode Python type | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 642 |            - [ ] Writing unicode literals | 
 | 643 |                - [ ] Obscurity: -U switch | 
 | 644 |            - [ ] Built-ins | 
 | 645 |                - [ ] unichr() | 
 | 646 |                - [ ] ord() | 
 | 647 |                - [ ] unicode() constructor | 
 | 648 |            - [ ] Unicode type | 
 | 649 |                - [ ] encode(), decode() methods | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 650 |        - [ ] Unicodedata module for character properties | 
 | 651 |        - [ ] I/O | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 652 |            - [ ] Reading/writing Unicode data into files | 
 | 653 |                - [ ] Byte-order marks | 
 | 654 |            - [ ] Unicode filenames | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 655 |        - [ ] Writing Unicode programs | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 656 |            - [ ] Do everything in Unicode | 
 | 657 |            - [ ] Declaring source code encodings (PEP 263) | 
| Georg Brandl | 116aa62 | 2007-08-15 14:28:22 +0000 | [diff] [blame] | 658 |        - [ ] Other issues | 
| Georg Brandl | a1c6a1c | 2009-01-03 21:26:05 +0000 | [diff] [blame] | 659 |            - [ ] Building Python (UCS2, UCS4) |