[2] | 1 | *****************
|
---|
| 2 | Unicode HOWTO
|
---|
| 3 | *****************
|
---|
| 4 |
|
---|
[391] | 5 | :Release: 1.03
|
---|
[2] | 6 |
|
---|
[391] | 7 | This HOWTO discusses Python 2.x's support for Unicode, and explains
|
---|
| 8 | various problems that people commonly encounter when trying to work
|
---|
| 9 | with Unicode. For the Python 3 version, see
|
---|
| 10 | <http://docs.python.org/py3k/howto/unicode.html>.
|
---|
[2] | 11 |
|
---|
| 12 | Introduction to Unicode
|
---|
| 13 | =======================
|
---|
| 14 |
|
---|
| 15 | History of Character Codes
|
---|
| 16 | --------------------------
|
---|
| 17 |
|
---|
| 18 | In 1968, the American Standard Code for Information Interchange, better known by
|
---|
| 19 | its acronym ASCII, was standardized. ASCII defined numeric codes for various
|
---|
| 20 | characters, with the numeric values running from 0 to
|
---|
| 21 | 127. For example, the lowercase letter 'a' is assigned 97 as its code
|
---|
| 22 | value.
|
---|
| 23 |
|
---|
| 24 | ASCII was an American-developed standard, so it only defined unaccented
|
---|
| 25 | characters. There was an 'e', but no 'é' or 'Ã'. This meant that languages
|
---|
| 26 | which required accented characters couldn't be faithfully represented in ASCII.
|
---|
| 27 | (Actually the missing accents matter for English, too, which contains words such
|
---|
| 28 | as 'naïve' and 'café', and some publications have house styles which require
|
---|
| 29 | spellings such as 'coöperate'.)
|
---|
| 30 |
|
---|
| 31 | For a while people just wrote programs that didn't display accents. I remember
|
---|
| 32 | looking at Apple ][ BASIC programs, published in French-language publications in
|
---|
| 33 | the mid-1980s, that had lines like these::
|
---|
| 34 |
|
---|
| 35 | PRINT "FICHIER EST COMPLETE."
|
---|
| 36 | PRINT "CARACTERE NON ACCEPTE."
|
---|
| 37 |
|
---|
| 38 | Those messages should contain accents, and they just look wrong to someone who
|
---|
| 39 | can read French.
|
---|
| 40 |
|
---|
| 41 | In the 1980s, almost all personal computers were 8-bit, meaning that bytes could
|
---|
| 42 | hold values ranging from 0 to 255. ASCII codes only went up to 127, so some
|
---|
| 43 | machines assigned values between 128 and 255 to accented characters. Different
|
---|
| 44 | machines had different codes, however, which led to problems exchanging files.
|
---|
| 45 | Eventually various commonly used sets of values for the 128-255 range emerged.
|
---|
| 46 | Some were true standards, defined by the International Standards Organization,
|
---|
| 47 | and some were **de facto** conventions that were invented by one company or
|
---|
| 48 | another and managed to catch on.
|
---|
| 49 |
|
---|
| 50 | 255 characters aren't very many. For example, you can't fit both the accented
|
---|
| 51 | characters used in Western Europe and the Cyrillic alphabet used for Russian
|
---|
| 52 | into the 128-255 range because there are more than 127 such characters.
|
---|
| 53 |
|
---|
| 54 | You could write files using different codes (all your Russian files in a coding
|
---|
| 55 | system called KOI8, all your French files in a different coding system called
|
---|
| 56 | Latin1), but what if you wanted to write a French document that quotes some
|
---|
| 57 | Russian text? In the 1980s people began to want to solve this problem, and the
|
---|
| 58 | Unicode standardization effort began.
|
---|
| 59 |
|
---|
| 60 | Unicode started out using 16-bit characters instead of 8-bit characters. 16
|
---|
| 61 | bits means you have 2^16 = 65,536 distinct values available, making it possible
|
---|
| 62 | to represent many different characters from many different alphabets; an initial
|
---|
| 63 | goal was to have Unicode contain the alphabets for every single human language.
|
---|
| 64 | It turns out that even 16 bits isn't enough to meet that goal, and the modern
|
---|
| 65 | Unicode specification uses a wider range of codes, 0-1,114,111 (0x10ffff in
|
---|
| 66 | base-16).
|
---|
| 67 |
|
---|
| 68 | There's a related ISO standard, ISO 10646. Unicode and ISO 10646 were
|
---|
| 69 | originally separate efforts, but the specifications were merged with the 1.1
|
---|
| 70 | revision of Unicode.
|
---|
| 71 |
|
---|
| 72 | (This discussion of Unicode's history is highly simplified. I don't think the
|
---|
| 73 | average Python programmer needs to worry about the historical details; consult
|
---|
| 74 | the Unicode consortium site listed in the References for more information.)
|
---|
| 75 |
|
---|
| 76 |
|
---|
| 77 | Definitions
|
---|
| 78 | -----------
|
---|
| 79 |
|
---|
| 80 | A **character** is the smallest possible component of a text. 'A', 'B', 'C',
|
---|
| 81 | etc., are all different characters. So are 'Ã' and 'Ã'. Characters are
|
---|
| 82 | abstractions, and vary depending on the language or context you're talking
|
---|
| 83 | about. For example, the symbol for ohms (âŠ) is usually drawn much like the
|
---|
| 84 | capital letter omega (Ω) in the Greek alphabet (they may even be the same in
|
---|
| 85 | some fonts), but these are two different characters that have different
|
---|
| 86 | meanings.
|
---|
| 87 |
|
---|
| 88 | The Unicode standard describes how characters are represented by **code
|
---|
| 89 | points**. A code point is an integer value, usually denoted in base 16. In the
|
---|
| 90 | standard, a code point is written using the notation U+12ca to mean the
|
---|
| 91 | character with value 0x12ca (4810 decimal). The Unicode standard contains a lot
|
---|
| 92 | of tables listing characters and their corresponding code points::
|
---|
| 93 |
|
---|
| 94 | 0061 'a'; LATIN SMALL LETTER A
|
---|
| 95 | 0062 'b'; LATIN SMALL LETTER B
|
---|
| 96 | 0063 'c'; LATIN SMALL LETTER C
|
---|
| 97 | ...
|
---|
| 98 | 007B '{'; LEFT CURLY BRACKET
|
---|
| 99 |
|
---|
| 100 | Strictly, these definitions imply that it's meaningless to say 'this is
|
---|
| 101 | character U+12ca'. U+12ca is a code point, which represents some particular
|
---|
| 102 | character; in this case, it represents the character 'ETHIOPIC SYLLABLE WI'. In
|
---|
| 103 | informal contexts, this distinction between code points and characters will
|
---|
| 104 | sometimes be forgotten.
|
---|
| 105 |
|
---|
| 106 | A character is represented on a screen or on paper by a set of graphical
|
---|
| 107 | elements that's called a **glyph**. The glyph for an uppercase A, for example,
|
---|
| 108 | is two diagonal strokes and a horizontal stroke, though the exact details will
|
---|
| 109 | depend on the font being used. Most Python code doesn't need to worry about
|
---|
| 110 | glyphs; figuring out the correct glyph to display is generally the job of a GUI
|
---|
| 111 | toolkit or a terminal's font renderer.
|
---|
| 112 |
|
---|
| 113 |
|
---|
| 114 | Encodings
|
---|
| 115 | ---------
|
---|
| 116 |
|
---|
| 117 | To summarize the previous section: a Unicode string is a sequence of code
|
---|
| 118 | points, which are numbers from 0 to 0x10ffff. This sequence needs to be
|
---|
| 119 | represented as a set of bytes (meaning, values from 0-255) in memory. The rules
|
---|
| 120 | for translating a Unicode string into a sequence of bytes are called an
|
---|
| 121 | **encoding**.
|
---|
| 122 |
|
---|
| 123 | The first encoding you might think of is an array of 32-bit integers. In this
|
---|
| 124 | representation, the string "Python" would look like this::
|
---|
| 125 |
|
---|
| 126 | P y t h o n
|
---|
| 127 | 0x50 00 00 00 79 00 00 00 74 00 00 00 68 00 00 00 6f 00 00 00 6e 00 00 00
|
---|
| 128 | 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
---|
| 129 |
|
---|
| 130 | This representation is straightforward but using it presents a number of
|
---|
| 131 | problems.
|
---|
| 132 |
|
---|
| 133 | 1. It's not portable; different processors order the bytes differently.
|
---|
| 134 |
|
---|
| 135 | 2. It's very wasteful of space. In most texts, the majority of the code points
|
---|
| 136 | are less than 127, or less than 255, so a lot of space is occupied by zero
|
---|
| 137 | bytes. The above string takes 24 bytes compared to the 6 bytes needed for an
|
---|
| 138 | ASCII representation. Increased RAM usage doesn't matter too much (desktop
|
---|
| 139 | computers have megabytes of RAM, and strings aren't usually that large), but
|
---|
| 140 | expanding our usage of disk and network bandwidth by a factor of 4 is
|
---|
| 141 | intolerable.
|
---|
| 142 |
|
---|
| 143 | 3. It's not compatible with existing C functions such as ``strlen()``, so a new
|
---|
| 144 | family of wide string functions would need to be used.
|
---|
| 145 |
|
---|
| 146 | 4. Many Internet standards are defined in terms of textual data, and can't
|
---|
| 147 | handle content with embedded zero bytes.
|
---|
| 148 |
|
---|
[391] | 149 | Generally people don't use this encoding, instead choosing other
|
---|
| 150 | encodings that are more efficient and convenient. UTF-8 is probably
|
---|
| 151 | the most commonly supported encoding; it will be discussed below.
|
---|
[2] | 152 |
|
---|
| 153 | Encodings don't have to handle every possible Unicode character, and most
|
---|
| 154 | encodings don't. For example, Python's default encoding is the 'ascii'
|
---|
| 155 | encoding. The rules for converting a Unicode string into the ASCII encoding are
|
---|
| 156 | simple; for each code point:
|
---|
| 157 |
|
---|
| 158 | 1. If the code point is < 128, each byte is the same as the value of the code
|
---|
| 159 | point.
|
---|
| 160 |
|
---|
| 161 | 2. If the code point is 128 or greater, the Unicode string can't be represented
|
---|
| 162 | in this encoding. (Python raises a :exc:`UnicodeEncodeError` exception in this
|
---|
| 163 | case.)
|
---|
| 164 |
|
---|
| 165 | Latin-1, also known as ISO-8859-1, is a similar encoding. Unicode code points
|
---|
| 166 | 0-255 are identical to the Latin-1 values, so converting to this encoding simply
|
---|
| 167 | requires converting code points to byte values; if a code point larger than 255
|
---|
| 168 | is encountered, the string can't be encoded into Latin-1.
|
---|
| 169 |
|
---|
| 170 | Encodings don't have to be simple one-to-one mappings like Latin-1. Consider
|
---|
| 171 | IBM's EBCDIC, which was used on IBM mainframes. Letter values weren't in one
|
---|
| 172 | block: 'a' through 'i' had values from 129 to 137, but 'j' through 'r' were 145
|
---|
| 173 | through 153. If you wanted to use EBCDIC as an encoding, you'd probably use
|
---|
| 174 | some sort of lookup table to perform the conversion, but this is largely an
|
---|
| 175 | internal detail.
|
---|
| 176 |
|
---|
| 177 | UTF-8 is one of the most commonly used encodings. UTF stands for "Unicode
|
---|
| 178 | Transformation Format", and the '8' means that 8-bit numbers are used in the
|
---|
| 179 | encoding. (There's also a UTF-16 encoding, but it's less frequently used than
|
---|
| 180 | UTF-8.) UTF-8 uses the following rules:
|
---|
| 181 |
|
---|
| 182 | 1. If the code point is <128, it's represented by the corresponding byte value.
|
---|
| 183 | 2. If the code point is between 128 and 0x7ff, it's turned into two byte values
|
---|
| 184 | between 128 and 255.
|
---|
| 185 | 3. Code points >0x7ff are turned into three- or four-byte sequences, where each
|
---|
| 186 | byte of the sequence is between 128 and 255.
|
---|
| 187 |
|
---|
| 188 | UTF-8 has several convenient properties:
|
---|
| 189 |
|
---|
| 190 | 1. It can handle any Unicode code point.
|
---|
| 191 | 2. A Unicode string is turned into a string of bytes containing no embedded zero
|
---|
| 192 | bytes. This avoids byte-ordering issues, and means UTF-8 strings can be
|
---|
| 193 | processed by C functions such as ``strcpy()`` and sent through protocols that
|
---|
| 194 | can't handle zero bytes.
|
---|
| 195 | 3. A string of ASCII text is also valid UTF-8 text.
|
---|
| 196 | 4. UTF-8 is fairly compact; the majority of code points are turned into two
|
---|
| 197 | bytes, and values less than 128 occupy only a single byte.
|
---|
| 198 | 5. If bytes are corrupted or lost, it's possible to determine the start of the
|
---|
| 199 | next UTF-8-encoded code point and resynchronize. It's also unlikely that
|
---|
| 200 | random 8-bit data will look like valid UTF-8.
|
---|
| 201 |
|
---|
| 202 |
|
---|
| 203 |
|
---|
| 204 | References
|
---|
| 205 | ----------
|
---|
| 206 |
|
---|
| 207 | The Unicode Consortium site at <http://www.unicode.org> has character charts, a
|
---|
| 208 | glossary, and PDF versions of the Unicode specification. Be prepared for some
|
---|
| 209 | difficult reading. <http://www.unicode.org/history/> is a chronology of the
|
---|
| 210 | origin and development of Unicode.
|
---|
| 211 |
|
---|
| 212 | To help understand the standard, Jukka Korpela has written an introductory guide
|
---|
| 213 | to reading the Unicode character tables, available at
|
---|
| 214 | <http://www.cs.tut.fi/~jkorpela/unicode/guide.html>.
|
---|
| 215 |
|
---|
| 216 | Another good introductory article was written by Joel Spolsky
|
---|
| 217 | <http://www.joelonsoftware.com/articles/Unicode.html>.
|
---|
| 218 | If this introduction didn't make things clear to you, you should try reading this
|
---|
| 219 | alternate article before continuing.
|
---|
| 220 |
|
---|
| 221 | .. Jason Orendorff XXX http://www.jorendorff.com/articles/unicode/ is broken
|
---|
| 222 |
|
---|
| 223 | Wikipedia entries are often helpful; see the entries for "character encoding"
|
---|
| 224 | <http://en.wikipedia.org/wiki/Character_encoding> and UTF-8
|
---|
| 225 | <http://en.wikipedia.org/wiki/UTF-8>, for example.
|
---|
| 226 |
|
---|
| 227 |
|
---|
[391] | 228 | Python 2.x's Unicode Support
|
---|
| 229 | ============================
|
---|
[2] | 230 |
|
---|
| 231 | Now that you've learned the rudiments of Unicode, we can look at Python's
|
---|
| 232 | Unicode features.
|
---|
| 233 |
|
---|
| 234 |
|
---|
| 235 | The Unicode Type
|
---|
| 236 | ----------------
|
---|
| 237 |
|
---|
| 238 | Unicode strings are expressed as instances of the :class:`unicode` type, one of
|
---|
| 239 | Python's repertoire of built-in types. It derives from an abstract type called
|
---|
| 240 | :class:`basestring`, which is also an ancestor of the :class:`str` type; you can
|
---|
| 241 | therefore check if a value is a string type with ``isinstance(value,
|
---|
| 242 | basestring)``. Under the hood, Python represents Unicode strings as either 16-
|
---|
| 243 | or 32-bit integers, depending on how the Python interpreter was compiled.
|
---|
| 244 |
|
---|
| 245 | The :func:`unicode` constructor has the signature ``unicode(string[, encoding,
|
---|
| 246 | errors])``. All of its arguments should be 8-bit strings. The first argument
|
---|
| 247 | is converted to Unicode using the specified encoding; if you leave off the
|
---|
| 248 | ``encoding`` argument, the ASCII encoding is used for the conversion, so
|
---|
| 249 | characters greater than 127 will be treated as errors::
|
---|
| 250 |
|
---|
| 251 | >>> unicode('abcdef')
|
---|
| 252 | u'abcdef'
|
---|
| 253 | >>> s = unicode('abcdef')
|
---|
| 254 | >>> type(s)
|
---|
| 255 | <type 'unicode'>
|
---|
[391] | 256 | >>> unicode('abcdef' + chr(255)) #doctest: +NORMALIZE_WHITESPACE
|
---|
[2] | 257 | Traceback (most recent call last):
|
---|
[391] | 258 | ...
|
---|
[2] | 259 | UnicodeDecodeError: 'ascii' codec can't decode byte 0xff in position 6:
|
---|
[391] | 260 | ordinal not in range(128)
|
---|
[2] | 261 |
|
---|
| 262 | The ``errors`` argument specifies the response when the input string can't be
|
---|
| 263 | converted according to the encoding's rules. Legal values for this argument are
|
---|
| 264 | 'strict' (raise a ``UnicodeDecodeError`` exception), 'replace' (add U+FFFD,
|
---|
| 265 | 'REPLACEMENT CHARACTER'), or 'ignore' (just leave the character out of the
|
---|
| 266 | Unicode result). The following examples show the differences::
|
---|
| 267 |
|
---|
[391] | 268 | >>> unicode('\x80abc', errors='strict') #doctest: +NORMALIZE_WHITESPACE
|
---|
[2] | 269 | Traceback (most recent call last):
|
---|
[391] | 270 | ...
|
---|
[2] | 271 | UnicodeDecodeError: 'ascii' codec can't decode byte 0x80 in position 0:
|
---|
[391] | 272 | ordinal not in range(128)
|
---|
[2] | 273 | >>> unicode('\x80abc', errors='replace')
|
---|
| 274 | u'\ufffdabc'
|
---|
| 275 | >>> unicode('\x80abc', errors='ignore')
|
---|
| 276 | u'abc'
|
---|
| 277 |
|
---|
[391] | 278 | Encodings are specified as strings containing the encoding's name. Python 2.7
|
---|
[2] | 279 | comes with roughly 100 different encodings; see the Python Library Reference at
|
---|
| 280 | :ref:`standard-encodings` for a list. Some encodings
|
---|
| 281 | have multiple names; for example, 'latin-1', 'iso_8859_1' and '8859' are all
|
---|
| 282 | synonyms for the same encoding.
|
---|
| 283 |
|
---|
| 284 | One-character Unicode strings can also be created with the :func:`unichr`
|
---|
| 285 | built-in function, which takes integers and returns a Unicode string of length 1
|
---|
| 286 | that contains the corresponding code point. The reverse operation is the
|
---|
| 287 | built-in :func:`ord` function that takes a one-character Unicode string and
|
---|
| 288 | returns the code point value::
|
---|
| 289 |
|
---|
| 290 | >>> unichr(40960)
|
---|
| 291 | u'\ua000'
|
---|
| 292 | >>> ord(u'\ua000')
|
---|
| 293 | 40960
|
---|
| 294 |
|
---|
| 295 | Instances of the :class:`unicode` type have many of the same methods as the
|
---|
| 296 | 8-bit string type for operations such as searching and formatting::
|
---|
| 297 |
|
---|
| 298 | >>> s = u'Was ever feather so lightly blown to and fro as this multitude?'
|
---|
| 299 | >>> s.count('e')
|
---|
| 300 | 5
|
---|
| 301 | >>> s.find('feather')
|
---|
| 302 | 9
|
---|
| 303 | >>> s.find('bird')
|
---|
| 304 | -1
|
---|
| 305 | >>> s.replace('feather', 'sand')
|
---|
| 306 | u'Was ever sand so lightly blown to and fro as this multitude?'
|
---|
| 307 | >>> s.upper()
|
---|
| 308 | u'WAS EVER FEATHER SO LIGHTLY BLOWN TO AND FRO AS THIS MULTITUDE?'
|
---|
| 309 |
|
---|
| 310 | Note that the arguments to these methods can be Unicode strings or 8-bit
|
---|
| 311 | strings. 8-bit strings will be converted to Unicode before carrying out the
|
---|
| 312 | operation; Python's default ASCII encoding will be used, so characters greater
|
---|
| 313 | than 127 will cause an exception::
|
---|
| 314 |
|
---|
[391] | 315 | >>> s.find('Was\x9f') #doctest: +NORMALIZE_WHITESPACE
|
---|
[2] | 316 | Traceback (most recent call last):
|
---|
[391] | 317 | ...
|
---|
| 318 | UnicodeDecodeError: 'ascii' codec can't decode byte 0x9f in position 3:
|
---|
| 319 | ordinal not in range(128)
|
---|
[2] | 320 | >>> s.find(u'Was\x9f')
|
---|
| 321 | -1
|
---|
| 322 |
|
---|
| 323 | Much Python code that operates on strings will therefore work with Unicode
|
---|
| 324 | strings without requiring any changes to the code. (Input and output code needs
|
---|
| 325 | more updating for Unicode; more on this later.)
|
---|
| 326 |
|
---|
| 327 | Another important method is ``.encode([encoding], [errors='strict'])``, which
|
---|
| 328 | returns an 8-bit string version of the Unicode string, encoded in the requested
|
---|
| 329 | encoding. The ``errors`` parameter is the same as the parameter of the
|
---|
| 330 | ``unicode()`` constructor, with one additional possibility; as well as 'strict',
|
---|
| 331 | 'ignore', and 'replace', you can also pass 'xmlcharrefreplace' which uses XML's
|
---|
| 332 | character references. The following example shows the different results::
|
---|
| 333 |
|
---|
| 334 | >>> u = unichr(40960) + u'abcd' + unichr(1972)
|
---|
| 335 | >>> u.encode('utf-8')
|
---|
| 336 | '\xea\x80\x80abcd\xde\xb4'
|
---|
[391] | 337 | >>> u.encode('ascii') #doctest: +NORMALIZE_WHITESPACE
|
---|
[2] | 338 | Traceback (most recent call last):
|
---|
[391] | 339 | ...
|
---|
| 340 | UnicodeEncodeError: 'ascii' codec can't encode character u'\ua000' in
|
---|
| 341 | position 0: ordinal not in range(128)
|
---|
[2] | 342 | >>> u.encode('ascii', 'ignore')
|
---|
| 343 | 'abcd'
|
---|
| 344 | >>> u.encode('ascii', 'replace')
|
---|
| 345 | '?abcd?'
|
---|
| 346 | >>> u.encode('ascii', 'xmlcharrefreplace')
|
---|
| 347 | 'ꀀabcd޴'
|
---|
| 348 |
|
---|
| 349 | Python's 8-bit strings have a ``.decode([encoding], [errors])`` method that
|
---|
| 350 | interprets the string using the given encoding::
|
---|
| 351 |
|
---|
| 352 | >>> u = unichr(40960) + u'abcd' + unichr(1972) # Assemble a string
|
---|
| 353 | >>> utf8_version = u.encode('utf-8') # Encode as UTF-8
|
---|
| 354 | >>> type(utf8_version), utf8_version
|
---|
| 355 | (<type 'str'>, '\xea\x80\x80abcd\xde\xb4')
|
---|
| 356 | >>> u2 = utf8_version.decode('utf-8') # Decode using UTF-8
|
---|
| 357 | >>> u == u2 # The two strings match
|
---|
| 358 | True
|
---|
| 359 |
|
---|
| 360 | The low-level routines for registering and accessing the available encodings are
|
---|
| 361 | found in the :mod:`codecs` module. However, the encoding and decoding functions
|
---|
| 362 | returned by this module are usually more low-level than is comfortable, so I'm
|
---|
| 363 | not going to describe the :mod:`codecs` module here. If you need to implement a
|
---|
| 364 | completely new encoding, you'll need to learn about the :mod:`codecs` module
|
---|
| 365 | interfaces, but implementing encodings is a specialized task that also won't be
|
---|
| 366 | covered here. Consult the Python documentation to learn more about this module.
|
---|
| 367 |
|
---|
| 368 | The most commonly used part of the :mod:`codecs` module is the
|
---|
| 369 | :func:`codecs.open` function which will be discussed in the section on input and
|
---|
| 370 | output.
|
---|
| 371 |
|
---|
| 372 |
|
---|
| 373 | Unicode Literals in Python Source Code
|
---|
| 374 | --------------------------------------
|
---|
| 375 |
|
---|
| 376 | In Python source code, Unicode literals are written as strings prefixed with the
|
---|
| 377 | 'u' or 'U' character: ``u'abcdefghijk'``. Specific code points can be written
|
---|
| 378 | using the ``\u`` escape sequence, which is followed by four hex digits giving
|
---|
| 379 | the code point. The ``\U`` escape sequence is similar, but expects 8 hex
|
---|
| 380 | digits, not 4.
|
---|
| 381 |
|
---|
| 382 | Unicode literals can also use the same escape sequences as 8-bit strings,
|
---|
| 383 | including ``\x``, but ``\x`` only takes two hex digits so it can't express an
|
---|
| 384 | arbitrary code point. Octal escapes can go up to U+01ff, which is octal 777.
|
---|
| 385 |
|
---|
| 386 | ::
|
---|
| 387 |
|
---|
| 388 | >>> s = u"a\xac\u1234\u20ac\U00008000"
|
---|
[391] | 389 | ... # ^^^^ two-digit hex escape
|
---|
| 390 | ... # ^^^^^^ four-digit Unicode escape
|
---|
| 391 | ... # ^^^^^^^^^^ eight-digit Unicode escape
|
---|
[2] | 392 | >>> for c in s: print ord(c),
|
---|
| 393 | ...
|
---|
| 394 | 97 172 4660 8364 32768
|
---|
| 395 |
|
---|
| 396 | Using escape sequences for code points greater than 127 is fine in small doses,
|
---|
| 397 | but becomes an annoyance if you're using many accented characters, as you would
|
---|
| 398 | in a program with messages in French or some other accent-using language. You
|
---|
| 399 | can also assemble strings using the :func:`unichr` built-in function, but this is
|
---|
| 400 | even more tedious.
|
---|
| 401 |
|
---|
| 402 | Ideally, you'd want to be able to write literals in your language's natural
|
---|
| 403 | encoding. You could then edit Python source code with your favorite editor
|
---|
| 404 | which would display the accented characters naturally, and have the right
|
---|
| 405 | characters used at runtime.
|
---|
| 406 |
|
---|
| 407 | Python supports writing Unicode literals in any encoding, but you have to
|
---|
| 408 | declare the encoding being used. This is done by including a special comment as
|
---|
| 409 | either the first or second line of the source file::
|
---|
| 410 |
|
---|
| 411 | #!/usr/bin/env python
|
---|
| 412 | # -*- coding: latin-1 -*-
|
---|
| 413 |
|
---|
| 414 | u = u'abcdé'
|
---|
| 415 | print ord(u[-1])
|
---|
| 416 |
|
---|
| 417 | The syntax is inspired by Emacs's notation for specifying variables local to a
|
---|
| 418 | file. Emacs supports many different variables, but Python only supports
|
---|
| 419 | 'coding'. The ``-*-`` symbols indicate to Emacs that the comment is special;
|
---|
| 420 | they have no significance to Python but are a convention. Python looks for
|
---|
| 421 | ``coding: name`` or ``coding=name`` in the comment.
|
---|
| 422 |
|
---|
| 423 | If you don't include such a comment, the default encoding used will be ASCII.
|
---|
| 424 | Versions of Python before 2.4 were Euro-centric and assumed Latin-1 as a default
|
---|
| 425 | encoding for string literals; in Python 2.4, characters greater than 127 still
|
---|
| 426 | work but result in a warning. For example, the following program has no
|
---|
| 427 | encoding declaration::
|
---|
| 428 |
|
---|
| 429 | #!/usr/bin/env python
|
---|
| 430 | u = u'abcdé'
|
---|
| 431 | print ord(u[-1])
|
---|
| 432 |
|
---|
| 433 | When you run it with Python 2.4, it will output the following warning::
|
---|
| 434 |
|
---|
[391] | 435 | amk:~$ python2.4 p263.py
|
---|
[2] | 436 | sys:1: DeprecationWarning: Non-ASCII character '\xe9'
|
---|
| 437 | in file p263.py on line 2, but no encoding declared;
|
---|
| 438 | see http://www.python.org/peps/pep-0263.html for details
|
---|
| 439 |
|
---|
[391] | 440 | Python 2.5 and higher are stricter and will produce a syntax error::
|
---|
[2] | 441 |
|
---|
[391] | 442 | amk:~$ python2.5 p263.py
|
---|
| 443 | File "/tmp/p263.py", line 2
|
---|
| 444 | SyntaxError: Non-ASCII character '\xc3' in file /tmp/p263.py
|
---|
| 445 | on line 2, but no encoding declared; see
|
---|
| 446 | http://www.python.org/peps/pep-0263.html for details
|
---|
| 447 |
|
---|
| 448 |
|
---|
[2] | 449 | Unicode Properties
|
---|
| 450 | ------------------
|
---|
| 451 |
|
---|
| 452 | The Unicode specification includes a database of information about code points.
|
---|
| 453 | For each code point that's defined, the information includes the character's
|
---|
| 454 | name, its category, the numeric value if applicable (Unicode has characters
|
---|
| 455 | representing the Roman numerals and fractions such as one-third and
|
---|
| 456 | four-fifths). There are also properties related to the code point's use in
|
---|
| 457 | bidirectional text and other display-related properties.
|
---|
| 458 |
|
---|
| 459 | The following program displays some information about several characters, and
|
---|
| 460 | prints the numeric value of one particular character::
|
---|
| 461 |
|
---|
| 462 | import unicodedata
|
---|
| 463 |
|
---|
| 464 | u = unichr(233) + unichr(0x0bf2) + unichr(3972) + unichr(6000) + unichr(13231)
|
---|
| 465 |
|
---|
| 466 | for i, c in enumerate(u):
|
---|
| 467 | print i, '%04x' % ord(c), unicodedata.category(c),
|
---|
| 468 | print unicodedata.name(c)
|
---|
| 469 |
|
---|
| 470 | # Get numeric value of second character
|
---|
| 471 | print unicodedata.numeric(u[1])
|
---|
| 472 |
|
---|
| 473 | When run, this prints::
|
---|
| 474 |
|
---|
| 475 | 0 00e9 Ll LATIN SMALL LETTER E WITH ACUTE
|
---|
| 476 | 1 0bf2 No TAMIL NUMBER ONE THOUSAND
|
---|
| 477 | 2 0f84 Mn TIBETAN MARK HALANTA
|
---|
| 478 | 3 1770 Lo TAGBANWA LETTER SA
|
---|
| 479 | 4 33af So SQUARE RAD OVER S SQUARED
|
---|
| 480 | 1000.0
|
---|
| 481 |
|
---|
| 482 | The category codes are abbreviations describing the nature of the character.
|
---|
| 483 | These are grouped into categories such as "Letter", "Number", "Punctuation", or
|
---|
| 484 | "Symbol", which in turn are broken up into subcategories. To take the codes
|
---|
| 485 | from the above output, ``'Ll'`` means 'Letter, lowercase', ``'No'`` means
|
---|
| 486 | "Number, other", ``'Mn'`` is "Mark, nonspacing", and ``'So'`` is "Symbol,
|
---|
| 487 | other". See
|
---|
[391] | 488 | <http://www.unicode.org/reports/tr44/#General_Category_Values> for a
|
---|
[2] | 489 | list of category codes.
|
---|
| 490 |
|
---|
| 491 | References
|
---|
| 492 | ----------
|
---|
| 493 |
|
---|
| 494 | The Unicode and 8-bit string types are described in the Python library reference
|
---|
| 495 | at :ref:`typesseq`.
|
---|
| 496 |
|
---|
| 497 | The documentation for the :mod:`unicodedata` module.
|
---|
| 498 |
|
---|
| 499 | The documentation for the :mod:`codecs` module.
|
---|
| 500 |
|
---|
| 501 | Marc-André Lemburg gave a presentation at EuroPython 2002 titled "Python and
|
---|
| 502 | Unicode". A PDF version of his slides is available at
|
---|
| 503 | <http://downloads.egenix.com/python/Unicode-EPC2002-Talk.pdf>, and is an
|
---|
| 504 | excellent overview of the design of Python's Unicode features.
|
---|
| 505 |
|
---|
| 506 |
|
---|
| 507 | Reading and Writing Unicode Data
|
---|
| 508 | ================================
|
---|
| 509 |
|
---|
| 510 | Once you've written some code that works with Unicode data, the next problem is
|
---|
| 511 | input/output. How do you get Unicode strings into your program, and how do you
|
---|
| 512 | convert Unicode into a form suitable for storage or transmission?
|
---|
| 513 |
|
---|
| 514 | It's possible that you may not need to do anything depending on your input
|
---|
| 515 | sources and output destinations; you should check whether the libraries used in
|
---|
| 516 | your application support Unicode natively. XML parsers often return Unicode
|
---|
| 517 | data, for example. Many relational databases also support Unicode-valued
|
---|
| 518 | columns and can return Unicode values from an SQL query.
|
---|
| 519 |
|
---|
| 520 | Unicode data is usually converted to a particular encoding before it gets
|
---|
| 521 | written to disk or sent over a socket. It's possible to do all the work
|
---|
| 522 | yourself: open a file, read an 8-bit string from it, and convert the string with
|
---|
| 523 | ``unicode(str, encoding)``. However, the manual approach is not recommended.
|
---|
| 524 |
|
---|
| 525 | One problem is the multi-byte nature of encodings; one Unicode character can be
|
---|
| 526 | represented by several bytes. If you want to read the file in arbitrary-sized
|
---|
| 527 | chunks (say, 1K or 4K), you need to write error-handling code to catch the case
|
---|
| 528 | where only part of the bytes encoding a single Unicode character are read at the
|
---|
| 529 | end of a chunk. One solution would be to read the entire file into memory and
|
---|
| 530 | then perform the decoding, but that prevents you from working with files that
|
---|
| 531 | are extremely large; if you need to read a 2Gb file, you need 2Gb of RAM.
|
---|
| 532 | (More, really, since for at least a moment you'd need to have both the encoded
|
---|
| 533 | string and its Unicode version in memory.)
|
---|
| 534 |
|
---|
| 535 | The solution would be to use the low-level decoding interface to catch the case
|
---|
| 536 | of partial coding sequences. The work of implementing this has already been
|
---|
| 537 | done for you: the :mod:`codecs` module includes a version of the :func:`open`
|
---|
| 538 | function that returns a file-like object that assumes the file's contents are in
|
---|
| 539 | a specified encoding and accepts Unicode parameters for methods such as
|
---|
| 540 | ``.read()`` and ``.write()``.
|
---|
| 541 |
|
---|
| 542 | The function's parameters are ``open(filename, mode='rb', encoding=None,
|
---|
| 543 | errors='strict', buffering=1)``. ``mode`` can be ``'r'``, ``'w'``, or ``'a'``,
|
---|
| 544 | just like the corresponding parameter to the regular built-in ``open()``
|
---|
| 545 | function; add a ``'+'`` to update the file. ``buffering`` is similarly parallel
|
---|
| 546 | to the standard function's parameter. ``encoding`` is a string giving the
|
---|
| 547 | encoding to use; if it's left as ``None``, a regular Python file object that
|
---|
| 548 | accepts 8-bit strings is returned. Otherwise, a wrapper object is returned, and
|
---|
| 549 | data written to or read from the wrapper object will be converted as needed.
|
---|
| 550 | ``errors`` specifies the action for encoding errors and can be one of the usual
|
---|
| 551 | values of 'strict', 'ignore', and 'replace'.
|
---|
| 552 |
|
---|
| 553 | Reading Unicode from a file is therefore simple::
|
---|
| 554 |
|
---|
| 555 | import codecs
|
---|
| 556 | f = codecs.open('unicode.rst', encoding='utf-8')
|
---|
| 557 | for line in f:
|
---|
| 558 | print repr(line)
|
---|
| 559 |
|
---|
| 560 | It's also possible to open files in update mode, allowing both reading and
|
---|
| 561 | writing::
|
---|
| 562 |
|
---|
| 563 | f = codecs.open('test', encoding='utf-8', mode='w+')
|
---|
| 564 | f.write(u'\u4500 blah blah blah\n')
|
---|
| 565 | f.seek(0)
|
---|
| 566 | print repr(f.readline()[:1])
|
---|
| 567 | f.close()
|
---|
| 568 |
|
---|
| 569 | Unicode character U+FEFF is used as a byte-order mark (BOM), and is often
|
---|
| 570 | written as the first character of a file in order to assist with autodetection
|
---|
| 571 | of the file's byte ordering. Some encodings, such as UTF-16, expect a BOM to be
|
---|
| 572 | present at the start of a file; when such an encoding is used, the BOM will be
|
---|
| 573 | automatically written as the first character and will be silently dropped when
|
---|
| 574 | the file is read. There are variants of these encodings, such as 'utf-16-le'
|
---|
| 575 | and 'utf-16-be' for little-endian and big-endian encodings, that specify one
|
---|
| 576 | particular byte ordering and don't skip the BOM.
|
---|
| 577 |
|
---|
| 578 |
|
---|
| 579 | Unicode filenames
|
---|
| 580 | -----------------
|
---|
| 581 |
|
---|
| 582 | Most of the operating systems in common use today support filenames that contain
|
---|
| 583 | arbitrary Unicode characters. Usually this is implemented by converting the
|
---|
| 584 | Unicode string into some encoding that varies depending on the system. For
|
---|
| 585 | example, Mac OS X uses UTF-8 while Windows uses a configurable encoding; on
|
---|
| 586 | Windows, Python uses the name "mbcs" to refer to whatever the currently
|
---|
| 587 | configured encoding is. On Unix systems, there will only be a filesystem
|
---|
| 588 | encoding if you've set the ``LANG`` or ``LC_CTYPE`` environment variables; if
|
---|
| 589 | you haven't, the default encoding is ASCII.
|
---|
| 590 |
|
---|
| 591 | The :func:`sys.getfilesystemencoding` function returns the encoding to use on
|
---|
| 592 | your current system, in case you want to do the encoding manually, but there's
|
---|
| 593 | not much reason to bother. When opening a file for reading or writing, you can
|
---|
| 594 | usually just provide the Unicode string as the filename, and it will be
|
---|
| 595 | automatically converted to the right encoding for you::
|
---|
| 596 |
|
---|
| 597 | filename = u'filename\u4500abc'
|
---|
| 598 | f = open(filename, 'w')
|
---|
| 599 | f.write('blah\n')
|
---|
| 600 | f.close()
|
---|
| 601 |
|
---|
| 602 | Functions in the :mod:`os` module such as :func:`os.stat` will also accept Unicode
|
---|
| 603 | filenames.
|
---|
| 604 |
|
---|
| 605 | :func:`os.listdir`, which returns filenames, raises an issue: should it return
|
---|
| 606 | the Unicode version of filenames, or should it return 8-bit strings containing
|
---|
| 607 | the encoded versions? :func:`os.listdir` will do both, depending on whether you
|
---|
| 608 | provided the directory path as an 8-bit string or a Unicode string. If you pass
|
---|
| 609 | a Unicode string as the path, filenames will be decoded using the filesystem's
|
---|
| 610 | encoding and a list of Unicode strings will be returned, while passing an 8-bit
|
---|
| 611 | path will return the 8-bit versions of the filenames. For example, assuming the
|
---|
| 612 | default filesystem encoding is UTF-8, running the following program::
|
---|
| 613 |
|
---|
| 614 | fn = u'filename\u4500abc'
|
---|
| 615 | f = open(fn, 'w')
|
---|
| 616 | f.close()
|
---|
| 617 |
|
---|
| 618 | import os
|
---|
| 619 | print os.listdir('.')
|
---|
| 620 | print os.listdir(u'.')
|
---|
| 621 |
|
---|
| 622 | will produce the following output::
|
---|
| 623 |
|
---|
| 624 | amk:~$ python t.py
|
---|
| 625 | ['.svn', 'filename\xe4\x94\x80abc', ...]
|
---|
| 626 | [u'.svn', u'filename\u4500abc', ...]
|
---|
| 627 |
|
---|
| 628 | The first list contains UTF-8-encoded filenames, and the second list contains
|
---|
| 629 | the Unicode versions.
|
---|
| 630 |
|
---|
| 631 |
|
---|
| 632 |
|
---|
| 633 | Tips for Writing Unicode-aware Programs
|
---|
| 634 | ---------------------------------------
|
---|
| 635 |
|
---|
| 636 | This section provides some suggestions on writing software that deals with
|
---|
| 637 | Unicode.
|
---|
| 638 |
|
---|
| 639 | The most important tip is:
|
---|
| 640 |
|
---|
| 641 | Software should only work with Unicode strings internally, converting to a
|
---|
| 642 | particular encoding on output.
|
---|
| 643 |
|
---|
| 644 | If you attempt to write processing functions that accept both Unicode and 8-bit
|
---|
| 645 | strings, you will find your program vulnerable to bugs wherever you combine the
|
---|
| 646 | two different kinds of strings. Python's default encoding is ASCII, so whenever
|
---|
| 647 | a character with an ASCII value > 127 is in the input data, you'll get a
|
---|
| 648 | :exc:`UnicodeDecodeError` because that character can't be handled by the ASCII
|
---|
| 649 | encoding.
|
---|
| 650 |
|
---|
| 651 | It's easy to miss such problems if you only test your software with data that
|
---|
| 652 | doesn't contain any accents; everything will seem to work, but there's actually
|
---|
| 653 | a bug in your program waiting for the first user who attempts to use characters
|
---|
| 654 | > 127. A second tip, therefore, is:
|
---|
| 655 |
|
---|
| 656 | Include characters > 127 and, even better, characters > 255 in your test
|
---|
| 657 | data.
|
---|
| 658 |
|
---|
| 659 | When using data coming from a web browser or some other untrusted source, a
|
---|
| 660 | common technique is to check for illegal characters in a string before using the
|
---|
| 661 | string in a generated command line or storing it in a database. If you're doing
|
---|
| 662 | this, be careful to check the string once it's in the form that will be used or
|
---|
| 663 | stored; it's possible for encodings to be used to disguise characters. This is
|
---|
| 664 | especially true if the input data also specifies the encoding; many encodings
|
---|
| 665 | leave the commonly checked-for characters alone, but Python includes some
|
---|
| 666 | encodings such as ``'base64'`` that modify every single character.
|
---|
| 667 |
|
---|
| 668 | For example, let's say you have a content management system that takes a Unicode
|
---|
| 669 | filename, and you want to disallow paths with a '/' character. You might write
|
---|
| 670 | this code::
|
---|
| 671 |
|
---|
| 672 | def read_file (filename, encoding):
|
---|
| 673 | if '/' in filename:
|
---|
| 674 | raise ValueError("'/' not allowed in filenames")
|
---|
| 675 | unicode_name = filename.decode(encoding)
|
---|
| 676 | f = open(unicode_name, 'r')
|
---|
| 677 | # ... return contents of file ...
|
---|
| 678 |
|
---|
| 679 | However, if an attacker could specify the ``'base64'`` encoding, they could pass
|
---|
| 680 | ``'L2V0Yy9wYXNzd2Q='``, which is the base-64 encoded form of the string
|
---|
| 681 | ``'/etc/passwd'``, to read a system file. The above code looks for ``'/'``
|
---|
| 682 | characters in the encoded form and misses the dangerous character in the
|
---|
| 683 | resulting decoded form.
|
---|
| 684 |
|
---|
| 685 | References
|
---|
| 686 | ----------
|
---|
| 687 |
|
---|
| 688 | The PDF slides for Marc-André Lemburg's presentation "Writing Unicode-aware
|
---|
| 689 | Applications in Python" are available at
|
---|
| 690 | <http://downloads.egenix.com/python/LSM2005-Developing-Unicode-aware-applications-in-Python.pdf>
|
---|
| 691 | and discuss questions of character encodings as well as how to internationalize
|
---|
| 692 | and localize an application.
|
---|
| 693 |
|
---|
| 694 |
|
---|
| 695 | Revision History and Acknowledgements
|
---|
| 696 | =====================================
|
---|
| 697 |
|
---|
| 698 | Thanks to the following people who have noted errors or offered suggestions on
|
---|
| 699 | this article: Nicholas Bastin, Marius Gedminas, Kent Johnson, Ken Krugler,
|
---|
| 700 | Marc-André Lemburg, Martin von Löwis, Chad Whitacre.
|
---|
| 701 |
|
---|
| 702 | Version 1.0: posted August 5 2005.
|
---|
| 703 |
|
---|
| 704 | Version 1.01: posted August 7 2005. Corrects factual and markup errors; adds
|
---|
| 705 | several links.
|
---|
| 706 |
|
---|
| 707 | Version 1.02: posted August 16 2005. Corrects factual errors.
|
---|
| 708 |
|
---|
[391] | 709 | Version 1.03: posted June 20 2010. Notes that Python 3.x is not covered,
|
---|
| 710 | and that the HOWTO only covers 2.x.
|
---|
[2] | 711 |
|
---|
[391] | 712 |
|
---|
| 713 | .. comment Describe Python 3.x support (new section? new document?)
|
---|
[2] | 714 | .. comment Additional topic: building Python w/ UCS2 or UCS4 support
|
---|
| 715 | .. comment Describe obscure -U switch somewhere?
|
---|
| 716 | .. comment Describe use of codecs.StreamRecoder and StreamReaderWriter
|
---|
| 717 |
|
---|
| 718 | .. comment
|
---|
| 719 | Original outline:
|
---|
| 720 |
|
---|
| 721 | - [ ] Unicode introduction
|
---|
| 722 | - [ ] ASCII
|
---|
| 723 | - [ ] Terms
|
---|
| 724 | - [ ] Character
|
---|
| 725 | - [ ] Code point
|
---|
| 726 | - [ ] Encodings
|
---|
| 727 | - [ ] Common encodings: ASCII, Latin-1, UTF-8
|
---|
| 728 | - [ ] Unicode Python type
|
---|
| 729 | - [ ] Writing unicode literals
|
---|
| 730 | - [ ] Obscurity: -U switch
|
---|
| 731 | - [ ] Built-ins
|
---|
| 732 | - [ ] unichr()
|
---|
| 733 | - [ ] ord()
|
---|
| 734 | - [ ] unicode() constructor
|
---|
| 735 | - [ ] Unicode type
|
---|
| 736 | - [ ] encode(), decode() methods
|
---|
| 737 | - [ ] Unicodedata module for character properties
|
---|
| 738 | - [ ] I/O
|
---|
| 739 | - [ ] Reading/writing Unicode data into files
|
---|
| 740 | - [ ] Byte-order marks
|
---|
| 741 | - [ ] Unicode filenames
|
---|
| 742 | - [ ] Writing Unicode programs
|
---|
| 743 | - [ ] Do everything in Unicode
|
---|
| 744 | - [ ] Declaring source code encodings (PEP 263)
|
---|
| 745 | - [ ] Other issues
|
---|
| 746 | - [ ] Building Python (UCS2, UCS4)
|
---|