| /* |
| |
| Unicode implementation based on original code by Fredrik Lundh, |
| modified by Marc-Andre Lemburg <mal@lemburg.com> according to the |
| Unicode Integration Proposal (see file Misc/unicode.txt). |
| |
| Major speed upgrades to the method implementations at the Reykjavik |
| NeedForSpeed sprint, by Fredrik Lundh and Andrew Dalke. |
| |
| Copyright (c) Corporation for National Research Initiatives. |
| |
| -------------------------------------------------------------------- |
| The original string type implementation is: |
| |
| Copyright (c) 1999 by Secret Labs AB |
| Copyright (c) 1999 by Fredrik Lundh |
| |
| By obtaining, using, and/or copying this software and/or its |
| associated documentation, you agree that you have read, understood, |
| and will comply with the following terms and conditions: |
| |
| Permission to use, copy, modify, and distribute this software and its |
| associated documentation for any purpose and without fee is hereby |
| granted, provided that the above copyright notice appears in all |
| copies, and that both that copyright notice and this permission notice |
| appear in supporting documentation, and that the name of Secret Labs |
| AB or the author not be used in advertising or publicity pertaining to |
| distribution of the software without specific, written prior |
| permission. |
| |
| SECRET LABS AB AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH REGARD TO |
| THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND |
| FITNESS. IN NO EVENT SHALL SECRET LABS AB OR THE AUTHOR BE LIABLE FOR |
| ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
| WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN |
| ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT |
| OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
| -------------------------------------------------------------------- |
| |
| */ |
| |
| #define PY_SSIZE_T_CLEAN |
| #include "Python.h" |
| #include "ucnhash.h" |
| |
| #ifdef MS_WINDOWS |
| #include <windows.h> |
| #endif |
| |
| /* Limit for the Unicode object free list */ |
| |
| #define PyUnicode_MAXFREELIST 1024 |
| |
| /* Limit for the Unicode object free list stay alive optimization. |
| |
| The implementation will keep allocated Unicode memory intact for |
| all objects on the free list having a size less than this |
| limit. This reduces malloc() overhead for small Unicode objects. |
| |
| At worst this will result in PyUnicode_MAXFREELIST * |
| (sizeof(PyUnicodeObject) + KEEPALIVE_SIZE_LIMIT + |
| malloc()-overhead) bytes of unused garbage. |
| |
| Setting the limit to 0 effectively turns the feature off. |
| |
| Note: This is an experimental feature ! If you get core dumps when |
| using Unicode objects, turn this feature off. |
| |
| */ |
| |
| #define KEEPALIVE_SIZE_LIMIT 9 |
| |
| /* Endianness switches; defaults to little endian */ |
| |
| #ifdef WORDS_BIGENDIAN |
| # define BYTEORDER_IS_BIG_ENDIAN |
| #else |
| # define BYTEORDER_IS_LITTLE_ENDIAN |
| #endif |
| |
| /* --- Globals ------------------------------------------------------------ |
| |
| NOTE: In the interpreter's initialization phase, some globals are currently |
| initialized dynamically as needed. In the process Unicode objects may |
| be created before the Unicode type is ready. |
| |
| */ |
| |
| |
| #ifdef __cplusplus |
| extern "C" { |
| #endif |
| |
| /* This dictionary holds all interned unicode strings. Note that references |
| to strings in this dictionary are *not* counted in the string's ob_refcnt. |
| When the interned string reaches a refcnt of 0 the string deallocation |
| function will delete the reference from this dictionary. |
| |
| Another way to look at this is that to say that the actual reference |
| count of a string is: s->ob_refcnt + (s->state ? 2 : 0) |
| */ |
| static PyObject *interned = NULL; |
| |
| /* Free list for Unicode objects */ |
| static PyUnicodeObject *free_list = NULL; |
| static int numfree = 0; |
| |
| /* The empty Unicode object is shared to improve performance. */ |
| static PyUnicodeObject *unicode_empty = NULL; |
| |
| #define _Py_RETURN_UNICODE_EMPTY() \ |
| do { \ |
| if (unicode_empty != NULL) \ |
| Py_INCREF(unicode_empty); \ |
| else { \ |
| unicode_empty = _PyUnicode_New(0); \ |
| if (unicode_empty != NULL) \ |
| Py_INCREF(unicode_empty); \ |
| } \ |
| return (PyObject *)unicode_empty; \ |
| } while (0) |
| |
| /* Single character Unicode strings in the Latin-1 range are being |
| shared as well. */ |
| static PyUnicodeObject *unicode_latin1[256] = {NULL}; |
| |
| /* Fast detection of the most frequent whitespace characters */ |
| const unsigned char _Py_ascii_whitespace[] = { |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| /* case 0x0009: * CHARACTER TABULATION */ |
| /* case 0x000A: * LINE FEED */ |
| /* case 0x000B: * LINE TABULATION */ |
| /* case 0x000C: * FORM FEED */ |
| /* case 0x000D: * CARRIAGE RETURN */ |
| 0, 1, 1, 1, 1, 1, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| /* case 0x001C: * FILE SEPARATOR */ |
| /* case 0x001D: * GROUP SEPARATOR */ |
| /* case 0x001E: * RECORD SEPARATOR */ |
| /* case 0x001F: * UNIT SEPARATOR */ |
| 0, 0, 0, 0, 1, 1, 1, 1, |
| /* case 0x0020: * SPACE */ |
| 1, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0 |
| }; |
| |
| static PyObject *unicode_encode_call_errorhandler(const char *errors, |
| PyObject **errorHandler,const char *encoding, const char *reason, |
| const Py_UNICODE *unicode, Py_ssize_t size, PyObject **exceptionObject, |
| Py_ssize_t startpos, Py_ssize_t endpos, Py_ssize_t *newpos); |
| |
| static void raise_encode_exception(PyObject **exceptionObject, |
| const char *encoding, |
| const Py_UNICODE *unicode, Py_ssize_t size, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason); |
| |
| /* Same for linebreaks */ |
| static unsigned char ascii_linebreak[] = { |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| /* 0x000A, * LINE FEED */ |
| /* 0x000B, * LINE TABULATION */ |
| /* 0x000C, * FORM FEED */ |
| /* 0x000D, * CARRIAGE RETURN */ |
| 0, 0, 1, 1, 1, 1, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| /* 0x001C, * FILE SEPARATOR */ |
| /* 0x001D, * GROUP SEPARATOR */ |
| /* 0x001E, * RECORD SEPARATOR */ |
| 0, 0, 0, 0, 1, 1, 1, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0 |
| }; |
| |
| |
| Py_UNICODE |
| PyUnicode_GetMax(void) |
| { |
| #ifdef Py_UNICODE_WIDE |
| return 0x10FFFF; |
| #else |
| /* This is actually an illegal character, so it should |
| not be passed to unichr. */ |
| return 0xFFFF; |
| #endif |
| } |
| |
| /* --- Bloom Filters ----------------------------------------------------- */ |
| |
| /* stuff to implement simple "bloom filters" for Unicode characters. |
| to keep things simple, we use a single bitmask, using the least 5 |
| bits from each unicode characters as the bit index. */ |
| |
| /* the linebreak mask is set up by Unicode_Init below */ |
| |
| #if LONG_BIT >= 128 |
| #define BLOOM_WIDTH 128 |
| #elif LONG_BIT >= 64 |
| #define BLOOM_WIDTH 64 |
| #elif LONG_BIT >= 32 |
| #define BLOOM_WIDTH 32 |
| #else |
| #error "LONG_BIT is smaller than 32" |
| #endif |
| |
| #define BLOOM_MASK unsigned long |
| |
| static BLOOM_MASK bloom_linebreak = ~(BLOOM_MASK)0; |
| |
| #define BLOOM_ADD(mask, ch) ((mask |= (1UL << ((ch) & (BLOOM_WIDTH - 1))))) |
| #define BLOOM(mask, ch) ((mask & (1UL << ((ch) & (BLOOM_WIDTH - 1))))) |
| |
| #define BLOOM_LINEBREAK(ch) \ |
| ((ch) < 128U ? ascii_linebreak[(ch)] : \ |
| (BLOOM(bloom_linebreak, (ch)) && Py_UNICODE_ISLINEBREAK(ch))) |
| |
| Py_LOCAL_INLINE(BLOOM_MASK) make_bloom_mask(Py_UNICODE* ptr, Py_ssize_t len) |
| { |
| /* calculate simple bloom-style bitmask for a given unicode string */ |
| |
| BLOOM_MASK mask; |
| Py_ssize_t i; |
| |
| mask = 0; |
| for (i = 0; i < len; i++) |
| BLOOM_ADD(mask, ptr[i]); |
| |
| return mask; |
| } |
| |
| Py_LOCAL_INLINE(int) unicode_member(Py_UNICODE chr, Py_UNICODE* set, Py_ssize_t setlen) |
| { |
| Py_ssize_t i; |
| |
| for (i = 0; i < setlen; i++) |
| if (set[i] == chr) |
| return 1; |
| |
| return 0; |
| } |
| |
| #define BLOOM_MEMBER(mask, chr, set, setlen) \ |
| BLOOM(mask, chr) && unicode_member(chr, set, setlen) |
| |
| /* --- Unicode Object ----------------------------------------------------- */ |
| |
| static |
| int unicode_resize(register PyUnicodeObject *unicode, |
| Py_ssize_t length) |
| { |
| void *oldstr; |
| |
| /* Shortcut if there's nothing much to do. */ |
| if (unicode->length == length) |
| goto reset; |
| |
| /* Resizing shared object (unicode_empty or single character |
| objects) in-place is not allowed. Use PyUnicode_Resize() |
| instead ! */ |
| |
| if (unicode == unicode_empty || |
| (unicode->length == 1 && |
| unicode->str[0] < 256U && |
| unicode_latin1[unicode->str[0]] == unicode)) { |
| PyErr_SetString(PyExc_SystemError, |
| "can't resize shared str objects"); |
| return -1; |
| } |
| |
| /* We allocate one more byte to make sure the string is Ux0000 terminated. |
| The overallocation is also used by fastsearch, which assumes that it's |
| safe to look at str[length] (without making any assumptions about what |
| it contains). */ |
| |
| oldstr = unicode->str; |
| unicode->str = PyObject_REALLOC(unicode->str, |
| sizeof(Py_UNICODE) * (length + 1)); |
| if (!unicode->str) { |
| unicode->str = (Py_UNICODE *)oldstr; |
| PyErr_NoMemory(); |
| return -1; |
| } |
| unicode->str[length] = 0; |
| unicode->length = length; |
| |
| reset: |
| /* Reset the object caches */ |
| if (unicode->defenc) { |
| Py_CLEAR(unicode->defenc); |
| } |
| unicode->hash = -1; |
| |
| return 0; |
| } |
| |
| /* We allocate one more byte to make sure the string is |
| Ux0000 terminated; some code (e.g. new_identifier) |
| relies on that. |
| |
| XXX This allocator could further be enhanced by assuring that the |
| free list never reduces its size below 1. |
| |
| */ |
| |
| static |
| PyUnicodeObject *_PyUnicode_New(Py_ssize_t length) |
| { |
| register PyUnicodeObject *unicode; |
| |
| /* Optimization for empty strings */ |
| if (length == 0 && unicode_empty != NULL) { |
| Py_INCREF(unicode_empty); |
| return unicode_empty; |
| } |
| |
| /* Ensure we won't overflow the size. */ |
| if (length > ((PY_SSIZE_T_MAX / sizeof(Py_UNICODE)) - 1)) { |
| return (PyUnicodeObject *)PyErr_NoMemory(); |
| } |
| |
| /* Unicode freelist & memory allocation */ |
| if (free_list) { |
| unicode = free_list; |
| free_list = *(PyUnicodeObject **)unicode; |
| numfree--; |
| if (unicode->str) { |
| /* Keep-Alive optimization: we only upsize the buffer, |
| never downsize it. */ |
| if ((unicode->length < length) && |
| unicode_resize(unicode, length) < 0) { |
| PyObject_DEL(unicode->str); |
| unicode->str = NULL; |
| } |
| } |
| else { |
| size_t new_size = sizeof(Py_UNICODE) * ((size_t)length + 1); |
| unicode->str = (Py_UNICODE*) PyObject_MALLOC(new_size); |
| } |
| PyObject_INIT(unicode, &PyUnicode_Type); |
| } |
| else { |
| size_t new_size; |
| unicode = PyObject_New(PyUnicodeObject, &PyUnicode_Type); |
| if (unicode == NULL) |
| return NULL; |
| new_size = sizeof(Py_UNICODE) * ((size_t)length + 1); |
| unicode->str = (Py_UNICODE*) PyObject_MALLOC(new_size); |
| } |
| |
| if (!unicode->str) { |
| PyErr_NoMemory(); |
| goto onError; |
| } |
| /* Initialize the first element to guard against cases where |
| * the caller fails before initializing str -- unicode_resize() |
| * reads str[0], and the Keep-Alive optimization can keep memory |
| * allocated for str alive across a call to unicode_dealloc(unicode). |
| * We don't want unicode_resize to read uninitialized memory in |
| * that case. |
| */ |
| unicode->str[0] = 0; |
| unicode->str[length] = 0; |
| unicode->length = length; |
| unicode->hash = -1; |
| unicode->state = 0; |
| unicode->defenc = NULL; |
| return unicode; |
| |
| onError: |
| /* XXX UNREF/NEWREF interface should be more symmetrical */ |
| _Py_DEC_REFTOTAL; |
| _Py_ForgetReference((PyObject *)unicode); |
| PyObject_Del(unicode); |
| return NULL; |
| } |
| |
| static |
| void unicode_dealloc(register PyUnicodeObject *unicode) |
| { |
| switch (PyUnicode_CHECK_INTERNED(unicode)) { |
| case SSTATE_NOT_INTERNED: |
| break; |
| |
| case SSTATE_INTERNED_MORTAL: |
| /* revive dead object temporarily for DelItem */ |
| Py_REFCNT(unicode) = 3; |
| if (PyDict_DelItem(interned, (PyObject *)unicode) != 0) |
| Py_FatalError( |
| "deletion of interned string failed"); |
| break; |
| |
| case SSTATE_INTERNED_IMMORTAL: |
| Py_FatalError("Immortal interned string died."); |
| |
| default: |
| Py_FatalError("Inconsistent interned string state."); |
| } |
| |
| if (PyUnicode_CheckExact(unicode) && |
| numfree < PyUnicode_MAXFREELIST) { |
| /* Keep-Alive optimization */ |
| if (unicode->length >= KEEPALIVE_SIZE_LIMIT) { |
| PyObject_DEL(unicode->str); |
| unicode->str = NULL; |
| unicode->length = 0; |
| } |
| if (unicode->defenc) { |
| Py_CLEAR(unicode->defenc); |
| } |
| /* Add to free list */ |
| *(PyUnicodeObject **)unicode = free_list; |
| free_list = unicode; |
| numfree++; |
| } |
| else { |
| PyObject_DEL(unicode->str); |
| Py_XDECREF(unicode->defenc); |
| Py_TYPE(unicode)->tp_free((PyObject *)unicode); |
| } |
| } |
| |
| static |
| int _PyUnicode_Resize(PyUnicodeObject **unicode, Py_ssize_t length) |
| { |
| register PyUnicodeObject *v; |
| |
| /* Argument checks */ |
| if (unicode == NULL) { |
| PyErr_BadInternalCall(); |
| return -1; |
| } |
| v = *unicode; |
| if (v == NULL || !PyUnicode_Check(v) || Py_REFCNT(v) != 1 || length < 0) { |
| PyErr_BadInternalCall(); |
| return -1; |
| } |
| |
| /* Resizing unicode_empty and single character objects is not |
| possible since these are being shared. We simply return a fresh |
| copy with the same Unicode content. */ |
| if (v->length != length && |
| (v == unicode_empty || v->length == 1)) { |
| PyUnicodeObject *w = _PyUnicode_New(length); |
| if (w == NULL) |
| return -1; |
| Py_UNICODE_COPY(w->str, v->str, |
| length < v->length ? length : v->length); |
| Py_DECREF(*unicode); |
| *unicode = w; |
| return 0; |
| } |
| |
| /* Note that we don't have to modify *unicode for unshared Unicode |
| objects, since we can modify them in-place. */ |
| return unicode_resize(v, length); |
| } |
| |
| int PyUnicode_Resize(PyObject **unicode, Py_ssize_t length) |
| { |
| return _PyUnicode_Resize((PyUnicodeObject **)unicode, length); |
| } |
| |
| PyObject *PyUnicode_FromUnicode(const Py_UNICODE *u, |
| Py_ssize_t size) |
| { |
| PyUnicodeObject *unicode; |
| |
| /* If the Unicode data is known at construction time, we can apply |
| some optimizations which share commonly used objects. */ |
| if (u != NULL) { |
| |
| /* Optimization for empty strings */ |
| if (size == 0) |
| _Py_RETURN_UNICODE_EMPTY(); |
| |
| /* Single character Unicode objects in the Latin-1 range are |
| shared when using this constructor */ |
| if (size == 1 && *u < 256) { |
| unicode = unicode_latin1[*u]; |
| if (!unicode) { |
| unicode = _PyUnicode_New(1); |
| if (!unicode) |
| return NULL; |
| unicode->str[0] = *u; |
| unicode_latin1[*u] = unicode; |
| } |
| Py_INCREF(unicode); |
| return (PyObject *)unicode; |
| } |
| } |
| |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| |
| /* Copy the Unicode data into the new object */ |
| if (u != NULL) |
| Py_UNICODE_COPY(unicode->str, u, size); |
| |
| return (PyObject *)unicode; |
| } |
| |
| PyObject *PyUnicode_FromStringAndSize(const char *u, Py_ssize_t size) |
| { |
| PyUnicodeObject *unicode; |
| |
| if (size < 0) { |
| PyErr_SetString(PyExc_SystemError, |
| "Negative size passed to PyUnicode_FromStringAndSize"); |
| return NULL; |
| } |
| |
| /* If the Unicode data is known at construction time, we can apply |
| some optimizations which share commonly used objects. |
| Also, this means the input must be UTF-8, so fall back to the |
| UTF-8 decoder at the end. */ |
| if (u != NULL) { |
| |
| /* Optimization for empty strings */ |
| if (size == 0) |
| _Py_RETURN_UNICODE_EMPTY(); |
| |
| /* Single characters are shared when using this constructor. |
| Restrict to ASCII, since the input must be UTF-8. */ |
| if (size == 1 && Py_CHARMASK(*u) < 128) { |
| unicode = unicode_latin1[Py_CHARMASK(*u)]; |
| if (!unicode) { |
| unicode = _PyUnicode_New(1); |
| if (!unicode) |
| return NULL; |
| unicode->str[0] = Py_CHARMASK(*u); |
| unicode_latin1[Py_CHARMASK(*u)] = unicode; |
| } |
| Py_INCREF(unicode); |
| return (PyObject *)unicode; |
| } |
| |
| return PyUnicode_DecodeUTF8(u, size, NULL); |
| } |
| |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| |
| return (PyObject *)unicode; |
| } |
| |
| PyObject *PyUnicode_FromString(const char *u) |
| { |
| size_t size = strlen(u); |
| if (size > PY_SSIZE_T_MAX) { |
| PyErr_SetString(PyExc_OverflowError, "input too long"); |
| return NULL; |
| } |
| |
| return PyUnicode_FromStringAndSize(u, size); |
| } |
| |
| #ifdef HAVE_WCHAR_H |
| |
| #if (Py_UNICODE_SIZE == 2) && defined(SIZEOF_WCHAR_T) && (SIZEOF_WCHAR_T == 4) |
| # define CONVERT_WCHAR_TO_SURROGATES |
| #endif |
| |
| #ifdef CONVERT_WCHAR_TO_SURROGATES |
| |
| /* Here sizeof(wchar_t) is 4 but Py_UNICODE_SIZE == 2, so we need |
| to convert from UTF32 to UTF16. */ |
| |
| PyObject *PyUnicode_FromWideChar(register const wchar_t *w, |
| Py_ssize_t size) |
| { |
| PyUnicodeObject *unicode; |
| register Py_ssize_t i; |
| Py_ssize_t alloc; |
| const wchar_t *orig_w; |
| |
| if (w == NULL) { |
| if (size == 0) |
| return PyUnicode_FromStringAndSize(NULL, 0); |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| |
| if (size == -1) { |
| size = wcslen(w); |
| } |
| |
| alloc = size; |
| orig_w = w; |
| for (i = size; i > 0; i--) { |
| if (*w > 0xFFFF) |
| alloc++; |
| w++; |
| } |
| w = orig_w; |
| unicode = _PyUnicode_New(alloc); |
| if (!unicode) |
| return NULL; |
| |
| /* Copy the wchar_t data into the new object */ |
| { |
| register Py_UNICODE *u; |
| u = PyUnicode_AS_UNICODE(unicode); |
| for (i = size; i > 0; i--) { |
| if (*w > 0xFFFF) { |
| wchar_t ordinal = *w++; |
| ordinal -= 0x10000; |
| *u++ = 0xD800 | (ordinal >> 10); |
| *u++ = 0xDC00 | (ordinal & 0x3FF); |
| } |
| else |
| *u++ = *w++; |
| } |
| } |
| return (PyObject *)unicode; |
| } |
| |
| #else |
| |
| PyObject *PyUnicode_FromWideChar(register const wchar_t *w, |
| Py_ssize_t size) |
| { |
| PyUnicodeObject *unicode; |
| |
| if (w == NULL) { |
| if (size == 0) |
| return PyUnicode_FromStringAndSize(NULL, 0); |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| |
| if (size == -1) { |
| size = wcslen(w); |
| } |
| |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| |
| /* Copy the wchar_t data into the new object */ |
| #if Py_UNICODE_SIZE == SIZEOF_WCHAR_T |
| memcpy(unicode->str, w, size * sizeof(wchar_t)); |
| #else |
| { |
| register Py_UNICODE *u; |
| register Py_ssize_t i; |
| u = PyUnicode_AS_UNICODE(unicode); |
| for (i = size; i > 0; i--) |
| *u++ = *w++; |
| } |
| #endif |
| |
| return (PyObject *)unicode; |
| } |
| |
| #endif /* CONVERT_WCHAR_TO_SURROGATES */ |
| |
| #undef CONVERT_WCHAR_TO_SURROGATES |
| |
| static void |
| makefmt(char *fmt, int longflag, int longlongflag, int size_tflag, |
| int zeropad, int width, int precision, char c) |
| { |
| *fmt++ = '%'; |
| if (width) { |
| if (zeropad) |
| *fmt++ = '0'; |
| fmt += sprintf(fmt, "%d", width); |
| } |
| if (precision) |
| fmt += sprintf(fmt, ".%d", precision); |
| if (longflag) |
| *fmt++ = 'l'; |
| else if (longlongflag) { |
| /* longlongflag should only ever be nonzero on machines with |
| HAVE_LONG_LONG defined */ |
| #ifdef HAVE_LONG_LONG |
| char *f = PY_FORMAT_LONG_LONG; |
| while (*f) |
| *fmt++ = *f++; |
| #else |
| /* we shouldn't ever get here */ |
| assert(0); |
| *fmt++ = 'l'; |
| #endif |
| } |
| else if (size_tflag) { |
| char *f = PY_FORMAT_SIZE_T; |
| while (*f) |
| *fmt++ = *f++; |
| } |
| *fmt++ = c; |
| *fmt = '\0'; |
| } |
| |
| #define appendstring(string) {for (copy = string;*copy;) *s++ = *copy++;} |
| |
| /* size of fixed-size buffer for formatting single arguments */ |
| #define ITEM_BUFFER_LEN 21 |
| /* maximum number of characters required for output of %ld. 21 characters |
| allows for 64-bit integers (in decimal) and an optional sign. */ |
| #define MAX_LONG_CHARS 21 |
| /* maximum number of characters required for output of %lld. |
| We need at most ceil(log10(256)*SIZEOF_LONG_LONG) digits, |
| plus 1 for the sign. 53/22 is an upper bound for log10(256). */ |
| #define MAX_LONG_LONG_CHARS (2 + (SIZEOF_LONG_LONG*53-1) / 22) |
| |
| PyObject * |
| PyUnicode_FromFormatV(const char *format, va_list vargs) |
| { |
| va_list count; |
| Py_ssize_t callcount = 0; |
| PyObject **callresults = NULL; |
| PyObject **callresult = NULL; |
| Py_ssize_t n = 0; |
| int width = 0; |
| int precision = 0; |
| int zeropad; |
| const char* f; |
| Py_UNICODE *s; |
| PyObject *string; |
| /* used by sprintf */ |
| char buffer[ITEM_BUFFER_LEN+1]; |
| /* use abuffer instead of buffer, if we need more space |
| * (which can happen if there's a format specifier with width). */ |
| char *abuffer = NULL; |
| char *realbuffer; |
| Py_ssize_t abuffersize = 0; |
| char fmt[61]; /* should be enough for %0width.precisionlld */ |
| const char *copy; |
| |
| Py_VA_COPY(count, vargs); |
| /* step 1: count the number of %S/%R/%A/%s format specifications |
| * (we call PyObject_Str()/PyObject_Repr()/PyObject_ASCII()/ |
| * PyUnicode_DecodeUTF8() for these objects once during step 3 and put the |
| * result in an array) */ |
| for (f = format; *f; f++) { |
| if (*f == '%') { |
| if (*(f+1)=='%') |
| continue; |
| if (*(f+1)=='S' || *(f+1)=='R' || *(f+1)=='A' || *(f+1) == 'V') |
| ++callcount; |
| while (Py_ISDIGIT((unsigned)*f)) |
| width = (width*10) + *f++ - '0'; |
| while (*++f && *f != '%' && !Py_ISALPHA((unsigned)*f)) |
| ; |
| if (*f == 's') |
| ++callcount; |
| } |
| else if (128 <= (unsigned char)*f) { |
| PyErr_Format(PyExc_ValueError, |
| "PyUnicode_FromFormatV() expects an ASCII-encoded format " |
| "string, got a non-ASCII byte: 0x%02x", |
| (unsigned char)*f); |
| return NULL; |
| } |
| } |
| /* step 2: allocate memory for the results of |
| * PyObject_Str()/PyObject_Repr()/PyUnicode_DecodeUTF8() calls */ |
| if (callcount) { |
| callresults = PyObject_Malloc(sizeof(PyObject *)*callcount); |
| if (!callresults) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| callresult = callresults; |
| } |
| /* step 3: figure out how large a buffer we need */ |
| for (f = format; *f; f++) { |
| if (*f == '%') { |
| #ifdef HAVE_LONG_LONG |
| int longlongflag = 0; |
| #endif |
| const char* p = f; |
| width = 0; |
| while (Py_ISDIGIT((unsigned)*f)) |
| width = (width*10) + *f++ - '0'; |
| while (*++f && *f != '%' && !Py_ISALPHA((unsigned)*f)) |
| ; |
| |
| /* skip the 'l' or 'z' in {%ld, %zd, %lu, %zu} since |
| * they don't affect the amount of space we reserve. |
| */ |
| if (*f == 'l') { |
| if (f[1] == 'd' || f[1] == 'u') { |
| ++f; |
| } |
| #ifdef HAVE_LONG_LONG |
| else if (f[1] == 'l' && |
| (f[2] == 'd' || f[2] == 'u')) { |
| longlongflag = 1; |
| f += 2; |
| } |
| #endif |
| } |
| else if (*f == 'z' && (f[1] == 'd' || f[1] == 'u')) { |
| ++f; |
| } |
| |
| switch (*f) { |
| case 'c': |
| { |
| #ifndef Py_UNICODE_WIDE |
| int ordinal = va_arg(count, int); |
| if (ordinal > 0xffff) |
| n += 2; |
| else |
| n++; |
| #else |
| (void)va_arg(count, int); |
| n++; |
| #endif |
| break; |
| } |
| case '%': |
| n++; |
| break; |
| case 'd': case 'u': case 'i': case 'x': |
| (void) va_arg(count, int); |
| #ifdef HAVE_LONG_LONG |
| if (longlongflag) { |
| if (width < MAX_LONG_LONG_CHARS) |
| width = MAX_LONG_LONG_CHARS; |
| } |
| else |
| #endif |
| /* MAX_LONG_CHARS is enough to hold a 64-bit integer, |
| including sign. Decimal takes the most space. This |
| isn't enough for octal. If a width is specified we |
| need more (which we allocate later). */ |
| if (width < MAX_LONG_CHARS) |
| width = MAX_LONG_CHARS; |
| n += width; |
| /* XXX should allow for large precision here too. */ |
| if (abuffersize < width) |
| abuffersize = width; |
| break; |
| case 's': |
| { |
| /* UTF-8 */ |
| const char *s = va_arg(count, const char*); |
| PyObject *str = PyUnicode_DecodeUTF8(s, strlen(s), "replace"); |
| if (!str) |
| goto fail; |
| n += PyUnicode_GET_SIZE(str); |
| /* Remember the str and switch to the next slot */ |
| *callresult++ = str; |
| break; |
| } |
| case 'U': |
| { |
| PyObject *obj = va_arg(count, PyObject *); |
| assert(obj && PyUnicode_Check(obj)); |
| n += PyUnicode_GET_SIZE(obj); |
| break; |
| } |
| case 'V': |
| { |
| PyObject *obj = va_arg(count, PyObject *); |
| const char *str = va_arg(count, const char *); |
| PyObject *str_obj; |
| assert(obj || str); |
| assert(!obj || PyUnicode_Check(obj)); |
| if (obj) { |
| n += PyUnicode_GET_SIZE(obj); |
| *callresult++ = NULL; |
| } |
| else { |
| str_obj = PyUnicode_DecodeUTF8(str, strlen(str), "replace"); |
| if (!str_obj) |
| goto fail; |
| n += PyUnicode_GET_SIZE(str_obj); |
| *callresult++ = str_obj; |
| } |
| break; |
| } |
| case 'S': |
| { |
| PyObject *obj = va_arg(count, PyObject *); |
| PyObject *str; |
| assert(obj); |
| str = PyObject_Str(obj); |
| if (!str) |
| goto fail; |
| n += PyUnicode_GET_SIZE(str); |
| /* Remember the str and switch to the next slot */ |
| *callresult++ = str; |
| break; |
| } |
| case 'R': |
| { |
| PyObject *obj = va_arg(count, PyObject *); |
| PyObject *repr; |
| assert(obj); |
| repr = PyObject_Repr(obj); |
| if (!repr) |
| goto fail; |
| n += PyUnicode_GET_SIZE(repr); |
| /* Remember the repr and switch to the next slot */ |
| *callresult++ = repr; |
| break; |
| } |
| case 'A': |
| { |
| PyObject *obj = va_arg(count, PyObject *); |
| PyObject *ascii; |
| assert(obj); |
| ascii = PyObject_ASCII(obj); |
| if (!ascii) |
| goto fail; |
| n += PyUnicode_GET_SIZE(ascii); |
| /* Remember the repr and switch to the next slot */ |
| *callresult++ = ascii; |
| break; |
| } |
| case 'p': |
| (void) va_arg(count, int); |
| /* maximum 64-bit pointer representation: |
| * 0xffffffffffffffff |
| * so 19 characters is enough. |
| * XXX I count 18 -- what's the extra for? |
| */ |
| n += 19; |
| break; |
| default: |
| /* if we stumble upon an unknown |
| formatting code, copy the rest of |
| the format string to the output |
| string. (we cannot just skip the |
| code, since there's no way to know |
| what's in the argument list) */ |
| n += strlen(p); |
| goto expand; |
| } |
| } else |
| n++; |
| } |
| expand: |
| if (abuffersize > ITEM_BUFFER_LEN) { |
| /* add 1 for sprintf's trailing null byte */ |
| abuffer = PyObject_Malloc(abuffersize + 1); |
| if (!abuffer) { |
| PyErr_NoMemory(); |
| goto fail; |
| } |
| realbuffer = abuffer; |
| } |
| else |
| realbuffer = buffer; |
| /* step 4: fill the buffer */ |
| /* Since we've analyzed how much space we need for the worst case, |
| we don't have to resize the string. |
| There can be no errors beyond this point. */ |
| string = PyUnicode_FromUnicode(NULL, n); |
| if (!string) |
| goto fail; |
| |
| s = PyUnicode_AS_UNICODE(string); |
| callresult = callresults; |
| |
| for (f = format; *f; f++) { |
| if (*f == '%') { |
| const char* p = f++; |
| int longflag = 0; |
| int longlongflag = 0; |
| int size_tflag = 0; |
| zeropad = (*f == '0'); |
| /* parse the width.precision part */ |
| width = 0; |
| while (Py_ISDIGIT((unsigned)*f)) |
| width = (width*10) + *f++ - '0'; |
| precision = 0; |
| if (*f == '.') { |
| f++; |
| while (Py_ISDIGIT((unsigned)*f)) |
| precision = (precision*10) + *f++ - '0'; |
| } |
| /* Handle %ld, %lu, %lld and %llu. */ |
| if (*f == 'l') { |
| if (f[1] == 'd' || f[1] == 'u') { |
| longflag = 1; |
| ++f; |
| } |
| #ifdef HAVE_LONG_LONG |
| else if (f[1] == 'l' && |
| (f[2] == 'd' || f[2] == 'u')) { |
| longlongflag = 1; |
| f += 2; |
| } |
| #endif |
| } |
| /* handle the size_t flag. */ |
| if (*f == 'z' && (f[1] == 'd' || f[1] == 'u')) { |
| size_tflag = 1; |
| ++f; |
| } |
| |
| switch (*f) { |
| case 'c': |
| { |
| int ordinal = va_arg(vargs, int); |
| #ifndef Py_UNICODE_WIDE |
| if (ordinal > 0xffff) { |
| ordinal -= 0x10000; |
| *s++ = 0xD800 | (ordinal >> 10); |
| *s++ = 0xDC00 | (ordinal & 0x3FF); |
| } else |
| #endif |
| *s++ = ordinal; |
| break; |
| } |
| case 'd': |
| makefmt(fmt, longflag, longlongflag, size_tflag, zeropad, |
| width, precision, 'd'); |
| if (longflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, long)); |
| #ifdef HAVE_LONG_LONG |
| else if (longlongflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, PY_LONG_LONG)); |
| #endif |
| else if (size_tflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, Py_ssize_t)); |
| else |
| sprintf(realbuffer, fmt, va_arg(vargs, int)); |
| appendstring(realbuffer); |
| break; |
| case 'u': |
| makefmt(fmt, longflag, longlongflag, size_tflag, zeropad, |
| width, precision, 'u'); |
| if (longflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, unsigned long)); |
| #ifdef HAVE_LONG_LONG |
| else if (longlongflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, |
| unsigned PY_LONG_LONG)); |
| #endif |
| else if (size_tflag) |
| sprintf(realbuffer, fmt, va_arg(vargs, size_t)); |
| else |
| sprintf(realbuffer, fmt, va_arg(vargs, unsigned int)); |
| appendstring(realbuffer); |
| break; |
| case 'i': |
| makefmt(fmt, 0, 0, 0, zeropad, width, precision, 'i'); |
| sprintf(realbuffer, fmt, va_arg(vargs, int)); |
| appendstring(realbuffer); |
| break; |
| case 'x': |
| makefmt(fmt, 0, 0, 0, zeropad, width, precision, 'x'); |
| sprintf(realbuffer, fmt, va_arg(vargs, int)); |
| appendstring(realbuffer); |
| break; |
| case 's': |
| { |
| /* unused, since we already have the result */ |
| (void) va_arg(vargs, char *); |
| Py_UNICODE_COPY(s, PyUnicode_AS_UNICODE(*callresult), |
| PyUnicode_GET_SIZE(*callresult)); |
| s += PyUnicode_GET_SIZE(*callresult); |
| /* We're done with the unicode()/repr() => forget it */ |
| Py_DECREF(*callresult); |
| /* switch to next unicode()/repr() result */ |
| ++callresult; |
| break; |
| } |
| case 'U': |
| { |
| PyObject *obj = va_arg(vargs, PyObject *); |
| Py_ssize_t size = PyUnicode_GET_SIZE(obj); |
| Py_UNICODE_COPY(s, PyUnicode_AS_UNICODE(obj), size); |
| s += size; |
| break; |
| } |
| case 'V': |
| { |
| PyObject *obj = va_arg(vargs, PyObject *); |
| va_arg(vargs, const char *); |
| if (obj) { |
| Py_ssize_t size = PyUnicode_GET_SIZE(obj); |
| Py_UNICODE_COPY(s, PyUnicode_AS_UNICODE(obj), size); |
| s += size; |
| } else { |
| Py_UNICODE_COPY(s, PyUnicode_AS_UNICODE(*callresult), |
| PyUnicode_GET_SIZE(*callresult)); |
| s += PyUnicode_GET_SIZE(*callresult); |
| Py_DECREF(*callresult); |
| } |
| ++callresult; |
| break; |
| } |
| case 'S': |
| case 'R': |
| case 'A': |
| { |
| Py_UNICODE *ucopy; |
| Py_ssize_t usize; |
| Py_ssize_t upos; |
| /* unused, since we already have the result */ |
| (void) va_arg(vargs, PyObject *); |
| ucopy = PyUnicode_AS_UNICODE(*callresult); |
| usize = PyUnicode_GET_SIZE(*callresult); |
| for (upos = 0; upos<usize;) |
| *s++ = ucopy[upos++]; |
| /* We're done with the unicode()/repr() => forget it */ |
| Py_DECREF(*callresult); |
| /* switch to next unicode()/repr() result */ |
| ++callresult; |
| break; |
| } |
| case 'p': |
| sprintf(buffer, "%p", va_arg(vargs, void*)); |
| /* %p is ill-defined: ensure leading 0x. */ |
| if (buffer[1] == 'X') |
| buffer[1] = 'x'; |
| else if (buffer[1] != 'x') { |
| memmove(buffer+2, buffer, strlen(buffer)+1); |
| buffer[0] = '0'; |
| buffer[1] = 'x'; |
| } |
| appendstring(buffer); |
| break; |
| case '%': |
| *s++ = '%'; |
| break; |
| default: |
| appendstring(p); |
| goto end; |
| } |
| } |
| else |
| *s++ = *f; |
| } |
| |
| end: |
| if (callresults) |
| PyObject_Free(callresults); |
| if (abuffer) |
| PyObject_Free(abuffer); |
| PyUnicode_Resize(&string, s - PyUnicode_AS_UNICODE(string)); |
| return string; |
| fail: |
| if (callresults) { |
| PyObject **callresult2 = callresults; |
| while (callresult2 < callresult) { |
| Py_XDECREF(*callresult2); |
| ++callresult2; |
| } |
| PyObject_Free(callresults); |
| } |
| if (abuffer) |
| PyObject_Free(abuffer); |
| return NULL; |
| } |
| |
| #undef appendstring |
| |
| PyObject * |
| PyUnicode_FromFormat(const char *format, ...) |
| { |
| PyObject* ret; |
| va_list vargs; |
| |
| #ifdef HAVE_STDARG_PROTOTYPES |
| va_start(vargs, format); |
| #else |
| va_start(vargs); |
| #endif |
| ret = PyUnicode_FromFormatV(format, vargs); |
| va_end(vargs); |
| return ret; |
| } |
| |
| /* Helper function for PyUnicode_AsWideChar() and PyUnicode_AsWideCharString(): |
| convert a Unicode object to a wide character string. |
| |
| - If w is NULL: return the number of wide characters (including the null |
| character) required to convert the unicode object. Ignore size argument. |
| |
| - Otherwise: return the number of wide characters (excluding the null |
| character) written into w. Write at most size wide characters (including |
| the null character). */ |
| static Py_ssize_t |
| unicode_aswidechar(PyUnicodeObject *unicode, |
| wchar_t *w, |
| Py_ssize_t size) |
| { |
| #if Py_UNICODE_SIZE == SIZEOF_WCHAR_T |
| Py_ssize_t res; |
| if (w != NULL) { |
| res = PyUnicode_GET_SIZE(unicode); |
| if (size > res) |
| size = res + 1; |
| else |
| res = size; |
| memcpy(w, unicode->str, size * sizeof(wchar_t)); |
| return res; |
| } |
| else |
| return PyUnicode_GET_SIZE(unicode) + 1; |
| #elif Py_UNICODE_SIZE == 2 && SIZEOF_WCHAR_T == 4 |
| register const Py_UNICODE *u; |
| const Py_UNICODE *uend; |
| const wchar_t *worig, *wend; |
| Py_ssize_t nchar; |
| |
| u = PyUnicode_AS_UNICODE(unicode); |
| uend = u + PyUnicode_GET_SIZE(unicode); |
| if (w != NULL) { |
| worig = w; |
| wend = w + size; |
| while (u != uend && w != wend) { |
| if (0xD800 <= u[0] && u[0] <= 0xDBFF |
| && 0xDC00 <= u[1] && u[1] <= 0xDFFF) |
| { |
| *w = (((u[0] & 0x3FF) << 10) | (u[1] & 0x3FF)) + 0x10000; |
| u += 2; |
| } |
| else { |
| *w = *u; |
| u++; |
| } |
| w++; |
| } |
| if (w != wend) |
| *w = L'\0'; |
| return w - worig; |
| } |
| else { |
| nchar = 1; /* null character at the end */ |
| while (u != uend) { |
| if (0xD800 <= u[0] && u[0] <= 0xDBFF |
| && 0xDC00 <= u[1] && u[1] <= 0xDFFF) |
| u += 2; |
| else |
| u++; |
| nchar++; |
| } |
| } |
| return nchar; |
| #elif Py_UNICODE_SIZE == 4 && SIZEOF_WCHAR_T == 2 |
| register Py_UNICODE *u, *uend, ordinal; |
| register Py_ssize_t i; |
| wchar_t *worig, *wend; |
| Py_ssize_t nchar; |
| |
| u = PyUnicode_AS_UNICODE(unicode); |
| uend = u + PyUnicode_GET_SIZE(u); |
| if (w != NULL) { |
| worig = w; |
| wend = w + size; |
| while (u != uend && w != wend) { |
| ordinal = *u; |
| if (ordinal > 0xffff) { |
| ordinal -= 0x10000; |
| *w++ = 0xD800 | (ordinal >> 10); |
| *w++ = 0xDC00 | (ordinal & 0x3FF); |
| } |
| else |
| *w++ = ordinal; |
| u++; |
| } |
| if (w != wend) |
| *w = 0; |
| return w - worig; |
| } |
| else { |
| nchar = 1; /* null character */ |
| while (u != uend) { |
| if (*u > 0xffff) |
| nchar += 2; |
| else |
| nchar++; |
| u++; |
| } |
| return nchar; |
| } |
| #else |
| # error "unsupported wchar_t and Py_UNICODE sizes, see issue #8670" |
| #endif |
| } |
| |
| Py_ssize_t |
| PyUnicode_AsWideChar(PyObject *unicode, |
| wchar_t *w, |
| Py_ssize_t size) |
| { |
| if (unicode == NULL) { |
| PyErr_BadInternalCall(); |
| return -1; |
| } |
| return unicode_aswidechar((PyUnicodeObject*)unicode, w, size); |
| } |
| |
| wchar_t* |
| PyUnicode_AsWideCharString(PyObject *unicode, |
| Py_ssize_t *size) |
| { |
| wchar_t* buffer; |
| Py_ssize_t buflen; |
| |
| if (unicode == NULL) { |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| |
| buflen = unicode_aswidechar((PyUnicodeObject *)unicode, NULL, 0); |
| if (PY_SSIZE_T_MAX / sizeof(wchar_t) < buflen) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| |
| buffer = PyMem_MALLOC(buflen * sizeof(wchar_t)); |
| if (buffer == NULL) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| buflen = unicode_aswidechar((PyUnicodeObject *)unicode, buffer, buflen); |
| if (size != NULL) |
| *size = buflen; |
| return buffer; |
| } |
| |
| #endif |
| |
| PyObject *PyUnicode_FromOrdinal(int ordinal) |
| { |
| Py_UNICODE s[2]; |
| |
| if (ordinal < 0 || ordinal > 0x10ffff) { |
| PyErr_SetString(PyExc_ValueError, |
| "chr() arg not in range(0x110000)"); |
| return NULL; |
| } |
| |
| #ifndef Py_UNICODE_WIDE |
| if (ordinal > 0xffff) { |
| ordinal -= 0x10000; |
| s[0] = 0xD800 | (ordinal >> 10); |
| s[1] = 0xDC00 | (ordinal & 0x3FF); |
| return PyUnicode_FromUnicode(s, 2); |
| } |
| #endif |
| |
| s[0] = (Py_UNICODE)ordinal; |
| return PyUnicode_FromUnicode(s, 1); |
| } |
| |
| PyObject *PyUnicode_FromObject(register PyObject *obj) |
| { |
| /* XXX Perhaps we should make this API an alias of |
| PyObject_Str() instead ?! */ |
| if (PyUnicode_CheckExact(obj)) { |
| Py_INCREF(obj); |
| return obj; |
| } |
| if (PyUnicode_Check(obj)) { |
| /* For a Unicode subtype that's not a Unicode object, |
| return a true Unicode object with the same data. */ |
| return PyUnicode_FromUnicode(PyUnicode_AS_UNICODE(obj), |
| PyUnicode_GET_SIZE(obj)); |
| } |
| PyErr_Format(PyExc_TypeError, |
| "Can't convert '%.100s' object to str implicitly", |
| Py_TYPE(obj)->tp_name); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_FromEncodedObject(register PyObject *obj, |
| const char *encoding, |
| const char *errors) |
| { |
| Py_buffer buffer; |
| PyObject *v; |
| |
| if (obj == NULL) { |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| |
| /* Decoding bytes objects is the most common case and should be fast */ |
| if (PyBytes_Check(obj)) { |
| if (PyBytes_GET_SIZE(obj) == 0) |
| _Py_RETURN_UNICODE_EMPTY(); |
| v = PyUnicode_Decode( |
| PyBytes_AS_STRING(obj), PyBytes_GET_SIZE(obj), |
| encoding, errors); |
| return v; |
| } |
| |
| if (PyUnicode_Check(obj)) { |
| PyErr_SetString(PyExc_TypeError, |
| "decoding str is not supported"); |
| return NULL; |
| } |
| |
| /* Retrieve a bytes buffer view through the PEP 3118 buffer interface */ |
| if (PyObject_GetBuffer(obj, &buffer, PyBUF_SIMPLE) < 0) { |
| PyErr_Format(PyExc_TypeError, |
| "coercing to str: need bytes, bytearray " |
| "or buffer-like object, %.80s found", |
| Py_TYPE(obj)->tp_name); |
| return NULL; |
| } |
| |
| if (buffer.len == 0) { |
| PyBuffer_Release(&buffer); |
| _Py_RETURN_UNICODE_EMPTY(); |
| } |
| |
| v = PyUnicode_Decode((char*) buffer.buf, buffer.len, encoding, errors); |
| PyBuffer_Release(&buffer); |
| return v; |
| } |
| |
| /* Convert encoding to lower case and replace '_' with '-' in order to |
| catch e.g. UTF_8. Return 0 on error (encoding is longer than lower_len-1), |
| 1 on success. */ |
| int |
| _Py_normalize_encoding(const char *encoding, |
| char *lower, |
| size_t lower_len) |
| { |
| const char *e; |
| char *l; |
| char *l_end; |
| |
| e = encoding; |
| l = lower; |
| l_end = &lower[lower_len - 1]; |
| while (*e) { |
| if (l == l_end) |
| return 0; |
| if (Py_ISUPPER(*e)) { |
| *l++ = Py_TOLOWER(*e++); |
| } |
| else if (*e == '_') { |
| *l++ = '-'; |
| e++; |
| } |
| else { |
| *l++ = *e++; |
| } |
| } |
| *l = '\0'; |
| return 1; |
| } |
| |
| PyObject *PyUnicode_Decode(const char *s, |
| Py_ssize_t size, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *buffer = NULL, *unicode; |
| Py_buffer info; |
| char lower[11]; /* Enough for any encoding shortcut */ |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Shortcuts for common default encodings */ |
| if (_Py_normalize_encoding(encoding, lower, sizeof(lower))) { |
| if (strcmp(lower, "utf-8") == 0) |
| return PyUnicode_DecodeUTF8(s, size, errors); |
| else if ((strcmp(lower, "latin-1") == 0) || |
| (strcmp(lower, "iso-8859-1") == 0)) |
| return PyUnicode_DecodeLatin1(s, size, errors); |
| #if defined(MS_WINDOWS) && defined(HAVE_USABLE_WCHAR_T) |
| else if (strcmp(lower, "mbcs") == 0) |
| return PyUnicode_DecodeMBCS(s, size, errors); |
| #endif |
| else if (strcmp(lower, "ascii") == 0) |
| return PyUnicode_DecodeASCII(s, size, errors); |
| else if (strcmp(lower, "utf-16") == 0) |
| return PyUnicode_DecodeUTF16(s, size, errors, 0); |
| else if (strcmp(lower, "utf-32") == 0) |
| return PyUnicode_DecodeUTF32(s, size, errors, 0); |
| } |
| |
| /* Decode via the codec registry */ |
| buffer = NULL; |
| if (PyBuffer_FillInfo(&info, NULL, (void *)s, size, 1, PyBUF_FULL_RO) < 0) |
| goto onError; |
| buffer = PyMemoryView_FromBuffer(&info); |
| if (buffer == NULL) |
| goto onError; |
| unicode = PyCodec_Decode(buffer, encoding, errors); |
| if (unicode == NULL) |
| goto onError; |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_Format(PyExc_TypeError, |
| "decoder did not return a str object (type=%.400s)", |
| Py_TYPE(unicode)->tp_name); |
| Py_DECREF(unicode); |
| goto onError; |
| } |
| Py_DECREF(buffer); |
| return unicode; |
| |
| onError: |
| Py_XDECREF(buffer); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_AsDecodedObject(PyObject *unicode, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v; |
| |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Decode via the codec registry */ |
| v = PyCodec_Decode(unicode, encoding, errors); |
| if (v == NULL) |
| goto onError; |
| return v; |
| |
| onError: |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_AsDecodedUnicode(PyObject *unicode, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v; |
| |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Decode via the codec registry */ |
| v = PyCodec_Decode(unicode, encoding, errors); |
| if (v == NULL) |
| goto onError; |
| if (!PyUnicode_Check(v)) { |
| PyErr_Format(PyExc_TypeError, |
| "decoder did not return a str object (type=%.400s)", |
| Py_TYPE(v)->tp_name); |
| Py_DECREF(v); |
| goto onError; |
| } |
| return v; |
| |
| onError: |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_Encode(const Py_UNICODE *s, |
| Py_ssize_t size, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v, *unicode; |
| |
| unicode = PyUnicode_FromUnicode(s, size); |
| if (unicode == NULL) |
| return NULL; |
| v = PyUnicode_AsEncodedString(unicode, encoding, errors); |
| Py_DECREF(unicode); |
| return v; |
| } |
| |
| PyObject *PyUnicode_AsEncodedObject(PyObject *unicode, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v; |
| |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Encode via the codec registry */ |
| v = PyCodec_Encode(unicode, encoding, errors); |
| if (v == NULL) |
| goto onError; |
| return v; |
| |
| onError: |
| return NULL; |
| } |
| |
| PyObject * |
| PyUnicode_EncodeFSDefault(PyObject *unicode) |
| { |
| #if defined(MS_WINDOWS) && defined(HAVE_USABLE_WCHAR_T) |
| return PyUnicode_EncodeMBCS(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| #elif defined(__APPLE__) |
| return PyUnicode_EncodeUTF8(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| "surrogateescape"); |
| #else |
| PyInterpreterState *interp = PyThreadState_GET()->interp; |
| /* Bootstrap check: if the filesystem codec is implemented in Python, we |
| cannot use it to encode and decode filenames before it is loaded. Load |
| the Python codec requires to encode at least its own filename. Use the C |
| version of the locale codec until the codec registry is initialized and |
| the Python codec is loaded. |
| |
| Py_FileSystemDefaultEncoding is shared between all interpreters, we |
| cannot only rely on it: check also interp->fscodec_initialized for |
| subinterpreters. */ |
| if (Py_FileSystemDefaultEncoding && interp->fscodec_initialized) { |
| return PyUnicode_AsEncodedString(unicode, |
| Py_FileSystemDefaultEncoding, |
| "surrogateescape"); |
| } |
| else { |
| /* locale encoding with surrogateescape */ |
| wchar_t *wchar; |
| char *bytes; |
| PyObject *bytes_obj; |
| size_t error_pos; |
| |
| wchar = PyUnicode_AsWideCharString(unicode, NULL); |
| if (wchar == NULL) |
| return NULL; |
| bytes = _Py_wchar2char(wchar, &error_pos); |
| if (bytes == NULL) { |
| if (error_pos != (size_t)-1) { |
| char *errmsg = strerror(errno); |
| PyObject *exc = NULL; |
| if (errmsg == NULL) |
| errmsg = "Py_wchar2char() failed"; |
| raise_encode_exception(&exc, |
| "filesystemencoding", |
| PyUnicode_AS_UNICODE(unicode), PyUnicode_GET_SIZE(unicode), |
| error_pos, error_pos+1, |
| errmsg); |
| Py_XDECREF(exc); |
| } |
| else |
| PyErr_NoMemory(); |
| PyMem_Free(wchar); |
| return NULL; |
| } |
| PyMem_Free(wchar); |
| |
| bytes_obj = PyBytes_FromString(bytes); |
| PyMem_Free(bytes); |
| return bytes_obj; |
| } |
| #endif |
| } |
| |
| PyObject *PyUnicode_AsEncodedString(PyObject *unicode, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v; |
| char lower[11]; /* Enough for any encoding shortcut */ |
| |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Shortcuts for common default encodings */ |
| if (_Py_normalize_encoding(encoding, lower, sizeof(lower))) { |
| if (strcmp(lower, "utf-8") == 0) |
| return PyUnicode_EncodeUTF8(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| errors); |
| else if ((strcmp(lower, "latin-1") == 0) || |
| (strcmp(lower, "iso-8859-1") == 0)) |
| return PyUnicode_EncodeLatin1(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| errors); |
| #if defined(MS_WINDOWS) && defined(HAVE_USABLE_WCHAR_T) |
| else if (strcmp(lower, "mbcs") == 0) |
| return PyUnicode_EncodeMBCS(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| errors); |
| #endif |
| else if (strcmp(lower, "ascii") == 0) |
| return PyUnicode_EncodeASCII(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| errors); |
| } |
| /* During bootstrap, we may need to find the encodings |
| package, to load the file system encoding, and require the |
| file system encoding in order to load the encodings |
| package. |
| |
| Break out of this dependency by assuming that the path to |
| the encodings module is ASCII-only. XXX could try wcstombs |
| instead, if the file system encoding is the locale's |
| encoding. */ |
| if (Py_FileSystemDefaultEncoding && |
| strcmp(encoding, Py_FileSystemDefaultEncoding) == 0 && |
| !PyThreadState_GET()->interp->codecs_initialized) |
| return PyUnicode_EncodeASCII(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| errors); |
| |
| /* Encode via the codec registry */ |
| v = PyCodec_Encode(unicode, encoding, errors); |
| if (v == NULL) |
| return NULL; |
| |
| /* The normal path */ |
| if (PyBytes_Check(v)) |
| return v; |
| |
| /* If the codec returns a buffer, raise a warning and convert to bytes */ |
| if (PyByteArray_Check(v)) { |
| int error; |
| PyObject *b; |
| |
| error = PyErr_WarnFormat(PyExc_RuntimeWarning, 1, |
| "encoder %s returned bytearray instead of bytes", |
| encoding); |
| if (error) { |
| Py_DECREF(v); |
| return NULL; |
| } |
| |
| b = PyBytes_FromStringAndSize(PyByteArray_AS_STRING(v), Py_SIZE(v)); |
| Py_DECREF(v); |
| return b; |
| } |
| |
| PyErr_Format(PyExc_TypeError, |
| "encoder did not return a bytes object (type=%.400s)", |
| Py_TYPE(v)->tp_name); |
| Py_DECREF(v); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_AsEncodedUnicode(PyObject *unicode, |
| const char *encoding, |
| const char *errors) |
| { |
| PyObject *v; |
| |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| |
| if (encoding == NULL) |
| encoding = PyUnicode_GetDefaultEncoding(); |
| |
| /* Encode via the codec registry */ |
| v = PyCodec_Encode(unicode, encoding, errors); |
| if (v == NULL) |
| goto onError; |
| if (!PyUnicode_Check(v)) { |
| PyErr_Format(PyExc_TypeError, |
| "encoder did not return an str object (type=%.400s)", |
| Py_TYPE(v)->tp_name); |
| Py_DECREF(v); |
| goto onError; |
| } |
| return v; |
| |
| onError: |
| return NULL; |
| } |
| |
| PyObject *_PyUnicode_AsDefaultEncodedString(PyObject *unicode, |
| const char *errors) |
| { |
| PyObject *v = ((PyUnicodeObject *)unicode)->defenc; |
| if (v) |
| return v; |
| if (errors != NULL) |
| Py_FatalError("non-NULL encoding in _PyUnicode_AsDefaultEncodedString"); |
| v = PyUnicode_EncodeUTF8(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| if (!v) |
| return NULL; |
| ((PyUnicodeObject *)unicode)->defenc = v; |
| return v; |
| } |
| |
| PyObject* |
| PyUnicode_DecodeFSDefault(const char *s) { |
| Py_ssize_t size = (Py_ssize_t)strlen(s); |
| return PyUnicode_DecodeFSDefaultAndSize(s, size); |
| } |
| |
| PyObject* |
| PyUnicode_DecodeFSDefaultAndSize(const char *s, Py_ssize_t size) |
| { |
| #if defined(MS_WINDOWS) && defined(HAVE_USABLE_WCHAR_T) |
| return PyUnicode_DecodeMBCS(s, size, NULL); |
| #elif defined(__APPLE__) |
| return PyUnicode_DecodeUTF8(s, size, "surrogateescape"); |
| #else |
| PyInterpreterState *interp = PyThreadState_GET()->interp; |
| /* Bootstrap check: if the filesystem codec is implemented in Python, we |
| cannot use it to encode and decode filenames before it is loaded. Load |
| the Python codec requires to encode at least its own filename. Use the C |
| version of the locale codec until the codec registry is initialized and |
| the Python codec is loaded. |
| |
| Py_FileSystemDefaultEncoding is shared between all interpreters, we |
| cannot only rely on it: check also interp->fscodec_initialized for |
| subinterpreters. */ |
| if (Py_FileSystemDefaultEncoding && interp->fscodec_initialized) { |
| return PyUnicode_Decode(s, size, |
| Py_FileSystemDefaultEncoding, |
| "surrogateescape"); |
| } |
| else { |
| /* locale encoding with surrogateescape */ |
| wchar_t *wchar; |
| PyObject *unicode; |
| size_t len; |
| |
| if (s[size] != '\0' || size != strlen(s)) { |
| PyErr_SetString(PyExc_TypeError, "embedded NUL character"); |
| return NULL; |
| } |
| |
| wchar = _Py_char2wchar(s, &len); |
| if (wchar == NULL) |
| return PyErr_NoMemory(); |
| |
| unicode = PyUnicode_FromWideChar(wchar, len); |
| PyMem_Free(wchar); |
| return unicode; |
| } |
| #endif |
| } |
| |
| |
| int |
| _PyUnicode_HasNULChars(PyObject* s) |
| { |
| static PyObject *nul = NULL; |
| |
| if (nul == NULL) |
| nul = PyUnicode_FromStringAndSize("\0", 1); |
| if (nul == NULL) |
| return -1; |
| return PyUnicode_Contains(s, nul); |
| } |
| |
| |
| int |
| PyUnicode_FSConverter(PyObject* arg, void* addr) |
| { |
| PyObject *output = NULL; |
| Py_ssize_t size; |
| void *data; |
| if (arg == NULL) { |
| Py_DECREF(*(PyObject**)addr); |
| return 1; |
| } |
| if (PyBytes_Check(arg)) { |
| output = arg; |
| Py_INCREF(output); |
| } |
| else { |
| arg = PyUnicode_FromObject(arg); |
| if (!arg) |
| return 0; |
| output = PyUnicode_EncodeFSDefault(arg); |
| Py_DECREF(arg); |
| if (!output) |
| return 0; |
| if (!PyBytes_Check(output)) { |
| Py_DECREF(output); |
| PyErr_SetString(PyExc_TypeError, "encoder failed to return bytes"); |
| return 0; |
| } |
| } |
| size = PyBytes_GET_SIZE(output); |
| data = PyBytes_AS_STRING(output); |
| if (size != strlen(data)) { |
| PyErr_SetString(PyExc_TypeError, "embedded NUL character"); |
| Py_DECREF(output); |
| return 0; |
| } |
| *(PyObject**)addr = output; |
| return Py_CLEANUP_SUPPORTED; |
| } |
| |
| |
| int |
| PyUnicode_FSDecoder(PyObject* arg, void* addr) |
| { |
| PyObject *output = NULL; |
| Py_ssize_t size; |
| void *data; |
| if (arg == NULL) { |
| Py_DECREF(*(PyObject**)addr); |
| return 1; |
| } |
| if (PyUnicode_Check(arg)) { |
| output = arg; |
| Py_INCREF(output); |
| } |
| else { |
| arg = PyBytes_FromObject(arg); |
| if (!arg) |
| return 0; |
| output = PyUnicode_DecodeFSDefaultAndSize(PyBytes_AS_STRING(arg), |
| PyBytes_GET_SIZE(arg)); |
| Py_DECREF(arg); |
| if (!output) |
| return 0; |
| if (!PyUnicode_Check(output)) { |
| Py_DECREF(output); |
| PyErr_SetString(PyExc_TypeError, "decoder failed to return unicode"); |
| return 0; |
| } |
| } |
| size = PyUnicode_GET_SIZE(output); |
| data = PyUnicode_AS_UNICODE(output); |
| if (size != Py_UNICODE_strlen(data)) { |
| PyErr_SetString(PyExc_TypeError, "embedded NUL character"); |
| Py_DECREF(output); |
| return 0; |
| } |
| *(PyObject**)addr = output; |
| return Py_CLEANUP_SUPPORTED; |
| } |
| |
| |
| char* |
| _PyUnicode_AsStringAndSize(PyObject *unicode, Py_ssize_t *psize) |
| { |
| PyObject *bytes; |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| bytes = _PyUnicode_AsDefaultEncodedString(unicode, NULL); |
| if (bytes == NULL) |
| return NULL; |
| if (psize != NULL) |
| *psize = PyBytes_GET_SIZE(bytes); |
| return PyBytes_AS_STRING(bytes); |
| } |
| |
| char* |
| _PyUnicode_AsString(PyObject *unicode) |
| { |
| return _PyUnicode_AsStringAndSize(unicode, NULL); |
| } |
| |
| Py_UNICODE *PyUnicode_AsUnicode(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| return PyUnicode_AS_UNICODE(unicode); |
| |
| onError: |
| return NULL; |
| } |
| |
| Py_ssize_t PyUnicode_GetSize(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| goto onError; |
| } |
| return PyUnicode_GET_SIZE(unicode); |
| |
| onError: |
| return -1; |
| } |
| |
| const char *PyUnicode_GetDefaultEncoding(void) |
| { |
| return "utf-8"; |
| } |
| |
| /* create or adjust a UnicodeDecodeError */ |
| static void |
| make_decode_exception(PyObject **exceptionObject, |
| const char *encoding, |
| const char *input, Py_ssize_t length, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason) |
| { |
| if (*exceptionObject == NULL) { |
| *exceptionObject = PyUnicodeDecodeError_Create( |
| encoding, input, length, startpos, endpos, reason); |
| } |
| else { |
| if (PyUnicodeDecodeError_SetStart(*exceptionObject, startpos)) |
| goto onError; |
| if (PyUnicodeDecodeError_SetEnd(*exceptionObject, endpos)) |
| goto onError; |
| if (PyUnicodeDecodeError_SetReason(*exceptionObject, reason)) |
| goto onError; |
| } |
| return; |
| |
| onError: |
| Py_DECREF(*exceptionObject); |
| *exceptionObject = NULL; |
| } |
| |
| /* error handling callback helper: |
| build arguments, call the callback and check the arguments, |
| if no exception occurred, copy the replacement to the output |
| and adjust various state variables. |
| return 0 on success, -1 on error |
| */ |
| |
| static |
| int unicode_decode_call_errorhandler(const char *errors, PyObject **errorHandler, |
| const char *encoding, const char *reason, |
| const char **input, const char **inend, Py_ssize_t *startinpos, |
| Py_ssize_t *endinpos, PyObject **exceptionObject, const char **inptr, |
| PyUnicodeObject **output, Py_ssize_t *outpos, Py_UNICODE **outptr) |
| { |
| static char *argparse = "O!n;decoding error handler must return (str, int) tuple"; |
| |
| PyObject *restuple = NULL; |
| PyObject *repunicode = NULL; |
| Py_ssize_t outsize = PyUnicode_GET_SIZE(*output); |
| Py_ssize_t insize; |
| Py_ssize_t requiredsize; |
| Py_ssize_t newpos; |
| Py_UNICODE *repptr; |
| PyObject *inputobj = NULL; |
| Py_ssize_t repsize; |
| int res = -1; |
| |
| if (*errorHandler == NULL) { |
| *errorHandler = PyCodec_LookupError(errors); |
| if (*errorHandler == NULL) |
| goto onError; |
| } |
| |
| make_decode_exception(exceptionObject, |
| encoding, |
| *input, *inend - *input, |
| *startinpos, *endinpos, |
| reason); |
| if (*exceptionObject == NULL) |
| goto onError; |
| |
| restuple = PyObject_CallFunctionObjArgs(*errorHandler, *exceptionObject, NULL); |
| if (restuple == NULL) |
| goto onError; |
| if (!PyTuple_Check(restuple)) { |
| PyErr_SetString(PyExc_TypeError, &argparse[4]); |
| goto onError; |
| } |
| if (!PyArg_ParseTuple(restuple, argparse, &PyUnicode_Type, &repunicode, &newpos)) |
| goto onError; |
| |
| /* Copy back the bytes variables, which might have been modified by the |
| callback */ |
| inputobj = PyUnicodeDecodeError_GetObject(*exceptionObject); |
| if (!inputobj) |
| goto onError; |
| if (!PyBytes_Check(inputobj)) { |
| PyErr_Format(PyExc_TypeError, "exception attribute object must be bytes"); |
| } |
| *input = PyBytes_AS_STRING(inputobj); |
| insize = PyBytes_GET_SIZE(inputobj); |
| *inend = *input + insize; |
| /* we can DECREF safely, as the exception has another reference, |
| so the object won't go away. */ |
| Py_DECREF(inputobj); |
| |
| if (newpos<0) |
| newpos = insize+newpos; |
| if (newpos<0 || newpos>insize) { |
| PyErr_Format(PyExc_IndexError, "position %zd from error handler out of bounds", newpos); |
| goto onError; |
| } |
| |
| /* need more space? (at least enough for what we |
| have+the replacement+the rest of the string (starting |
| at the new input position), so we won't have to check space |
| when there are no errors in the rest of the string) */ |
| repptr = PyUnicode_AS_UNICODE(repunicode); |
| repsize = PyUnicode_GET_SIZE(repunicode); |
| requiredsize = *outpos + repsize + insize-newpos; |
| if (requiredsize > outsize) { |
| if (requiredsize<2*outsize) |
| requiredsize = 2*outsize; |
| if (_PyUnicode_Resize(output, requiredsize) < 0) |
| goto onError; |
| *outptr = PyUnicode_AS_UNICODE(*output) + *outpos; |
| } |
| *endinpos = newpos; |
| *inptr = *input + newpos; |
| Py_UNICODE_COPY(*outptr, repptr, repsize); |
| *outptr += repsize; |
| *outpos += repsize; |
| |
| /* we made it! */ |
| res = 0; |
| |
| onError: |
| Py_XDECREF(restuple); |
| return res; |
| } |
| |
| /* --- UTF-7 Codec -------------------------------------------------------- */ |
| |
| /* See RFC2152 for details. We encode conservatively and decode liberally. */ |
| |
| /* Three simple macros defining base-64. */ |
| |
| /* Is c a base-64 character? */ |
| |
| #define IS_BASE64(c) \ |
| (((c) >= 'A' && (c) <= 'Z') || \ |
| ((c) >= 'a' && (c) <= 'z') || \ |
| ((c) >= '0' && (c) <= '9') || \ |
| (c) == '+' || (c) == '/') |
| |
| /* given that c is a base-64 character, what is its base-64 value? */ |
| |
| #define FROM_BASE64(c) \ |
| (((c) >= 'A' && (c) <= 'Z') ? (c) - 'A' : \ |
| ((c) >= 'a' && (c) <= 'z') ? (c) - 'a' + 26 : \ |
| ((c) >= '0' && (c) <= '9') ? (c) - '0' + 52 : \ |
| (c) == '+' ? 62 : 63) |
| |
| /* What is the base-64 character of the bottom 6 bits of n? */ |
| |
| #define TO_BASE64(n) \ |
| ("ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"[(n) & 0x3f]) |
| |
| /* DECODE_DIRECT: this byte encountered in a UTF-7 string should be |
| * decoded as itself. We are permissive on decoding; the only ASCII |
| * byte not decoding to itself is the + which begins a base64 |
| * string. */ |
| |
| #define DECODE_DIRECT(c) \ |
| ((c) <= 127 && (c) != '+') |
| |
| /* The UTF-7 encoder treats ASCII characters differently according to |
| * whether they are Set D, Set O, Whitespace, or special (i.e. none of |
| * the above). See RFC2152. This array identifies these different |
| * sets: |
| * 0 : "Set D" |
| * alphanumeric and '(),-./:? |
| * 1 : "Set O" |
| * !"#$%&*;<=>@[]^_`{|} |
| * 2 : "whitespace" |
| * ht nl cr sp |
| * 3 : special (must be base64 encoded) |
| * everything else (i.e. +\~ and non-printing codes 0-8 11-12 14-31 127) |
| */ |
| |
| static |
| char utf7_category[128] = { |
| /* nul soh stx etx eot enq ack bel bs ht nl vt np cr so si */ |
| 3, 3, 3, 3, 3, 3, 3, 3, 3, 2, 2, 3, 3, 2, 3, 3, |
| /* dle dc1 dc2 dc3 dc4 nak syn etb can em sub esc fs gs rs us */ |
| 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, |
| /* sp ! " # $ % & ' ( ) * + , - . / */ |
| 2, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 3, 0, 0, 0, 0, |
| /* 0 1 2 3 4 5 6 7 8 9 : ; < = > ? */ |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, |
| /* @ A B C D E F G H I J K L M N O */ |
| 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, |
| /* P Q R S T U V W X Y Z [ \ ] ^ _ */ |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 3, 1, 1, 1, |
| /* ` a b c d e f g h i j k l m n o */ |
| 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, |
| /* p q r s t u v w x y z { | } ~ del */ |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 3, 3, |
| }; |
| |
| /* ENCODE_DIRECT: this character should be encoded as itself. The |
| * answer depends on whether we are encoding set O as itself, and also |
| * on whether we are encoding whitespace as itself. RFC2152 makes it |
| * clear that the answers to these questions vary between |
| * applications, so this code needs to be flexible. */ |
| |
| #define ENCODE_DIRECT(c, directO, directWS) \ |
| ((c) < 128 && (c) > 0 && \ |
| ((utf7_category[(c)] == 0) || \ |
| (directWS && (utf7_category[(c)] == 2)) || \ |
| (directO && (utf7_category[(c)] == 1)))) |
| |
| PyObject *PyUnicode_DecodeUTF7(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| return PyUnicode_DecodeUTF7Stateful(s, size, errors, NULL); |
| } |
| |
| /* The decoder. The only state we preserve is our read position, |
| * i.e. how many characters we have consumed. So if we end in the |
| * middle of a shift sequence we have to back off the read position |
| * and the output to the beginning of the sequence, otherwise we lose |
| * all the shift state (seen bits, number of bits seen, high |
| * surrogate). */ |
| |
| PyObject *PyUnicode_DecodeUTF7Stateful(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| Py_ssize_t *consumed) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| const char *e; |
| PyUnicodeObject *unicode; |
| Py_UNICODE *p; |
| const char *errmsg = ""; |
| int inShift = 0; |
| Py_UNICODE *shiftOutStart; |
| unsigned int base64bits = 0; |
| unsigned long base64buffer = 0; |
| Py_UNICODE surrogate = 0; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| if (size == 0) { |
| if (consumed) |
| *consumed = 0; |
| return (PyObject *)unicode; |
| } |
| |
| p = unicode->str; |
| shiftOutStart = p; |
| e = s + size; |
| |
| while (s < e) { |
| Py_UNICODE ch; |
| restart: |
| ch = (unsigned char) *s; |
| |
| if (inShift) { /* in a base-64 section */ |
| if (IS_BASE64(ch)) { /* consume a base-64 character */ |
| base64buffer = (base64buffer << 6) | FROM_BASE64(ch); |
| base64bits += 6; |
| s++; |
| if (base64bits >= 16) { |
| /* we have enough bits for a UTF-16 value */ |
| Py_UNICODE outCh = (Py_UNICODE) |
| (base64buffer >> (base64bits-16)); |
| base64bits -= 16; |
| base64buffer &= (1 << base64bits) - 1; /* clear high bits */ |
| if (surrogate) { |
| /* expecting a second surrogate */ |
| if (outCh >= 0xDC00 && outCh <= 0xDFFF) { |
| #ifdef Py_UNICODE_WIDE |
| *p++ = (((surrogate & 0x3FF)<<10) |
| | (outCh & 0x3FF)) + 0x10000; |
| #else |
| *p++ = surrogate; |
| *p++ = outCh; |
| #endif |
| surrogate = 0; |
| continue; |
| } |
| else { |
| *p++ = surrogate; |
| surrogate = 0; |
| } |
| } |
| if (outCh >= 0xD800 && outCh <= 0xDBFF) { |
| /* first surrogate */ |
| surrogate = outCh; |
| } |
| else { |
| *p++ = outCh; |
| } |
| } |
| } |
| else { /* now leaving a base-64 section */ |
| inShift = 0; |
| s++; |
| if (surrogate) { |
| *p++ = surrogate; |
| surrogate = 0; |
| } |
| if (base64bits > 0) { /* left-over bits */ |
| if (base64bits >= 6) { |
| /* We've seen at least one base-64 character */ |
| errmsg = "partial character in shift sequence"; |
| goto utf7Error; |
| } |
| else { |
| /* Some bits remain; they should be zero */ |
| if (base64buffer != 0) { |
| errmsg = "non-zero padding bits in shift sequence"; |
| goto utf7Error; |
| } |
| } |
| } |
| if (ch != '-') { |
| /* '-' is absorbed; other terminating |
| characters are preserved */ |
| *p++ = ch; |
| } |
| } |
| } |
| else if ( ch == '+' ) { |
| startinpos = s-starts; |
| s++; /* consume '+' */ |
| if (s < e && *s == '-') { /* '+-' encodes '+' */ |
| s++; |
| *p++ = '+'; |
| } |
| else { /* begin base64-encoded section */ |
| inShift = 1; |
| shiftOutStart = p; |
| base64bits = 0; |
| } |
| } |
| else if (DECODE_DIRECT(ch)) { /* character decodes as itself */ |
| *p++ = ch; |
| s++; |
| } |
| else { |
| startinpos = s-starts; |
| s++; |
| errmsg = "unexpected special character"; |
| goto utf7Error; |
| } |
| continue; |
| utf7Error: |
| outpos = p-PyUnicode_AS_UNICODE(unicode); |
| endinpos = s-starts; |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "utf7", errmsg, |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &unicode, &outpos, &p)) |
| goto onError; |
| } |
| |
| /* end of string */ |
| |
| if (inShift && !consumed) { /* in shift sequence, no more to follow */ |
| /* if we're in an inconsistent state, that's an error */ |
| if (surrogate || |
| (base64bits >= 6) || |
| (base64bits > 0 && base64buffer != 0)) { |
| outpos = p-PyUnicode_AS_UNICODE(unicode); |
| endinpos = size; |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "utf7", "unterminated shift sequence", |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &unicode, &outpos, &p)) |
| goto onError; |
| if (s < e) |
| goto restart; |
| } |
| } |
| |
| /* return state */ |
| if (consumed) { |
| if (inShift) { |
| p = shiftOutStart; /* back off output */ |
| *consumed = startinpos; |
| } |
| else { |
| *consumed = s-starts; |
| } |
| } |
| |
| if (_PyUnicode_Resize(&unicode, p - PyUnicode_AS_UNICODE(unicode)) < 0) |
| goto onError; |
| |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)unicode; |
| |
| onError: |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| Py_DECREF(unicode); |
| return NULL; |
| } |
| |
| |
| PyObject *PyUnicode_EncodeUTF7(const Py_UNICODE *s, |
| Py_ssize_t size, |
| int base64SetO, |
| int base64WhiteSpace, |
| const char *errors) |
| { |
| PyObject *v; |
| /* It might be possible to tighten this worst case */ |
| Py_ssize_t allocated = 8 * size; |
| int inShift = 0; |
| Py_ssize_t i = 0; |
| unsigned int base64bits = 0; |
| unsigned long base64buffer = 0; |
| char * out; |
| char * start; |
| |
| if (size == 0) |
| return PyBytes_FromStringAndSize(NULL, 0); |
| |
| if (allocated / 8 != size) |
| return PyErr_NoMemory(); |
| |
| v = PyBytes_FromStringAndSize(NULL, allocated); |
| if (v == NULL) |
| return NULL; |
| |
| start = out = PyBytes_AS_STRING(v); |
| for (;i < size; ++i) { |
| Py_UNICODE ch = s[i]; |
| |
| if (inShift) { |
| if (ENCODE_DIRECT(ch, !base64SetO, !base64WhiteSpace)) { |
| /* shifting out */ |
| if (base64bits) { /* output remaining bits */ |
| *out++ = TO_BASE64(base64buffer << (6-base64bits)); |
| base64buffer = 0; |
| base64bits = 0; |
| } |
| inShift = 0; |
| /* Characters not in the BASE64 set implicitly unshift the sequence |
| so no '-' is required, except if the character is itself a '-' */ |
| if (IS_BASE64(ch) || ch == '-') { |
| *out++ = '-'; |
| } |
| *out++ = (char) ch; |
| } |
| else { |
| goto encode_char; |
| } |
| } |
| else { /* not in a shift sequence */ |
| if (ch == '+') { |
| *out++ = '+'; |
| *out++ = '-'; |
| } |
| else if (ENCODE_DIRECT(ch, !base64SetO, !base64WhiteSpace)) { |
| *out++ = (char) ch; |
| } |
| else { |
| *out++ = '+'; |
| inShift = 1; |
| goto encode_char; |
| } |
| } |
| continue; |
| encode_char: |
| #ifdef Py_UNICODE_WIDE |
| if (ch >= 0x10000) { |
| /* code first surrogate */ |
| base64bits += 16; |
| base64buffer = (base64buffer << 16) | 0xd800 | ((ch-0x10000) >> 10); |
| while (base64bits >= 6) { |
| *out++ = TO_BASE64(base64buffer >> (base64bits-6)); |
| base64bits -= 6; |
| } |
| /* prepare second surrogate */ |
| ch = 0xDC00 | ((ch-0x10000) & 0x3FF); |
| } |
| #endif |
| base64bits += 16; |
| base64buffer = (base64buffer << 16) | ch; |
| while (base64bits >= 6) { |
| *out++ = TO_BASE64(base64buffer >> (base64bits-6)); |
| base64bits -= 6; |
| } |
| } |
| if (base64bits) |
| *out++= TO_BASE64(base64buffer << (6-base64bits) ); |
| if (inShift) |
| *out++ = '-'; |
| if (_PyBytes_Resize(&v, out - start) < 0) |
| return NULL; |
| return v; |
| } |
| |
| #undef IS_BASE64 |
| #undef FROM_BASE64 |
| #undef TO_BASE64 |
| #undef DECODE_DIRECT |
| #undef ENCODE_DIRECT |
| |
| /* --- UTF-8 Codec -------------------------------------------------------- */ |
| |
| static |
| char utf8_code_length[256] = { |
| /* Map UTF-8 encoded prefix byte to sequence length. Zero means |
| illegal prefix. See RFC 3629 for details */ |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, /* 00-0F */ |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, /* 70-7F */ |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, /* 80-8F */ |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, /* B0-BF */ |
| 0, 0, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, /* C0-C1 + C2-CF */ |
| 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, /* D0-DF */ |
| 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, /* E0-EF */ |
| 4, 4, 4, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 /* F0-F4 + F5-FF */ |
| }; |
| |
| PyObject *PyUnicode_DecodeUTF8(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| return PyUnicode_DecodeUTF8Stateful(s, size, errors, NULL); |
| } |
| |
| /* Mask to check or force alignment of a pointer to C 'long' boundaries */ |
| #define LONG_PTR_MASK (size_t) (SIZEOF_LONG - 1) |
| |
| /* Mask to quickly check whether a C 'long' contains a |
| non-ASCII, UTF8-encoded char. */ |
| #if (SIZEOF_LONG == 8) |
| # define ASCII_CHAR_MASK 0x8080808080808080L |
| #elif (SIZEOF_LONG == 4) |
| # define ASCII_CHAR_MASK 0x80808080L |
| #else |
| # error C 'long' size should be either 4 or 8! |
| #endif |
| |
| PyObject *PyUnicode_DecodeUTF8Stateful(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| Py_ssize_t *consumed) |
| { |
| const char *starts = s; |
| int n; |
| int k; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| const char *e, *aligned_end; |
| PyUnicodeObject *unicode; |
| Py_UNICODE *p; |
| const char *errmsg = ""; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| /* Note: size will always be longer than the resulting Unicode |
| character count */ |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| if (size == 0) { |
| if (consumed) |
| *consumed = 0; |
| return (PyObject *)unicode; |
| } |
| |
| /* Unpack UTF-8 encoded data */ |
| p = unicode->str; |
| e = s + size; |
| aligned_end = (const char *) ((size_t) e & ~LONG_PTR_MASK); |
| |
| while (s < e) { |
| Py_UCS4 ch = (unsigned char)*s; |
| |
| if (ch < 0x80) { |
| /* Fast path for runs of ASCII characters. Given that common UTF-8 |
| input will consist of an overwhelming majority of ASCII |
| characters, we try to optimize for this case by checking |
| as many characters as a C 'long' can contain. |
| First, check if we can do an aligned read, as most CPUs have |
| a penalty for unaligned reads. |
| */ |
| if (!((size_t) s & LONG_PTR_MASK)) { |
| /* Help register allocation */ |
| register const char *_s = s; |
| register Py_UNICODE *_p = p; |
| while (_s < aligned_end) { |
| /* Read a whole long at a time (either 4 or 8 bytes), |
| and do a fast unrolled copy if it only contains ASCII |
| characters. */ |
| unsigned long data = *(unsigned long *) _s; |
| if (data & ASCII_CHAR_MASK) |
| break; |
| _p[0] = (unsigned char) _s[0]; |
| _p[1] = (unsigned char) _s[1]; |
| _p[2] = (unsigned char) _s[2]; |
| _p[3] = (unsigned char) _s[3]; |
| #if (SIZEOF_LONG == 8) |
| _p[4] = (unsigned char) _s[4]; |
| _p[5] = (unsigned char) _s[5]; |
| _p[6] = (unsigned char) _s[6]; |
| _p[7] = (unsigned char) _s[7]; |
| #endif |
| _s += SIZEOF_LONG; |
| _p += SIZEOF_LONG; |
| } |
| s = _s; |
| p = _p; |
| if (s == e) |
| break; |
| ch = (unsigned char)*s; |
| } |
| } |
| |
| if (ch < 0x80) { |
| *p++ = (Py_UNICODE)ch; |
| s++; |
| continue; |
| } |
| |
| n = utf8_code_length[ch]; |
| |
| if (s + n > e) { |
| if (consumed) |
| break; |
| else { |
| errmsg = "unexpected end of data"; |
| startinpos = s-starts; |
| endinpos = startinpos+1; |
| for (k=1; (k < size-startinpos) && ((s[k]&0xC0) == 0x80); k++) |
| endinpos++; |
| goto utf8Error; |
| } |
| } |
| |
| switch (n) { |
| |
| case 0: |
| errmsg = "invalid start byte"; |
| startinpos = s-starts; |
| endinpos = startinpos+1; |
| goto utf8Error; |
| |
| case 1: |
| errmsg = "internal error"; |
| startinpos = s-starts; |
| endinpos = startinpos+1; |
| goto utf8Error; |
| |
| case 2: |
| if ((s[1] & 0xc0) != 0x80) { |
| errmsg = "invalid continuation byte"; |
| startinpos = s-starts; |
| endinpos = startinpos + 1; |
| goto utf8Error; |
| } |
| ch = ((s[0] & 0x1f) << 6) + (s[1] & 0x3f); |
| assert ((ch > 0x007F) && (ch <= 0x07FF)); |
| *p++ = (Py_UNICODE)ch; |
| break; |
| |
| case 3: |
| /* Decoding UTF-8 sequences in range \xed\xa0\x80-\xed\xbf\xbf |
| will result in surrogates in range d800-dfff. Surrogates are |
| not valid UTF-8 so they are rejected. |
| See http://www.unicode.org/versions/Unicode5.2.0/ch03.pdf |
| (table 3-7) and http://www.rfc-editor.org/rfc/rfc3629.txt */ |
| if ((s[1] & 0xc0) != 0x80 || |
| (s[2] & 0xc0) != 0x80 || |
| ((unsigned char)s[0] == 0xE0 && |
| (unsigned char)s[1] < 0xA0) || |
| ((unsigned char)s[0] == 0xED && |
| (unsigned char)s[1] > 0x9F)) { |
| errmsg = "invalid continuation byte"; |
| startinpos = s-starts; |
| endinpos = startinpos + 1; |
| |
| /* if s[1] first two bits are 1 and 0, then the invalid |
| continuation byte is s[2], so increment endinpos by 1, |
| if not, s[1] is invalid and endinpos doesn't need to |
| be incremented. */ |
| if ((s[1] & 0xC0) == 0x80) |
| endinpos++; |
| goto utf8Error; |
| } |
| ch = ((s[0] & 0x0f) << 12) + ((s[1] & 0x3f) << 6) + (s[2] & 0x3f); |
| assert ((ch > 0x07FF) && (ch <= 0xFFFF)); |
| *p++ = (Py_UNICODE)ch; |
| break; |
| |
| case 4: |
| if ((s[1] & 0xc0) != 0x80 || |
| (s[2] & 0xc0) != 0x80 || |
| (s[3] & 0xc0) != 0x80 || |
| ((unsigned char)s[0] == 0xF0 && |
| (unsigned char)s[1] < 0x90) || |
| ((unsigned char)s[0] == 0xF4 && |
| (unsigned char)s[1] > 0x8F)) { |
| errmsg = "invalid continuation byte"; |
| startinpos = s-starts; |
| endinpos = startinpos + 1; |
| if ((s[1] & 0xC0) == 0x80) { |
| endinpos++; |
| if ((s[2] & 0xC0) == 0x80) |
| endinpos++; |
| } |
| goto utf8Error; |
| } |
| ch = ((s[0] & 0x7) << 18) + ((s[1] & 0x3f) << 12) + |
| ((s[2] & 0x3f) << 6) + (s[3] & 0x3f); |
| assert ((ch > 0xFFFF) && (ch <= 0x10ffff)); |
| |
| #ifdef Py_UNICODE_WIDE |
| *p++ = (Py_UNICODE)ch; |
| #else |
| /* compute and append the two surrogates: */ |
| |
| /* translate from 10000..10FFFF to 0..FFFF */ |
| ch -= 0x10000; |
| |
| /* high surrogate = top 10 bits added to D800 */ |
| *p++ = (Py_UNICODE)(0xD800 + (ch >> 10)); |
| |
| /* low surrogate = bottom 10 bits added to DC00 */ |
| *p++ = (Py_UNICODE)(0xDC00 + (ch & 0x03FF)); |
| #endif |
| break; |
| } |
| s += n; |
| continue; |
| |
| utf8Error: |
| outpos = p-PyUnicode_AS_UNICODE(unicode); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "utf-8", errmsg, |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &unicode, &outpos, &p)) |
| goto onError; |
| aligned_end = (const char *) ((size_t) e & ~LONG_PTR_MASK); |
| } |
| if (consumed) |
| *consumed = s-starts; |
| |
| /* Adjust length */ |
| if (_PyUnicode_Resize(&unicode, p - unicode->str) < 0) |
| goto onError; |
| |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)unicode; |
| |
| onError: |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| Py_DECREF(unicode); |
| return NULL; |
| } |
| |
| #undef ASCII_CHAR_MASK |
| |
| #ifdef __APPLE__ |
| |
| /* Simplified UTF-8 decoder using surrogateescape error handler, |
| used to decode the command line arguments on Mac OS X. |
| |
| Return a pointer to a newly allocated wide character string (use |
| PyMem_Free() to free the memory), or NULL on memory allocation error. */ |
| |
| wchar_t* |
| _Py_DecodeUTF8_surrogateescape(const char *s, Py_ssize_t size) |
| { |
| int n; |
| const char *e; |
| wchar_t *unicode, *p; |
| |
| /* Note: size will always be longer than the resulting Unicode |
| character count */ |
| if (PY_SSIZE_T_MAX / sizeof(wchar_t) < (size + 1)) |
| return NULL; |
| unicode = PyMem_Malloc((size + 1) * sizeof(wchar_t)); |
| if (!unicode) |
| return NULL; |
| |
| /* Unpack UTF-8 encoded data */ |
| p = unicode; |
| e = s + size; |
| while (s < e) { |
| Py_UCS4 ch = (unsigned char)*s; |
| |
| if (ch < 0x80) { |
| *p++ = (wchar_t)ch; |
| s++; |
| continue; |
| } |
| |
| n = utf8_code_length[ch]; |
| if (s + n > e) { |
| goto surrogateescape; |
| } |
| |
| switch (n) { |
| case 0: |
| case 1: |
| goto surrogateescape; |
| |
| case 2: |
| if ((s[1] & 0xc0) != 0x80) |
| goto surrogateescape; |
| ch = ((s[0] & 0x1f) << 6) + (s[1] & 0x3f); |
| assert ((ch > 0x007F) && (ch <= 0x07FF)); |
| *p++ = (wchar_t)ch; |
| break; |
| |
| case 3: |
| /* Decoding UTF-8 sequences in range \xed\xa0\x80-\xed\xbf\xbf |
| will result in surrogates in range d800-dfff. Surrogates are |
| not valid UTF-8 so they are rejected. |
| See http://www.unicode.org/versions/Unicode5.2.0/ch03.pdf |
| (table 3-7) and http://www.rfc-editor.org/rfc/rfc3629.txt */ |
| if ((s[1] & 0xc0) != 0x80 || |
| (s[2] & 0xc0) != 0x80 || |
| ((unsigned char)s[0] == 0xE0 && |
| (unsigned char)s[1] < 0xA0) || |
| ((unsigned char)s[0] == 0xED && |
| (unsigned char)s[1] > 0x9F)) { |
| |
| goto surrogateescape; |
| } |
| ch = ((s[0] & 0x0f) << 12) + ((s[1] & 0x3f) << 6) + (s[2] & 0x3f); |
| assert ((ch > 0x07FF) && (ch <= 0xFFFF)); |
| *p++ = (Py_UNICODE)ch; |
| break; |
| |
| case 4: |
| if ((s[1] & 0xc0) != 0x80 || |
| (s[2] & 0xc0) != 0x80 || |
| (s[3] & 0xc0) != 0x80 || |
| ((unsigned char)s[0] == 0xF0 && |
| (unsigned char)s[1] < 0x90) || |
| ((unsigned char)s[0] == 0xF4 && |
| (unsigned char)s[1] > 0x8F)) { |
| goto surrogateescape; |
| } |
| ch = ((s[0] & 0x7) << 18) + ((s[1] & 0x3f) << 12) + |
| ((s[2] & 0x3f) << 6) + (s[3] & 0x3f); |
| assert ((ch > 0xFFFF) && (ch <= 0x10ffff)); |
| |
| #if SIZEOF_WCHAR_T == 4 |
| *p++ = (wchar_t)ch; |
| #else |
| /* compute and append the two surrogates: */ |
| |
| /* translate from 10000..10FFFF to 0..FFFF */ |
| ch -= 0x10000; |
| |
| /* high surrogate = top 10 bits added to D800 */ |
| *p++ = (wchar_t)(0xD800 + (ch >> 10)); |
| |
| /* low surrogate = bottom 10 bits added to DC00 */ |
| *p++ = (wchar_t)(0xDC00 + (ch & 0x03FF)); |
| #endif |
| break; |
| } |
| s += n; |
| continue; |
| |
| surrogateescape: |
| *p++ = 0xDC00 + ch; |
| s++; |
| } |
| *p = L'\0'; |
| return unicode; |
| } |
| |
| #endif /* __APPLE__ */ |
| |
| /* Allocation strategy: if the string is short, convert into a stack buffer |
| and allocate exactly as much space needed at the end. Else allocate the |
| maximum possible needed (4 result bytes per Unicode character), and return |
| the excess memory at the end. |
| */ |
| PyObject * |
| PyUnicode_EncodeUTF8(const Py_UNICODE *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| #define MAX_SHORT_UNICHARS 300 /* largest size we'll do on the stack */ |
| |
| Py_ssize_t i; /* index into s of next input byte */ |
| PyObject *result; /* result string object */ |
| char *p; /* next free byte in output buffer */ |
| Py_ssize_t nallocated; /* number of result bytes allocated */ |
| Py_ssize_t nneeded; /* number of result bytes needed */ |
| char stackbuf[MAX_SHORT_UNICHARS * 4]; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| assert(s != NULL); |
| assert(size >= 0); |
| |
| if (size <= MAX_SHORT_UNICHARS) { |
| /* Write into the stack buffer; nallocated can't overflow. |
| * At the end, we'll allocate exactly as much heap space as it |
| * turns out we need. |
| */ |
| nallocated = Py_SAFE_DOWNCAST(sizeof(stackbuf), size_t, int); |
| result = NULL; /* will allocate after we're done */ |
| p = stackbuf; |
| } |
| else { |
| /* Overallocate on the heap, and give the excess back at the end. */ |
| nallocated = size * 4; |
| if (nallocated / 4 != size) /* overflow! */ |
| return PyErr_NoMemory(); |
| result = PyBytes_FromStringAndSize(NULL, nallocated); |
| if (result == NULL) |
| return NULL; |
| p = PyBytes_AS_STRING(result); |
| } |
| |
| for (i = 0; i < size;) { |
| Py_UCS4 ch = s[i++]; |
| |
| if (ch < 0x80) |
| /* Encode ASCII */ |
| *p++ = (char) ch; |
| |
| else if (ch < 0x0800) { |
| /* Encode Latin-1 */ |
| *p++ = (char)(0xc0 | (ch >> 6)); |
| *p++ = (char)(0x80 | (ch & 0x3f)); |
| } else if (0xD800 <= ch && ch <= 0xDFFF) { |
| #ifndef Py_UNICODE_WIDE |
| /* Special case: check for high and low surrogate */ |
| if (ch <= 0xDBFF && i != size && 0xDC00 <= s[i] && s[i] <= 0xDFFF) { |
| Py_UCS4 ch2 = s[i]; |
| /* Combine the two surrogates to form a UCS4 value */ |
| ch = ((ch - 0xD800) << 10 | (ch2 - 0xDC00)) + 0x10000; |
| i++; |
| |
| /* Encode UCS4 Unicode ordinals */ |
| *p++ = (char)(0xf0 | (ch >> 18)); |
| *p++ = (char)(0x80 | ((ch >> 12) & 0x3f)); |
| *p++ = (char)(0x80 | ((ch >> 6) & 0x3f)); |
| *p++ = (char)(0x80 | (ch & 0x3f)); |
| } else { |
| #endif |
| Py_ssize_t newpos; |
| PyObject *rep; |
| Py_ssize_t repsize, k; |
| rep = unicode_encode_call_errorhandler |
| (errors, &errorHandler, "utf-8", "surrogates not allowed", |
| s, size, &exc, i-1, i, &newpos); |
| if (!rep) |
| goto error; |
| |
| if (PyBytes_Check(rep)) |
| repsize = PyBytes_GET_SIZE(rep); |
| else |
| repsize = PyUnicode_GET_SIZE(rep); |
| |
| if (repsize > 4) { |
| Py_ssize_t offset; |
| |
| if (result == NULL) |
| offset = p - stackbuf; |
| else |
| offset = p - PyBytes_AS_STRING(result); |
| |
| if (nallocated > PY_SSIZE_T_MAX - repsize + 4) { |
| /* integer overflow */ |
| PyErr_NoMemory(); |
| goto error; |
| } |
| nallocated += repsize - 4; |
| if (result != NULL) { |
| if (_PyBytes_Resize(&result, nallocated) < 0) |
| goto error; |
| } else { |
| result = PyBytes_FromStringAndSize(NULL, nallocated); |
| if (result == NULL) |
| goto error; |
| Py_MEMCPY(PyBytes_AS_STRING(result), stackbuf, offset); |
| } |
| p = PyBytes_AS_STRING(result) + offset; |
| } |
| |
| if (PyBytes_Check(rep)) { |
| char *prep = PyBytes_AS_STRING(rep); |
| for(k = repsize; k > 0; k--) |
| *p++ = *prep++; |
| } else /* rep is unicode */ { |
| Py_UNICODE *prep = PyUnicode_AS_UNICODE(rep); |
| Py_UNICODE c; |
| |
| for(k=0; k<repsize; k++) { |
| c = prep[k]; |
| if (0x80 <= c) { |
| raise_encode_exception(&exc, "utf-8", s, size, |
| i-1, i, "surrogates not allowed"); |
| goto error; |
| } |
| *p++ = (char)prep[k]; |
| } |
| } |
| Py_DECREF(rep); |
| #ifndef Py_UNICODE_WIDE |
| } |
| #endif |
| } else if (ch < 0x10000) { |
| *p++ = (char)(0xe0 | (ch >> 12)); |
| *p++ = (char)(0x80 | ((ch >> 6) & 0x3f)); |
| *p++ = (char)(0x80 | (ch & 0x3f)); |
| } else /* ch >= 0x10000 */ { |
| /* Encode UCS4 Unicode ordinals */ |
| *p++ = (char)(0xf0 | (ch >> 18)); |
| *p++ = (char)(0x80 | ((ch >> 12) & 0x3f)); |
| *p++ = (char)(0x80 | ((ch >> 6) & 0x3f)); |
| *p++ = (char)(0x80 | (ch & 0x3f)); |
| } |
| } |
| |
| if (result == NULL) { |
| /* This was stack allocated. */ |
| nneeded = p - stackbuf; |
| assert(nneeded <= nallocated); |
| result = PyBytes_FromStringAndSize(stackbuf, nneeded); |
| } |
| else { |
| /* Cut back to size actually needed. */ |
| nneeded = p - PyBytes_AS_STRING(result); |
| assert(nneeded <= nallocated); |
| _PyBytes_Resize(&result, nneeded); |
| } |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return result; |
| error: |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| Py_XDECREF(result); |
| return NULL; |
| |
| #undef MAX_SHORT_UNICHARS |
| } |
| |
| PyObject *PyUnicode_AsUTF8String(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeUTF8(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| } |
| |
| /* --- UTF-32 Codec ------------------------------------------------------- */ |
| |
| PyObject * |
| PyUnicode_DecodeUTF32(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| int *byteorder) |
| { |
| return PyUnicode_DecodeUTF32Stateful(s, size, errors, byteorder, NULL); |
| } |
| |
| PyObject * |
| PyUnicode_DecodeUTF32Stateful(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| int *byteorder, |
| Py_ssize_t *consumed) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| PyUnicodeObject *unicode; |
| Py_UNICODE *p; |
| #ifndef Py_UNICODE_WIDE |
| int pairs = 0; |
| const unsigned char *qq; |
| #else |
| const int pairs = 0; |
| #endif |
| const unsigned char *q, *e; |
| int bo = 0; /* assume native ordering by default */ |
| const char *errmsg = ""; |
| /* Offsets from q for retrieving bytes in the right order. */ |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| int iorder[] = {0, 1, 2, 3}; |
| #else |
| int iorder[] = {3, 2, 1, 0}; |
| #endif |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| q = (unsigned char *)s; |
| e = q + size; |
| |
| if (byteorder) |
| bo = *byteorder; |
| |
| /* Check for BOM marks (U+FEFF) in the input and adjust current |
| byte order setting accordingly. In native mode, the leading BOM |
| mark is skipped, in all other modes, it is copied to the output |
| stream as-is (giving a ZWNBSP character). */ |
| if (bo == 0) { |
| if (size >= 4) { |
| const Py_UCS4 bom = (q[iorder[3]] << 24) | (q[iorder[2]] << 16) | |
| (q[iorder[1]] << 8) | q[iorder[0]]; |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| if (bom == 0x0000FEFF) { |
| q += 4; |
| bo = -1; |
| } |
| else if (bom == 0xFFFE0000) { |
| q += 4; |
| bo = 1; |
| } |
| #else |
| if (bom == 0x0000FEFF) { |
| q += 4; |
| bo = 1; |
| } |
| else if (bom == 0xFFFE0000) { |
| q += 4; |
| bo = -1; |
| } |
| #endif |
| } |
| } |
| |
| if (bo == -1) { |
| /* force LE */ |
| iorder[0] = 0; |
| iorder[1] = 1; |
| iorder[2] = 2; |
| iorder[3] = 3; |
| } |
| else if (bo == 1) { |
| /* force BE */ |
| iorder[0] = 3; |
| iorder[1] = 2; |
| iorder[2] = 1; |
| iorder[3] = 0; |
| } |
| |
| /* On narrow builds we split characters outside the BMP into two |
| codepoints => count how much extra space we need. */ |
| #ifndef Py_UNICODE_WIDE |
| for (qq = q; e - qq >= 4; qq += 4) |
| if (qq[iorder[2]] != 0 || qq[iorder[3]] != 0) |
| pairs++; |
| #endif |
| |
| /* This might be one to much, because of a BOM */ |
| unicode = _PyUnicode_New((size+3)/4+pairs); |
| if (!unicode) |
| return NULL; |
| if (size == 0) |
| return (PyObject *)unicode; |
| |
| /* Unpack UTF-32 encoded data */ |
| p = unicode->str; |
| |
| while (q < e) { |
| Py_UCS4 ch; |
| /* remaining bytes at the end? (size should be divisible by 4) */ |
| if (e-q<4) { |
| if (consumed) |
| break; |
| errmsg = "truncated data"; |
| startinpos = ((const char *)q)-starts; |
| endinpos = ((const char *)e)-starts; |
| goto utf32Error; |
| /* The remaining input chars are ignored if the callback |
| chooses to skip the input */ |
| } |
| ch = (q[iorder[3]] << 24) | (q[iorder[2]] << 16) | |
| (q[iorder[1]] << 8) | q[iorder[0]]; |
| |
| if (ch >= 0x110000) |
| { |
| errmsg = "codepoint not in range(0x110000)"; |
| startinpos = ((const char *)q)-starts; |
| endinpos = startinpos+4; |
| goto utf32Error; |
| } |
| #ifndef Py_UNICODE_WIDE |
| if (ch >= 0x10000) |
| { |
| *p++ = 0xD800 | ((ch-0x10000) >> 10); |
| *p++ = 0xDC00 | ((ch-0x10000) & 0x3FF); |
| } |
| else |
| #endif |
| *p++ = ch; |
| q += 4; |
| continue; |
| utf32Error: |
| outpos = p-PyUnicode_AS_UNICODE(unicode); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "utf32", errmsg, |
| &starts, (const char **)&e, &startinpos, &endinpos, &exc, (const char **)&q, |
| &unicode, &outpos, &p)) |
| goto onError; |
| } |
| |
| if (byteorder) |
| *byteorder = bo; |
| |
| if (consumed) |
| *consumed = (const char *)q-starts; |
| |
| /* Adjust length */ |
| if (_PyUnicode_Resize(&unicode, p - unicode->str) < 0) |
| goto onError; |
| |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)unicode; |
| |
| onError: |
| Py_DECREF(unicode); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| PyObject * |
| PyUnicode_EncodeUTF32(const Py_UNICODE *s, |
| Py_ssize_t size, |
| const char *errors, |
| int byteorder) |
| { |
| PyObject *v; |
| unsigned char *p; |
| Py_ssize_t nsize, bytesize; |
| #ifndef Py_UNICODE_WIDE |
| Py_ssize_t i, pairs; |
| #else |
| const int pairs = 0; |
| #endif |
| /* Offsets from p for storing byte pairs in the right order. */ |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| int iorder[] = {0, 1, 2, 3}; |
| #else |
| int iorder[] = {3, 2, 1, 0}; |
| #endif |
| |
| #define STORECHAR(CH) \ |
| do { \ |
| p[iorder[3]] = ((CH) >> 24) & 0xff; \ |
| p[iorder[2]] = ((CH) >> 16) & 0xff; \ |
| p[iorder[1]] = ((CH) >> 8) & 0xff; \ |
| p[iorder[0]] = (CH) & 0xff; \ |
| p += 4; \ |
| } while(0) |
| |
| /* In narrow builds we can output surrogate pairs as one codepoint, |
| so we need less space. */ |
| #ifndef Py_UNICODE_WIDE |
| for (i = pairs = 0; i < size-1; i++) |
| if (0xD800 <= s[i] && s[i] <= 0xDBFF && |
| 0xDC00 <= s[i+1] && s[i+1] <= 0xDFFF) |
| pairs++; |
| #endif |
| nsize = (size - pairs + (byteorder == 0)); |
| bytesize = nsize * 4; |
| if (bytesize / 4 != nsize) |
| return PyErr_NoMemory(); |
| v = PyBytes_FromStringAndSize(NULL, bytesize); |
| if (v == NULL) |
| return NULL; |
| |
| p = (unsigned char *)PyBytes_AS_STRING(v); |
| if (byteorder == 0) |
| STORECHAR(0xFEFF); |
| if (size == 0) |
| goto done; |
| |
| if (byteorder == -1) { |
| /* force LE */ |
| iorder[0] = 0; |
| iorder[1] = 1; |
| iorder[2] = 2; |
| iorder[3] = 3; |
| } |
| else if (byteorder == 1) { |
| /* force BE */ |
| iorder[0] = 3; |
| iorder[1] = 2; |
| iorder[2] = 1; |
| iorder[3] = 0; |
| } |
| |
| while (size-- > 0) { |
| Py_UCS4 ch = *s++; |
| #ifndef Py_UNICODE_WIDE |
| if (0xD800 <= ch && ch <= 0xDBFF && size > 0) { |
| Py_UCS4 ch2 = *s; |
| if (0xDC00 <= ch2 && ch2 <= 0xDFFF) { |
| ch = (((ch & 0x3FF)<<10) | (ch2 & 0x3FF)) + 0x10000; |
| s++; |
| size--; |
| } |
| } |
| #endif |
| STORECHAR(ch); |
| } |
| |
| done: |
| return v; |
| #undef STORECHAR |
| } |
| |
| PyObject *PyUnicode_AsUTF32String(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeUTF32(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL, |
| 0); |
| } |
| |
| /* --- UTF-16 Codec ------------------------------------------------------- */ |
| |
| PyObject * |
| PyUnicode_DecodeUTF16(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| int *byteorder) |
| { |
| return PyUnicode_DecodeUTF16Stateful(s, size, errors, byteorder, NULL); |
| } |
| |
| /* Two masks for fast checking of whether a C 'long' may contain |
| UTF16-encoded surrogate characters. This is an efficient heuristic, |
| assuming that non-surrogate characters with a code point >= 0x8000 are |
| rare in most input. |
| FAST_CHAR_MASK is used when the input is in native byte ordering, |
| SWAPPED_FAST_CHAR_MASK when the input is in byteswapped ordering. |
| */ |
| #if (SIZEOF_LONG == 8) |
| # define FAST_CHAR_MASK 0x8000800080008000L |
| # define SWAPPED_FAST_CHAR_MASK 0x0080008000800080L |
| #elif (SIZEOF_LONG == 4) |
| # define FAST_CHAR_MASK 0x80008000L |
| # define SWAPPED_FAST_CHAR_MASK 0x00800080L |
| #else |
| # error C 'long' size should be either 4 or 8! |
| #endif |
| |
| PyObject * |
| PyUnicode_DecodeUTF16Stateful(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| int *byteorder, |
| Py_ssize_t *consumed) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| PyUnicodeObject *unicode; |
| Py_UNICODE *p; |
| const unsigned char *q, *e, *aligned_end; |
| int bo = 0; /* assume native ordering by default */ |
| int native_ordering = 0; |
| const char *errmsg = ""; |
| /* Offsets from q for retrieving byte pairs in the right order. */ |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| int ihi = 1, ilo = 0; |
| #else |
| int ihi = 0, ilo = 1; |
| #endif |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| /* Note: size will always be longer than the resulting Unicode |
| character count */ |
| unicode = _PyUnicode_New(size); |
| if (!unicode) |
| return NULL; |
| if (size == 0) |
| return (PyObject *)unicode; |
| |
| /* Unpack UTF-16 encoded data */ |
| p = unicode->str; |
| q = (unsigned char *)s; |
| e = q + size; |
| |
| if (byteorder) |
| bo = *byteorder; |
| |
| /* Check for BOM marks (U+FEFF) in the input and adjust current |
| byte order setting accordingly. In native mode, the leading BOM |
| mark is skipped, in all other modes, it is copied to the output |
| stream as-is (giving a ZWNBSP character). */ |
| if (bo == 0) { |
| if (size >= 2) { |
| const Py_UNICODE bom = (q[ihi] << 8) | q[ilo]; |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| if (bom == 0xFEFF) { |
| q += 2; |
| bo = -1; |
| } |
| else if (bom == 0xFFFE) { |
| q += 2; |
| bo = 1; |
| } |
| #else |
| if (bom == 0xFEFF) { |
| q += 2; |
| bo = 1; |
| } |
| else if (bom == 0xFFFE) { |
| q += 2; |
| bo = -1; |
| } |
| #endif |
| } |
| } |
| |
| if (bo == -1) { |
| /* force LE */ |
| ihi = 1; |
| ilo = 0; |
| } |
| else if (bo == 1) { |
| /* force BE */ |
| ihi = 0; |
| ilo = 1; |
| } |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| native_ordering = ilo < ihi; |
| #else |
| native_ordering = ilo > ihi; |
| #endif |
| |
| aligned_end = (const unsigned char *) ((size_t) e & ~LONG_PTR_MASK); |
| while (1) { |
| Py_UNICODE ch; |
| if (e - q < 2) { |
| /* remaining byte at the end? (size should be even) */ |
| if (q == e || consumed) |
| break; |
| errmsg = "truncated data"; |
| startinpos = ((const char *)q) - starts; |
| endinpos = ((const char *)e) - starts; |
| outpos = p - PyUnicode_AS_UNICODE(unicode); |
| goto utf16Error; |
| /* The remaining input chars are ignored if the callback |
| chooses to skip the input */ |
| } |
| /* First check for possible aligned read of a C 'long'. Unaligned |
| reads are more expensive, better to defer to another iteration. */ |
| if (!((size_t) q & LONG_PTR_MASK)) { |
| /* Fast path for runs of non-surrogate chars. */ |
| register const unsigned char *_q = q; |
| Py_UNICODE *_p = p; |
| if (native_ordering) { |
| /* Native ordering is simple: as long as the input cannot |
| possibly contain a surrogate char, do an unrolled copy |
| of several 16-bit code points to the target object. |
| The non-surrogate check is done on several input bytes |
| at a time (as many as a C 'long' can contain). */ |
| while (_q < aligned_end) { |
| unsigned long data = * (unsigned long *) _q; |
| if (data & FAST_CHAR_MASK) |
| break; |
| _p[0] = ((unsigned short *) _q)[0]; |
| _p[1] = ((unsigned short *) _q)[1]; |
| #if (SIZEOF_LONG == 8) |
| _p[2] = ((unsigned short *) _q)[2]; |
| _p[3] = ((unsigned short *) _q)[3]; |
| #endif |
| _q += SIZEOF_LONG; |
| _p += SIZEOF_LONG / 2; |
| } |
| } |
| else { |
| /* Byteswapped ordering is similar, but we must decompose |
| the copy bytewise, and take care of zero'ing out the |
| upper bytes if the target object is in 32-bit units |
| (that is, in UCS-4 builds). */ |
| while (_q < aligned_end) { |
| unsigned long data = * (unsigned long *) _q; |
| if (data & SWAPPED_FAST_CHAR_MASK) |
| break; |
| /* Zero upper bytes in UCS-4 builds */ |
| #if (Py_UNICODE_SIZE > 2) |
| _p[0] = 0; |
| _p[1] = 0; |
| #if (SIZEOF_LONG == 8) |
| _p[2] = 0; |
| _p[3] = 0; |
| #endif |
| #endif |
| /* Issue #4916; UCS-4 builds on big endian machines must |
| fill the two last bytes of each 4-byte unit. */ |
| #if (!defined(BYTEORDER_IS_LITTLE_ENDIAN) && Py_UNICODE_SIZE > 2) |
| # define OFF 2 |
| #else |
| # define OFF 0 |
| #endif |
| ((unsigned char *) _p)[OFF + 1] = _q[0]; |
| ((unsigned char *) _p)[OFF + 0] = _q[1]; |
| ((unsigned char *) _p)[OFF + 1 + Py_UNICODE_SIZE] = _q[2]; |
| ((unsigned char *) _p)[OFF + 0 + Py_UNICODE_SIZE] = _q[3]; |
| #if (SIZEOF_LONG == 8) |
| ((unsigned char *) _p)[OFF + 1 + 2 * Py_UNICODE_SIZE] = _q[4]; |
| ((unsigned char *) _p)[OFF + 0 + 2 * Py_UNICODE_SIZE] = _q[5]; |
| ((unsigned char *) _p)[OFF + 1 + 3 * Py_UNICODE_SIZE] = _q[6]; |
| ((unsigned char *) _p)[OFF + 0 + 3 * Py_UNICODE_SIZE] = _q[7]; |
| #endif |
| #undef OFF |
| _q += SIZEOF_LONG; |
| _p += SIZEOF_LONG / 2; |
| } |
| } |
| p = _p; |
| q = _q; |
| if (e - q < 2) |
| continue; |
| } |
| ch = (q[ihi] << 8) | q[ilo]; |
| |
| q += 2; |
| |
| if (ch < 0xD800 || ch > 0xDFFF) { |
| *p++ = ch; |
| continue; |
| } |
| |
| /* UTF-16 code pair: */ |
| if (e - q < 2) { |
| q -= 2; |
| if (consumed) |
| break; |
| errmsg = "unexpected end of data"; |
| startinpos = ((const char *)q) - starts; |
| endinpos = ((const char *)e) - starts; |
| goto utf16Error; |
| } |
| if (0xD800 <= ch && ch <= 0xDBFF) { |
| Py_UNICODE ch2 = (q[ihi] << 8) | q[ilo]; |
| q += 2; |
| if (0xDC00 <= ch2 && ch2 <= 0xDFFF) { |
| #ifndef Py_UNICODE_WIDE |
| *p++ = ch; |
| *p++ = ch2; |
| #else |
| *p++ = (((ch & 0x3FF)<<10) | (ch2 & 0x3FF)) + 0x10000; |
| #endif |
| continue; |
| } |
| else { |
| errmsg = "illegal UTF-16 surrogate"; |
| startinpos = (((const char *)q)-4)-starts; |
| endinpos = startinpos+2; |
| goto utf16Error; |
| } |
| |
| } |
| errmsg = "illegal encoding"; |
| startinpos = (((const char *)q)-2)-starts; |
| endinpos = startinpos+2; |
| /* Fall through to report the error */ |
| |
| utf16Error: |
| outpos = p - PyUnicode_AS_UNICODE(unicode); |
| if (unicode_decode_call_errorhandler( |
| errors, |
| &errorHandler, |
| "utf16", errmsg, |
| &starts, |
| (const char **)&e, |
| &startinpos, |
| &endinpos, |
| &exc, |
| (const char **)&q, |
| &unicode, |
| &outpos, |
| &p)) |
| goto onError; |
| /* Update data because unicode_decode_call_errorhandler might have |
| changed the input object. */ |
| aligned_end = (const unsigned char *) ((size_t) e & ~LONG_PTR_MASK); |
| } |
| |
| if (byteorder) |
| *byteorder = bo; |
| |
| if (consumed) |
| *consumed = (const char *)q-starts; |
| |
| /* Adjust length */ |
| if (_PyUnicode_Resize(&unicode, p - unicode->str) < 0) |
| goto onError; |
| |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)unicode; |
| |
| onError: |
| Py_DECREF(unicode); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| #undef FAST_CHAR_MASK |
| #undef SWAPPED_FAST_CHAR_MASK |
| |
| PyObject * |
| PyUnicode_EncodeUTF16(const Py_UNICODE *s, |
| Py_ssize_t size, |
| const char *errors, |
| int byteorder) |
| { |
| PyObject *v; |
| unsigned char *p; |
| Py_ssize_t nsize, bytesize; |
| #ifdef Py_UNICODE_WIDE |
| Py_ssize_t i, pairs; |
| #else |
| const int pairs = 0; |
| #endif |
| /* Offsets from p for storing byte pairs in the right order. */ |
| #ifdef BYTEORDER_IS_LITTLE_ENDIAN |
| int ihi = 1, ilo = 0; |
| #else |
| int ihi = 0, ilo = 1; |
| #endif |
| |
| #define STORECHAR(CH) \ |
| do { \ |
| p[ihi] = ((CH) >> 8) & 0xff; \ |
| p[ilo] = (CH) & 0xff; \ |
| p += 2; \ |
| } while(0) |
| |
| #ifdef Py_UNICODE_WIDE |
| for (i = pairs = 0; i < size; i++) |
| if (s[i] >= 0x10000) |
| pairs++; |
| #endif |
| /* 2 * (size + pairs + (byteorder == 0)) */ |
| if (size > PY_SSIZE_T_MAX || |
| size > PY_SSIZE_T_MAX - pairs - (byteorder == 0)) |
| return PyErr_NoMemory(); |
| nsize = size + pairs + (byteorder == 0); |
| bytesize = nsize * 2; |
| if (bytesize / 2 != nsize) |
| return PyErr_NoMemory(); |
| v = PyBytes_FromStringAndSize(NULL, bytesize); |
| if (v == NULL) |
| return NULL; |
| |
| p = (unsigned char *)PyBytes_AS_STRING(v); |
| if (byteorder == 0) |
| STORECHAR(0xFEFF); |
| if (size == 0) |
| goto done; |
| |
| if (byteorder == -1) { |
| /* force LE */ |
| ihi = 1; |
| ilo = 0; |
| } |
| else if (byteorder == 1) { |
| /* force BE */ |
| ihi = 0; |
| ilo = 1; |
| } |
| |
| while (size-- > 0) { |
| Py_UNICODE ch = *s++; |
| Py_UNICODE ch2 = 0; |
| #ifdef Py_UNICODE_WIDE |
| if (ch >= 0x10000) { |
| ch2 = 0xDC00 | ((ch-0x10000) & 0x3FF); |
| ch = 0xD800 | ((ch-0x10000) >> 10); |
| } |
| #endif |
| STORECHAR(ch); |
| if (ch2) |
| STORECHAR(ch2); |
| } |
| |
| done: |
| return v; |
| #undef STORECHAR |
| } |
| |
| PyObject *PyUnicode_AsUTF16String(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeUTF16(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL, |
| 0); |
| } |
| |
| /* --- Unicode Escape Codec ----------------------------------------------- */ |
| |
| static _PyUnicode_Name_CAPI *ucnhash_CAPI = NULL; |
| |
| PyObject *PyUnicode_DecodeUnicodeEscape(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| const char *end; |
| char* message; |
| Py_UCS4 chr = 0xffffffff; /* in case 'getcode' messes up */ |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| /* Escaped strings will always be longer than the resulting |
| Unicode string, so we start with size here and then reduce the |
| length after conversion to the true value. |
| (but if the error callback returns a long replacement string |
| we'll have to allocate more space) */ |
| v = _PyUnicode_New(size); |
| if (v == NULL) |
| goto onError; |
| if (size == 0) |
| return (PyObject *)v; |
| |
| p = PyUnicode_AS_UNICODE(v); |
| end = s + size; |
| |
| while (s < end) { |
| unsigned char c; |
| Py_UNICODE x; |
| int digits; |
| |
| /* Non-escape characters are interpreted as Unicode ordinals */ |
| if (*s != '\\') { |
| *p++ = (unsigned char) *s++; |
| continue; |
| } |
| |
| startinpos = s-starts; |
| /* \ - Escapes */ |
| s++; |
| c = *s++; |
| if (s > end) |
| c = '\0'; /* Invalid after \ */ |
| switch (c) { |
| |
| /* \x escapes */ |
| case '\n': break; |
| case '\\': *p++ = '\\'; break; |
| case '\'': *p++ = '\''; break; |
| case '\"': *p++ = '\"'; break; |
| case 'b': *p++ = '\b'; break; |
| case 'f': *p++ = '\014'; break; /* FF */ |
| case 't': *p++ = '\t'; break; |
| case 'n': *p++ = '\n'; break; |
| case 'r': *p++ = '\r'; break; |
| case 'v': *p++ = '\013'; break; /* VT */ |
| case 'a': *p++ = '\007'; break; /* BEL, not classic C */ |
| |
| /* \OOO (octal) escapes */ |
| case '0': case '1': case '2': case '3': |
| case '4': case '5': case '6': case '7': |
| x = s[-1] - '0'; |
| if (s < end && '0' <= *s && *s <= '7') { |
| x = (x<<3) + *s++ - '0'; |
| if (s < end && '0' <= *s && *s <= '7') |
| x = (x<<3) + *s++ - '0'; |
| } |
| *p++ = x; |
| break; |
| |
| /* hex escapes */ |
| /* \xXX */ |
| case 'x': |
| digits = 2; |
| message = "truncated \\xXX escape"; |
| goto hexescape; |
| |
| /* \uXXXX */ |
| case 'u': |
| digits = 4; |
| message = "truncated \\uXXXX escape"; |
| goto hexescape; |
| |
| /* \UXXXXXXXX */ |
| case 'U': |
| digits = 8; |
| message = "truncated \\UXXXXXXXX escape"; |
| hexescape: |
| chr = 0; |
| if (end - s < digits) { |
| /* count only hex digits */ |
| for (; s < end; ++s) { |
| c = (unsigned char)*s; |
| if (!Py_ISXDIGIT(c)) |
| goto error; |
| } |
| goto error; |
| } |
| for (; digits--; ++s) { |
| c = (unsigned char)*s; |
| if (!Py_ISXDIGIT(c)) |
| goto error; |
| chr = (chr<<4) & ~0xF; |
| if (c >= '0' && c <= '9') |
| chr += c - '0'; |
| else if (c >= 'a' && c <= 'f') |
| chr += 10 + c - 'a'; |
| else |
| chr += 10 + c - 'A'; |
| } |
| if (chr == 0xffffffff && PyErr_Occurred()) |
| /* _decoding_error will have already written into the |
| target buffer. */ |
| break; |
| store: |
| /* when we get here, chr is a 32-bit unicode character */ |
| if (chr <= 0xffff) |
| /* UCS-2 character */ |
| *p++ = (Py_UNICODE) chr; |
| else if (chr <= 0x10ffff) { |
| /* UCS-4 character. Either store directly, or as |
| surrogate pair. */ |
| #ifdef Py_UNICODE_WIDE |
| *p++ = chr; |
| #else |
| chr -= 0x10000L; |
| *p++ = 0xD800 + (Py_UNICODE) (chr >> 10); |
| *p++ = 0xDC00 + (Py_UNICODE) (chr & 0x03FF); |
| #endif |
| } else { |
| message = "illegal Unicode character"; |
| goto error; |
| } |
| break; |
| |
| /* \N{name} */ |
| case 'N': |
| message = "malformed \\N character escape"; |
| if (ucnhash_CAPI == NULL) { |
| /* load the unicode data module */ |
| ucnhash_CAPI = (_PyUnicode_Name_CAPI *)PyCapsule_Import(PyUnicodeData_CAPSULE_NAME, 1); |
| if (ucnhash_CAPI == NULL) |
| goto ucnhashError; |
| } |
| if (*s == '{') { |
| const char *start = s+1; |
| /* look for the closing brace */ |
| while (*s != '}' && s < end) |
| s++; |
| if (s > start && s < end && *s == '}') { |
| /* found a name. look it up in the unicode database */ |
| message = "unknown Unicode character name"; |
| s++; |
| if (s - start - 1 <= INT_MAX && |
| ucnhash_CAPI->getcode(NULL, start, (int)(s-start-1), &chr)) |
| goto store; |
| } |
| } |
| goto error; |
| |
| default: |
| if (s > end) { |
| message = "\\ at end of string"; |
| s--; |
| goto error; |
| } |
| else { |
| *p++ = '\\'; |
| *p++ = (unsigned char)s[-1]; |
| } |
| break; |
| } |
| continue; |
| |
| error: |
| endinpos = s-starts; |
| outpos = p-PyUnicode_AS_UNICODE(v); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "unicodeescape", message, |
| &starts, &end, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) |
| goto onError; |
| continue; |
| } |
| if (_PyUnicode_Resize(&v, p - PyUnicode_AS_UNICODE(v)) < 0) |
| goto onError; |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)v; |
| |
| ucnhashError: |
| PyErr_SetString( |
| PyExc_UnicodeError, |
| "\\N escapes not supported (can't load unicodedata module)" |
| ); |
| Py_XDECREF(v); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| |
| onError: |
| Py_XDECREF(v); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| /* Return a Unicode-Escape string version of the Unicode object. |
| |
| If quotes is true, the string is enclosed in u"" or u'' quotes as |
| appropriate. |
| |
| */ |
| |
| Py_LOCAL_INLINE(const Py_UNICODE *) findchar(const Py_UNICODE *s, |
| Py_ssize_t size, |
| Py_UNICODE ch) |
| { |
| /* like wcschr, but doesn't stop at NULL characters */ |
| |
| while (size-- > 0) { |
| if (*s == ch) |
| return s; |
| s++; |
| } |
| |
| return NULL; |
| } |
| |
| static const char *hexdigits = "0123456789abcdef"; |
| |
| PyObject *PyUnicode_EncodeUnicodeEscape(const Py_UNICODE *s, |
| Py_ssize_t size) |
| { |
| PyObject *repr; |
| char *p; |
| |
| #ifdef Py_UNICODE_WIDE |
| const Py_ssize_t expandsize = 10; |
| #else |
| const Py_ssize_t expandsize = 6; |
| #endif |
| |
| /* XXX(nnorwitz): rather than over-allocating, it would be |
| better to choose a different scheme. Perhaps scan the |
| first N-chars of the string and allocate based on that size. |
| */ |
| /* Initial allocation is based on the longest-possible unichr |
| escape. |
| |
| In wide (UTF-32) builds '\U00xxxxxx' is 10 chars per source |
| unichr, so in this case it's the longest unichr escape. In |
| narrow (UTF-16) builds this is five chars per source unichr |
| since there are two unichrs in the surrogate pair, so in narrow |
| (UTF-16) builds it's not the longest unichr escape. |
| |
| In wide or narrow builds '\uxxxx' is 6 chars per source unichr, |
| so in the narrow (UTF-16) build case it's the longest unichr |
| escape. |
| */ |
| |
| if (size == 0) |
| return PyBytes_FromStringAndSize(NULL, 0); |
| |
| if (size > (PY_SSIZE_T_MAX - 2 - 1) / expandsize) |
| return PyErr_NoMemory(); |
| |
| repr = PyBytes_FromStringAndSize(NULL, |
| 2 |
| + expandsize*size |
| + 1); |
| if (repr == NULL) |
| return NULL; |
| |
| p = PyBytes_AS_STRING(repr); |
| |
| while (size-- > 0) { |
| Py_UNICODE ch = *s++; |
| |
| /* Escape backslashes */ |
| if (ch == '\\') { |
| *p++ = '\\'; |
| *p++ = (char) ch; |
| continue; |
| } |
| |
| #ifdef Py_UNICODE_WIDE |
| /* Map 21-bit characters to '\U00xxxxxx' */ |
| else if (ch >= 0x10000) { |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = hexdigits[(ch >> 28) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 24) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 20) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 16) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 12) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 8) & 0x0000000F]; |
| *p++ = hexdigits[(ch >> 4) & 0x0000000F]; |
| *p++ = hexdigits[ch & 0x0000000F]; |
| continue; |
| } |
| #else |
| /* Map UTF-16 surrogate pairs to '\U00xxxxxx' */ |
| else if (ch >= 0xD800 && ch < 0xDC00) { |
| Py_UNICODE ch2; |
| Py_UCS4 ucs; |
| |
| ch2 = *s++; |
| size--; |
| if (ch2 >= 0xDC00 && ch2 <= 0xDFFF) { |
| ucs = (((ch & 0x03FF) << 10) | (ch2 & 0x03FF)) + 0x00010000; |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = hexdigits[(ucs >> 28) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 24) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 20) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 16) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 12) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 8) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 4) & 0x0000000F]; |
| *p++ = hexdigits[ucs & 0x0000000F]; |
| continue; |
| } |
| /* Fall through: isolated surrogates are copied as-is */ |
| s--; |
| size++; |
| } |
| #endif |
| |
| /* Map 16-bit characters to '\uxxxx' */ |
| if (ch >= 256) { |
| *p++ = '\\'; |
| *p++ = 'u'; |
| *p++ = hexdigits[(ch >> 12) & 0x000F]; |
| *p++ = hexdigits[(ch >> 8) & 0x000F]; |
| *p++ = hexdigits[(ch >> 4) & 0x000F]; |
| *p++ = hexdigits[ch & 0x000F]; |
| } |
| |
| /* Map special whitespace to '\t', \n', '\r' */ |
| else if (ch == '\t') { |
| *p++ = '\\'; |
| *p++ = 't'; |
| } |
| else if (ch == '\n') { |
| *p++ = '\\'; |
| *p++ = 'n'; |
| } |
| else if (ch == '\r') { |
| *p++ = '\\'; |
| *p++ = 'r'; |
| } |
| |
| /* Map non-printable US ASCII to '\xhh' */ |
| else if (ch < ' ' || ch >= 0x7F) { |
| *p++ = '\\'; |
| *p++ = 'x'; |
| *p++ = hexdigits[(ch >> 4) & 0x000F]; |
| *p++ = hexdigits[ch & 0x000F]; |
| } |
| |
| /* Copy everything else as-is */ |
| else |
| *p++ = (char) ch; |
| } |
| |
| assert(p - PyBytes_AS_STRING(repr) > 0); |
| if (_PyBytes_Resize(&repr, p - PyBytes_AS_STRING(repr)) < 0) |
| return NULL; |
| return repr; |
| } |
| |
| PyObject *PyUnicode_AsUnicodeEscapeString(PyObject *unicode) |
| { |
| PyObject *s; |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| s = PyUnicode_EncodeUnicodeEscape(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode)); |
| return s; |
| } |
| |
| /* --- Raw Unicode Escape Codec ------------------------------------------- */ |
| |
| PyObject *PyUnicode_DecodeRawUnicodeEscape(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| const char *end; |
| const char *bs; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| /* Escaped strings will always be longer than the resulting |
| Unicode string, so we start with size here and then reduce the |
| length after conversion to the true value. (But decoding error |
| handler might have to resize the string) */ |
| v = _PyUnicode_New(size); |
| if (v == NULL) |
| goto onError; |
| if (size == 0) |
| return (PyObject *)v; |
| p = PyUnicode_AS_UNICODE(v); |
| end = s + size; |
| while (s < end) { |
| unsigned char c; |
| Py_UCS4 x; |
| int i; |
| int count; |
| |
| /* Non-escape characters are interpreted as Unicode ordinals */ |
| if (*s != '\\') { |
| *p++ = (unsigned char)*s++; |
| continue; |
| } |
| startinpos = s-starts; |
| |
| /* \u-escapes are only interpreted iff the number of leading |
| backslashes if odd */ |
| bs = s; |
| for (;s < end;) { |
| if (*s != '\\') |
| break; |
| *p++ = (unsigned char)*s++; |
| } |
| if (((s - bs) & 1) == 0 || |
| s >= end || |
| (*s != 'u' && *s != 'U')) { |
| continue; |
| } |
| p--; |
| count = *s=='u' ? 4 : 8; |
| s++; |
| |
| /* \uXXXX with 4 hex digits, \Uxxxxxxxx with 8 */ |
| outpos = p-PyUnicode_AS_UNICODE(v); |
| for (x = 0, i = 0; i < count; ++i, ++s) { |
| c = (unsigned char)*s; |
| if (!Py_ISXDIGIT(c)) { |
| endinpos = s-starts; |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "rawunicodeescape", "truncated \\uXXXX", |
| &starts, &end, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) |
| goto onError; |
| goto nextByte; |
| } |
| x = (x<<4) & ~0xF; |
| if (c >= '0' && c <= '9') |
| x += c - '0'; |
| else if (c >= 'a' && c <= 'f') |
| x += 10 + c - 'a'; |
| else |
| x += 10 + c - 'A'; |
| } |
| if (x <= 0xffff) |
| /* UCS-2 character */ |
| *p++ = (Py_UNICODE) x; |
| else if (x <= 0x10ffff) { |
| /* UCS-4 character. Either store directly, or as |
| surrogate pair. */ |
| #ifdef Py_UNICODE_WIDE |
| *p++ = (Py_UNICODE) x; |
| #else |
| x -= 0x10000L; |
| *p++ = 0xD800 + (Py_UNICODE) (x >> 10); |
| *p++ = 0xDC00 + (Py_UNICODE) (x & 0x03FF); |
| #endif |
| } else { |
| endinpos = s-starts; |
| outpos = p-PyUnicode_AS_UNICODE(v); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "rawunicodeescape", "\\Uxxxxxxxx out of range", |
| &starts, &end, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) |
| goto onError; |
| } |
| nextByte: |
| ; |
| } |
| if (_PyUnicode_Resize(&v, p - PyUnicode_AS_UNICODE(v)) < 0) |
| goto onError; |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)v; |
| |
| onError: |
| Py_XDECREF(v); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_EncodeRawUnicodeEscape(const Py_UNICODE *s, |
| Py_ssize_t size) |
| { |
| PyObject *repr; |
| char *p; |
| char *q; |
| |
| #ifdef Py_UNICODE_WIDE |
| const Py_ssize_t expandsize = 10; |
| #else |
| const Py_ssize_t expandsize = 6; |
| #endif |
| |
| if (size > PY_SSIZE_T_MAX / expandsize) |
| return PyErr_NoMemory(); |
| |
| repr = PyBytes_FromStringAndSize(NULL, expandsize * size); |
| if (repr == NULL) |
| return NULL; |
| if (size == 0) |
| return repr; |
| |
| p = q = PyBytes_AS_STRING(repr); |
| while (size-- > 0) { |
| Py_UNICODE ch = *s++; |
| #ifdef Py_UNICODE_WIDE |
| /* Map 32-bit characters to '\Uxxxxxxxx' */ |
| if (ch >= 0x10000) { |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = hexdigits[(ch >> 28) & 0xf]; |
| *p++ = hexdigits[(ch >> 24) & 0xf]; |
| *p++ = hexdigits[(ch >> 20) & 0xf]; |
| *p++ = hexdigits[(ch >> 16) & 0xf]; |
| *p++ = hexdigits[(ch >> 12) & 0xf]; |
| *p++ = hexdigits[(ch >> 8) & 0xf]; |
| *p++ = hexdigits[(ch >> 4) & 0xf]; |
| *p++ = hexdigits[ch & 15]; |
| } |
| else |
| #else |
| /* Map UTF-16 surrogate pairs to '\U00xxxxxx' */ |
| if (ch >= 0xD800 && ch < 0xDC00) { |
| Py_UNICODE ch2; |
| Py_UCS4 ucs; |
| |
| ch2 = *s++; |
| size--; |
| if (ch2 >= 0xDC00 && ch2 <= 0xDFFF) { |
| ucs = (((ch & 0x03FF) << 10) | (ch2 & 0x03FF)) + 0x00010000; |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = hexdigits[(ucs >> 28) & 0xf]; |
| *p++ = hexdigits[(ucs >> 24) & 0xf]; |
| *p++ = hexdigits[(ucs >> 20) & 0xf]; |
| *p++ = hexdigits[(ucs >> 16) & 0xf]; |
| *p++ = hexdigits[(ucs >> 12) & 0xf]; |
| *p++ = hexdigits[(ucs >> 8) & 0xf]; |
| *p++ = hexdigits[(ucs >> 4) & 0xf]; |
| *p++ = hexdigits[ucs & 0xf]; |
| continue; |
| } |
| /* Fall through: isolated surrogates are copied as-is */ |
| s--; |
| size++; |
| } |
| #endif |
| /* Map 16-bit characters to '\uxxxx' */ |
| if (ch >= 256) { |
| *p++ = '\\'; |
| *p++ = 'u'; |
| *p++ = hexdigits[(ch >> 12) & 0xf]; |
| *p++ = hexdigits[(ch >> 8) & 0xf]; |
| *p++ = hexdigits[(ch >> 4) & 0xf]; |
| *p++ = hexdigits[ch & 15]; |
| } |
| /* Copy everything else as-is */ |
| else |
| *p++ = (char) ch; |
| } |
| size = p - q; |
| |
| assert(size > 0); |
| if (_PyBytes_Resize(&repr, size) < 0) |
| return NULL; |
| return repr; |
| } |
| |
| PyObject *PyUnicode_AsRawUnicodeEscapeString(PyObject *unicode) |
| { |
| PyObject *s; |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| s = PyUnicode_EncodeRawUnicodeEscape(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode)); |
| |
| return s; |
| } |
| |
| /* --- Unicode Internal Codec ------------------------------------------- */ |
| |
| PyObject *_PyUnicode_DecodeUnicodeInternal(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| const char *end; |
| const char *reason; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| #ifdef Py_UNICODE_WIDE |
| Py_UNICODE unimax = PyUnicode_GetMax(); |
| #endif |
| |
| /* XXX overflow detection missing */ |
| v = _PyUnicode_New((size+Py_UNICODE_SIZE-1)/ Py_UNICODE_SIZE); |
| if (v == NULL) |
| goto onError; |
| if (PyUnicode_GetSize((PyObject *)v) == 0) |
| return (PyObject *)v; |
| p = PyUnicode_AS_UNICODE(v); |
| end = s + size; |
| |
| while (s < end) { |
| memcpy(p, s, sizeof(Py_UNICODE)); |
| /* We have to sanity check the raw data, otherwise doom looms for |
| some malformed UCS-4 data. */ |
| if ( |
| #ifdef Py_UNICODE_WIDE |
| *p > unimax || *p < 0 || |
| #endif |
| end-s < Py_UNICODE_SIZE |
| ) |
| { |
| startinpos = s - starts; |
| if (end-s < Py_UNICODE_SIZE) { |
| endinpos = end-starts; |
| reason = "truncated input"; |
| } |
| else { |
| endinpos = s - starts + Py_UNICODE_SIZE; |
| reason = "illegal code point (> 0x10FFFF)"; |
| } |
| outpos = p - PyUnicode_AS_UNICODE(v); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "unicode_internal", reason, |
| &starts, &end, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) { |
| goto onError; |
| } |
| } |
| else { |
| p++; |
| s += Py_UNICODE_SIZE; |
| } |
| } |
| |
| if (_PyUnicode_Resize(&v, p - PyUnicode_AS_UNICODE(v)) < 0) |
| goto onError; |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)v; |
| |
| onError: |
| Py_XDECREF(v); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| /* --- Latin-1 Codec ------------------------------------------------------ */ |
| |
| PyObject *PyUnicode_DecodeLatin1(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| const char *e, *unrolled_end; |
| |
| /* Latin-1 is equivalent to the first 256 ordinals in Unicode. */ |
| if (size == 1) { |
| Py_UNICODE r = *(unsigned char*)s; |
| return PyUnicode_FromUnicode(&r, 1); |
| } |
| |
| v = _PyUnicode_New(size); |
| if (v == NULL) |
| goto onError; |
| if (size == 0) |
| return (PyObject *)v; |
| p = PyUnicode_AS_UNICODE(v); |
| e = s + size; |
| /* Unrolling the copy makes it much faster by reducing the looping |
| overhead. This is similar to what many memcpy() implementations do. */ |
| unrolled_end = e - 4; |
| while (s < unrolled_end) { |
| p[0] = (unsigned char) s[0]; |
| p[1] = (unsigned char) s[1]; |
| p[2] = (unsigned char) s[2]; |
| p[3] = (unsigned char) s[3]; |
| s += 4; |
| p += 4; |
| } |
| while (s < e) |
| *p++ = (unsigned char) *s++; |
| return (PyObject *)v; |
| |
| onError: |
| Py_XDECREF(v); |
| return NULL; |
| } |
| |
| /* create or adjust a UnicodeEncodeError */ |
| static void make_encode_exception(PyObject **exceptionObject, |
| const char *encoding, |
| const Py_UNICODE *unicode, Py_ssize_t size, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason) |
| { |
| if (*exceptionObject == NULL) { |
| *exceptionObject = PyUnicodeEncodeError_Create( |
| encoding, unicode, size, startpos, endpos, reason); |
| } |
| else { |
| if (PyUnicodeEncodeError_SetStart(*exceptionObject, startpos)) |
| goto onError; |
| if (PyUnicodeEncodeError_SetEnd(*exceptionObject, endpos)) |
| goto onError; |
| if (PyUnicodeEncodeError_SetReason(*exceptionObject, reason)) |
| goto onError; |
| return; |
| onError: |
| Py_DECREF(*exceptionObject); |
| *exceptionObject = NULL; |
| } |
| } |
| |
| /* raises a UnicodeEncodeError */ |
| static void raise_encode_exception(PyObject **exceptionObject, |
| const char *encoding, |
| const Py_UNICODE *unicode, Py_ssize_t size, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason) |
| { |
| make_encode_exception(exceptionObject, |
| encoding, unicode, size, startpos, endpos, reason); |
| if (*exceptionObject != NULL) |
| PyCodec_StrictErrors(*exceptionObject); |
| } |
| |
| /* error handling callback helper: |
| build arguments, call the callback and check the arguments, |
| put the result into newpos and return the replacement string, which |
| has to be freed by the caller */ |
| static PyObject *unicode_encode_call_errorhandler(const char *errors, |
| PyObject **errorHandler, |
| const char *encoding, const char *reason, |
| const Py_UNICODE *unicode, Py_ssize_t size, PyObject **exceptionObject, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| Py_ssize_t *newpos) |
| { |
| static char *argparse = "On;encoding error handler must return (str/bytes, int) tuple"; |
| |
| PyObject *restuple; |
| PyObject *resunicode; |
| |
| if (*errorHandler == NULL) { |
| *errorHandler = PyCodec_LookupError(errors); |
| if (*errorHandler == NULL) |
| return NULL; |
| } |
| |
| make_encode_exception(exceptionObject, |
| encoding, unicode, size, startpos, endpos, reason); |
| if (*exceptionObject == NULL) |
| return NULL; |
| |
| restuple = PyObject_CallFunctionObjArgs( |
| *errorHandler, *exceptionObject, NULL); |
| if (restuple == NULL) |
| return NULL; |
| if (!PyTuple_Check(restuple)) { |
| PyErr_SetString(PyExc_TypeError, &argparse[3]); |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| if (!PyArg_ParseTuple(restuple, argparse, |
| &resunicode, newpos)) { |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| if (!PyUnicode_Check(resunicode) && !PyBytes_Check(resunicode)) { |
| PyErr_SetString(PyExc_TypeError, &argparse[3]); |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| if (*newpos<0) |
| *newpos = size+*newpos; |
| if (*newpos<0 || *newpos>size) { |
| PyErr_Format(PyExc_IndexError, "position %zd from error handler out of bounds", *newpos); |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| Py_INCREF(resunicode); |
| Py_DECREF(restuple); |
| return resunicode; |
| } |
| |
| static PyObject *unicode_encode_ucs1(const Py_UNICODE *p, |
| Py_ssize_t size, |
| const char *errors, |
| int limit) |
| { |
| /* output object */ |
| PyObject *res; |
| /* pointers to the beginning and end+1 of input */ |
| const Py_UNICODE *startp = p; |
| const Py_UNICODE *endp = p + size; |
| /* pointer to the beginning of the unencodable characters */ |
| /* const Py_UNICODE *badp = NULL; */ |
| /* pointer into the output */ |
| char *str; |
| /* current output position */ |
| Py_ssize_t ressize; |
| const char *encoding = (limit == 256) ? "latin-1" : "ascii"; |
| const char *reason = (limit == 256) ? "ordinal not in range(256)" : "ordinal not in range(128)"; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| /* the following variable is used for caching string comparisons |
| * -1=not initialized, 0=unknown, 1=strict, 2=replace, 3=ignore, 4=xmlcharrefreplace */ |
| int known_errorHandler = -1; |
| |
| /* allocate enough for a simple encoding without |
| replacements, if we need more, we'll resize */ |
| if (size == 0) |
| return PyBytes_FromStringAndSize(NULL, 0); |
| res = PyBytes_FromStringAndSize(NULL, size); |
| if (res == NULL) |
| return NULL; |
| str = PyBytes_AS_STRING(res); |
| ressize = size; |
| |
| while (p<endp) { |
| Py_UNICODE c = *p; |
| |
| /* can we encode this? */ |
| if (c<limit) { |
| /* no overflow check, because we know that the space is enough */ |
| *str++ = (char)c; |
| ++p; |
| } |
| else { |
| Py_ssize_t unicodepos = p-startp; |
| Py_ssize_t requiredsize; |
| PyObject *repunicode; |
| Py_ssize_t repsize; |
| Py_ssize_t newpos; |
| Py_ssize_t respos; |
| Py_UNICODE *uni2; |
| /* startpos for collecting unencodable chars */ |
| const Py_UNICODE *collstart = p; |
| const Py_UNICODE *collend = p; |
| /* find all unecodable characters */ |
| while ((collend < endp) && ((*collend)>=limit)) |
| ++collend; |
| /* cache callback name lookup (if not done yet, i.e. it's the first error) */ |
| if (known_errorHandler==-1) { |
| if ((errors==NULL) || (!strcmp(errors, "strict"))) |
| known_errorHandler = 1; |
| else if (!strcmp(errors, "replace")) |
| known_errorHandler = 2; |
| else if (!strcmp(errors, "ignore")) |
| known_errorHandler = 3; |
| else if (!strcmp(errors, "xmlcharrefreplace")) |
| known_errorHandler = 4; |
| else |
| known_errorHandler = 0; |
| } |
| switch (known_errorHandler) { |
| case 1: /* strict */ |
| raise_encode_exception(&exc, encoding, startp, size, collstart-startp, collend-startp, reason); |
| goto onError; |
| case 2: /* replace */ |
| while (collstart++<collend) |
| *str++ = '?'; /* fall through */ |
| case 3: /* ignore */ |
| p = collend; |
| break; |
| case 4: /* xmlcharrefreplace */ |
| respos = str - PyBytes_AS_STRING(res); |
| /* determine replacement size (temporarily (mis)uses p) */ |
| for (p = collstart, repsize = 0; p < collend; ++p) { |
| if (*p<10) |
| repsize += 2+1+1; |
| else if (*p<100) |
| repsize += 2+2+1; |
| else if (*p<1000) |
| repsize += 2+3+1; |
| else if (*p<10000) |
| repsize += 2+4+1; |
| #ifndef Py_UNICODE_WIDE |
| else |
| repsize += 2+5+1; |
| #else |
| else if (*p<100000) |
| repsize += 2+5+1; |
| else if (*p<1000000) |
| repsize += 2+6+1; |
| else |
| repsize += 2+7+1; |
| #endif |
| } |
| requiredsize = respos+repsize+(endp-collend); |
| if (requiredsize > ressize) { |
| if (requiredsize<2*ressize) |
| requiredsize = 2*ressize; |
| if (_PyBytes_Resize(&res, requiredsize)) |
| goto onError; |
| str = PyBytes_AS_STRING(res) + respos; |
| ressize = requiredsize; |
| } |
| /* generate replacement (temporarily (mis)uses p) */ |
| for (p = collstart; p < collend; ++p) { |
| str += sprintf(str, "&#%d;", (int)*p); |
| } |
| p = collend; |
| break; |
| default: |
| repunicode = unicode_encode_call_errorhandler(errors, &errorHandler, |
| encoding, reason, startp, size, &exc, |
| collstart-startp, collend-startp, &newpos); |
| if (repunicode == NULL) |
| goto onError; |
| if (PyBytes_Check(repunicode)) { |
| /* Directly copy bytes result to output. */ |
| repsize = PyBytes_Size(repunicode); |
| if (repsize > 1) { |
| /* Make room for all additional bytes. */ |
| respos = str - PyBytes_AS_STRING(res); |
| if (_PyBytes_Resize(&res, ressize+repsize-1)) { |
| Py_DECREF(repunicode); |
| goto onError; |
| } |
| str = PyBytes_AS_STRING(res) + respos; |
| ressize += repsize-1; |
| } |
| memcpy(str, PyBytes_AsString(repunicode), repsize); |
| str += repsize; |
| p = startp + newpos; |
| Py_DECREF(repunicode); |
| break; |
| } |
| /* need more space? (at least enough for what we |
| have+the replacement+the rest of the string, so |
| we won't have to check space for encodable characters) */ |
| respos = str - PyBytes_AS_STRING(res); |
| repsize = PyUnicode_GET_SIZE(repunicode); |
| requiredsize = respos+repsize+(endp-collend); |
| if (requiredsize > ressize) { |
| if (requiredsize<2*ressize) |
| requiredsize = 2*ressize; |
| if (_PyBytes_Resize(&res, requiredsize)) { |
| Py_DECREF(repunicode); |
| goto onError; |
| } |
| str = PyBytes_AS_STRING(res) + respos; |
| ressize = requiredsize; |
| } |
| /* check if there is anything unencodable in the replacement |
| and copy it to the output */ |
| for (uni2 = PyUnicode_AS_UNICODE(repunicode);repsize-->0; ++uni2, ++str) { |
| c = *uni2; |
| if (c >= limit) { |
| raise_encode_exception(&exc, encoding, startp, size, |
| unicodepos, unicodepos+1, reason); |
| Py_DECREF(repunicode); |
| goto onError; |
| } |
| *str = (char)c; |
| } |
| p = startp + newpos; |
| Py_DECREF(repunicode); |
| } |
| } |
| } |
| /* Resize if we allocated to much */ |
| size = str - PyBytes_AS_STRING(res); |
| if (size < ressize) { /* If this falls res will be NULL */ |
| assert(size >= 0); |
| if (_PyBytes_Resize(&res, size) < 0) |
| goto onError; |
| } |
| |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return res; |
| |
| onError: |
| Py_XDECREF(res); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_EncodeLatin1(const Py_UNICODE *p, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| return unicode_encode_ucs1(p, size, errors, 256); |
| } |
| |
| PyObject *PyUnicode_AsLatin1String(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeLatin1(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| } |
| |
| /* --- 7-bit ASCII Codec -------------------------------------------------- */ |
| |
| PyObject *PyUnicode_DecodeASCII(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| const char *starts = s; |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| const char *e; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| |
| /* ASCII is equivalent to the first 128 ordinals in Unicode. */ |
| if (size == 1 && *(unsigned char*)s < 128) { |
| Py_UNICODE r = *(unsigned char*)s; |
| return PyUnicode_FromUnicode(&r, 1); |
| } |
| |
| v = _PyUnicode_New(size); |
| if (v == NULL) |
| goto onError; |
| if (size == 0) |
| return (PyObject *)v; |
| p = PyUnicode_AS_UNICODE(v); |
| e = s + size; |
| while (s < e) { |
| register unsigned char c = (unsigned char)*s; |
| if (c < 128) { |
| *p++ = c; |
| ++s; |
| } |
| else { |
| startinpos = s-starts; |
| endinpos = startinpos + 1; |
| outpos = p - (Py_UNICODE *)PyUnicode_AS_UNICODE(v); |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "ascii", "ordinal not in range(128)", |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) |
| goto onError; |
| } |
| } |
| if (p - PyUnicode_AS_UNICODE(v) < PyUnicode_GET_SIZE(v)) |
| if (_PyUnicode_Resize(&v, p - PyUnicode_AS_UNICODE(v)) < 0) |
| goto onError; |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)v; |
| |
| onError: |
| Py_XDECREF(v); |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_EncodeASCII(const Py_UNICODE *p, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| return unicode_encode_ucs1(p, size, errors, 128); |
| } |
| |
| PyObject *PyUnicode_AsASCIIString(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeASCII(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| } |
| |
| #if defined(MS_WINDOWS) && defined(HAVE_USABLE_WCHAR_T) |
| |
| /* --- MBCS codecs for Windows -------------------------------------------- */ |
| |
| #if SIZEOF_INT < SIZEOF_SIZE_T |
| #define NEED_RETRY |
| #endif |
| |
| /* XXX This code is limited to "true" double-byte encodings, as |
| a) it assumes an incomplete character consists of a single byte, and |
| b) IsDBCSLeadByte (probably) does not work for non-DBCS multi-byte |
| encodings, see IsDBCSLeadByteEx documentation. */ |
| |
| static int is_dbcs_lead_byte(const char *s, int offset) |
| { |
| const char *curr = s + offset; |
| |
| if (IsDBCSLeadByte(*curr)) { |
| const char *prev = CharPrev(s, curr); |
| return (prev == curr) || !IsDBCSLeadByte(*prev) || (curr - prev == 2); |
| } |
| return 0; |
| } |
| |
| /* |
| * Decode MBCS string into unicode object. If 'final' is set, converts |
| * trailing lead-byte too. Returns consumed size if succeed, -1 otherwise. |
| */ |
| static int decode_mbcs(PyUnicodeObject **v, |
| const char *s, /* MBCS string */ |
| int size, /* sizeof MBCS string */ |
| int final, |
| const char *errors) |
| { |
| Py_UNICODE *p; |
| Py_ssize_t n; |
| DWORD usize; |
| DWORD flags; |
| |
| assert(size >= 0); |
| |
| /* check and handle 'errors' arg */ |
| if (errors==NULL || strcmp(errors, "strict")==0) |
| flags = MB_ERR_INVALID_CHARS; |
| else if (strcmp(errors, "ignore")==0) |
| flags = 0; |
| else { |
| PyErr_Format(PyExc_ValueError, |
| "mbcs encoding does not support errors='%s'", |
| errors); |
| return -1; |
| } |
| |
| /* Skip trailing lead-byte unless 'final' is set */ |
| if (!final && size >= 1 && is_dbcs_lead_byte(s, size - 1)) |
| --size; |
| |
| /* First get the size of the result */ |
| if (size > 0) { |
| usize = MultiByteToWideChar(CP_ACP, flags, s, size, NULL, 0); |
| if (usize==0) |
| goto mbcs_decode_error; |
| } else |
| usize = 0; |
| |
| if (*v == NULL) { |
| /* Create unicode object */ |
| *v = _PyUnicode_New(usize); |
| if (*v == NULL) |
| return -1; |
| n = 0; |
| } |
| else { |
| /* Extend unicode object */ |
| n = PyUnicode_GET_SIZE(*v); |
| if (_PyUnicode_Resize(v, n + usize) < 0) |
| return -1; |
| } |
| |
| /* Do the conversion */ |
| if (usize > 0) { |
| p = PyUnicode_AS_UNICODE(*v) + n; |
| if (0 == MultiByteToWideChar(CP_ACP, flags, s, size, p, usize)) { |
| goto mbcs_decode_error; |
| } |
| } |
| return size; |
| |
| mbcs_decode_error: |
| /* If the last error was ERROR_NO_UNICODE_TRANSLATION, then |
| we raise a UnicodeDecodeError - else it is a 'generic' |
| windows error |
| */ |
| if (GetLastError()==ERROR_NO_UNICODE_TRANSLATION) { |
| /* Ideally, we should get reason from FormatMessage - this |
| is the Windows 2000 English version of the message |
| */ |
| PyObject *exc = NULL; |
| const char *reason = "No mapping for the Unicode character exists " |
| "in the target multi-byte code page."; |
| make_decode_exception(&exc, "mbcs", s, size, 0, 0, reason); |
| if (exc != NULL) { |
| PyCodec_StrictErrors(exc); |
| Py_DECREF(exc); |
| } |
| } else { |
| PyErr_SetFromWindowsErrWithFilename(0, NULL); |
| } |
| return -1; |
| } |
| |
| PyObject *PyUnicode_DecodeMBCSStateful(const char *s, |
| Py_ssize_t size, |
| const char *errors, |
| Py_ssize_t *consumed) |
| { |
| PyUnicodeObject *v = NULL; |
| int done; |
| |
| if (consumed) |
| *consumed = 0; |
| |
| #ifdef NEED_RETRY |
| retry: |
| if (size > INT_MAX) |
| done = decode_mbcs(&v, s, INT_MAX, 0, errors); |
| else |
| #endif |
| done = decode_mbcs(&v, s, (int)size, !consumed, errors); |
| |
| if (done < 0) { |
| Py_XDECREF(v); |
| return NULL; |
| } |
| |
| if (consumed) |
| *consumed += done; |
| |
| #ifdef NEED_RETRY |
| if (size > INT_MAX) { |
| s += done; |
| size -= done; |
| goto retry; |
| } |
| #endif |
| |
| return (PyObject *)v; |
| } |
| |
| PyObject *PyUnicode_DecodeMBCS(const char *s, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| return PyUnicode_DecodeMBCSStateful(s, size, errors, NULL); |
| } |
| |
| /* |
| * Convert unicode into string object (MBCS). |
| * Returns 0 if succeed, -1 otherwise. |
| */ |
| static int encode_mbcs(PyObject **repr, |
| const Py_UNICODE *p, /* unicode */ |
| int size, /* size of unicode */ |
| const char* errors) |
| { |
| BOOL usedDefaultChar = FALSE; |
| BOOL *pusedDefaultChar; |
| int mbcssize; |
| Py_ssize_t n; |
| PyObject *exc = NULL; |
| DWORD flags; |
| |
| assert(size >= 0); |
| |
| /* check and handle 'errors' arg */ |
| if (errors==NULL || strcmp(errors, "strict")==0) { |
| flags = WC_NO_BEST_FIT_CHARS; |
| pusedDefaultChar = &usedDefaultChar; |
| } else if (strcmp(errors, "replace")==0) { |
| flags = 0; |
| pusedDefaultChar = NULL; |
| } else { |
| PyErr_Format(PyExc_ValueError, |
| "mbcs encoding does not support errors='%s'", |
| errors); |
| return -1; |
| } |
| |
| /* First get the size of the result */ |
| if (size > 0) { |
| mbcssize = WideCharToMultiByte(CP_ACP, flags, p, size, NULL, 0, |
| NULL, pusedDefaultChar); |
| if (mbcssize == 0) { |
| PyErr_SetFromWindowsErrWithFilename(0, NULL); |
| return -1; |
| } |
| /* If we used a default char, then we failed! */ |
| if (pusedDefaultChar && *pusedDefaultChar) |
| goto mbcs_encode_error; |
| } else { |
| mbcssize = 0; |
| } |
| |
| if (*repr == NULL) { |
| /* Create string object */ |
| *repr = PyBytes_FromStringAndSize(NULL, mbcssize); |
| if (*repr == NULL) |
| return -1; |
| n = 0; |
| } |
| else { |
| /* Extend string object */ |
| n = PyBytes_Size(*repr); |
| if (_PyBytes_Resize(repr, n + mbcssize) < 0) |
| return -1; |
| } |
| |
| /* Do the conversion */ |
| if (size > 0) { |
| char *s = PyBytes_AS_STRING(*repr) + n; |
| if (0 == WideCharToMultiByte(CP_ACP, flags, p, size, s, mbcssize, |
| NULL, pusedDefaultChar)) { |
| PyErr_SetFromWindowsErrWithFilename(0, NULL); |
| return -1; |
| } |
| if (pusedDefaultChar && *pusedDefaultChar) |
| goto mbcs_encode_error; |
| } |
| return 0; |
| |
| mbcs_encode_error: |
| raise_encode_exception(&exc, "mbcs", p, size, 0, 0, "invalid character"); |
| Py_XDECREF(exc); |
| return -1; |
| } |
| |
| PyObject *PyUnicode_EncodeMBCS(const Py_UNICODE *p, |
| Py_ssize_t size, |
| const char *errors) |
| { |
| PyObject *repr = NULL; |
| int ret; |
| |
| #ifdef NEED_RETRY |
| retry: |
| if (size > INT_MAX) |
| ret = encode_mbcs(&repr, p, INT_MAX, errors); |
| else |
| #endif |
| ret = encode_mbcs(&repr, p, (int)size, errors); |
| |
| if (ret < 0) { |
| Py_XDECREF(repr); |
| return NULL; |
| } |
| |
| #ifdef NEED_RETRY |
| if (size > INT_MAX) { |
| p += INT_MAX; |
| size -= INT_MAX; |
| goto retry; |
| } |
| #endif |
| |
| return repr; |
| } |
| |
| PyObject *PyUnicode_AsMBCSString(PyObject *unicode) |
| { |
| if (!PyUnicode_Check(unicode)) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeMBCS(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| NULL); |
| } |
| |
| #undef NEED_RETRY |
| |
| #endif /* MS_WINDOWS */ |
| |
| /* --- Character Mapping Codec -------------------------------------------- */ |
| |
| PyObject *PyUnicode_DecodeCharmap(const char *s, |
| Py_ssize_t size, |
| PyObject *mapping, |
| const char *errors) |
| { |
| const char *starts = s; |
| Py_ssize_t startinpos; |
| Py_ssize_t endinpos; |
| Py_ssize_t outpos; |
| const char *e; |
| PyUnicodeObject *v; |
| Py_UNICODE *p; |
| Py_ssize_t extrachars = 0; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| Py_UNICODE *mapstring = NULL; |
| Py_ssize_t maplen = 0; |
| |
| /* Default to Latin-1 */ |
| if (mapping == NULL) |
| return PyUnicode_DecodeLatin1(s, size, errors); |
| |
| v = _PyUnicode_New(size); |
| if (v == NULL) |
| goto onError; |
| if (size == 0) |
| return (PyObject *)v; |
| p = PyUnicode_AS_UNICODE(v); |
| e = s + size; |
| if (PyUnicode_CheckExact(mapping)) { |
| mapstring = PyUnicode_AS_UNICODE(mapping); |
| maplen = PyUnicode_GET_SIZE(mapping); |
| while (s < e) { |
| unsigned char ch = *s; |
| Py_UNICODE x = 0xfffe; /* illegal value */ |
| |
| if (ch < maplen) |
| x = mapstring[ch]; |
| |
| if (x == 0xfffe) { |
| /* undefined mapping */ |
| outpos = p-PyUnicode_AS_UNICODE(v); |
| startinpos = s-starts; |
| endinpos = startinpos+1; |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "charmap", "character maps to <undefined>", |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) { |
| goto onError; |
| } |
| continue; |
| } |
| *p++ = x; |
| ++s; |
| } |
| } |
| else { |
| while (s < e) { |
| unsigned char ch = *s; |
| PyObject *w, *x; |
| |
| /* Get mapping (char ordinal -> integer, Unicode char or None) */ |
| w = PyLong_FromLong((long)ch); |
| if (w == NULL) |
| goto onError; |
| x = PyObject_GetItem(mapping, w); |
| Py_DECREF(w); |
| if (x == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_LookupError)) { |
| /* No mapping found means: mapping is undefined. */ |
| PyErr_Clear(); |
| goto Undefined; |
| } else |
| goto onError; |
| } |
| |
| /* Apply mapping */ |
| if (x == Py_None) |
| goto Undefined; |
| if (PyLong_Check(x)) { |
| long value = PyLong_AS_LONG(x); |
| if (value == 0xFFFE) |
| goto Undefined; |
| if (value < 0 || value > 0x10FFFF) { |
| PyErr_SetString(PyExc_TypeError, |
| "character mapping must be in range(0x110000)"); |
| Py_DECREF(x); |
| goto onError; |
| } |
| |
| #ifndef Py_UNICODE_WIDE |
| if (value > 0xFFFF) { |
| /* see the code for 1-n mapping below */ |
| if (extrachars < 2) { |
| /* resize first */ |
| Py_ssize_t oldpos = p - PyUnicode_AS_UNICODE(v); |
| Py_ssize_t needed = 10 - extrachars; |
| extrachars += needed; |
| /* XXX overflow detection missing */ |
| if (_PyUnicode_Resize(&v, |
| PyUnicode_GET_SIZE(v) + needed) < 0) { |
| Py_DECREF(x); |
| goto onError; |
| } |
| p = PyUnicode_AS_UNICODE(v) + oldpos; |
| } |
| value -= 0x10000; |
| *p++ = 0xD800 | (value >> 10); |
| *p++ = 0xDC00 | (value & 0x3FF); |
| extrachars -= 2; |
| } |
| else |
| #endif |
| *p++ = (Py_UNICODE)value; |
| } |
| else if (PyUnicode_Check(x)) { |
| Py_ssize_t targetsize = PyUnicode_GET_SIZE(x); |
| |
| if (targetsize == 1) { |
| /* 1-1 mapping */ |
| Py_UNICODE value = *PyUnicode_AS_UNICODE(x); |
| if (value == 0xFFFE) |
| goto Undefined; |
| *p++ = value; |
| } |
| else if (targetsize > 1) { |
| /* 1-n mapping */ |
| if (targetsize > extrachars) { |
| /* resize first */ |
| Py_ssize_t oldpos = p - PyUnicode_AS_UNICODE(v); |
| Py_ssize_t needed = (targetsize - extrachars) + \ |
| (targetsize << 2); |
| extrachars += needed; |
| /* XXX overflow detection missing */ |
| if (_PyUnicode_Resize(&v, |
| PyUnicode_GET_SIZE(v) + needed) < 0) { |
| Py_DECREF(x); |
| goto onError; |
| } |
| p = PyUnicode_AS_UNICODE(v) + oldpos; |
| } |
| Py_UNICODE_COPY(p, |
| PyUnicode_AS_UNICODE(x), |
| targetsize); |
| p += targetsize; |
| extrachars -= targetsize; |
| } |
| /* 1-0 mapping: skip the character */ |
| } |
| else { |
| /* wrong return value */ |
| PyErr_SetString(PyExc_TypeError, |
| "character mapping must return integer, None or str"); |
| Py_DECREF(x); |
| goto onError; |
| } |
| Py_DECREF(x); |
| ++s; |
| continue; |
| Undefined: |
| /* undefined mapping */ |
| Py_XDECREF(x); |
| outpos = p-PyUnicode_AS_UNICODE(v); |
| startinpos = s-starts; |
| endinpos = startinpos+1; |
| if (unicode_decode_call_errorhandler( |
| errors, &errorHandler, |
| "charmap", "character maps to <undefined>", |
| &starts, &e, &startinpos, &endinpos, &exc, &s, |
| &v, &outpos, &p)) { |
| goto onError; |
| } |
| } |
| } |
| if (p - PyUnicode_AS_UNICODE(v) < PyUnicode_GET_SIZE(v)) |
| if (_PyUnicode_Resize(&v, p - PyUnicode_AS_UNICODE(v)) < 0) |
| goto onError; |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| return (PyObject *)v; |
| |
| onError: |
| Py_XDECREF(errorHandler); |
| Py_XDECREF(exc); |
| Py_XDECREF(v); |
| return NULL; |
| } |
| |
| /* Charmap encoding: the lookup table */ |
| |
| struct encoding_map{ |
| PyObject_HEAD |
| unsigned char level1[32]; |
| int count2, count3; |
| unsigned char level23[1]; |
| }; |
| |
| static PyObject* |
| encoding_map_size(PyObject *obj, PyObject* args) |
| { |
| struct encoding_map *map = (struct encoding_map*)obj; |
| return PyLong_FromLong(sizeof(*map) - 1 + 16*map->count2 + |
| 128*map->count3); |
| } |
| |
| static PyMethodDef encoding_map_methods[] = { |
| {"size", encoding_map_size, METH_NOARGS, |
| PyDoc_STR("Return the size (in bytes) of this object") }, |
| { 0 } |
| }; |
| |
| static void |
| encoding_map_dealloc(PyObject* o) |
| { |
| PyObject_FREE(o); |
| } |
| |
| static PyTypeObject EncodingMapType = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "EncodingMap", /*tp_name*/ |
| sizeof(struct encoding_map), /*tp_basicsize*/ |
| 0, /*tp_itemsize*/ |
| /* methods */ |
| encoding_map_dealloc, /*tp_dealloc*/ |
| 0, /*tp_print*/ |
| 0, /*tp_getattr*/ |
| 0, /*tp_setattr*/ |
| 0, /*tp_reserved*/ |
| 0, /*tp_repr*/ |
| 0, /*tp_as_number*/ |
| 0, /*tp_as_sequence*/ |
| 0, /*tp_as_mapping*/ |
| 0, /*tp_hash*/ |
| 0, /*tp_call*/ |
| 0, /*tp_str*/ |
| 0, /*tp_getattro*/ |
| 0, /*tp_setattro*/ |
| 0, /*tp_as_buffer*/ |
| Py_TPFLAGS_DEFAULT, /*tp_flags*/ |
| 0, /*tp_doc*/ |
| 0, /*tp_traverse*/ |
| 0, /*tp_clear*/ |
| 0, /*tp_richcompare*/ |
| 0, /*tp_weaklistoffset*/ |
| 0, /*tp_iter*/ |
| 0, /*tp_iternext*/ |
| encoding_map_methods, /*tp_methods*/ |
| 0, /*tp_members*/ |
| 0, /*tp_getset*/ |
| 0, /*tp_base*/ |
| 0, /*tp_dict*/ |
| 0, /*tp_descr_get*/ |
| 0, /*tp_descr_set*/ |
| 0, /*tp_dictoffset*/ |
| 0, /*tp_init*/ |
| 0, /*tp_alloc*/ |
| 0, /*tp_new*/ |
| 0, /*tp_free*/ |
| 0, /*tp_is_gc*/ |
| }; |
| |
| PyObject* |
| PyUnicode_BuildEncodingMap(PyObject* string) |
| { |
| Py_UNICODE *decode; |
| PyObject *result; |
| struct encoding_map *mresult; |
| int i; |
| int need_dict = 0; |
| unsigned char level1[32]; |
| unsigned char level2[512]; |
| unsigned char *mlevel1, *mlevel2, *mlevel3; |
| int count2 = 0, count3 = 0; |
| |
| if (!PyUnicode_Check(string) || PyUnicode_GetSize(string) != 256) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| decode = PyUnicode_AS_UNICODE(string); |
| memset(level1, 0xFF, sizeof level1); |
| memset(level2, 0xFF, sizeof level2); |
| |
| /* If there isn't a one-to-one mapping of NULL to \0, |
| or if there are non-BMP characters, we need to use |
| a mapping dictionary. */ |
| if (decode[0] != 0) |
| need_dict = 1; |
| for (i = 1; i < 256; i++) { |
| int l1, l2; |
| if (decode[i] == 0 |
| #ifdef Py_UNICODE_WIDE |
| || decode[i] > 0xFFFF |
| #endif |
| ) { |
| need_dict = 1; |
| break; |
| } |
| if (decode[i] == 0xFFFE) |
| /* unmapped character */ |
| continue; |
| l1 = decode[i] >> 11; |
| l2 = decode[i] >> 7; |
| if (level1[l1] == 0xFF) |
| level1[l1] = count2++; |
| if (level2[l2] == 0xFF) |
| level2[l2] = count3++; |
| } |
| |
| if (count2 >= 0xFF || count3 >= 0xFF) |
| need_dict = 1; |
| |
| if (need_dict) { |
| PyObject *result = PyDict_New(); |
| PyObject *key, *value; |
| if (!result) |
| return NULL; |
| for (i = 0; i < 256; i++) { |
| key = value = NULL; |
| key = PyLong_FromLong(decode[i]); |
| value = PyLong_FromLong(i); |
| if (!key || !value) |
| goto failed1; |
| if (PyDict_SetItem(result, key, value) == -1) |
| goto failed1; |
| Py_DECREF(key); |
| Py_DECREF(value); |
| } |
| return result; |
| failed1: |
| Py_XDECREF(key); |
| Py_XDECREF(value); |
| Py_DECREF(result); |
| return NULL; |
| } |
| |
| /* Create a three-level trie */ |
| result = PyObject_MALLOC(sizeof(struct encoding_map) + |
| 16*count2 + 128*count3 - 1); |
| if (!result) |
| return PyErr_NoMemory(); |
| PyObject_Init(result, &EncodingMapType); |
| mresult = (struct encoding_map*)result; |
| mresult->count2 = count2; |
| mresult->count3 = count3; |
| mlevel1 = mresult->level1; |
| mlevel2 = mresult->level23; |
| mlevel3 = mresult->level23 + 16*count2; |
| memcpy(mlevel1, level1, 32); |
| memset(mlevel2, 0xFF, 16*count2); |
| memset(mlevel3, 0, 128*count3); |
| count3 = 0; |
| for (i = 1; i < 256; i++) { |
| int o1, o2, o3, i2, i3; |
| if (decode[i] == 0xFFFE) |
| /* unmapped character */ |
| continue; |
| o1 = decode[i]>>11; |
| o2 = (decode[i]>>7) & 0xF; |
| i2 = 16*mlevel1[o1] + o2; |
| if (mlevel2[i2] == 0xFF) |
| mlevel2[i2] = count3++; |
| o3 = decode[i] & 0x7F; |
| i3 = 128*mlevel2[i2] + o3; |
| mlevel3[i3] = i; |
| } |
| return result; |
| } |
| |
| static int |
| encoding_map_lookup(Py_UNICODE c, PyObject *mapping) |
| { |
| struct encoding_map *map = (struct encoding_map*)mapping; |
| int l1 = c>>11; |
| int l2 = (c>>7) & 0xF; |
| int l3 = c & 0x7F; |
| int i; |
| |
| #ifdef Py_UNICODE_WIDE |
| if (c > 0xFFFF) { |
| return -1; |
| } |
| #endif |
| if (c == 0) |
| return 0; |
| /* level 1*/ |
| i = map->level1[l1]; |
| if (i == 0xFF) { |
| return -1; |
| } |
| /* level 2*/ |
| i = map->level23[16*i+l2]; |
| if (i == 0xFF) { |
| return -1; |
| } |
| /* level 3 */ |
| i = map->level23[16*map->count2 + 128*i + l3]; |
| if (i == 0) { |
| return -1; |
| } |
| return i; |
| } |
| |
| /* Lookup the character ch in the mapping. If the character |
| can't be found, Py_None is returned (or NULL, if another |
| error occurred). */ |
| static PyObject *charmapencode_lookup(Py_UNICODE c, PyObject *mapping) |
| { |
| PyObject *w = PyLong_FromLong((long)c); |
| PyObject *x; |
| |
| if (w == NULL) |
| return NULL; |
| x = PyObject_GetItem(mapping, w); |
| Py_DECREF(w); |
| if (x == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_LookupError)) { |
| /* No mapping found means: mapping is undefined. */ |
| PyErr_Clear(); |
| x = Py_None; |
| Py_INCREF(x); |
| return x; |
| } else |
| return NULL; |
| } |
| else if (x == Py_None) |
| return x; |
| else if (PyLong_Check(x)) { |
| long value = PyLong_AS_LONG(x); |
| if (value < 0 || value > 255) { |
| PyErr_SetString(PyExc_TypeError, |
| "character mapping must be in range(256)"); |
| Py_DECREF(x); |
| return NULL; |
| } |
| return x; |
| } |
| else if (PyBytes_Check(x)) |
| return x; |
| else { |
| /* wrong return value */ |
| PyErr_Format(PyExc_TypeError, |
| "character mapping must return integer, bytes or None, not %.400s", |
| x->ob_type->tp_name); |
| Py_DECREF(x); |
| return NULL; |
| } |
| } |
| |
| static int |
| charmapencode_resize(PyObject **outobj, Py_ssize_t *outpos, Py_ssize_t requiredsize) |
| { |
| Py_ssize_t outsize = PyBytes_GET_SIZE(*outobj); |
| /* exponentially overallocate to minimize reallocations */ |
| if (requiredsize < 2*outsize) |
| requiredsize = 2*outsize; |
| if (_PyBytes_Resize(outobj, requiredsize)) |
| return -1; |
| return 0; |
| } |
| |
| typedef enum charmapencode_result { |
| enc_SUCCESS, enc_FAILED, enc_EXCEPTION |
| }charmapencode_result; |
| /* lookup the character, put the result in the output string and adjust |
| various state variables. Resize the output bytes object if not enough |
| space is available. Return a new reference to the object that |
| was put in the output buffer, or Py_None, if the mapping was undefined |
| (in which case no character was written) or NULL, if a |
| reallocation error occurred. The caller must decref the result */ |
| static |
| charmapencode_result charmapencode_output(Py_UNICODE c, PyObject *mapping, |
| PyObject **outobj, Py_ssize_t *outpos) |
| { |
| PyObject *rep; |
| char *outstart; |
| Py_ssize_t outsize = PyBytes_GET_SIZE(*outobj); |
| |
| if (Py_TYPE(mapping) == &EncodingMapType) { |
| int res = encoding_map_lookup(c, mapping); |
| Py_ssize_t requiredsize = *outpos+1; |
| if (res == -1) |
| return enc_FAILED; |
| if (outsize<requiredsize) |
| if (charmapencode_resize(outobj, outpos, requiredsize)) |
| return enc_EXCEPTION; |
| outstart = PyBytes_AS_STRING(*outobj); |
| outstart[(*outpos)++] = (char)res; |
| return enc_SUCCESS; |
| } |
| |
| rep = charmapencode_lookup(c, mapping); |
| if (rep==NULL) |
| return enc_EXCEPTION; |
| else if (rep==Py_None) { |
| Py_DECREF(rep); |
| return enc_FAILED; |
| } else { |
| if (PyLong_Check(rep)) { |
| Py_ssize_t requiredsize = *outpos+1; |
| if (outsize<requiredsize) |
| if (charmapencode_resize(outobj, outpos, requiredsize)) { |
| Py_DECREF(rep); |
| return enc_EXCEPTION; |
| } |
| outstart = PyBytes_AS_STRING(*outobj); |
| outstart[(*outpos)++] = (char)PyLong_AS_LONG(rep); |
| } |
| else { |
| const char *repchars = PyBytes_AS_STRING(rep); |
| Py_ssize_t repsize = PyBytes_GET_SIZE(rep); |
| Py_ssize_t requiredsize = *outpos+repsize; |
| if (outsize<requiredsize) |
| if (charmapencode_resize(outobj, outpos, requiredsize)) { |
| Py_DECREF(rep); |
| return enc_EXCEPTION; |
| } |
| outstart = PyBytes_AS_STRING(*outobj); |
| memcpy(outstart + *outpos, repchars, repsize); |
| *outpos += repsize; |
| } |
| } |
| Py_DECREF(rep); |
| return enc_SUCCESS; |
| } |
| |
| /* handle an error in PyUnicode_EncodeCharmap |
| Return 0 on success, -1 on error */ |
| static |
| int charmap_encoding_error( |
| const Py_UNICODE *p, Py_ssize_t size, Py_ssize_t *inpos, PyObject *mapping, |
| PyObject **exceptionObject, |
| int *known_errorHandler, PyObject **errorHandler, const char *errors, |
| PyObject **res, Py_ssize_t *respos) |
| { |
| PyObject *repunicode = NULL; /* initialize to prevent gcc warning */ |
| Py_ssize_t repsize; |
| Py_ssize_t newpos; |
| Py_UNICODE *uni2; |
| /* startpos for collecting unencodable chars */ |
| Py_ssize_t collstartpos = *inpos; |
| Py_ssize_t collendpos = *inpos+1; |
| Py_ssize_t collpos; |
| char *encoding = "charmap"; |
| char *reason = "character maps to <undefined>"; |
| charmapencode_result x; |
| |
| /* find all unencodable characters */ |
| while (collendpos < size) { |
| PyObject *rep; |
| if (Py_TYPE(mapping) == &EncodingMapType) { |
| int res = encoding_map_lookup(p[collendpos], mapping); |
| if (res != -1) |
| break; |
| ++collendpos; |
| continue; |
| } |
| |
| rep = charmapencode_lookup(p[collendpos], mapping); |
| if (rep==NULL) |
| return -1; |
| else if (rep!=Py_None) { |
| Py_DECREF(rep); |
| break; |
| } |
| Py_DECREF(rep); |
| ++collendpos; |
| } |
| /* cache callback name lookup |
| * (if not done yet, i.e. it's the first error) */ |
| if (*known_errorHandler==-1) { |
| if ((errors==NULL) || (!strcmp(errors, "strict"))) |
| *known_errorHandler = 1; |
| else if (!strcmp(errors, "replace")) |
| *known_errorHandler = 2; |
| else if (!strcmp(errors, "ignore")) |
| *known_errorHandler = 3; |
| else if (!strcmp(errors, "xmlcharrefreplace")) |
| *known_errorHandler = 4; |
| else |
| *known_errorHandler = 0; |
| } |
| switch (*known_errorHandler) { |
| case 1: /* strict */ |
| raise_encode_exception(exceptionObject, encoding, p, size, collstartpos, collendpos, reason); |
| return -1; |
| case 2: /* replace */ |
| for (collpos = collstartpos; collpos<collendpos; ++collpos) { |
| x = charmapencode_output('?', mapping, res, respos); |
| if (x==enc_EXCEPTION) { |
| return -1; |
| } |
| else if (x==enc_FAILED) { |
| raise_encode_exception(exceptionObject, encoding, p, size, collstartpos, collendpos, reason); |
| return -1; |
| } |
| } |
| /* fall through */ |
| case 3: /* ignore */ |
| *inpos = collendpos; |
| break; |
| case 4: /* xmlcharrefreplace */ |
| /* generate replacement (temporarily (mis)uses p) */ |
| for (collpos = collstartpos; collpos < collendpos; ++collpos) { |
| char buffer[2+29+1+1]; |
| char *cp; |
| sprintf(buffer, "&#%d;", (int)p[collpos]); |
| for (cp = buffer; *cp; ++cp) { |
| x = charmapencode_output(*cp, mapping, res, respos); |
| if (x==enc_EXCEPTION) |
| return -1; |
| else if (x==enc_FAILED) { |
| raise_encode_exception(exceptionObject, encoding, p, size, collstartpos, collendpos, reason); |
| return -1; |
| } |
| } |
| } |
| *inpos = collendpos; |
| break; |
| default: |
| repunicode = unicode_encode_call_errorhandler(errors, errorHandler, |
| encoding, reason, p, size, exceptionObject, |
| collstartpos, collendpos, &newpos); |
| if (repunicode == NULL) |
| return -1; |
| if (PyBytes_Check(repunicode)) { |
| /* Directly copy bytes result to output. */ |
| Py_ssize_t outsize = PyBytes_Size(*res); |
| Py_ssize_t requiredsize; |
| repsize = PyBytes_Size(repunicode); |
| requiredsize = *respos + repsize; |
| if (requiredsize > outsize) |
| /* Make room for all additional bytes. */ |
| if (charmapencode_resize(res, respos, requiredsize)) { |
| Py_DECREF(repunicode); |
| return -1; |
| } |
| memcpy(PyBytes_AsString(*res) + *respos, |
| PyBytes_AsString(repunicode), repsize); |
| *respos += repsize; |
| *inpos = newpos; |
| Py_DECREF(repunicode); |
| break; |
| } |
| /* generate replacement */ |
| repsize = PyUnicode_GET_SIZE(repunicode); |
| for (uni2 = PyUnicode_AS_UNICODE(repunicode); repsize-->0; ++uni2) { |
| x = charmapencode_output(*uni2, mapping, res, respos); |
| if (x==enc_EXCEPTION) { |
| return -1; |
| } |
| else if (x==enc_FAILED) { |
| Py_DECREF(repunicode); |
| raise_encode_exception(exceptionObject, encoding, p, size, collstartpos, collendpos, reason); |
| return -1; |
| } |
| } |
| *inpos = newpos; |
| Py_DECREF(repunicode); |
| } |
| return 0; |
| } |
| |
| PyObject *PyUnicode_EncodeCharmap(const Py_UNICODE *p, |
| Py_ssize_t size, |
| PyObject *mapping, |
| const char *errors) |
| { |
| /* output object */ |
| PyObject *res = NULL; |
| /* current input position */ |
| Py_ssize_t inpos = 0; |
| /* current output position */ |
| Py_ssize_t respos = 0; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| /* the following variable is used for caching string comparisons |
| * -1=not initialized, 0=unknown, 1=strict, 2=replace, |
| * 3=ignore, 4=xmlcharrefreplace */ |
| int known_errorHandler = -1; |
| |
| /* Default to Latin-1 */ |
| if (mapping == NULL) |
| return PyUnicode_EncodeLatin1(p, size, errors); |
| |
| /* allocate enough for a simple encoding without |
| replacements, if we need more, we'll resize */ |
| res = PyBytes_FromStringAndSize(NULL, size); |
| if (res == NULL) |
| goto onError; |
| if (size == 0) |
| return res; |
| |
| while (inpos<size) { |
| /* try to encode it */ |
| charmapencode_result x = charmapencode_output(p[inpos], mapping, &res, &respos); |
| if (x==enc_EXCEPTION) /* error */ |
| goto onError; |
| if (x==enc_FAILED) { /* unencodable character */ |
| if (charmap_encoding_error(p, size, &inpos, mapping, |
| &exc, |
| &known_errorHandler, &errorHandler, errors, |
| &res, &respos)) { |
| goto onError; |
| } |
| } |
| else |
| /* done with this character => adjust input position */ |
| ++inpos; |
| } |
| |
| /* Resize if we allocated to much */ |
| if (respos<PyBytes_GET_SIZE(res)) |
| if (_PyBytes_Resize(&res, respos) < 0) |
| goto onError; |
| |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return res; |
| |
| onError: |
| Py_XDECREF(res); |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_AsCharmapString(PyObject *unicode, |
| PyObject *mapping) |
| { |
| if (!PyUnicode_Check(unicode) || mapping == NULL) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| return PyUnicode_EncodeCharmap(PyUnicode_AS_UNICODE(unicode), |
| PyUnicode_GET_SIZE(unicode), |
| mapping, |
| NULL); |
| } |
| |
| /* create or adjust a UnicodeTranslateError */ |
| static void make_translate_exception(PyObject **exceptionObject, |
| const Py_UNICODE *unicode, Py_ssize_t size, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason) |
| { |
| if (*exceptionObject == NULL) { |
| *exceptionObject = PyUnicodeTranslateError_Create( |
| unicode, size, startpos, endpos, reason); |
| } |
| else { |
| if (PyUnicodeTranslateError_SetStart(*exceptionObject, startpos)) |
| goto onError; |
| if (PyUnicodeTranslateError_SetEnd(*exceptionObject, endpos)) |
| goto onError; |
| if (PyUnicodeTranslateError_SetReason(*exceptionObject, reason)) |
| goto onError; |
| return; |
| onError: |
| Py_DECREF(*exceptionObject); |
| *exceptionObject = NULL; |
| } |
| } |
| |
| /* raises a UnicodeTranslateError */ |
| static void raise_translate_exception(PyObject **exceptionObject, |
| const Py_UNICODE *unicode, Py_ssize_t size, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| const char *reason) |
| { |
| make_translate_exception(exceptionObject, |
| unicode, size, startpos, endpos, reason); |
| if (*exceptionObject != NULL) |
| PyCodec_StrictErrors(*exceptionObject); |
| } |
| |
| /* error handling callback helper: |
| build arguments, call the callback and check the arguments, |
| put the result into newpos and return the replacement string, which |
| has to be freed by the caller */ |
| static PyObject *unicode_translate_call_errorhandler(const char *errors, |
| PyObject **errorHandler, |
| const char *reason, |
| const Py_UNICODE *unicode, Py_ssize_t size, PyObject **exceptionObject, |
| Py_ssize_t startpos, Py_ssize_t endpos, |
| Py_ssize_t *newpos) |
| { |
| static char *argparse = "O!n;translating error handler must return (str, int) tuple"; |
| |
| Py_ssize_t i_newpos; |
| PyObject *restuple; |
| PyObject *resunicode; |
| |
| if (*errorHandler == NULL) { |
| *errorHandler = PyCodec_LookupError(errors); |
| if (*errorHandler == NULL) |
| return NULL; |
| } |
| |
| make_translate_exception(exceptionObject, |
| unicode, size, startpos, endpos, reason); |
| if (*exceptionObject == NULL) |
| return NULL; |
| |
| restuple = PyObject_CallFunctionObjArgs( |
| *errorHandler, *exceptionObject, NULL); |
| if (restuple == NULL) |
| return NULL; |
| if (!PyTuple_Check(restuple)) { |
| PyErr_SetString(PyExc_TypeError, &argparse[4]); |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| if (!PyArg_ParseTuple(restuple, argparse, &PyUnicode_Type, |
| &resunicode, &i_newpos)) { |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| if (i_newpos<0) |
| *newpos = size+i_newpos; |
| else |
| *newpos = i_newpos; |
| if (*newpos<0 || *newpos>size) { |
| PyErr_Format(PyExc_IndexError, "position %zd from error handler out of bounds", *newpos); |
| Py_DECREF(restuple); |
| return NULL; |
| } |
| Py_INCREF(resunicode); |
| Py_DECREF(restuple); |
| return resunicode; |
| } |
| |
| /* Lookup the character ch in the mapping and put the result in result, |
| which must be decrefed by the caller. |
| Return 0 on success, -1 on error */ |
| static |
| int charmaptranslate_lookup(Py_UNICODE c, PyObject *mapping, PyObject **result) |
| { |
| PyObject *w = PyLong_FromLong((long)c); |
| PyObject *x; |
| |
| if (w == NULL) |
| return -1; |
| x = PyObject_GetItem(mapping, w); |
| Py_DECREF(w); |
| if (x == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_LookupError)) { |
| /* No mapping found means: use 1:1 mapping. */ |
| PyErr_Clear(); |
| *result = NULL; |
| return 0; |
| } else |
| return -1; |
| } |
| else if (x == Py_None) { |
| *result = x; |
| return 0; |
| } |
| else if (PyLong_Check(x)) { |
| long value = PyLong_AS_LONG(x); |
| long max = PyUnicode_GetMax(); |
| if (value < 0 || value > max) { |
| PyErr_Format(PyExc_TypeError, |
| "character mapping must be in range(0x%x)", max+1); |
| Py_DECREF(x); |
| return -1; |
| } |
| *result = x; |
| return 0; |
| } |
| else if (PyUnicode_Check(x)) { |
| *result = x; |
| return 0; |
| } |
| else { |
| /* wrong return value */ |
| PyErr_SetString(PyExc_TypeError, |
| "character mapping must return integer, None or str"); |
| Py_DECREF(x); |
| return -1; |
| } |
| } |
| /* ensure that *outobj is at least requiredsize characters long, |
| if not reallocate and adjust various state variables. |
| Return 0 on success, -1 on error */ |
| static |
| int charmaptranslate_makespace(PyObject **outobj, Py_UNICODE **outp, |
| Py_ssize_t requiredsize) |
| { |
| Py_ssize_t oldsize = PyUnicode_GET_SIZE(*outobj); |
| if (requiredsize > oldsize) { |
| /* remember old output position */ |
| Py_ssize_t outpos = *outp-PyUnicode_AS_UNICODE(*outobj); |
| /* exponentially overallocate to minimize reallocations */ |
| if (requiredsize < 2 * oldsize) |
| requiredsize = 2 * oldsize; |
| if (PyUnicode_Resize(outobj, requiredsize) < 0) |
| return -1; |
| *outp = PyUnicode_AS_UNICODE(*outobj) + outpos; |
| } |
| return 0; |
| } |
| /* lookup the character, put the result in the output string and adjust |
| various state variables. Return a new reference to the object that |
| was put in the output buffer in *result, or Py_None, if the mapping was |
| undefined (in which case no character was written). |
| The called must decref result. |
| Return 0 on success, -1 on error. */ |
| static |
| int charmaptranslate_output(const Py_UNICODE *startinp, const Py_UNICODE *curinp, |
| Py_ssize_t insize, PyObject *mapping, PyObject **outobj, Py_UNICODE **outp, |
| PyObject **res) |
| { |
| if (charmaptranslate_lookup(*curinp, mapping, res)) |
| return -1; |
| if (*res==NULL) { |
| /* not found => default to 1:1 mapping */ |
| *(*outp)++ = *curinp; |
| } |
| else if (*res==Py_None) |
| ; |
| else if (PyLong_Check(*res)) { |
| /* no overflow check, because we know that the space is enough */ |
| *(*outp)++ = (Py_UNICODE)PyLong_AS_LONG(*res); |
| } |
| else if (PyUnicode_Check(*res)) { |
| Py_ssize_t repsize = PyUnicode_GET_SIZE(*res); |
| if (repsize==1) { |
| /* no overflow check, because we know that the space is enough */ |
| *(*outp)++ = *PyUnicode_AS_UNICODE(*res); |
| } |
| else if (repsize!=0) { |
| /* more than one character */ |
| Py_ssize_t requiredsize = (*outp-PyUnicode_AS_UNICODE(*outobj)) + |
| (insize - (curinp-startinp)) + |
| repsize - 1; |
| if (charmaptranslate_makespace(outobj, outp, requiredsize)) |
| return -1; |
| memcpy(*outp, PyUnicode_AS_UNICODE(*res), sizeof(Py_UNICODE)*repsize); |
| *outp += repsize; |
| } |
| } |
| else |
| return -1; |
| return 0; |
| } |
| |
| PyObject *PyUnicode_TranslateCharmap(const Py_UNICODE *p, |
| Py_ssize_t size, |
| PyObject *mapping, |
| const char *errors) |
| { |
| /* output object */ |
| PyObject *res = NULL; |
| /* pointers to the beginning and end+1 of input */ |
| const Py_UNICODE *startp = p; |
| const Py_UNICODE *endp = p + size; |
| /* pointer into the output */ |
| Py_UNICODE *str; |
| /* current output position */ |
| Py_ssize_t respos = 0; |
| char *reason = "character maps to <undefined>"; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| /* the following variable is used for caching string comparisons |
| * -1=not initialized, 0=unknown, 1=strict, 2=replace, |
| * 3=ignore, 4=xmlcharrefreplace */ |
| int known_errorHandler = -1; |
| |
| if (mapping == NULL) { |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| |
| /* allocate enough for a simple 1:1 translation without |
| replacements, if we need more, we'll resize */ |
| res = PyUnicode_FromUnicode(NULL, size); |
| if (res == NULL) |
| goto onError; |
| if (size == 0) |
| return res; |
| str = PyUnicode_AS_UNICODE(res); |
| |
| while (p<endp) { |
| /* try to encode it */ |
| PyObject *x = NULL; |
| if (charmaptranslate_output(startp, p, size, mapping, &res, &str, &x)) { |
| Py_XDECREF(x); |
| goto onError; |
| } |
| Py_XDECREF(x); |
| if (x!=Py_None) /* it worked => adjust input pointer */ |
| ++p; |
| else { /* untranslatable character */ |
| PyObject *repunicode = NULL; /* initialize to prevent gcc warning */ |
| Py_ssize_t repsize; |
| Py_ssize_t newpos; |
| Py_UNICODE *uni2; |
| /* startpos for collecting untranslatable chars */ |
| const Py_UNICODE *collstart = p; |
| const Py_UNICODE *collend = p+1; |
| const Py_UNICODE *coll; |
| |
| /* find all untranslatable characters */ |
| while (collend < endp) { |
| if (charmaptranslate_lookup(*collend, mapping, &x)) |
| goto onError; |
| Py_XDECREF(x); |
| if (x!=Py_None) |
| break; |
| ++collend; |
| } |
| /* cache callback name lookup |
| * (if not done yet, i.e. it's the first error) */ |
| if (known_errorHandler==-1) { |
| if ((errors==NULL) || (!strcmp(errors, "strict"))) |
| known_errorHandler = 1; |
| else if (!strcmp(errors, "replace")) |
| known_errorHandler = 2; |
| else if (!strcmp(errors, "ignore")) |
| known_errorHandler = 3; |
| else if (!strcmp(errors, "xmlcharrefreplace")) |
| known_errorHandler = 4; |
| else |
| known_errorHandler = 0; |
| } |
| switch (known_errorHandler) { |
| case 1: /* strict */ |
| raise_translate_exception(&exc, startp, size, collstart-startp, collend-startp, reason); |
| goto onError; |
| case 2: /* replace */ |
| /* No need to check for space, this is a 1:1 replacement */ |
| for (coll = collstart; coll<collend; ++coll) |
| *str++ = '?'; |
| /* fall through */ |
| case 3: /* ignore */ |
| p = collend; |
| break; |
| case 4: /* xmlcharrefreplace */ |
| /* generate replacement (temporarily (mis)uses p) */ |
| for (p = collstart; p < collend; ++p) { |
| char buffer[2+29+1+1]; |
| char *cp; |
| sprintf(buffer, "&#%d;", (int)*p); |
| if (charmaptranslate_makespace(&res, &str, |
| (str-PyUnicode_AS_UNICODE(res))+strlen(buffer)+(endp-collend))) |
| goto onError; |
| for (cp = buffer; *cp; ++cp) |
| *str++ = *cp; |
| } |
| p = collend; |
| break; |
| default: |
| repunicode = unicode_translate_call_errorhandler(errors, &errorHandler, |
| reason, startp, size, &exc, |
| collstart-startp, collend-startp, &newpos); |
| if (repunicode == NULL) |
| goto onError; |
| /* generate replacement */ |
| repsize = PyUnicode_GET_SIZE(repunicode); |
| if (charmaptranslate_makespace(&res, &str, |
| (str-PyUnicode_AS_UNICODE(res))+repsize+(endp-collend))) { |
| Py_DECREF(repunicode); |
| goto onError; |
| } |
| for (uni2 = PyUnicode_AS_UNICODE(repunicode); repsize-->0; ++uni2) |
| *str++ = *uni2; |
| p = startp + newpos; |
| Py_DECREF(repunicode); |
| } |
| } |
| } |
| /* Resize if we allocated to much */ |
| respos = str-PyUnicode_AS_UNICODE(res); |
| if (respos<PyUnicode_GET_SIZE(res)) { |
| if (PyUnicode_Resize(&res, respos) < 0) |
| goto onError; |
| } |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return res; |
| |
| onError: |
| Py_XDECREF(res); |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return NULL; |
| } |
| |
| PyObject *PyUnicode_Translate(PyObject *str, |
| PyObject *mapping, |
| const char *errors) |
| { |
| PyObject *result; |
| |
| str = PyUnicode_FromObject(str); |
| if (str == NULL) |
| goto onError; |
| result = PyUnicode_TranslateCharmap(PyUnicode_AS_UNICODE(str), |
| PyUnicode_GET_SIZE(str), |
| mapping, |
| errors); |
| Py_DECREF(str); |
| return result; |
| |
| onError: |
| Py_XDECREF(str); |
| return NULL; |
| } |
| |
| PyObject * |
| PyUnicode_TransformDecimalToASCII(Py_UNICODE *s, |
| Py_ssize_t length) |
| { |
| PyObject *result; |
| Py_UNICODE *p; /* write pointer into result */ |
| Py_ssize_t i; |
| /* Copy to a new string */ |
| result = (PyObject *)_PyUnicode_New(length); |
| Py_UNICODE_COPY(PyUnicode_AS_UNICODE(result), s, length); |
| if (result == NULL) |
| return result; |
| p = PyUnicode_AS_UNICODE(result); |
| /* Iterate over code points */ |
| for (i = 0; i < length; i++) { |
| Py_UNICODE ch =s[i]; |
| if (ch > 127) { |
| int decimal = Py_UNICODE_TODECIMAL(ch); |
| if (decimal >= 0) |
| p[i] = '0' + decimal; |
| } |
| } |
| return result; |
| } |
| /* --- Decimal Encoder ---------------------------------------------------- */ |
| |
| int PyUnicode_EncodeDecimal(Py_UNICODE *s, |
| Py_ssize_t length, |
| char *output, |
| const char *errors) |
| { |
| Py_UNICODE *p, *end; |
| PyObject *errorHandler = NULL; |
| PyObject *exc = NULL; |
| const char *encoding = "decimal"; |
| const char *reason = "invalid decimal Unicode string"; |
| /* the following variable is used for caching string comparisons |
| * -1=not initialized, 0=unknown, 1=strict, 2=replace, 3=ignore, 4=xmlcharrefreplace */ |
| int known_errorHandler = -1; |
| |
| if (output == NULL) { |
| PyErr_BadArgument(); |
| return -1; |
| } |
| |
| p = s; |
| end = s + length; |
| while (p < end) { |
| register Py_UNICODE ch = *p; |
| int decimal; |
| PyObject *repunicode; |
| Py_ssize_t repsize; |
| Py_ssize_t newpos; |
| Py_UNICODE *uni2; |
| Py_UNICODE *collstart; |
| Py_UNICODE *collend; |
| |
| if (Py_UNICODE_ISSPACE(ch)) { |
| *output++ = ' '; |
| ++p; |
| continue; |
| } |
| decimal = Py_UNICODE_TODECIMAL(ch); |
| if (decimal >= 0) { |
| *output++ = '0' + decimal; |
| ++p; |
| continue; |
| } |
| if (0 < ch && ch < 256) { |
| *output++ = (char)ch; |
| ++p; |
| continue; |
| } |
| /* All other characters are considered unencodable */ |
| collstart = p; |
| for (collend = p+1; collend < end; collend++) { |
| if ((0 < *collend && *collend < 256) || |
| Py_UNICODE_ISSPACE(*collend) || |
| 0 <= Py_UNICODE_TODECIMAL(*collend)) |
| break; |
| } |
| /* cache callback name lookup |
| * (if not done yet, i.e. it's the first error) */ |
| if (known_errorHandler==-1) { |
| if ((errors==NULL) || (!strcmp(errors, "strict"))) |
| known_errorHandler = 1; |
| else if (!strcmp(errors, "replace")) |
| known_errorHandler = 2; |
| else if (!strcmp(errors, "ignore")) |
| known_errorHandler = 3; |
| else if (!strcmp(errors, "xmlcharrefreplace")) |
| known_errorHandler = 4; |
| else |
| known_errorHandler = 0; |
| } |
| switch (known_errorHandler) { |
| case 1: /* strict */ |
| raise_encode_exception(&exc, encoding, s, length, collstart-s, collend-s, reason); |
| goto onError; |
| case 2: /* replace */ |
| for (p = collstart; p < collend; ++p) |
| *output++ = '?'; |
| /* fall through */ |
| case 3: /* ignore */ |
| p = collend; |
| break; |
| case 4: /* xmlcharrefreplace */ |
| /* generate replacement (temporarily (mis)uses p) */ |
| for (p = collstart; p < collend; ++p) |
| output += sprintf(output, "&#%d;", (int)*p); |
| p = collend; |
| break; |
| default: |
| repunicode = unicode_encode_call_errorhandler(errors, &errorHandler, |
| encoding, reason, s, length, &exc, |
| collstart-s, collend-s, &newpos); |
| if (repunicode == NULL) |
| goto onError; |
| if (!PyUnicode_Check(repunicode)) { |
| /* Byte results not supported, since they have no decimal property. */ |
| PyErr_SetString(PyExc_TypeError, "error handler should return unicode"); |
| Py_DECREF(repunicode); |
| goto onError; |
| } |
| /* generate replacement */ |
| repsize = PyUnicode_GET_SIZE(repunicode); |
| for (uni2 = PyUnicode_AS_UNICODE(repunicode); repsize-->0; ++uni2) { |
| Py_UNICODE ch = *uni2; |
| if (Py_UNICODE_ISSPACE(ch)) |
| *output++ = ' '; |
| else { |
| decimal = Py_UNICODE_TODECIMAL(ch); |
| if (decimal >= 0) |
| *output++ = '0' + decimal; |
| else if (0 < ch && ch < 256) |
| *output++ = (char)ch; |
| else { |
| Py_DECREF(repunicode); |
| raise_encode_exception(&exc, encoding, |
| s, length, collstart-s, collend-s, reason); |
| goto onError; |
| } |
| } |
| } |
| p = s + newpos; |
| Py_DECREF(repunicode); |
| } |
| } |
| /* 0-terminate the output string */ |
| *output++ = '\0'; |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return 0; |
| |
| onError: |
| Py_XDECREF(exc); |
| Py_XDECREF(errorHandler); |
| return -1; |
| } |
| |
| /* --- Helpers ------------------------------------------------------------ */ |
| |
| #include "stringlib/unicodedefs.h" |
| #include "stringlib/fastsearch.h" |
| |
| #include "stringlib/count.h" |
| #include "stringlib/find.h" |
| #include "stringlib/partition.h" |
| #include "stringlib/split.h" |
| |
| #define _Py_InsertThousandsGrouping _PyUnicode_InsertThousandsGrouping |
| #define _Py_InsertThousandsGroupingLocale _PyUnicode_InsertThousandsGroupingLocale |
| #include "stringlib/localeutil.h" |
| |
| /* helper macro to fixup start/end slice values */ |
| #define ADJUST_INDICES(start, end, len) \ |
| if (end > len) \ |
| end = len; \ |
| else if (end < 0) { \ |
| end += len; \ |
| if (end < 0) \ |
| end = 0; \ |
| } \ |
| if (start < 0) { \ |
| start += len; \ |
| if (start < 0) \ |
| start = 0; \ |
| } |
| |
| /* _Py_UNICODE_NEXT is a private macro used to retrieve the character pointed |
| * by 'ptr', possibly combining surrogate pairs on narrow builds. |
| * 'ptr' and 'end' must be Py_UNICODE*, with 'ptr' pointing at the character |
| * that should be returned and 'end' pointing to the end of the buffer. |
| * ('end' is used on narrow builds to detect a lone surrogate at the |
| * end of the buffer that should be returned unchanged.) |
| * The ptr and end arguments should be side-effect free and ptr must an lvalue. |
| * The type of the returned char is always Py_UCS4. |
| * |
| * Note: the macro advances ptr to next char, so it might have side-effects |
| * (especially if used with other macros). |
| */ |
| |
| /* helper macros used by _Py_UNICODE_NEXT */ |
| #define _Py_UNICODE_IS_HIGH_SURROGATE(ch) (0xD800 <= ch && ch <= 0xDBFF) |
| #define _Py_UNICODE_IS_LOW_SURROGATE(ch) (0xDC00 <= ch && ch <= 0xDFFF) |
| /* Join two surrogate characters and return a single Py_UCS4 value. */ |
| #define _Py_UNICODE_JOIN_SURROGATES(high, low) \ |
| (((((Py_UCS4)(high) & 0x03FF) << 10) | \ |
| ((Py_UCS4)(low) & 0x03FF)) + 0x10000) |
| |
| #ifdef Py_UNICODE_WIDE |
| #define _Py_UNICODE_NEXT(ptr, end) *(ptr)++ |
| #else |
| #define _Py_UNICODE_NEXT(ptr, end) \ |
| (((_Py_UNICODE_IS_HIGH_SURROGATE(*(ptr)) && (ptr) < (end)) && \ |
| _Py_UNICODE_IS_LOW_SURROGATE((ptr)[1])) ? \ |
| ((ptr) += 2,_Py_UNICODE_JOIN_SURROGATES((ptr)[-2], (ptr)[-1])) : \ |
| (Py_UCS4)*(ptr)++) |
| #endif |
| |
| Py_ssize_t PyUnicode_Count(PyObject *str, |
| PyObject *substr, |
| Py_ssize_t start, |
| Py_ssize_t end) |
| { |
| Py_ssize_t result; |
| PyUnicodeObject* str_obj; |
| PyUnicodeObject* sub_obj; |
| |
| str_obj = (PyUnicodeObject*) PyUnicode_FromObject(str); |
| if (!str_obj) |
| return -1; |
| sub_obj = (PyUnicodeObject*) PyUnicode_FromObject(substr); |
| if (!sub_obj) { |
| Py_DECREF(str_obj); |
| return -1; |
| } |
| |
| ADJUST_INDICES(start, end, str_obj->length); |
| result = stringlib_count( |
| str_obj->str + start, end - start, sub_obj->str, sub_obj->length, |
| PY_SSIZE_T_MAX |
| ); |
| |
| Py_DECREF(sub_obj); |
| Py_DECREF(str_obj); |
| |
| return result; |
| } |
| |
| Py_ssize_t PyUnicode_Find(PyObject *str, |
| PyObject *sub, |
| Py_ssize_t start, |
| Py_ssize_t end, |
| int direction) |
| { |
| Py_ssize_t result; |
| |
| str = PyUnicode_FromObject(str); |
| if (!str) |
| return -2; |
| sub = PyUnicode_FromObject(sub); |
| if (!sub) { |
| Py_DECREF(str); |
| return -2; |
| } |
| |
| if (direction > 0) |
| result = stringlib_find_slice( |
| PyUnicode_AS_UNICODE(str), PyUnicode_GET_SIZE(str), |
| PyUnicode_AS_UNICODE(sub), PyUnicode_GET_SIZE(sub), |
| start, end |
| ); |
| else |
| result = stringlib_rfind_slice( |
| PyUnicode_AS_UNICODE(str), PyUnicode_GET_SIZE(str), |
| PyUnicode_AS_UNICODE(sub), PyUnicode_GET_SIZE(sub), |
| start, end |
| ); |
| |
| Py_DECREF(str); |
| Py_DECREF(sub); |
| |
| return result; |
| } |
| |
| static |
| int tailmatch(PyUnicodeObject *self, |
| PyUnicodeObject *substring, |
| Py_ssize_t start, |
| Py_ssize_t end, |
| int direction) |
| { |
| if (substring->length == 0) |
| return 1; |
| |
| ADJUST_INDICES(start, end, self->length); |
| end -= substring->length; |
| if (end < start) |
| return 0; |
| |
| if (direction > 0) { |
| if (Py_UNICODE_MATCH(self, end, substring)) |
| return 1; |
| } else { |
| if (Py_UNICODE_MATCH(self, start, substring)) |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| Py_ssize_t PyUnicode_Tailmatch(PyObject *str, |
| PyObject *substr, |
| Py_ssize_t start, |
| Py_ssize_t end, |
| int direction) |
| { |
| Py_ssize_t result; |
| |
| str = PyUnicode_FromObject(str); |
| if (str == NULL) |
| return -1; |
| substr = PyUnicode_FromObject(substr); |
| if (substr == NULL) { |
| Py_DECREF(str); |
| return -1; |
| } |
| |
| result = tailmatch((PyUnicodeObject *)str, |
| (PyUnicodeObject *)substr, |
| start, end, direction); |
| Py_DECREF(str); |
| Py_DECREF(substr); |
| return result; |
| } |
| |
| /* Apply fixfct filter to the Unicode object self and return a |
| reference to the modified object */ |
| |
| static |
| PyObject *fixup(PyUnicodeObject *self, |
| int (*fixfct)(PyUnicodeObject *s)) |
| { |
| |
| PyUnicodeObject *u; |
| |
| u = (PyUnicodeObject*) PyUnicode_FromUnicode(NULL, self->length); |
| if (u == NULL) |
| return NULL; |
| |
| Py_UNICODE_COPY(u->str, self->str, self->length); |
| |
| if (!fixfct(u) && PyUnicode_CheckExact(self)) { |
| /* fixfct should return TRUE if it modified the buffer. If |
| FALSE, return a reference to the original buffer instead |
| (to save space, not time) */ |
| Py_INCREF(self); |
| Py_DECREF(u); |
| return (PyObject*) self; |
| } |
| return (PyObject*) u; |
| } |
| |
| static |
| int fixupper(PyUnicodeObject *self) |
| { |
| Py_ssize_t len = self->length; |
| Py_UNICODE *s = self->str; |
| int status = 0; |
| |
| while (len-- > 0) { |
| register Py_UNICODE ch; |
| |
| ch = Py_UNICODE_TOUPPER(*s); |
| if (ch != *s) { |
| status = 1; |
| *s = ch; |
| } |
| s++; |
| } |
| |
| return status; |
| } |
| |
| static |
| int fixlower(PyUnicodeObject *self) |
| { |
| Py_ssize_t len = self->length; |
| Py_UNICODE *s = self->str; |
| int status = 0; |
| |
| while (len-- > 0) { |
| register Py_UNICODE ch; |
| |
| ch = Py_UNICODE_TOLOWER(*s); |
| if (ch != *s) { |
| status = 1; |
| *s = ch; |
| } |
| s++; |
| } |
| |
| return status; |
| } |
| |
| static |
| int fixswapcase(PyUnicodeObject *self) |
| { |
| Py_ssize_t len = self->length; |
| Py_UNICODE *s = self->str; |
| int status = 0; |
| |
| while (len-- > 0) { |
| if (Py_UNICODE_ISUPPER(*s)) { |
| *s = Py_UNICODE_TOLOWER(*s); |
| status = 1; |
| } else if (Py_UNICODE_ISLOWER(*s)) { |
| *s = Py_UNICODE_TOUPPER(*s); |
| status = 1; |
| } |
| s++; |
| } |
| |
| return status; |
| } |
| |
| static |
| int fixcapitalize(PyUnicodeObject *self) |
| { |
| Py_ssize_t len = self->length; |
| Py_UNICODE *s = self->str; |
| int status = 0; |
| |
| if (len == 0) |
| return 0; |
| if (!Py_UNICODE_ISUPPER(*s)) { |
| *s = Py_UNICODE_TOUPPER(*s); |
| status = 1; |
| } |
| s++; |
| while (--len > 0) { |
| if (!Py_UNICODE_ISLOWER(*s)) { |
| *s = Py_UNICODE_TOLOWER(*s); |
| status = 1; |
| } |
| s++; |
| } |
| return status; |
| } |
| |
| static |
| int fixtitle(PyUnicodeObject *self) |
| { |
| register Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register Py_UNICODE *e; |
| int previous_is_cased; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1) { |
| Py_UNICODE ch = Py_UNICODE_TOTITLE(*p); |
| if (*p != ch) { |
| *p = ch; |
| return 1; |
| } |
| else |
| return 0; |
| } |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| previous_is_cased = 0; |
| for (; p < e; p++) { |
| register const Py_UNICODE ch = *p; |
| |
| if (previous_is_cased) |
| *p = Py_UNICODE_TOLOWER(ch); |
| else |
| *p = Py_UNICODE_TOTITLE(ch); |
| |
| if (Py_UNICODE_ISLOWER(ch) || |
| Py_UNICODE_ISUPPER(ch) || |
| Py_UNICODE_ISTITLE(ch)) |
| previous_is_cased = 1; |
| else |
| previous_is_cased = 0; |
| } |
| return 1; |
| } |
| |
| PyObject * |
| PyUnicode_Join(PyObject *separator, PyObject *seq) |
| { |
| const Py_UNICODE blank = ' '; |
| const Py_UNICODE *sep = ␣ |
| Py_ssize_t seplen = 1; |
| PyUnicodeObject *res = NULL; /* the result */ |
| Py_UNICODE *res_p; /* pointer to free byte in res's string area */ |
| PyObject *fseq; /* PySequence_Fast(seq) */ |
| Py_ssize_t seqlen; /* len(fseq) -- number of items in sequence */ |
| PyObject **items; |
| PyObject *item; |
| Py_ssize_t sz, i; |
| |
| fseq = PySequence_Fast(seq, ""); |
| if (fseq == NULL) { |
| return NULL; |
| } |
| |
| /* NOTE: the following code can't call back into Python code, |
| * so we are sure that fseq won't be mutated. |
| */ |
| |
| seqlen = PySequence_Fast_GET_SIZE(fseq); |
| /* If empty sequence, return u"". */ |
| if (seqlen == 0) { |
| res = _PyUnicode_New(0); /* empty sequence; return u"" */ |
| goto Done; |
| } |
| items = PySequence_Fast_ITEMS(fseq); |
| /* If singleton sequence with an exact Unicode, return that. */ |
| if (seqlen == 1) { |
| item = items[0]; |
| if (PyUnicode_CheckExact(item)) { |
| Py_INCREF(item); |
| res = (PyUnicodeObject *)item; |
| goto Done; |
| } |
| } |
| else { |
| /* Set up sep and seplen */ |
| if (separator == NULL) { |
| sep = ␣ |
| seplen = 1; |
| } |
| else { |
| if (!PyUnicode_Check(separator)) { |
| PyErr_Format(PyExc_TypeError, |
| "separator: expected str instance," |
| " %.80s found", |
| Py_TYPE(separator)->tp_name); |
| goto onError; |
| } |
| sep = PyUnicode_AS_UNICODE(separator); |
| seplen = PyUnicode_GET_SIZE(separator); |
| } |
| } |
| |
| /* There are at least two things to join, or else we have a subclass |
| * of str in the sequence. |
| * Do a pre-pass to figure out the total amount of space we'll |
| * need (sz), and see whether all argument are strings. |
| */ |
| sz = 0; |
| for (i = 0; i < seqlen; i++) { |
| const Py_ssize_t old_sz = sz; |
| item = items[i]; |
| if (!PyUnicode_Check(item)) { |
| PyErr_Format(PyExc_TypeError, |
| "sequence item %zd: expected str instance," |
| " %.80s found", |
| i, Py_TYPE(item)->tp_name); |
| goto onError; |
| } |
| sz += PyUnicode_GET_SIZE(item); |
| if (i != 0) |
| sz += seplen; |
| if (sz < old_sz || sz > PY_SSIZE_T_MAX) { |
| PyErr_SetString(PyExc_OverflowError, |
| "join() result is too long for a Python string"); |
| goto onError; |
| } |
| } |
| |
| res = _PyUnicode_New(sz); |
| if (res == NULL) |
| goto onError; |
| |
| /* Catenate everything. */ |
| res_p = PyUnicode_AS_UNICODE(res); |
| for (i = 0; i < seqlen; ++i) { |
| Py_ssize_t itemlen; |
| item = items[i]; |
| itemlen = PyUnicode_GET_SIZE(item); |
| /* Copy item, and maybe the separator. */ |
| if (i) { |
| Py_UNICODE_COPY(res_p, sep, seplen); |
| res_p += seplen; |
| } |
| Py_UNICODE_COPY(res_p, PyUnicode_AS_UNICODE(item), itemlen); |
| res_p += itemlen; |
| } |
| |
| Done: |
| Py_DECREF(fseq); |
| return (PyObject *)res; |
| |
| onError: |
| Py_DECREF(fseq); |
| Py_XDECREF(res); |
| return NULL; |
| } |
| |
| static |
| PyUnicodeObject *pad(PyUnicodeObject *self, |
| Py_ssize_t left, |
| Py_ssize_t right, |
| Py_UNICODE fill) |
| { |
| PyUnicodeObject *u; |
| |
| if (left < 0) |
| left = 0; |
| if (right < 0) |
| right = 0; |
| |
| if (left == 0 && right == 0 && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return self; |
| } |
| |
| if (left > PY_SSIZE_T_MAX - self->length || |
| right > PY_SSIZE_T_MAX - (left + self->length)) { |
| PyErr_SetString(PyExc_OverflowError, "padded string is too long"); |
| return NULL; |
| } |
| u = _PyUnicode_New(left + self->length + right); |
| if (u) { |
| if (left) |
| Py_UNICODE_FILL(u->str, fill, left); |
| Py_UNICODE_COPY(u->str + left, self->str, self->length); |
| if (right) |
| Py_UNICODE_FILL(u->str + left + self->length, fill, right); |
| } |
| |
| return u; |
| } |
| |
| PyObject *PyUnicode_Splitlines(PyObject *string, int keepends) |
| { |
| PyObject *list; |
| |
| string = PyUnicode_FromObject(string); |
| if (string == NULL) |
| return NULL; |
| |
| list = stringlib_splitlines( |
| (PyObject*) string, PyUnicode_AS_UNICODE(string), |
| PyUnicode_GET_SIZE(string), keepends); |
| |
| Py_DECREF(string); |
| return list; |
| } |
| |
| static |
| PyObject *split(PyUnicodeObject *self, |
| PyUnicodeObject *substring, |
| Py_ssize_t maxcount) |
| { |
| if (maxcount < 0) |
| maxcount = PY_SSIZE_T_MAX; |
| |
| if (substring == NULL) |
| return stringlib_split_whitespace( |
| (PyObject*) self, self->str, self->length, maxcount |
| ); |
| |
| return stringlib_split( |
| (PyObject*) self, self->str, self->length, |
| substring->str, substring->length, |
| maxcount |
| ); |
| } |
| |
| static |
| PyObject *rsplit(PyUnicodeObject *self, |
| PyUnicodeObject *substring, |
| Py_ssize_t maxcount) |
| { |
| if (maxcount < 0) |
| maxcount = PY_SSIZE_T_MAX; |
| |
| if (substring == NULL) |
| return stringlib_rsplit_whitespace( |
| (PyObject*) self, self->str, self->length, maxcount |
| ); |
| |
| return stringlib_rsplit( |
| (PyObject*) self, self->str, self->length, |
| substring->str, substring->length, |
| maxcount |
| ); |
| } |
| |
| static |
| PyObject *replace(PyUnicodeObject *self, |
| PyUnicodeObject *str1, |
| PyUnicodeObject *str2, |
| Py_ssize_t maxcount) |
| { |
| PyUnicodeObject *u; |
| |
| if (maxcount < 0) |
| maxcount = PY_SSIZE_T_MAX; |
| else if (maxcount == 0 || self->length == 0) |
| goto nothing; |
| |
| if (str1->length == str2->length) { |
| Py_ssize_t i; |
| /* same length */ |
| if (str1->length == 0) |
| goto nothing; |
| if (str1->length == 1) { |
| /* replace characters */ |
| Py_UNICODE u1, u2; |
| if (!findchar(self->str, self->length, str1->str[0])) |
| goto nothing; |
| u = (PyUnicodeObject*) PyUnicode_FromUnicode(NULL, self->length); |
| if (!u) |
| return NULL; |
| Py_UNICODE_COPY(u->str, self->str, self->length); |
| u1 = str1->str[0]; |
| u2 = str2->str[0]; |
| for (i = 0; i < u->length; i++) |
| if (u->str[i] == u1) { |
| if (--maxcount < 0) |
| break; |
| u->str[i] = u2; |
| } |
| } else { |
| i = stringlib_find( |
| self->str, self->length, str1->str, str1->length, 0 |
| ); |
| if (i < 0) |
| goto nothing; |
| u = (PyUnicodeObject*) PyUnicode_FromUnicode(NULL, self->length); |
| if (!u) |
| return NULL; |
| Py_UNICODE_COPY(u->str, self->str, self->length); |
| |
| /* change everything in-place, starting with this one */ |
| Py_UNICODE_COPY(u->str+i, str2->str, str2->length); |
| i += str1->length; |
| |
| while ( --maxcount > 0) { |
| i = stringlib_find(self->str+i, self->length-i, |
| str1->str, str1->length, |
| i); |
| if (i == -1) |
| break; |
| Py_UNICODE_COPY(u->str+i, str2->str, str2->length); |
| i += str1->length; |
| } |
| } |
| } else { |
| |
| Py_ssize_t n, i, j; |
| Py_ssize_t product, new_size, delta; |
| Py_UNICODE *p; |
| |
| /* replace strings */ |
| n = stringlib_count(self->str, self->length, str1->str, str1->length, |
| maxcount); |
| if (n == 0) |
| goto nothing; |
| /* new_size = self->length + n * (str2->length - str1->length)); */ |
| delta = (str2->length - str1->length); |
| if (delta == 0) { |
| new_size = self->length; |
| } else { |
| product = n * (str2->length - str1->length); |
| if ((product / (str2->length - str1->length)) != n) { |
| PyErr_SetString(PyExc_OverflowError, |
| "replace string is too long"); |
| return NULL; |
| } |
| new_size = self->length + product; |
| if (new_size < 0) { |
| PyErr_SetString(PyExc_OverflowError, |
| "replace string is too long"); |
| return NULL; |
| } |
| } |
| u = _PyUnicode_New(new_size); |
| if (!u) |
| return NULL; |
| i = 0; |
| p = u->str; |
| if (str1->length > 0) { |
| while (n-- > 0) { |
| /* look for next match */ |
| j = stringlib_find(self->str+i, self->length-i, |
| str1->str, str1->length, |
| i); |
| if (j == -1) |
| break; |
| else if (j > i) { |
| /* copy unchanged part [i:j] */ |
| Py_UNICODE_COPY(p, self->str+i, j-i); |
| p += j - i; |
| } |
| /* copy substitution string */ |
| if (str2->length > 0) { |
| Py_UNICODE_COPY(p, str2->str, str2->length); |
| p += str2->length; |
| } |
| i = j + str1->length; |
| } |
| if (i < self->length) |
| /* copy tail [i:] */ |
| Py_UNICODE_COPY(p, self->str+i, self->length-i); |
| } else { |
| /* interleave */ |
| while (n > 0) { |
| Py_UNICODE_COPY(p, str2->str, str2->length); |
| p += str2->length; |
| if (--n <= 0) |
| break; |
| *p++ = self->str[i++]; |
| } |
| Py_UNICODE_COPY(p, self->str+i, self->length-i); |
| } |
| } |
| return (PyObject *) u; |
| |
| nothing: |
| /* nothing to replace; return original string (when possible) */ |
| if (PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject *) self; |
| } |
| return PyUnicode_FromUnicode(self->str, self->length); |
| } |
| |
| /* --- Unicode Object Methods --------------------------------------------- */ |
| |
| PyDoc_STRVAR(title__doc__, |
| "S.title() -> str\n\ |
| \n\ |
| Return a titlecased version of S, i.e. words start with title case\n\ |
| characters, all remaining cased characters have lower case."); |
| |
| static PyObject* |
| unicode_title(PyUnicodeObject *self) |
| { |
| return fixup(self, fixtitle); |
| } |
| |
| PyDoc_STRVAR(capitalize__doc__, |
| "S.capitalize() -> str\n\ |
| \n\ |
| Return a capitalized version of S, i.e. make the first character\n\ |
| have upper case and the rest lower case."); |
| |
| static PyObject* |
| unicode_capitalize(PyUnicodeObject *self) |
| { |
| return fixup(self, fixcapitalize); |
| } |
| |
| #if 0 |
| PyDoc_STRVAR(capwords__doc__, |
| "S.capwords() -> str\n\ |
| \n\ |
| Apply .capitalize() to all words in S and return the result with\n\ |
| normalized whitespace (all whitespace strings are replaced by ' ')."); |
| |
| static PyObject* |
| unicode_capwords(PyUnicodeObject *self) |
| { |
| PyObject *list; |
| PyObject *item; |
| Py_ssize_t i; |
| |
| /* Split into words */ |
| list = split(self, NULL, -1); |
| if (!list) |
| return NULL; |
| |
| /* Capitalize each word */ |
| for (i = 0; i < PyList_GET_SIZE(list); i++) { |
| item = fixup((PyUnicodeObject *)PyList_GET_ITEM(list, i), |
| fixcapitalize); |
| if (item == NULL) |
| goto onError; |
| Py_DECREF(PyList_GET_ITEM(list, i)); |
| PyList_SET_ITEM(list, i, item); |
| } |
| |
| /* Join the words to form a new string */ |
| item = PyUnicode_Join(NULL, list); |
| |
| onError: |
| Py_DECREF(list); |
| return (PyObject *)item; |
| } |
| #endif |
| |
| /* Argument converter. Coerces to a single unicode character */ |
| |
| static int |
| convert_uc(PyObject *obj, void *addr) |
| { |
| Py_UNICODE *fillcharloc = (Py_UNICODE *)addr; |
| PyObject *uniobj; |
| Py_UNICODE *unistr; |
| |
| uniobj = PyUnicode_FromObject(obj); |
| if (uniobj == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "The fill character cannot be converted to Unicode"); |
| return 0; |
| } |
| if (PyUnicode_GET_SIZE(uniobj) != 1) { |
| PyErr_SetString(PyExc_TypeError, |
| "The fill character must be exactly one character long"); |
| Py_DECREF(uniobj); |
| return 0; |
| } |
| unistr = PyUnicode_AS_UNICODE(uniobj); |
| *fillcharloc = unistr[0]; |
| Py_DECREF(uniobj); |
| return 1; |
| } |
| |
| PyDoc_STRVAR(center__doc__, |
| "S.center(width[, fillchar]) -> str\n\ |
| \n\ |
| Return S centered in a string of length width. Padding is\n\ |
| done using the specified fill character (default is a space)"); |
| |
| static PyObject * |
| unicode_center(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_ssize_t marg, left; |
| Py_ssize_t width; |
| Py_UNICODE fillchar = ' '; |
| |
| if (!PyArg_ParseTuple(args, "n|O&:center", &width, convert_uc, &fillchar)) |
| return NULL; |
| |
| if (self->length >= width && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*) self; |
| } |
| |
| marg = width - self->length; |
| left = marg / 2 + (marg & width & 1); |
| |
| return (PyObject*) pad(self, left, marg - left, fillchar); |
| } |
| |
| #if 0 |
| |
| /* This code should go into some future Unicode collation support |
| module. The basic comparison should compare ordinals on a naive |
| basis (this is what Java does and thus Jython too). */ |
| |
| /* speedy UTF-16 code point order comparison */ |
| /* gleaned from: */ |
| /* http://www-4.ibm.com/software/developer/library/utf16.html?dwzone=unicode */ |
| |
| static short utf16Fixup[32] = |
| { |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0, 0, 0, 0, 0, |
| 0, 0, 0, 0x2000, -0x800, -0x800, -0x800, -0x800 |
| }; |
| |
| static int |
| unicode_compare(PyUnicodeObject *str1, PyUnicodeObject *str2) |
| { |
| Py_ssize_t len1, len2; |
| |
| Py_UNICODE *s1 = str1->str; |
| Py_UNICODE *s2 = str2->str; |
| |
| len1 = str1->length; |
| len2 = str2->length; |
| |
| while (len1 > 0 && len2 > 0) { |
| Py_UNICODE c1, c2; |
| |
| c1 = *s1++; |
| c2 = *s2++; |
| |
| if (c1 > (1<<11) * 26) |
| c1 += utf16Fixup[c1>>11]; |
| if (c2 > (1<<11) * 26) |
| c2 += utf16Fixup[c2>>11]; |
| /* now c1 and c2 are in UTF-32-compatible order */ |
| |
| if (c1 != c2) |
| return (c1 < c2) ? -1 : 1; |
| |
| len1--; len2--; |
| } |
| |
| return (len1 < len2) ? -1 : (len1 != len2); |
| } |
| |
| #else |
| |
| static int |
| unicode_compare(PyUnicodeObject *str1, PyUnicodeObject *str2) |
| { |
| register Py_ssize_t len1, len2; |
| |
| Py_UNICODE *s1 = str1->str; |
| Py_UNICODE *s2 = str2->str; |
| |
| len1 = str1->length; |
| len2 = str2->length; |
| |
| while (len1 > 0 && len2 > 0) { |
| Py_UNICODE c1, c2; |
| |
| c1 = *s1++; |
| c2 = *s2++; |
| |
| if (c1 != c2) |
| return (c1 < c2) ? -1 : 1; |
| |
| len1--; len2--; |
| } |
| |
| return (len1 < len2) ? -1 : (len1 != len2); |
| } |
| |
| #endif |
| |
| int PyUnicode_Compare(PyObject *left, |
| PyObject *right) |
| { |
| if (PyUnicode_Check(left) && PyUnicode_Check(right)) |
| return unicode_compare((PyUnicodeObject *)left, |
| (PyUnicodeObject *)right); |
| PyErr_Format(PyExc_TypeError, |
| "Can't compare %.100s and %.100s", |
| left->ob_type->tp_name, |
| right->ob_type->tp_name); |
| return -1; |
| } |
| |
| int |
| PyUnicode_CompareWithASCIIString(PyObject* uni, const char* str) |
| { |
| int i; |
| Py_UNICODE *id; |
| assert(PyUnicode_Check(uni)); |
| id = PyUnicode_AS_UNICODE(uni); |
| /* Compare Unicode string and source character set string */ |
| for (i = 0; id[i] && str[i]; i++) |
| if (id[i] != str[i]) |
| return ((int)id[i] < (int)str[i]) ? -1 : 1; |
| /* This check keeps Python strings that end in '\0' from comparing equal |
| to C strings identical up to that point. */ |
| if (PyUnicode_GET_SIZE(uni) != i || id[i]) |
| return 1; /* uni is longer */ |
| if (str[i]) |
| return -1; /* str is longer */ |
| return 0; |
| } |
| |
| |
| #define TEST_COND(cond) \ |
| ((cond) ? Py_True : Py_False) |
| |
| PyObject *PyUnicode_RichCompare(PyObject *left, |
| PyObject *right, |
| int op) |
| { |
| int result; |
| |
| if (PyUnicode_Check(left) && PyUnicode_Check(right)) { |
| PyObject *v; |
| if (((PyUnicodeObject *) left)->length != |
| ((PyUnicodeObject *) right)->length) { |
| if (op == Py_EQ) { |
| Py_INCREF(Py_False); |
| return Py_False; |
| } |
| if (op == Py_NE) { |
| Py_INCREF(Py_True); |
| return Py_True; |
| } |
| } |
| if (left == right) |
| result = 0; |
| else |
| result = unicode_compare((PyUnicodeObject *)left, |
| (PyUnicodeObject *)right); |
| |
| /* Convert the return value to a Boolean */ |
| switch (op) { |
| case Py_EQ: |
| v = TEST_COND(result == 0); |
| break; |
| case Py_NE: |
| v = TEST_COND(result != 0); |
| break; |
| case Py_LE: |
| v = TEST_COND(result <= 0); |
| break; |
| case Py_GE: |
| v = TEST_COND(result >= 0); |
| break; |
| case Py_LT: |
| v = TEST_COND(result == -1); |
| break; |
| case Py_GT: |
| v = TEST_COND(result == 1); |
| break; |
| default: |
| PyErr_BadArgument(); |
| return NULL; |
| } |
| Py_INCREF(v); |
| return v; |
| } |
| |
| Py_INCREF(Py_NotImplemented); |
| return Py_NotImplemented; |
| } |
| |
| int PyUnicode_Contains(PyObject *container, |
| PyObject *element) |
| { |
| PyObject *str, *sub; |
| int result; |
| |
| /* Coerce the two arguments */ |
| sub = PyUnicode_FromObject(element); |
| if (!sub) { |
| PyErr_Format(PyExc_TypeError, |
| "'in <string>' requires string as left operand, not %s", |
| element->ob_type->tp_name); |
| return -1; |
| } |
| |
| str = PyUnicode_FromObject(container); |
| if (!str) { |
| Py_DECREF(sub); |
| return -1; |
| } |
| |
| result = stringlib_contains_obj(str, sub); |
| |
| Py_DECREF(str); |
| Py_DECREF(sub); |
| |
| return result; |
| } |
| |
| /* Concat to string or Unicode object giving a new Unicode object. */ |
| |
| PyObject *PyUnicode_Concat(PyObject *left, |
| PyObject *right) |
| { |
| PyUnicodeObject *u = NULL, *v = NULL, *w; |
| |
| /* Coerce the two arguments */ |
| u = (PyUnicodeObject *)PyUnicode_FromObject(left); |
| if (u == NULL) |
| goto onError; |
| v = (PyUnicodeObject *)PyUnicode_FromObject(right); |
| if (v == NULL) |
| goto onError; |
| |
| /* Shortcuts */ |
| if (v == unicode_empty) { |
| Py_DECREF(v); |
| return (PyObject *)u; |
| } |
| if (u == unicode_empty) { |
| Py_DECREF(u); |
| return (PyObject *)v; |
| } |
| |
| /* Concat the two Unicode strings */ |
| w = _PyUnicode_New(u->length + v->length); |
| if (w == NULL) |
| goto onError; |
| Py_UNICODE_COPY(w->str, u->str, u->length); |
| Py_UNICODE_COPY(w->str + u->length, v->str, v->length); |
| |
| Py_DECREF(u); |
| Py_DECREF(v); |
| return (PyObject *)w; |
| |
| onError: |
| Py_XDECREF(u); |
| Py_XDECREF(v); |
| return NULL; |
| } |
| |
| void |
| PyUnicode_Append(PyObject **pleft, PyObject *right) |
| { |
| PyObject *new; |
| if (*pleft == NULL) |
| return; |
| if (right == NULL || !PyUnicode_Check(*pleft)) { |
| Py_DECREF(*pleft); |
| *pleft = NULL; |
| return; |
| } |
| new = PyUnicode_Concat(*pleft, right); |
| Py_DECREF(*pleft); |
| *pleft = new; |
| } |
| |
| void |
| PyUnicode_AppendAndDel(PyObject **pleft, PyObject *right) |
| { |
| PyUnicode_Append(pleft, right); |
| Py_XDECREF(right); |
| } |
| |
| PyDoc_STRVAR(count__doc__, |
| "S.count(sub[, start[, end]]) -> int\n\ |
| \n\ |
| Return the number of non-overlapping occurrences of substring sub in\n\ |
| string S[start:end]. Optional arguments start and end are\n\ |
| interpreted as in slice notation."); |
| |
| static PyObject * |
| unicode_count(PyUnicodeObject *self, PyObject *args) |
| { |
| PyUnicodeObject *substring; |
| Py_ssize_t start = 0; |
| Py_ssize_t end = PY_SSIZE_T_MAX; |
| PyObject *result; |
| |
| if (!stringlib_parse_args_finds_unicode("count", args, &substring, |
| &start, &end)) |
| return NULL; |
| |
| ADJUST_INDICES(start, end, self->length); |
| result = PyLong_FromSsize_t( |
| stringlib_count(self->str + start, end - start, |
| substring->str, substring->length, |
| PY_SSIZE_T_MAX) |
| ); |
| |
| Py_DECREF(substring); |
| |
| return result; |
| } |
| |
| PyDoc_STRVAR(encode__doc__, |
| "S.encode(encoding='utf-8', errors='strict') -> bytes\n\ |
| \n\ |
| Encode S using the codec registered for encoding. Default encoding\n\ |
| is 'utf-8'. errors may be given to set a different error\n\ |
| handling scheme. Default is 'strict' meaning that encoding errors raise\n\ |
| a UnicodeEncodeError. Other possible values are 'ignore', 'replace' and\n\ |
| 'xmlcharrefreplace' as well as any other name registered with\n\ |
| codecs.register_error that can handle UnicodeEncodeErrors."); |
| |
| static PyObject * |
| unicode_encode(PyUnicodeObject *self, PyObject *args, PyObject *kwargs) |
| { |
| static char *kwlist[] = {"encoding", "errors", 0}; |
| char *encoding = NULL; |
| char *errors = NULL; |
| |
| if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|ss:encode", |
| kwlist, &encoding, &errors)) |
| return NULL; |
| return PyUnicode_AsEncodedString((PyObject *)self, encoding, errors); |
| } |
| |
| PyDoc_STRVAR(expandtabs__doc__, |
| "S.expandtabs([tabsize]) -> str\n\ |
| \n\ |
| Return a copy of S where all tab characters are expanded using spaces.\n\ |
| If tabsize is not given, a tab size of 8 characters is assumed."); |
| |
| static PyObject* |
| unicode_expandtabs(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_UNICODE *e; |
| Py_UNICODE *p; |
| Py_UNICODE *q; |
| Py_UNICODE *qe; |
| Py_ssize_t i, j, incr; |
| PyUnicodeObject *u; |
| int tabsize = 8; |
| |
| if (!PyArg_ParseTuple(args, "|i:expandtabs", &tabsize)) |
| return NULL; |
| |
| /* First pass: determine size of output string */ |
| i = 0; /* chars up to and including most recent \n or \r */ |
| j = 0; /* chars since most recent \n or \r (use in tab calculations) */ |
| e = self->str + self->length; /* end of input */ |
| for (p = self->str; p < e; p++) |
| if (*p == '\t') { |
| if (tabsize > 0) { |
| incr = tabsize - (j % tabsize); /* cannot overflow */ |
| if (j > PY_SSIZE_T_MAX - incr) |
| goto overflow1; |
| j += incr; |
| } |
| } |
| else { |
| if (j > PY_SSIZE_T_MAX - 1) |
| goto overflow1; |
| j++; |
| if (*p == '\n' || *p == '\r') { |
| if (i > PY_SSIZE_T_MAX - j) |
| goto overflow1; |
| i += j; |
| j = 0; |
| } |
| } |
| |
| if (i > PY_SSIZE_T_MAX - j) |
| goto overflow1; |
| |
| /* Second pass: create output string and fill it */ |
| u = _PyUnicode_New(i + j); |
| if (!u) |
| return NULL; |
| |
| j = 0; /* same as in first pass */ |
| q = u->str; /* next output char */ |
| qe = u->str + u->length; /* end of output */ |
| |
| for (p = self->str; p < e; p++) |
| if (*p == '\t') { |
| if (tabsize > 0) { |
| i = tabsize - (j % tabsize); |
| j += i; |
| while (i--) { |
| if (q >= qe) |
| goto overflow2; |
| *q++ = ' '; |
| } |
| } |
| } |
| else { |
| if (q >= qe) |
| goto overflow2; |
| *q++ = *p; |
| j++; |
| if (*p == '\n' || *p == '\r') |
| j = 0; |
| } |
| |
| return (PyObject*) u; |
| |
| overflow2: |
| Py_DECREF(u); |
| overflow1: |
| PyErr_SetString(PyExc_OverflowError, "new string is too long"); |
| return NULL; |
| } |
| |
| PyDoc_STRVAR(find__doc__, |
| "S.find(sub[, start[, end]]) -> int\n\ |
| \n\ |
| Return the lowest index in S where substring sub is found,\n\ |
| such that sub is contained within S[start:end]. Optional\n\ |
| arguments start and end are interpreted as in slice notation.\n\ |
| \n\ |
| Return -1 on failure."); |
| |
| static PyObject * |
| unicode_find(PyUnicodeObject *self, PyObject *args) |
| { |
| PyUnicodeObject *substring; |
| Py_ssize_t start; |
| Py_ssize_t end; |
| Py_ssize_t result; |
| |
| if (!stringlib_parse_args_finds_unicode("find", args, &substring, |
| &start, &end)) |
| return NULL; |
| |
| result = stringlib_find_slice( |
| PyUnicode_AS_UNICODE(self), PyUnicode_GET_SIZE(self), |
| PyUnicode_AS_UNICODE(substring), PyUnicode_GET_SIZE(substring), |
| start, end |
| ); |
| |
| Py_DECREF(substring); |
| |
| return PyLong_FromSsize_t(result); |
| } |
| |
| static PyObject * |
| unicode_getitem(PyUnicodeObject *self, Py_ssize_t index) |
| { |
| if (index < 0 || index >= self->length) { |
| PyErr_SetString(PyExc_IndexError, "string index out of range"); |
| return NULL; |
| } |
| |
| return (PyObject*) PyUnicode_FromUnicode(&self->str[index], 1); |
| } |
| |
| /* Believe it or not, this produces the same value for ASCII strings |
| as string_hash(). */ |
| static Py_hash_t |
| unicode_hash(PyUnicodeObject *self) |
| { |
| Py_ssize_t len; |
| Py_UNICODE *p; |
| Py_uhash_t x; /* Unsigned for defined overflow behavior. */ |
| |
| #ifdef Py_DEBUG |
| assert(_Py_HashSecret_Initialized); |
| #endif |
| if (self->hash != -1) |
| return self->hash; |
| len = Py_SIZE(self); |
| /* |
| We make the hash of the empty string be 0, rather than using |
| (prefix ^ suffix), since this slightly obfuscates the hash secret |
| */ |
| if (len == 0) { |
| self->hash = 0; |
| return 0; |
| } |
| p = self->str; |
| x = _Py_HashSecret.prefix; |
| x ^= *p << 7; |
| while (--len >= 0) |
| x = (_PyHASH_MULTIPLIER*x) ^ *p++; |
| x ^= Py_SIZE(self); |
| x ^= _Py_HashSecret.suffix; |
| if (x == -1) |
| x = -2; |
| self->hash = x; |
| return x; |
| } |
| |
| PyDoc_STRVAR(index__doc__, |
| "S.index(sub[, start[, end]]) -> int\n\ |
| \n\ |
| Like S.find() but raise ValueError when the substring is not found."); |
| |
| static PyObject * |
| unicode_index(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_ssize_t result; |
| PyUnicodeObject *substring; |
| Py_ssize_t start; |
| Py_ssize_t end; |
| |
| if (!stringlib_parse_args_finds_unicode("index", args, &substring, |
| &start, &end)) |
| return NULL; |
| |
| result = stringlib_find_slice( |
| PyUnicode_AS_UNICODE(self), PyUnicode_GET_SIZE(self), |
| PyUnicode_AS_UNICODE(substring), PyUnicode_GET_SIZE(substring), |
| start, end |
| ); |
| |
| Py_DECREF(substring); |
| |
| if (result < 0) { |
| PyErr_SetString(PyExc_ValueError, "substring not found"); |
| return NULL; |
| } |
| |
| return PyLong_FromSsize_t(result); |
| } |
| |
| PyDoc_STRVAR(islower__doc__, |
| "S.islower() -> bool\n\ |
| \n\ |
| Return True if all cased characters in S are lowercase and there is\n\ |
| at least one cased character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_islower(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| int cased; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1) |
| return PyBool_FromLong(Py_UNICODE_ISLOWER(*p)); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| cased = 0; |
| while (p < e) { |
| const Py_UCS4 ch = _Py_UNICODE_NEXT(p, e); |
| |
| if (Py_UNICODE_ISUPPER(ch) || Py_UNICODE_ISTITLE(ch)) |
| return PyBool_FromLong(0); |
| else if (!cased && Py_UNICODE_ISLOWER(ch)) |
| cased = 1; |
| } |
| return PyBool_FromLong(cased); |
| } |
| |
| PyDoc_STRVAR(isupper__doc__, |
| "S.isupper() -> bool\n\ |
| \n\ |
| Return True if all cased characters in S are uppercase and there is\n\ |
| at least one cased character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_isupper(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| int cased; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1) |
| return PyBool_FromLong(Py_UNICODE_ISUPPER(*p) != 0); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| cased = 0; |
| while (p < e) { |
| const Py_UCS4 ch = _Py_UNICODE_NEXT(p, e); |
| |
| if (Py_UNICODE_ISLOWER(ch) || Py_UNICODE_ISTITLE(ch)) |
| return PyBool_FromLong(0); |
| else if (!cased && Py_UNICODE_ISUPPER(ch)) |
| cased = 1; |
| } |
| return PyBool_FromLong(cased); |
| } |
| |
| PyDoc_STRVAR(istitle__doc__, |
| "S.istitle() -> bool\n\ |
| \n\ |
| Return True if S is a titlecased string and there is at least one\n\ |
| character in S, i.e. upper- and titlecase characters may only\n\ |
| follow uncased characters and lowercase characters only cased ones.\n\ |
| Return False otherwise."); |
| |
| static PyObject* |
| unicode_istitle(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| int cased, previous_is_cased; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1) |
| return PyBool_FromLong((Py_UNICODE_ISTITLE(*p) != 0) || |
| (Py_UNICODE_ISUPPER(*p) != 0)); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| cased = 0; |
| previous_is_cased = 0; |
| while (p < e) { |
| const Py_UCS4 ch = _Py_UNICODE_NEXT(p, e); |
| |
| if (Py_UNICODE_ISUPPER(ch) || Py_UNICODE_ISTITLE(ch)) { |
| if (previous_is_cased) |
| return PyBool_FromLong(0); |
| previous_is_cased = 1; |
| cased = 1; |
| } |
| else if (Py_UNICODE_ISLOWER(ch)) { |
| if (!previous_is_cased) |
| return PyBool_FromLong(0); |
| previous_is_cased = 1; |
| cased = 1; |
| } |
| else |
| previous_is_cased = 0; |
| } |
| return PyBool_FromLong(cased); |
| } |
| |
| PyDoc_STRVAR(isspace__doc__, |
| "S.isspace() -> bool\n\ |
| \n\ |
| Return True if all characters in S are whitespace\n\ |
| and there is at least one character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_isspace(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISSPACE(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| const Py_UCS4 ch = _Py_UNICODE_NEXT(p, e); |
| if (!Py_UNICODE_ISSPACE(ch)) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| PyDoc_STRVAR(isalpha__doc__, |
| "S.isalpha() -> bool\n\ |
| \n\ |
| Return True if all characters in S are alphabetic\n\ |
| and there is at least one character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_isalpha(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISALPHA(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| if (!Py_UNICODE_ISALPHA(_Py_UNICODE_NEXT(p, e))) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| PyDoc_STRVAR(isalnum__doc__, |
| "S.isalnum() -> bool\n\ |
| \n\ |
| Return True if all characters in S are alphanumeric\n\ |
| and there is at least one character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_isalnum(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISALNUM(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| const Py_UCS4 ch = _Py_UNICODE_NEXT(p, e); |
| if (!Py_UNICODE_ISALNUM(ch)) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| PyDoc_STRVAR(isdecimal__doc__, |
| "S.isdecimal() -> bool\n\ |
| \n\ |
| Return True if there are only decimal characters in S,\n\ |
| False otherwise."); |
| |
| static PyObject* |
| unicode_isdecimal(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISDECIMAL(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| if (!Py_UNICODE_ISDECIMAL(_Py_UNICODE_NEXT(p, e))) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| PyDoc_STRVAR(isdigit__doc__, |
| "S.isdigit() -> bool\n\ |
| \n\ |
| Return True if all characters in S are digits\n\ |
| and there is at least one character in S, False otherwise."); |
| |
| static PyObject* |
| unicode_isdigit(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISDIGIT(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| if (!Py_UNICODE_ISDIGIT(_Py_UNICODE_NEXT(p, e))) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| PyDoc_STRVAR(isnumeric__doc__, |
| "S.isnumeric() -> bool\n\ |
| \n\ |
| Return True if there are only numeric characters in S,\n\ |
| False otherwise."); |
| |
| static PyObject* |
| unicode_isnumeric(PyUnicodeObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && |
| Py_UNICODE_ISNUMERIC(*p)) |
| return PyBool_FromLong(1); |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return PyBool_FromLong(0); |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| if (!Py_UNICODE_ISNUMERIC(_Py_UNICODE_NEXT(p, e))) |
| return PyBool_FromLong(0); |
| } |
| return PyBool_FromLong(1); |
| } |
| |
| int |
| PyUnicode_IsIdentifier(PyObject *self) |
| { |
| const Py_UNICODE *p = PyUnicode_AS_UNICODE((PyUnicodeObject*)self); |
| const Py_UNICODE *e; |
| Py_UCS4 first; |
| |
| /* Special case for empty strings */ |
| if (PyUnicode_GET_SIZE(self) == 0) |
| return 0; |
| |
| /* PEP 3131 says that the first character must be in |
| XID_Start and subsequent characters in XID_Continue, |
| and for the ASCII range, the 2.x rules apply (i.e |
| start with letters and underscore, continue with |
| letters, digits, underscore). However, given the current |
| definition of XID_Start and XID_Continue, it is sufficient |
| to check just for these, except that _ must be allowed |
| as starting an identifier. */ |
| e = p + PyUnicode_GET_SIZE(self); |
| first = _Py_UNICODE_NEXT(p, e); |
| if (!_PyUnicode_IsXidStart(first) && first != 0x5F /* LOW LINE */) |
| return 0; |
| |
| while (p < e) |
| if (!_PyUnicode_IsXidContinue(_Py_UNICODE_NEXT(p, e))) |
| return 0; |
| return 1; |
| } |
| |
| PyDoc_STRVAR(isidentifier__doc__, |
| "S.isidentifier() -> bool\n\ |
| \n\ |
| Return True if S is a valid identifier according\n\ |
| to the language definition."); |
| |
| static PyObject* |
| unicode_isidentifier(PyObject *self) |
| { |
| return PyBool_FromLong(PyUnicode_IsIdentifier(self)); |
| } |
| |
| PyDoc_STRVAR(isprintable__doc__, |
| "S.isprintable() -> bool\n\ |
| \n\ |
| Return True if all characters in S are considered\n\ |
| printable in repr() or S is empty, False otherwise."); |
| |
| static PyObject* |
| unicode_isprintable(PyObject *self) |
| { |
| register const Py_UNICODE *p = PyUnicode_AS_UNICODE(self); |
| register const Py_UNICODE *e; |
| |
| /* Shortcut for single character strings */ |
| if (PyUnicode_GET_SIZE(self) == 1 && Py_UNICODE_ISPRINTABLE(*p)) { |
| Py_RETURN_TRUE; |
| } |
| |
| e = p + PyUnicode_GET_SIZE(self); |
| while (p < e) { |
| if (!Py_UNICODE_ISPRINTABLE(_Py_UNICODE_NEXT(p, e))) { |
| Py_RETURN_FALSE; |
| } |
| } |
| Py_RETURN_TRUE; |
| } |
| |
| PyDoc_STRVAR(join__doc__, |
| "S.join(iterable) -> str\n\ |
| \n\ |
| Return a string which is the concatenation of the strings in the\n\ |
| iterable. The separator between elements is S."); |
| |
| static PyObject* |
| unicode_join(PyObject *self, PyObject *data) |
| { |
| return PyUnicode_Join(self, data); |
| } |
| |
| static Py_ssize_t |
| unicode_length(PyUnicodeObject *self) |
| { |
| return self->length; |
| } |
| |
| PyDoc_STRVAR(ljust__doc__, |
| "S.ljust(width[, fillchar]) -> str\n\ |
| \n\ |
| Return S left-justified in a Unicode string of length width. Padding is\n\ |
| done using the specified fill character (default is a space)."); |
| |
| static PyObject * |
| unicode_ljust(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_ssize_t width; |
| Py_UNICODE fillchar = ' '; |
| |
| if (!PyArg_ParseTuple(args, "n|O&:ljust", &width, convert_uc, &fillchar)) |
| return NULL; |
| |
| if (self->length >= width && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*) self; |
| } |
| |
| return (PyObject*) pad(self, 0, width - self->length, fillchar); |
| } |
| |
| PyDoc_STRVAR(lower__doc__, |
| "S.lower() -> str\n\ |
| \n\ |
| Return a copy of the string S converted to lowercase."); |
| |
| static PyObject* |
| unicode_lower(PyUnicodeObject *self) |
| { |
| return fixup(self, fixlower); |
| } |
| |
| #define LEFTSTRIP 0 |
| #define RIGHTSTRIP 1 |
| #define BOTHSTRIP 2 |
| |
| /* Arrays indexed by above */ |
| static const char *stripformat[] = {"|O:lstrip", "|O:rstrip", "|O:strip"}; |
| |
| #define STRIPNAME(i) (stripformat[i]+3) |
| |
| /* externally visible for str.strip(unicode) */ |
| PyObject * |
| _PyUnicode_XStrip(PyUnicodeObject *self, int striptype, PyObject *sepobj) |
| { |
| Py_UNICODE *s = PyUnicode_AS_UNICODE(self); |
| Py_ssize_t len = PyUnicode_GET_SIZE(self); |
| Py_UNICODE *sep = PyUnicode_AS_UNICODE(sepobj); |
| Py_ssize_t seplen = PyUnicode_GET_SIZE(sepobj); |
| Py_ssize_t i, j; |
| |
| BLOOM_MASK sepmask = make_bloom_mask(sep, seplen); |
| |
| i = 0; |
| if (striptype != RIGHTSTRIP) { |
| while (i < len && BLOOM_MEMBER(sepmask, s[i], sep, seplen)) { |
| i++; |
| } |
| } |
| |
| j = len; |
| if (striptype != LEFTSTRIP) { |
| do { |
| j--; |
| } while (j >= i && BLOOM_MEMBER(sepmask, s[j], sep, seplen)); |
| j++; |
| } |
| |
| if (i == 0 && j == len && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*)self; |
| } |
| else |
| return PyUnicode_FromUnicode(s+i, j-i); |
| } |
| |
| |
| static PyObject * |
| do_strip(PyUnicodeObject *self, int striptype) |
| { |
| Py_UNICODE *s = PyUnicode_AS_UNICODE(self); |
| Py_ssize_t len = PyUnicode_GET_SIZE(self), i, j; |
| |
| i = 0; |
| if (striptype != RIGHTSTRIP) { |
| while (i < len && Py_UNICODE_ISSPACE(s[i])) { |
| i++; |
| } |
| } |
| |
| j = len; |
| if (striptype != LEFTSTRIP) { |
| do { |
| j--; |
| } while (j >= i && Py_UNICODE_ISSPACE(s[j])); |
| j++; |
| } |
| |
| if (i == 0 && j == len && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*)self; |
| } |
| else |
| return PyUnicode_FromUnicode(s+i, j-i); |
| } |
| |
| |
| static PyObject * |
| do_argstrip(PyUnicodeObject *self, int striptype, PyObject *args) |
| { |
| PyObject *sep = NULL; |
| |
| if (!PyArg_ParseTuple(args, (char *)stripformat[striptype], &sep)) |
| return NULL; |
| |
| if (sep != NULL && sep != Py_None) { |
| if (PyUnicode_Check(sep)) |
| return _PyUnicode_XStrip(self, striptype, sep); |
| else { |
| PyErr_Format(PyExc_TypeError, |
| "%s arg must be None or str", |
| STRIPNAME(striptype)); |
| return NULL; |
| } |
| } |
| |
| return do_strip(self, striptype); |
| } |
| |
| |
| PyDoc_STRVAR(strip__doc__, |
| "S.strip([chars]) -> str\n\ |
| \n\ |
| Return a copy of the string S with leading and trailing\n\ |
| whitespace removed.\n\ |
| If chars is given and not None, remove characters in chars instead."); |
| |
| static PyObject * |
| unicode_strip(PyUnicodeObject *self, PyObject *args) |
| { |
| if (PyTuple_GET_SIZE(args) == 0) |
| return do_strip(self, BOTHSTRIP); /* Common case */ |
| else |
| return do_argstrip(self, BOTHSTRIP, args); |
| } |
| |
| |
| PyDoc_STRVAR(lstrip__doc__, |
| "S.lstrip([chars]) -> str\n\ |
| \n\ |
| Return a copy of the string S with leading whitespace removed.\n\ |
| If chars is given and not None, remove characters in chars instead."); |
| |
| static PyObject * |
| unicode_lstrip(PyUnicodeObject *self, PyObject *args) |
| { |
| if (PyTuple_GET_SIZE(args) == 0) |
| return do_strip(self, LEFTSTRIP); /* Common case */ |
| else |
| return do_argstrip(self, LEFTSTRIP, args); |
| } |
| |
| |
| PyDoc_STRVAR(rstrip__doc__, |
| "S.rstrip([chars]) -> str\n\ |
| \n\ |
| Return a copy of the string S with trailing whitespace removed.\n\ |
| If chars is given and not None, remove characters in chars instead."); |
| |
| static PyObject * |
| unicode_rstrip(PyUnicodeObject *self, PyObject *args) |
| { |
| if (PyTuple_GET_SIZE(args) == 0) |
| return do_strip(self, RIGHTSTRIP); /* Common case */ |
| else |
| return do_argstrip(self, RIGHTSTRIP, args); |
| } |
| |
| |
| static PyObject* |
| unicode_repeat(PyUnicodeObject *str, Py_ssize_t len) |
| { |
| PyUnicodeObject *u; |
| Py_UNICODE *p; |
| Py_ssize_t nchars; |
| size_t nbytes; |
| |
| if (len < 1) |
| _Py_RETURN_UNICODE_EMPTY(); |
| |
| if (len == 1 && PyUnicode_CheckExact(str)) { |
| /* no repeat, return original string */ |
| Py_INCREF(str); |
| return (PyObject*) str; |
| } |
| |
| /* ensure # of chars needed doesn't overflow int and # of bytes |
| * needed doesn't overflow size_t |
| */ |
| nchars = len * str->length; |
| if (nchars / len != str->length) { |
| PyErr_SetString(PyExc_OverflowError, |
| "repeated string is too long"); |
| return NULL; |
| } |
| nbytes = (nchars + 1) * sizeof(Py_UNICODE); |
| if (nbytes / sizeof(Py_UNICODE) != (size_t)(nchars + 1)) { |
| PyErr_SetString(PyExc_OverflowError, |
| "repeated string is too long"); |
| return NULL; |
| } |
| u = _PyUnicode_New(nchars); |
| if (!u) |
| return NULL; |
| |
| p = u->str; |
| |
| if (str->length == 1) { |
| Py_UNICODE_FILL(p, str->str[0], len); |
| } else { |
| Py_ssize_t done = str->length; /* number of characters copied this far */ |
| Py_UNICODE_COPY(p, str->str, str->length); |
| while (done < nchars) { |
| Py_ssize_t n = (done <= nchars-done) ? done : nchars-done; |
| Py_UNICODE_COPY(p+done, p, n); |
| done += n; |
| } |
| } |
| |
| return (PyObject*) u; |
| } |
| |
| PyObject *PyUnicode_Replace(PyObject *obj, |
| PyObject *subobj, |
| PyObject *replobj, |
| Py_ssize_t maxcount) |
| { |
| PyObject *self; |
| PyObject *str1; |
| PyObject *str2; |
| PyObject *result; |
| |
| self = PyUnicode_FromObject(obj); |
| if (self == NULL) |
| return NULL; |
| str1 = PyUnicode_FromObject(subobj); |
| if (str1 == NULL) { |
| Py_DECREF(self); |
| return NULL; |
| } |
| str2 = PyUnicode_FromObject(replobj); |
| if (str2 == NULL) { |
| Py_DECREF(self); |
| Py_DECREF(str1); |
| return NULL; |
| } |
| result = replace((PyUnicodeObject *)self, |
| (PyUnicodeObject *)str1, |
| (PyUnicodeObject *)str2, |
| maxcount); |
| Py_DECREF(self); |
| Py_DECREF(str1); |
| Py_DECREF(str2); |
| return result; |
| } |
| |
| PyDoc_STRVAR(replace__doc__, |
| "S.replace(old, new[, count]) -> str\n\ |
| \n\ |
| Return a copy of S with all occurrences of substring\n\ |
| old replaced by new. If the optional argument count is\n\ |
| given, only the first count occurrences are replaced."); |
| |
| static PyObject* |
| unicode_replace(PyUnicodeObject *self, PyObject *args) |
| { |
| PyUnicodeObject *str1; |
| PyUnicodeObject *str2; |
| Py_ssize_t maxcount = -1; |
| PyObject *result; |
| |
| if (!PyArg_ParseTuple(args, "OO|n:replace", &str1, &str2, &maxcount)) |
| return NULL; |
| str1 = (PyUnicodeObject *)PyUnicode_FromObject((PyObject *)str1); |
| if (str1 == NULL) |
| return NULL; |
| str2 = (PyUnicodeObject *)PyUnicode_FromObject((PyObject *)str2); |
| if (str2 == NULL) { |
| Py_DECREF(str1); |
| return NULL; |
| } |
| |
| result = replace(self, str1, str2, maxcount); |
| |
| Py_DECREF(str1); |
| Py_DECREF(str2); |
| return result; |
| } |
| |
| static |
| PyObject *unicode_repr(PyObject *unicode) |
| { |
| PyObject *repr; |
| Py_UNICODE *p; |
| Py_UNICODE *s = PyUnicode_AS_UNICODE(unicode); |
| Py_ssize_t size = PyUnicode_GET_SIZE(unicode); |
| |
| /* XXX(nnorwitz): rather than over-allocating, it would be |
| better to choose a different scheme. Perhaps scan the |
| first N-chars of the string and allocate based on that size. |
| */ |
| /* Initial allocation is based on the longest-possible unichr |
| escape. |
| |
| In wide (UTF-32) builds '\U00xxxxxx' is 10 chars per source |
| unichr, so in this case it's the longest unichr escape. In |
| narrow (UTF-16) builds this is five chars per source unichr |
| since there are two unichrs in the surrogate pair, so in narrow |
| (UTF-16) builds it's not the longest unichr escape. |
| |
| In wide or narrow builds '\uxxxx' is 6 chars per source unichr, |
| so in the narrow (UTF-16) build case it's the longest unichr |
| escape. |
| */ |
| |
| repr = PyUnicode_FromUnicode(NULL, |
| 2 /* quotes */ |
| #ifdef Py_UNICODE_WIDE |
| + 10*size |
| #else |
| + 6*size |
| #endif |
| + 1); |
| if (repr == NULL) |
| return NULL; |
| |
| p = PyUnicode_AS_UNICODE(repr); |
| |
| /* Add quote */ |
| *p++ = (findchar(s, size, '\'') && |
| !findchar(s, size, '"')) ? '"' : '\''; |
| while (size-- > 0) { |
| Py_UNICODE ch = *s++; |
| |
| /* Escape quotes and backslashes */ |
| if ((ch == PyUnicode_AS_UNICODE(repr)[0]) || (ch == '\\')) { |
| *p++ = '\\'; |
| *p++ = ch; |
| continue; |
| } |
| |
| /* Map special whitespace to '\t', \n', '\r' */ |
| if (ch == '\t') { |
| *p++ = '\\'; |
| *p++ = 't'; |
| } |
| else if (ch == '\n') { |
| *p++ = '\\'; |
| *p++ = 'n'; |
| } |
| else if (ch == '\r') { |
| *p++ = '\\'; |
| *p++ = 'r'; |
| } |
| |
| /* Map non-printable US ASCII to '\xhh' */ |
| else if (ch < ' ' || ch == 0x7F) { |
| *p++ = '\\'; |
| *p++ = 'x'; |
| *p++ = hexdigits[(ch >> 4) & 0x000F]; |
| *p++ = hexdigits[ch & 0x000F]; |
| } |
| |
| /* Copy ASCII characters as-is */ |
| else if (ch < 0x7F) { |
| *p++ = ch; |
| } |
| |
| /* Non-ASCII characters */ |
| else { |
| Py_UCS4 ucs = ch; |
| |
| #ifndef Py_UNICODE_WIDE |
| Py_UNICODE ch2 = 0; |
| /* Get code point from surrogate pair */ |
| if (size > 0) { |
| ch2 = *s; |
| if (ch >= 0xD800 && ch < 0xDC00 && ch2 >= 0xDC00 |
| && ch2 <= 0xDFFF) { |
| ucs = (((ch & 0x03FF) << 10) | (ch2 & 0x03FF)) |
| + 0x00010000; |
| s++; |
| size--; |
| } |
| } |
| #endif |
| /* Map Unicode whitespace and control characters |
| (categories Z* and C* except ASCII space) |
| */ |
| if (!Py_UNICODE_ISPRINTABLE(ucs)) { |
| /* Map 8-bit characters to '\xhh' */ |
| if (ucs <= 0xff) { |
| *p++ = '\\'; |
| *p++ = 'x'; |
| *p++ = hexdigits[(ch >> 4) & 0x000F]; |
| *p++ = hexdigits[ch & 0x000F]; |
| } |
| /* Map 21-bit characters to '\U00xxxxxx' */ |
| else if (ucs >= 0x10000) { |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = hexdigits[(ucs >> 28) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 24) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 20) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 16) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 12) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 8) & 0x0000000F]; |
| *p++ = hexdigits[(ucs >> 4) & 0x0000000F]; |
| *p++ = hexdigits[ucs & 0x0000000F]; |
| } |
| /* Map 16-bit characters to '\uxxxx' */ |
| else { |
| *p++ = '\\'; |
| *p++ = 'u'; |
| *p++ = hexdigits[(ucs >> 12) & 0x000F]; |
| *p++ = hexdigits[(ucs >> 8) & 0x000F]; |
| *p++ = hexdigits[(ucs >> 4) & 0x000F]; |
| *p++ = hexdigits[ucs & 0x000F]; |
| } |
| } |
| /* Copy characters as-is */ |
| else { |
| *p++ = ch; |
| #ifndef Py_UNICODE_WIDE |
| if (ucs >= 0x10000) |
| *p++ = ch2; |
| #endif |
| } |
| } |
| } |
| /* Add quote */ |
| *p++ = PyUnicode_AS_UNICODE(repr)[0]; |
| |
| *p = '\0'; |
| PyUnicode_Resize(&repr, p - PyUnicode_AS_UNICODE(repr)); |
| return repr; |
| } |
| |
| PyDoc_STRVAR(rfind__doc__, |
| "S.rfind(sub[, start[, end]]) -> int\n\ |
| \n\ |
| Return the highest index in S where substring sub is found,\n\ |
| such that sub is contained within S[start:end]. Optional\n\ |
| arguments start and end are interpreted as in slice notation.\n\ |
| \n\ |
| Return -1 on failure."); |
| |
| static PyObject * |
| unicode_rfind(PyUnicodeObject *self, PyObject *args) |
| { |
| PyUnicodeObject *substring; |
| Py_ssize_t start; |
| Py_ssize_t end; |
| Py_ssize_t result; |
| |
| if (!stringlib_parse_args_finds_unicode("rfind", args, &substring, |
| &start, &end)) |
| return NULL; |
| |
| result = stringlib_rfind_slice( |
| PyUnicode_AS_UNICODE(self), PyUnicode_GET_SIZE(self), |
| PyUnicode_AS_UNICODE(substring), PyUnicode_GET_SIZE(substring), |
| start, end |
| ); |
| |
| Py_DECREF(substring); |
| |
| return PyLong_FromSsize_t(result); |
| } |
| |
| PyDoc_STRVAR(rindex__doc__, |
| "S.rindex(sub[, start[, end]]) -> int\n\ |
| \n\ |
| Like S.rfind() but raise ValueError when the substring is not found."); |
| |
| static PyObject * |
| unicode_rindex(PyUnicodeObject *self, PyObject *args) |
| { |
| PyUnicodeObject *substring; |
| Py_ssize_t start; |
| Py_ssize_t end; |
| Py_ssize_t result; |
| |
| if (!stringlib_parse_args_finds_unicode("rindex", args, &substring, |
| &start, &end)) |
| return NULL; |
| |
| result = stringlib_rfind_slice( |
| PyUnicode_AS_UNICODE(self), PyUnicode_GET_SIZE(self), |
| PyUnicode_AS_UNICODE(substring), PyUnicode_GET_SIZE(substring), |
| start, end |
| ); |
| |
| Py_DECREF(substring); |
| |
| if (result < 0) { |
| PyErr_SetString(PyExc_ValueError, "substring not found"); |
| return NULL; |
| } |
| return PyLong_FromSsize_t(result); |
| } |
| |
| PyDoc_STRVAR(rjust__doc__, |
| "S.rjust(width[, fillchar]) -> str\n\ |
| \n\ |
| Return S right-justified in a string of length width. Padding is\n\ |
| done using the specified fill character (default is a space)."); |
| |
| static PyObject * |
| unicode_rjust(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_ssize_t width; |
| Py_UNICODE fillchar = ' '; |
| |
| if (!PyArg_ParseTuple(args, "n|O&:rjust", &width, convert_uc, &fillchar)) |
| return NULL; |
| |
| if (self->length >= width && PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*) self; |
| } |
| |
| return (PyObject*) pad(self, width - self->length, 0, fillchar); |
| } |
| |
| PyObject *PyUnicode_Split(PyObject *s, |
| PyObject *sep, |
| Py_ssize_t maxsplit) |
| { |
| PyObject *result; |
| |
| s = PyUnicode_FromObject(s); |
| if (s == NULL) |
| return NULL; |
| if (sep != NULL) { |
| sep = PyUnicode_FromObject(sep); |
| if (sep == NULL) { |
| Py_DECREF(s); |
| return NULL; |
| } |
| } |
| |
| result = split((PyUnicodeObject *)s, (PyUnicodeObject *)sep, maxsplit); |
| |
| Py_DECREF(s); |
| Py_XDECREF(sep); |
| return result; |
| } |
| |
| PyDoc_STRVAR(split__doc__, |
| "S.split([sep[, maxsplit]]) -> list of strings\n\ |
| \n\ |
| Return a list of the words in S, using sep as the\n\ |
| delimiter string. If maxsplit is given, at most maxsplit\n\ |
| splits are done. If sep is not specified or is None, any\n\ |
| whitespace string is a separator and empty strings are\n\ |
| removed from the result."); |
| |
| static PyObject* |
| unicode_split(PyUnicodeObject *self, PyObject *args) |
| { |
| PyObject *substring = Py_None; |
| Py_ssize_t maxcount = -1; |
| |
| if (!PyArg_ParseTuple(args, "|On:split", &substring, &maxcount)) |
| return NULL; |
| |
| if (substring == Py_None) |
| return split(self, NULL, maxcount); |
| else if (PyUnicode_Check(substring)) |
| return split(self, (PyUnicodeObject *)substring, maxcount); |
| else |
| return PyUnicode_Split((PyObject *)self, substring, maxcount); |
| } |
| |
| PyObject * |
| PyUnicode_Partition(PyObject *str_in, PyObject *sep_in) |
| { |
| PyObject* str_obj; |
| PyObject* sep_obj; |
| PyObject* out; |
| |
| str_obj = PyUnicode_FromObject(str_in); |
| if (!str_obj) |
| return NULL; |
| sep_obj = PyUnicode_FromObject(sep_in); |
| if (!sep_obj) { |
| Py_DECREF(str_obj); |
| return NULL; |
| } |
| |
| out = stringlib_partition( |
| str_obj, PyUnicode_AS_UNICODE(str_obj), PyUnicode_GET_SIZE(str_obj), |
| sep_obj, PyUnicode_AS_UNICODE(sep_obj), PyUnicode_GET_SIZE(sep_obj) |
| ); |
| |
| Py_DECREF(sep_obj); |
| Py_DECREF(str_obj); |
| |
| return out; |
| } |
| |
| |
| PyObject * |
| PyUnicode_RPartition(PyObject *str_in, PyObject *sep_in) |
| { |
| PyObject* str_obj; |
| PyObject* sep_obj; |
| PyObject* out; |
| |
| str_obj = PyUnicode_FromObject(str_in); |
| if (!str_obj) |
| return NULL; |
| sep_obj = PyUnicode_FromObject(sep_in); |
| if (!sep_obj) { |
| Py_DECREF(str_obj); |
| return NULL; |
| } |
| |
| out = stringlib_rpartition( |
| str_obj, PyUnicode_AS_UNICODE(str_obj), PyUnicode_GET_SIZE(str_obj), |
| sep_obj, PyUnicode_AS_UNICODE(sep_obj), PyUnicode_GET_SIZE(sep_obj) |
| ); |
| |
| Py_DECREF(sep_obj); |
| Py_DECREF(str_obj); |
| |
| return out; |
| } |
| |
| PyDoc_STRVAR(partition__doc__, |
| "S.partition(sep) -> (head, sep, tail)\n\ |
| \n\ |
| Search for the separator sep in S, and return the part before it,\n\ |
| the separator itself, and the part after it. If the separator is not\n\ |
| found, return S and two empty strings."); |
| |
| static PyObject* |
| unicode_partition(PyUnicodeObject *self, PyObject *separator) |
| { |
| return PyUnicode_Partition((PyObject *)self, separator); |
| } |
| |
| PyDoc_STRVAR(rpartition__doc__, |
| "S.rpartition(sep) -> (head, sep, tail)\n\ |
| \n\ |
| Search for the separator sep in S, starting at the end of S, and return\n\ |
| the part before it, the separator itself, and the part after it. If the\n\ |
| separator is not found, return two empty strings and S."); |
| |
| static PyObject* |
| unicode_rpartition(PyUnicodeObject *self, PyObject *separator) |
| { |
| return PyUnicode_RPartition((PyObject *)self, separator); |
| } |
| |
| PyObject *PyUnicode_RSplit(PyObject *s, |
| PyObject *sep, |
| Py_ssize_t maxsplit) |
| { |
| PyObject *result; |
| |
| s = PyUnicode_FromObject(s); |
| if (s == NULL) |
| return NULL; |
| if (sep != NULL) { |
| sep = PyUnicode_FromObject(sep); |
| if (sep == NULL) { |
| Py_DECREF(s); |
| return NULL; |
| } |
| } |
| |
| result = rsplit((PyUnicodeObject *)s, (PyUnicodeObject *)sep, maxsplit); |
| |
| Py_DECREF(s); |
| Py_XDECREF(sep); |
| return result; |
| } |
| |
| PyDoc_STRVAR(rsplit__doc__, |
| "S.rsplit([sep[, maxsplit]]) -> list of strings\n\ |
| \n\ |
| Return a list of the words in S, using sep as the\n\ |
| delimiter string, starting at the end of the string and\n\ |
| working to the front. If maxsplit is given, at most maxsplit\n\ |
| splits are done. If sep is not specified, any whitespace string\n\ |
| is a separator."); |
| |
| static PyObject* |
| unicode_rsplit(PyUnicodeObject *self, PyObject *args) |
| { |
| PyObject *substring = Py_None; |
| Py_ssize_t maxcount = -1; |
| |
| if (!PyArg_ParseTuple(args, "|On:rsplit", &substring, &maxcount)) |
| return NULL; |
| |
| if (substring == Py_None) |
| return rsplit(self, NULL, maxcount); |
| else if (PyUnicode_Check(substring)) |
| return rsplit(self, (PyUnicodeObject *)substring, maxcount); |
| else |
| return PyUnicode_RSplit((PyObject *)self, substring, maxcount); |
| } |
| |
| PyDoc_STRVAR(splitlines__doc__, |
| "S.splitlines([keepends]) -> list of strings\n\ |
| \n\ |
| Return a list of the lines in S, breaking at line boundaries.\n\ |
| Line breaks are not included in the resulting list unless keepends\n\ |
| is given and true."); |
| |
| static PyObject* |
| unicode_splitlines(PyUnicodeObject *self, PyObject *args) |
| { |
| int keepends = 0; |
| |
| if (!PyArg_ParseTuple(args, "|i:splitlines", &keepends)) |
| return NULL; |
| |
| return PyUnicode_Splitlines((PyObject *)self, keepends); |
| } |
| |
| static |
| PyObject *unicode_str(PyObject *self) |
| { |
| if (PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return self; |
| } else |
| /* Subtype -- return genuine unicode string with the same value. */ |
| return PyUnicode_FromUnicode(PyUnicode_AS_UNICODE(self), |
| PyUnicode_GET_SIZE(self)); |
| } |
| |
| PyDoc_STRVAR(swapcase__doc__, |
| "S.swapcase() -> str\n\ |
| \n\ |
| Return a copy of S with uppercase characters converted to lowercase\n\ |
| and vice versa."); |
| |
| static PyObject* |
| unicode_swapcase(PyUnicodeObject *self) |
| { |
| return fixup(self, fixswapcase); |
| } |
| |
| PyDoc_STRVAR(maketrans__doc__, |
| "str.maketrans(x[, y[, z]]) -> dict (static method)\n\ |
| \n\ |
| Return a translation table usable for str.translate().\n\ |
| If there is only one argument, it must be a dictionary mapping Unicode\n\ |
| ordinals (integers) or characters to Unicode ordinals, strings or None.\n\ |
| Character keys will be then converted to ordinals.\n\ |
| If there are two arguments, they must be strings of equal length, and\n\ |
| in the resulting dictionary, each character in x will be mapped to the\n\ |
| character at the same position in y. If there is a third argument, it\n\ |
| must be a string, whose characters will be mapped to None in the result."); |
| |
| static PyObject* |
| unicode_maketrans(PyUnicodeObject *null, PyObject *args) |
| { |
| PyObject *x, *y = NULL, *z = NULL; |
| PyObject *new = NULL, *key, *value; |
| Py_ssize_t i = 0; |
| int res; |
| |
| if (!PyArg_ParseTuple(args, "O|UU:maketrans", &x, &y, &z)) |
| return NULL; |
| new = PyDict_New(); |
| if (!new) |
| return NULL; |
| if (y != NULL) { |
| /* x must be a string too, of equal length */ |
| Py_ssize_t ylen = PyUnicode_GET_SIZE(y); |
| if (!PyUnicode_Check(x)) { |
| PyErr_SetString(PyExc_TypeError, "first maketrans argument must " |
| "be a string if there is a second argument"); |
| goto err; |
| } |
| if (PyUnicode_GET_SIZE(x) != ylen) { |
| PyErr_SetString(PyExc_ValueError, "the first two maketrans " |
| "arguments must have equal length"); |
| goto err; |
| } |
| /* create entries for translating chars in x to those in y */ |
| for (i = 0; i < PyUnicode_GET_SIZE(x); i++) { |
| key = PyLong_FromLong(PyUnicode_AS_UNICODE(x)[i]); |
| if (!key) |
| goto err; |
| value = PyLong_FromLong(PyUnicode_AS_UNICODE(y)[i]); |
| if (!value) { |
| Py_DECREF(key); |
| goto err; |
| } |
| res = PyDict_SetItem(new, key, value); |
| Py_DECREF(key); |
| Py_DECREF(value); |
| if (res < 0) |
| goto err; |
| } |
| /* create entries for deleting chars in z */ |
| if (z != NULL) { |
| for (i = 0; i < PyUnicode_GET_SIZE(z); i++) { |
| key = PyLong_FromLong(PyUnicode_AS_UNICODE(z)[i]); |
| if (!key) |
| goto err; |
| res = PyDict_SetItem(new, key, Py_None); |
| Py_DECREF(key); |
| if (res < 0) |
| goto err; |
| } |
| } |
| } else { |
| /* x must be a dict */ |
| if (!PyDict_CheckExact(x)) { |
| PyErr_SetString(PyExc_TypeError, "if you give only one argument " |
| "to maketrans it must be a dict"); |
| goto err; |
| } |
| /* copy entries into the new dict, converting string keys to int keys */ |
| while (PyDict_Next(x, &i, &key, &value)) { |
| if (PyUnicode_Check(key)) { |
| /* convert string keys to integer keys */ |
| PyObject *newkey; |
| if (PyUnicode_GET_SIZE(key) != 1) { |
| PyErr_SetString(PyExc_ValueError, "string keys in translate " |
| "table must be of length 1"); |
| goto err; |
| } |
| newkey = PyLong_FromLong(PyUnicode_AS_UNICODE(key)[0]); |
| if (!newkey) |
| goto err; |
| res = PyDict_SetItem(new, newkey, value); |
| Py_DECREF(newkey); |
| if (res < 0) |
| goto err; |
| } else if (PyLong_Check(key)) { |
| /* just keep integer keys */ |
| if (PyDict_SetItem(new, key, value) < 0) |
| goto err; |
| } else { |
| PyErr_SetString(PyExc_TypeError, "keys in translate table must " |
| "be strings or integers"); |
| goto err; |
| } |
| } |
| } |
| return new; |
| err: |
| Py_DECREF(new); |
| return NULL; |
| } |
| |
| PyDoc_STRVAR(translate__doc__, |
| "S.translate(table) -> str\n\ |
| \n\ |
| Return a copy of the string S, where all characters have been mapped\n\ |
| through the given translation table, which must be a mapping of\n\ |
| Unicode ordinals to Unicode ordinals, strings, or None.\n\ |
| Unmapped characters are left untouched. Characters mapped to None\n\ |
| are deleted."); |
| |
| static PyObject* |
| unicode_translate(PyUnicodeObject *self, PyObject *table) |
| { |
| return PyUnicode_TranslateCharmap(self->str, self->length, table, "ignore"); |
| } |
| |
| PyDoc_STRVAR(upper__doc__, |
| "S.upper() -> str\n\ |
| \n\ |
| Return a copy of S converted to uppercase."); |
| |
| static PyObject* |
| unicode_upper(PyUnicodeObject *self) |
| { |
| return fixup(self, fixupper); |
| } |
| |
| PyDoc_STRVAR(zfill__doc__, |
| "S.zfill(width) -> str\n\ |
| \n\ |
| Pad a numeric string S with zeros on the left, to fill a field\n\ |
| of the specified width. The string S is never truncated."); |
| |
| static PyObject * |
| unicode_zfill(PyUnicodeObject *self, PyObject *args) |
| { |
| Py_ssize_t fill; |
| PyUnicodeObject *u; |
| |
| Py_ssize_t width; |
| if (!PyArg_ParseTuple(args, "n:zfill", &width)) |
| return NULL; |
| |
| if (self->length >= width) { |
| if (PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject*) self; |
| } |
| else |
| return PyUnicode_FromUnicode( |
| PyUnicode_AS_UNICODE(self), |
| PyUnicode_GET_SIZE(self) |
| ); |
| } |
| |
| fill = width - self->length; |
| |
| u = pad(self, fill, 0, '0'); |
| |
| if (u == NULL) |
| return NULL; |
| |
| if (u->str[fill] == '+' || u->str[fill] == '-') { |
| /* move sign to beginning of string */ |
| u->str[0] = u->str[fill]; |
| u->str[fill] = '0'; |
| } |
| |
| return (PyObject*) u; |
| } |
| |
| #if 0 |
| static PyObject* |
| unicode_freelistsize(PyUnicodeObject *self) |
| { |
| return PyLong_FromLong(numfree); |
| } |
| |
| static PyObject * |
| unicode__decimal2ascii(PyObject *self) |
| { |
| return PyUnicode_TransformDecimalToASCII(PyUnicode_AS_UNICODE(self), |
| PyUnicode_GET_SIZE(self)); |
| } |
| #endif |
| |
| PyDoc_STRVAR(startswith__doc__, |
| "S.startswith(prefix[, start[, end]]) -> bool\n\ |
| \n\ |
| Return True if S starts with the specified prefix, False otherwise.\n\ |
| With optional start, test S beginning at that position.\n\ |
| With optional end, stop comparing S at that position.\n\ |
| prefix can also be a tuple of strings to try."); |
| |
| static PyObject * |
| unicode_startswith(PyUnicodeObject *self, |
| PyObject *args) |
| { |
| PyObject *subobj; |
| PyUnicodeObject *substring; |
| Py_ssize_t start = 0; |
| Py_ssize_t end = PY_SSIZE_T_MAX; |
| int result; |
| |
| if (!stringlib_parse_args_finds("startswith", args, &subobj, &start, &end)) |
| return NULL; |
| if (PyTuple_Check(subobj)) { |
| Py_ssize_t i; |
| for (i = 0; i < PyTuple_GET_SIZE(subobj); i++) { |
| substring = (PyUnicodeObject *)PyUnicode_FromObject( |
| PyTuple_GET_ITEM(subobj, i)); |
| if (substring == NULL) |
| return NULL; |
| result = tailmatch(self, substring, start, end, -1); |
| Py_DECREF(substring); |
| if (result) { |
| Py_RETURN_TRUE; |
| } |
| } |
| /* nothing matched */ |
| Py_RETURN_FALSE; |
| } |
| substring = (PyUnicodeObject *)PyUnicode_FromObject(subobj); |
| if (substring == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_TypeError)) |
| PyErr_Format(PyExc_TypeError, "startswith first arg must be str or " |
| "a tuple of str, not %s", Py_TYPE(subobj)->tp_name); |
| return NULL; |
| } |
| result = tailmatch(self, substring, start, end, -1); |
| Py_DECREF(substring); |
| return PyBool_FromLong(result); |
| } |
| |
| |
| PyDoc_STRVAR(endswith__doc__, |
| "S.endswith(suffix[, start[, end]]) -> bool\n\ |
| \n\ |
| Return True if S ends with the specified suffix, False otherwise.\n\ |
| With optional start, test S beginning at that position.\n\ |
| With optional end, stop comparing S at that position.\n\ |
| suffix can also be a tuple of strings to try."); |
| |
| static PyObject * |
| unicode_endswith(PyUnicodeObject *self, |
| PyObject *args) |
| { |
| PyObject *subobj; |
| PyUnicodeObject *substring; |
| Py_ssize_t start = 0; |
| Py_ssize_t end = PY_SSIZE_T_MAX; |
| int result; |
| |
| if (!stringlib_parse_args_finds("endswith", args, &subobj, &start, &end)) |
| return NULL; |
| if (PyTuple_Check(subobj)) { |
| Py_ssize_t i; |
| for (i = 0; i < PyTuple_GET_SIZE(subobj); i++) { |
| substring = (PyUnicodeObject *)PyUnicode_FromObject( |
| PyTuple_GET_ITEM(subobj, i)); |
| if (substring == NULL) |
| return NULL; |
| result = tailmatch(self, substring, start, end, +1); |
| Py_DECREF(substring); |
| if (result) { |
| Py_RETURN_TRUE; |
| } |
| } |
| Py_RETURN_FALSE; |
| } |
| substring = (PyUnicodeObject *)PyUnicode_FromObject(subobj); |
| if (substring == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_TypeError)) |
| PyErr_Format(PyExc_TypeError, "endswith first arg must be str or " |
| "a tuple of str, not %s", Py_TYPE(subobj)->tp_name); |
| return NULL; |
| } |
| result = tailmatch(self, substring, start, end, +1); |
| Py_DECREF(substring); |
| return PyBool_FromLong(result); |
| } |
| |
| #include "stringlib/string_format.h" |
| |
| PyDoc_STRVAR(format__doc__, |
| "S.format(*args, **kwargs) -> str\n\ |
| \n\ |
| Return a formatted version of S, using substitutions from args and kwargs.\n\ |
| The substitutions are identified by braces ('{' and '}')."); |
| |
| PyDoc_STRVAR(format_map__doc__, |
| "S.format_map(mapping) -> str\n\ |
| \n\ |
| Return a formatted version of S, using substitutions from mapping.\n\ |
| The substitutions are identified by braces ('{' and '}')."); |
| |
| static PyObject * |
| unicode__format__(PyObject* self, PyObject* args) |
| { |
| PyObject *format_spec; |
| |
| if (!PyArg_ParseTuple(args, "U:__format__", &format_spec)) |
| return NULL; |
| |
| return _PyUnicode_FormatAdvanced(self, |
| PyUnicode_AS_UNICODE(format_spec), |
| PyUnicode_GET_SIZE(format_spec)); |
| } |
| |
| PyDoc_STRVAR(p_format__doc__, |
| "S.__format__(format_spec) -> str\n\ |
| \n\ |
| Return a formatted version of S as described by format_spec."); |
| |
| static PyObject * |
| unicode__sizeof__(PyUnicodeObject *v) |
| { |
| return PyLong_FromSsize_t(sizeof(PyUnicodeObject) + |
| sizeof(Py_UNICODE) * (v->length + 1)); |
| } |
| |
| PyDoc_STRVAR(sizeof__doc__, |
| "S.__sizeof__() -> size of S in memory, in bytes"); |
| |
| static PyObject * |
| unicode_getnewargs(PyUnicodeObject *v) |
| { |
| return Py_BuildValue("(u#)", v->str, v->length); |
| } |
| |
| static PyMethodDef unicode_methods[] = { |
| {"encode", (PyCFunction) unicode_encode, METH_VARARGS | METH_KEYWORDS, encode__doc__}, |
| {"replace", (PyCFunction) unicode_replace, METH_VARARGS, replace__doc__}, |
| {"split", (PyCFunction) unicode_split, METH_VARARGS, split__doc__}, |
| {"rsplit", (PyCFunction) unicode_rsplit, METH_VARARGS, rsplit__doc__}, |
| {"join", (PyCFunction) unicode_join, METH_O, join__doc__}, |
| {"capitalize", (PyCFunction) unicode_capitalize, METH_NOARGS, capitalize__doc__}, |
| {"title", (PyCFunction) unicode_title, METH_NOARGS, title__doc__}, |
| {"center", (PyCFunction) unicode_center, METH_VARARGS, center__doc__}, |
| {"count", (PyCFunction) unicode_count, METH_VARARGS, count__doc__}, |
| {"expandtabs", (PyCFunction) unicode_expandtabs, METH_VARARGS, expandtabs__doc__}, |
| {"find", (PyCFunction) unicode_find, METH_VARARGS, find__doc__}, |
| {"partition", (PyCFunction) unicode_partition, METH_O, partition__doc__}, |
| {"index", (PyCFunction) unicode_index, METH_VARARGS, index__doc__}, |
| {"ljust", (PyCFunction) unicode_ljust, METH_VARARGS, ljust__doc__}, |
| {"lower", (PyCFunction) unicode_lower, METH_NOARGS, lower__doc__}, |
| {"lstrip", (PyCFunction) unicode_lstrip, METH_VARARGS, lstrip__doc__}, |
| {"rfind", (PyCFunction) unicode_rfind, METH_VARARGS, rfind__doc__}, |
| {"rindex", (PyCFunction) unicode_rindex, METH_VARARGS, rindex__doc__}, |
| {"rjust", (PyCFunction) unicode_rjust, METH_VARARGS, rjust__doc__}, |
| {"rstrip", (PyCFunction) unicode_rstrip, METH_VARARGS, rstrip__doc__}, |
| {"rpartition", (PyCFunction) unicode_rpartition, METH_O, rpartition__doc__}, |
| {"splitlines", (PyCFunction) unicode_splitlines, METH_VARARGS, splitlines__doc__}, |
| {"strip", (PyCFunction) unicode_strip, METH_VARARGS, strip__doc__}, |
| {"swapcase", (PyCFunction) unicode_swapcase, METH_NOARGS, swapcase__doc__}, |
| {"translate", (PyCFunction) unicode_translate, METH_O, translate__doc__}, |
| {"upper", (PyCFunction) unicode_upper, METH_NOARGS, upper__doc__}, |
| {"startswith", (PyCFunction) unicode_startswith, METH_VARARGS, startswith__doc__}, |
| {"endswith", (PyCFunction) unicode_endswith, METH_VARARGS, endswith__doc__}, |
| {"islower", (PyCFunction) unicode_islower, METH_NOARGS, islower__doc__}, |
| {"isupper", (PyCFunction) unicode_isupper, METH_NOARGS, isupper__doc__}, |
| {"istitle", (PyCFunction) unicode_istitle, METH_NOARGS, istitle__doc__}, |
| {"isspace", (PyCFunction) unicode_isspace, METH_NOARGS, isspace__doc__}, |
| {"isdecimal", (PyCFunction) unicode_isdecimal, METH_NOARGS, isdecimal__doc__}, |
| {"isdigit", (PyCFunction) unicode_isdigit, METH_NOARGS, isdigit__doc__}, |
| {"isnumeric", (PyCFunction) unicode_isnumeric, METH_NOARGS, isnumeric__doc__}, |
| {"isalpha", (PyCFunction) unicode_isalpha, METH_NOARGS, isalpha__doc__}, |
| {"isalnum", (PyCFunction) unicode_isalnum, METH_NOARGS, isalnum__doc__}, |
| {"isidentifier", (PyCFunction) unicode_isidentifier, METH_NOARGS, isidentifier__doc__}, |
| {"isprintable", (PyCFunction) unicode_isprintable, METH_NOARGS, isprintable__doc__}, |
| {"zfill", (PyCFunction) unicode_zfill, METH_VARARGS, zfill__doc__}, |
| {"format", (PyCFunction) do_string_format, METH_VARARGS | METH_KEYWORDS, format__doc__}, |
| {"format_map", (PyCFunction) do_string_format_map, METH_O, format_map__doc__}, |
| {"__format__", (PyCFunction) unicode__format__, METH_VARARGS, p_format__doc__}, |
| {"maketrans", (PyCFunction) unicode_maketrans, |
| METH_VARARGS | METH_STATIC, maketrans__doc__}, |
| {"__sizeof__", (PyCFunction) unicode__sizeof__, METH_NOARGS, sizeof__doc__}, |
| #if 0 |
| {"capwords", (PyCFunction) unicode_capwords, METH_NOARGS, capwords__doc__}, |
| #endif |
| |
| #if 0 |
| /* These methods are just used for debugging the implementation. */ |
| {"freelistsize", (PyCFunction) unicode_freelistsize, METH_NOARGS}, |
| {"_decimal2ascii", (PyCFunction) unicode__decimal2ascii, METH_NOARGS}, |
| #endif |
| |
| {"__getnewargs__", (PyCFunction)unicode_getnewargs, METH_NOARGS}, |
| {NULL, NULL} |
| }; |
| |
| static PyObject * |
| unicode_mod(PyObject *v, PyObject *w) |
| { |
| if (!PyUnicode_Check(v)) { |
| Py_INCREF(Py_NotImplemented); |
| return Py_NotImplemented; |
| } |
| return PyUnicode_Format(v, w); |
| } |
| |
| static PyNumberMethods unicode_as_number = { |
| 0, /*nb_add*/ |
| 0, /*nb_subtract*/ |
| 0, /*nb_multiply*/ |
| unicode_mod, /*nb_remainder*/ |
| }; |
| |
| static PySequenceMethods unicode_as_sequence = { |
| (lenfunc) unicode_length, /* sq_length */ |
| PyUnicode_Concat, /* sq_concat */ |
| (ssizeargfunc) unicode_repeat, /* sq_repeat */ |
| (ssizeargfunc) unicode_getitem, /* sq_item */ |
| 0, /* sq_slice */ |
| 0, /* sq_ass_item */ |
| 0, /* sq_ass_slice */ |
| PyUnicode_Contains, /* sq_contains */ |
| }; |
| |
| static PyObject* |
| unicode_subscript(PyUnicodeObject* self, PyObject* item) |
| { |
| if (PyIndex_Check(item)) { |
| Py_ssize_t i = PyNumber_AsSsize_t(item, PyExc_IndexError); |
| if (i == -1 && PyErr_Occurred()) |
| return NULL; |
| if (i < 0) |
| i += PyUnicode_GET_SIZE(self); |
| return unicode_getitem(self, i); |
| } else if (PySlice_Check(item)) { |
| Py_ssize_t start, stop, step, slicelength, cur, i; |
| Py_UNICODE* source_buf; |
| Py_UNICODE* result_buf; |
| PyObject* result; |
| |
| if (PySlice_GetIndicesEx(item, PyUnicode_GET_SIZE(self), |
| &start, &stop, &step, &slicelength) < 0) { |
| return NULL; |
| } |
| |
| if (slicelength <= 0) { |
| return PyUnicode_FromUnicode(NULL, 0); |
| } else if (start == 0 && step == 1 && slicelength == self->length && |
| PyUnicode_CheckExact(self)) { |
| Py_INCREF(self); |
| return (PyObject *)self; |
| } else if (step == 1) { |
| return PyUnicode_FromUnicode(self->str + start, slicelength); |
| } else { |
| source_buf = PyUnicode_AS_UNICODE((PyObject*)self); |
| result_buf = (Py_UNICODE *)PyObject_MALLOC(slicelength* |
| sizeof(Py_UNICODE)); |
| |
| if (result_buf == NULL) |
| return PyErr_NoMemory(); |
| |
| for (cur = start, i = 0; i < slicelength; cur += step, i++) { |
| result_buf[i] = source_buf[cur]; |
| } |
| |
| result = PyUnicode_FromUnicode(result_buf, slicelength); |
| PyObject_FREE(result_buf); |
| return result; |
| } |
| } else { |
| PyErr_SetString(PyExc_TypeError, "string indices must be integers"); |
| return NULL; |
| } |
| } |
| |
| static PyMappingMethods unicode_as_mapping = { |
| (lenfunc)unicode_length, /* mp_length */ |
| (binaryfunc)unicode_subscript, /* mp_subscript */ |
| (objobjargproc)0, /* mp_ass_subscript */ |
| }; |
| |
| |
| /* Helpers for PyUnicode_Format() */ |
| |
| static PyObject * |
| getnextarg(PyObject *args, Py_ssize_t arglen, Py_ssize_t *p_argidx) |
| { |
| Py_ssize_t argidx = *p_argidx; |
| if (argidx < arglen) { |
| (*p_argidx)++; |
| if (arglen < 0) |
| return args; |
| else |
| return PyTuple_GetItem(args, argidx); |
| } |
| PyErr_SetString(PyExc_TypeError, |
| "not enough arguments for format string"); |
| return NULL; |
| } |
| |
| /* Returns a new reference to a PyUnicode object, or NULL on failure. */ |
| |
| static PyObject * |
| formatfloat(PyObject *v, int flags, int prec, int type) |
| { |
| char *p; |
| PyObject *result; |
| double x; |
| |
| x = PyFloat_AsDouble(v); |
| if (x == -1.0 && PyErr_Occurred()) |
| return NULL; |
| |
| if (prec < 0) |
| prec = 6; |
| |
| p = PyOS_double_to_string(x, type, prec, |
| (flags & F_ALT) ? Py_DTSF_ALT : 0, NULL); |
| if (p == NULL) |
| return NULL; |
| result = PyUnicode_FromStringAndSize(p, strlen(p)); |
| PyMem_Free(p); |
| return result; |
| } |
| |
| static PyObject* |
| formatlong(PyObject *val, int flags, int prec, int type) |
| { |
| char *buf; |
| int len; |
| PyObject *str; /* temporary string object. */ |
| PyObject *result; |
| |
| str = _PyBytes_FormatLong(val, flags, prec, type, &buf, &len); |
| if (!str) |
| return NULL; |
| result = PyUnicode_FromStringAndSize(buf, len); |
| Py_DECREF(str); |
| return result; |
| } |
| |
| static int |
| formatchar(Py_UNICODE *buf, |
| size_t buflen, |
| PyObject *v) |
| { |
| /* presume that the buffer is at least 3 characters long */ |
| if (PyUnicode_Check(v)) { |
| if (PyUnicode_GET_SIZE(v) == 1) { |
| buf[0] = PyUnicode_AS_UNICODE(v)[0]; |
| buf[1] = '\0'; |
| return 1; |
| } |
| #ifndef Py_UNICODE_WIDE |
| if (PyUnicode_GET_SIZE(v) == 2) { |
| /* Decode a valid surrogate pair */ |
| int c0 = PyUnicode_AS_UNICODE(v)[0]; |
| int c1 = PyUnicode_AS_UNICODE(v)[1]; |
| if (0xD800 <= c0 && c0 <= 0xDBFF && |
| 0xDC00 <= c1 && c1 <= 0xDFFF) { |
| buf[0] = c0; |
| buf[1] = c1; |
| buf[2] = '\0'; |
| return 2; |
| } |
| } |
| #endif |
| goto onError; |
| } |
| else { |
| /* Integer input truncated to a character */ |
| long x; |
| x = PyLong_AsLong(v); |
| if (x == -1 && PyErr_Occurred()) |
| goto onError; |
| |
| if (x < 0 || x > 0x10ffff) { |
| PyErr_SetString(PyExc_OverflowError, |
| "%c arg not in range(0x110000)"); |
| return -1; |
| } |
| |
| #ifndef Py_UNICODE_WIDE |
| if (x > 0xffff) { |
| x -= 0x10000; |
| buf[0] = (Py_UNICODE)(0xD800 | (x >> 10)); |
| buf[1] = (Py_UNICODE)(0xDC00 | (x & 0x3FF)); |
| return 2; |
| } |
| #endif |
| buf[0] = (Py_UNICODE) x; |
| buf[1] = '\0'; |
| return 1; |
| } |
| |
| onError: |
| PyErr_SetString(PyExc_TypeError, |
| "%c requires int or char"); |
| return -1; |
| } |
| |
| /* fmt%(v1,v2,...) is roughly equivalent to sprintf(fmt, v1, v2, ...) |
| FORMATBUFLEN is the length of the buffer in which chars are formatted. |
| */ |
| #define FORMATBUFLEN (size_t)10 |
| |
| PyObject *PyUnicode_Format(PyObject *format, |
| PyObject *args) |
| { |
| Py_UNICODE *fmt, *res; |
| Py_ssize_t fmtcnt, rescnt, reslen, arglen, argidx; |
| int args_owned = 0; |
| PyUnicodeObject *result = NULL; |
| PyObject *dict = NULL; |
| PyObject *uformat; |
| |
| if (format == NULL || args == NULL) { |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| uformat = PyUnicode_FromObject(format); |
| if (uformat == NULL) |
| return NULL; |
| fmt = PyUnicode_AS_UNICODE(uformat); |
| fmtcnt = PyUnicode_GET_SIZE(uformat); |
| |
| reslen = rescnt = fmtcnt + 100; |
| result = _PyUnicode_New(reslen); |
| if (result == NULL) |
| goto onError; |
| res = PyUnicode_AS_UNICODE(result); |
| |
| if (PyTuple_Check(args)) { |
| arglen = PyTuple_Size(args); |
| argidx = 0; |
| } |
| else { |
| arglen = -1; |
| argidx = -2; |
| } |
| if (PyMapping_Check(args) && !PyTuple_Check(args) && !PyUnicode_Check(args)) |
| dict = args; |
| |
| while (--fmtcnt >= 0) { |
| if (*fmt != '%') { |
| if (--rescnt < 0) { |
| rescnt = fmtcnt + 100; |
| reslen += rescnt; |
| if (_PyUnicode_Resize(&result, reslen) < 0) |
| goto onError; |
| res = PyUnicode_AS_UNICODE(result) + reslen - rescnt; |
| --rescnt; |
| } |
| *res++ = *fmt++; |
| } |
| else { |
| /* Got a format specifier */ |
| int flags = 0; |
| Py_ssize_t width = -1; |
| int prec = -1; |
| Py_UNICODE c = '\0'; |
| Py_UNICODE fill; |
| int isnumok; |
| PyObject *v = NULL; |
| PyObject *temp = NULL; |
| Py_UNICODE *pbuf; |
| Py_UNICODE sign; |
| Py_ssize_t len; |
| Py_UNICODE formatbuf[FORMATBUFLEN]; /* For formatchar() */ |
| |
| fmt++; |
| if (*fmt == '(') { |
| Py_UNICODE *keystart; |
| Py_ssize_t keylen; |
| PyObject *key; |
| int pcount = 1; |
| |
| if (dict == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "format requires a mapping"); |
| goto onError; |
| } |
| ++fmt; |
| --fmtcnt; |
| keystart = fmt; |
| /* Skip over balanced parentheses */ |
| while (pcount > 0 && --fmtcnt >= 0) { |
| if (*fmt == ')') |
| --pcount; |
| else if (*fmt == '(') |
| ++pcount; |
| fmt++; |
| } |
| keylen = fmt - keystart - 1; |
| if (fmtcnt < 0 || pcount > 0) { |
| PyErr_SetString(PyExc_ValueError, |
| "incomplete format key"); |
| goto onError; |
| } |
| #if 0 |
| /* keys are converted to strings using UTF-8 and |
| then looked up since Python uses strings to hold |
| variables names etc. in its namespaces and we |
| wouldn't want to break common idioms. */ |
| key = PyUnicode_EncodeUTF8(keystart, |
| keylen, |
| NULL); |
| #else |
| key = PyUnicode_FromUnicode(keystart, keylen); |
| #endif |
| if (key == NULL) |
| goto onError; |
| if (args_owned) { |
| Py_DECREF(args); |
| args_owned = 0; |
| } |
| args = PyObject_GetItem(dict, key); |
| Py_DECREF(key); |
| if (args == NULL) { |
| goto onError; |
| } |
| args_owned = 1; |
| arglen = -1; |
| argidx = -2; |
| } |
| while (--fmtcnt >= 0) { |
| switch (c = *fmt++) { |
| case '-': flags |= F_LJUST; continue; |
| case '+': flags |= F_SIGN; continue; |
| case ' ': flags |= F_BLANK; continue; |
| case '#': flags |= F_ALT; continue; |
| case '0': flags |= F_ZERO; continue; |
| } |
| break; |
| } |
| if (c == '*') { |
| v = getnextarg(args, arglen, &argidx); |
| if (v == NULL) |
| goto onError; |
| if (!PyLong_Check(v)) { |
| PyErr_SetString(PyExc_TypeError, |
| "* wants int"); |
| goto onError; |
| } |
| width = PyLong_AsSsize_t(v); |
| if (width == -1 && PyErr_Occurred()) |
| goto onError; |
| if (width < 0) { |
| flags |= F_LJUST; |
| width = -width; |
| } |
| if (--fmtcnt >= 0) |
| c = *fmt++; |
| } |
| else if (c >= '0' && c <= '9') { |
| width = c - '0'; |
| while (--fmtcnt >= 0) { |
| c = *fmt++; |
| if (c < '0' || c > '9') |
| break; |
| if (width > (PY_SSIZE_T_MAX - ((int)c - '0')) / 10) { |
| PyErr_SetString(PyExc_ValueError, |
| "width too big"); |
| goto onError; |
| } |
| width = width*10 + (c - '0'); |
| } |
| } |
| if (c == '.') { |
| prec = 0; |
| if (--fmtcnt >= 0) |
| c = *fmt++; |
| if (c == '*') { |
| v = getnextarg(args, arglen, &argidx); |
| if (v == NULL) |
| goto onError; |
| if (!PyLong_Check(v)) { |
| PyErr_SetString(PyExc_TypeError, |
| "* wants int"); |
| goto onError; |
| } |
| prec = _PyLong_AsInt(v); |
| if (prec == -1 && PyErr_Occurred()) |
| goto onError; |
| if (prec < 0) |
| prec = 0; |
| if (--fmtcnt >= 0) |
| c = *fmt++; |
| } |
| else if (c >= '0' && c <= '9') { |
| prec = c - '0'; |
| while (--fmtcnt >= 0) { |
| c = *fmt++; |
| if (c < '0' || c > '9') |
| break; |
| if (prec > (INT_MAX - ((int)c - '0')) / 10) { |
| PyErr_SetString(PyExc_ValueError, |
| "prec too big"); |
| goto onError; |
| } |
| prec = prec*10 + (c - '0'); |
| } |
| } |
| } /* prec */ |
| if (fmtcnt >= 0) { |
| if (c == 'h' || c == 'l' || c == 'L') { |
| if (--fmtcnt >= 0) |
| c = *fmt++; |
| } |
| } |
| if (fmtcnt < 0) { |
| PyErr_SetString(PyExc_ValueError, |
| "incomplete format"); |
| goto onError; |
| } |
| if (c != '%') { |
| v = getnextarg(args, arglen, &argidx); |
| if (v == NULL) |
| goto onError; |
| } |
| sign = 0; |
| fill = ' '; |
| switch (c) { |
| |
| case '%': |
| pbuf = formatbuf; |
| /* presume that buffer length is at least 1 */ |
| pbuf[0] = '%'; |
| len = 1; |
| break; |
| |
| case 's': |
| case 'r': |
| case 'a': |
| if (PyUnicode_CheckExact(v) && c == 's') { |
| temp = v; |
| Py_INCREF(temp); |
| } |
| else { |
| if (c == 's') |
| temp = PyObject_Str(v); |
| else if (c == 'r') |
| temp = PyObject_Repr(v); |
| else |
| temp = PyObject_ASCII(v); |
| if (temp == NULL) |
| goto onError; |
| if (PyUnicode_Check(temp)) |
| /* nothing to do */; |
| else { |
| Py_DECREF(temp); |
| PyErr_SetString(PyExc_TypeError, |
| "%s argument has non-string str()"); |
| goto onError; |
| } |
| } |
| pbuf = PyUnicode_AS_UNICODE(temp); |
| len = PyUnicode_GET_SIZE(temp); |
| if (prec >= 0 && len > prec) |
| len = prec; |
| break; |
| |
| case 'i': |
| case 'd': |
| case 'u': |
| case 'o': |
| case 'x': |
| case 'X': |
| isnumok = 0; |
| if (PyNumber_Check(v)) { |
| PyObject *iobj=NULL; |
| |
| if (PyLong_Check(v)) { |
| iobj = v; |
| Py_INCREF(iobj); |
| } |
| else { |
| iobj = PyNumber_Long(v); |
| } |
| if (iobj!=NULL) { |
| if (PyLong_Check(iobj)) { |
| isnumok = 1; |
| temp = formatlong(iobj, flags, prec, (c == 'i'? 'd': c)); |
| Py_DECREF(iobj); |
| if (!temp) |
| goto onError; |
| pbuf = PyUnicode_AS_UNICODE(temp); |
| len = PyUnicode_GET_SIZE(temp); |
| sign = 1; |
| } |
| else { |
| Py_DECREF(iobj); |
| } |
| } |
| } |
| if (!isnumok) { |
| PyErr_Format(PyExc_TypeError, |
| "%%%c format: a number is required, " |
| "not %.200s", (char)c, Py_TYPE(v)->tp_name); |
| goto onError; |
| } |
| if (flags & F_ZERO) |
| fill = '0'; |
| break; |
| |
| case 'e': |
| case 'E': |
| case 'f': |
| case 'F': |
| case 'g': |
| case 'G': |
| temp = formatfloat(v, flags, prec, c); |
| if (!temp) |
| goto onError; |
| pbuf = PyUnicode_AS_UNICODE(temp); |
| len = PyUnicode_GET_SIZE(temp); |
| sign = 1; |
| if (flags & F_ZERO) |
| fill = '0'; |
| break; |
| |
| case 'c': |
| pbuf = formatbuf; |
| len = formatchar(pbuf, sizeof(formatbuf)/sizeof(Py_UNICODE), v); |
| if (len < 0) |
| goto onError; |
| break; |
| |
| default: |
| PyErr_Format(PyExc_ValueError, |
| "unsupported format character '%c' (0x%x) " |
| "at index %zd", |
| (31<=c && c<=126) ? (char)c : '?', |
| (int)c, |
| (Py_ssize_t)(fmt - 1 - |
| PyUnicode_AS_UNICODE(uformat))); |
| goto onError; |
| } |
| if (sign) { |
| if (*pbuf == '-' || *pbuf == '+') { |
| sign = *pbuf++; |
| len--; |
| } |
| else if (flags & F_SIGN) |
| sign = '+'; |
| else if (flags & F_BLANK) |
| sign = ' '; |
| else |
| sign = 0; |
| } |
| if (width < len) |
| width = len; |
| if (rescnt - (sign != 0) < width) { |
| reslen -= rescnt; |
| rescnt = width + fmtcnt + 100; |
| reslen += rescnt; |
| if (reslen < 0) { |
| Py_XDECREF(temp); |
| PyErr_NoMemory(); |
| goto onError; |
| } |
| if (_PyUnicode_Resize(&result, reslen) < 0) { |
| Py_XDECREF(temp); |
| goto onError; |
| } |
| res = PyUnicode_AS_UNICODE(result) |
| + reslen - rescnt; |
| } |
| if (sign) { |
| if (fill != ' ') |
| *res++ = sign; |
| rescnt--; |
| if (width > len) |
| width--; |
| } |
| if ((flags & F_ALT) && (c == 'x' || c == 'X' || c == 'o')) { |
| assert(pbuf[0] == '0'); |
| assert(pbuf[1] == c); |
| if (fill != ' ') { |
| *res++ = *pbuf++; |
| *res++ = *pbuf++; |
| } |
| rescnt -= 2; |
| width -= 2; |
| if (width < 0) |
| width = 0; |
| len -= 2; |
| } |
| if (width > len && !(flags & F_LJUST)) { |
| do { |
| --rescnt; |
| *res++ = fill; |
| } while (--width > len); |
| } |
| if (fill == ' ') { |
| if (sign) |
| *res++ = sign; |
| if ((flags & F_ALT) && (c == 'x' || c == 'X' || c == 'o')) { |
| assert(pbuf[0] == '0'); |
| assert(pbuf[1] == c); |
| *res++ = *pbuf++; |
| *res++ = *pbuf++; |
| } |
| } |
| Py_UNICODE_COPY(res, pbuf, len); |
| res += len; |
| rescnt -= len; |
| while (--width >= len) { |
| --rescnt; |
| *res++ = ' '; |
| } |
| if (dict && (argidx < arglen) && c != '%') { |
| PyErr_SetString(PyExc_TypeError, |
| "not all arguments converted during string formatting"); |
| Py_XDECREF(temp); |
| goto onError; |
| } |
| Py_XDECREF(temp); |
| } /* '%' */ |
| } /* until end */ |
| if (argidx < arglen && !dict) { |
| PyErr_SetString(PyExc_TypeError, |
| "not all arguments converted during string formatting"); |
| goto onError; |
| } |
| |
| if (_PyUnicode_Resize(&result, reslen - rescnt) < 0) |
| goto onError; |
| if (args_owned) { |
| Py_DECREF(args); |
| } |
| Py_DECREF(uformat); |
| return (PyObject *)result; |
| |
| onError: |
| Py_XDECREF(result); |
| Py_DECREF(uformat); |
| if (args_owned) { |
| Py_DECREF(args); |
| } |
| return NULL; |
| } |
| |
| static PyObject * |
| unicode_subtype_new(PyTypeObject *type, PyObject *args, PyObject *kwds); |
| |
| static PyObject * |
| unicode_new(PyTypeObject *type, PyObject *args, PyObject *kwds) |
| { |
| PyObject *x = NULL; |
| static char *kwlist[] = {"object", "encoding", "errors", 0}; |
| char *encoding = NULL; |
| char *errors = NULL; |
| |
| if (type != &PyUnicode_Type) |
| return unicode_subtype_new(type, args, kwds); |
| if (!PyArg_ParseTupleAndKeywords(args, kwds, "|Oss:str", |
| kwlist, &x, &encoding, &errors)) |
| return NULL; |
| if (x == NULL) |
| return (PyObject *)_PyUnicode_New(0); |
| if (encoding == NULL && errors == NULL) |
| return PyObject_Str(x); |
| else |
| return PyUnicode_FromEncodedObject(x, encoding, errors); |
| } |
| |
| static PyObject * |
| unicode_subtype_new(PyTypeObject *type, PyObject *args, PyObject *kwds) |
| { |
| PyUnicodeObject *tmp, *pnew; |
| Py_ssize_t n; |
| |
| assert(PyType_IsSubtype(type, &PyUnicode_Type)); |
| tmp = (PyUnicodeObject *)unicode_new(&PyUnicode_Type, args, kwds); |
| if (tmp == NULL) |
| return NULL; |
| assert(PyUnicode_Check(tmp)); |
| pnew = (PyUnicodeObject *) type->tp_alloc(type, n = tmp->length); |
| if (pnew == NULL) { |
| Py_DECREF(tmp); |
| return NULL; |
| } |
| pnew->str = (Py_UNICODE*) PyObject_MALLOC(sizeof(Py_UNICODE) * (n+1)); |
| if (pnew->str == NULL) { |
| _Py_ForgetReference((PyObject *)pnew); |
| PyObject_Del(pnew); |
| Py_DECREF(tmp); |
| return PyErr_NoMemory(); |
| } |
| Py_UNICODE_COPY(pnew->str, tmp->str, n+1); |
| pnew->length = n; |
| pnew->hash = tmp->hash; |
| Py_DECREF(tmp); |
| return (PyObject *)pnew; |
| } |
| |
| PyDoc_STRVAR(unicode_doc, |
| "str(object='') -> str\n\ |
| str(bytes_or_buffer[, encoding[, errors]]) -> str\n\ |
| \n\ |
| Create a new string object from the given object. If encoding or\n\ |
| errors is specified, then the object must expose a data buffer\n\ |
| that will be decoded using the given encoding and error handler.\n\ |
| Otherwise, returns the result of object.__str__() (if defined)\n\ |
| or repr(object).\n\ |
| encoding defaults to sys.getdefaultencoding().\n\ |
| errors defaults to 'strict'."); |
| |
| static PyObject *unicode_iter(PyObject *seq); |
| |
| PyTypeObject PyUnicode_Type = { |
| PyVarObject_HEAD_INIT(&PyType_Type, 0) |
| "str", /* tp_name */ |
| sizeof(PyUnicodeObject), /* tp_size */ |
| 0, /* tp_itemsize */ |
| /* Slots */ |
| (destructor)unicode_dealloc, /* tp_dealloc */ |
| 0, /* tp_print */ |
| 0, /* tp_getattr */ |
| 0, /* tp_setattr */ |
| 0, /* tp_reserved */ |
| unicode_repr, /* tp_repr */ |
| &unicode_as_number, /* tp_as_number */ |
| &unicode_as_sequence, /* tp_as_sequence */ |
| &unicode_as_mapping, /* tp_as_mapping */ |
| (hashfunc) unicode_hash, /* tp_hash*/ |
| 0, /* tp_call*/ |
| (reprfunc) unicode_str, /* tp_str */ |
| PyObject_GenericGetAttr, /* tp_getattro */ |
| 0, /* tp_setattro */ |
| 0, /* tp_as_buffer */ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | |
| Py_TPFLAGS_UNICODE_SUBCLASS, /* tp_flags */ |
| unicode_doc, /* tp_doc */ |
| 0, /* tp_traverse */ |
| 0, /* tp_clear */ |
| PyUnicode_RichCompare, /* tp_richcompare */ |
| 0, /* tp_weaklistoffset */ |
| unicode_iter, /* tp_iter */ |
| 0, /* tp_iternext */ |
| unicode_methods, /* tp_methods */ |
| 0, /* tp_members */ |
| 0, /* tp_getset */ |
| &PyBaseObject_Type, /* tp_base */ |
| 0, /* tp_dict */ |
| 0, /* tp_descr_get */ |
| 0, /* tp_descr_set */ |
| 0, /* tp_dictoffset */ |
| 0, /* tp_init */ |
| 0, /* tp_alloc */ |
| unicode_new, /* tp_new */ |
| PyObject_Del, /* tp_free */ |
| }; |
| |
| /* Initialize the Unicode implementation */ |
| |
| void _PyUnicode_Init(void) |
| { |
| /* XXX - move this array to unicodectype.c ? */ |
| Py_UNICODE linebreak[] = { |
| 0x000A, /* LINE FEED */ |
| 0x000D, /* CARRIAGE RETURN */ |
| 0x001C, /* FILE SEPARATOR */ |
| 0x001D, /* GROUP SEPARATOR */ |
| 0x001E, /* RECORD SEPARATOR */ |
| 0x0085, /* NEXT LINE */ |
| 0x2028, /* LINE SEPARATOR */ |
| 0x2029, /* PARAGRAPH SEPARATOR */ |
| }; |
| |
| /* Init the implementation */ |
| if (!unicode_empty) { |
| unicode_empty = _PyUnicode_New(0); |
| if (!unicode_empty) |
| return; |
| } |
| |
| if (PyType_Ready(&PyUnicode_Type) < 0) |
| Py_FatalError("Can't initialize 'unicode'"); |
| |
| /* initialize the linebreak bloom filter */ |
| bloom_linebreak = make_bloom_mask( |
| linebreak, sizeof(linebreak) / sizeof(linebreak[0]) |
| ); |
| |
| PyType_Ready(&EncodingMapType); |
| |
| if (PyType_Ready(&PyFieldNameIter_Type) < 0) |
| Py_FatalError("Can't initialize field name iterator type"); |
| |
| if (PyType_Ready(&PyFormatterIter_Type) < 0) |
| Py_FatalError("Can't initialize formatter iter type"); |
| } |
| |
| /* Finalize the Unicode implementation */ |
| |
| int |
| PyUnicode_ClearFreeList(void) |
| { |
| int freelist_size = numfree; |
| PyUnicodeObject *u; |
| |
| for (u = free_list; u != NULL;) { |
| PyUnicodeObject *v = u; |
| u = *(PyUnicodeObject **)u; |
| if (v->str) |
| PyObject_DEL(v->str); |
| Py_XDECREF(v->defenc); |
| PyObject_Del(v); |
| numfree--; |
| } |
| free_list = NULL; |
| assert(numfree == 0); |
| return freelist_size; |
| } |
| |
| void |
| _PyUnicode_Fini(void) |
| { |
| int i; |
| |
| Py_CLEAR(unicode_empty); |
| |
| for (i = 0; i < 256; i++) |
| Py_CLEAR(unicode_latin1[i]); |
| |
| (void)PyUnicode_ClearFreeList(); |
| } |
| |
| void |
| PyUnicode_InternInPlace(PyObject **p) |
| { |
| register PyUnicodeObject *s = (PyUnicodeObject *)(*p); |
| PyObject *t; |
| if (s == NULL || !PyUnicode_Check(s)) |
| Py_FatalError( |
| "PyUnicode_InternInPlace: unicode strings only please!"); |
| /* If it's a subclass, we don't really know what putting |
| it in the interned dict might do. */ |
| if (!PyUnicode_CheckExact(s)) |
| return; |
| if (PyUnicode_CHECK_INTERNED(s)) |
| return; |
| if (interned == NULL) { |
| interned = PyDict_New(); |
| if (interned == NULL) { |
| PyErr_Clear(); /* Don't leave an exception */ |
| return; |
| } |
| } |
| /* It might be that the GetItem call fails even |
| though the key is present in the dictionary, |
| namely when this happens during a stack overflow. */ |
| Py_ALLOW_RECURSION |
| t = PyDict_GetItem(interned, (PyObject *)s); |
| Py_END_ALLOW_RECURSION |
| |
| if (t) { |
| Py_INCREF(t); |
| Py_DECREF(*p); |
| *p = t; |
| return; |
| } |
| |
| PyThreadState_GET()->recursion_critical = 1; |
| if (PyDict_SetItem(interned, (PyObject *)s, (PyObject *)s) < 0) { |
| PyErr_Clear(); |
| PyThreadState_GET()->recursion_critical = 0; |
| return; |
| } |
| PyThreadState_GET()->recursion_critical = 0; |
| /* The two references in interned are not counted by refcnt. |
| The deallocator will take care of this */ |
| Py_REFCNT(s) -= 2; |
| PyUnicode_CHECK_INTERNED(s) = SSTATE_INTERNED_MORTAL; |
| } |
| |
| void |
| PyUnicode_InternImmortal(PyObject **p) |
| { |
| PyUnicode_InternInPlace(p); |
| if (PyUnicode_CHECK_INTERNED(*p) != SSTATE_INTERNED_IMMORTAL) { |
| PyUnicode_CHECK_INTERNED(*p) = SSTATE_INTERNED_IMMORTAL; |
| Py_INCREF(*p); |
| } |
| } |
| |
| PyObject * |
| PyUnicode_InternFromString(const char *cp) |
| { |
| PyObject *s = PyUnicode_FromString(cp); |
| if (s == NULL) |
| return NULL; |
| PyUnicode_InternInPlace(&s); |
| return s; |
| } |
| |
| void _Py_ReleaseInternedUnicodeStrings(void) |
| { |
| PyObject *keys; |
| PyUnicodeObject *s; |
| Py_ssize_t i, n; |
| Py_ssize_t immortal_size = 0, mortal_size = 0; |
| |
| if (interned == NULL || !PyDict_Check(interned)) |
| return; |
| keys = PyDict_Keys(interned); |
| if (keys == NULL || !PyList_Check(keys)) { |
| PyErr_Clear(); |
| return; |
| } |
| |
| /* Since _Py_ReleaseInternedUnicodeStrings() is intended to help a leak |
| detector, interned unicode strings are not forcibly deallocated; |
| rather, we give them their stolen references back, and then clear |
| and DECREF the interned dict. */ |
| |
| n = PyList_GET_SIZE(keys); |
| fprintf(stderr, "releasing %" PY_FORMAT_SIZE_T "d interned strings\n", |
| n); |
| for (i = 0; i < n; i++) { |
| s = (PyUnicodeObject *) PyList_GET_ITEM(keys, i); |
| switch (s->state) { |
| case SSTATE_NOT_INTERNED: |
| /* XXX Shouldn't happen */ |
| break; |
| case SSTATE_INTERNED_IMMORTAL: |
| Py_REFCNT(s) += 1; |
| immortal_size += s->length; |
| break; |
| case SSTATE_INTERNED_MORTAL: |
| Py_REFCNT(s) += 2; |
| mortal_size += s->length; |
| break; |
| default: |
| Py_FatalError("Inconsistent interned string state."); |
| } |
| s->state = SSTATE_NOT_INTERNED; |
| } |
| fprintf(stderr, "total size of all interned strings: " |
| "%" PY_FORMAT_SIZE_T "d/%" PY_FORMAT_SIZE_T "d " |
| "mortal/immortal\n", mortal_size, immortal_size); |
| Py_DECREF(keys); |
| PyDict_Clear(interned); |
| Py_CLEAR(interned); |
| } |
| |
| |
| /********************* Unicode Iterator **************************/ |
| |
| typedef struct { |
| PyObject_HEAD |
| Py_ssize_t it_index; |
| PyUnicodeObject *it_seq; /* Set to NULL when iterator is exhausted */ |
| } unicodeiterobject; |
| |
| static void |
| unicodeiter_dealloc(unicodeiterobject *it) |
| { |
| _PyObject_GC_UNTRACK(it); |
| Py_XDECREF(it->it_seq); |
| PyObject_GC_Del(it); |
| } |
| |
| static int |
| unicodeiter_traverse(unicodeiterobject *it, visitproc visit, void *arg) |
| { |
| Py_VISIT(it->it_seq); |
| return 0; |
| } |
| |
| static PyObject * |
| unicodeiter_next(unicodeiterobject *it) |
| { |
| PyUnicodeObject *seq; |
| PyObject *item; |
| |
| assert(it != NULL); |
| seq = it->it_seq; |
| if (seq == NULL) |
| return NULL; |
| assert(PyUnicode_Check(seq)); |
| |
| if (it->it_index < PyUnicode_GET_SIZE(seq)) { |
| item = PyUnicode_FromUnicode( |
| PyUnicode_AS_UNICODE(seq)+it->it_index, 1); |
| if (item != NULL) |
| ++it->it_index; |
| return item; |
| } |
| |
| Py_DECREF(seq); |
| it->it_seq = NULL; |
| return NULL; |
| } |
| |
| static PyObject * |
| unicodeiter_len(unicodeiterobject *it) |
| { |
| Py_ssize_t len = 0; |
| if (it->it_seq) |
| len = PyUnicode_GET_SIZE(it->it_seq) - it->it_index; |
| return PyLong_FromSsize_t(len); |
| } |
| |
| PyDoc_STRVAR(length_hint_doc, "Private method returning an estimate of len(list(it))."); |
| |
| static PyMethodDef unicodeiter_methods[] = { |
| {"__length_hint__", (PyCFunction)unicodeiter_len, METH_NOARGS, |
| length_hint_doc}, |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| PyTypeObject PyUnicodeIter_Type = { |
| PyVarObject_HEAD_INIT(&PyType_Type, 0) |
| "str_iterator", /* tp_name */ |
| sizeof(unicodeiterobject), /* tp_basicsize */ |
| 0, /* tp_itemsize */ |
| /* methods */ |
| (destructor)unicodeiter_dealloc, /* tp_dealloc */ |
| 0, /* tp_print */ |
| 0, /* tp_getattr */ |
| 0, /* tp_setattr */ |
| 0, /* tp_reserved */ |
| 0, /* tp_repr */ |
| 0, /* tp_as_number */ |
| 0, /* tp_as_sequence */ |
| 0, /* tp_as_mapping */ |
| 0, /* tp_hash */ |
| 0, /* tp_call */ |
| 0, /* tp_str */ |
| PyObject_GenericGetAttr, /* tp_getattro */ |
| 0, /* tp_setattro */ |
| 0, /* tp_as_buffer */ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_HAVE_GC,/* tp_flags */ |
| 0, /* tp_doc */ |
| (traverseproc)unicodeiter_traverse, /* tp_traverse */ |
| 0, /* tp_clear */ |
| 0, /* tp_richcompare */ |
| 0, /* tp_weaklistoffset */ |
| PyObject_SelfIter, /* tp_iter */ |
| (iternextfunc)unicodeiter_next, /* tp_iternext */ |
| unicodeiter_methods, /* tp_methods */ |
| 0, |
| }; |
| |
| static PyObject * |
| unicode_iter(PyObject *seq) |
| { |
| unicodeiterobject *it; |
| |
| if (!PyUnicode_Check(seq)) { |
| PyErr_BadInternalCall(); |
| return NULL; |
| } |
| it = PyObject_GC_New(unicodeiterobject, &PyUnicodeIter_Type); |
| if (it == NULL) |
| return NULL; |
| it->it_index = 0; |
| Py_INCREF(seq); |
| it->it_seq = (PyUnicodeObject *)seq; |
| _PyObject_GC_TRACK(it); |
| return (PyObject *)it; |
| } |
| |
| size_t |
| Py_UNICODE_strlen(const Py_UNICODE *u) |
| { |
| int res = 0; |
| while(*u++) |
| res++; |
| return res; |
| } |
| |
| Py_UNICODE* |
| Py_UNICODE_strcpy(Py_UNICODE *s1, const Py_UNICODE *s2) |
| { |
| Py_UNICODE *u = s1; |
| while ((*u++ = *s2++)); |
| return s1; |
| } |
| |
| Py_UNICODE* |
| Py_UNICODE_strncpy(Py_UNICODE *s1, const Py_UNICODE *s2, size_t n) |
| { |
| Py_UNICODE *u = s1; |
| while ((*u++ = *s2++)) |
| if (n-- == 0) |
| break; |
| return s1; |
| } |
| |
| Py_UNICODE* |
| Py_UNICODE_strcat(Py_UNICODE *s1, const Py_UNICODE *s2) |
| { |
| Py_UNICODE *u1 = s1; |
| u1 += Py_UNICODE_strlen(u1); |
| Py_UNICODE_strcpy(u1, s2); |
| return s1; |
| } |
| |
| int |
| Py_UNICODE_strcmp(const Py_UNICODE *s1, const Py_UNICODE *s2) |
| { |
| while (*s1 && *s2 && *s1 == *s2) |
| s1++, s2++; |
| if (*s1 && *s2) |
| return (*s1 < *s2) ? -1 : +1; |
| if (*s1) |
| return 1; |
| if (*s2) |
| return -1; |
| return 0; |
| } |
| |
| int |
| Py_UNICODE_strncmp(const Py_UNICODE *s1, const Py_UNICODE *s2, size_t n) |
| { |
| register Py_UNICODE u1, u2; |
| for (; n != 0; n--) { |
| u1 = *s1; |
| u2 = *s2; |
| if (u1 != u2) |
| return (u1 < u2) ? -1 : +1; |
| if (u1 == '\0') |
| return 0; |
| s1++; |
| s2++; |
| } |
| return 0; |
| } |
| |
| Py_UNICODE* |
| Py_UNICODE_strchr(const Py_UNICODE *s, Py_UNICODE c) |
| { |
| const Py_UNICODE *p; |
| for (p = s; *p; p++) |
| if (*p == c) |
| return (Py_UNICODE*)p; |
| return NULL; |
| } |
| |
| Py_UNICODE* |
| Py_UNICODE_strrchr(const Py_UNICODE *s, Py_UNICODE c) |
| { |
| const Py_UNICODE *p; |
| p = s + Py_UNICODE_strlen(s); |
| while (p != s) { |
| p--; |
| if (*p == c) |
| return (Py_UNICODE*)p; |
| } |
| return NULL; |
| } |
| |
| Py_UNICODE* |
| PyUnicode_AsUnicodeCopy(PyObject *object) |
| { |
| PyUnicodeObject *unicode = (PyUnicodeObject *)object; |
| Py_UNICODE *copy; |
| Py_ssize_t size; |
| |
| /* Ensure we won't overflow the size. */ |
| if (PyUnicode_GET_SIZE(unicode) > ((PY_SSIZE_T_MAX / sizeof(Py_UNICODE)) - 1)) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| size = PyUnicode_GET_SIZE(unicode) + 1; /* copy the nul character */ |
| size *= sizeof(Py_UNICODE); |
| copy = PyMem_Malloc(size); |
| if (copy == NULL) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| memcpy(copy, PyUnicode_AS_UNICODE(unicode), size); |
| return copy; |
| } |
| |
| /* A _string module, to export formatter_parser and formatter_field_name_split |
| to the string.Formatter class implemented in Python. */ |
| |
| static PyMethodDef _string_methods[] = { |
| {"formatter_field_name_split", (PyCFunction) formatter_field_name_split, |
| METH_O, PyDoc_STR("split the argument as a field name")}, |
| {"formatter_parser", (PyCFunction) formatter_parser, |
| METH_O, PyDoc_STR("parse the argument as a format string")}, |
| {NULL, NULL} |
| }; |
| |
| static struct PyModuleDef _string_module = { |
| PyModuleDef_HEAD_INIT, |
| "_string", |
| PyDoc_STR("string helper module"), |
| 0, |
| _string_methods, |
| NULL, |
| NULL, |
| NULL, |
| NULL |
| }; |
| |
| PyMODINIT_FUNC |
| PyInit__string(void) |
| { |
| return PyModule_Create(&_string_module); |
| } |
| |
| |
| #ifdef __cplusplus |
| } |
| #endif |