experimental UCS-4 support: added USE_UCS4_STORAGE define to
unicodeobject.h, which forces sizeof(Py_UNICODE) == sizeof(Py_UCS4).
(this may be good enough for platforms that doesn't have a 16-bit
type. the UTF-16 codecs don't work, though)
diff --git a/Include/unicodeobject.h b/Include/unicodeobject.h
index f91a5a0..e330fd1 100644
--- a/Include/unicodeobject.h
+++ b/Include/unicodeobject.h
@@ -58,6 +58,19 @@
/* --- Internal Unicode Format -------------------------------------------- */
+/* experimental UCS-4 support. enable at your own risk! */
+#undef USE_UCS4_STORAGE
+
+/*
+ * Use this typedef when you need to represent a UTF-16 surrogate pair
+ * as single unsigned integer.
+ */
+#if SIZEOF_INT >= 4
+typedef unsigned int Py_UCS4;
+#elif SIZEOF_LONG >= 4
+typedef unsigned long Py_UCS4;
+#endif
+
/* Set these flags if the platform has "wchar.h", "wctype.h" and the
wchar_t type is a 16-bit unsigned type */
/* #define HAVE_WCHAR_H */
@@ -66,8 +79,8 @@
/* Defaults for various platforms */
#ifndef HAVE_USABLE_WCHAR_T
-/* Windows has a usable wchar_t type */
-# if defined(MS_WIN32)
+/* Windows has a usable wchar_t type (unless we're using UCS-4) */
+# if defined(MS_WIN32) && !defined(USE_UCS4_STORAGE)
# define HAVE_USABLE_WCHAR_T
# endif
@@ -105,19 +118,13 @@
If a short is not 16 bits on your platform, you have to fix the
typedef below, or the module initialization code will complain. */
+#ifdef USE_UCS4_STORAGE
+typedef Py_UCS4 Py_UNICODE;
+#else
typedef unsigned short Py_UNICODE;
-
#endif
-/*
- * Use this typedef when you need to represent a UTF-16 surrogate pair
- * as single unsigned integer.
- */
-#if SIZEOF_INT >= 4
-typedef unsigned int Py_UCS4;
-#elif SIZEOF_LONG >= 4
-typedef unsigned long Py_UCS4;
-#endif
+#endif
/* --- Internal Unicode Operations ---------------------------------------- */