| #include "Python.h" |
| #include "structmember.h" |
| |
| PyDoc_STRVAR(pickle_module_doc, |
| "Optimized C implementation for the Python pickle module."); |
| |
| /*[clinic input] |
| module _pickle |
| class _pickle.Pickler "PicklerObject *" "&Pickler_Type" |
| class _pickle.PicklerMemoProxy "PicklerMemoProxyObject *" "&PicklerMemoProxyType" |
| class _pickle.Unpickler "UnpicklerObject *" "&Unpickler_Type" |
| class _pickle.UnpicklerMemoProxy "UnpicklerMemoProxyObject *" "&UnpicklerMemoProxyType" |
| [clinic start generated code]*/ |
| /*[clinic end generated code: output=da39a3ee5e6b4b0d input=4b3e113468a58e6c]*/ |
| |
| /* Bump this when new opcodes are added to the pickle protocol. */ |
| enum { |
| HIGHEST_PROTOCOL = 4, |
| DEFAULT_PROTOCOL = 3 |
| }; |
| |
| /* Pickle opcodes. These must be kept updated with pickle.py. |
| Extensive docs are in pickletools.py. */ |
| enum opcode { |
| MARK = '(', |
| STOP = '.', |
| POP = '0', |
| POP_MARK = '1', |
| DUP = '2', |
| FLOAT = 'F', |
| INT = 'I', |
| BININT = 'J', |
| BININT1 = 'K', |
| LONG = 'L', |
| BININT2 = 'M', |
| NONE = 'N', |
| PERSID = 'P', |
| BINPERSID = 'Q', |
| REDUCE = 'R', |
| STRING = 'S', |
| BINSTRING = 'T', |
| SHORT_BINSTRING = 'U', |
| UNICODE = 'V', |
| BINUNICODE = 'X', |
| APPEND = 'a', |
| BUILD = 'b', |
| GLOBAL = 'c', |
| DICT = 'd', |
| EMPTY_DICT = '}', |
| APPENDS = 'e', |
| GET = 'g', |
| BINGET = 'h', |
| INST = 'i', |
| LONG_BINGET = 'j', |
| LIST = 'l', |
| EMPTY_LIST = ']', |
| OBJ = 'o', |
| PUT = 'p', |
| BINPUT = 'q', |
| LONG_BINPUT = 'r', |
| SETITEM = 's', |
| TUPLE = 't', |
| EMPTY_TUPLE = ')', |
| SETITEMS = 'u', |
| BINFLOAT = 'G', |
| |
| /* Protocol 2. */ |
| PROTO = '\x80', |
| NEWOBJ = '\x81', |
| EXT1 = '\x82', |
| EXT2 = '\x83', |
| EXT4 = '\x84', |
| TUPLE1 = '\x85', |
| TUPLE2 = '\x86', |
| TUPLE3 = '\x87', |
| NEWTRUE = '\x88', |
| NEWFALSE = '\x89', |
| LONG1 = '\x8a', |
| LONG4 = '\x8b', |
| |
| /* Protocol 3 (Python 3.x) */ |
| BINBYTES = 'B', |
| SHORT_BINBYTES = 'C', |
| |
| /* Protocol 4 */ |
| SHORT_BINUNICODE = '\x8c', |
| BINUNICODE8 = '\x8d', |
| BINBYTES8 = '\x8e', |
| EMPTY_SET = '\x8f', |
| ADDITEMS = '\x90', |
| FROZENSET = '\x91', |
| NEWOBJ_EX = '\x92', |
| STACK_GLOBAL = '\x93', |
| MEMOIZE = '\x94', |
| FRAME = '\x95' |
| }; |
| |
| enum { |
| /* Keep in synch with pickle.Pickler._BATCHSIZE. This is how many elements |
| batch_list/dict() pumps out before doing APPENDS/SETITEMS. Nothing will |
| break if this gets out of synch with pickle.py, but it's unclear that would |
| help anything either. */ |
| BATCHSIZE = 1000, |
| |
| /* Nesting limit until Pickler, when running in "fast mode", starts |
| checking for self-referential data-structures. */ |
| FAST_NESTING_LIMIT = 50, |
| |
| /* Initial size of the write buffer of Pickler. */ |
| WRITE_BUF_SIZE = 4096, |
| |
| /* Prefetch size when unpickling (disabled on unpeekable streams) */ |
| PREFETCH = 8192 * 16, |
| |
| FRAME_SIZE_TARGET = 64 * 1024, |
| |
| FRAME_HEADER_SIZE = 9 |
| }; |
| |
| /*************************************************************************/ |
| |
| /* State of the pickle module, per PEP 3121. */ |
| typedef struct { |
| /* Exception classes for pickle. */ |
| PyObject *PickleError; |
| PyObject *PicklingError; |
| PyObject *UnpicklingError; |
| |
| /* copyreg.dispatch_table, {type_object: pickling_function} */ |
| PyObject *dispatch_table; |
| |
| /* For the extension opcodes EXT1, EXT2 and EXT4. */ |
| |
| /* copyreg._extension_registry, {(module_name, function_name): code} */ |
| PyObject *extension_registry; |
| /* copyreg._extension_cache, {code: object} */ |
| PyObject *extension_cache; |
| /* copyreg._inverted_registry, {code: (module_name, function_name)} */ |
| PyObject *inverted_registry; |
| |
| /* Import mappings for compatibility with Python 2.x */ |
| |
| /* _compat_pickle.NAME_MAPPING, |
| {(oldmodule, oldname): (newmodule, newname)} */ |
| PyObject *name_mapping_2to3; |
| /* _compat_pickle.IMPORT_MAPPING, {oldmodule: newmodule} */ |
| PyObject *import_mapping_2to3; |
| /* Same, but with REVERSE_NAME_MAPPING / REVERSE_IMPORT_MAPPING */ |
| PyObject *name_mapping_3to2; |
| PyObject *import_mapping_3to2; |
| |
| /* codecs.encode, used for saving bytes in older protocols */ |
| PyObject *codecs_encode; |
| /* builtins.getattr, used for saving nested names with protocol < 4 */ |
| PyObject *getattr; |
| } PickleState; |
| |
| /* Forward declaration of the _pickle module definition. */ |
| static struct PyModuleDef _picklemodule; |
| |
| /* Given a module object, get its per-module state. */ |
| static PickleState * |
| _Pickle_GetState(PyObject *module) |
| { |
| return (PickleState *)PyModule_GetState(module); |
| } |
| |
| /* Find the module instance imported in the currently running sub-interpreter |
| and get its state. */ |
| static PickleState * |
| _Pickle_GetGlobalState(void) |
| { |
| return _Pickle_GetState(PyState_FindModule(&_picklemodule)); |
| } |
| |
| /* Clear the given pickle module state. */ |
| static void |
| _Pickle_ClearState(PickleState *st) |
| { |
| Py_CLEAR(st->PickleError); |
| Py_CLEAR(st->PicklingError); |
| Py_CLEAR(st->UnpicklingError); |
| Py_CLEAR(st->dispatch_table); |
| Py_CLEAR(st->extension_registry); |
| Py_CLEAR(st->extension_cache); |
| Py_CLEAR(st->inverted_registry); |
| Py_CLEAR(st->name_mapping_2to3); |
| Py_CLEAR(st->import_mapping_2to3); |
| Py_CLEAR(st->name_mapping_3to2); |
| Py_CLEAR(st->import_mapping_3to2); |
| Py_CLEAR(st->codecs_encode); |
| Py_CLEAR(st->getattr); |
| } |
| |
| /* Initialize the given pickle module state. */ |
| static int |
| _Pickle_InitState(PickleState *st) |
| { |
| PyObject *builtins; |
| PyObject *copyreg = NULL; |
| PyObject *compat_pickle = NULL; |
| PyObject *codecs = NULL; |
| |
| builtins = PyEval_GetBuiltins(); |
| if (builtins == NULL) |
| goto error; |
| st->getattr = PyDict_GetItemString(builtins, "getattr"); |
| if (st->getattr == NULL) |
| goto error; |
| Py_INCREF(st->getattr); |
| |
| copyreg = PyImport_ImportModule("copyreg"); |
| if (!copyreg) |
| goto error; |
| st->dispatch_table = PyObject_GetAttrString(copyreg, "dispatch_table"); |
| if (!st->dispatch_table) |
| goto error; |
| if (!PyDict_CheckExact(st->dispatch_table)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "copyreg.dispatch_table should be a dict, not %.200s", |
| Py_TYPE(st->dispatch_table)->tp_name); |
| goto error; |
| } |
| st->extension_registry = \ |
| PyObject_GetAttrString(copyreg, "_extension_registry"); |
| if (!st->extension_registry) |
| goto error; |
| if (!PyDict_CheckExact(st->extension_registry)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "copyreg._extension_registry should be a dict, " |
| "not %.200s", Py_TYPE(st->extension_registry)->tp_name); |
| goto error; |
| } |
| st->inverted_registry = \ |
| PyObject_GetAttrString(copyreg, "_inverted_registry"); |
| if (!st->inverted_registry) |
| goto error; |
| if (!PyDict_CheckExact(st->inverted_registry)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "copyreg._inverted_registry should be a dict, " |
| "not %.200s", Py_TYPE(st->inverted_registry)->tp_name); |
| goto error; |
| } |
| st->extension_cache = PyObject_GetAttrString(copyreg, "_extension_cache"); |
| if (!st->extension_cache) |
| goto error; |
| if (!PyDict_CheckExact(st->extension_cache)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "copyreg._extension_cache should be a dict, " |
| "not %.200s", Py_TYPE(st->extension_cache)->tp_name); |
| goto error; |
| } |
| Py_CLEAR(copyreg); |
| |
| /* Load the 2.x -> 3.x stdlib module mapping tables */ |
| compat_pickle = PyImport_ImportModule("_compat_pickle"); |
| if (!compat_pickle) |
| goto error; |
| st->name_mapping_2to3 = \ |
| PyObject_GetAttrString(compat_pickle, "NAME_MAPPING"); |
| if (!st->name_mapping_2to3) |
| goto error; |
| if (!PyDict_CheckExact(st->name_mapping_2to3)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.NAME_MAPPING should be a dict, not %.200s", |
| Py_TYPE(st->name_mapping_2to3)->tp_name); |
| goto error; |
| } |
| st->import_mapping_2to3 = \ |
| PyObject_GetAttrString(compat_pickle, "IMPORT_MAPPING"); |
| if (!st->import_mapping_2to3) |
| goto error; |
| if (!PyDict_CheckExact(st->import_mapping_2to3)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.IMPORT_MAPPING should be a dict, " |
| "not %.200s", Py_TYPE(st->import_mapping_2to3)->tp_name); |
| goto error; |
| } |
| /* ... and the 3.x -> 2.x mapping tables */ |
| st->name_mapping_3to2 = \ |
| PyObject_GetAttrString(compat_pickle, "REVERSE_NAME_MAPPING"); |
| if (!st->name_mapping_3to2) |
| goto error; |
| if (!PyDict_CheckExact(st->name_mapping_3to2)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.REVERSE_NAME_MAPPING should be a dict, " |
| "not %.200s", Py_TYPE(st->name_mapping_3to2)->tp_name); |
| goto error; |
| } |
| st->import_mapping_3to2 = \ |
| PyObject_GetAttrString(compat_pickle, "REVERSE_IMPORT_MAPPING"); |
| if (!st->import_mapping_3to2) |
| goto error; |
| if (!PyDict_CheckExact(st->import_mapping_3to2)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.REVERSE_IMPORT_MAPPING should be a dict, " |
| "not %.200s", Py_TYPE(st->import_mapping_3to2)->tp_name); |
| goto error; |
| } |
| Py_CLEAR(compat_pickle); |
| |
| codecs = PyImport_ImportModule("codecs"); |
| if (codecs == NULL) |
| goto error; |
| st->codecs_encode = PyObject_GetAttrString(codecs, "encode"); |
| if (st->codecs_encode == NULL) { |
| goto error; |
| } |
| if (!PyCallable_Check(st->codecs_encode)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "codecs.encode should be a callable, not %.200s", |
| Py_TYPE(st->codecs_encode)->tp_name); |
| goto error; |
| } |
| Py_CLEAR(codecs); |
| |
| return 0; |
| |
| error: |
| Py_CLEAR(copyreg); |
| Py_CLEAR(compat_pickle); |
| Py_CLEAR(codecs); |
| _Pickle_ClearState(st); |
| return -1; |
| } |
| |
| /* Helper for calling a function with a single argument quickly. |
| |
| This function steals the reference of the given argument. */ |
| static PyObject * |
| _Pickle_FastCall(PyObject *func, PyObject *obj) |
| { |
| PyObject *result; |
| PyObject *arg_tuple = PyTuple_New(1); |
| |
| /* Note: this function used to reuse the argument tuple. This used to give |
| a slight performance boost with older pickle implementations where many |
| unbuffered reads occurred (thus needing many function calls). |
| |
| However, this optimization was removed because it was too complicated |
| to get right. It abused the C API for tuples to mutate them which led |
| to subtle reference counting and concurrency bugs. Furthermore, the |
| introduction of protocol 4 and the prefetching optimization via peek() |
| significantly reduced the number of function calls we do. Thus, the |
| benefits became marginal at best. */ |
| |
| if (arg_tuple == NULL) { |
| Py_DECREF(obj); |
| return NULL; |
| } |
| PyTuple_SET_ITEM(arg_tuple, 0, obj); |
| result = PyObject_Call(func, arg_tuple, NULL); |
| Py_CLEAR(arg_tuple); |
| return result; |
| } |
| |
| /*************************************************************************/ |
| |
| static int |
| stack_underflow(void) |
| { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, "unpickling stack underflow"); |
| return -1; |
| } |
| |
| /* Internal data type used as the unpickling stack. */ |
| typedef struct { |
| PyObject_VAR_HEAD |
| PyObject **data; |
| Py_ssize_t allocated; /* number of slots in data allocated */ |
| } Pdata; |
| |
| static void |
| Pdata_dealloc(Pdata *self) |
| { |
| Py_ssize_t i = Py_SIZE(self); |
| while (--i >= 0) { |
| Py_DECREF(self->data[i]); |
| } |
| PyMem_FREE(self->data); |
| PyObject_Del(self); |
| } |
| |
| static PyTypeObject Pdata_Type = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "_pickle.Pdata", /*tp_name*/ |
| sizeof(Pdata), /*tp_basicsize*/ |
| sizeof(PyObject *), /*tp_itemsize*/ |
| (destructor)Pdata_dealloc, /*tp_dealloc*/ |
| }; |
| |
| static PyObject * |
| Pdata_New(void) |
| { |
| Pdata *self; |
| |
| if (!(self = PyObject_New(Pdata, &Pdata_Type))) |
| return NULL; |
| Py_SIZE(self) = 0; |
| self->allocated = 8; |
| self->data = PyMem_MALLOC(self->allocated * sizeof(PyObject *)); |
| if (self->data) |
| return (PyObject *)self; |
| Py_DECREF(self); |
| return PyErr_NoMemory(); |
| } |
| |
| |
| /* Retain only the initial clearto items. If clearto >= the current |
| * number of items, this is a (non-erroneous) NOP. |
| */ |
| static int |
| Pdata_clear(Pdata *self, Py_ssize_t clearto) |
| { |
| Py_ssize_t i = Py_SIZE(self); |
| |
| if (clearto < 0) |
| return stack_underflow(); |
| if (clearto >= i) |
| return 0; |
| |
| while (--i >= clearto) { |
| Py_CLEAR(self->data[i]); |
| } |
| Py_SIZE(self) = clearto; |
| return 0; |
| } |
| |
| static int |
| Pdata_grow(Pdata *self) |
| { |
| PyObject **data = self->data; |
| size_t allocated = (size_t)self->allocated; |
| size_t new_allocated; |
| |
| new_allocated = (allocated >> 3) + 6; |
| /* check for integer overflow */ |
| if (new_allocated > (size_t)PY_SSIZE_T_MAX - allocated) |
| goto nomemory; |
| new_allocated += allocated; |
| PyMem_RESIZE(data, PyObject *, new_allocated); |
| if (data == NULL) |
| goto nomemory; |
| |
| self->data = data; |
| self->allocated = (Py_ssize_t)new_allocated; |
| return 0; |
| |
| nomemory: |
| PyErr_NoMemory(); |
| return -1; |
| } |
| |
| /* D is a Pdata*. Pop the topmost element and store it into V, which |
| * must be an lvalue holding PyObject*. On stack underflow, UnpicklingError |
| * is raised and V is set to NULL. |
| */ |
| static PyObject * |
| Pdata_pop(Pdata *self) |
| { |
| if (Py_SIZE(self) == 0) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, "bad pickle data"); |
| return NULL; |
| } |
| return self->data[--Py_SIZE(self)]; |
| } |
| #define PDATA_POP(D, V) do { (V) = Pdata_pop((D)); } while (0) |
| |
| static int |
| Pdata_push(Pdata *self, PyObject *obj) |
| { |
| if (Py_SIZE(self) == self->allocated && Pdata_grow(self) < 0) { |
| return -1; |
| } |
| self->data[Py_SIZE(self)++] = obj; |
| return 0; |
| } |
| |
| /* Push an object on stack, transferring its ownership to the stack. */ |
| #define PDATA_PUSH(D, O, ER) do { \ |
| if (Pdata_push((D), (O)) < 0) return (ER); } while(0) |
| |
| /* Push an object on stack, adding a new reference to the object. */ |
| #define PDATA_APPEND(D, O, ER) do { \ |
| Py_INCREF((O)); \ |
| if (Pdata_push((D), (O)) < 0) return (ER); } while(0) |
| |
| static PyObject * |
| Pdata_poptuple(Pdata *self, Py_ssize_t start) |
| { |
| PyObject *tuple; |
| Py_ssize_t len, i, j; |
| |
| len = Py_SIZE(self) - start; |
| tuple = PyTuple_New(len); |
| if (tuple == NULL) |
| return NULL; |
| for (i = start, j = 0; j < len; i++, j++) |
| PyTuple_SET_ITEM(tuple, j, self->data[i]); |
| |
| Py_SIZE(self) = start; |
| return tuple; |
| } |
| |
| static PyObject * |
| Pdata_poplist(Pdata *self, Py_ssize_t start) |
| { |
| PyObject *list; |
| Py_ssize_t len, i, j; |
| |
| len = Py_SIZE(self) - start; |
| list = PyList_New(len); |
| if (list == NULL) |
| return NULL; |
| for (i = start, j = 0; j < len; i++, j++) |
| PyList_SET_ITEM(list, j, self->data[i]); |
| |
| Py_SIZE(self) = start; |
| return list; |
| } |
| |
| typedef struct { |
| PyObject *me_key; |
| Py_ssize_t me_value; |
| } PyMemoEntry; |
| |
| typedef struct { |
| Py_ssize_t mt_mask; |
| Py_ssize_t mt_used; |
| Py_ssize_t mt_allocated; |
| PyMemoEntry *mt_table; |
| } PyMemoTable; |
| |
| typedef struct PicklerObject { |
| PyObject_HEAD |
| PyMemoTable *memo; /* Memo table, keep track of the seen |
| objects to support self-referential objects |
| pickling. */ |
| PyObject *pers_func; /* persistent_id() method, can be NULL */ |
| PyObject *dispatch_table; /* private dispatch_table, can be NULL */ |
| |
| PyObject *write; /* write() method of the output stream. */ |
| PyObject *output_buffer; /* Write into a local bytearray buffer before |
| flushing to the stream. */ |
| Py_ssize_t output_len; /* Length of output_buffer. */ |
| Py_ssize_t max_output_len; /* Allocation size of output_buffer. */ |
| int proto; /* Pickle protocol number, >= 0 */ |
| int bin; /* Boolean, true if proto > 0 */ |
| int framing; /* True when framing is enabled, proto >= 4 */ |
| Py_ssize_t frame_start; /* Position in output_buffer where the |
| where the current frame begins. -1 if there |
| is no frame currently open. */ |
| |
| Py_ssize_t buf_size; /* Size of the current buffered pickle data */ |
| int fast; /* Enable fast mode if set to a true value. |
| The fast mode disable the usage of memo, |
| therefore speeding the pickling process by |
| not generating superfluous PUT opcodes. It |
| should not be used if with self-referential |
| objects. */ |
| int fast_nesting; |
| int fix_imports; /* Indicate whether Pickler should fix |
| the name of globals for Python 2.x. */ |
| PyObject *fast_memo; |
| } PicklerObject; |
| |
| typedef struct UnpicklerObject { |
| PyObject_HEAD |
| Pdata *stack; /* Pickle data stack, store unpickled objects. */ |
| |
| /* The unpickler memo is just an array of PyObject *s. Using a dict |
| is unnecessary, since the keys are contiguous ints. */ |
| PyObject **memo; |
| Py_ssize_t memo_size; /* Capacity of the memo array */ |
| Py_ssize_t memo_len; /* Number of objects in the memo */ |
| |
| PyObject *pers_func; /* persistent_load() method, can be NULL. */ |
| |
| Py_buffer buffer; |
| char *input_buffer; |
| char *input_line; |
| Py_ssize_t input_len; |
| Py_ssize_t next_read_idx; |
| Py_ssize_t prefetched_idx; /* index of first prefetched byte */ |
| |
| PyObject *read; /* read() method of the input stream. */ |
| PyObject *readline; /* readline() method of the input stream. */ |
| PyObject *peek; /* peek() method of the input stream, or NULL */ |
| |
| char *encoding; /* Name of the encoding to be used for |
| decoding strings pickled using Python |
| 2.x. The default value is "ASCII" */ |
| char *errors; /* Name of errors handling scheme to used when |
| decoding strings. The default value is |
| "strict". */ |
| Py_ssize_t *marks; /* Mark stack, used for unpickling container |
| objects. */ |
| Py_ssize_t num_marks; /* Number of marks in the mark stack. */ |
| Py_ssize_t marks_size; /* Current allocated size of the mark stack. */ |
| int proto; /* Protocol of the pickle loaded. */ |
| int fix_imports; /* Indicate whether Unpickler should fix |
| the name of globals pickled by Python 2.x. */ |
| } UnpicklerObject; |
| |
| typedef struct { |
| PyObject_HEAD |
| PicklerObject *pickler; /* Pickler whose memo table we're proxying. */ |
| } PicklerMemoProxyObject; |
| |
| typedef struct { |
| PyObject_HEAD |
| UnpicklerObject *unpickler; |
| } UnpicklerMemoProxyObject; |
| |
| /* Forward declarations */ |
| static int save(PicklerObject *, PyObject *, int); |
| static int save_reduce(PicklerObject *, PyObject *, PyObject *); |
| static PyTypeObject Pickler_Type; |
| static PyTypeObject Unpickler_Type; |
| |
| #include "clinic/_pickle.c.h" |
| |
| /************************************************************************* |
| A custom hashtable mapping void* to Python ints. This is used by the pickler |
| for memoization. Using a custom hashtable rather than PyDict allows us to skip |
| a bunch of unnecessary object creation. This makes a huge performance |
| difference. */ |
| |
| #define MT_MINSIZE 8 |
| #define PERTURB_SHIFT 5 |
| |
| |
| static PyMemoTable * |
| PyMemoTable_New(void) |
| { |
| PyMemoTable *memo = PyMem_MALLOC(sizeof(PyMemoTable)); |
| if (memo == NULL) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| |
| memo->mt_used = 0; |
| memo->mt_allocated = MT_MINSIZE; |
| memo->mt_mask = MT_MINSIZE - 1; |
| memo->mt_table = PyMem_MALLOC(MT_MINSIZE * sizeof(PyMemoEntry)); |
| if (memo->mt_table == NULL) { |
| PyMem_FREE(memo); |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| memset(memo->mt_table, 0, MT_MINSIZE * sizeof(PyMemoEntry)); |
| |
| return memo; |
| } |
| |
| static PyMemoTable * |
| PyMemoTable_Copy(PyMemoTable *self) |
| { |
| Py_ssize_t i; |
| PyMemoTable *new = PyMemoTable_New(); |
| if (new == NULL) |
| return NULL; |
| |
| new->mt_used = self->mt_used; |
| new->mt_allocated = self->mt_allocated; |
| new->mt_mask = self->mt_mask; |
| /* The table we get from _New() is probably smaller than we wanted. |
| Free it and allocate one that's the right size. */ |
| PyMem_FREE(new->mt_table); |
| new->mt_table = PyMem_NEW(PyMemoEntry, self->mt_allocated); |
| if (new->mt_table == NULL) { |
| PyMem_FREE(new); |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| for (i = 0; i < self->mt_allocated; i++) { |
| Py_XINCREF(self->mt_table[i].me_key); |
| } |
| memcpy(new->mt_table, self->mt_table, |
| sizeof(PyMemoEntry) * self->mt_allocated); |
| |
| return new; |
| } |
| |
| static Py_ssize_t |
| PyMemoTable_Size(PyMemoTable *self) |
| { |
| return self->mt_used; |
| } |
| |
| static int |
| PyMemoTable_Clear(PyMemoTable *self) |
| { |
| Py_ssize_t i = self->mt_allocated; |
| |
| while (--i >= 0) { |
| Py_XDECREF(self->mt_table[i].me_key); |
| } |
| self->mt_used = 0; |
| memset(self->mt_table, 0, self->mt_allocated * sizeof(PyMemoEntry)); |
| return 0; |
| } |
| |
| static void |
| PyMemoTable_Del(PyMemoTable *self) |
| { |
| if (self == NULL) |
| return; |
| PyMemoTable_Clear(self); |
| |
| PyMem_FREE(self->mt_table); |
| PyMem_FREE(self); |
| } |
| |
| /* Since entries cannot be deleted from this hashtable, _PyMemoTable_Lookup() |
| can be considerably simpler than dictobject.c's lookdict(). */ |
| static PyMemoEntry * |
| _PyMemoTable_Lookup(PyMemoTable *self, PyObject *key) |
| { |
| size_t i; |
| size_t perturb; |
| size_t mask = (size_t)self->mt_mask; |
| PyMemoEntry *table = self->mt_table; |
| PyMemoEntry *entry; |
| Py_hash_t hash = (Py_hash_t)key >> 3; |
| |
| i = hash & mask; |
| entry = &table[i]; |
| if (entry->me_key == NULL || entry->me_key == key) |
| return entry; |
| |
| for (perturb = hash; ; perturb >>= PERTURB_SHIFT) { |
| i = (i << 2) + i + perturb + 1; |
| entry = &table[i & mask]; |
| if (entry->me_key == NULL || entry->me_key == key) |
| return entry; |
| } |
| assert(0); /* Never reached */ |
| return NULL; |
| } |
| |
| /* Returns -1 on failure, 0 on success. */ |
| static int |
| _PyMemoTable_ResizeTable(PyMemoTable *self, Py_ssize_t min_size) |
| { |
| PyMemoEntry *oldtable = NULL; |
| PyMemoEntry *oldentry, *newentry; |
| Py_ssize_t new_size = MT_MINSIZE; |
| Py_ssize_t to_process; |
| |
| assert(min_size > 0); |
| |
| /* Find the smallest valid table size >= min_size. */ |
| while (new_size < min_size && new_size > 0) |
| new_size <<= 1; |
| if (new_size <= 0) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| /* new_size needs to be a power of two. */ |
| assert((new_size & (new_size - 1)) == 0); |
| |
| /* Allocate new table. */ |
| oldtable = self->mt_table; |
| self->mt_table = PyMem_NEW(PyMemoEntry, new_size); |
| if (self->mt_table == NULL) { |
| self->mt_table = oldtable; |
| PyErr_NoMemory(); |
| return -1; |
| } |
| self->mt_allocated = new_size; |
| self->mt_mask = new_size - 1; |
| memset(self->mt_table, 0, sizeof(PyMemoEntry) * new_size); |
| |
| /* Copy entries from the old table. */ |
| to_process = self->mt_used; |
| for (oldentry = oldtable; to_process > 0; oldentry++) { |
| if (oldentry->me_key != NULL) { |
| to_process--; |
| /* newentry is a pointer to a chunk of the new |
| mt_table, so we're setting the key:value pair |
| in-place. */ |
| newentry = _PyMemoTable_Lookup(self, oldentry->me_key); |
| newentry->me_key = oldentry->me_key; |
| newentry->me_value = oldentry->me_value; |
| } |
| } |
| |
| /* Deallocate the old table. */ |
| PyMem_FREE(oldtable); |
| return 0; |
| } |
| |
| /* Returns NULL on failure, a pointer to the value otherwise. */ |
| static Py_ssize_t * |
| PyMemoTable_Get(PyMemoTable *self, PyObject *key) |
| { |
| PyMemoEntry *entry = _PyMemoTable_Lookup(self, key); |
| if (entry->me_key == NULL) |
| return NULL; |
| return &entry->me_value; |
| } |
| |
| /* Returns -1 on failure, 0 on success. */ |
| static int |
| PyMemoTable_Set(PyMemoTable *self, PyObject *key, Py_ssize_t value) |
| { |
| PyMemoEntry *entry; |
| |
| assert(key != NULL); |
| |
| entry = _PyMemoTable_Lookup(self, key); |
| if (entry->me_key != NULL) { |
| entry->me_value = value; |
| return 0; |
| } |
| Py_INCREF(key); |
| entry->me_key = key; |
| entry->me_value = value; |
| self->mt_used++; |
| |
| /* If we added a key, we can safely resize. Otherwise just return! |
| * If used >= 2/3 size, adjust size. Normally, this quaduples the size. |
| * |
| * Quadrupling the size improves average table sparseness |
| * (reducing collisions) at the cost of some memory. It also halves |
| * the number of expensive resize operations in a growing memo table. |
| * |
| * Very large memo tables (over 50K items) use doubling instead. |
| * This may help applications with severe memory constraints. |
| */ |
| if (!(self->mt_used * 3 >= (self->mt_mask + 1) * 2)) |
| return 0; |
| return _PyMemoTable_ResizeTable(self, |
| (self->mt_used > 50000 ? 2 : 4) * self->mt_used); |
| } |
| |
| #undef MT_MINSIZE |
| #undef PERTURB_SHIFT |
| |
| /*************************************************************************/ |
| |
| |
| static int |
| _Pickler_ClearBuffer(PicklerObject *self) |
| { |
| Py_XSETREF(self->output_buffer, |
| PyBytes_FromStringAndSize(NULL, self->max_output_len)); |
| if (self->output_buffer == NULL) |
| return -1; |
| self->output_len = 0; |
| self->frame_start = -1; |
| return 0; |
| } |
| |
| static void |
| _write_size64(char *out, size_t value) |
| { |
| size_t i; |
| |
| assert(sizeof(size_t) <= 8); |
| |
| for (i = 0; i < sizeof(size_t); i++) { |
| out[i] = (unsigned char)((value >> (8 * i)) & 0xff); |
| } |
| for (i = sizeof(size_t); i < 8; i++) { |
| out[i] = 0; |
| } |
| } |
| |
| static void |
| _Pickler_WriteFrameHeader(PicklerObject *self, char *qdata, size_t frame_len) |
| { |
| qdata[0] = FRAME; |
| _write_size64(qdata + 1, frame_len); |
| } |
| |
| static int |
| _Pickler_CommitFrame(PicklerObject *self) |
| { |
| size_t frame_len; |
| char *qdata; |
| |
| if (!self->framing || self->frame_start == -1) |
| return 0; |
| frame_len = self->output_len - self->frame_start - FRAME_HEADER_SIZE; |
| qdata = PyBytes_AS_STRING(self->output_buffer) + self->frame_start; |
| _Pickler_WriteFrameHeader(self, qdata, frame_len); |
| self->frame_start = -1; |
| return 0; |
| } |
| |
| static int |
| _Pickler_OpcodeBoundary(PicklerObject *self) |
| { |
| Py_ssize_t frame_len; |
| |
| if (!self->framing || self->frame_start == -1) |
| return 0; |
| frame_len = self->output_len - self->frame_start - FRAME_HEADER_SIZE; |
| if (frame_len >= FRAME_SIZE_TARGET) |
| return _Pickler_CommitFrame(self); |
| else |
| return 0; |
| } |
| |
| static PyObject * |
| _Pickler_GetString(PicklerObject *self) |
| { |
| PyObject *output_buffer = self->output_buffer; |
| |
| assert(self->output_buffer != NULL); |
| |
| if (_Pickler_CommitFrame(self)) |
| return NULL; |
| |
| self->output_buffer = NULL; |
| /* Resize down to exact size */ |
| if (_PyBytes_Resize(&output_buffer, self->output_len) < 0) |
| return NULL; |
| return output_buffer; |
| } |
| |
| static int |
| _Pickler_FlushToFile(PicklerObject *self) |
| { |
| PyObject *output, *result; |
| |
| assert(self->write != NULL); |
| |
| /* This will commit the frame first */ |
| output = _Pickler_GetString(self); |
| if (output == NULL) |
| return -1; |
| |
| result = _Pickle_FastCall(self->write, output); |
| Py_XDECREF(result); |
| return (result == NULL) ? -1 : 0; |
| } |
| |
| static Py_ssize_t |
| _Pickler_Write(PicklerObject *self, const char *s, Py_ssize_t data_len) |
| { |
| Py_ssize_t i, n, required; |
| char *buffer; |
| int need_new_frame; |
| |
| assert(s != NULL); |
| need_new_frame = (self->framing && self->frame_start == -1); |
| |
| if (need_new_frame) |
| n = data_len + FRAME_HEADER_SIZE; |
| else |
| n = data_len; |
| |
| required = self->output_len + n; |
| if (required > self->max_output_len) { |
| /* Make place in buffer for the pickle chunk */ |
| if (self->output_len >= PY_SSIZE_T_MAX / 2 - n) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| self->max_output_len = (self->output_len + n) / 2 * 3; |
| if (_PyBytes_Resize(&self->output_buffer, self->max_output_len) < 0) |
| return -1; |
| } |
| buffer = PyBytes_AS_STRING(self->output_buffer); |
| if (need_new_frame) { |
| /* Setup new frame */ |
| Py_ssize_t frame_start = self->output_len; |
| self->frame_start = frame_start; |
| for (i = 0; i < FRAME_HEADER_SIZE; i++) { |
| /* Write an invalid value, for debugging */ |
| buffer[frame_start + i] = 0xFE; |
| } |
| self->output_len += FRAME_HEADER_SIZE; |
| } |
| if (data_len < 8) { |
| /* This is faster than memcpy when the string is short. */ |
| for (i = 0; i < data_len; i++) { |
| buffer[self->output_len + i] = s[i]; |
| } |
| } |
| else { |
| memcpy(buffer + self->output_len, s, data_len); |
| } |
| self->output_len += data_len; |
| return data_len; |
| } |
| |
| static PicklerObject * |
| _Pickler_New(void) |
| { |
| PicklerObject *self; |
| |
| self = PyObject_GC_New(PicklerObject, &Pickler_Type); |
| if (self == NULL) |
| return NULL; |
| |
| self->pers_func = NULL; |
| self->dispatch_table = NULL; |
| self->write = NULL; |
| self->proto = 0; |
| self->bin = 0; |
| self->framing = 0; |
| self->frame_start = -1; |
| self->fast = 0; |
| self->fast_nesting = 0; |
| self->fix_imports = 0; |
| self->fast_memo = NULL; |
| self->max_output_len = WRITE_BUF_SIZE; |
| self->output_len = 0; |
| |
| self->memo = PyMemoTable_New(); |
| self->output_buffer = PyBytes_FromStringAndSize(NULL, |
| self->max_output_len); |
| |
| if (self->memo == NULL || self->output_buffer == NULL) { |
| Py_DECREF(self); |
| return NULL; |
| } |
| return self; |
| } |
| |
| static int |
| _Pickler_SetProtocol(PicklerObject *self, PyObject *protocol, int fix_imports) |
| { |
| long proto; |
| |
| if (protocol == NULL || protocol == Py_None) { |
| proto = DEFAULT_PROTOCOL; |
| } |
| else { |
| proto = PyLong_AsLong(protocol); |
| if (proto < 0) { |
| if (proto == -1 && PyErr_Occurred()) |
| return -1; |
| proto = HIGHEST_PROTOCOL; |
| } |
| else if (proto > HIGHEST_PROTOCOL) { |
| PyErr_Format(PyExc_ValueError, "pickle protocol must be <= %d", |
| HIGHEST_PROTOCOL); |
| return -1; |
| } |
| } |
| self->proto = (int)proto; |
| self->bin = proto > 0; |
| self->fix_imports = fix_imports && proto < 3; |
| return 0; |
| } |
| |
| /* Returns -1 (with an exception set) on failure, 0 on success. This may |
| be called once on a freshly created Pickler. */ |
| static int |
| _Pickler_SetOutputStream(PicklerObject *self, PyObject *file) |
| { |
| _Py_IDENTIFIER(write); |
| assert(file != NULL); |
| self->write = _PyObject_GetAttrId(file, &PyId_write); |
| if (self->write == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) |
| PyErr_SetString(PyExc_TypeError, |
| "file must have a 'write' attribute"); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* Returns the size of the input on success, -1 on failure. This takes its |
| own reference to `input`. */ |
| static Py_ssize_t |
| _Unpickler_SetStringInput(UnpicklerObject *self, PyObject *input) |
| { |
| if (self->buffer.buf != NULL) |
| PyBuffer_Release(&self->buffer); |
| if (PyObject_GetBuffer(input, &self->buffer, PyBUF_CONTIG_RO) < 0) |
| return -1; |
| self->input_buffer = self->buffer.buf; |
| self->input_len = self->buffer.len; |
| self->next_read_idx = 0; |
| self->prefetched_idx = self->input_len; |
| return self->input_len; |
| } |
| |
| static int |
| _Unpickler_SkipConsumed(UnpicklerObject *self) |
| { |
| Py_ssize_t consumed; |
| PyObject *r; |
| |
| consumed = self->next_read_idx - self->prefetched_idx; |
| if (consumed <= 0) |
| return 0; |
| |
| assert(self->peek); /* otherwise we did something wrong */ |
| /* This makes a useless copy... */ |
| r = PyObject_CallFunction(self->read, "n", consumed); |
| if (r == NULL) |
| return -1; |
| Py_DECREF(r); |
| |
| self->prefetched_idx = self->next_read_idx; |
| return 0; |
| } |
| |
| static const Py_ssize_t READ_WHOLE_LINE = -1; |
| |
| /* If reading from a file, we need to only pull the bytes we need, since there |
| may be multiple pickle objects arranged contiguously in the same input |
| buffer. |
| |
| If `n` is READ_WHOLE_LINE, read a whole line. Otherwise, read up to `n` |
| bytes from the input stream/buffer. |
| |
| Update the unpickler's input buffer with the newly-read data. Returns -1 on |
| failure; on success, returns the number of bytes read from the file. |
| |
| On success, self->input_len will be 0; this is intentional so that when |
| unpickling from a file, the "we've run out of data" code paths will trigger, |
| causing the Unpickler to go back to the file for more data. Use the returned |
| size to tell you how much data you can process. */ |
| static Py_ssize_t |
| _Unpickler_ReadFromFile(UnpicklerObject *self, Py_ssize_t n) |
| { |
| PyObject *data; |
| Py_ssize_t read_size; |
| |
| assert(self->read != NULL); |
| |
| if (_Unpickler_SkipConsumed(self) < 0) |
| return -1; |
| |
| if (n == READ_WHOLE_LINE) { |
| PyObject *empty_tuple = PyTuple_New(0); |
| data = PyObject_Call(self->readline, empty_tuple, NULL); |
| Py_DECREF(empty_tuple); |
| } |
| else { |
| PyObject *len; |
| /* Prefetch some data without advancing the file pointer, if possible */ |
| if (self->peek && n < PREFETCH) { |
| len = PyLong_FromSsize_t(PREFETCH); |
| if (len == NULL) |
| return -1; |
| data = _Pickle_FastCall(self->peek, len); |
| if (data == NULL) { |
| if (!PyErr_ExceptionMatches(PyExc_NotImplementedError)) |
| return -1; |
| /* peek() is probably not supported by the given file object */ |
| PyErr_Clear(); |
| Py_CLEAR(self->peek); |
| } |
| else { |
| read_size = _Unpickler_SetStringInput(self, data); |
| Py_DECREF(data); |
| self->prefetched_idx = 0; |
| if (n <= read_size) |
| return n; |
| } |
| } |
| len = PyLong_FromSsize_t(n); |
| if (len == NULL) |
| return -1; |
| data = _Pickle_FastCall(self->read, len); |
| } |
| if (data == NULL) |
| return -1; |
| |
| read_size = _Unpickler_SetStringInput(self, data); |
| Py_DECREF(data); |
| return read_size; |
| } |
| |
| /* Read `n` bytes from the unpickler's data source, storing the result in `*s`. |
| |
| This should be used for all data reads, rather than accessing the unpickler's |
| input buffer directly. This method deals correctly with reading from input |
| streams, which the input buffer doesn't deal with. |
| |
| Note that when reading from a file-like object, self->next_read_idx won't |
| be updated (it should remain at 0 for the entire unpickling process). You |
| should use this function's return value to know how many bytes you can |
| consume. |
| |
| Returns -1 (with an exception set) on failure. On success, return the |
| number of chars read. */ |
| static Py_ssize_t |
| _Unpickler_Read(UnpicklerObject *self, char **s, Py_ssize_t n) |
| { |
| Py_ssize_t num_read; |
| |
| *s = NULL; |
| if (self->next_read_idx > PY_SSIZE_T_MAX - n) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, |
| "read would overflow (invalid bytecode)"); |
| return -1; |
| } |
| if (self->next_read_idx + n <= self->input_len) { |
| *s = self->input_buffer + self->next_read_idx; |
| self->next_read_idx += n; |
| return n; |
| } |
| if (!self->read) { |
| PyErr_Format(PyExc_EOFError, "Ran out of input"); |
| return -1; |
| } |
| num_read = _Unpickler_ReadFromFile(self, n); |
| if (num_read < 0) |
| return -1; |
| if (num_read < n) { |
| PyErr_Format(PyExc_EOFError, "Ran out of input"); |
| return -1; |
| } |
| *s = self->input_buffer; |
| self->next_read_idx = n; |
| return n; |
| } |
| |
| static Py_ssize_t |
| _Unpickler_CopyLine(UnpicklerObject *self, char *line, Py_ssize_t len, |
| char **result) |
| { |
| char *input_line = PyMem_Realloc(self->input_line, len + 1); |
| if (input_line == NULL) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| |
| memcpy(input_line, line, len); |
| input_line[len] = '\0'; |
| self->input_line = input_line; |
| *result = self->input_line; |
| return len; |
| } |
| |
| /* Read a line from the input stream/buffer. If we run off the end of the input |
| before hitting \n, return the data we found. |
| |
| Returns the number of chars read, or -1 on failure. */ |
| static Py_ssize_t |
| _Unpickler_Readline(UnpicklerObject *self, char **result) |
| { |
| Py_ssize_t i, num_read; |
| |
| for (i = self->next_read_idx; i < self->input_len; i++) { |
| if (self->input_buffer[i] == '\n') { |
| char *line_start = self->input_buffer + self->next_read_idx; |
| num_read = i - self->next_read_idx + 1; |
| self->next_read_idx = i + 1; |
| return _Unpickler_CopyLine(self, line_start, num_read, result); |
| } |
| } |
| if (self->read) { |
| num_read = _Unpickler_ReadFromFile(self, READ_WHOLE_LINE); |
| if (num_read < 0) |
| return -1; |
| self->next_read_idx = num_read; |
| return _Unpickler_CopyLine(self, self->input_buffer, num_read, result); |
| } |
| |
| /* If we get here, we've run off the end of the input string. Return the |
| remaining string and let the caller figure it out. */ |
| *result = self->input_buffer + self->next_read_idx; |
| num_read = i - self->next_read_idx; |
| self->next_read_idx = i; |
| return num_read; |
| } |
| |
| /* Returns -1 (with an exception set) on failure, 0 on success. The memo array |
| will be modified in place. */ |
| static int |
| _Unpickler_ResizeMemoList(UnpicklerObject *self, Py_ssize_t new_size) |
| { |
| Py_ssize_t i; |
| |
| assert(new_size > self->memo_size); |
| |
| PyMem_RESIZE(self->memo, PyObject *, new_size); |
| if (self->memo == NULL) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| for (i = self->memo_size; i < new_size; i++) |
| self->memo[i] = NULL; |
| self->memo_size = new_size; |
| return 0; |
| } |
| |
| /* Returns NULL if idx is out of bounds. */ |
| static PyObject * |
| _Unpickler_MemoGet(UnpicklerObject *self, Py_ssize_t idx) |
| { |
| if (idx < 0 || idx >= self->memo_size) |
| return NULL; |
| |
| return self->memo[idx]; |
| } |
| |
| /* Returns -1 (with an exception set) on failure, 0 on success. |
| This takes its own reference to `value`. */ |
| static int |
| _Unpickler_MemoPut(UnpicklerObject *self, Py_ssize_t idx, PyObject *value) |
| { |
| PyObject *old_item; |
| |
| if (idx >= self->memo_size) { |
| if (_Unpickler_ResizeMemoList(self, idx * 2) < 0) |
| return -1; |
| assert(idx < self->memo_size); |
| } |
| Py_INCREF(value); |
| old_item = self->memo[idx]; |
| self->memo[idx] = value; |
| if (old_item != NULL) { |
| Py_DECREF(old_item); |
| } |
| else { |
| self->memo_len++; |
| } |
| return 0; |
| } |
| |
| static PyObject ** |
| _Unpickler_NewMemo(Py_ssize_t new_size) |
| { |
| PyObject **memo = PyMem_NEW(PyObject *, new_size); |
| if (memo == NULL) { |
| PyErr_NoMemory(); |
| return NULL; |
| } |
| memset(memo, 0, new_size * sizeof(PyObject *)); |
| return memo; |
| } |
| |
| /* Free the unpickler's memo, taking care to decref any items left in it. */ |
| static void |
| _Unpickler_MemoCleanup(UnpicklerObject *self) |
| { |
| Py_ssize_t i; |
| PyObject **memo = self->memo; |
| |
| if (self->memo == NULL) |
| return; |
| self->memo = NULL; |
| i = self->memo_size; |
| while (--i >= 0) { |
| Py_XDECREF(memo[i]); |
| } |
| PyMem_FREE(memo); |
| } |
| |
| static UnpicklerObject * |
| _Unpickler_New(void) |
| { |
| UnpicklerObject *self; |
| |
| self = PyObject_GC_New(UnpicklerObject, &Unpickler_Type); |
| if (self == NULL) |
| return NULL; |
| |
| self->pers_func = NULL; |
| self->input_buffer = NULL; |
| self->input_line = NULL; |
| self->input_len = 0; |
| self->next_read_idx = 0; |
| self->prefetched_idx = 0; |
| self->read = NULL; |
| self->readline = NULL; |
| self->peek = NULL; |
| self->encoding = NULL; |
| self->errors = NULL; |
| self->marks = NULL; |
| self->num_marks = 0; |
| self->marks_size = 0; |
| self->proto = 0; |
| self->fix_imports = 0; |
| memset(&self->buffer, 0, sizeof(Py_buffer)); |
| self->memo_size = 32; |
| self->memo_len = 0; |
| self->memo = _Unpickler_NewMemo(self->memo_size); |
| self->stack = (Pdata *)Pdata_New(); |
| |
| if (self->memo == NULL || self->stack == NULL) { |
| Py_DECREF(self); |
| return NULL; |
| } |
| |
| return self; |
| } |
| |
| /* Returns -1 (with an exception set) on failure, 0 on success. This may |
| be called once on a freshly created Pickler. */ |
| static int |
| _Unpickler_SetInputStream(UnpicklerObject *self, PyObject *file) |
| { |
| _Py_IDENTIFIER(peek); |
| _Py_IDENTIFIER(read); |
| _Py_IDENTIFIER(readline); |
| |
| self->peek = _PyObject_GetAttrId(file, &PyId_peek); |
| if (self->peek == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) |
| PyErr_Clear(); |
| else |
| return -1; |
| } |
| self->read = _PyObject_GetAttrId(file, &PyId_read); |
| self->readline = _PyObject_GetAttrId(file, &PyId_readline); |
| if (self->readline == NULL || self->read == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) |
| PyErr_SetString(PyExc_TypeError, |
| "file must have 'read' and 'readline' attributes"); |
| Py_CLEAR(self->read); |
| Py_CLEAR(self->readline); |
| Py_CLEAR(self->peek); |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* Returns -1 (with an exception set) on failure, 0 on success. This may |
| be called once on a freshly created Pickler. */ |
| static int |
| _Unpickler_SetInputEncoding(UnpicklerObject *self, |
| const char *encoding, |
| const char *errors) |
| { |
| if (encoding == NULL) |
| encoding = "ASCII"; |
| if (errors == NULL) |
| errors = "strict"; |
| |
| self->encoding = _PyMem_Strdup(encoding); |
| self->errors = _PyMem_Strdup(errors); |
| if (self->encoding == NULL || self->errors == NULL) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* Generate a GET opcode for an object stored in the memo. */ |
| static int |
| memo_get(PicklerObject *self, PyObject *key) |
| { |
| Py_ssize_t *value; |
| char pdata[30]; |
| Py_ssize_t len; |
| |
| value = PyMemoTable_Get(self->memo, key); |
| if (value == NULL) { |
| PyErr_SetObject(PyExc_KeyError, key); |
| return -1; |
| } |
| |
| if (!self->bin) { |
| pdata[0] = GET; |
| PyOS_snprintf(pdata + 1, sizeof(pdata) - 1, |
| "%" PY_FORMAT_SIZE_T "d\n", *value); |
| len = strlen(pdata); |
| } |
| else { |
| if (*value < 256) { |
| pdata[0] = BINGET; |
| pdata[1] = (unsigned char)(*value & 0xff); |
| len = 2; |
| } |
| else if ((size_t)*value <= 0xffffffffUL) { |
| pdata[0] = LONG_BINGET; |
| pdata[1] = (unsigned char)(*value & 0xff); |
| pdata[2] = (unsigned char)((*value >> 8) & 0xff); |
| pdata[3] = (unsigned char)((*value >> 16) & 0xff); |
| pdata[4] = (unsigned char)((*value >> 24) & 0xff); |
| len = 5; |
| } |
| else { /* unlikely */ |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->PicklingError, |
| "memo id too large for LONG_BINGET"); |
| return -1; |
| } |
| } |
| |
| if (_Pickler_Write(self, pdata, len) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| /* Store an object in the memo, assign it a new unique ID based on the number |
| of objects currently stored in the memo and generate a PUT opcode. */ |
| static int |
| memo_put(PicklerObject *self, PyObject *obj) |
| { |
| char pdata[30]; |
| Py_ssize_t len; |
| Py_ssize_t idx; |
| |
| const char memoize_op = MEMOIZE; |
| |
| if (self->fast) |
| return 0; |
| |
| idx = PyMemoTable_Size(self->memo); |
| if (PyMemoTable_Set(self->memo, obj, idx) < 0) |
| return -1; |
| |
| if (self->proto >= 4) { |
| if (_Pickler_Write(self, &memoize_op, 1) < 0) |
| return -1; |
| return 0; |
| } |
| else if (!self->bin) { |
| pdata[0] = PUT; |
| PyOS_snprintf(pdata + 1, sizeof(pdata) - 1, |
| "%" PY_FORMAT_SIZE_T "d\n", idx); |
| len = strlen(pdata); |
| } |
| else { |
| if (idx < 256) { |
| pdata[0] = BINPUT; |
| pdata[1] = (unsigned char)idx; |
| len = 2; |
| } |
| else if ((size_t)idx <= 0xffffffffUL) { |
| pdata[0] = LONG_BINPUT; |
| pdata[1] = (unsigned char)(idx & 0xff); |
| pdata[2] = (unsigned char)((idx >> 8) & 0xff); |
| pdata[3] = (unsigned char)((idx >> 16) & 0xff); |
| pdata[4] = (unsigned char)((idx >> 24) & 0xff); |
| len = 5; |
| } |
| else { /* unlikely */ |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->PicklingError, |
| "memo id too large for LONG_BINPUT"); |
| return -1; |
| } |
| } |
| if (_Pickler_Write(self, pdata, len) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| static PyObject * |
| get_dotted_path(PyObject *obj, PyObject *name) { |
| _Py_static_string(PyId_dot, "."); |
| _Py_static_string(PyId_locals, "<locals>"); |
| PyObject *dotted_path; |
| Py_ssize_t i, n; |
| |
| dotted_path = PyUnicode_Split(name, _PyUnicode_FromId(&PyId_dot), -1); |
| if (dotted_path == NULL) |
| return NULL; |
| n = PyList_GET_SIZE(dotted_path); |
| assert(n >= 1); |
| for (i = 0; i < n; i++) { |
| PyObject *subpath = PyList_GET_ITEM(dotted_path, i); |
| PyObject *result = PyUnicode_RichCompare( |
| subpath, _PyUnicode_FromId(&PyId_locals), Py_EQ); |
| int is_equal = (result == Py_True); |
| assert(PyBool_Check(result)); |
| Py_DECREF(result); |
| if (is_equal) { |
| if (obj == NULL) |
| PyErr_Format(PyExc_AttributeError, |
| "Can't pickle local object %R", name); |
| else |
| PyErr_Format(PyExc_AttributeError, |
| "Can't pickle local attribute %R on %R", name, obj); |
| Py_DECREF(dotted_path); |
| return NULL; |
| } |
| } |
| return dotted_path; |
| } |
| |
| static PyObject * |
| get_deep_attribute(PyObject *obj, PyObject *names, PyObject **pparent) |
| { |
| Py_ssize_t i, n; |
| PyObject *parent = NULL; |
| |
| assert(PyList_CheckExact(names)); |
| Py_INCREF(obj); |
| n = PyList_GET_SIZE(names); |
| for (i = 0; i < n; i++) { |
| PyObject *name = PyList_GET_ITEM(names, i); |
| Py_XDECREF(parent); |
| parent = obj; |
| obj = PyObject_GetAttr(parent, name); |
| if (obj == NULL) { |
| Py_DECREF(parent); |
| return NULL; |
| } |
| } |
| if (pparent != NULL) |
| *pparent = parent; |
| else |
| Py_XDECREF(parent); |
| return obj; |
| } |
| |
| static void |
| reformat_attribute_error(PyObject *obj, PyObject *name) |
| { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) { |
| PyErr_Clear(); |
| PyErr_Format(PyExc_AttributeError, |
| "Can't get attribute %R on %R", name, obj); |
| } |
| } |
| |
| |
| static PyObject * |
| getattribute(PyObject *obj, PyObject *name, int allow_qualname) |
| { |
| PyObject *dotted_path, *attr; |
| |
| if (allow_qualname) { |
| dotted_path = get_dotted_path(obj, name); |
| if (dotted_path == NULL) |
| return NULL; |
| attr = get_deep_attribute(obj, dotted_path, NULL); |
| Py_DECREF(dotted_path); |
| } |
| else |
| attr = PyObject_GetAttr(obj, name); |
| if (attr == NULL) |
| reformat_attribute_error(obj, name); |
| return attr; |
| } |
| |
| static PyObject * |
| whichmodule(PyObject *global, PyObject *dotted_path) |
| { |
| PyObject *module_name; |
| PyObject *modules_dict; |
| PyObject *module; |
| Py_ssize_t i; |
| _Py_IDENTIFIER(__module__); |
| _Py_IDENTIFIER(modules); |
| _Py_IDENTIFIER(__main__); |
| |
| module_name = _PyObject_GetAttrId(global, &PyId___module__); |
| |
| if (module_name == NULL) { |
| if (!PyErr_ExceptionMatches(PyExc_AttributeError)) |
| return NULL; |
| PyErr_Clear(); |
| } |
| else { |
| /* In some rare cases (e.g., bound methods of extension types), |
| __module__ can be None. If it is so, then search sys.modules for |
| the module of global. */ |
| if (module_name != Py_None) |
| return module_name; |
| Py_CLEAR(module_name); |
| } |
| assert(module_name == NULL); |
| |
| /* Fallback on walking sys.modules */ |
| modules_dict = _PySys_GetObjectId(&PyId_modules); |
| if (modules_dict == NULL) { |
| PyErr_SetString(PyExc_RuntimeError, "unable to get sys.modules"); |
| return NULL; |
| } |
| |
| i = 0; |
| while (PyDict_Next(modules_dict, &i, &module_name, &module)) { |
| PyObject *candidate; |
| if (PyUnicode_Check(module_name) && |
| !PyUnicode_CompareWithASCIIString(module_name, "__main__")) |
| continue; |
| if (module == Py_None) |
| continue; |
| |
| candidate = get_deep_attribute(module, dotted_path, NULL); |
| if (candidate == NULL) { |
| if (!PyErr_ExceptionMatches(PyExc_AttributeError)) |
| return NULL; |
| PyErr_Clear(); |
| continue; |
| } |
| |
| if (candidate == global) { |
| Py_INCREF(module_name); |
| Py_DECREF(candidate); |
| return module_name; |
| } |
| Py_DECREF(candidate); |
| } |
| |
| /* If no module is found, use __main__. */ |
| module_name = _PyUnicode_FromId(&PyId___main__); |
| Py_INCREF(module_name); |
| return module_name; |
| } |
| |
| /* fast_save_enter() and fast_save_leave() are guards against recursive |
| objects when Pickler is used with the "fast mode" (i.e., with object |
| memoization disabled). If the nesting of a list or dict object exceed |
| FAST_NESTING_LIMIT, these guards will start keeping an internal |
| reference to the seen list or dict objects and check whether these objects |
| are recursive. These are not strictly necessary, since save() has a |
| hard-coded recursion limit, but they give a nicer error message than the |
| typical RuntimeError. */ |
| static int |
| fast_save_enter(PicklerObject *self, PyObject *obj) |
| { |
| /* if fast_nesting < 0, we're doing an error exit. */ |
| if (++self->fast_nesting >= FAST_NESTING_LIMIT) { |
| PyObject *key = NULL; |
| if (self->fast_memo == NULL) { |
| self->fast_memo = PyDict_New(); |
| if (self->fast_memo == NULL) { |
| self->fast_nesting = -1; |
| return 0; |
| } |
| } |
| key = PyLong_FromVoidPtr(obj); |
| if (key == NULL) |
| return 0; |
| if (PyDict_GetItemWithError(self->fast_memo, key)) { |
| Py_DECREF(key); |
| PyErr_Format(PyExc_ValueError, |
| "fast mode: can't pickle cyclic objects " |
| "including object type %.200s at %p", |
| obj->ob_type->tp_name, obj); |
| self->fast_nesting = -1; |
| return 0; |
| } |
| if (PyErr_Occurred()) { |
| return 0; |
| } |
| if (PyDict_SetItem(self->fast_memo, key, Py_None) < 0) { |
| Py_DECREF(key); |
| self->fast_nesting = -1; |
| return 0; |
| } |
| Py_DECREF(key); |
| } |
| return 1; |
| } |
| |
| static int |
| fast_save_leave(PicklerObject *self, PyObject *obj) |
| { |
| if (self->fast_nesting-- >= FAST_NESTING_LIMIT) { |
| PyObject *key = PyLong_FromVoidPtr(obj); |
| if (key == NULL) |
| return 0; |
| if (PyDict_DelItem(self->fast_memo, key) < 0) { |
| Py_DECREF(key); |
| return 0; |
| } |
| Py_DECREF(key); |
| } |
| return 1; |
| } |
| |
| static int |
| save_none(PicklerObject *self, PyObject *obj) |
| { |
| const char none_op = NONE; |
| if (_Pickler_Write(self, &none_op, 1) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| static int |
| save_bool(PicklerObject *self, PyObject *obj) |
| { |
| if (self->proto >= 2) { |
| const char bool_op = (obj == Py_True) ? NEWTRUE : NEWFALSE; |
| if (_Pickler_Write(self, &bool_op, 1) < 0) |
| return -1; |
| } |
| else { |
| /* These aren't opcodes -- they're ways to pickle bools before protocol 2 |
| * so that unpicklers written before bools were introduced unpickle them |
| * as ints, but unpicklers after can recognize that bools were intended. |
| * Note that protocol 2 added direct ways to pickle bools. |
| */ |
| const char *bool_str = (obj == Py_True) ? "I01\n" : "I00\n"; |
| if (_Pickler_Write(self, bool_str, strlen(bool_str)) < 0) |
| return -1; |
| } |
| return 0; |
| } |
| |
| static int |
| save_long(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *repr = NULL; |
| Py_ssize_t size; |
| long val; |
| int status = 0; |
| |
| const char long_op = LONG; |
| |
| val= PyLong_AsLong(obj); |
| if (val == -1 && PyErr_Occurred()) { |
| /* out of range for int pickling */ |
| PyErr_Clear(); |
| } |
| else if (self->bin && |
| (sizeof(long) <= 4 || |
| (val <= 0x7fffffffL && val >= (-0x7fffffffL - 1)))) { |
| /* result fits in a signed 4-byte integer. |
| |
| Note: we can't use -0x80000000L in the above condition because some |
| compilers (e.g., MSVC) will promote 0x80000000L to an unsigned type |
| before applying the unary minus when sizeof(long) <= 4. The |
| resulting value stays unsigned which is commonly not what we want, |
| so MSVC happily warns us about it. However, that result would have |
| been fine because we guard for sizeof(long) <= 4 which turns the |
| condition true in that particular case. */ |
| char pdata[32]; |
| Py_ssize_t len = 0; |
| |
| pdata[1] = (unsigned char)(val & 0xff); |
| pdata[2] = (unsigned char)((val >> 8) & 0xff); |
| pdata[3] = (unsigned char)((val >> 16) & 0xff); |
| pdata[4] = (unsigned char)((val >> 24) & 0xff); |
| |
| if ((pdata[4] == 0) && (pdata[3] == 0)) { |
| if (pdata[2] == 0) { |
| pdata[0] = BININT1; |
| len = 2; |
| } |
| else { |
| pdata[0] = BININT2; |
| len = 3; |
| } |
| } |
| else { |
| pdata[0] = BININT; |
| len = 5; |
| } |
| |
| if (_Pickler_Write(self, pdata, len) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| if (self->proto >= 2) { |
| /* Linear-time pickling. */ |
| size_t nbits; |
| size_t nbytes; |
| unsigned char *pdata; |
| char header[5]; |
| int i; |
| int sign = _PyLong_Sign(obj); |
| |
| if (sign == 0) { |
| header[0] = LONG1; |
| header[1] = 0; /* It's 0 -- an empty bytestring. */ |
| if (_Pickler_Write(self, header, 2) < 0) |
| goto error; |
| return 0; |
| } |
| nbits = _PyLong_NumBits(obj); |
| if (nbits == (size_t)-1 && PyErr_Occurred()) |
| goto error; |
| /* How many bytes do we need? There are nbits >> 3 full |
| * bytes of data, and nbits & 7 leftover bits. If there |
| * are any leftover bits, then we clearly need another |
| * byte. Wnat's not so obvious is that we *probably* |
| * need another byte even if there aren't any leftovers: |
| * the most-significant bit of the most-significant byte |
| * acts like a sign bit, and it's usually got a sense |
| * opposite of the one we need. The exception is ints |
| * of the form -(2**(8*j-1)) for j > 0. Such an int is |
| * its own 256's-complement, so has the right sign bit |
| * even without the extra byte. That's a pain to check |
| * for in advance, though, so we always grab an extra |
| * byte at the start, and cut it back later if possible. |
| */ |
| nbytes = (nbits >> 3) + 1; |
| if (nbytes > 0x7fffffffL) { |
| PyErr_SetString(PyExc_OverflowError, |
| "int too large to pickle"); |
| goto error; |
| } |
| repr = PyBytes_FromStringAndSize(NULL, (Py_ssize_t)nbytes); |
| if (repr == NULL) |
| goto error; |
| pdata = (unsigned char *)PyBytes_AS_STRING(repr); |
| i = _PyLong_AsByteArray((PyLongObject *)obj, |
| pdata, nbytes, |
| 1 /* little endian */ , 1 /* signed */ ); |
| if (i < 0) |
| goto error; |
| /* If the int is negative, this may be a byte more than |
| * needed. This is so iff the MSB is all redundant sign |
| * bits. |
| */ |
| if (sign < 0 && |
| nbytes > 1 && |
| pdata[nbytes - 1] == 0xff && |
| (pdata[nbytes - 2] & 0x80) != 0) { |
| nbytes--; |
| } |
| |
| if (nbytes < 256) { |
| header[0] = LONG1; |
| header[1] = (unsigned char)nbytes; |
| size = 2; |
| } |
| else { |
| header[0] = LONG4; |
| size = (Py_ssize_t) nbytes; |
| for (i = 1; i < 5; i++) { |
| header[i] = (unsigned char)(size & 0xff); |
| size >>= 8; |
| } |
| size = 5; |
| } |
| if (_Pickler_Write(self, header, size) < 0 || |
| _Pickler_Write(self, (char *)pdata, (int)nbytes) < 0) |
| goto error; |
| } |
| else { |
| char *string; |
| |
| /* proto < 2: write the repr and newline. This is quadratic-time (in |
| the number of digits), in both directions. We add a trailing 'L' |
| to the repr, for compatibility with Python 2.x. */ |
| |
| repr = PyObject_Repr(obj); |
| if (repr == NULL) |
| goto error; |
| |
| string = _PyUnicode_AsStringAndSize(repr, &size); |
| if (string == NULL) |
| goto error; |
| |
| if (_Pickler_Write(self, &long_op, 1) < 0 || |
| _Pickler_Write(self, string, size) < 0 || |
| _Pickler_Write(self, "L\n", 2) < 0) |
| goto error; |
| } |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| Py_XDECREF(repr); |
| |
| return status; |
| } |
| |
| static int |
| save_float(PicklerObject *self, PyObject *obj) |
| { |
| double x = PyFloat_AS_DOUBLE((PyFloatObject *)obj); |
| |
| if (self->bin) { |
| char pdata[9]; |
| pdata[0] = BINFLOAT; |
| if (_PyFloat_Pack8(x, (unsigned char *)&pdata[1], 0) < 0) |
| return -1; |
| if (_Pickler_Write(self, pdata, 9) < 0) |
| return -1; |
| } |
| else { |
| int result = -1; |
| char *buf = NULL; |
| char op = FLOAT; |
| |
| if (_Pickler_Write(self, &op, 1) < 0) |
| goto done; |
| |
| buf = PyOS_double_to_string(x, 'r', 0, Py_DTSF_ADD_DOT_0, NULL); |
| if (!buf) { |
| PyErr_NoMemory(); |
| goto done; |
| } |
| |
| if (_Pickler_Write(self, buf, strlen(buf)) < 0) |
| goto done; |
| |
| if (_Pickler_Write(self, "\n", 1) < 0) |
| goto done; |
| |
| result = 0; |
| done: |
| PyMem_Free(buf); |
| return result; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| save_bytes(PicklerObject *self, PyObject *obj) |
| { |
| if (self->proto < 3) { |
| /* Older pickle protocols do not have an opcode for pickling bytes |
| objects. Therefore, we need to fake the copy protocol (i.e., |
| the __reduce__ method) to permit bytes object unpickling. |
| |
| Here we use a hack to be compatible with Python 2. Since in Python |
| 2 'bytes' is just an alias for 'str' (which has different |
| parameters than the actual bytes object), we use codecs.encode |
| to create the appropriate 'str' object when unpickled using |
| Python 2 *and* the appropriate 'bytes' object when unpickled |
| using Python 3. Again this is a hack and we don't need to do this |
| with newer protocols. */ |
| PyObject *reduce_value = NULL; |
| int status; |
| |
| if (PyBytes_GET_SIZE(obj) == 0) { |
| reduce_value = Py_BuildValue("(O())", (PyObject*)&PyBytes_Type); |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyObject *unicode_str = |
| PyUnicode_DecodeLatin1(PyBytes_AS_STRING(obj), |
| PyBytes_GET_SIZE(obj), |
| "strict"); |
| _Py_IDENTIFIER(latin1); |
| |
| if (unicode_str == NULL) |
| return -1; |
| reduce_value = Py_BuildValue("(O(OO))", |
| st->codecs_encode, unicode_str, |
| _PyUnicode_FromId(&PyId_latin1)); |
| Py_DECREF(unicode_str); |
| } |
| |
| if (reduce_value == NULL) |
| return -1; |
| |
| /* save_reduce() will memoize the object automatically. */ |
| status = save_reduce(self, reduce_value, obj); |
| Py_DECREF(reduce_value); |
| return status; |
| } |
| else { |
| Py_ssize_t size; |
| char header[9]; |
| Py_ssize_t len; |
| |
| size = PyBytes_GET_SIZE(obj); |
| if (size < 0) |
| return -1; |
| |
| if (size <= 0xff) { |
| header[0] = SHORT_BINBYTES; |
| header[1] = (unsigned char)size; |
| len = 2; |
| } |
| else if ((size_t)size <= 0xffffffffUL) { |
| header[0] = BINBYTES; |
| header[1] = (unsigned char)(size & 0xff); |
| header[2] = (unsigned char)((size >> 8) & 0xff); |
| header[3] = (unsigned char)((size >> 16) & 0xff); |
| header[4] = (unsigned char)((size >> 24) & 0xff); |
| len = 5; |
| } |
| else if (self->proto >= 4) { |
| header[0] = BINBYTES8; |
| _write_size64(header + 1, size); |
| len = 9; |
| } |
| else { |
| PyErr_SetString(PyExc_OverflowError, |
| "cannot serialize a bytes object larger than 4 GiB"); |
| return -1; /* string too large */ |
| } |
| |
| if (_Pickler_Write(self, header, len) < 0) |
| return -1; |
| |
| if (_Pickler_Write(self, PyBytes_AS_STRING(obj), size) < 0) |
| return -1; |
| |
| if (memo_put(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| } |
| |
| /* A copy of PyUnicode_EncodeRawUnicodeEscape() that also translates |
| backslash and newline characters to \uXXXX escapes. */ |
| static PyObject * |
| raw_unicode_escape(PyObject *obj) |
| { |
| PyObject *repr; |
| char *p; |
| Py_ssize_t i, size; |
| size_t expandsize; |
| void *data; |
| unsigned int kind; |
| |
| if (PyUnicode_READY(obj)) |
| return NULL; |
| |
| size = PyUnicode_GET_LENGTH(obj); |
| data = PyUnicode_DATA(obj); |
| kind = PyUnicode_KIND(obj); |
| if (kind == PyUnicode_4BYTE_KIND) |
| expandsize = 10; |
| else |
| expandsize = 6; |
| |
| if ((size_t)size > (size_t)PY_SSIZE_T_MAX / expandsize) |
| return PyErr_NoMemory(); |
| repr = PyBytes_FromStringAndSize(NULL, expandsize * size); |
| if (repr == NULL) |
| return NULL; |
| if (size == 0) |
| return repr; |
| assert(Py_REFCNT(repr) == 1); |
| |
| p = PyBytes_AS_STRING(repr); |
| for (i=0; i < size; i++) { |
| Py_UCS4 ch = PyUnicode_READ(kind, data, i); |
| /* Map 32-bit characters to '\Uxxxxxxxx' */ |
| if (ch >= 0x10000) { |
| *p++ = '\\'; |
| *p++ = 'U'; |
| *p++ = Py_hexdigits[(ch >> 28) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 24) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 20) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 16) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 12) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 8) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 4) & 0xf]; |
| *p++ = Py_hexdigits[ch & 15]; |
| } |
| /* Map 16-bit characters to '\uxxxx' */ |
| else if (ch >= 256 || ch == '\\' || ch == '\n') { |
| *p++ = '\\'; |
| *p++ = 'u'; |
| *p++ = Py_hexdigits[(ch >> 12) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 8) & 0xf]; |
| *p++ = Py_hexdigits[(ch >> 4) & 0xf]; |
| *p++ = Py_hexdigits[ch & 15]; |
| } |
| /* Copy everything else as-is */ |
| else |
| *p++ = (char) ch; |
| } |
| size = p - PyBytes_AS_STRING(repr); |
| if (_PyBytes_Resize(&repr, size) < 0) |
| return NULL; |
| return repr; |
| } |
| |
| static int |
| write_utf8(PicklerObject *self, char *data, Py_ssize_t size) |
| { |
| char header[9]; |
| Py_ssize_t len; |
| |
| assert(size >= 0); |
| if (size <= 0xff && self->proto >= 4) { |
| header[0] = SHORT_BINUNICODE; |
| header[1] = (unsigned char)(size & 0xff); |
| len = 2; |
| } |
| else if ((size_t)size <= 0xffffffffUL) { |
| header[0] = BINUNICODE; |
| header[1] = (unsigned char)(size & 0xff); |
| header[2] = (unsigned char)((size >> 8) & 0xff); |
| header[3] = (unsigned char)((size >> 16) & 0xff); |
| header[4] = (unsigned char)((size >> 24) & 0xff); |
| len = 5; |
| } |
| else if (self->proto >= 4) { |
| header[0] = BINUNICODE8; |
| _write_size64(header + 1, size); |
| len = 9; |
| } |
| else { |
| PyErr_SetString(PyExc_OverflowError, |
| "cannot serialize a string larger than 4GiB"); |
| return -1; |
| } |
| |
| if (_Pickler_Write(self, header, len) < 0) |
| return -1; |
| if (_Pickler_Write(self, data, size) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| static int |
| write_unicode_binary(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *encoded = NULL; |
| Py_ssize_t size; |
| char *data; |
| int r; |
| |
| if (PyUnicode_READY(obj)) |
| return -1; |
| |
| data = PyUnicode_AsUTF8AndSize(obj, &size); |
| if (data != NULL) |
| return write_utf8(self, data, size); |
| |
| /* Issue #8383: for strings with lone surrogates, fallback on the |
| "surrogatepass" error handler. */ |
| PyErr_Clear(); |
| encoded = PyUnicode_AsEncodedString(obj, "utf-8", "surrogatepass"); |
| if (encoded == NULL) |
| return -1; |
| |
| r = write_utf8(self, PyBytes_AS_STRING(encoded), |
| PyBytes_GET_SIZE(encoded)); |
| Py_DECREF(encoded); |
| return r; |
| } |
| |
| static int |
| save_unicode(PicklerObject *self, PyObject *obj) |
| { |
| if (self->bin) { |
| if (write_unicode_binary(self, obj) < 0) |
| return -1; |
| } |
| else { |
| PyObject *encoded; |
| Py_ssize_t size; |
| const char unicode_op = UNICODE; |
| |
| encoded = raw_unicode_escape(obj); |
| if (encoded == NULL) |
| return -1; |
| |
| if (_Pickler_Write(self, &unicode_op, 1) < 0) { |
| Py_DECREF(encoded); |
| return -1; |
| } |
| |
| size = PyBytes_GET_SIZE(encoded); |
| if (_Pickler_Write(self, PyBytes_AS_STRING(encoded), size) < 0) { |
| Py_DECREF(encoded); |
| return -1; |
| } |
| Py_DECREF(encoded); |
| |
| if (_Pickler_Write(self, "\n", 1) < 0) |
| return -1; |
| } |
| if (memo_put(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| /* A helper for save_tuple. Push the len elements in tuple t on the stack. */ |
| static int |
| store_tuple_elements(PicklerObject *self, PyObject *t, Py_ssize_t len) |
| { |
| Py_ssize_t i; |
| |
| assert(PyTuple_Size(t) == len); |
| |
| for (i = 0; i < len; i++) { |
| PyObject *element = PyTuple_GET_ITEM(t, i); |
| |
| if (element == NULL) |
| return -1; |
| if (save(self, element, 0) < 0) |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* Tuples are ubiquitous in the pickle protocols, so many techniques are |
| * used across protocols to minimize the space needed to pickle them. |
| * Tuples are also the only builtin immutable type that can be recursive |
| * (a tuple can be reached from itself), and that requires some subtle |
| * magic so that it works in all cases. IOW, this is a long routine. |
| */ |
| static int |
| save_tuple(PicklerObject *self, PyObject *obj) |
| { |
| Py_ssize_t len, i; |
| |
| const char mark_op = MARK; |
| const char tuple_op = TUPLE; |
| const char pop_op = POP; |
| const char pop_mark_op = POP_MARK; |
| const char len2opcode[] = {EMPTY_TUPLE, TUPLE1, TUPLE2, TUPLE3}; |
| |
| if ((len = PyTuple_Size(obj)) < 0) |
| return -1; |
| |
| if (len == 0) { |
| char pdata[2]; |
| |
| if (self->proto) { |
| pdata[0] = EMPTY_TUPLE; |
| len = 1; |
| } |
| else { |
| pdata[0] = MARK; |
| pdata[1] = TUPLE; |
| len = 2; |
| } |
| if (_Pickler_Write(self, pdata, len) < 0) |
| return -1; |
| return 0; |
| } |
| |
| /* The tuple isn't in the memo now. If it shows up there after |
| * saving the tuple elements, the tuple must be recursive, in |
| * which case we'll pop everything we put on the stack, and fetch |
| * its value from the memo. |
| */ |
| if (len <= 3 && self->proto >= 2) { |
| /* Use TUPLE{1,2,3} opcodes. */ |
| if (store_tuple_elements(self, obj, len) < 0) |
| return -1; |
| |
| if (PyMemoTable_Get(self->memo, obj)) { |
| /* pop the len elements */ |
| for (i = 0; i < len; i++) |
| if (_Pickler_Write(self, &pop_op, 1) < 0) |
| return -1; |
| /* fetch from memo */ |
| if (memo_get(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| else { /* Not recursive. */ |
| if (_Pickler_Write(self, len2opcode + len, 1) < 0) |
| return -1; |
| } |
| goto memoize; |
| } |
| |
| /* proto < 2 and len > 0, or proto >= 2 and len > 3. |
| * Generate MARK e1 e2 ... TUPLE |
| */ |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| return -1; |
| |
| if (store_tuple_elements(self, obj, len) < 0) |
| return -1; |
| |
| if (PyMemoTable_Get(self->memo, obj)) { |
| /* pop the stack stuff we pushed */ |
| if (self->bin) { |
| if (_Pickler_Write(self, &pop_mark_op, 1) < 0) |
| return -1; |
| } |
| else { |
| /* Note that we pop one more than len, to remove |
| * the MARK too. |
| */ |
| for (i = 0; i <= len; i++) |
| if (_Pickler_Write(self, &pop_op, 1) < 0) |
| return -1; |
| } |
| /* fetch from memo */ |
| if (memo_get(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| else { /* Not recursive. */ |
| if (_Pickler_Write(self, &tuple_op, 1) < 0) |
| return -1; |
| } |
| |
| memoize: |
| if (memo_put(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| /* iter is an iterator giving items, and we batch up chunks of |
| * MARK item item ... item APPENDS |
| * opcode sequences. Calling code should have arranged to first create an |
| * empty list, or list-like object, for the APPENDS to operate on. |
| * Returns 0 on success, <0 on error. |
| */ |
| static int |
| batch_list(PicklerObject *self, PyObject *iter) |
| { |
| PyObject *obj = NULL; |
| PyObject *firstitem = NULL; |
| int i, n; |
| |
| const char mark_op = MARK; |
| const char append_op = APPEND; |
| const char appends_op = APPENDS; |
| |
| assert(iter != NULL); |
| |
| /* XXX: I think this function could be made faster by avoiding the |
| iterator interface and fetching objects directly from list using |
| PyList_GET_ITEM. |
| */ |
| |
| if (self->proto == 0) { |
| /* APPENDS isn't available; do one at a time. */ |
| for (;;) { |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| return -1; |
| break; |
| } |
| i = save(self, obj, 0); |
| Py_DECREF(obj); |
| if (i < 0) |
| return -1; |
| if (_Pickler_Write(self, &append_op, 1) < 0) |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* proto > 0: write in batches of BATCHSIZE. */ |
| do { |
| /* Get first item */ |
| firstitem = PyIter_Next(iter); |
| if (firstitem == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| |
| /* nothing more to add */ |
| break; |
| } |
| |
| /* Try to get a second item */ |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| |
| /* Only one item to write */ |
| if (save(self, firstitem, 0) < 0) |
| goto error; |
| if (_Pickler_Write(self, &append_op, 1) < 0) |
| goto error; |
| Py_CLEAR(firstitem); |
| break; |
| } |
| |
| /* More than one item to write */ |
| |
| /* Pump out MARK, items, APPENDS. */ |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| goto error; |
| |
| if (save(self, firstitem, 0) < 0) |
| goto error; |
| Py_CLEAR(firstitem); |
| n = 1; |
| |
| /* Fetch and save up to BATCHSIZE items */ |
| while (obj) { |
| if (save(self, obj, 0) < 0) |
| goto error; |
| Py_CLEAR(obj); |
| n += 1; |
| |
| if (n == BATCHSIZE) |
| break; |
| |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| break; |
| } |
| } |
| |
| if (_Pickler_Write(self, &appends_op, 1) < 0) |
| goto error; |
| |
| } while (n == BATCHSIZE); |
| return 0; |
| |
| error: |
| Py_XDECREF(firstitem); |
| Py_XDECREF(obj); |
| return -1; |
| } |
| |
| /* This is a variant of batch_list() above, specialized for lists (with no |
| * support for list subclasses). Like batch_list(), we batch up chunks of |
| * MARK item item ... item APPENDS |
| * opcode sequences. Calling code should have arranged to first create an |
| * empty list, or list-like object, for the APPENDS to operate on. |
| * Returns 0 on success, -1 on error. |
| * |
| * This version is considerably faster than batch_list(), if less general. |
| * |
| * Note that this only works for protocols > 0. |
| */ |
| static int |
| batch_list_exact(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *item = NULL; |
| Py_ssize_t this_batch, total; |
| |
| const char append_op = APPEND; |
| const char appends_op = APPENDS; |
| const char mark_op = MARK; |
| |
| assert(obj != NULL); |
| assert(self->proto > 0); |
| assert(PyList_CheckExact(obj)); |
| |
| if (PyList_GET_SIZE(obj) == 1) { |
| item = PyList_GET_ITEM(obj, 0); |
| if (save(self, item, 0) < 0) |
| return -1; |
| if (_Pickler_Write(self, &append_op, 1) < 0) |
| return -1; |
| return 0; |
| } |
| |
| /* Write in batches of BATCHSIZE. */ |
| total = 0; |
| do { |
| this_batch = 0; |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| return -1; |
| while (total < PyList_GET_SIZE(obj)) { |
| item = PyList_GET_ITEM(obj, total); |
| if (save(self, item, 0) < 0) |
| return -1; |
| total++; |
| if (++this_batch == BATCHSIZE) |
| break; |
| } |
| if (_Pickler_Write(self, &appends_op, 1) < 0) |
| return -1; |
| |
| } while (total < PyList_GET_SIZE(obj)); |
| |
| return 0; |
| } |
| |
| static int |
| save_list(PicklerObject *self, PyObject *obj) |
| { |
| char header[3]; |
| Py_ssize_t len; |
| int status = 0; |
| |
| if (self->fast && !fast_save_enter(self, obj)) |
| goto error; |
| |
| /* Create an empty list. */ |
| if (self->bin) { |
| header[0] = EMPTY_LIST; |
| len = 1; |
| } |
| else { |
| header[0] = MARK; |
| header[1] = LIST; |
| len = 2; |
| } |
| |
| if (_Pickler_Write(self, header, len) < 0) |
| goto error; |
| |
| /* Get list length, and bow out early if empty. */ |
| if ((len = PyList_Size(obj)) < 0) |
| goto error; |
| |
| if (memo_put(self, obj) < 0) |
| goto error; |
| |
| if (len != 0) { |
| /* Materialize the list elements. */ |
| if (PyList_CheckExact(obj) && self->proto > 0) { |
| if (Py_EnterRecursiveCall(" while pickling an object")) |
| goto error; |
| status = batch_list_exact(self, obj); |
| Py_LeaveRecursiveCall(); |
| } else { |
| PyObject *iter = PyObject_GetIter(obj); |
| if (iter == NULL) |
| goto error; |
| |
| if (Py_EnterRecursiveCall(" while pickling an object")) { |
| Py_DECREF(iter); |
| goto error; |
| } |
| status = batch_list(self, iter); |
| Py_LeaveRecursiveCall(); |
| Py_DECREF(iter); |
| } |
| } |
| if (0) { |
| error: |
| status = -1; |
| } |
| |
| if (self->fast && !fast_save_leave(self, obj)) |
| status = -1; |
| |
| return status; |
| } |
| |
| /* iter is an iterator giving (key, value) pairs, and we batch up chunks of |
| * MARK key value ... key value SETITEMS |
| * opcode sequences. Calling code should have arranged to first create an |
| * empty dict, or dict-like object, for the SETITEMS to operate on. |
| * Returns 0 on success, <0 on error. |
| * |
| * This is very much like batch_list(). The difference between saving |
| * elements directly, and picking apart two-tuples, is so long-winded at |
| * the C level, though, that attempts to combine these routines were too |
| * ugly to bear. |
| */ |
| static int |
| batch_dict(PicklerObject *self, PyObject *iter) |
| { |
| PyObject *obj = NULL; |
| PyObject *firstitem = NULL; |
| int i, n; |
| |
| const char mark_op = MARK; |
| const char setitem_op = SETITEM; |
| const char setitems_op = SETITEMS; |
| |
| assert(iter != NULL); |
| |
| if (self->proto == 0) { |
| /* SETITEMS isn't available; do one at a time. */ |
| for (;;) { |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| return -1; |
| break; |
| } |
| if (!PyTuple_Check(obj) || PyTuple_Size(obj) != 2) { |
| PyErr_SetString(PyExc_TypeError, "dict items " |
| "iterator must return 2-tuples"); |
| return -1; |
| } |
| i = save(self, PyTuple_GET_ITEM(obj, 0), 0); |
| if (i >= 0) |
| i = save(self, PyTuple_GET_ITEM(obj, 1), 0); |
| Py_DECREF(obj); |
| if (i < 0) |
| return -1; |
| if (_Pickler_Write(self, &setitem_op, 1) < 0) |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* proto > 0: write in batches of BATCHSIZE. */ |
| do { |
| /* Get first item */ |
| firstitem = PyIter_Next(iter); |
| if (firstitem == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| |
| /* nothing more to add */ |
| break; |
| } |
| if (!PyTuple_Check(firstitem) || PyTuple_Size(firstitem) != 2) { |
| PyErr_SetString(PyExc_TypeError, "dict items " |
| "iterator must return 2-tuples"); |
| goto error; |
| } |
| |
| /* Try to get a second item */ |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| |
| /* Only one item to write */ |
| if (save(self, PyTuple_GET_ITEM(firstitem, 0), 0) < 0) |
| goto error; |
| if (save(self, PyTuple_GET_ITEM(firstitem, 1), 0) < 0) |
| goto error; |
| if (_Pickler_Write(self, &setitem_op, 1) < 0) |
| goto error; |
| Py_CLEAR(firstitem); |
| break; |
| } |
| |
| /* More than one item to write */ |
| |
| /* Pump out MARK, items, SETITEMS. */ |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| goto error; |
| |
| if (save(self, PyTuple_GET_ITEM(firstitem, 0), 0) < 0) |
| goto error; |
| if (save(self, PyTuple_GET_ITEM(firstitem, 1), 0) < 0) |
| goto error; |
| Py_CLEAR(firstitem); |
| n = 1; |
| |
| /* Fetch and save up to BATCHSIZE items */ |
| while (obj) { |
| if (!PyTuple_Check(obj) || PyTuple_Size(obj) != 2) { |
| PyErr_SetString(PyExc_TypeError, "dict items " |
| "iterator must return 2-tuples"); |
| goto error; |
| } |
| if (save(self, PyTuple_GET_ITEM(obj, 0), 0) < 0 || |
| save(self, PyTuple_GET_ITEM(obj, 1), 0) < 0) |
| goto error; |
| Py_CLEAR(obj); |
| n += 1; |
| |
| if (n == BATCHSIZE) |
| break; |
| |
| obj = PyIter_Next(iter); |
| if (obj == NULL) { |
| if (PyErr_Occurred()) |
| goto error; |
| break; |
| } |
| } |
| |
| if (_Pickler_Write(self, &setitems_op, 1) < 0) |
| goto error; |
| |
| } while (n == BATCHSIZE); |
| return 0; |
| |
| error: |
| Py_XDECREF(firstitem); |
| Py_XDECREF(obj); |
| return -1; |
| } |
| |
| /* This is a variant of batch_dict() above that specializes for dicts, with no |
| * support for dict subclasses. Like batch_dict(), we batch up chunks of |
| * MARK key value ... key value SETITEMS |
| * opcode sequences. Calling code should have arranged to first create an |
| * empty dict, or dict-like object, for the SETITEMS to operate on. |
| * Returns 0 on success, -1 on error. |
| * |
| * Note that this currently doesn't work for protocol 0. |
| */ |
| static int |
| batch_dict_exact(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *key = NULL, *value = NULL; |
| int i; |
| Py_ssize_t dict_size, ppos = 0; |
| |
| const char mark_op = MARK; |
| const char setitem_op = SETITEM; |
| const char setitems_op = SETITEMS; |
| |
| assert(obj != NULL); |
| assert(self->proto > 0); |
| |
| dict_size = PyDict_Size(obj); |
| |
| /* Special-case len(d) == 1 to save space. */ |
| if (dict_size == 1) { |
| PyDict_Next(obj, &ppos, &key, &value); |
| if (save(self, key, 0) < 0) |
| return -1; |
| if (save(self, value, 0) < 0) |
| return -1; |
| if (_Pickler_Write(self, &setitem_op, 1) < 0) |
| return -1; |
| return 0; |
| } |
| |
| /* Write in batches of BATCHSIZE. */ |
| do { |
| i = 0; |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| return -1; |
| while (PyDict_Next(obj, &ppos, &key, &value)) { |
| if (save(self, key, 0) < 0) |
| return -1; |
| if (save(self, value, 0) < 0) |
| return -1; |
| if (++i == BATCHSIZE) |
| break; |
| } |
| if (_Pickler_Write(self, &setitems_op, 1) < 0) |
| return -1; |
| if (PyDict_Size(obj) != dict_size) { |
| PyErr_Format( |
| PyExc_RuntimeError, |
| "dictionary changed size during iteration"); |
| return -1; |
| } |
| |
| } while (i == BATCHSIZE); |
| return 0; |
| } |
| |
| static int |
| save_dict(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *items, *iter; |
| char header[3]; |
| Py_ssize_t len; |
| int status = 0; |
| |
| if (self->fast && !fast_save_enter(self, obj)) |
| goto error; |
| |
| /* Create an empty dict. */ |
| if (self->bin) { |
| header[0] = EMPTY_DICT; |
| len = 1; |
| } |
| else { |
| header[0] = MARK; |
| header[1] = DICT; |
| len = 2; |
| } |
| |
| if (_Pickler_Write(self, header, len) < 0) |
| goto error; |
| |
| /* Get dict size, and bow out early if empty. */ |
| if ((len = PyDict_Size(obj)) < 0) |
| goto error; |
| |
| if (memo_put(self, obj) < 0) |
| goto error; |
| |
| if (len != 0) { |
| /* Save the dict items. */ |
| if (PyDict_CheckExact(obj) && self->proto > 0) { |
| /* We can take certain shortcuts if we know this is a dict and |
| not a dict subclass. */ |
| if (Py_EnterRecursiveCall(" while pickling an object")) |
| goto error; |
| status = batch_dict_exact(self, obj); |
| Py_LeaveRecursiveCall(); |
| } else { |
| _Py_IDENTIFIER(items); |
| |
| items = _PyObject_CallMethodId(obj, &PyId_items, "()"); |
| if (items == NULL) |
| goto error; |
| iter = PyObject_GetIter(items); |
| Py_DECREF(items); |
| if (iter == NULL) |
| goto error; |
| if (Py_EnterRecursiveCall(" while pickling an object")) { |
| Py_DECREF(iter); |
| goto error; |
| } |
| status = batch_dict(self, iter); |
| Py_LeaveRecursiveCall(); |
| Py_DECREF(iter); |
| } |
| } |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| |
| if (self->fast && !fast_save_leave(self, obj)) |
| status = -1; |
| |
| return status; |
| } |
| |
| static int |
| save_set(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *item; |
| int i; |
| Py_ssize_t set_size, ppos = 0; |
| Py_hash_t hash; |
| |
| const char empty_set_op = EMPTY_SET; |
| const char mark_op = MARK; |
| const char additems_op = ADDITEMS; |
| |
| if (self->proto < 4) { |
| PyObject *items; |
| PyObject *reduce_value; |
| int status; |
| |
| items = PySequence_List(obj); |
| if (items == NULL) { |
| return -1; |
| } |
| reduce_value = Py_BuildValue("(O(O))", (PyObject*)&PySet_Type, items); |
| Py_DECREF(items); |
| if (reduce_value == NULL) { |
| return -1; |
| } |
| /* save_reduce() will memoize the object automatically. */ |
| status = save_reduce(self, reduce_value, obj); |
| Py_DECREF(reduce_value); |
| return status; |
| } |
| |
| if (_Pickler_Write(self, &empty_set_op, 1) < 0) |
| return -1; |
| |
| if (memo_put(self, obj) < 0) |
| return -1; |
| |
| set_size = PySet_GET_SIZE(obj); |
| if (set_size == 0) |
| return 0; /* nothing to do */ |
| |
| /* Write in batches of BATCHSIZE. */ |
| do { |
| i = 0; |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| return -1; |
| while (_PySet_NextEntry(obj, &ppos, &item, &hash)) { |
| if (save(self, item, 0) < 0) |
| return -1; |
| if (++i == BATCHSIZE) |
| break; |
| } |
| if (_Pickler_Write(self, &additems_op, 1) < 0) |
| return -1; |
| if (PySet_GET_SIZE(obj) != set_size) { |
| PyErr_Format( |
| PyExc_RuntimeError, |
| "set changed size during iteration"); |
| return -1; |
| } |
| } while (i == BATCHSIZE); |
| |
| return 0; |
| } |
| |
| static int |
| save_frozenset(PicklerObject *self, PyObject *obj) |
| { |
| PyObject *iter; |
| |
| const char mark_op = MARK; |
| const char frozenset_op = FROZENSET; |
| |
| if (self->fast && !fast_save_enter(self, obj)) |
| return -1; |
| |
| if (self->proto < 4) { |
| PyObject *items; |
| PyObject *reduce_value; |
| int status; |
| |
| items = PySequence_List(obj); |
| if (items == NULL) { |
| return -1; |
| } |
| reduce_value = Py_BuildValue("(O(O))", (PyObject*)&PyFrozenSet_Type, |
| items); |
| Py_DECREF(items); |
| if (reduce_value == NULL) { |
| return -1; |
| } |
| /* save_reduce() will memoize the object automatically. */ |
| status = save_reduce(self, reduce_value, obj); |
| Py_DECREF(reduce_value); |
| return status; |
| } |
| |
| if (_Pickler_Write(self, &mark_op, 1) < 0) |
| return -1; |
| |
| iter = PyObject_GetIter(obj); |
| if (iter == NULL) { |
| return -1; |
| } |
| for (;;) { |
| PyObject *item; |
| |
| item = PyIter_Next(iter); |
| if (item == NULL) { |
| if (PyErr_Occurred()) { |
| Py_DECREF(iter); |
| return -1; |
| } |
| break; |
| } |
| if (save(self, item, 0) < 0) { |
| Py_DECREF(item); |
| Py_DECREF(iter); |
| return -1; |
| } |
| Py_DECREF(item); |
| } |
| Py_DECREF(iter); |
| |
| /* If the object is already in the memo, this means it is |
| recursive. In this case, throw away everything we put on the |
| stack, and fetch the object back from the memo. */ |
| if (PyMemoTable_Get(self->memo, obj)) { |
| const char pop_mark_op = POP_MARK; |
| |
| if (_Pickler_Write(self, &pop_mark_op, 1) < 0) |
| return -1; |
| if (memo_get(self, obj) < 0) |
| return -1; |
| return 0; |
| } |
| |
| if (_Pickler_Write(self, &frozenset_op, 1) < 0) |
| return -1; |
| if (memo_put(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| static int |
| fix_imports(PyObject **module_name, PyObject **global_name) |
| { |
| PyObject *key; |
| PyObject *item; |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| key = PyTuple_Pack(2, *module_name, *global_name); |
| if (key == NULL) |
| return -1; |
| item = PyDict_GetItemWithError(st->name_mapping_3to2, key); |
| Py_DECREF(key); |
| if (item) { |
| PyObject *fixed_module_name; |
| PyObject *fixed_global_name; |
| |
| if (!PyTuple_Check(item) || PyTuple_GET_SIZE(item) != 2) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.REVERSE_NAME_MAPPING values " |
| "should be 2-tuples, not %.200s", |
| Py_TYPE(item)->tp_name); |
| return -1; |
| } |
| fixed_module_name = PyTuple_GET_ITEM(item, 0); |
| fixed_global_name = PyTuple_GET_ITEM(item, 1); |
| if (!PyUnicode_Check(fixed_module_name) || |
| !PyUnicode_Check(fixed_global_name)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.REVERSE_NAME_MAPPING values " |
| "should be pairs of str, not (%.200s, %.200s)", |
| Py_TYPE(fixed_module_name)->tp_name, |
| Py_TYPE(fixed_global_name)->tp_name); |
| return -1; |
| } |
| |
| Py_CLEAR(*module_name); |
| Py_CLEAR(*global_name); |
| Py_INCREF(fixed_module_name); |
| Py_INCREF(fixed_global_name); |
| *module_name = fixed_module_name; |
| *global_name = fixed_global_name; |
| return 0; |
| } |
| else if (PyErr_Occurred()) { |
| return -1; |
| } |
| |
| item = PyDict_GetItemWithError(st->import_mapping_3to2, *module_name); |
| if (item) { |
| if (!PyUnicode_Check(item)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.REVERSE_IMPORT_MAPPING values " |
| "should be strings, not %.200s", |
| Py_TYPE(item)->tp_name); |
| return -1; |
| } |
| Py_INCREF(item); |
| Py_XSETREF(*module_name, item); |
| } |
| else if (PyErr_Occurred()) { |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| save_global(PicklerObject *self, PyObject *obj, PyObject *name) |
| { |
| PyObject *global_name = NULL; |
| PyObject *module_name = NULL; |
| PyObject *module = NULL; |
| PyObject *parent = NULL; |
| PyObject *dotted_path = NULL; |
| PyObject *lastname = NULL; |
| PyObject *cls; |
| PickleState *st = _Pickle_GetGlobalState(); |
| int status = 0; |
| _Py_IDENTIFIER(__name__); |
| _Py_IDENTIFIER(__qualname__); |
| |
| const char global_op = GLOBAL; |
| |
| if (name) { |
| Py_INCREF(name); |
| global_name = name; |
| } |
| else { |
| global_name = _PyObject_GetAttrId(obj, &PyId___qualname__); |
| if (global_name == NULL) { |
| if (!PyErr_ExceptionMatches(PyExc_AttributeError)) |
| goto error; |
| PyErr_Clear(); |
| } |
| if (global_name == NULL) { |
| global_name = _PyObject_GetAttrId(obj, &PyId___name__); |
| if (global_name == NULL) |
| goto error; |
| } |
| } |
| |
| dotted_path = get_dotted_path(module, global_name); |
| if (dotted_path == NULL) |
| goto error; |
| module_name = whichmodule(obj, dotted_path); |
| if (module_name == NULL) |
| goto error; |
| |
| /* XXX: Change to use the import C API directly with level=0 to disallow |
| relative imports. |
| |
| XXX: PyImport_ImportModuleLevel could be used. However, this bypasses |
| builtins.__import__. Therefore, _pickle, unlike pickle.py, will ignore |
| custom import functions (IMHO, this would be a nice security |
| feature). The import C API would need to be extended to support the |
| extra parameters of __import__ to fix that. */ |
| module = PyImport_Import(module_name); |
| if (module == NULL) { |
| PyErr_Format(st->PicklingError, |
| "Can't pickle %R: import of module %R failed", |
| obj, module_name); |
| goto error; |
| } |
| lastname = PyList_GET_ITEM(dotted_path, PyList_GET_SIZE(dotted_path)-1); |
| Py_INCREF(lastname); |
| cls = get_deep_attribute(module, dotted_path, &parent); |
| Py_CLEAR(dotted_path); |
| if (cls == NULL) { |
| PyErr_Format(st->PicklingError, |
| "Can't pickle %R: attribute lookup %S on %S failed", |
| obj, global_name, module_name); |
| goto error; |
| } |
| if (cls != obj) { |
| Py_DECREF(cls); |
| PyErr_Format(st->PicklingError, |
| "Can't pickle %R: it's not the same object as %S.%S", |
| obj, module_name, global_name); |
| goto error; |
| } |
| Py_DECREF(cls); |
| |
| if (self->proto >= 2) { |
| /* See whether this is in the extension registry, and if |
| * so generate an EXT opcode. |
| */ |
| PyObject *extension_key; |
| PyObject *code_obj; /* extension code as Python object */ |
| long code; /* extension code as C value */ |
| char pdata[5]; |
| Py_ssize_t n; |
| |
| extension_key = PyTuple_Pack(2, module_name, global_name); |
| if (extension_key == NULL) { |
| goto error; |
| } |
| code_obj = PyDict_GetItemWithError(st->extension_registry, |
| extension_key); |
| Py_DECREF(extension_key); |
| /* The object is not registered in the extension registry. |
| This is the most likely code path. */ |
| if (code_obj == NULL) { |
| if (PyErr_Occurred()) { |
| goto error; |
| } |
| goto gen_global; |
| } |
| |
| /* XXX: pickle.py doesn't check neither the type, nor the range |
| of the value returned by the extension_registry. It should for |
| consistency. */ |
| |
| /* Verify code_obj has the right type and value. */ |
| if (!PyLong_Check(code_obj)) { |
| PyErr_Format(st->PicklingError, |
| "Can't pickle %R: extension code %R isn't an integer", |
| obj, code_obj); |
| goto error; |
| } |
| code = PyLong_AS_LONG(code_obj); |
| if (code <= 0 || code > 0x7fffffffL) { |
| if (!PyErr_Occurred()) |
| PyErr_Format(st->PicklingError, "Can't pickle %R: extension " |
| "code %ld is out of range", obj, code); |
| goto error; |
| } |
| |
| /* Generate an EXT opcode. */ |
| if (code <= 0xff) { |
| pdata[0] = EXT1; |
| pdata[1] = (unsigned char)code; |
| n = 2; |
| } |
| else if (code <= 0xffff) { |
| pdata[0] = EXT2; |
| pdata[1] = (unsigned char)(code & 0xff); |
| pdata[2] = (unsigned char)((code >> 8) & 0xff); |
| n = 3; |
| } |
| else { |
| pdata[0] = EXT4; |
| pdata[1] = (unsigned char)(code & 0xff); |
| pdata[2] = (unsigned char)((code >> 8) & 0xff); |
| pdata[3] = (unsigned char)((code >> 16) & 0xff); |
| pdata[4] = (unsigned char)((code >> 24) & 0xff); |
| n = 5; |
| } |
| |
| if (_Pickler_Write(self, pdata, n) < 0) |
| goto error; |
| } |
| else { |
| gen_global: |
| if (parent == module) { |
| Py_INCREF(lastname); |
| Py_DECREF(global_name); |
| global_name = lastname; |
| } |
| if (self->proto >= 4) { |
| const char stack_global_op = STACK_GLOBAL; |
| |
| if (save(self, module_name, 0) < 0) |
| goto error; |
| if (save(self, global_name, 0) < 0) |
| goto error; |
| |
| if (_Pickler_Write(self, &stack_global_op, 1) < 0) |
| goto error; |
| } |
| else if (parent != module) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyObject *reduce_value = Py_BuildValue("(O(OO))", |
| st->getattr, parent, lastname); |
| status = save_reduce(self, reduce_value, NULL); |
| Py_DECREF(reduce_value); |
| if (status < 0) |
| goto error; |
| } |
| else { |
| /* Generate a normal global opcode if we are using a pickle |
| protocol < 4, or if the object is not registered in the |
| extension registry. */ |
| PyObject *encoded; |
| PyObject *(*unicode_encoder)(PyObject *); |
| |
| if (_Pickler_Write(self, &global_op, 1) < 0) |
| goto error; |
| |
| /* For protocol < 3 and if the user didn't request against doing |
| so, we convert module names to the old 2.x module names. */ |
| if (self->proto < 3 && self->fix_imports) { |
| if (fix_imports(&module_name, &global_name) < 0) { |
| goto error; |
| } |
| } |
| |
| /* Since Python 3.0 now supports non-ASCII identifiers, we encode |
| both the module name and the global name using UTF-8. We do so |
| only when we are using the pickle protocol newer than version |
| 3. This is to ensure compatibility with older Unpickler running |
| on Python 2.x. */ |
| if (self->proto == 3) { |
| unicode_encoder = PyUnicode_AsUTF8String; |
| } |
| else { |
| unicode_encoder = PyUnicode_AsASCIIString; |
| } |
| encoded = unicode_encoder(module_name); |
| if (encoded == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_UnicodeEncodeError)) |
| PyErr_Format(st->PicklingError, |
| "can't pickle module identifier '%S' using " |
| "pickle protocol %i", |
| module_name, self->proto); |
| goto error; |
| } |
| if (_Pickler_Write(self, PyBytes_AS_STRING(encoded), |
| PyBytes_GET_SIZE(encoded)) < 0) { |
| Py_DECREF(encoded); |
| goto error; |
| } |
| Py_DECREF(encoded); |
| if(_Pickler_Write(self, "\n", 1) < 0) |
| goto error; |
| |
| /* Save the name of the module. */ |
| encoded = unicode_encoder(global_name); |
| if (encoded == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_UnicodeEncodeError)) |
| PyErr_Format(st->PicklingError, |
| "can't pickle global identifier '%S' using " |
| "pickle protocol %i", |
| global_name, self->proto); |
| goto error; |
| } |
| if (_Pickler_Write(self, PyBytes_AS_STRING(encoded), |
| PyBytes_GET_SIZE(encoded)) < 0) { |
| Py_DECREF(encoded); |
| goto error; |
| } |
| Py_DECREF(encoded); |
| if (_Pickler_Write(self, "\n", 1) < 0) |
| goto error; |
| } |
| /* Memoize the object. */ |
| if (memo_put(self, obj) < 0) |
| goto error; |
| } |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| Py_XDECREF(module_name); |
| Py_XDECREF(global_name); |
| Py_XDECREF(module); |
| Py_XDECREF(parent); |
| Py_XDECREF(dotted_path); |
| Py_XDECREF(lastname); |
| |
| return status; |
| } |
| |
| static int |
| save_singleton_type(PicklerObject *self, PyObject *obj, PyObject *singleton) |
| { |
| PyObject *reduce_value; |
| int status; |
| |
| reduce_value = Py_BuildValue("O(O)", &PyType_Type, singleton); |
| if (reduce_value == NULL) { |
| return -1; |
| } |
| status = save_reduce(self, reduce_value, obj); |
| Py_DECREF(reduce_value); |
| return status; |
| } |
| |
| static int |
| save_type(PicklerObject *self, PyObject *obj) |
| { |
| if (obj == (PyObject *)&_PyNone_Type) { |
| return save_singleton_type(self, obj, Py_None); |
| } |
| else if (obj == (PyObject *)&PyEllipsis_Type) { |
| return save_singleton_type(self, obj, Py_Ellipsis); |
| } |
| else if (obj == (PyObject *)&_PyNotImplemented_Type) { |
| return save_singleton_type(self, obj, Py_NotImplemented); |
| } |
| return save_global(self, obj, NULL); |
| } |
| |
| static int |
| save_pers(PicklerObject *self, PyObject *obj, PyObject *func) |
| { |
| PyObject *pid = NULL; |
| int status = 0; |
| |
| const char persid_op = PERSID; |
| const char binpersid_op = BINPERSID; |
| |
| Py_INCREF(obj); |
| pid = _Pickle_FastCall(func, obj); |
| if (pid == NULL) |
| return -1; |
| |
| if (pid != Py_None) { |
| if (self->bin) { |
| if (save(self, pid, 1) < 0 || |
| _Pickler_Write(self, &binpersid_op, 1) < 0) |
| goto error; |
| } |
| else { |
| PyObject *pid_str = NULL; |
| char *pid_ascii_bytes; |
| Py_ssize_t size; |
| |
| pid_str = PyObject_Str(pid); |
| if (pid_str == NULL) |
| goto error; |
| |
| /* XXX: Should it check whether the persistent id only contains |
| ASCII characters? And what if the pid contains embedded |
| newlines? */ |
| pid_ascii_bytes = _PyUnicode_AsStringAndSize(pid_str, &size); |
| Py_DECREF(pid_str); |
| if (pid_ascii_bytes == NULL) |
| goto error; |
| |
| if (_Pickler_Write(self, &persid_op, 1) < 0 || |
| _Pickler_Write(self, pid_ascii_bytes, size) < 0 || |
| _Pickler_Write(self, "\n", 1) < 0) |
| goto error; |
| } |
| status = 1; |
| } |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| Py_XDECREF(pid); |
| |
| return status; |
| } |
| |
| static PyObject * |
| get_class(PyObject *obj) |
| { |
| PyObject *cls; |
| _Py_IDENTIFIER(__class__); |
| |
| cls = _PyObject_GetAttrId(obj, &PyId___class__); |
| if (cls == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) { |
| PyErr_Clear(); |
| cls = (PyObject *) Py_TYPE(obj); |
| Py_INCREF(cls); |
| } |
| } |
| return cls; |
| } |
| |
| /* We're saving obj, and args is the 2-thru-5 tuple returned by the |
| * appropriate __reduce__ method for obj. |
| */ |
| static int |
| save_reduce(PicklerObject *self, PyObject *args, PyObject *obj) |
| { |
| PyObject *callable; |
| PyObject *argtup; |
| PyObject *state = NULL; |
| PyObject *listitems = Py_None; |
| PyObject *dictitems = Py_None; |
| PickleState *st = _Pickle_GetGlobalState(); |
| Py_ssize_t size; |
| int use_newobj = 0, use_newobj_ex = 0; |
| |
| const char reduce_op = REDUCE; |
| const char build_op = BUILD; |
| const char newobj_op = NEWOBJ; |
| const char newobj_ex_op = NEWOBJ_EX; |
| |
| size = PyTuple_Size(args); |
| if (size < 2 || size > 5) { |
| PyErr_SetString(st->PicklingError, "tuple returned by " |
| "__reduce__ must contain 2 through 5 elements"); |
| return -1; |
| } |
| |
| if (!PyArg_UnpackTuple(args, "save_reduce", 2, 5, |
| &callable, &argtup, &state, &listitems, &dictitems)) |
| return -1; |
| |
| if (!PyCallable_Check(callable)) { |
| PyErr_SetString(st->PicklingError, "first item of the tuple " |
| "returned by __reduce__ must be callable"); |
| return -1; |
| } |
| if (!PyTuple_Check(argtup)) { |
| PyErr_SetString(st->PicklingError, "second item of the tuple " |
| "returned by __reduce__ must be a tuple"); |
| return -1; |
| } |
| |
| if (state == Py_None) |
| state = NULL; |
| |
| if (listitems == Py_None) |
| listitems = NULL; |
| else if (!PyIter_Check(listitems)) { |
| PyErr_Format(st->PicklingError, "fourth element of the tuple " |
| "returned by __reduce__ must be an iterator, not %s", |
| Py_TYPE(listitems)->tp_name); |
| return -1; |
| } |
| |
| if (dictitems == Py_None) |
| dictitems = NULL; |
| else if (!PyIter_Check(dictitems)) { |
| PyErr_Format(st->PicklingError, "fifth element of the tuple " |
| "returned by __reduce__ must be an iterator, not %s", |
| Py_TYPE(dictitems)->tp_name); |
| return -1; |
| } |
| |
| if (self->proto >= 2) { |
| PyObject *name; |
| _Py_IDENTIFIER(__name__); |
| |
| name = _PyObject_GetAttrId(callable, &PyId___name__); |
| if (name == NULL) { |
| if (!PyErr_ExceptionMatches(PyExc_AttributeError)) { |
| return -1; |
| } |
| PyErr_Clear(); |
| } |
| else if (PyUnicode_Check(name)) { |
| if (self->proto >= 4) { |
| _Py_IDENTIFIER(__newobj_ex__); |
| use_newobj_ex = PyUnicode_Compare( |
| name, _PyUnicode_FromId(&PyId___newobj_ex__)) == 0; |
| } |
| if (!use_newobj_ex) { |
| _Py_IDENTIFIER(__newobj__); |
| use_newobj = PyUnicode_Compare( |
| name, _PyUnicode_FromId(&PyId___newobj__)) == 0; |
| } |
| } |
| Py_XDECREF(name); |
| } |
| |
| if (use_newobj_ex) { |
| PyObject *cls; |
| PyObject *args; |
| PyObject *kwargs; |
| |
| if (Py_SIZE(argtup) != 3) { |
| PyErr_Format(st->PicklingError, |
| "length of the NEWOBJ_EX argument tuple must be " |
| "exactly 3, not %zd", Py_SIZE(argtup)); |
| return -1; |
| } |
| |
| cls = PyTuple_GET_ITEM(argtup, 0); |
| if (!PyType_Check(cls)) { |
| PyErr_Format(st->PicklingError, |
| "first item from NEWOBJ_EX argument tuple must " |
| "be a class, not %.200s", Py_TYPE(cls)->tp_name); |
| return -1; |
| } |
| args = PyTuple_GET_ITEM(argtup, 1); |
| if (!PyTuple_Check(args)) { |
| PyErr_Format(st->PicklingError, |
| "second item from NEWOBJ_EX argument tuple must " |
| "be a tuple, not %.200s", Py_TYPE(args)->tp_name); |
| return -1; |
| } |
| kwargs = PyTuple_GET_ITEM(argtup, 2); |
| if (!PyDict_Check(kwargs)) { |
| PyErr_Format(st->PicklingError, |
| "third item from NEWOBJ_EX argument tuple must " |
| "be a dict, not %.200s", Py_TYPE(kwargs)->tp_name); |
| return -1; |
| } |
| |
| if (save(self, cls, 0) < 0 || |
| save(self, args, 0) < 0 || |
| save(self, kwargs, 0) < 0 || |
| _Pickler_Write(self, &newobj_ex_op, 1) < 0) { |
| return -1; |
| } |
| } |
| else if (use_newobj) { |
| PyObject *cls; |
| PyObject *newargtup; |
| PyObject *obj_class; |
| int p; |
| |
| /* Sanity checks. */ |
| if (Py_SIZE(argtup) < 1) { |
| PyErr_SetString(st->PicklingError, "__newobj__ arglist is empty"); |
| return -1; |
| } |
| |
| cls = PyTuple_GET_ITEM(argtup, 0); |
| if (!PyType_Check(cls)) { |
| PyErr_SetString(st->PicklingError, "args[0] from " |
| "__newobj__ args is not a type"); |
| return -1; |
| } |
| |
| if (obj != NULL) { |
| obj_class = get_class(obj); |
| p = obj_class != cls; /* true iff a problem */ |
| Py_DECREF(obj_class); |
| if (p) { |
| PyErr_SetString(st->PicklingError, "args[0] from " |
| "__newobj__ args has the wrong class"); |
| return -1; |
| } |
| } |
| /* XXX: These calls save() are prone to infinite recursion. Imagine |
| what happen if the value returned by the __reduce__() method of |
| some extension type contains another object of the same type. Ouch! |
| |
| Here is a quick example, that I ran into, to illustrate what I |
| mean: |
| |
| >>> import pickle, copyreg |
| >>> copyreg.dispatch_table.pop(complex) |
| >>> pickle.dumps(1+2j) |
| Traceback (most recent call last): |
| ... |
| RecursionError: maximum recursion depth exceeded |
| |
| Removing the complex class from copyreg.dispatch_table made the |
| __reduce_ex__() method emit another complex object: |
| |
| >>> (1+1j).__reduce_ex__(2) |
| (<function __newobj__ at 0xb7b71c3c>, |
| (<class 'complex'>, (1+1j)), None, None, None) |
| |
| Thus when save() was called on newargstup (the 2nd item) recursion |
| ensued. Of course, the bug was in the complex class which had a |
| broken __getnewargs__() that emitted another complex object. But, |
| the point, here, is it is quite easy to end up with a broken reduce |
| function. */ |
| |
| /* Save the class and its __new__ arguments. */ |
| if (save(self, cls, 0) < 0) |
| return -1; |
| |
| newargtup = PyTuple_GetSlice(argtup, 1, Py_SIZE(argtup)); |
| if (newargtup == NULL) |
| return -1; |
| |
| p = save(self, newargtup, 0); |
| Py_DECREF(newargtup); |
| if (p < 0) |
| return -1; |
| |
| /* Add NEWOBJ opcode. */ |
| if (_Pickler_Write(self, &newobj_op, 1) < 0) |
| return -1; |
| } |
| else { /* Not using NEWOBJ. */ |
| if (save(self, callable, 0) < 0 || |
| save(self, argtup, 0) < 0 || |
| _Pickler_Write(self, &reduce_op, 1) < 0) |
| return -1; |
| } |
| |
| /* obj can be NULL when save_reduce() is used directly. A NULL obj means |
| the caller do not want to memoize the object. Not particularly useful, |
| but that is to mimic the behavior save_reduce() in pickle.py when |
| obj is None. */ |
| if (obj != NULL) { |
| /* If the object is already in the memo, this means it is |
| recursive. In this case, throw away everything we put on the |
| stack, and fetch the object back from the memo. */ |
| if (PyMemoTable_Get(self->memo, obj)) { |
| const char pop_op = POP; |
| |
| if (_Pickler_Write(self, &pop_op, 1) < 0) |
| return -1; |
| if (memo_get(self, obj) < 0) |
| return -1; |
| |
| return 0; |
| } |
| else if (memo_put(self, obj) < 0) |
| return -1; |
| } |
| |
| if (listitems && batch_list(self, listitems) < 0) |
| return -1; |
| |
| if (dictitems && batch_dict(self, dictitems) < 0) |
| return -1; |
| |
| if (state) { |
| if (save(self, state, 0) < 0 || |
| _Pickler_Write(self, &build_op, 1) < 0) |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| save(PicklerObject *self, PyObject *obj, int pers_save) |
| { |
| PyTypeObject *type; |
| PyObject *reduce_func = NULL; |
| PyObject *reduce_value = NULL; |
| int status = 0; |
| |
| if (_Pickler_OpcodeBoundary(self) < 0) |
| return -1; |
| |
| if (Py_EnterRecursiveCall(" while pickling an object")) |
| return -1; |
| |
| /* The extra pers_save argument is necessary to avoid calling save_pers() |
| on its returned object. */ |
| if (!pers_save && self->pers_func) { |
| /* save_pers() returns: |
| -1 to signal an error; |
| 0 if it did nothing successfully; |
| 1 if a persistent id was saved. |
| */ |
| if ((status = save_pers(self, obj, self->pers_func)) != 0) |
| goto done; |
| } |
| |
| type = Py_TYPE(obj); |
| |
| /* The old cPickle had an optimization that used switch-case statement |
| dispatching on the first letter of the type name. This has was removed |
| since benchmarks shown that this optimization was actually slowing |
| things down. */ |
| |
| /* Atom types; these aren't memoized, so don't check the memo. */ |
| |
| if (obj == Py_None) { |
| status = save_none(self, obj); |
| goto done; |
| } |
| else if (obj == Py_False || obj == Py_True) { |
| status = save_bool(self, obj); |
| goto done; |
| } |
| else if (type == &PyLong_Type) { |
| status = save_long(self, obj); |
| goto done; |
| } |
| else if (type == &PyFloat_Type) { |
| status = save_float(self, obj); |
| goto done; |
| } |
| |
| /* Check the memo to see if it has the object. If so, generate |
| a GET (or BINGET) opcode, instead of pickling the object |
| once again. */ |
| if (PyMemoTable_Get(self->memo, obj)) { |
| if (memo_get(self, obj) < 0) |
| goto error; |
| goto done; |
| } |
| |
| if (type == &PyBytes_Type) { |
| status = save_bytes(self, obj); |
| goto done; |
| } |
| else if (type == &PyUnicode_Type) { |
| status = save_unicode(self, obj); |
| goto done; |
| } |
| else if (type == &PyDict_Type) { |
| status = save_dict(self, obj); |
| goto done; |
| } |
| else if (type == &PySet_Type) { |
| status = save_set(self, obj); |
| goto done; |
| } |
| else if (type == &PyFrozenSet_Type) { |
| status = save_frozenset(self, obj); |
| goto done; |
| } |
| else if (type == &PyList_Type) { |
| status = save_list(self, obj); |
| goto done; |
| } |
| else if (type == &PyTuple_Type) { |
| status = save_tuple(self, obj); |
| goto done; |
| } |
| else if (type == &PyType_Type) { |
| status = save_type(self, obj); |
| goto done; |
| } |
| else if (type == &PyFunction_Type) { |
| status = save_global(self, obj, NULL); |
| goto done; |
| } |
| |
| /* XXX: This part needs some unit tests. */ |
| |
| /* Get a reduction callable, and call it. This may come from |
| * self.dispatch_table, copyreg.dispatch_table, the object's |
| * __reduce_ex__ method, or the object's __reduce__ method. |
| */ |
| if (self->dispatch_table == NULL) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| reduce_func = PyDict_GetItemWithError(st->dispatch_table, |
| (PyObject *)type); |
| if (reduce_func == NULL) { |
| if (PyErr_Occurred()) { |
| goto error; |
| } |
| } else { |
| /* PyDict_GetItemWithError() returns a borrowed reference. |
| Increase the reference count to be consistent with |
| PyObject_GetItem and _PyObject_GetAttrId used below. */ |
| Py_INCREF(reduce_func); |
| } |
| } else { |
| reduce_func = PyObject_GetItem(self->dispatch_table, |
| (PyObject *)type); |
| if (reduce_func == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_KeyError)) |
| PyErr_Clear(); |
| else |
| goto error; |
| } |
| } |
| if (reduce_func != NULL) { |
| Py_INCREF(obj); |
| reduce_value = _Pickle_FastCall(reduce_func, obj); |
| } |
| else if (PyType_IsSubtype(type, &PyType_Type)) { |
| status = save_global(self, obj, NULL); |
| goto done; |
| } |
| else { |
| _Py_IDENTIFIER(__reduce__); |
| _Py_IDENTIFIER(__reduce_ex__); |
| |
| |
| /* XXX: If the __reduce__ method is defined, __reduce_ex__ is |
| automatically defined as __reduce__. While this is convenient, this |
| make it impossible to know which method was actually called. Of |
| course, this is not a big deal. But still, it would be nice to let |
| the user know which method was called when something go |
| wrong. Incidentally, this means if __reduce_ex__ is not defined, we |
| don't actually have to check for a __reduce__ method. */ |
| |
| /* Check for a __reduce_ex__ method. */ |
| reduce_func = _PyObject_GetAttrId(obj, &PyId___reduce_ex__); |
| if (reduce_func != NULL) { |
| PyObject *proto; |
| proto = PyLong_FromLong(self->proto); |
| if (proto != NULL) { |
| reduce_value = _Pickle_FastCall(reduce_func, proto); |
| } |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) { |
| PyErr_Clear(); |
| } |
| else { |
| goto error; |
| } |
| /* Check for a __reduce__ method. */ |
| reduce_func = _PyObject_GetAttrId(obj, &PyId___reduce__); |
| if (reduce_func != NULL) { |
| PyObject *empty_tuple = PyTuple_New(0); |
| reduce_value = PyObject_Call(reduce_func, empty_tuple, |
| NULL); |
| Py_DECREF(empty_tuple); |
| } |
| else { |
| PyErr_Format(st->PicklingError, |
| "can't pickle '%.200s' object: %R", |
| type->tp_name, obj); |
| goto error; |
| } |
| } |
| } |
| |
| if (reduce_value == NULL) |
| goto error; |
| |
| if (PyUnicode_Check(reduce_value)) { |
| status = save_global(self, obj, reduce_value); |
| goto done; |
| } |
| |
| if (!PyTuple_Check(reduce_value)) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->PicklingError, |
| "__reduce__ must return a string or tuple"); |
| goto error; |
| } |
| |
| status = save_reduce(self, reduce_value, obj); |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| done: |
| |
| Py_LeaveRecursiveCall(); |
| Py_XDECREF(reduce_func); |
| Py_XDECREF(reduce_value); |
| |
| return status; |
| } |
| |
| static int |
| dump(PicklerObject *self, PyObject *obj) |
| { |
| const char stop_op = STOP; |
| |
| if (self->proto >= 2) { |
| char header[2]; |
| |
| header[0] = PROTO; |
| assert(self->proto >= 0 && self->proto < 256); |
| header[1] = (unsigned char)self->proto; |
| if (_Pickler_Write(self, header, 2) < 0) |
| return -1; |
| if (self->proto >= 4) |
| self->framing = 1; |
| } |
| |
| if (save(self, obj, 0) < 0 || |
| _Pickler_Write(self, &stop_op, 1) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Pickler.clear_memo |
| |
| Clears the pickler's "memo". |
| |
| The memo is the data structure that remembers which objects the |
| pickler has already seen, so that shared or recursive objects are |
| pickled by reference and not by value. This method is useful when |
| re-using picklers. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_Pickler_clear_memo_impl(PicklerObject *self) |
| /*[clinic end generated code: output=8665c8658aaa094b input=01bdad52f3d93e56]*/ |
| { |
| if (self->memo) |
| PyMemoTable_Clear(self->memo); |
| |
| Py_RETURN_NONE; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Pickler.dump |
| |
| obj: object |
| / |
| |
| Write a pickled representation of the given object to the open file. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_Pickler_dump(PicklerObject *self, PyObject *obj) |
| /*[clinic end generated code: output=87ecad1261e02ac7 input=552eb1c0f52260d9]*/ |
| { |
| /* Check whether the Pickler was initialized correctly (issue3664). |
| Developers often forget to call __init__() in their subclasses, which |
| would trigger a segfault without this check. */ |
| if (self->write == NULL) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_Format(st->PicklingError, |
| "Pickler.__init__() was not called by %s.__init__()", |
| Py_TYPE(self)->tp_name); |
| return NULL; |
| } |
| |
| if (_Pickler_ClearBuffer(self) < 0) |
| return NULL; |
| |
| if (dump(self, obj) < 0) |
| return NULL; |
| |
| if (_Pickler_FlushToFile(self) < 0) |
| return NULL; |
| |
| Py_RETURN_NONE; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Pickler.__sizeof__ -> Py_ssize_t |
| |
| Returns size in memory, in bytes. |
| [clinic start generated code]*/ |
| |
| static Py_ssize_t |
| _pickle_Pickler___sizeof___impl(PicklerObject *self) |
| /*[clinic end generated code: output=106edb3123f332e1 input=8cbbec9bd5540d42]*/ |
| { |
| Py_ssize_t res, s; |
| |
| res = _PyObject_SIZE(Py_TYPE(self)); |
| if (self->memo != NULL) { |
| res += sizeof(PyMemoTable); |
| res += self->memo->mt_allocated * sizeof(PyMemoEntry); |
| } |
| if (self->output_buffer != NULL) { |
| s = _PySys_GetSizeOf(self->output_buffer); |
| if (s == -1) |
| return -1; |
| res += s; |
| } |
| return res; |
| } |
| |
| static struct PyMethodDef Pickler_methods[] = { |
| _PICKLE_PICKLER_DUMP_METHODDEF |
| _PICKLE_PICKLER_CLEAR_MEMO_METHODDEF |
| _PICKLE_PICKLER___SIZEOF___METHODDEF |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| static void |
| Pickler_dealloc(PicklerObject *self) |
| { |
| PyObject_GC_UnTrack(self); |
| |
| Py_XDECREF(self->output_buffer); |
| Py_XDECREF(self->write); |
| Py_XDECREF(self->pers_func); |
| Py_XDECREF(self->dispatch_table); |
| Py_XDECREF(self->fast_memo); |
| |
| PyMemoTable_Del(self->memo); |
| |
| Py_TYPE(self)->tp_free((PyObject *)self); |
| } |
| |
| static int |
| Pickler_traverse(PicklerObject *self, visitproc visit, void *arg) |
| { |
| Py_VISIT(self->write); |
| Py_VISIT(self->pers_func); |
| Py_VISIT(self->dispatch_table); |
| Py_VISIT(self->fast_memo); |
| return 0; |
| } |
| |
| static int |
| Pickler_clear(PicklerObject *self) |
| { |
| Py_CLEAR(self->output_buffer); |
| Py_CLEAR(self->write); |
| Py_CLEAR(self->pers_func); |
| Py_CLEAR(self->dispatch_table); |
| Py_CLEAR(self->fast_memo); |
| |
| if (self->memo != NULL) { |
| PyMemoTable *memo = self->memo; |
| self->memo = NULL; |
| PyMemoTable_Del(memo); |
| } |
| return 0; |
| } |
| |
| |
| /*[clinic input] |
| |
| _pickle.Pickler.__init__ |
| |
| file: object |
| protocol: object = NULL |
| fix_imports: bool = True |
| |
| This takes a binary file for writing a pickle data stream. |
| |
| The optional *protocol* argument tells the pickler to use the given |
| protocol; supported protocols are 0, 1, 2, 3 and 4. The default |
| protocol is 3; a backward-incompatible protocol designed for Python 3. |
| |
| Specifying a negative protocol version selects the highest protocol |
| version supported. The higher the protocol used, the more recent the |
| version of Python needed to read the pickle produced. |
| |
| The *file* argument must have a write() method that accepts a single |
| bytes argument. It can thus be a file object opened for binary |
| writing, an io.BytesIO instance, or any other custom object that meets |
| this interface. |
| |
| If *fix_imports* is True and protocol is less than 3, pickle will try |
| to map the new Python 3 names to the old module names used in Python |
| 2, so that the pickle data stream is readable with Python 2. |
| [clinic start generated code]*/ |
| |
| static int |
| _pickle_Pickler___init___impl(PicklerObject *self, PyObject *file, |
| PyObject *protocol, int fix_imports) |
| /*[clinic end generated code: output=b5f31078dab17fb0 input=4faabdbc763c2389]*/ |
| { |
| _Py_IDENTIFIER(persistent_id); |
| _Py_IDENTIFIER(dispatch_table); |
| |
| /* In case of multiple __init__() calls, clear previous content. */ |
| if (self->write != NULL) |
| (void)Pickler_clear(self); |
| |
| if (_Pickler_SetProtocol(self, protocol, fix_imports) < 0) |
| return -1; |
| |
| if (_Pickler_SetOutputStream(self, file) < 0) |
| return -1; |
| |
| /* memo and output_buffer may have already been created in _Pickler_New */ |
| if (self->memo == NULL) { |
| self->memo = PyMemoTable_New(); |
| if (self->memo == NULL) |
| return -1; |
| } |
| self->output_len = 0; |
| if (self->output_buffer == NULL) { |
| self->max_output_len = WRITE_BUF_SIZE; |
| self->output_buffer = PyBytes_FromStringAndSize(NULL, |
| self->max_output_len); |
| if (self->output_buffer == NULL) |
| return -1; |
| } |
| |
| self->fast = 0; |
| self->fast_nesting = 0; |
| self->fast_memo = NULL; |
| self->pers_func = NULL; |
| if (_PyObject_HasAttrId((PyObject *)self, &PyId_persistent_id)) { |
| self->pers_func = _PyObject_GetAttrId((PyObject *)self, |
| &PyId_persistent_id); |
| if (self->pers_func == NULL) |
| return -1; |
| } |
| self->dispatch_table = NULL; |
| if (_PyObject_HasAttrId((PyObject *)self, &PyId_dispatch_table)) { |
| self->dispatch_table = _PyObject_GetAttrId((PyObject *)self, |
| &PyId_dispatch_table); |
| if (self->dispatch_table == NULL) |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| |
| /* Define a proxy object for the Pickler's internal memo object. This is to |
| * avoid breaking code like: |
| * pickler.memo.clear() |
| * and |
| * pickler.memo = saved_memo |
| * Is this a good idea? Not really, but we don't want to break code that uses |
| * it. Note that we don't implement the entire mapping API here. This is |
| * intentional, as these should be treated as black-box implementation details. |
| */ |
| |
| /*[clinic input] |
| _pickle.PicklerMemoProxy.clear |
| |
| Remove all items from memo. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_PicklerMemoProxy_clear_impl(PicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=5fb9370d48ae8b05 input=ccc186dacd0f1405]*/ |
| { |
| if (self->pickler->memo) |
| PyMemoTable_Clear(self->pickler->memo); |
| Py_RETURN_NONE; |
| } |
| |
| /*[clinic input] |
| _pickle.PicklerMemoProxy.copy |
| |
| Copy the memo to a new object. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_PicklerMemoProxy_copy_impl(PicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=bb83a919d29225ef input=b73043485ac30b36]*/ |
| { |
| Py_ssize_t i; |
| PyMemoTable *memo; |
| PyObject *new_memo = PyDict_New(); |
| if (new_memo == NULL) |
| return NULL; |
| |
| memo = self->pickler->memo; |
| for (i = 0; i < memo->mt_allocated; ++i) { |
| PyMemoEntry entry = memo->mt_table[i]; |
| if (entry.me_key != NULL) { |
| int status; |
| PyObject *key, *value; |
| |
| key = PyLong_FromVoidPtr(entry.me_key); |
| value = Py_BuildValue("nO", entry.me_value, entry.me_key); |
| |
| if (key == NULL || value == NULL) { |
| Py_XDECREF(key); |
| Py_XDECREF(value); |
| goto error; |
| } |
| status = PyDict_SetItem(new_memo, key, value); |
| Py_DECREF(key); |
| Py_DECREF(value); |
| if (status < 0) |
| goto error; |
| } |
| } |
| return new_memo; |
| |
| error: |
| Py_XDECREF(new_memo); |
| return NULL; |
| } |
| |
| /*[clinic input] |
| _pickle.PicklerMemoProxy.__reduce__ |
| |
| Implement pickle support. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_PicklerMemoProxy___reduce___impl(PicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=bebba1168863ab1d input=2f7c540e24b7aae4]*/ |
| { |
| PyObject *reduce_value, *dict_args; |
| PyObject *contents = _pickle_PicklerMemoProxy_copy_impl(self); |
| if (contents == NULL) |
| return NULL; |
| |
| reduce_value = PyTuple_New(2); |
| if (reduce_value == NULL) { |
| Py_DECREF(contents); |
| return NULL; |
| } |
| dict_args = PyTuple_New(1); |
| if (dict_args == NULL) { |
| Py_DECREF(contents); |
| Py_DECREF(reduce_value); |
| return NULL; |
| } |
| PyTuple_SET_ITEM(dict_args, 0, contents); |
| Py_INCREF((PyObject *)&PyDict_Type); |
| PyTuple_SET_ITEM(reduce_value, 0, (PyObject *)&PyDict_Type); |
| PyTuple_SET_ITEM(reduce_value, 1, dict_args); |
| return reduce_value; |
| } |
| |
| static PyMethodDef picklerproxy_methods[] = { |
| _PICKLE_PICKLERMEMOPROXY_CLEAR_METHODDEF |
| _PICKLE_PICKLERMEMOPROXY_COPY_METHODDEF |
| _PICKLE_PICKLERMEMOPROXY___REDUCE___METHODDEF |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| static void |
| PicklerMemoProxy_dealloc(PicklerMemoProxyObject *self) |
| { |
| PyObject_GC_UnTrack(self); |
| Py_XDECREF(self->pickler); |
| PyObject_GC_Del((PyObject *)self); |
| } |
| |
| static int |
| PicklerMemoProxy_traverse(PicklerMemoProxyObject *self, |
| visitproc visit, void *arg) |
| { |
| Py_VISIT(self->pickler); |
| return 0; |
| } |
| |
| static int |
| PicklerMemoProxy_clear(PicklerMemoProxyObject *self) |
| { |
| Py_CLEAR(self->pickler); |
| return 0; |
| } |
| |
| static PyTypeObject PicklerMemoProxyType = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "_pickle.PicklerMemoProxy", /*tp_name*/ |
| sizeof(PicklerMemoProxyObject), /*tp_basicsize*/ |
| 0, |
| (destructor)PicklerMemoProxy_dealloc, /* tp_dealloc */ |
| 0, /* tp_print */ |
| 0, /* tp_getattr */ |
| 0, /* tp_setattr */ |
| 0, /* tp_compare */ |
| 0, /* tp_repr */ |
| 0, /* tp_as_number */ |
| 0, /* tp_as_sequence */ |
| 0, /* tp_as_mapping */ |
| PyObject_HashNotImplemented, /* tp_hash */ |
| 0, /* tp_call */ |
| 0, /* tp_str */ |
| PyObject_GenericGetAttr, /* tp_getattro */ |
| PyObject_GenericSetAttr, /* tp_setattro */ |
| 0, /* tp_as_buffer */ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC, |
| 0, /* tp_doc */ |
| (traverseproc)PicklerMemoProxy_traverse, /* tp_traverse */ |
| (inquiry)PicklerMemoProxy_clear, /* tp_clear */ |
| 0, /* tp_richcompare */ |
| 0, /* tp_weaklistoffset */ |
| 0, /* tp_iter */ |
| 0, /* tp_iternext */ |
| picklerproxy_methods, /* tp_methods */ |
| }; |
| |
| static PyObject * |
| PicklerMemoProxy_New(PicklerObject *pickler) |
| { |
| PicklerMemoProxyObject *self; |
| |
| self = PyObject_GC_New(PicklerMemoProxyObject, &PicklerMemoProxyType); |
| if (self == NULL) |
| return NULL; |
| Py_INCREF(pickler); |
| self->pickler = pickler; |
| PyObject_GC_Track(self); |
| return (PyObject *)self; |
| } |
| |
| /*****************************************************************************/ |
| |
| static PyObject * |
| Pickler_get_memo(PicklerObject *self) |
| { |
| return PicklerMemoProxy_New(self); |
| } |
| |
| static int |
| Pickler_set_memo(PicklerObject *self, PyObject *obj) |
| { |
| PyMemoTable *new_memo = NULL; |
| |
| if (obj == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "attribute deletion is not supported"); |
| return -1; |
| } |
| |
| if (Py_TYPE(obj) == &PicklerMemoProxyType) { |
| PicklerObject *pickler = |
| ((PicklerMemoProxyObject *)obj)->pickler; |
| |
| new_memo = PyMemoTable_Copy(pickler->memo); |
| if (new_memo == NULL) |
| return -1; |
| } |
| else if (PyDict_Check(obj)) { |
| Py_ssize_t i = 0; |
| PyObject *key, *value; |
| |
| new_memo = PyMemoTable_New(); |
| if (new_memo == NULL) |
| return -1; |
| |
| while (PyDict_Next(obj, &i, &key, &value)) { |
| Py_ssize_t memo_id; |
| PyObject *memo_obj; |
| |
| if (!PyTuple_Check(value) || Py_SIZE(value) != 2) { |
| PyErr_SetString(PyExc_TypeError, |
| "'memo' values must be 2-item tuples"); |
| goto error; |
| } |
| memo_id = PyLong_AsSsize_t(PyTuple_GET_ITEM(value, 0)); |
| if (memo_id == -1 && PyErr_Occurred()) |
| goto error; |
| memo_obj = PyTuple_GET_ITEM(value, 1); |
| if (PyMemoTable_Set(new_memo, memo_obj, memo_id) < 0) |
| goto error; |
| } |
| } |
| else { |
| PyErr_Format(PyExc_TypeError, |
| "'memo' attribute must be an PicklerMemoProxy object" |
| "or dict, not %.200s", Py_TYPE(obj)->tp_name); |
| return -1; |
| } |
| |
| PyMemoTable_Del(self->memo); |
| self->memo = new_memo; |
| |
| return 0; |
| |
| error: |
| if (new_memo) |
| PyMemoTable_Del(new_memo); |
| return -1; |
| } |
| |
| static PyObject * |
| Pickler_get_persid(PicklerObject *self) |
| { |
| if (self->pers_func == NULL) |
| PyErr_SetString(PyExc_AttributeError, "persistent_id"); |
| else |
| Py_INCREF(self->pers_func); |
| return self->pers_func; |
| } |
| |
| static int |
| Pickler_set_persid(PicklerObject *self, PyObject *value) |
| { |
| PyObject *tmp; |
| |
| if (value == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "attribute deletion is not supported"); |
| return -1; |
| } |
| if (!PyCallable_Check(value)) { |
| PyErr_SetString(PyExc_TypeError, |
| "persistent_id must be a callable taking one argument"); |
| return -1; |
| } |
| |
| tmp = self->pers_func; |
| Py_INCREF(value); |
| self->pers_func = value; |
| Py_XDECREF(tmp); /* self->pers_func can be NULL, so be careful. */ |
| |
| return 0; |
| } |
| |
| static PyMemberDef Pickler_members[] = { |
| {"bin", T_INT, offsetof(PicklerObject, bin)}, |
| {"fast", T_INT, offsetof(PicklerObject, fast)}, |
| {"dispatch_table", T_OBJECT_EX, offsetof(PicklerObject, dispatch_table)}, |
| {NULL} |
| }; |
| |
| static PyGetSetDef Pickler_getsets[] = { |
| {"memo", (getter)Pickler_get_memo, |
| (setter)Pickler_set_memo}, |
| {"persistent_id", (getter)Pickler_get_persid, |
| (setter)Pickler_set_persid}, |
| {NULL} |
| }; |
| |
| static PyTypeObject Pickler_Type = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "_pickle.Pickler" , /*tp_name*/ |
| sizeof(PicklerObject), /*tp_basicsize*/ |
| 0, /*tp_itemsize*/ |
| (destructor)Pickler_dealloc, /*tp_dealloc*/ |
| 0, /*tp_print*/ |
| 0, /*tp_getattr*/ |
| 0, /*tp_setattr*/ |
| 0, /*tp_reserved*/ |
| 0, /*tp_repr*/ |
| 0, /*tp_as_number*/ |
| 0, /*tp_as_sequence*/ |
| 0, /*tp_as_mapping*/ |
| 0, /*tp_hash*/ |
| 0, /*tp_call*/ |
| 0, /*tp_str*/ |
| 0, /*tp_getattro*/ |
| 0, /*tp_setattro*/ |
| 0, /*tp_as_buffer*/ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC, |
| _pickle_Pickler___init____doc__, /*tp_doc*/ |
| (traverseproc)Pickler_traverse, /*tp_traverse*/ |
| (inquiry)Pickler_clear, /*tp_clear*/ |
| 0, /*tp_richcompare*/ |
| 0, /*tp_weaklistoffset*/ |
| 0, /*tp_iter*/ |
| 0, /*tp_iternext*/ |
| Pickler_methods, /*tp_methods*/ |
| Pickler_members, /*tp_members*/ |
| Pickler_getsets, /*tp_getset*/ |
| 0, /*tp_base*/ |
| 0, /*tp_dict*/ |
| 0, /*tp_descr_get*/ |
| 0, /*tp_descr_set*/ |
| 0, /*tp_dictoffset*/ |
| _pickle_Pickler___init__, /*tp_init*/ |
| PyType_GenericAlloc, /*tp_alloc*/ |
| PyType_GenericNew, /*tp_new*/ |
| PyObject_GC_Del, /*tp_free*/ |
| 0, /*tp_is_gc*/ |
| }; |
| |
| /* Temporary helper for calling self.find_class(). |
| |
| XXX: It would be nice to able to avoid Python function call overhead, by |
| using directly the C version of find_class(), when find_class() is not |
| overridden by a subclass. Although, this could become rather hackish. A |
| simpler optimization would be to call the C function when self is not a |
| subclass instance. */ |
| static PyObject * |
| find_class(UnpicklerObject *self, PyObject *module_name, PyObject *global_name) |
| { |
| _Py_IDENTIFIER(find_class); |
| |
| return _PyObject_CallMethodId((PyObject *)self, &PyId_find_class, "OO", |
| module_name, global_name); |
| } |
| |
| static Py_ssize_t |
| marker(UnpicklerObject *self) |
| { |
| PickleState *st = _Pickle_GetGlobalState(); |
| if (self->num_marks < 1) { |
| PyErr_SetString(st->UnpicklingError, "could not find MARK"); |
| return -1; |
| } |
| |
| return self->marks[--self->num_marks]; |
| } |
| |
| static int |
| load_none(UnpicklerObject *self) |
| { |
| PDATA_APPEND(self->stack, Py_None, -1); |
| return 0; |
| } |
| |
| static int |
| bad_readline(void) |
| { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, "pickle data was truncated"); |
| return -1; |
| } |
| |
| static int |
| load_int(UnpicklerObject *self) |
| { |
| PyObject *value; |
| char *endptr, *s; |
| Py_ssize_t len; |
| long x; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| |
| errno = 0; |
| /* XXX: Should the base argument of strtol() be explicitly set to 10? |
| XXX(avassalotti): Should this uses PyOS_strtol()? */ |
| x = strtol(s, &endptr, 0); |
| |
| if (errno || (*endptr != '\n' && *endptr != '\0')) { |
| /* Hm, maybe we've got something long. Let's try reading |
| * it as a Python int object. */ |
| errno = 0; |
| /* XXX: Same thing about the base here. */ |
| value = PyLong_FromString(s, NULL, 0); |
| if (value == NULL) { |
| PyErr_SetString(PyExc_ValueError, |
| "could not convert string to int"); |
| return -1; |
| } |
| } |
| else { |
| if (len == 3 && (x == 0 || x == 1)) { |
| if ((value = PyBool_FromLong(x)) == NULL) |
| return -1; |
| } |
| else { |
| if ((value = PyLong_FromLong(x)) == NULL) |
| return -1; |
| } |
| } |
| |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_bool(UnpicklerObject *self, PyObject *boolean) |
| { |
| assert(boolean == Py_True || boolean == Py_False); |
| PDATA_APPEND(self->stack, boolean, -1); |
| return 0; |
| } |
| |
| /* s contains x bytes of an unsigned little-endian integer. Return its value |
| * as a C Py_ssize_t, or -1 if it's higher than PY_SSIZE_T_MAX. |
| */ |
| static Py_ssize_t |
| calc_binsize(char *bytes, int nbytes) |
| { |
| unsigned char *s = (unsigned char *)bytes; |
| int i; |
| size_t x = 0; |
| |
| if (nbytes > (int)sizeof(size_t)) { |
| /* Check for integer overflow. BINBYTES8 and BINUNICODE8 opcodes |
| * have 64-bit size that can't be represented on 32-bit platform. |
| */ |
| for (i = (int)sizeof(size_t); i < nbytes; i++) { |
| if (s[i]) |
| return -1; |
| } |
| nbytes = (int)sizeof(size_t); |
| } |
| for (i = 0; i < nbytes; i++) { |
| x |= (size_t) s[i] << (8 * i); |
| } |
| |
| if (x > PY_SSIZE_T_MAX) |
| return -1; |
| else |
| return (Py_ssize_t) x; |
| } |
| |
| /* s contains x bytes of a little-endian integer. Return its value as a |
| * C int. Obscure: when x is 1 or 2, this is an unsigned little-endian |
| * int, but when x is 4 it's a signed one. This is a historical source |
| * of x-platform bugs. |
| */ |
| static long |
| calc_binint(char *bytes, int nbytes) |
| { |
| unsigned char *s = (unsigned char *)bytes; |
| Py_ssize_t i; |
| long x = 0; |
| |
| for (i = 0; i < nbytes; i++) { |
| x |= (long)s[i] << (8 * i); |
| } |
| |
| /* Unlike BININT1 and BININT2, BININT (more accurately BININT4) |
| * is signed, so on a box with longs bigger than 4 bytes we need |
| * to extend a BININT's sign bit to the full width. |
| */ |
| if (SIZEOF_LONG > 4 && nbytes == 4) { |
| x |= -(x & (1L << 31)); |
| } |
| |
| return x; |
| } |
| |
| static int |
| load_binintx(UnpicklerObject *self, char *s, int size) |
| { |
| PyObject *value; |
| long x; |
| |
| x = calc_binint(s, size); |
| |
| if ((value = PyLong_FromLong(x)) == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_binint(UnpicklerObject *self) |
| { |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 4) < 0) |
| return -1; |
| |
| return load_binintx(self, s, 4); |
| } |
| |
| static int |
| load_binint1(UnpicklerObject *self) |
| { |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 1) < 0) |
| return -1; |
| |
| return load_binintx(self, s, 1); |
| } |
| |
| static int |
| load_binint2(UnpicklerObject *self) |
| { |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 2) < 0) |
| return -1; |
| |
| return load_binintx(self, s, 2); |
| } |
| |
| static int |
| load_long(UnpicklerObject *self) |
| { |
| PyObject *value; |
| char *s; |
| Py_ssize_t len; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| |
| /* s[len-2] will usually be 'L' (and s[len-1] is '\n'); we need to remove |
| the 'L' before calling PyLong_FromString. In order to maintain |
| compatibility with Python 3.0.0, we don't actually *require* |
| the 'L' to be present. */ |
| if (s[len-2] == 'L') |
| s[len-2] = '\0'; |
| /* XXX: Should the base argument explicitly set to 10? */ |
| value = PyLong_FromString(s, NULL, 0); |
| if (value == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| /* 'size' bytes contain the # of bytes of little-endian 256's-complement |
| * data following. |
| */ |
| static int |
| load_counted_long(UnpicklerObject *self, int size) |
| { |
| PyObject *value; |
| char *nbytes; |
| char *pdata; |
| |
| assert(size == 1 || size == 4); |
| if (_Unpickler_Read(self, &nbytes, size) < 0) |
| return -1; |
| |
| size = calc_binint(nbytes, size); |
| if (size < 0) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| /* Corrupt or hostile pickle -- we never write one like this */ |
| PyErr_SetString(st->UnpicklingError, |
| "LONG pickle has negative byte count"); |
| return -1; |
| } |
| |
| if (size == 0) |
| value = PyLong_FromLong(0L); |
| else { |
| /* Read the raw little-endian bytes and convert. */ |
| if (_Unpickler_Read(self, &pdata, size) < 0) |
| return -1; |
| value = _PyLong_FromByteArray((unsigned char *)pdata, (size_t)size, |
| 1 /* little endian */ , 1 /* signed */ ); |
| } |
| if (value == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_float(UnpicklerObject *self) |
| { |
| PyObject *value; |
| char *endptr, *s; |
| Py_ssize_t len; |
| double d; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| |
| errno = 0; |
| d = PyOS_string_to_double(s, &endptr, PyExc_OverflowError); |
| if (d == -1.0 && PyErr_Occurred()) |
| return -1; |
| if ((endptr[0] != '\n') && (endptr[0] != '\0')) { |
| PyErr_SetString(PyExc_ValueError, "could not convert string to float"); |
| return -1; |
| } |
| value = PyFloat_FromDouble(d); |
| if (value == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_binfloat(UnpicklerObject *self) |
| { |
| PyObject *value; |
| double x; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 8) < 0) |
| return -1; |
| |
| x = _PyFloat_Unpack8((unsigned char *)s, 0); |
| if (x == -1.0 && PyErr_Occurred()) |
| return -1; |
| |
| if ((value = PyFloat_FromDouble(x)) == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_string(UnpicklerObject *self) |
| { |
| PyObject *bytes; |
| PyObject *obj; |
| Py_ssize_t len; |
| char *s, *p; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| /* Strip the newline */ |
| len--; |
| /* Strip outermost quotes */ |
| if (len >= 2 && s[0] == s[len - 1] && (s[0] == '\'' || s[0] == '"')) { |
| p = s + 1; |
| len -= 2; |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, |
| "the STRING opcode argument must be quoted"); |
| return -1; |
| } |
| assert(len >= 0); |
| |
| /* Use the PyBytes API to decode the string, since that is what is used |
| to encode, and then coerce the result to Unicode. */ |
| bytes = PyBytes_DecodeEscape(p, len, NULL, 0, NULL); |
| if (bytes == NULL) |
| return -1; |
| |
| /* Leave the Python 2.x strings as bytes if the *encoding* given to the |
| Unpickler was 'bytes'. Otherwise, convert them to unicode. */ |
| if (strcmp(self->encoding, "bytes") == 0) { |
| obj = bytes; |
| } |
| else { |
| obj = PyUnicode_FromEncodedObject(bytes, self->encoding, self->errors); |
| Py_DECREF(bytes); |
| if (obj == NULL) { |
| return -1; |
| } |
| } |
| |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_counted_binstring(UnpicklerObject *self, int nbytes) |
| { |
| PyObject *obj; |
| Py_ssize_t size; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, nbytes) < 0) |
| return -1; |
| |
| size = calc_binsize(s, nbytes); |
| if (size < 0) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_Format(st->UnpicklingError, |
| "BINSTRING exceeds system's maximum size of %zd bytes", |
| PY_SSIZE_T_MAX); |
| return -1; |
| } |
| |
| if (_Unpickler_Read(self, &s, size) < 0) |
| return -1; |
| |
| /* Convert Python 2.x strings to bytes if the *encoding* given to the |
| Unpickler was 'bytes'. Otherwise, convert them to unicode. */ |
| if (strcmp(self->encoding, "bytes") == 0) { |
| obj = PyBytes_FromStringAndSize(s, size); |
| } |
| else { |
| obj = PyUnicode_Decode(s, size, self->encoding, self->errors); |
| } |
| if (obj == NULL) { |
| return -1; |
| } |
| |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_counted_binbytes(UnpicklerObject *self, int nbytes) |
| { |
| PyObject *bytes; |
| Py_ssize_t size; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, nbytes) < 0) |
| return -1; |
| |
| size = calc_binsize(s, nbytes); |
| if (size < 0) { |
| PyErr_Format(PyExc_OverflowError, |
| "BINBYTES exceeds system's maximum size of %zd bytes", |
| PY_SSIZE_T_MAX); |
| return -1; |
| } |
| |
| if (_Unpickler_Read(self, &s, size) < 0) |
| return -1; |
| |
| bytes = PyBytes_FromStringAndSize(s, size); |
| if (bytes == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, bytes, -1); |
| return 0; |
| } |
| |
| static int |
| load_unicode(UnpicklerObject *self) |
| { |
| PyObject *str; |
| Py_ssize_t len; |
| char *s; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 1) |
| return bad_readline(); |
| |
| str = PyUnicode_DecodeRawUnicodeEscape(s, len - 1, NULL); |
| if (str == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, str, -1); |
| return 0; |
| } |
| |
| static int |
| load_counted_binunicode(UnpicklerObject *self, int nbytes) |
| { |
| PyObject *str; |
| Py_ssize_t size; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, nbytes) < 0) |
| return -1; |
| |
| size = calc_binsize(s, nbytes); |
| if (size < 0) { |
| PyErr_Format(PyExc_OverflowError, |
| "BINUNICODE exceeds system's maximum size of %zd bytes", |
| PY_SSIZE_T_MAX); |
| return -1; |
| } |
| |
| if (_Unpickler_Read(self, &s, size) < 0) |
| return -1; |
| |
| str = PyUnicode_DecodeUTF8(s, size, "surrogatepass"); |
| if (str == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, str, -1); |
| return 0; |
| } |
| |
| static int |
| load_counted_tuple(UnpicklerObject *self, Py_ssize_t len) |
| { |
| PyObject *tuple; |
| |
| if (Py_SIZE(self->stack) < len) |
| return stack_underflow(); |
| |
| tuple = Pdata_poptuple(self->stack, Py_SIZE(self->stack) - len); |
| if (tuple == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, tuple, -1); |
| return 0; |
| } |
| |
| static int |
| load_tuple(UnpicklerObject *self) |
| { |
| Py_ssize_t i; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| |
| return load_counted_tuple(self, Py_SIZE(self->stack) - i); |
| } |
| |
| static int |
| load_empty_list(UnpicklerObject *self) |
| { |
| PyObject *list; |
| |
| if ((list = PyList_New(0)) == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, list, -1); |
| return 0; |
| } |
| |
| static int |
| load_empty_dict(UnpicklerObject *self) |
| { |
| PyObject *dict; |
| |
| if ((dict = PyDict_New()) == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, dict, -1); |
| return 0; |
| } |
| |
| static int |
| load_empty_set(UnpicklerObject *self) |
| { |
| PyObject *set; |
| |
| if ((set = PySet_New(NULL)) == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, set, -1); |
| return 0; |
| } |
| |
| static int |
| load_list(UnpicklerObject *self) |
| { |
| PyObject *list; |
| Py_ssize_t i; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| |
| list = Pdata_poplist(self->stack, i); |
| if (list == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, list, -1); |
| return 0; |
| } |
| |
| static int |
| load_dict(UnpicklerObject *self) |
| { |
| PyObject *dict, *key, *value; |
| Py_ssize_t i, j, k; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| j = Py_SIZE(self->stack); |
| |
| if ((dict = PyDict_New()) == NULL) |
| return -1; |
| |
| for (k = i + 1; k < j; k += 2) { |
| key = self->stack->data[k - 1]; |
| value = self->stack->data[k]; |
| if (PyDict_SetItem(dict, key, value) < 0) { |
| Py_DECREF(dict); |
| return -1; |
| } |
| } |
| Pdata_clear(self->stack, i); |
| PDATA_PUSH(self->stack, dict, -1); |
| return 0; |
| } |
| |
| static int |
| load_frozenset(UnpicklerObject *self) |
| { |
| PyObject *items; |
| PyObject *frozenset; |
| Py_ssize_t i; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| |
| items = Pdata_poptuple(self->stack, i); |
| if (items == NULL) |
| return -1; |
| |
| frozenset = PyFrozenSet_New(items); |
| Py_DECREF(items); |
| if (frozenset == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, frozenset, -1); |
| return 0; |
| } |
| |
| static PyObject * |
| instantiate(PyObject *cls, PyObject *args) |
| { |
| PyObject *result = NULL; |
| _Py_IDENTIFIER(__getinitargs__); |
| /* Caller must assure args are a tuple. Normally, args come from |
| Pdata_poptuple which packs objects from the top of the stack |
| into a newly created tuple. */ |
| assert(PyTuple_Check(args)); |
| if (Py_SIZE(args) > 0 || !PyType_Check(cls) || |
| _PyObject_HasAttrId(cls, &PyId___getinitargs__)) { |
| result = PyObject_CallObject(cls, args); |
| } |
| else { |
| _Py_IDENTIFIER(__new__); |
| |
| result = _PyObject_CallMethodId(cls, &PyId___new__, "O", cls); |
| } |
| return result; |
| } |
| |
| static int |
| load_obj(UnpicklerObject *self) |
| { |
| PyObject *cls, *args, *obj = NULL; |
| Py_ssize_t i; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| |
| if (Py_SIZE(self->stack) - i < 1) |
| return stack_underflow(); |
| |
| args = Pdata_poptuple(self->stack, i + 1); |
| if (args == NULL) |
| return -1; |
| |
| PDATA_POP(self->stack, cls); |
| if (cls) { |
| obj = instantiate(cls, args); |
| Py_DECREF(cls); |
| } |
| Py_DECREF(args); |
| if (obj == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_inst(UnpicklerObject *self) |
| { |
| PyObject *cls = NULL; |
| PyObject *args = NULL; |
| PyObject *obj = NULL; |
| PyObject *module_name; |
| PyObject *class_name; |
| Py_ssize_t len; |
| Py_ssize_t i; |
| char *s; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| |
| /* Here it is safe to use PyUnicode_DecodeASCII(), even though non-ASCII |
| identifiers are permitted in Python 3.0, since the INST opcode is only |
| supported by older protocols on Python 2.x. */ |
| module_name = PyUnicode_DecodeASCII(s, len - 1, "strict"); |
| if (module_name == NULL) |
| return -1; |
| |
| if ((len = _Unpickler_Readline(self, &s)) >= 0) { |
| if (len < 2) { |
| Py_DECREF(module_name); |
| return bad_readline(); |
| } |
| class_name = PyUnicode_DecodeASCII(s, len - 1, "strict"); |
| if (class_name != NULL) { |
| cls = find_class(self, module_name, class_name); |
| Py_DECREF(class_name); |
| } |
| } |
| Py_DECREF(module_name); |
| |
| if (cls == NULL) |
| return -1; |
| |
| if ((args = Pdata_poptuple(self->stack, i)) != NULL) { |
| obj = instantiate(cls, args); |
| Py_DECREF(args); |
| } |
| Py_DECREF(cls); |
| |
| if (obj == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_newobj(UnpicklerObject *self) |
| { |
| PyObject *args = NULL; |
| PyObject *clsraw = NULL; |
| PyTypeObject *cls; /* clsraw cast to its true type */ |
| PyObject *obj; |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| /* Stack is ... cls argtuple, and we want to call |
| * cls.__new__(cls, *argtuple). |
| */ |
| PDATA_POP(self->stack, args); |
| if (args == NULL) |
| goto error; |
| if (!PyTuple_Check(args)) { |
| PyErr_SetString(st->UnpicklingError, |
| "NEWOBJ expected an arg " "tuple."); |
| goto error; |
| } |
| |
| PDATA_POP(self->stack, clsraw); |
| cls = (PyTypeObject *)clsraw; |
| if (cls == NULL) |
| goto error; |
| if (!PyType_Check(cls)) { |
| PyErr_SetString(st->UnpicklingError, "NEWOBJ class argument " |
| "isn't a type object"); |
| goto error; |
| } |
| if (cls->tp_new == NULL) { |
| PyErr_SetString(st->UnpicklingError, "NEWOBJ class argument " |
| "has NULL tp_new"); |
| goto error; |
| } |
| |
| /* Call __new__. */ |
| obj = cls->tp_new(cls, args, NULL); |
| if (obj == NULL) |
| goto error; |
| |
| Py_DECREF(args); |
| Py_DECREF(clsraw); |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| |
| error: |
| Py_XDECREF(args); |
| Py_XDECREF(clsraw); |
| return -1; |
| } |
| |
| static int |
| load_newobj_ex(UnpicklerObject *self) |
| { |
| PyObject *cls, *args, *kwargs; |
| PyObject *obj; |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| PDATA_POP(self->stack, kwargs); |
| if (kwargs == NULL) { |
| return -1; |
| } |
| PDATA_POP(self->stack, args); |
| if (args == NULL) { |
| Py_DECREF(kwargs); |
| return -1; |
| } |
| PDATA_POP(self->stack, cls); |
| if (cls == NULL) { |
| Py_DECREF(kwargs); |
| Py_DECREF(args); |
| return -1; |
| } |
| |
| if (!PyType_Check(cls)) { |
| Py_DECREF(kwargs); |
| Py_DECREF(args); |
| PyErr_Format(st->UnpicklingError, |
| "NEWOBJ_EX class argument must be a type, not %.200s", |
| Py_TYPE(cls)->tp_name); |
| Py_DECREF(cls); |
| return -1; |
| } |
| |
| if (((PyTypeObject *)cls)->tp_new == NULL) { |
| Py_DECREF(kwargs); |
| Py_DECREF(args); |
| Py_DECREF(cls); |
| PyErr_SetString(st->UnpicklingError, |
| "NEWOBJ_EX class argument doesn't have __new__"); |
| return -1; |
| } |
| obj = ((PyTypeObject *)cls)->tp_new((PyTypeObject *)cls, args, kwargs); |
| Py_DECREF(kwargs); |
| Py_DECREF(args); |
| Py_DECREF(cls); |
| if (obj == NULL) { |
| return -1; |
| } |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_global(UnpicklerObject *self) |
| { |
| PyObject *global = NULL; |
| PyObject *module_name; |
| PyObject *global_name; |
| Py_ssize_t len; |
| char *s; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| module_name = PyUnicode_DecodeUTF8(s, len - 1, "strict"); |
| if (!module_name) |
| return -1; |
| |
| if ((len = _Unpickler_Readline(self, &s)) >= 0) { |
| if (len < 2) { |
| Py_DECREF(module_name); |
| return bad_readline(); |
| } |
| global_name = PyUnicode_DecodeUTF8(s, len - 1, "strict"); |
| if (global_name) { |
| global = find_class(self, module_name, global_name); |
| Py_DECREF(global_name); |
| } |
| } |
| Py_DECREF(module_name); |
| |
| if (global == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, global, -1); |
| return 0; |
| } |
| |
| static int |
| load_stack_global(UnpicklerObject *self) |
| { |
| PyObject *global; |
| PyObject *module_name; |
| PyObject *global_name; |
| |
| PDATA_POP(self->stack, global_name); |
| PDATA_POP(self->stack, module_name); |
| if (module_name == NULL || !PyUnicode_CheckExact(module_name) || |
| global_name == NULL || !PyUnicode_CheckExact(global_name)) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, "STACK_GLOBAL requires str"); |
| Py_XDECREF(global_name); |
| Py_XDECREF(module_name); |
| return -1; |
| } |
| global = find_class(self, module_name, global_name); |
| Py_DECREF(global_name); |
| Py_DECREF(module_name); |
| if (global == NULL) |
| return -1; |
| PDATA_PUSH(self->stack, global, -1); |
| return 0; |
| } |
| |
| static int |
| load_persid(UnpicklerObject *self) |
| { |
| PyObject *pid; |
| Py_ssize_t len; |
| char *s; |
| |
| if (self->pers_func) { |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 1) |
| return bad_readline(); |
| |
| pid = PyBytes_FromStringAndSize(s, len - 1); |
| if (pid == NULL) |
| return -1; |
| |
| /* This does not leak since _Pickle_FastCall() steals the reference |
| to pid first. */ |
| pid = _Pickle_FastCall(self->pers_func, pid); |
| if (pid == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, pid, -1); |
| return 0; |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, |
| "A load persistent id instruction was encountered,\n" |
| "but no persistent_load function was specified."); |
| return -1; |
| } |
| } |
| |
| static int |
| load_binpersid(UnpicklerObject *self) |
| { |
| PyObject *pid; |
| |
| if (self->pers_func) { |
| PDATA_POP(self->stack, pid); |
| if (pid == NULL) |
| return -1; |
| |
| /* This does not leak since _Pickle_FastCall() steals the |
| reference to pid first. */ |
| pid = _Pickle_FastCall(self->pers_func, pid); |
| if (pid == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, pid, -1); |
| return 0; |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, |
| "A load persistent id instruction was encountered,\n" |
| "but no persistent_load function was specified."); |
| return -1; |
| } |
| } |
| |
| static int |
| load_pop(UnpicklerObject *self) |
| { |
| Py_ssize_t len = Py_SIZE(self->stack); |
| |
| /* Note that we split the (pickle.py) stack into two stacks, |
| * an object stack and a mark stack. We have to be clever and |
| * pop the right one. We do this by looking at the top of the |
| * mark stack first, and only signalling a stack underflow if |
| * the object stack is empty and the mark stack doesn't match |
| * our expectations. |
| */ |
| if (self->num_marks > 0 && self->marks[self->num_marks - 1] == len) { |
| self->num_marks--; |
| } else if (len > 0) { |
| len--; |
| Py_DECREF(self->stack->data[len]); |
| Py_SIZE(self->stack) = len; |
| } else { |
| return stack_underflow(); |
| } |
| return 0; |
| } |
| |
| static int |
| load_pop_mark(UnpicklerObject *self) |
| { |
| Py_ssize_t i; |
| |
| if ((i = marker(self)) < 0) |
| return -1; |
| |
| Pdata_clear(self->stack, i); |
| |
| return 0; |
| } |
| |
| static int |
| load_dup(UnpicklerObject *self) |
| { |
| PyObject *last; |
| Py_ssize_t len; |
| |
| if ((len = Py_SIZE(self->stack)) <= 0) |
| return stack_underflow(); |
| last = self->stack->data[len - 1]; |
| PDATA_APPEND(self->stack, last, -1); |
| return 0; |
| } |
| |
| static int |
| load_get(UnpicklerObject *self) |
| { |
| PyObject *key, *value; |
| Py_ssize_t idx; |
| Py_ssize_t len; |
| char *s; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| |
| key = PyLong_FromString(s, NULL, 10); |
| if (key == NULL) |
| return -1; |
| idx = PyLong_AsSsize_t(key); |
| if (idx == -1 && PyErr_Occurred()) { |
| Py_DECREF(key); |
| return -1; |
| } |
| |
| value = _Unpickler_MemoGet(self, idx); |
| if (value == NULL) { |
| if (!PyErr_Occurred()) |
| PyErr_SetObject(PyExc_KeyError, key); |
| Py_DECREF(key); |
| return -1; |
| } |
| Py_DECREF(key); |
| |
| PDATA_APPEND(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_binget(UnpicklerObject *self) |
| { |
| PyObject *value; |
| Py_ssize_t idx; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 1) < 0) |
| return -1; |
| |
| idx = Py_CHARMASK(s[0]); |
| |
| value = _Unpickler_MemoGet(self, idx); |
| if (value == NULL) { |
| PyObject *key = PyLong_FromSsize_t(idx); |
| if (key != NULL) { |
| PyErr_SetObject(PyExc_KeyError, key); |
| Py_DECREF(key); |
| } |
| return -1; |
| } |
| |
| PDATA_APPEND(self->stack, value, -1); |
| return 0; |
| } |
| |
| static int |
| load_long_binget(UnpicklerObject *self) |
| { |
| PyObject *value; |
| Py_ssize_t idx; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 4) < 0) |
| return -1; |
| |
| idx = calc_binsize(s, 4); |
| |
| value = _Unpickler_MemoGet(self, idx); |
| if (value == NULL) { |
| PyObject *key = PyLong_FromSsize_t(idx); |
| if (key != NULL) { |
| PyErr_SetObject(PyExc_KeyError, key); |
| Py_DECREF(key); |
| } |
| return -1; |
| } |
| |
| PDATA_APPEND(self->stack, value, -1); |
| return 0; |
| } |
| |
| /* Push an object from the extension registry (EXT[124]). nbytes is |
| * the number of bytes following the opcode, holding the index (code) value. |
| */ |
| static int |
| load_extension(UnpicklerObject *self, int nbytes) |
| { |
| char *codebytes; /* the nbytes bytes after the opcode */ |
| long code; /* calc_binint returns long */ |
| PyObject *py_code; /* code as a Python int */ |
| PyObject *obj; /* the object to push */ |
| PyObject *pair; /* (module_name, class_name) */ |
| PyObject *module_name, *class_name; |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| assert(nbytes == 1 || nbytes == 2 || nbytes == 4); |
| if (_Unpickler_Read(self, &codebytes, nbytes) < 0) |
| return -1; |
| code = calc_binint(codebytes, nbytes); |
| if (code <= 0) { /* note that 0 is forbidden */ |
| /* Corrupt or hostile pickle. */ |
| PyErr_SetString(st->UnpicklingError, "EXT specifies code <= 0"); |
| return -1; |
| } |
| |
| /* Look for the code in the cache. */ |
| py_code = PyLong_FromLong(code); |
| if (py_code == NULL) |
| return -1; |
| obj = PyDict_GetItemWithError(st->extension_cache, py_code); |
| if (obj != NULL) { |
| /* Bingo. */ |
| Py_DECREF(py_code); |
| PDATA_APPEND(self->stack, obj, -1); |
| return 0; |
| } |
| if (PyErr_Occurred()) { |
| Py_DECREF(py_code); |
| return -1; |
| } |
| |
| /* Look up the (module_name, class_name) pair. */ |
| pair = PyDict_GetItemWithError(st->inverted_registry, py_code); |
| if (pair == NULL) { |
| Py_DECREF(py_code); |
| if (!PyErr_Occurred()) { |
| PyErr_Format(PyExc_ValueError, "unregistered extension " |
| "code %ld", code); |
| } |
| return -1; |
| } |
| /* Since the extension registry is manipulable via Python code, |
| * confirm that pair is really a 2-tuple of strings. |
| */ |
| if (!PyTuple_Check(pair) || PyTuple_Size(pair) != 2 || |
| !PyUnicode_Check(module_name = PyTuple_GET_ITEM(pair, 0)) || |
| !PyUnicode_Check(class_name = PyTuple_GET_ITEM(pair, 1))) { |
| Py_DECREF(py_code); |
| PyErr_Format(PyExc_ValueError, "_inverted_registry[%ld] " |
| "isn't a 2-tuple of strings", code); |
| return -1; |
| } |
| /* Load the object. */ |
| obj = find_class(self, module_name, class_name); |
| if (obj == NULL) { |
| Py_DECREF(py_code); |
| return -1; |
| } |
| /* Cache code -> obj. */ |
| code = PyDict_SetItem(st->extension_cache, py_code, obj); |
| Py_DECREF(py_code); |
| if (code < 0) { |
| Py_DECREF(obj); |
| return -1; |
| } |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| static int |
| load_put(UnpicklerObject *self) |
| { |
| PyObject *key, *value; |
| Py_ssize_t idx; |
| Py_ssize_t len; |
| char *s; |
| |
| if ((len = _Unpickler_Readline(self, &s)) < 0) |
| return -1; |
| if (len < 2) |
| return bad_readline(); |
| if (Py_SIZE(self->stack) <= 0) |
| return stack_underflow(); |
| value = self->stack->data[Py_SIZE(self->stack) - 1]; |
| |
| key = PyLong_FromString(s, NULL, 10); |
| if (key == NULL) |
| return -1; |
| idx = PyLong_AsSsize_t(key); |
| Py_DECREF(key); |
| if (idx < 0) { |
| if (!PyErr_Occurred()) |
| PyErr_SetString(PyExc_ValueError, |
| "negative PUT argument"); |
| return -1; |
| } |
| |
| return _Unpickler_MemoPut(self, idx, value); |
| } |
| |
| static int |
| load_binput(UnpicklerObject *self) |
| { |
| PyObject *value; |
| Py_ssize_t idx; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 1) < 0) |
| return -1; |
| |
| if (Py_SIZE(self->stack) <= 0) |
| return stack_underflow(); |
| value = self->stack->data[Py_SIZE(self->stack) - 1]; |
| |
| idx = Py_CHARMASK(s[0]); |
| |
| return _Unpickler_MemoPut(self, idx, value); |
| } |
| |
| static int |
| load_long_binput(UnpicklerObject *self) |
| { |
| PyObject *value; |
| Py_ssize_t idx; |
| char *s; |
| |
| if (_Unpickler_Read(self, &s, 4) < 0) |
| return -1; |
| |
| if (Py_SIZE(self->stack) <= 0) |
| return stack_underflow(); |
| value = self->stack->data[Py_SIZE(self->stack) - 1]; |
| |
| idx = calc_binsize(s, 4); |
| if (idx < 0) { |
| PyErr_SetString(PyExc_ValueError, |
| "negative LONG_BINPUT argument"); |
| return -1; |
| } |
| |
| return _Unpickler_MemoPut(self, idx, value); |
| } |
| |
| static int |
| load_memoize(UnpicklerObject *self) |
| { |
| PyObject *value; |
| |
| if (Py_SIZE(self->stack) <= 0) |
| return stack_underflow(); |
| value = self->stack->data[Py_SIZE(self->stack) - 1]; |
| |
| return _Unpickler_MemoPut(self, self->memo_len, value); |
| } |
| |
| static int |
| do_append(UnpicklerObject *self, Py_ssize_t x) |
| { |
| PyObject *value; |
| PyObject *list; |
| Py_ssize_t len, i; |
| |
| len = Py_SIZE(self->stack); |
| if (x > len || x <= 0) |
| return stack_underflow(); |
| if (len == x) /* nothing to do */ |
| return 0; |
| |
| list = self->stack->data[x - 1]; |
| |
| if (PyList_Check(list)) { |
| PyObject *slice; |
| Py_ssize_t list_len; |
| int ret; |
| |
| slice = Pdata_poplist(self->stack, x); |
| if (!slice) |
| return -1; |
| list_len = PyList_GET_SIZE(list); |
| ret = PyList_SetSlice(list, list_len, list_len, slice); |
| Py_DECREF(slice); |
| return ret; |
| } |
| else { |
| PyObject *append_func; |
| _Py_IDENTIFIER(append); |
| |
| append_func = _PyObject_GetAttrId(list, &PyId_append); |
| if (append_func == NULL) |
| return -1; |
| for (i = x; i < len; i++) { |
| PyObject *result; |
| |
| value = self->stack->data[i]; |
| result = _Pickle_FastCall(append_func, value); |
| if (result == NULL) { |
| Pdata_clear(self->stack, i + 1); |
| Py_SIZE(self->stack) = x; |
| Py_DECREF(append_func); |
| return -1; |
| } |
| Py_DECREF(result); |
| } |
| Py_SIZE(self->stack) = x; |
| Py_DECREF(append_func); |
| } |
| |
| return 0; |
| } |
| |
| static int |
| load_append(UnpicklerObject *self) |
| { |
| if (Py_SIZE(self->stack) - 1 <= 0) |
| return stack_underflow(); |
| return do_append(self, Py_SIZE(self->stack) - 1); |
| } |
| |
| static int |
| load_appends(UnpicklerObject *self) |
| { |
| Py_ssize_t i = marker(self); |
| if (i < 0) |
| return -1; |
| return do_append(self, i); |
| } |
| |
| static int |
| do_setitems(UnpicklerObject *self, Py_ssize_t x) |
| { |
| PyObject *value, *key; |
| PyObject *dict; |
| Py_ssize_t len, i; |
| int status = 0; |
| |
| len = Py_SIZE(self->stack); |
| if (x > len || x <= 0) |
| return stack_underflow(); |
| if (len == x) /* nothing to do */ |
| return 0; |
| if ((len - x) % 2 != 0) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| /* Currupt or hostile pickle -- we never write one like this. */ |
| PyErr_SetString(st->UnpicklingError, |
| "odd number of items for SETITEMS"); |
| return -1; |
| } |
| |
| /* Here, dict does not actually need to be a PyDict; it could be anything |
| that supports the __setitem__ attribute. */ |
| dict = self->stack->data[x - 1]; |
| |
| for (i = x + 1; i < len; i += 2) { |
| key = self->stack->data[i - 1]; |
| value = self->stack->data[i]; |
| if (PyObject_SetItem(dict, key, value) < 0) { |
| status = -1; |
| break; |
| } |
| } |
| |
| Pdata_clear(self->stack, x); |
| return status; |
| } |
| |
| static int |
| load_setitem(UnpicklerObject *self) |
| { |
| return do_setitems(self, Py_SIZE(self->stack) - 2); |
| } |
| |
| static int |
| load_setitems(UnpicklerObject *self) |
| { |
| Py_ssize_t i = marker(self); |
| if (i < 0) |
| return -1; |
| return do_setitems(self, i); |
| } |
| |
| static int |
| load_additems(UnpicklerObject *self) |
| { |
| PyObject *set; |
| Py_ssize_t mark, len, i; |
| |
| mark = marker(self); |
| if (mark < 0) |
| return -1; |
| len = Py_SIZE(self->stack); |
| if (mark > len || mark <= 0) |
| return stack_underflow(); |
| if (len == mark) /* nothing to do */ |
| return 0; |
| |
| set = self->stack->data[mark - 1]; |
| |
| if (PySet_Check(set)) { |
| PyObject *items; |
| int status; |
| |
| items = Pdata_poptuple(self->stack, mark); |
| if (items == NULL) |
| return -1; |
| |
| status = _PySet_Update(set, items); |
| Py_DECREF(items); |
| return status; |
| } |
| else { |
| PyObject *add_func; |
| _Py_IDENTIFIER(add); |
| |
| add_func = _PyObject_GetAttrId(set, &PyId_add); |
| if (add_func == NULL) |
| return -1; |
| for (i = mark; i < len; i++) { |
| PyObject *result; |
| PyObject *item; |
| |
| item = self->stack->data[i]; |
| result = _Pickle_FastCall(add_func, item); |
| if (result == NULL) { |
| Pdata_clear(self->stack, i + 1); |
| Py_SIZE(self->stack) = mark; |
| return -1; |
| } |
| Py_DECREF(result); |
| } |
| Py_SIZE(self->stack) = mark; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| load_build(UnpicklerObject *self) |
| { |
| PyObject *state, *inst, *slotstate; |
| PyObject *setstate; |
| int status = 0; |
| _Py_IDENTIFIER(__setstate__); |
| |
| /* Stack is ... instance, state. We want to leave instance at |
| * the stack top, possibly mutated via instance.__setstate__(state). |
| */ |
| if (Py_SIZE(self->stack) < 2) |
| return stack_underflow(); |
| |
| PDATA_POP(self->stack, state); |
| if (state == NULL) |
| return -1; |
| |
| inst = self->stack->data[Py_SIZE(self->stack) - 1]; |
| |
| setstate = _PyObject_GetAttrId(inst, &PyId___setstate__); |
| if (setstate == NULL) { |
| if (PyErr_ExceptionMatches(PyExc_AttributeError)) |
| PyErr_Clear(); |
| else { |
| Py_DECREF(state); |
| return -1; |
| } |
| } |
| else { |
| PyObject *result; |
| |
| /* The explicit __setstate__ is responsible for everything. */ |
| result = _Pickle_FastCall(setstate, state); |
| Py_DECREF(setstate); |
| if (result == NULL) |
| return -1; |
| Py_DECREF(result); |
| return 0; |
| } |
| |
| /* A default __setstate__. First see whether state embeds a |
| * slot state dict too (a proto 2 addition). |
| */ |
| if (PyTuple_Check(state) && Py_SIZE(state) == 2) { |
| PyObject *tmp = state; |
| |
| state = PyTuple_GET_ITEM(tmp, 0); |
| slotstate = PyTuple_GET_ITEM(tmp, 1); |
| Py_INCREF(state); |
| Py_INCREF(slotstate); |
| Py_DECREF(tmp); |
| } |
| else |
| slotstate = NULL; |
| |
| /* Set inst.__dict__ from the state dict (if any). */ |
| if (state != Py_None) { |
| PyObject *dict; |
| PyObject *d_key, *d_value; |
| Py_ssize_t i; |
| _Py_IDENTIFIER(__dict__); |
| |
| if (!PyDict_Check(state)) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, "state is not a dictionary"); |
| goto error; |
| } |
| dict = _PyObject_GetAttrId(inst, &PyId___dict__); |
| if (dict == NULL) |
| goto error; |
| |
| i = 0; |
| while (PyDict_Next(state, &i, &d_key, &d_value)) { |
| /* normally the keys for instance attributes are |
| interned. we should try to do that here. */ |
| Py_INCREF(d_key); |
| if (PyUnicode_CheckExact(d_key)) |
| PyUnicode_InternInPlace(&d_key); |
| if (PyObject_SetItem(dict, d_key, d_value) < 0) { |
| Py_DECREF(d_key); |
| goto error; |
| } |
| Py_DECREF(d_key); |
| } |
| Py_DECREF(dict); |
| } |
| |
| /* Also set instance attributes from the slotstate dict (if any). */ |
| if (slotstate != NULL) { |
| PyObject *d_key, *d_value; |
| Py_ssize_t i; |
| |
| if (!PyDict_Check(slotstate)) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_SetString(st->UnpicklingError, |
| "slot state is not a dictionary"); |
| goto error; |
| } |
| i = 0; |
| while (PyDict_Next(slotstate, &i, &d_key, &d_value)) { |
| if (PyObject_SetAttr(inst, d_key, d_value) < 0) |
| goto error; |
| } |
| } |
| |
| if (0) { |
| error: |
| status = -1; |
| } |
| |
| Py_DECREF(state); |
| Py_XDECREF(slotstate); |
| return status; |
| } |
| |
| static int |
| load_mark(UnpicklerObject *self) |
| { |
| |
| /* Note that we split the (pickle.py) stack into two stacks, an |
| * object stack and a mark stack. Here we push a mark onto the |
| * mark stack. |
| */ |
| |
| if ((self->num_marks + 1) >= self->marks_size) { |
| size_t alloc; |
| |
| /* Use the size_t type to check for overflow. */ |
| alloc = ((size_t)self->num_marks << 1) + 20; |
| if (alloc > (PY_SSIZE_T_MAX / sizeof(Py_ssize_t)) || |
| alloc <= ((size_t)self->num_marks + 1)) { |
| PyErr_NoMemory(); |
| return -1; |
| } |
| |
| if (self->marks == NULL) |
| self->marks = PyMem_NEW(Py_ssize_t, alloc); |
| else |
| PyMem_RESIZE(self->marks, Py_ssize_t, alloc); |
| if (self->marks == NULL) { |
| self->marks_size = 0; |
| PyErr_NoMemory(); |
| return -1; |
| } |
| self->marks_size = (Py_ssize_t)alloc; |
| } |
| |
| self->marks[self->num_marks++] = Py_SIZE(self->stack); |
| |
| return 0; |
| } |
| |
| static int |
| load_reduce(UnpicklerObject *self) |
| { |
| PyObject *callable = NULL; |
| PyObject *argtup = NULL; |
| PyObject *obj = NULL; |
| |
| PDATA_POP(self->stack, argtup); |
| if (argtup == NULL) |
| return -1; |
| PDATA_POP(self->stack, callable); |
| if (callable) { |
| obj = PyObject_CallObject(callable, argtup); |
| Py_DECREF(callable); |
| } |
| Py_DECREF(argtup); |
| |
| if (obj == NULL) |
| return -1; |
| |
| PDATA_PUSH(self->stack, obj, -1); |
| return 0; |
| } |
| |
| /* Just raises an error if we don't know the protocol specified. PROTO |
| * is the first opcode for protocols >= 2. |
| */ |
| static int |
| load_proto(UnpicklerObject *self) |
| { |
| char *s; |
| int i; |
| |
| if (_Unpickler_Read(self, &s, 1) < 0) |
| return -1; |
| |
| i = (unsigned char)s[0]; |
| if (i <= HIGHEST_PROTOCOL) { |
| self->proto = i; |
| return 0; |
| } |
| |
| PyErr_Format(PyExc_ValueError, "unsupported pickle protocol: %d", i); |
| return -1; |
| } |
| |
| static int |
| load_frame(UnpicklerObject *self) |
| { |
| char *s; |
| Py_ssize_t frame_len; |
| |
| if (_Unpickler_Read(self, &s, 8) < 0) |
| return -1; |
| |
| frame_len = calc_binsize(s, 8); |
| if (frame_len < 0) { |
| PyErr_Format(PyExc_OverflowError, |
| "FRAME length exceeds system's maximum of %zd bytes", |
| PY_SSIZE_T_MAX); |
| return -1; |
| } |
| |
| if (_Unpickler_Read(self, &s, frame_len) < 0) |
| return -1; |
| |
| /* Rewind to start of frame */ |
| self->next_read_idx -= frame_len; |
| return 0; |
| } |
| |
| static PyObject * |
| load(UnpicklerObject *self) |
| { |
| PyObject *value = NULL; |
| char *s = NULL; |
| |
| self->num_marks = 0; |
| self->proto = 0; |
| if (Py_SIZE(self->stack)) |
| Pdata_clear(self->stack, 0); |
| |
| /* Convenient macros for the dispatch while-switch loop just below. */ |
| #define OP(opcode, load_func) \ |
| case opcode: if (load_func(self) < 0) break; continue; |
| |
| #define OP_ARG(opcode, load_func, arg) \ |
| case opcode: if (load_func(self, (arg)) < 0) break; continue; |
| |
| while (1) { |
| if (_Unpickler_Read(self, &s, 1) < 0) |
| break; |
| |
| switch ((enum opcode)s[0]) { |
| OP(NONE, load_none) |
| OP(BININT, load_binint) |
| OP(BININT1, load_binint1) |
| OP(BININT2, load_binint2) |
| OP(INT, load_int) |
| OP(LONG, load_long) |
| OP_ARG(LONG1, load_counted_long, 1) |
| OP_ARG(LONG4, load_counted_long, 4) |
| OP(FLOAT, load_float) |
| OP(BINFLOAT, load_binfloat) |
| OP_ARG(SHORT_BINBYTES, load_counted_binbytes, 1) |
| OP_ARG(BINBYTES, load_counted_binbytes, 4) |
| OP_ARG(BINBYTES8, load_counted_binbytes, 8) |
| OP_ARG(SHORT_BINSTRING, load_counted_binstring, 1) |
| OP_ARG(BINSTRING, load_counted_binstring, 4) |
| OP(STRING, load_string) |
| OP(UNICODE, load_unicode) |
| OP_ARG(SHORT_BINUNICODE, load_counted_binunicode, 1) |
| OP_ARG(BINUNICODE, load_counted_binunicode, 4) |
| OP_ARG(BINUNICODE8, load_counted_binunicode, 8) |
| OP_ARG(EMPTY_TUPLE, load_counted_tuple, 0) |
| OP_ARG(TUPLE1, load_counted_tuple, 1) |
| OP_ARG(TUPLE2, load_counted_tuple, 2) |
| OP_ARG(TUPLE3, load_counted_tuple, 3) |
| OP(TUPLE, load_tuple) |
| OP(EMPTY_LIST, load_empty_list) |
| OP(LIST, load_list) |
| OP(EMPTY_DICT, load_empty_dict) |
| OP(DICT, load_dict) |
| OP(EMPTY_SET, load_empty_set) |
| OP(ADDITEMS, load_additems) |
| OP(FROZENSET, load_frozenset) |
| OP(OBJ, load_obj) |
| OP(INST, load_inst) |
| OP(NEWOBJ, load_newobj) |
| OP(NEWOBJ_EX, load_newobj_ex) |
| OP(GLOBAL, load_global) |
| OP(STACK_GLOBAL, load_stack_global) |
| OP(APPEND, load_append) |
| OP(APPENDS, load_appends) |
| OP(BUILD, load_build) |
| OP(DUP, load_dup) |
| OP(BINGET, load_binget) |
| OP(LONG_BINGET, load_long_binget) |
| OP(GET, load_get) |
| OP(MARK, load_mark) |
| OP(BINPUT, load_binput) |
| OP(LONG_BINPUT, load_long_binput) |
| OP(PUT, load_put) |
| OP(MEMOIZE, load_memoize) |
| OP(POP, load_pop) |
| OP(POP_MARK, load_pop_mark) |
| OP(SETITEM, load_setitem) |
| OP(SETITEMS, load_setitems) |
| OP(PERSID, load_persid) |
| OP(BINPERSID, load_binpersid) |
| OP(REDUCE, load_reduce) |
| OP(PROTO, load_proto) |
| OP(FRAME, load_frame) |
| OP_ARG(EXT1, load_extension, 1) |
| OP_ARG(EXT2, load_extension, 2) |
| OP_ARG(EXT4, load_extension, 4) |
| OP_ARG(NEWTRUE, load_bool, Py_True) |
| OP_ARG(NEWFALSE, load_bool, Py_False) |
| |
| case STOP: |
| break; |
| |
| default: |
| if (s[0] == '\0') { |
| PyErr_SetNone(PyExc_EOFError); |
| } |
| else { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_Format(st->UnpicklingError, |
| "invalid load key, '%c'.", s[0]); |
| } |
| return NULL; |
| } |
| |
| break; /* and we are done! */ |
| } |
| |
| if (PyErr_Occurred()) { |
| return NULL; |
| } |
| |
| if (_Unpickler_SkipConsumed(self) < 0) |
| return NULL; |
| |
| PDATA_POP(self->stack, value); |
| return value; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Unpickler.load |
| |
| Load a pickle. |
| |
| Read a pickled object representation from the open file object given |
| in the constructor, and return the reconstituted object hierarchy |
| specified therein. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_Unpickler_load_impl(UnpicklerObject *self) |
| /*[clinic end generated code: output=fdcc488aad675b14 input=acbb91a42fa9b7b9]*/ |
| { |
| UnpicklerObject *unpickler = (UnpicklerObject*)self; |
| |
| /* Check whether the Unpickler was initialized correctly. This prevents |
| segfaulting if a subclass overridden __init__ with a function that does |
| not call Unpickler.__init__(). Here, we simply ensure that self->read |
| is not NULL. */ |
| if (unpickler->read == NULL) { |
| PickleState *st = _Pickle_GetGlobalState(); |
| PyErr_Format(st->UnpicklingError, |
| "Unpickler.__init__() was not called by %s.__init__()", |
| Py_TYPE(unpickler)->tp_name); |
| return NULL; |
| } |
| |
| return load(unpickler); |
| } |
| |
| /* The name of find_class() is misleading. In newer pickle protocols, this |
| function is used for loading any global (i.e., functions), not just |
| classes. The name is kept only for backward compatibility. */ |
| |
| /*[clinic input] |
| |
| _pickle.Unpickler.find_class |
| |
| module_name: object |
| global_name: object |
| / |
| |
| Return an object from a specified module. |
| |
| If necessary, the module will be imported. Subclasses may override |
| this method (e.g. to restrict unpickling of arbitrary classes and |
| functions). |
| |
| This method is called whenever a class or a function object is |
| needed. Both arguments passed are str objects. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_Unpickler_find_class_impl(UnpicklerObject *self, |
| PyObject *module_name, |
| PyObject *global_name) |
| /*[clinic end generated code: output=becc08d7f9ed41e3 input=e2e6a865de093ef4]*/ |
| { |
| PyObject *global; |
| PyObject *modules_dict; |
| PyObject *module; |
| _Py_IDENTIFIER(modules); |
| |
| /* Try to map the old names used in Python 2.x to the new ones used in |
| Python 3.x. We do this only with old pickle protocols and when the |
| user has not disabled the feature. */ |
| if (self->proto < 3 && self->fix_imports) { |
| PyObject *key; |
| PyObject *item; |
| PickleState *st = _Pickle_GetGlobalState(); |
| |
| /* Check if the global (i.e., a function or a class) was renamed |
| or moved to another module. */ |
| key = PyTuple_Pack(2, module_name, global_name); |
| if (key == NULL) |
| return NULL; |
| item = PyDict_GetItemWithError(st->name_mapping_2to3, key); |
| Py_DECREF(key); |
| if (item) { |
| if (!PyTuple_Check(item) || PyTuple_GET_SIZE(item) != 2) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.NAME_MAPPING values should be " |
| "2-tuples, not %.200s", Py_TYPE(item)->tp_name); |
| return NULL; |
| } |
| module_name = PyTuple_GET_ITEM(item, 0); |
| global_name = PyTuple_GET_ITEM(item, 1); |
| if (!PyUnicode_Check(module_name) || |
| !PyUnicode_Check(global_name)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.NAME_MAPPING values should be " |
| "pairs of str, not (%.200s, %.200s)", |
| Py_TYPE(module_name)->tp_name, |
| Py_TYPE(global_name)->tp_name); |
| return NULL; |
| } |
| } |
| else if (PyErr_Occurred()) { |
| return NULL; |
| } |
| else { |
| /* Check if the module was renamed. */ |
| item = PyDict_GetItemWithError(st->import_mapping_2to3, module_name); |
| if (item) { |
| if (!PyUnicode_Check(item)) { |
| PyErr_Format(PyExc_RuntimeError, |
| "_compat_pickle.IMPORT_MAPPING values should be " |
| "strings, not %.200s", Py_TYPE(item)->tp_name); |
| return NULL; |
| } |
| module_name = item; |
| } |
| else if (PyErr_Occurred()) { |
| return NULL; |
| } |
| } |
| } |
| |
| modules_dict = _PySys_GetObjectId(&PyId_modules); |
| if (modules_dict == NULL) { |
| PyErr_SetString(PyExc_RuntimeError, "unable to get sys.modules"); |
| return NULL; |
| } |
| |
| module = PyDict_GetItemWithError(modules_dict, module_name); |
| if (module == NULL) { |
| if (PyErr_Occurred()) |
| return NULL; |
| module = PyImport_Import(module_name); |
| if (module == NULL) |
| return NULL; |
| global = getattribute(module, global_name, self->proto >= 4); |
| Py_DECREF(module); |
| } |
| else { |
| global = getattribute(module, global_name, self->proto >= 4); |
| } |
| return global; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Unpickler.__sizeof__ -> Py_ssize_t |
| |
| Returns size in memory, in bytes. |
| [clinic start generated code]*/ |
| |
| static Py_ssize_t |
| _pickle_Unpickler___sizeof___impl(UnpicklerObject *self) |
| /*[clinic end generated code: output=119d9d03ad4c7651 input=13333471fdeedf5e]*/ |
| { |
| Py_ssize_t res; |
| |
| res = _PyObject_SIZE(Py_TYPE(self)); |
| if (self->memo != NULL) |
| res += self->memo_size * sizeof(PyObject *); |
| if (self->marks != NULL) |
| res += self->marks_size * sizeof(Py_ssize_t); |
| if (self->input_line != NULL) |
| res += strlen(self->input_line) + 1; |
| if (self->encoding != NULL) |
| res += strlen(self->encoding) + 1; |
| if (self->errors != NULL) |
| res += strlen(self->errors) + 1; |
| return res; |
| } |
| |
| static struct PyMethodDef Unpickler_methods[] = { |
| _PICKLE_UNPICKLER_LOAD_METHODDEF |
| _PICKLE_UNPICKLER_FIND_CLASS_METHODDEF |
| _PICKLE_UNPICKLER___SIZEOF___METHODDEF |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| static void |
| Unpickler_dealloc(UnpicklerObject *self) |
| { |
| PyObject_GC_UnTrack((PyObject *)self); |
| Py_XDECREF(self->readline); |
| Py_XDECREF(self->read); |
| Py_XDECREF(self->peek); |
| Py_XDECREF(self->stack); |
| Py_XDECREF(self->pers_func); |
| if (self->buffer.buf != NULL) { |
| PyBuffer_Release(&self->buffer); |
| self->buffer.buf = NULL; |
| } |
| |
| _Unpickler_MemoCleanup(self); |
| PyMem_Free(self->marks); |
| PyMem_Free(self->input_line); |
| PyMem_Free(self->encoding); |
| PyMem_Free(self->errors); |
| |
| Py_TYPE(self)->tp_free((PyObject *)self); |
| } |
| |
| static int |
| Unpickler_traverse(UnpicklerObject *self, visitproc visit, void *arg) |
| { |
| Py_VISIT(self->readline); |
| Py_VISIT(self->read); |
| Py_VISIT(self->peek); |
| Py_VISIT(self->stack); |
| Py_VISIT(self->pers_func); |
| return 0; |
| } |
| |
| static int |
| Unpickler_clear(UnpicklerObject *self) |
| { |
| Py_CLEAR(self->readline); |
| Py_CLEAR(self->read); |
| Py_CLEAR(self->peek); |
| Py_CLEAR(self->stack); |
| Py_CLEAR(self->pers_func); |
| if (self->buffer.buf != NULL) { |
| PyBuffer_Release(&self->buffer); |
| self->buffer.buf = NULL; |
| } |
| |
| _Unpickler_MemoCleanup(self); |
| PyMem_Free(self->marks); |
| self->marks = NULL; |
| PyMem_Free(self->input_line); |
| self->input_line = NULL; |
| PyMem_Free(self->encoding); |
| self->encoding = NULL; |
| PyMem_Free(self->errors); |
| self->errors = NULL; |
| |
| return 0; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.Unpickler.__init__ |
| |
| file: object |
| * |
| fix_imports: bool = True |
| encoding: str = 'ASCII' |
| errors: str = 'strict' |
| |
| This takes a binary file for reading a pickle data stream. |
| |
| The protocol version of the pickle is detected automatically, so no |
| protocol argument is needed. Bytes past the pickled object's |
| representation are ignored. |
| |
| The argument *file* must have two methods, a read() method that takes |
| an integer argument, and a readline() method that requires no |
| arguments. Both methods should return bytes. Thus *file* can be a |
| binary file object opened for reading, an io.BytesIO object, or any |
| other custom object that meets this interface. |
| |
| Optional keyword arguments are *fix_imports*, *encoding* and *errors*, |
| which are used to control compatiblity support for pickle stream |
| generated by Python 2. If *fix_imports* is True, pickle will try to |
| map the old Python 2 names to the new names used in Python 3. The |
| *encoding* and *errors* tell pickle how to decode 8-bit string |
| instances pickled by Python 2; these default to 'ASCII' and 'strict', |
| respectively. The *encoding* can be 'bytes' to read these 8-bit |
| string instances as bytes objects. |
| [clinic start generated code]*/ |
| |
| static int |
| _pickle_Unpickler___init___impl(UnpicklerObject *self, PyObject *file, |
| int fix_imports, const char *encoding, |
| const char *errors) |
| /*[clinic end generated code: output=e2c8ce748edc57b0 input=04ece661aa884837]*/ |
| { |
| _Py_IDENTIFIER(persistent_load); |
| |
| /* In case of multiple __init__() calls, clear previous content. */ |
| if (self->read != NULL) |
| (void)Unpickler_clear(self); |
| |
| if (_Unpickler_SetInputStream(self, file) < 0) |
| return -1; |
| |
| if (_Unpickler_SetInputEncoding(self, encoding, errors) < 0) |
| return -1; |
| |
| self->fix_imports = fix_imports; |
| if (self->fix_imports == -1) |
| return -1; |
| |
| if (_PyObject_HasAttrId((PyObject *)self, &PyId_persistent_load)) { |
| self->pers_func = _PyObject_GetAttrId((PyObject *)self, |
| &PyId_persistent_load); |
| if (self->pers_func == NULL) |
| return 1; |
| } |
| else { |
| self->pers_func = NULL; |
| } |
| |
| self->stack = (Pdata *)Pdata_New(); |
| if (self->stack == NULL) |
| return 1; |
| |
| self->memo_size = 32; |
| self->memo = _Unpickler_NewMemo(self->memo_size); |
| if (self->memo == NULL) |
| return -1; |
| |
| self->proto = 0; |
| |
| return 0; |
| } |
| |
| |
| /* Define a proxy object for the Unpickler's internal memo object. This is to |
| * avoid breaking code like: |
| * unpickler.memo.clear() |
| * and |
| * unpickler.memo = saved_memo |
| * Is this a good idea? Not really, but we don't want to break code that uses |
| * it. Note that we don't implement the entire mapping API here. This is |
| * intentional, as these should be treated as black-box implementation details. |
| * |
| * We do, however, have to implement pickling/unpickling support because of |
| * real-world code like cvs2svn. |
| */ |
| |
| /*[clinic input] |
| _pickle.UnpicklerMemoProxy.clear |
| |
| Remove all items from memo. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_UnpicklerMemoProxy_clear_impl(UnpicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=d20cd43f4ba1fb1f input=b1df7c52e7afd9bd]*/ |
| { |
| _Unpickler_MemoCleanup(self->unpickler); |
| self->unpickler->memo = _Unpickler_NewMemo(self->unpickler->memo_size); |
| if (self->unpickler->memo == NULL) |
| return NULL; |
| Py_RETURN_NONE; |
| } |
| |
| /*[clinic input] |
| _pickle.UnpicklerMemoProxy.copy |
| |
| Copy the memo to a new object. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_UnpicklerMemoProxy_copy_impl(UnpicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=e12af7e9bc1e4c77 input=97769247ce032c1d]*/ |
| { |
| Py_ssize_t i; |
| PyObject *new_memo = PyDict_New(); |
| if (new_memo == NULL) |
| return NULL; |
| |
| for (i = 0; i < self->unpickler->memo_size; i++) { |
| int status; |
| PyObject *key, *value; |
| |
| value = self->unpickler->memo[i]; |
| if (value == NULL) |
| continue; |
| |
| key = PyLong_FromSsize_t(i); |
| if (key == NULL) |
| goto error; |
| status = PyDict_SetItem(new_memo, key, value); |
| Py_DECREF(key); |
| if (status < 0) |
| goto error; |
| } |
| return new_memo; |
| |
| error: |
| Py_DECREF(new_memo); |
| return NULL; |
| } |
| |
| /*[clinic input] |
| _pickle.UnpicklerMemoProxy.__reduce__ |
| |
| Implement pickling support. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_UnpicklerMemoProxy___reduce___impl(UnpicklerMemoProxyObject *self) |
| /*[clinic end generated code: output=6da34ac048d94cca input=6920862413407199]*/ |
| { |
| PyObject *reduce_value; |
| PyObject *constructor_args; |
| PyObject *contents = _pickle_UnpicklerMemoProxy_copy_impl(self); |
| if (contents == NULL) |
| return NULL; |
| |
| reduce_value = PyTuple_New(2); |
| if (reduce_value == NULL) { |
| Py_DECREF(contents); |
| return NULL; |
| } |
| constructor_args = PyTuple_New(1); |
| if (constructor_args == NULL) { |
| Py_DECREF(contents); |
| Py_DECREF(reduce_value); |
| return NULL; |
| } |
| PyTuple_SET_ITEM(constructor_args, 0, contents); |
| Py_INCREF((PyObject *)&PyDict_Type); |
| PyTuple_SET_ITEM(reduce_value, 0, (PyObject *)&PyDict_Type); |
| PyTuple_SET_ITEM(reduce_value, 1, constructor_args); |
| return reduce_value; |
| } |
| |
| static PyMethodDef unpicklerproxy_methods[] = { |
| _PICKLE_UNPICKLERMEMOPROXY_CLEAR_METHODDEF |
| _PICKLE_UNPICKLERMEMOPROXY_COPY_METHODDEF |
| _PICKLE_UNPICKLERMEMOPROXY___REDUCE___METHODDEF |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| static void |
| UnpicklerMemoProxy_dealloc(UnpicklerMemoProxyObject *self) |
| { |
| PyObject_GC_UnTrack(self); |
| Py_XDECREF(self->unpickler); |
| PyObject_GC_Del((PyObject *)self); |
| } |
| |
| static int |
| UnpicklerMemoProxy_traverse(UnpicklerMemoProxyObject *self, |
| visitproc visit, void *arg) |
| { |
| Py_VISIT(self->unpickler); |
| return 0; |
| } |
| |
| static int |
| UnpicklerMemoProxy_clear(UnpicklerMemoProxyObject *self) |
| { |
| Py_CLEAR(self->unpickler); |
| return 0; |
| } |
| |
| static PyTypeObject UnpicklerMemoProxyType = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "_pickle.UnpicklerMemoProxy", /*tp_name*/ |
| sizeof(UnpicklerMemoProxyObject), /*tp_basicsize*/ |
| 0, |
| (destructor)UnpicklerMemoProxy_dealloc, /* tp_dealloc */ |
| 0, /* tp_print */ |
| 0, /* tp_getattr */ |
| 0, /* tp_setattr */ |
| 0, /* tp_compare */ |
| 0, /* tp_repr */ |
| 0, /* tp_as_number */ |
| 0, /* tp_as_sequence */ |
| 0, /* tp_as_mapping */ |
| PyObject_HashNotImplemented, /* tp_hash */ |
| 0, /* tp_call */ |
| 0, /* tp_str */ |
| PyObject_GenericGetAttr, /* tp_getattro */ |
| PyObject_GenericSetAttr, /* tp_setattro */ |
| 0, /* tp_as_buffer */ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC, |
| 0, /* tp_doc */ |
| (traverseproc)UnpicklerMemoProxy_traverse, /* tp_traverse */ |
| (inquiry)UnpicklerMemoProxy_clear, /* tp_clear */ |
| 0, /* tp_richcompare */ |
| 0, /* tp_weaklistoffset */ |
| 0, /* tp_iter */ |
| 0, /* tp_iternext */ |
| unpicklerproxy_methods, /* tp_methods */ |
| }; |
| |
| static PyObject * |
| UnpicklerMemoProxy_New(UnpicklerObject *unpickler) |
| { |
| UnpicklerMemoProxyObject *self; |
| |
| self = PyObject_GC_New(UnpicklerMemoProxyObject, |
| &UnpicklerMemoProxyType); |
| if (self == NULL) |
| return NULL; |
| Py_INCREF(unpickler); |
| self->unpickler = unpickler; |
| PyObject_GC_Track(self); |
| return (PyObject *)self; |
| } |
| |
| /*****************************************************************************/ |
| |
| |
| static PyObject * |
| Unpickler_get_memo(UnpicklerObject *self) |
| { |
| return UnpicklerMemoProxy_New(self); |
| } |
| |
| static int |
| Unpickler_set_memo(UnpicklerObject *self, PyObject *obj) |
| { |
| PyObject **new_memo; |
| Py_ssize_t new_memo_size = 0; |
| Py_ssize_t i; |
| |
| if (obj == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "attribute deletion is not supported"); |
| return -1; |
| } |
| |
| if (Py_TYPE(obj) == &UnpicklerMemoProxyType) { |
| UnpicklerObject *unpickler = |
| ((UnpicklerMemoProxyObject *)obj)->unpickler; |
| |
| new_memo_size = unpickler->memo_size; |
| new_memo = _Unpickler_NewMemo(new_memo_size); |
| if (new_memo == NULL) |
| return -1; |
| |
| for (i = 0; i < new_memo_size; i++) { |
| Py_XINCREF(unpickler->memo[i]); |
| new_memo[i] = unpickler->memo[i]; |
| } |
| } |
| else if (PyDict_Check(obj)) { |
| Py_ssize_t i = 0; |
| PyObject *key, *value; |
| |
| new_memo_size = PyDict_Size(obj); |
| new_memo = _Unpickler_NewMemo(new_memo_size); |
| if (new_memo == NULL) |
| return -1; |
| |
| while (PyDict_Next(obj, &i, &key, &value)) { |
| Py_ssize_t idx; |
| if (!PyLong_Check(key)) { |
| PyErr_SetString(PyExc_TypeError, |
| "memo key must be integers"); |
| goto error; |
| } |
| idx = PyLong_AsSsize_t(key); |
| if (idx == -1 && PyErr_Occurred()) |
| goto error; |
| if (idx < 0) { |
| PyErr_SetString(PyExc_ValueError, |
| "memo key must be positive integers."); |
| goto error; |
| } |
| if (_Unpickler_MemoPut(self, idx, value) < 0) |
| goto error; |
| } |
| } |
| else { |
| PyErr_Format(PyExc_TypeError, |
| "'memo' attribute must be an UnpicklerMemoProxy object" |
| "or dict, not %.200s", Py_TYPE(obj)->tp_name); |
| return -1; |
| } |
| |
| _Unpickler_MemoCleanup(self); |
| self->memo_size = new_memo_size; |
| self->memo = new_memo; |
| |
| return 0; |
| |
| error: |
| if (new_memo_size) { |
| i = new_memo_size; |
| while (--i >= 0) { |
| Py_XDECREF(new_memo[i]); |
| } |
| PyMem_FREE(new_memo); |
| } |
| return -1; |
| } |
| |
| static PyObject * |
| Unpickler_get_persload(UnpicklerObject *self) |
| { |
| if (self->pers_func == NULL) |
| PyErr_SetString(PyExc_AttributeError, "persistent_load"); |
| else |
| Py_INCREF(self->pers_func); |
| return self->pers_func; |
| } |
| |
| static int |
| Unpickler_set_persload(UnpicklerObject *self, PyObject *value) |
| { |
| PyObject *tmp; |
| |
| if (value == NULL) { |
| PyErr_SetString(PyExc_TypeError, |
| "attribute deletion is not supported"); |
| return -1; |
| } |
| if (!PyCallable_Check(value)) { |
| PyErr_SetString(PyExc_TypeError, |
| "persistent_load must be a callable taking " |
| "one argument"); |
| return -1; |
| } |
| |
| tmp = self->pers_func; |
| Py_INCREF(value); |
| self->pers_func = value; |
| Py_XDECREF(tmp); /* self->pers_func can be NULL, so be careful. */ |
| |
| return 0; |
| } |
| |
| static PyGetSetDef Unpickler_getsets[] = { |
| {"memo", (getter)Unpickler_get_memo, (setter)Unpickler_set_memo}, |
| {"persistent_load", (getter)Unpickler_get_persload, |
| (setter)Unpickler_set_persload}, |
| {NULL} |
| }; |
| |
| static PyTypeObject Unpickler_Type = { |
| PyVarObject_HEAD_INIT(NULL, 0) |
| "_pickle.Unpickler", /*tp_name*/ |
| sizeof(UnpicklerObject), /*tp_basicsize*/ |
| 0, /*tp_itemsize*/ |
| (destructor)Unpickler_dealloc, /*tp_dealloc*/ |
| 0, /*tp_print*/ |
| 0, /*tp_getattr*/ |
| 0, /*tp_setattr*/ |
| 0, /*tp_reserved*/ |
| 0, /*tp_repr*/ |
| 0, /*tp_as_number*/ |
| 0, /*tp_as_sequence*/ |
| 0, /*tp_as_mapping*/ |
| 0, /*tp_hash*/ |
| 0, /*tp_call*/ |
| 0, /*tp_str*/ |
| 0, /*tp_getattro*/ |
| 0, /*tp_setattro*/ |
| 0, /*tp_as_buffer*/ |
| Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC, |
| _pickle_Unpickler___init____doc__, /*tp_doc*/ |
| (traverseproc)Unpickler_traverse, /*tp_traverse*/ |
| (inquiry)Unpickler_clear, /*tp_clear*/ |
| 0, /*tp_richcompare*/ |
| 0, /*tp_weaklistoffset*/ |
| 0, /*tp_iter*/ |
| 0, /*tp_iternext*/ |
| Unpickler_methods, /*tp_methods*/ |
| 0, /*tp_members*/ |
| Unpickler_getsets, /*tp_getset*/ |
| 0, /*tp_base*/ |
| 0, /*tp_dict*/ |
| 0, /*tp_descr_get*/ |
| 0, /*tp_descr_set*/ |
| 0, /*tp_dictoffset*/ |
| _pickle_Unpickler___init__, /*tp_init*/ |
| PyType_GenericAlloc, /*tp_alloc*/ |
| PyType_GenericNew, /*tp_new*/ |
| PyObject_GC_Del, /*tp_free*/ |
| 0, /*tp_is_gc*/ |
| }; |
| |
| /*[clinic input] |
| |
| _pickle.dump |
| |
| obj: object |
| file: object |
| protocol: object = NULL |
| * |
| fix_imports: bool = True |
| |
| Write a pickled representation of obj to the open file object file. |
| |
| This is equivalent to ``Pickler(file, protocol).dump(obj)``, but may |
| be more efficient. |
| |
| The optional *protocol* argument tells the pickler to use the given |
| protocol supported protocols are 0, 1, 2, 3 and 4. The default |
| protocol is 3; a backward-incompatible protocol designed for Python 3. |
| |
| Specifying a negative protocol version selects the highest protocol |
| version supported. The higher the protocol used, the more recent the |
| version of Python needed to read the pickle produced. |
| |
| The *file* argument must have a write() method that accepts a single |
| bytes argument. It can thus be a file object opened for binary |
| writing, an io.BytesIO instance, or any other custom object that meets |
| this interface. |
| |
| If *fix_imports* is True and protocol is less than 3, pickle will try |
| to map the new Python 3 names to the old module names used in Python |
| 2, so that the pickle data stream is readable with Python 2. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_dump_impl(PyModuleDef *module, PyObject *obj, PyObject *file, |
| PyObject *protocol, int fix_imports) |
| /*[clinic end generated code: output=0de7dff89c406816 input=830f8a64cef6f042]*/ |
| { |
| PicklerObject *pickler = _Pickler_New(); |
| |
| if (pickler == NULL) |
| return NULL; |
| |
| if (_Pickler_SetProtocol(pickler, protocol, fix_imports) < 0) |
| goto error; |
| |
| if (_Pickler_SetOutputStream(pickler, file) < 0) |
| goto error; |
| |
| if (dump(pickler, obj) < 0) |
| goto error; |
| |
| if (_Pickler_FlushToFile(pickler) < 0) |
| goto error; |
| |
| Py_DECREF(pickler); |
| Py_RETURN_NONE; |
| |
| error: |
| Py_XDECREF(pickler); |
| return NULL; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.dumps |
| |
| obj: object |
| protocol: object = NULL |
| * |
| fix_imports: bool = True |
| |
| Return the pickled representation of the object as a bytes object. |
| |
| The optional *protocol* argument tells the pickler to use the given |
| protocol; supported protocols are 0, 1, 2, 3 and 4. The default |
| protocol is 3; a backward-incompatible protocol designed for Python 3. |
| |
| Specifying a negative protocol version selects the highest protocol |
| version supported. The higher the protocol used, the more recent the |
| version of Python needed to read the pickle produced. |
| |
| If *fix_imports* is True and *protocol* is less than 3, pickle will |
| try to map the new Python 3 names to the old module names used in |
| Python 2, so that the pickle data stream is readable with Python 2. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_dumps_impl(PyModuleDef *module, PyObject *obj, PyObject *protocol, |
| int fix_imports) |
| /*[clinic end generated code: output=daa380db56fe07b9 input=293dbeda181580b7]*/ |
| { |
| PyObject *result; |
| PicklerObject *pickler = _Pickler_New(); |
| |
| if (pickler == NULL) |
| return NULL; |
| |
| if (_Pickler_SetProtocol(pickler, protocol, fix_imports) < 0) |
| goto error; |
| |
| if (dump(pickler, obj) < 0) |
| goto error; |
| |
| result = _Pickler_GetString(pickler); |
| Py_DECREF(pickler); |
| return result; |
| |
| error: |
| Py_XDECREF(pickler); |
| return NULL; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.load |
| |
| file: object |
| * |
| fix_imports: bool = True |
| encoding: str = 'ASCII' |
| errors: str = 'strict' |
| |
| Read and return an object from the pickle data stored in a file. |
| |
| This is equivalent to ``Unpickler(file).load()``, but may be more |
| efficient. |
| |
| The protocol version of the pickle is detected automatically, so no |
| protocol argument is needed. Bytes past the pickled object's |
| representation are ignored. |
| |
| The argument *file* must have two methods, a read() method that takes |
| an integer argument, and a readline() method that requires no |
| arguments. Both methods should return bytes. Thus *file* can be a |
| binary file object opened for reading, an io.BytesIO object, or any |
| other custom object that meets this interface. |
| |
| Optional keyword arguments are *fix_imports*, *encoding* and *errors*, |
| which are used to control compatiblity support for pickle stream |
| generated by Python 2. If *fix_imports* is True, pickle will try to |
| map the old Python 2 names to the new names used in Python 3. The |
| *encoding* and *errors* tell pickle how to decode 8-bit string |
| instances pickled by Python 2; these default to 'ASCII' and 'strict', |
| respectively. The *encoding* can be 'bytes' to read these 8-bit |
| string instances as bytes objects. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_load_impl(PyModuleDef *module, PyObject *file, int fix_imports, |
| const char *encoding, const char *errors) |
| /*[clinic end generated code: output=798f1c57cb2b4eb1 input=2df7c7a1e6742204]*/ |
| { |
| PyObject *result; |
| UnpicklerObject *unpickler = _Unpickler_New(); |
| |
| if (unpickler == NULL) |
| return NULL; |
| |
| if (_Unpickler_SetInputStream(unpickler, file) < 0) |
| goto error; |
| |
| if (_Unpickler_SetInputEncoding(unpickler, encoding, errors) < 0) |
| goto error; |
| |
| unpickler->fix_imports = fix_imports; |
| |
| result = load(unpickler); |
| Py_DECREF(unpickler); |
| return result; |
| |
| error: |
| Py_XDECREF(unpickler); |
| return NULL; |
| } |
| |
| /*[clinic input] |
| |
| _pickle.loads |
| |
| data: object |
| * |
| fix_imports: bool = True |
| encoding: str = 'ASCII' |
| errors: str = 'strict' |
| |
| Read and return an object from the given pickle data. |
| |
| The protocol version of the pickle is detected automatically, so no |
| protocol argument is needed. Bytes past the pickled object's |
| representation are ignored. |
| |
| Optional keyword arguments are *fix_imports*, *encoding* and *errors*, |
| which are used to control compatiblity support for pickle stream |
| generated by Python 2. If *fix_imports* is True, pickle will try to |
| map the old Python 2 names to the new names used in Python 3. The |
| *encoding* and *errors* tell pickle how to decode 8-bit string |
| instances pickled by Python 2; these default to 'ASCII' and 'strict', |
| respectively. The *encoding* can be 'bytes' to read these 8-bit |
| string instances as bytes objects. |
| [clinic start generated code]*/ |
| |
| static PyObject * |
| _pickle_loads_impl(PyModuleDef *module, PyObject *data, int fix_imports, |
| const char *encoding, const char *errors) |
| /*[clinic end generated code: output=61e9cdb01e36a736 input=f57f0fdaa2b4cb8b]*/ |
| { |
| PyObject *result; |
| UnpicklerObject *unpickler = _Unpickler_New(); |
| |
| if (unpickler == NULL) |
| return NULL; |
| |
| if (_Unpickler_SetStringInput(unpickler, data) < 0) |
| goto error; |
| |
| if (_Unpickler_SetInputEncoding(unpickler, encoding, errors) < 0) |
| goto error; |
| |
| unpickler->fix_imports = fix_imports; |
| |
| result = load(unpickler); |
| Py_DECREF(unpickler); |
| return result; |
| |
| error: |
| Py_XDECREF(unpickler); |
| return NULL; |
| } |
| |
| static struct PyMethodDef pickle_methods[] = { |
| _PICKLE_DUMP_METHODDEF |
| _PICKLE_DUMPS_METHODDEF |
| _PICKLE_LOAD_METHODDEF |
| _PICKLE_LOADS_METHODDEF |
| {NULL, NULL} /* sentinel */ |
| }; |
| |
| static int |
| pickle_clear(PyObject *m) |
| { |
| _Pickle_ClearState(_Pickle_GetState(m)); |
| return 0; |
| } |
| |
| static void |
| pickle_free(PyObject *m) |
| { |
| _Pickle_ClearState(_Pickle_GetState(m)); |
| } |
| |
| static int |
| pickle_traverse(PyObject *m, visitproc visit, void *arg) |
| { |
| PickleState *st = _Pickle_GetState(m); |
| Py_VISIT(st->PickleError); |
| Py_VISIT(st->PicklingError); |
| Py_VISIT(st->UnpicklingError); |
| Py_VISIT(st->dispatch_table); |
| Py_VISIT(st->extension_registry); |
| Py_VISIT(st->extension_cache); |
| Py_VISIT(st->inverted_registry); |
| Py_VISIT(st->name_mapping_2to3); |
| Py_VISIT(st->import_mapping_2to3); |
| Py_VISIT(st->name_mapping_3to2); |
| Py_VISIT(st->import_mapping_3to2); |
| Py_VISIT(st->codecs_encode); |
| Py_VISIT(st->getattr); |
| return 0; |
| } |
| |
| static struct PyModuleDef _picklemodule = { |
| PyModuleDef_HEAD_INIT, |
| "_pickle", /* m_name */ |
| pickle_module_doc, /* m_doc */ |
| sizeof(PickleState), /* m_size */ |
| pickle_methods, /* m_methods */ |
| NULL, /* m_reload */ |
| pickle_traverse, /* m_traverse */ |
| pickle_clear, /* m_clear */ |
| (freefunc)pickle_free /* m_free */ |
| }; |
| |
| PyMODINIT_FUNC |
| PyInit__pickle(void) |
| { |
| PyObject *m; |
| PickleState *st; |
| |
| m = PyState_FindModule(&_picklemodule); |
| if (m) { |
| Py_INCREF(m); |
| return m; |
| } |
| |
| if (PyType_Ready(&Unpickler_Type) < 0) |
| return NULL; |
| if (PyType_Ready(&Pickler_Type) < 0) |
| return NULL; |
| if (PyType_Ready(&Pdata_Type) < 0) |
| return NULL; |
| if (PyType_Ready(&PicklerMemoProxyType) < 0) |
| return NULL; |
| if (PyType_Ready(&UnpicklerMemoProxyType) < 0) |
| return NULL; |
| |
| /* Create the module and add the functions. */ |
| m = PyModule_Create(&_picklemodule); |
| if (m == NULL) |
| return NULL; |
| |
| Py_INCREF(&Pickler_Type); |
| if (PyModule_AddObject(m, "Pickler", (PyObject *)&Pickler_Type) < 0) |
| return NULL; |
| Py_INCREF(&Unpickler_Type); |
| if (PyModule_AddObject(m, "Unpickler", (PyObject *)&Unpickler_Type) < 0) |
| return NULL; |
| |
| st = _Pickle_GetState(m); |
| |
| /* Initialize the exceptions. */ |
| st->PickleError = PyErr_NewException("_pickle.PickleError", NULL, NULL); |
| if (st->PickleError == NULL) |
| return NULL; |
| st->PicklingError = \ |
| PyErr_NewException("_pickle.PicklingError", st->PickleError, NULL); |
| if (st->PicklingError == NULL) |
| return NULL; |
| st->UnpicklingError = \ |
| PyErr_NewException("_pickle.UnpicklingError", st->PickleError, NULL); |
| if (st->UnpicklingError == NULL) |
| return NULL; |
| |
| Py_INCREF(st->PickleError); |
| if (PyModule_AddObject(m, "PickleError", st->PickleError) < 0) |
| return NULL; |
| Py_INCREF(st->PicklingError); |
| if (PyModule_AddObject(m, "PicklingError", st->PicklingError) < 0) |
| return NULL; |
| Py_INCREF(st->UnpicklingError); |
| if (PyModule_AddObject(m, "UnpicklingError", st->UnpicklingError) < 0) |
| return NULL; |
| |
| if (_Pickle_InitState(st) < 0) |
| return NULL; |
| |
| return m; |
| } |