|  | 
 | /* Parser-tokenizer link implementation */ | 
 |  | 
 | #include "Python.h" | 
 | #include "tokenizer.h" | 
 | #include "node.h" | 
 | #include "grammar.h" | 
 | #include "parser.h" | 
 | #include "parsetok.h" | 
 | #include "errcode.h" | 
 | #include "graminit.h" | 
 |  | 
 |  | 
 | /* Forward */ | 
 | static node *parsetok(struct tok_state *, grammar *, int, perrdetail *, int *); | 
 | static int initerr(perrdetail *err_ret, PyObject * filename); | 
 |  | 
 | typedef struct { | 
 |     struct { | 
 |         int lineno; | 
 |         char *comment; | 
 |     } *items; | 
 |     size_t size; | 
 |     size_t num_items; | 
 | } growable_comment_array; | 
 |  | 
 | static int | 
 | growable_comment_array_init(growable_comment_array *arr, size_t initial_size) { | 
 |     assert(initial_size > 0); | 
 |     arr->items = malloc(initial_size * sizeof(*arr->items)); | 
 |     arr->size = initial_size; | 
 |     arr->num_items = 0; | 
 |  | 
 |     return arr->items != NULL; | 
 | } | 
 |  | 
 | static int | 
 | growable_comment_array_add(growable_comment_array *arr, int lineno, char *comment) { | 
 |     if (arr->num_items >= arr->size) { | 
 |         arr->size *= 2; | 
 |         arr->items = realloc(arr->items, arr->size * sizeof(*arr->items)); | 
 |         if (!arr->items) { | 
 |             return 0; | 
 |         } | 
 |     } | 
 |  | 
 |     arr->items[arr->num_items].lineno = lineno; | 
 |     arr->items[arr->num_items].comment = comment; | 
 |     arr->num_items++; | 
 |     return 1; | 
 | } | 
 |  | 
 | static void | 
 | growable_comment_array_deallocate(growable_comment_array *arr) { | 
 |     for (unsigned i = 0; i < arr->num_items; i++) { | 
 |         PyObject_FREE(arr->items[i].comment); | 
 |     } | 
 |     free(arr->items); | 
 | } | 
 |  | 
 | /* Parse input coming from a string.  Return error code, print some errors. */ | 
 | node * | 
 | PyParser_ParseString(const char *s, grammar *g, int start, perrdetail *err_ret) | 
 | { | 
 |     return PyParser_ParseStringFlagsFilename(s, NULL, g, start, err_ret, 0); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseStringFlags(const char *s, grammar *g, int start, | 
 |                           perrdetail *err_ret, int flags) | 
 | { | 
 |     return PyParser_ParseStringFlagsFilename(s, NULL, | 
 |                                              g, start, err_ret, flags); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseStringFlagsFilename(const char *s, const char *filename, | 
 |                           grammar *g, int start, | 
 |                           perrdetail *err_ret, int flags) | 
 | { | 
 |     int iflags = flags; | 
 |     return PyParser_ParseStringFlagsFilenameEx(s, filename, g, start, | 
 |                                                err_ret, &iflags); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseStringObject(const char *s, PyObject *filename, | 
 |                            grammar *g, int start, | 
 |                            perrdetail *err_ret, int *flags) | 
 | { | 
 |     struct tok_state *tok; | 
 |     int exec_input = start == file_input; | 
 |  | 
 |     if (initerr(err_ret, filename) < 0) | 
 |         return NULL; | 
 |  | 
 |     if (PySys_Audit("compile", "yO", s, err_ret->filename) < 0) { | 
 |         err_ret->error = E_ERROR; | 
 |         return NULL; | 
 |     } | 
 |  | 
 |     if (*flags & PyPARSE_IGNORE_COOKIE) | 
 |         tok = PyTokenizer_FromUTF8(s, exec_input); | 
 |     else | 
 |         tok = PyTokenizer_FromString(s, exec_input); | 
 |     if (tok == NULL) { | 
 |         err_ret->error = PyErr_Occurred() ? E_DECODE : E_NOMEM; | 
 |         return NULL; | 
 |     } | 
 |     if (*flags & PyPARSE_TYPE_COMMENTS) { | 
 |         tok->type_comments = 1; | 
 |     } | 
 |  | 
 |     Py_INCREF(err_ret->filename); | 
 |     tok->filename = err_ret->filename; | 
 |     if (*flags & PyPARSE_ASYNC_HACKS) | 
 |         tok->async_hacks = 1; | 
 |     return parsetok(tok, g, start, err_ret, flags); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseStringFlagsFilenameEx(const char *s, const char *filename_str, | 
 |                           grammar *g, int start, | 
 |                           perrdetail *err_ret, int *flags) | 
 | { | 
 |     node *n; | 
 |     PyObject *filename = NULL; | 
 |     if (filename_str != NULL) { | 
 |         filename = PyUnicode_DecodeFSDefault(filename_str); | 
 |         if (filename == NULL) { | 
 |             err_ret->error = E_ERROR; | 
 |             return NULL; | 
 |         } | 
 |     } | 
 |     n = PyParser_ParseStringObject(s, filename, g, start, err_ret, flags); | 
 |     Py_XDECREF(filename); | 
 |     return n; | 
 | } | 
 |  | 
 | /* Parse input coming from a file.  Return error code, print some errors. */ | 
 |  | 
 | node * | 
 | PyParser_ParseFile(FILE *fp, const char *filename, grammar *g, int start, | 
 |                    const char *ps1, const char *ps2, | 
 |                    perrdetail *err_ret) | 
 | { | 
 |     return PyParser_ParseFileFlags(fp, filename, NULL, | 
 |                                    g, start, ps1, ps2, err_ret, 0); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseFileFlags(FILE *fp, const char *filename, const char *enc, | 
 |                         grammar *g, int start, | 
 |                         const char *ps1, const char *ps2, | 
 |                         perrdetail *err_ret, int flags) | 
 | { | 
 |     int iflags = flags; | 
 |     return PyParser_ParseFileFlagsEx(fp, filename, enc, g, start, ps1, | 
 |                                      ps2, err_ret, &iflags); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseFileObject(FILE *fp, PyObject *filename, | 
 |                          const char *enc, grammar *g, int start, | 
 |                          const char *ps1, const char *ps2, | 
 |                          perrdetail *err_ret, int *flags) | 
 | { | 
 |     struct tok_state *tok; | 
 |  | 
 |     if (initerr(err_ret, filename) < 0) | 
 |         return NULL; | 
 |  | 
 |     if (PySys_Audit("compile", "OO", Py_None, err_ret->filename) < 0) { | 
 |         return NULL; | 
 |     } | 
 |  | 
 |     if ((tok = PyTokenizer_FromFile(fp, enc, ps1, ps2)) == NULL) { | 
 |         err_ret->error = E_NOMEM; | 
 |         return NULL; | 
 |     } | 
 |     if (*flags & PyPARSE_TYPE_COMMENTS) { | 
 |         tok->type_comments = 1; | 
 |     } | 
 |     Py_INCREF(err_ret->filename); | 
 |     tok->filename = err_ret->filename; | 
 |     return parsetok(tok, g, start, err_ret, flags); | 
 | } | 
 |  | 
 | node * | 
 | PyParser_ParseFileFlagsEx(FILE *fp, const char *filename, | 
 |                           const char *enc, grammar *g, int start, | 
 |                           const char *ps1, const char *ps2, | 
 |                           perrdetail *err_ret, int *flags) | 
 | { | 
 |     node *n; | 
 |     PyObject *fileobj = NULL; | 
 |     if (filename != NULL) { | 
 |         fileobj = PyUnicode_DecodeFSDefault(filename); | 
 |         if (fileobj == NULL) { | 
 |             err_ret->error = E_ERROR; | 
 |             return NULL; | 
 |         } | 
 |     } | 
 |     n = PyParser_ParseFileObject(fp, fileobj, enc, g, | 
 |                                  start, ps1, ps2, err_ret, flags); | 
 |     Py_XDECREF(fileobj); | 
 |     return n; | 
 | } | 
 |  | 
 | #ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | 
 | #if 0 | 
 | static const char with_msg[] = | 
 | "%s:%d: Warning: 'with' will become a reserved keyword in Python 2.6\n"; | 
 |  | 
 | static const char as_msg[] = | 
 | "%s:%d: Warning: 'as' will become a reserved keyword in Python 2.6\n"; | 
 |  | 
 | static void | 
 | warn(const char *msg, const char *filename, int lineno) | 
 | { | 
 |     if (filename == NULL) | 
 |         filename = "<string>"; | 
 |     PySys_WriteStderr(msg, filename, lineno); | 
 | } | 
 | #endif | 
 | #endif | 
 |  | 
 | /* Parse input coming from the given tokenizer structure. | 
 |    Return error code. */ | 
 |  | 
 | static node * | 
 | parsetok(struct tok_state *tok, grammar *g, int start, perrdetail *err_ret, | 
 |          int *flags) | 
 | { | 
 |     parser_state *ps; | 
 |     node *n; | 
 |     int started = 0; | 
 |     int col_offset, end_col_offset; | 
 |     growable_comment_array type_ignores; | 
 |  | 
 |     if (!growable_comment_array_init(&type_ignores, 10)) { | 
 |         err_ret->error = E_NOMEM; | 
 |         PyTokenizer_Free(tok); | 
 |         return NULL; | 
 |     } | 
 |  | 
 |     if ((ps = PyParser_New(g, start)) == NULL) { | 
 |         err_ret->error = E_NOMEM; | 
 |         PyTokenizer_Free(tok); | 
 |         return NULL; | 
 |     } | 
 | #ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | 
 |     if (*flags & PyPARSE_BARRY_AS_BDFL) | 
 |         ps->p_flags |= CO_FUTURE_BARRY_AS_BDFL; | 
 |     if (*flags & PyPARSE_TYPE_COMMENTS) | 
 |         ps->p_flags |= PyCF_TYPE_COMMENTS; | 
 | #endif | 
 |  | 
 |     for (;;) { | 
 |         char *a, *b; | 
 |         int type; | 
 |         size_t len; | 
 |         char *str; | 
 |         col_offset = -1; | 
 |         int lineno; | 
 |         const char *line_start; | 
 |  | 
 |         type = PyTokenizer_Get(tok, &a, &b); | 
 |         if (type == ERRORTOKEN) { | 
 |             err_ret->error = tok->done; | 
 |             break; | 
 |         } | 
 |         if (type == ENDMARKER && started) { | 
 |             type = NEWLINE; /* Add an extra newline */ | 
 |             started = 0; | 
 |             /* Add the right number of dedent tokens, | 
 |                except if a certain flag is given -- | 
 |                codeop.py uses this. */ | 
 |             if (tok->indent && | 
 |                 !(*flags & PyPARSE_DONT_IMPLY_DEDENT)) | 
 |             { | 
 |                 tok->pendin = -tok->indent; | 
 |                 tok->indent = 0; | 
 |             } | 
 |         } | 
 |         else | 
 |             started = 1; | 
 |         len = (a != NULL && b != NULL) ? b - a : 0; | 
 |         str = (char *) PyObject_MALLOC(len + 1); | 
 |         if (str == NULL) { | 
 |             err_ret->error = E_NOMEM; | 
 |             break; | 
 |         } | 
 |         if (len > 0) | 
 |             strncpy(str, a, len); | 
 |         str[len] = '\0'; | 
 |  | 
 | #ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | 
 |         if (type == NOTEQUAL) { | 
 |             if (!(ps->p_flags & CO_FUTURE_BARRY_AS_BDFL) && | 
 |                             strcmp(str, "!=")) { | 
 |                 PyObject_FREE(str); | 
 |                 err_ret->error = E_SYNTAX; | 
 |                 break; | 
 |             } | 
 |             else if ((ps->p_flags & CO_FUTURE_BARRY_AS_BDFL) && | 
 |                             strcmp(str, "<>")) { | 
 |                 PyObject_FREE(str); | 
 |                 err_ret->expected = NOTEQUAL; | 
 |                 err_ret->error = E_SYNTAX; | 
 |                 break; | 
 |             } | 
 |         } | 
 | #endif | 
 |  | 
 |         /* Nodes of type STRING, especially multi line strings | 
 |            must be handled differently in order to get both | 
 |            the starting line number and the column offset right. | 
 |            (cf. issue 16806) */ | 
 |         lineno = type == STRING ? tok->first_lineno : tok->lineno; | 
 |         line_start = type == STRING ? tok->multi_line_start : tok->line_start; | 
 |         if (a != NULL && a >= line_start) { | 
 |             col_offset = Py_SAFE_DOWNCAST(a - line_start, | 
 |                                           intptr_t, int); | 
 |         } | 
 |         else { | 
 |             col_offset = -1; | 
 |         } | 
 |  | 
 |         if (b != NULL && b >= tok->line_start) { | 
 |             end_col_offset = Py_SAFE_DOWNCAST(b - tok->line_start, | 
 |                                               intptr_t, int); | 
 |         } | 
 |         else { | 
 |             end_col_offset = -1; | 
 |         } | 
 |  | 
 |         if (type == TYPE_IGNORE) { | 
 |             if (!growable_comment_array_add(&type_ignores, tok->lineno, str)) { | 
 |                 err_ret->error = E_NOMEM; | 
 |                 break; | 
 |             } | 
 |             continue; | 
 |         } | 
 |  | 
 |         if ((err_ret->error = | 
 |              PyParser_AddToken(ps, (int)type, str, | 
 |                                lineno, col_offset, tok->lineno, end_col_offset, | 
 |                                &(err_ret->expected))) != E_OK) { | 
 |             if (err_ret->error != E_DONE) { | 
 |                 PyObject_FREE(str); | 
 |                 err_ret->token = type; | 
 |             } | 
 |             break; | 
 |         } | 
 |     } | 
 |  | 
 |     if (err_ret->error == E_DONE) { | 
 |         n = ps->p_tree; | 
 |         ps->p_tree = NULL; | 
 |  | 
 |         if (n->n_type == file_input) { | 
 |             /* Put type_ignore nodes in the ENDMARKER of file_input. */ | 
 |             int num; | 
 |             node *ch; | 
 |             size_t i; | 
 |  | 
 |             num = NCH(n); | 
 |             ch = CHILD(n, num - 1); | 
 |             REQ(ch, ENDMARKER); | 
 |  | 
 |             for (i = 0; i < type_ignores.num_items; i++) { | 
 |                 int res = PyNode_AddChild(ch, TYPE_IGNORE, type_ignores.items[i].comment, | 
 |                                           type_ignores.items[i].lineno, 0, | 
 |                                           type_ignores.items[i].lineno, 0); | 
 |                 if (res != 0) { | 
 |                     err_ret->error = res; | 
 |                     PyNode_Free(n); | 
 |                     n = NULL; | 
 |                     break; | 
 |                 } | 
 |                 type_ignores.items[i].comment = NULL; | 
 |             } | 
 |         } | 
 |  | 
 |         /* Check that the source for a single input statement really | 
 |            is a single statement by looking at what is left in the | 
 |            buffer after parsing.  Trailing whitespace and comments | 
 |            are OK.  */ | 
 |         if (err_ret->error == E_DONE && start == single_input) { | 
 |             char *cur = tok->cur; | 
 |             char c = *tok->cur; | 
 |  | 
 |             for (;;) { | 
 |                 while (c == ' ' || c == '\t' || c == '\n' || c == '\014') | 
 |                     c = *++cur; | 
 |  | 
 |                 if (!c) | 
 |                     break; | 
 |  | 
 |                 if (c != '#') { | 
 |                     err_ret->error = E_BADSINGLE; | 
 |                     PyNode_Free(n); | 
 |                     n = NULL; | 
 |                     break; | 
 |                 } | 
 |  | 
 |                 /* Suck up comment. */ | 
 |                 while (c && c != '\n') | 
 |                     c = *++cur; | 
 |             } | 
 |         } | 
 |     } | 
 |     else | 
 |         n = NULL; | 
 |  | 
 |     growable_comment_array_deallocate(&type_ignores); | 
 |  | 
 | #ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | 
 |     *flags = ps->p_flags; | 
 | #endif | 
 |     PyParser_Delete(ps); | 
 |  | 
 |     if (n == NULL) { | 
 |         if (tok->done == E_EOF) | 
 |             err_ret->error = E_EOF; | 
 |         err_ret->lineno = tok->lineno; | 
 |         if (tok->buf != NULL) { | 
 |             size_t len; | 
 |             assert(tok->cur - tok->buf < INT_MAX); | 
 |             /* if we've managed to parse a token, point the offset to its start, | 
 |              * else use the current reading position of the tokenizer | 
 |              */ | 
 |             err_ret->offset = col_offset != -1 ? col_offset + 1 : ((int)(tok->cur - tok->buf)); | 
 |             len = tok->inp - tok->buf; | 
 |             err_ret->text = (char *) PyObject_MALLOC(len + 1); | 
 |             if (err_ret->text != NULL) { | 
 |                 if (len > 0) | 
 |                     strncpy(err_ret->text, tok->buf, len); | 
 |                 err_ret->text[len] = '\0'; | 
 |             } | 
 |         } | 
 |     } else if (tok->encoding != NULL) { | 
 |         /* 'nodes->n_str' uses PyObject_*, while 'tok->encoding' was | 
 |          * allocated using PyMem_ | 
 |          */ | 
 |         node* r = PyNode_New(encoding_decl); | 
 |         if (r) | 
 |             r->n_str = PyObject_MALLOC(strlen(tok->encoding)+1); | 
 |         if (!r || !r->n_str) { | 
 |             err_ret->error = E_NOMEM; | 
 |             if (r) | 
 |                 PyObject_FREE(r); | 
 |             n = NULL; | 
 |             goto done; | 
 |         } | 
 |         strcpy(r->n_str, tok->encoding); | 
 |         PyMem_FREE(tok->encoding); | 
 |         tok->encoding = NULL; | 
 |         r->n_nchildren = 1; | 
 |         r->n_child = n; | 
 |         n = r; | 
 |     } | 
 |  | 
 | done: | 
 |     PyTokenizer_Free(tok); | 
 |  | 
 |     if (n != NULL) { | 
 |         _PyNode_FinalizeEndPos(n); | 
 |     } | 
 |     return n; | 
 | } | 
 |  | 
 | static int | 
 | initerr(perrdetail *err_ret, PyObject *filename) | 
 | { | 
 |     err_ret->error = E_OK; | 
 |     err_ret->lineno = 0; | 
 |     err_ret->offset = 0; | 
 |     err_ret->text = NULL; | 
 |     err_ret->token = -1; | 
 |     err_ret->expected = -1; | 
 |     if (filename) { | 
 |         Py_INCREF(filename); | 
 |         err_ret->filename = filename; | 
 |     } | 
 |     else { | 
 |         err_ret->filename = PyUnicode_FromString("<string>"); | 
 |         if (err_ret->filename == NULL) { | 
 |             err_ret->error = E_ERROR; | 
 |             return -1; | 
 |         } | 
 |     } | 
 |     return 0; | 
 | } |