mirror of
https://github.com/python/cpython
synced 2024-11-02 08:22:14 +00:00
ff0ec52d3c
interactive EOF, which leaked *one* byte).
192 lines
4.5 KiB
C
192 lines
4.5 KiB
C
/***********************************************************
|
|
Copyright 1991-1995 by Stichting Mathematisch Centrum, Amsterdam,
|
|
The Netherlands.
|
|
|
|
All Rights Reserved
|
|
|
|
Permission to use, copy, modify, and distribute this software and its
|
|
documentation for any purpose and without fee is hereby granted,
|
|
provided that the above copyright notice appear in all copies and that
|
|
both that copyright notice and this permission notice appear in
|
|
supporting documentation, and that the names of Stichting Mathematisch
|
|
Centrum or CWI or Corporation for National Research Initiatives or
|
|
CNRI not be used in advertising or publicity pertaining to
|
|
distribution of the software without specific, written prior
|
|
permission.
|
|
|
|
While CWI is the initial source for this software, a modified version
|
|
is made available by the Corporation for National Research Initiatives
|
|
(CNRI) at the Internet address ftp://ftp.python.org.
|
|
|
|
STICHTING MATHEMATISCH CENTRUM AND CNRI DISCLAIM ALL WARRANTIES WITH
|
|
REGARD TO THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF
|
|
MERCHANTABILITY AND FITNESS, IN NO EVENT SHALL STICHTING MATHEMATISCH
|
|
CENTRUM OR CNRI BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL
|
|
DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR
|
|
PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER
|
|
TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR
|
|
PERFORMANCE OF THIS SOFTWARE.
|
|
|
|
******************************************************************/
|
|
|
|
/* Parser-tokenizer link implementation */
|
|
|
|
#include "pgenheaders.h"
|
|
#include "tokenizer.h"
|
|
#include "node.h"
|
|
#include "grammar.h"
|
|
#include "parser.h"
|
|
#include "parsetok.h"
|
|
#include "errcode.h"
|
|
|
|
|
|
/* Forward */
|
|
static node *parsetok Py_PROTO((struct tok_state *, grammar *, int,
|
|
perrdetail *));
|
|
|
|
/* Parse input coming from a string. Return error code, print some errors. */
|
|
|
|
node *
|
|
PyParser_ParseString(s, g, start, err_ret)
|
|
char *s;
|
|
grammar *g;
|
|
int start;
|
|
perrdetail *err_ret;
|
|
{
|
|
struct tok_state *tok;
|
|
|
|
err_ret->error = E_OK;
|
|
err_ret->filename = NULL;
|
|
err_ret->lineno = 0;
|
|
err_ret->offset = 0;
|
|
err_ret->text = NULL;
|
|
|
|
if ((tok = PyTokenizer_FromString(s)) == NULL) {
|
|
err_ret->error = E_NOMEM;
|
|
return NULL;
|
|
}
|
|
|
|
return parsetok(tok, g, start, err_ret);
|
|
}
|
|
|
|
|
|
/* Parse input coming from a file. Return error code, print some errors. */
|
|
|
|
node *
|
|
PyParser_ParseFile(fp, filename, g, start, ps1, ps2, err_ret)
|
|
FILE *fp;
|
|
char *filename;
|
|
grammar *g;
|
|
int start;
|
|
char *ps1, *ps2;
|
|
perrdetail *err_ret;
|
|
{
|
|
struct tok_state *tok;
|
|
|
|
err_ret->error = E_OK;
|
|
err_ret->filename = filename;
|
|
err_ret->lineno = 0;
|
|
err_ret->offset = 0;
|
|
err_ret->text = NULL;
|
|
|
|
if ((tok = PyTokenizer_FromFile(fp, ps1, ps2)) == NULL) {
|
|
err_ret->error = E_NOMEM;
|
|
return NULL;
|
|
}
|
|
|
|
#ifdef macintosh
|
|
{
|
|
int tabsize = guesstabsize(filename);
|
|
if (tabsize > 0)
|
|
tok->tabsize = tabsize;
|
|
}
|
|
#endif
|
|
|
|
return parsetok(tok, g, start, err_ret);
|
|
}
|
|
|
|
/* Parse input coming from the given tokenizer structure.
|
|
Return error code. */
|
|
|
|
static node *
|
|
parsetok(tok, g, start, err_ret)
|
|
struct tok_state *tok;
|
|
grammar *g;
|
|
int start;
|
|
perrdetail *err_ret;
|
|
{
|
|
parser_state *ps;
|
|
node *n;
|
|
int started = 0;
|
|
|
|
if ((ps = PyParser_New(g, start)) == NULL) {
|
|
fprintf(stderr, "no mem for new parser\n");
|
|
err_ret->error = E_NOMEM;
|
|
return NULL;
|
|
}
|
|
|
|
for (;;) {
|
|
char *a, *b;
|
|
int type;
|
|
int len;
|
|
char *str;
|
|
|
|
type = PyTokenizer_Get(tok, &a, &b);
|
|
if (type == ERRORTOKEN) {
|
|
err_ret->error = tok->done;
|
|
break;
|
|
}
|
|
if (type == ENDMARKER && started) {
|
|
type = NEWLINE; /* Add an extra newline */
|
|
started = 0;
|
|
}
|
|
else
|
|
started = 1;
|
|
len = b - a; /* XXX this may compute NULL - NULL */
|
|
str = PyMem_NEW(char, len + 1);
|
|
if (str == NULL) {
|
|
fprintf(stderr, "no mem for next token\n");
|
|
err_ret->error = E_NOMEM;
|
|
break;
|
|
}
|
|
if (len > 0)
|
|
strncpy(str, a, len);
|
|
str[len] = '\0';
|
|
if ((err_ret->error =
|
|
PyParser_AddToken(ps, (int)type, str,
|
|
tok->lineno)) != E_OK) {
|
|
if (err_ret->error != E_DONE)
|
|
PyMem_DEL(str);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (err_ret->error == E_DONE) {
|
|
n = ps->p_tree;
|
|
ps->p_tree = NULL;
|
|
}
|
|
else
|
|
n = NULL;
|
|
|
|
PyParser_Delete(ps);
|
|
|
|
if (n == NULL) {
|
|
if (tok->lineno <= 1 && tok->done == E_EOF)
|
|
err_ret->error = E_EOF;
|
|
err_ret->lineno = tok->lineno;
|
|
err_ret->offset = tok->cur - tok->buf;
|
|
if (tok->buf != NULL) {
|
|
int len = tok->inp - tok->buf;
|
|
err_ret->text = malloc(len + 1);
|
|
if (err_ret->text != NULL) {
|
|
if (len > 0)
|
|
strncpy(err_ret->text, tok->buf, len);
|
|
err_ret->text[len] = '\0';
|
|
}
|
|
}
|
|
}
|
|
|
|
PyTokenizer_Free(tok);
|
|
|
|
return n;
|
|
}
|