mirror of
https://github.com/python/cpython.git
synced 2025-08-02 16:13:13 +00:00
bpo-45434: Mark the PyTokenizer C API as private (GH-28924)
Rename PyTokenize functions to mark them as private: * PyTokenizer_FindEncodingFilename() => _PyTokenizer_FindEncodingFilename() * PyTokenizer_FromString() => _PyTokenizer_FromString() * PyTokenizer_FromFile() => _PyTokenizer_FromFile() * PyTokenizer_FromUTF8() => _PyTokenizer_FromUTF8() * PyTokenizer_Free() => _PyTokenizer_Free() * PyTokenizer_Get() => _PyTokenizer_Get() Remove the unused PyTokenizer_FindEncoding() function. import.c: remove unused #include "errcode.h".
This commit is contained in:
parent
3901c08114
commit
713bb19356
7 changed files with 40 additions and 48 deletions
|
@ -108,7 +108,7 @@ static char *
|
|||
error_ret(struct tok_state *tok) /* XXX */
|
||||
{
|
||||
tok->decoding_erred = 1;
|
||||
if (tok->fp != NULL && tok->buf != NULL) /* see PyTokenizer_Free */
|
||||
if (tok->fp != NULL && tok->buf != NULL) /* see _PyTokenizer_Free */
|
||||
PyMem_Free(tok->buf);
|
||||
tok->buf = tok->cur = tok->inp = NULL;
|
||||
tok->start = NULL;
|
||||
|
@ -702,7 +702,7 @@ decode_str(const char *input, int single, struct tok_state *tok)
|
|||
/* Set up tokenizer for string */
|
||||
|
||||
struct tok_state *
|
||||
PyTokenizer_FromString(const char *str, int exec_input)
|
||||
_PyTokenizer_FromString(const char *str, int exec_input)
|
||||
{
|
||||
struct tok_state *tok = tok_new();
|
||||
char *decoded;
|
||||
|
@ -711,7 +711,7 @@ PyTokenizer_FromString(const char *str, int exec_input)
|
|||
return NULL;
|
||||
decoded = decode_str(str, exec_input, tok);
|
||||
if (decoded == NULL) {
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
|
@ -723,7 +723,7 @@ PyTokenizer_FromString(const char *str, int exec_input)
|
|||
/* Set up tokenizer for UTF-8 string */
|
||||
|
||||
struct tok_state *
|
||||
PyTokenizer_FromUTF8(const char *str, int exec_input)
|
||||
_PyTokenizer_FromUTF8(const char *str, int exec_input)
|
||||
{
|
||||
struct tok_state *tok = tok_new();
|
||||
char *translated;
|
||||
|
@ -731,7 +731,7 @@ PyTokenizer_FromUTF8(const char *str, int exec_input)
|
|||
return NULL;
|
||||
tok->input = translated = translate_newlines(str, exec_input, tok);
|
||||
if (translated == NULL) {
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return NULL;
|
||||
}
|
||||
tok->decoding_state = STATE_NORMAL;
|
||||
|
@ -739,7 +739,7 @@ PyTokenizer_FromUTF8(const char *str, int exec_input)
|
|||
tok->str = translated;
|
||||
tok->encoding = new_string("utf-8", 5, tok);
|
||||
if (!tok->encoding) {
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
|
@ -751,14 +751,14 @@ PyTokenizer_FromUTF8(const char *str, int exec_input)
|
|||
/* Set up tokenizer for file */
|
||||
|
||||
struct tok_state *
|
||||
PyTokenizer_FromFile(FILE *fp, const char* enc,
|
||||
const char *ps1, const char *ps2)
|
||||
_PyTokenizer_FromFile(FILE *fp, const char* enc,
|
||||
const char *ps1, const char *ps2)
|
||||
{
|
||||
struct tok_state *tok = tok_new();
|
||||
if (tok == NULL)
|
||||
return NULL;
|
||||
if ((tok->buf = (char *)PyMem_Malloc(BUFSIZ)) == NULL) {
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return NULL;
|
||||
}
|
||||
tok->cur = tok->inp = tok->buf;
|
||||
|
@ -771,7 +771,7 @@ PyTokenizer_FromFile(FILE *fp, const char* enc,
|
|||
gets copied into the parse tree. */
|
||||
tok->encoding = new_string(enc, strlen(enc), tok);
|
||||
if (!tok->encoding) {
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return NULL;
|
||||
}
|
||||
tok->decoding_state = STATE_NORMAL;
|
||||
|
@ -782,7 +782,7 @@ PyTokenizer_FromFile(FILE *fp, const char* enc,
|
|||
/* Free a tok_state structure */
|
||||
|
||||
void
|
||||
PyTokenizer_Free(struct tok_state *tok)
|
||||
_PyTokenizer_Free(struct tok_state *tok)
|
||||
{
|
||||
if (tok->encoding != NULL) {
|
||||
PyMem_Free(tok->encoding);
|
||||
|
@ -2049,7 +2049,8 @@ tok_get(struct tok_state *tok, const char **p_start, const char **p_end)
|
|||
}
|
||||
|
||||
int
|
||||
PyTokenizer_Get(struct tok_state *tok, const char **p_start, const char **p_end)
|
||||
_PyTokenizer_Get(struct tok_state *tok,
|
||||
const char **p_start, const char **p_end)
|
||||
{
|
||||
int result = tok_get(tok, p_start, p_end);
|
||||
if (tok->decoding_erred) {
|
||||
|
@ -2062,7 +2063,7 @@ PyTokenizer_Get(struct tok_state *tok, const char **p_start, const char **p_end)
|
|||
/* Get the encoding of a Python file. Check for the coding cookie and check if
|
||||
the file starts with a BOM.
|
||||
|
||||
PyTokenizer_FindEncodingFilename() returns NULL when it can't find the
|
||||
_PyTokenizer_FindEncodingFilename() returns NULL when it can't find the
|
||||
encoding in the first or second line of the file (in which case the encoding
|
||||
should be assumed to be UTF-8).
|
||||
|
||||
|
@ -2070,7 +2071,7 @@ PyTokenizer_Get(struct tok_state *tok, const char **p_start, const char **p_end)
|
|||
by the caller. */
|
||||
|
||||
char *
|
||||
PyTokenizer_FindEncodingFilename(int fd, PyObject *filename)
|
||||
_PyTokenizer_FindEncodingFilename(int fd, PyObject *filename)
|
||||
{
|
||||
struct tok_state *tok;
|
||||
FILE *fp;
|
||||
|
@ -2087,7 +2088,7 @@ PyTokenizer_FindEncodingFilename(int fd, PyObject *filename)
|
|||
if (fp == NULL) {
|
||||
return NULL;
|
||||
}
|
||||
tok = PyTokenizer_FromFile(fp, NULL, NULL, NULL);
|
||||
tok = _PyTokenizer_FromFile(fp, NULL, NULL, NULL);
|
||||
if (tok == NULL) {
|
||||
fclose(fp);
|
||||
return NULL;
|
||||
|
@ -2100,12 +2101,12 @@ PyTokenizer_FindEncodingFilename(int fd, PyObject *filename)
|
|||
tok->filename = PyUnicode_FromString("<string>");
|
||||
if (tok->filename == NULL) {
|
||||
fclose(fp);
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return encoding;
|
||||
}
|
||||
}
|
||||
while (tok->lineno < 2 && tok->done == E_OK) {
|
||||
PyTokenizer_Get(tok, &p_start, &p_end);
|
||||
_PyTokenizer_Get(tok, &p_start, &p_end);
|
||||
}
|
||||
fclose(fp);
|
||||
if (tok->encoding) {
|
||||
|
@ -2114,18 +2115,11 @@ PyTokenizer_FindEncodingFilename(int fd, PyObject *filename)
|
|||
strcpy(encoding, tok->encoding);
|
||||
}
|
||||
}
|
||||
PyTokenizer_Free(tok);
|
||||
_PyTokenizer_Free(tok);
|
||||
return encoding;
|
||||
}
|
||||
|
||||
char *
|
||||
PyTokenizer_FindEncoding(int fd)
|
||||
{
|
||||
return PyTokenizer_FindEncodingFilename(fd, NULL);
|
||||
}
|
||||
|
||||
#ifdef Py_DEBUG
|
||||
|
||||
void
|
||||
tok_dump(int type, char *start, char *end)
|
||||
{
|
||||
|
@ -2133,5 +2127,4 @@ tok_dump(int type, char *start, char *end)
|
|||
if (type == NAME || type == NUMBER || type == STRING || type == OP)
|
||||
printf("(%.*s)", (int)(end - start), start);
|
||||
}
|
||||
|
||||
#endif
|
||||
#endif // Py_DEBUG
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue