mirror of
https://github.com/python/cpython.git
synced 2025-08-22 17:55:18 +00:00
bpo-45434: Mark the PyTokenizer C API as private (GH-28924)
Rename PyTokenize functions to mark them as private: * PyTokenizer_FindEncodingFilename() => _PyTokenizer_FindEncodingFilename() * PyTokenizer_FromString() => _PyTokenizer_FromString() * PyTokenizer_FromFile() => _PyTokenizer_FromFile() * PyTokenizer_FromUTF8() => _PyTokenizer_FromUTF8() * PyTokenizer_Free() => _PyTokenizer_Free() * PyTokenizer_Get() => _PyTokenizer_Get() Remove the unused PyTokenizer_FindEncoding() function. import.c: remove unused #include "errcode.h".
This commit is contained in:
parent
3901c08114
commit
713bb19356
7 changed files with 40 additions and 48 deletions
|
@ -47,7 +47,7 @@ tokenizeriter_new_impl(PyTypeObject *type, const char *source)
|
|||
if (filename == NULL) {
|
||||
return NULL;
|
||||
}
|
||||
self->tok = PyTokenizer_FromUTF8(source, 1);
|
||||
self->tok = _PyTokenizer_FromUTF8(source, 1);
|
||||
if (self->tok == NULL) {
|
||||
Py_DECREF(filename);
|
||||
return NULL;
|
||||
|
@ -61,7 +61,7 @@ tokenizeriter_next(tokenizeriterobject *it)
|
|||
{
|
||||
const char *start;
|
||||
const char *end;
|
||||
int type = PyTokenizer_Get(it->tok, &start, &end);
|
||||
int type = _PyTokenizer_Get(it->tok, &start, &end);
|
||||
if (type == ERRORTOKEN && PyErr_Occurred()) {
|
||||
return NULL;
|
||||
}
|
||||
|
@ -105,7 +105,7 @@ static void
|
|||
tokenizeriter_dealloc(tokenizeriterobject *it)
|
||||
{
|
||||
PyTypeObject *tp = Py_TYPE(it);
|
||||
PyTokenizer_Free(it->tok);
|
||||
_PyTokenizer_Free(it->tok);
|
||||
tp->tp_free(it);
|
||||
Py_DECREF(tp);
|
||||
}
|
||||
|
|
|
@ -11,7 +11,6 @@
|
|||
#include "pycore_interp.h" // _PyInterpreterState_ClearModules()
|
||||
#include "pycore_pystate.h" // _PyInterpreterState_GET()
|
||||
#include "pycore_sysmodule.h"
|
||||
#include "errcode.h"
|
||||
#include "marshal.h"
|
||||
#include "code.h"
|
||||
#include "importdl.h"
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
#define MAX_NTHREADS 100
|
||||
|
||||
/* Function from Parser/tokenizer.c */
|
||||
extern char * PyTokenizer_FindEncodingFilename(int, PyObject *);
|
||||
extern char* _PyTokenizer_FindEncodingFilename(int, PyObject *);
|
||||
|
||||
_Py_IDENTIFIER(TextIOWrapper);
|
||||
_Py_IDENTIFIER(close);
|
||||
|
@ -431,7 +431,7 @@ _Py_DisplaySourceLine(PyObject *f, PyObject *filename, int lineno, int indent, i
|
|||
Py_DECREF(binary);
|
||||
return 0;
|
||||
}
|
||||
found_encoding = PyTokenizer_FindEncodingFilename(fd, filename);
|
||||
found_encoding = _PyTokenizer_FindEncodingFilename(fd, filename);
|
||||
if (found_encoding == NULL)
|
||||
PyErr_Clear();
|
||||
encoding = (found_encoding != NULL) ? found_encoding : "utf-8";
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue