mirror of
https://github.com/python/cpython.git
synced 2025-10-22 22:53:06 +00:00

svn+ssh://pythondev@svn.python.org/python/trunk ........ r58096 | brett.cannon | 2007-09-10 23:38:27 +0200 (Mon, 10 Sep 2007) | 4 lines Fix a possible segfault from recursing too deep to get the repr of a list. Closes issue #1096. ........ r58097 | bill.janssen | 2007-09-10 23:51:02 +0200 (Mon, 10 Sep 2007) | 33 lines More work on SSL support. * Much expanded test suite: All protocols tested against all other protocols. All protocols tested with all certificate options. Tests for bad key and bad cert. Test of STARTTLS functionality. Test of RAND_* functions. * Fixes for threading/malloc bug. * Issue 1065 fixed: sslsocket class renamed to SSLSocket. sslerror class renamed to SSLError. Function "wrap_socket" now used to wrap an existing socket. * Issue 1583946 finally fixed: Support for subjectAltName added. Subject name now returned as proper DN list of RDNs. * SSLError exported from socket as "sslerror". * RAND_* functions properly exported from ssl.py. * Documentation improved: Example of how to create a self-signed certificate. Better indexing. ........ r58098 | guido.van.rossum | 2007-09-11 00:02:25 +0200 (Tue, 11 Sep 2007) | 9 lines Patch # 1140 (my code, approved by Effbot). Make sure the type of the return value of re.sub(x, y, z) is the type of y+x (i.e. unicode if either is unicode, str if they are both str) even if there are no substitutions or if x==z (which triggered various special cases in join_list()). Could be backported to 2.5; no need to port to 3.0. ........ r58099 | guido.van.rossum | 2007-09-11 00:36:02 +0200 (Tue, 11 Sep 2007) | 8 lines Patch # 1026 by Benjamin Aranguren (with Alex Martelli): Backport abc.py and isinstance/issubclass overloading to 2.6. I had to backport test_typechecks.py myself, and make one small change to abc.py to avoid duplicate work when x.__class__ and type(x) are the same. ........ r58100 | bill.janssen | 2007-09-11 01:41:24 +0200 (Tue, 11 Sep 2007) | 3 lines A better way of finding an open port to test with. ........ r58101 | bill.janssen | 2007-09-11 03:09:19 +0200 (Tue, 11 Sep 2007) | 4 lines Make sure test_ssl doesn't reference the ssl module in a context where it can't be imported. ........ r58102 | bill.janssen | 2007-09-11 04:42:07 +0200 (Tue, 11 Sep 2007) | 3 lines Fix some documentation bugs. ........ r58103 | nick.coghlan | 2007-09-11 16:01:18 +0200 (Tue, 11 Sep 2007) | 1 line Always use the -E flag when spawning subprocesses in test_cmd_line (Issue 1056) ........ r58106 | thomas.heller | 2007-09-11 21:17:48 +0200 (Tue, 11 Sep 2007) | 3 lines Disable some tests that fail on the 'ppc Debian unstable' buildbot to find out if they cause the segfault on the 'alpha Debian' machine. ........ r58108 | brett.cannon | 2007-09-11 23:02:28 +0200 (Tue, 11 Sep 2007) | 6 lines Generators had their throw() method allowing string exceptions. That's a no-no. Fixes issue #1147. Need to fix 2.5 to raise a proper warning if a string exception is passed in. ........ r58112 | georg.brandl | 2007-09-12 20:03:51 +0200 (Wed, 12 Sep 2007) | 3 lines New documentation page for the bdb module. (This doesn't need to be merged to Py3k.) ........ r58114 | georg.brandl | 2007-09-12 20:05:57 +0200 (Wed, 12 Sep 2007) | 2 lines Bug #1152: use non-deprecated name in example. ........ r58115 | georg.brandl | 2007-09-12 20:08:33 +0200 (Wed, 12 Sep 2007) | 2 lines Fix #1122: wrong return type documented for various _Size() functions. ........ r58117 | georg.brandl | 2007-09-12 20:10:56 +0200 (Wed, 12 Sep 2007) | 2 lines Fix #1139: PyFile_Encoding really is PyFile_SetEncoding. ........ r58119 | georg.brandl | 2007-09-12 20:29:18 +0200 (Wed, 12 Sep 2007) | 2 lines bug #1154: release memory allocated by "es" PyArg_ParseTuple format specifier. ........ r58121 | bill.janssen | 2007-09-12 20:52:05 +0200 (Wed, 12 Sep 2007) | 1 line root certificate for https://svn.python.org/, used in test_ssl ........ r58122 | georg.brandl | 2007-09-12 21:00:07 +0200 (Wed, 12 Sep 2007) | 3 lines Bug #1153: repr.repr() now doesn't require set and dictionary items to be orderable to properly represent them. ........ r58125 | georg.brandl | 2007-09-12 21:29:28 +0200 (Wed, 12 Sep 2007) | 4 lines #1120: put explicit version in the shebang lines of pydoc, idle and smtpd.py scripts that are installed by setup.py. That way, they work when only "make altinstall" is used. ........ r58139 | mark.summerfield | 2007-09-13 16:54:30 +0200 (Thu, 13 Sep 2007) | 9 lines Replaced variable o with obj in operator.rst because o is easy to confuse. Added a note about Python 3's collections.Mapping etc., above section that describes isMappingType() etc. Added xrefs between os, os.path, fileinput, and open(). ........ r58143 | facundo.batista | 2007-09-13 20:13:15 +0200 (Thu, 13 Sep 2007) | 7 lines Merged the decimal-branch (revisions 54886 to 58140). Decimal is now fully updated to the latests Decimal Specification (v1.66) and the latests test cases (v2.56). Thanks to Mark Dickinson for all his help during this process. ........ r58145 | facundo.batista | 2007-09-13 20:42:09 +0200 (Thu, 13 Sep 2007) | 7 lines Put the parameter watchexp back in (changed watchexp from an int to a bool). Also second argument to watchexp is now converted to Decimal, just as with all the other two-argument operations. Thanks Mark Dickinson. ........ r58147 | andrew.kuchling | 2007-09-14 00:49:34 +0200 (Fri, 14 Sep 2007) | 1 line Add various items ........ r58148 | andrew.kuchling | 2007-09-14 00:50:10 +0200 (Fri, 14 Sep 2007) | 1 line Make target unique ........ r58154 | facundo.batista | 2007-09-14 20:58:34 +0200 (Fri, 14 Sep 2007) | 3 lines Included the new functions, and new descriptions. ........ r58155 | thomas.heller | 2007-09-14 21:40:35 +0200 (Fri, 14 Sep 2007) | 2 lines ctypes.util.find_library uses dump(1) instead of objdump(1) on Solaris. Fixes issue #1777530; will backport to release25-maint. ........ r58159 | facundo.batista | 2007-09-14 23:29:52 +0200 (Fri, 14 Sep 2007) | 3 lines Some additions (examples and a bit on the tutorial). ........ r58160 | georg.brandl | 2007-09-15 18:53:36 +0200 (Sat, 15 Sep 2007) | 2 lines Remove bdb from the "undocumented modules" list. ........ r58164 | bill.janssen | 2007-09-17 00:06:00 +0200 (Mon, 17 Sep 2007) | 15 lines Add support for asyncore server-side SSL support. This requires adding the 'makefile' method to ssl.SSLSocket, and importing the requisite fakefile class from socket.py, and making the appropriate changes to it to make it use the SSL connection. Added sample HTTPS server to test_ssl.py, and test that uses it. Change SSL tests to use https://svn.python.org/, instead of www.sf.net and pop.gmail.com. Added utility function to ssl module, get_server_certificate, to wrap up the several things to be done to pull a certificate from a remote server. ........ r58173 | bill.janssen | 2007-09-17 01:16:46 +0200 (Mon, 17 Sep 2007) | 1 line use binary mode when reading files for testAsyncore to make Windows happy ........ r58175 | raymond.hettinger | 2007-09-17 02:55:00 +0200 (Mon, 17 Sep 2007) | 7 lines Sync-up named tuples with the latest version of the ASPN recipe. Allows optional commas in the field-name spec (help when named tuples are used in conjuction with sql queries). Adds the __fields__ attribute for introspection and to support conversion to dictionary form. Adds a __replace__() method similar to str.replace() but using a named field as a target. Clean-up spelling and presentation in doc-strings. ........ r58176 | brett.cannon | 2007-09-17 05:28:34 +0200 (Mon, 17 Sep 2007) | 5 lines Add a bunch of GIL release/acquire points in tp_print implementations and for PyObject_Print(). Closes issue #1164. ........ r58177 | sean.reifschneider | 2007-09-17 07:45:04 +0200 (Mon, 17 Sep 2007) | 2 lines issue1597011: Fix for bz2 module corner-case error due to error checking bug. ........ r58180 | facundo.batista | 2007-09-17 18:26:50 +0200 (Mon, 17 Sep 2007) | 3 lines Decimal is updated, :) ........ r58181 | facundo.batista | 2007-09-17 19:30:13 +0200 (Mon, 17 Sep 2007) | 5 lines The methods always return Decimal classes, even if they're executed through a subclass (thanks Mark Dickinson). Added a bit of testing for this. ........ r58183 | sean.reifschneider | 2007-09-17 22:53:21 +0200 (Mon, 17 Sep 2007) | 2 lines issue1082: Fixing platform and system for Vista. ........ r58185 | andrew.kuchling | 2007-09-18 03:36:16 +0200 (Tue, 18 Sep 2007) | 1 line Add item; sort properly ........ r58186 | raymond.hettinger | 2007-09-18 05:33:19 +0200 (Tue, 18 Sep 2007) | 1 line Handle corner cased on 0-tuples and 1-tuples. Add verbose option so people can see how it works. ........ r58192 | georg.brandl | 2007-09-18 09:24:40 +0200 (Tue, 18 Sep 2007) | 2 lines A bit of reordering, also show more subheadings in the lang ref index. ........ r58193 | facundo.batista | 2007-09-18 18:53:18 +0200 (Tue, 18 Sep 2007) | 4 lines Speed up of the various division operations (remainder, divide, divideint and divmod). Thanks Mark Dickinson. ........ r58197 | raymond.hettinger | 2007-09-19 00:18:02 +0200 (Wed, 19 Sep 2007) | 1 line Cleanup docs for NamedTuple. ........
2064 lines
48 KiB
C
2064 lines
48 KiB
C
/*
|
|
|
|
python-bz2 - python bz2 library interface
|
|
|
|
Copyright (c) 2002 Gustavo Niemeyer <niemeyer@conectiva.com>
|
|
Copyright (c) 2002 Python Software Foundation; All Rights Reserved
|
|
|
|
*/
|
|
|
|
#include "Python.h"
|
|
#include <stdio.h>
|
|
#include <bzlib.h>
|
|
#include "structmember.h"
|
|
|
|
#ifdef WITH_THREAD
|
|
#include "pythread.h"
|
|
#endif
|
|
|
|
static char __author__[] =
|
|
"The bz2 python module was written by:\n\
|
|
\n\
|
|
Gustavo Niemeyer <niemeyer@conectiva.com>\n\
|
|
";
|
|
|
|
/* Our very own off_t-like type, 64-bit if possible */
|
|
/* copied from Objects/fileobject.c */
|
|
#if !defined(HAVE_LARGEFILE_SUPPORT)
|
|
typedef off_t Py_off_t;
|
|
#elif SIZEOF_OFF_T >= 8
|
|
typedef off_t Py_off_t;
|
|
#elif SIZEOF_FPOS_T >= 8
|
|
typedef fpos_t Py_off_t;
|
|
#else
|
|
#error "Large file support, but neither off_t nor fpos_t is large enough."
|
|
#endif
|
|
|
|
#define BUF(v) PyBytes_AS_STRING(v)
|
|
|
|
#define MODE_CLOSED 0
|
|
#define MODE_READ 1
|
|
#define MODE_READ_EOF 2
|
|
#define MODE_WRITE 3
|
|
|
|
#define BZ2FileObject_Check(v) (Py_Type(v) == &BZ2File_Type)
|
|
|
|
|
|
#ifdef BZ_CONFIG_ERROR
|
|
|
|
#if SIZEOF_LONG >= 8
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
(((long)bzs->total_out_hi32 << 32) + bzs->total_out_lo32)
|
|
#elif SIZEOF_LONG_LONG >= 8
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
(((PY_LONG_LONG)bzs->total_out_hi32 << 32) + bzs->total_out_lo32)
|
|
#else
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
bzs->total_out_lo32
|
|
#endif
|
|
|
|
#else /* ! BZ_CONFIG_ERROR */
|
|
|
|
#define BZ2_bzRead bzRead
|
|
#define BZ2_bzReadOpen bzReadOpen
|
|
#define BZ2_bzReadClose bzReadClose
|
|
#define BZ2_bzWrite bzWrite
|
|
#define BZ2_bzWriteOpen bzWriteOpen
|
|
#define BZ2_bzWriteClose bzWriteClose
|
|
#define BZ2_bzCompress bzCompress
|
|
#define BZ2_bzCompressInit bzCompressInit
|
|
#define BZ2_bzCompressEnd bzCompressEnd
|
|
#define BZ2_bzDecompress bzDecompress
|
|
#define BZ2_bzDecompressInit bzDecompressInit
|
|
#define BZ2_bzDecompressEnd bzDecompressEnd
|
|
|
|
#define BZS_TOTAL_OUT(bzs) bzs->total_out
|
|
|
|
#endif /* ! BZ_CONFIG_ERROR */
|
|
|
|
|
|
#ifdef WITH_THREAD
|
|
#define ACQUIRE_LOCK(obj) PyThread_acquire_lock(obj->lock, 1)
|
|
#define RELEASE_LOCK(obj) PyThread_release_lock(obj->lock)
|
|
#else
|
|
#define ACQUIRE_LOCK(obj)
|
|
#define RELEASE_LOCK(obj)
|
|
#endif
|
|
|
|
/* Bits in f_newlinetypes */
|
|
#define NEWLINE_UNKNOWN 0 /* No newline seen, yet */
|
|
#define NEWLINE_CR 1 /* \r newline seen */
|
|
#define NEWLINE_LF 2 /* \n newline seen */
|
|
#define NEWLINE_CRLF 4 /* \r\n newline seen */
|
|
|
|
/* ===================================================================== */
|
|
/* Structure definitions. */
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
FILE *rawfp;
|
|
|
|
char* f_buf; /* Allocated readahead buffer */
|
|
char* f_bufend; /* Points after last occupied position */
|
|
char* f_bufptr; /* Current buffer position */
|
|
|
|
BZFILE *fp;
|
|
int mode;
|
|
Py_off_t pos;
|
|
Py_off_t size;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2FileObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
bz_stream bzs;
|
|
int running;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2CompObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
bz_stream bzs;
|
|
int running;
|
|
PyObject *unused_data;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2DecompObject;
|
|
|
|
/* ===================================================================== */
|
|
/* Utility functions. */
|
|
|
|
static int
|
|
Util_CatchBZ2Error(int bzerror)
|
|
{
|
|
int ret = 0;
|
|
switch(bzerror) {
|
|
case BZ_OK:
|
|
case BZ_STREAM_END:
|
|
break;
|
|
|
|
#ifdef BZ_CONFIG_ERROR
|
|
case BZ_CONFIG_ERROR:
|
|
PyErr_SetString(PyExc_SystemError,
|
|
"the bz2 library was not compiled "
|
|
"correctly");
|
|
ret = 1;
|
|
break;
|
|
#endif
|
|
|
|
case BZ_PARAM_ERROR:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"the bz2 library has received wrong "
|
|
"parameters");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_MEM_ERROR:
|
|
PyErr_NoMemory();
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_DATA_ERROR:
|
|
case BZ_DATA_ERROR_MAGIC:
|
|
PyErr_SetString(PyExc_IOError, "invalid data stream");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_IO_ERROR:
|
|
PyErr_SetString(PyExc_IOError, "unknown IO error");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_UNEXPECTED_EOF:
|
|
PyErr_SetString(PyExc_EOFError,
|
|
"compressed file ended before the "
|
|
"logical end-of-stream was detected");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_SEQUENCE_ERROR:
|
|
PyErr_SetString(PyExc_RuntimeError,
|
|
"wrong sequence of bz2 library "
|
|
"commands used");
|
|
ret = 1;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
#if BUFSIZ < 8192
|
|
#define SMALLCHUNK 8192
|
|
#else
|
|
#define SMALLCHUNK BUFSIZ
|
|
#endif
|
|
|
|
#if SIZEOF_INT < 4
|
|
#define BIGCHUNK (512 * 32)
|
|
#else
|
|
#define BIGCHUNK (512 * 1024)
|
|
#endif
|
|
|
|
/* This is a hacked version of Python's fileobject.c:new_buffersize(). */
|
|
static size_t
|
|
Util_NewBufferSize(size_t currentsize)
|
|
{
|
|
if (currentsize > SMALLCHUNK) {
|
|
/* Keep doubling until we reach BIGCHUNK;
|
|
then keep adding BIGCHUNK. */
|
|
if (currentsize <= BIGCHUNK)
|
|
return currentsize + currentsize;
|
|
else
|
|
return currentsize + BIGCHUNK;
|
|
}
|
|
return currentsize + SMALLCHUNK;
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:get_line(). */
|
|
static PyObject *
|
|
Util_GetLine(BZ2FileObject *f, int n)
|
|
{
|
|
char c;
|
|
char *buf, *end;
|
|
size_t total_v_size; /* total # of slots in buffer */
|
|
size_t used_v_size; /* # used slots in buffer */
|
|
size_t increment; /* amount to increment the buffer */
|
|
PyObject *v;
|
|
int bzerror;
|
|
int bytes_read;
|
|
|
|
total_v_size = n > 0 ? n : 100;
|
|
v = PyBytes_FromStringAndSize((char *)NULL, total_v_size);
|
|
if (v == NULL)
|
|
return NULL;
|
|
|
|
buf = BUF(v);
|
|
end = buf + total_v_size;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
do {
|
|
bytes_read = BZ2_bzRead(&bzerror, f->fp, &c, 1);
|
|
f->pos++;
|
|
if (bytes_read == 0)
|
|
break;
|
|
*buf++ = c;
|
|
} while (bzerror == BZ_OK && c != '\n' && buf != end);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
f->size = f->pos;
|
|
f->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(v);
|
|
return NULL;
|
|
}
|
|
if (c == '\n')
|
|
break;
|
|
/* Must be because buf == end */
|
|
if (n > 0)
|
|
break;
|
|
used_v_size = total_v_size;
|
|
increment = total_v_size >> 2; /* mild exponential growth */
|
|
total_v_size += increment;
|
|
if (total_v_size > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"line is longer than a Python string can hold");
|
|
Py_DECREF(v);
|
|
return NULL;
|
|
}
|
|
if (PyBytes_Resize(v, total_v_size) < 0) {
|
|
Py_DECREF(v);
|
|
return NULL;
|
|
}
|
|
buf = BUF(v) + used_v_size;
|
|
end = BUF(v) + total_v_size;
|
|
}
|
|
|
|
used_v_size = buf - BUF(v);
|
|
if (used_v_size != total_v_size) {
|
|
if (PyBytes_Resize(v, used_v_size) < 0) {
|
|
Py_DECREF(v);
|
|
v = NULL;
|
|
}
|
|
}
|
|
return v;
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:drop_readahead(). */
|
|
static void
|
|
Util_DropReadAhead(BZ2FileObject *f)
|
|
{
|
|
if (f->f_buf != NULL) {
|
|
PyMem_Free(f->f_buf);
|
|
f->f_buf = NULL;
|
|
}
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:readahead(). */
|
|
static int
|
|
Util_ReadAhead(BZ2FileObject *f, int bufsize)
|
|
{
|
|
int chunksize;
|
|
int bzerror;
|
|
|
|
if (f->f_buf != NULL) {
|
|
if((f->f_bufend - f->f_bufptr) >= 1)
|
|
return 0;
|
|
else
|
|
Util_DropReadAhead(f);
|
|
}
|
|
if (f->mode == MODE_READ_EOF) {
|
|
f->f_bufptr = f->f_buf;
|
|
f->f_bufend = f->f_buf;
|
|
return 0;
|
|
}
|
|
if ((f->f_buf = PyMem_Malloc(bufsize)) == NULL) {
|
|
return -1;
|
|
}
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = BZ2_bzRead(&bzerror, f->fp, f->f_buf, bufsize);
|
|
Py_END_ALLOW_THREADS
|
|
f->pos += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
f->size = f->pos;
|
|
f->mode = MODE_READ_EOF;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Util_DropReadAhead(f);
|
|
return -1;
|
|
}
|
|
f->f_bufptr = f->f_buf;
|
|
f->f_bufend = f->f_buf + chunksize;
|
|
return 0;
|
|
}
|
|
|
|
/* This is a hacked version of Python's
|
|
* fileobject.c:readahead_get_line_skip(). */
|
|
static PyBytesObject *
|
|
Util_ReadAheadGetLineSkip(BZ2FileObject *f, int skip, int bufsize)
|
|
{
|
|
PyBytesObject* s;
|
|
char *bufptr;
|
|
char *buf;
|
|
int len;
|
|
|
|
if (f->f_buf == NULL)
|
|
if (Util_ReadAhead(f, bufsize) < 0)
|
|
return NULL;
|
|
|
|
len = f->f_bufend - f->f_bufptr;
|
|
if (len == 0)
|
|
return (PyBytesObject *)
|
|
PyBytes_FromStringAndSize(NULL, skip);
|
|
bufptr = memchr(f->f_bufptr, '\n', len);
|
|
if (bufptr != NULL) {
|
|
bufptr++; /* Count the '\n' */
|
|
len = bufptr - f->f_bufptr;
|
|
s = (PyBytesObject *)
|
|
PyBytes_FromStringAndSize(NULL, skip+len);
|
|
if (s == NULL)
|
|
return NULL;
|
|
memcpy(PyBytes_AS_STRING(s)+skip, f->f_bufptr, len);
|
|
f->f_bufptr = bufptr;
|
|
if (bufptr == f->f_bufend)
|
|
Util_DropReadAhead(f);
|
|
} else {
|
|
bufptr = f->f_bufptr;
|
|
buf = f->f_buf;
|
|
f->f_buf = NULL; /* Force new readahead buffer */
|
|
s = Util_ReadAheadGetLineSkip(f, skip+len,
|
|
bufsize + (bufsize>>2));
|
|
if (s == NULL) {
|
|
PyMem_Free(buf);
|
|
return NULL;
|
|
}
|
|
memcpy(PyBytes_AS_STRING(s)+skip, bufptr, len);
|
|
PyMem_Free(buf);
|
|
}
|
|
return s;
|
|
}
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2File. */
|
|
|
|
PyDoc_STRVAR(BZ2File_read__doc__,
|
|
"read([size]) -> string\n\
|
|
\n\
|
|
Read at most size uncompressed bytes, returned as a string. If the size\n\
|
|
argument is negative or omitted, read until EOF is reached.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_read(). */
|
|
static PyObject *
|
|
BZ2File_read(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
long bytesrequested = -1;
|
|
size_t bytesread, buffersize, chunksize;
|
|
int bzerror;
|
|
PyObject *ret = NULL;
|
|
|
|
if (!PyArg_ParseTuple(args, "|l:read", &bytesrequested))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
ret = PyBytes_FromStringAndSize("", 0);
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if (bytesrequested < 0)
|
|
buffersize = Util_NewBufferSize((size_t)0);
|
|
else
|
|
buffersize = bytesrequested;
|
|
if (buffersize > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"requested number of bytes is "
|
|
"more than a Python string can hold");
|
|
goto cleanup;
|
|
}
|
|
ret = PyBytes_FromStringAndSize((char *)NULL, buffersize);
|
|
if (ret == NULL || buffersize == 0)
|
|
goto cleanup;
|
|
bytesread = 0;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = BZ2_bzRead(&bzerror, self->fp,
|
|
BUF(ret)+bytesread,
|
|
buffersize-bytesread);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
goto cleanup;
|
|
}
|
|
if (bytesrequested < 0) {
|
|
buffersize = Util_NewBufferSize(buffersize);
|
|
if (PyBytes_Resize(ret, buffersize) < 0) {
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
goto cleanup;
|
|
}
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
if (bytesread != buffersize) {
|
|
if (PyBytes_Resize(ret, bytesread) < 0) {
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
}
|
|
}
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_readline__doc__,
|
|
"readline([size]) -> string\n\
|
|
\n\
|
|
Return the next line from the file, as a string, retaining newline.\n\
|
|
A non-negative size argument will limit the maximum number of bytes to\n\
|
|
return (an incomplete line may be returned then). Return an empty\n\
|
|
string at EOF.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_readline(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
int sizehint = -1;
|
|
|
|
if (!PyArg_ParseTuple(args, "|i:readline", &sizehint))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
ret = PyBytes_FromStringAndSize("", 0);
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if (sizehint == 0)
|
|
ret = PyBytes_FromStringAndSize("", 0);
|
|
else
|
|
ret = Util_GetLine(self, (sizehint < 0) ? 0 : sizehint);
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_readlines__doc__,
|
|
"readlines([size]) -> list\n\
|
|
\n\
|
|
Call readline() repeatedly and return a list of lines read.\n\
|
|
The optional size argument, if given, is an approximate bound on the\n\
|
|
total number of bytes in the lines returned.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_readlines(). */
|
|
static PyObject *
|
|
BZ2File_readlines(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
long sizehint = 0;
|
|
PyObject *list = NULL;
|
|
PyObject *line;
|
|
char small_buffer[SMALLCHUNK];
|
|
char *buffer = small_buffer;
|
|
size_t buffersize = SMALLCHUNK;
|
|
PyObject *big_buffer = NULL;
|
|
size_t nfilled = 0;
|
|
size_t nread;
|
|
size_t totalread = 0;
|
|
char *p, *q, *end;
|
|
int err;
|
|
int shortread = 0;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "|l:readlines", &sizehint))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
list = PyList_New(0);
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if ((list = PyList_New(0)) == NULL)
|
|
goto cleanup;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
nread = BZ2_bzRead(&bzerror, self->fp,
|
|
buffer+nfilled, buffersize-nfilled);
|
|
self->pos += nread;
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
if (nread == 0) {
|
|
sizehint = 0;
|
|
break;
|
|
}
|
|
shortread = 1;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
error:
|
|
Py_DECREF(list);
|
|
list = NULL;
|
|
goto cleanup;
|
|
}
|
|
totalread += nread;
|
|
p = memchr(buffer+nfilled, '\n', nread);
|
|
if (!shortread && p == NULL) {
|
|
/* Need a larger buffer to fit this line */
|
|
nfilled += nread;
|
|
buffersize *= 2;
|
|
if (buffersize > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"line is longer than a Python string can hold");
|
|
goto error;
|
|
}
|
|
if (big_buffer == NULL) {
|
|
/* Create the big buffer */
|
|
big_buffer = PyBytes_FromStringAndSize(
|
|
NULL, buffersize);
|
|
if (big_buffer == NULL)
|
|
goto error;
|
|
buffer = PyBytes_AS_STRING(big_buffer);
|
|
memcpy(buffer, small_buffer, nfilled);
|
|
}
|
|
else {
|
|
/* Grow the big buffer */
|
|
if (PyBytes_Resize(big_buffer, buffersize) < 0){
|
|
Py_DECREF(big_buffer);
|
|
big_buffer = NULL;
|
|
goto error;
|
|
}
|
|
buffer = PyBytes_AS_STRING(big_buffer);
|
|
}
|
|
continue;
|
|
}
|
|
end = buffer+nfilled+nread;
|
|
q = buffer;
|
|
while (p != NULL) {
|
|
/* Process complete lines */
|
|
p++;
|
|
line = PyBytes_FromStringAndSize(q, p-q);
|
|
if (line == NULL)
|
|
goto error;
|
|
err = PyList_Append(list, line);
|
|
Py_DECREF(line);
|
|
if (err != 0)
|
|
goto error;
|
|
q = p;
|
|
p = memchr(q, '\n', end-q);
|
|
}
|
|
/* Move the remaining incomplete line to the start */
|
|
nfilled = end-q;
|
|
memmove(buffer, q, nfilled);
|
|
if (sizehint > 0)
|
|
if (totalread >= (size_t)sizehint)
|
|
break;
|
|
if (shortread) {
|
|
sizehint = 0;
|
|
break;
|
|
}
|
|
}
|
|
if (nfilled != 0) {
|
|
/* Partial last line */
|
|
line = PyBytes_FromStringAndSize(buffer, nfilled);
|
|
if (line == NULL)
|
|
goto error;
|
|
if (sizehint > 0) {
|
|
/* Need to complete the last line */
|
|
PyObject *rest = Util_GetLine(self, 0);
|
|
PyObject *new;
|
|
if (rest == NULL) {
|
|
Py_DECREF(line);
|
|
goto error;
|
|
}
|
|
new = PyBytes_Concat(line, rest);
|
|
Py_DECREF(line);
|
|
Py_DECREF(rest);
|
|
line = new;
|
|
if (line == NULL)
|
|
goto error;
|
|
}
|
|
err = PyList_Append(list, line);
|
|
Py_DECREF(line);
|
|
if (err != 0)
|
|
goto error;
|
|
}
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
if (big_buffer) {
|
|
Py_DECREF(big_buffer);
|
|
}
|
|
return list;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_write__doc__,
|
|
"write(data) -> None\n\
|
|
\n\
|
|
Write the 'data' string to file. Note that due to buffering, close() may\n\
|
|
be needed before the file on disk reflects the data written.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_write(). */
|
|
static PyObject *
|
|
BZ2File_write(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
char *buf;
|
|
int len;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#:write", &buf, &len))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_WRITE:
|
|
break;
|
|
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for writing");
|
|
goto cleanup;
|
|
}
|
|
|
|
Py_BEGIN_ALLOW_THREADS
|
|
BZ2_bzWrite (&bzerror, self->fp, buf, len);
|
|
self->pos += len;
|
|
Py_END_ALLOW_THREADS
|
|
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_writelines__doc__,
|
|
"writelines(sequence_of_strings) -> None\n\
|
|
\n\
|
|
Write the sequence of strings to the file. Note that newlines are not\n\
|
|
added. The sequence can be any iterable object producing strings. This is\n\
|
|
equivalent to calling write() for each string.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_writelines(). */
|
|
static PyObject *
|
|
BZ2File_writelines(BZ2FileObject *self, PyObject *seq)
|
|
{
|
|
#define CHUNKSIZE 1000
|
|
PyObject *list = NULL;
|
|
PyObject *iter = NULL;
|
|
PyObject *ret = NULL;
|
|
PyObject *line;
|
|
int i, j, index, len, islist;
|
|
int bzerror;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_WRITE:
|
|
break;
|
|
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto error;
|
|
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for writing");
|
|
goto error;
|
|
}
|
|
|
|
islist = PyList_Check(seq);
|
|
if (!islist) {
|
|
iter = PyObject_GetIter(seq);
|
|
if (iter == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"writelines() requires an iterable argument");
|
|
goto error;
|
|
}
|
|
list = PyList_New(CHUNKSIZE);
|
|
if (list == NULL)
|
|
goto error;
|
|
}
|
|
|
|
/* Strategy: slurp CHUNKSIZE lines into a private list,
|
|
checking that they are all strings, then write that list
|
|
without holding the interpreter lock, then come back for more. */
|
|
for (index = 0; ; index += CHUNKSIZE) {
|
|
if (islist) {
|
|
Py_XDECREF(list);
|
|
list = PyList_GetSlice(seq, index, index+CHUNKSIZE);
|
|
if (list == NULL)
|
|
goto error;
|
|
j = PyList_GET_SIZE(list);
|
|
}
|
|
else {
|
|
for (j = 0; j < CHUNKSIZE; j++) {
|
|
line = PyIter_Next(iter);
|
|
if (line == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
break;
|
|
}
|
|
PyList_SetItem(list, j, line);
|
|
}
|
|
}
|
|
if (j == 0)
|
|
break;
|
|
|
|
/* Check that all entries are indeed byte strings. If not,
|
|
apply the same rules as for file.write() and
|
|
convert the rets to strings. This is slow, but
|
|
seems to be the only way since all conversion APIs
|
|
could potentially execute Python code. */
|
|
for (i = 0; i < j; i++) {
|
|
PyObject *v = PyList_GET_ITEM(list, i);
|
|
if (!PyBytes_Check(v)) {
|
|
const char *buffer;
|
|
Py_ssize_t len;
|
|
if (PyObject_AsCharBuffer(v, &buffer, &len)) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"writelines() "
|
|
"argument must be "
|
|
"a sequence of "
|
|
"bytes objects");
|
|
goto error;
|
|
}
|
|
line = PyBytes_FromStringAndSize(buffer,
|
|
len);
|
|
if (line == NULL)
|
|
goto error;
|
|
Py_DECREF(v);
|
|
PyList_SET_ITEM(list, i, line);
|
|
}
|
|
}
|
|
|
|
/* Since we are releasing the global lock, the
|
|
following code may *not* execute Python code. */
|
|
Py_BEGIN_ALLOW_THREADS
|
|
for (i = 0; i < j; i++) {
|
|
line = PyList_GET_ITEM(list, i);
|
|
len = PyBytes_GET_SIZE(line);
|
|
BZ2_bzWrite (&bzerror, self->fp,
|
|
PyBytes_AS_STRING(line), len);
|
|
if (bzerror != BZ_OK) {
|
|
Py_BLOCK_THREADS
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
}
|
|
Py_END_ALLOW_THREADS
|
|
|
|
if (j < CHUNKSIZE)
|
|
break;
|
|
}
|
|
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(list);
|
|
Py_XDECREF(iter);
|
|
return ret;
|
|
#undef CHUNKSIZE
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_seek__doc__,
|
|
"seek(offset [, whence]) -> None\n\
|
|
\n\
|
|
Move to new file position. Argument offset is a byte count. Optional\n\
|
|
argument whence defaults to 0 (offset from start of file, offset\n\
|
|
should be >= 0); other values are 1 (move relative to current position,\n\
|
|
positive or negative), and 2 (move relative to end of file, usually\n\
|
|
negative, although many platforms allow seeking beyond the end of a file).\n\
|
|
\n\
|
|
Note that seeking of bz2 files is emulated, and depending on the parameters\n\
|
|
the operation may be extremely slow.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_seek(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
int where = 0;
|
|
PyObject *offobj;
|
|
Py_off_t offset;
|
|
char small_buffer[SMALLCHUNK];
|
|
char *buffer = small_buffer;
|
|
size_t buffersize = SMALLCHUNK;
|
|
Py_off_t bytesread = 0;
|
|
size_t readsize;
|
|
int chunksize;
|
|
int bzerror;
|
|
PyObject *ret = NULL;
|
|
|
|
if (!PyArg_ParseTuple(args, "O|i:seek", &offobj, &where))
|
|
return NULL;
|
|
#if !defined(HAVE_LARGEFILE_SUPPORT)
|
|
offset = PyInt_AsLong(offobj);
|
|
#else
|
|
offset = PyLong_Check(offobj) ?
|
|
PyLong_AsLongLong(offobj) : PyInt_AsLong(offobj);
|
|
#endif
|
|
if (PyErr_Occurred())
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
Util_DropReadAhead(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
break;
|
|
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"seek works only while reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if (where == 2) {
|
|
if (self->size == -1) {
|
|
assert(self->mode != MODE_READ_EOF);
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = BZ2_bzRead(&bzerror, self->fp,
|
|
buffer, buffersize);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
}
|
|
self->mode = MODE_READ_EOF;
|
|
self->size = self->pos;
|
|
bytesread = 0;
|
|
}
|
|
offset = self->size + offset;
|
|
} else if (where == 1) {
|
|
offset = self->pos + offset;
|
|
}
|
|
|
|
/* Before getting here, offset must be the absolute position the file
|
|
* pointer should be set to. */
|
|
|
|
if (offset >= self->pos) {
|
|
/* we can move forward */
|
|
offset -= self->pos;
|
|
} else {
|
|
/* we cannot move back, so rewind the stream */
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
rewind(self->rawfp);
|
|
self->pos = 0;
|
|
self->fp = BZ2_bzReadOpen(&bzerror, self->rawfp,
|
|
0, 0, NULL, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
self->mode = MODE_READ;
|
|
}
|
|
|
|
if (offset <= 0 || self->mode == MODE_READ_EOF)
|
|
goto exit;
|
|
|
|
/* Before getting here, offset must be set to the number of bytes
|
|
* to walk forward. */
|
|
for (;;) {
|
|
if (offset-bytesread > buffersize)
|
|
readsize = buffersize;
|
|
else
|
|
/* offset might be wider that readsize, but the result
|
|
* of the subtraction is bound by buffersize (see the
|
|
* condition above). buffersize is 8192. */
|
|
readsize = (size_t)(offset-bytesread);
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = BZ2_bzRead(&bzerror, self->fp, buffer, readsize);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
if (bytesread == offset)
|
|
break;
|
|
}
|
|
|
|
exit:
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_tell__doc__,
|
|
"tell() -> int\n\
|
|
\n\
|
|
Return the current file position, an integer (may be a long integer).\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_tell(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
}
|
|
|
|
#if !defined(HAVE_LARGEFILE_SUPPORT)
|
|
ret = PyInt_FromLong(self->pos);
|
|
#else
|
|
ret = PyLong_FromLongLong(self->pos);
|
|
#endif
|
|
|
|
cleanup:
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_close__doc__,
|
|
"close() -> None or (perhaps) an integer\n\
|
|
\n\
|
|
Close the file. Sets data attribute .closed to true. A closed file\n\
|
|
cannot be used for further I/O operations. close() may be called more\n\
|
|
than once without error.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_close(BZ2FileObject *self)
|
|
{
|
|
PyObject *ret = NULL;
|
|
int bzerror = BZ_OK;
|
|
|
|
if (self->mode == MODE_CLOSED) {
|
|
Py_RETURN_NONE;
|
|
}
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
break;
|
|
case MODE_WRITE:
|
|
BZ2_bzWriteClose(&bzerror, self->fp,
|
|
0, NULL, NULL);
|
|
break;
|
|
}
|
|
self->mode = MODE_CLOSED;
|
|
fclose(self->rawfp);
|
|
self->rawfp = NULL;
|
|
if (bzerror == BZ_OK) {
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
}
|
|
else {
|
|
Util_CatchBZ2Error(bzerror);
|
|
}
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
static PyObject *BZ2File_getiter(BZ2FileObject *self);
|
|
|
|
static PyMethodDef BZ2File_methods[] = {
|
|
{"read", (PyCFunction)BZ2File_read, METH_VARARGS, BZ2File_read__doc__},
|
|
{"readline", (PyCFunction)BZ2File_readline, METH_VARARGS, BZ2File_readline__doc__},
|
|
{"readlines", (PyCFunction)BZ2File_readlines, METH_VARARGS, BZ2File_readlines__doc__},
|
|
{"write", (PyCFunction)BZ2File_write, METH_VARARGS, BZ2File_write__doc__},
|
|
{"writelines", (PyCFunction)BZ2File_writelines, METH_O, BZ2File_writelines__doc__},
|
|
{"seek", (PyCFunction)BZ2File_seek, METH_VARARGS, BZ2File_seek__doc__},
|
|
{"tell", (PyCFunction)BZ2File_tell, METH_NOARGS, BZ2File_tell__doc__},
|
|
{"close", (PyCFunction)BZ2File_close, METH_NOARGS, BZ2File_close__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Getters and setters of BZ2File. */
|
|
|
|
static PyObject *
|
|
BZ2File_get_closed(BZ2FileObject *self, void *closure)
|
|
{
|
|
return PyInt_FromLong(self->mode == MODE_CLOSED);
|
|
}
|
|
|
|
static PyGetSetDef BZ2File_getset[] = {
|
|
{"closed", (getter)BZ2File_get_closed, NULL,
|
|
"True if the file is closed"},
|
|
{NULL} /* Sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2File_Type. */
|
|
|
|
static int
|
|
BZ2File_init(BZ2FileObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
static char *kwlist[] = {"filename", "mode", "buffering",
|
|
"compresslevel", 0};
|
|
char *name;
|
|
char *mode = "r";
|
|
int buffering = -1;
|
|
int compresslevel = 9;
|
|
int bzerror;
|
|
int mode_char = 0;
|
|
|
|
self->size = -1;
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "s|sii:BZ2File",
|
|
kwlist, &name, &mode, &buffering,
|
|
&compresslevel))
|
|
return -1;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
return -1;
|
|
}
|
|
|
|
for (;;) {
|
|
int error = 0;
|
|
switch (*mode) {
|
|
case 'r':
|
|
case 'w':
|
|
if (mode_char)
|
|
error = 1;
|
|
mode_char = *mode;
|
|
break;
|
|
|
|
case 'b':
|
|
break;
|
|
|
|
default:
|
|
error = 1;
|
|
break;
|
|
}
|
|
if (error) {
|
|
PyErr_Format(PyExc_ValueError,
|
|
"invalid mode char %c", *mode);
|
|
return -1;
|
|
}
|
|
mode++;
|
|
if (*mode == '\0')
|
|
break;
|
|
}
|
|
|
|
if (mode_char == 0) {
|
|
mode_char = 'r';
|
|
}
|
|
|
|
mode = (mode_char == 'r') ? "rb" : "wb";
|
|
|
|
self->rawfp = fopen(name, mode);
|
|
if (self->rawfp == NULL) {
|
|
PyErr_SetFromErrno(PyExc_IOError);
|
|
return -1;
|
|
}
|
|
/* XXX Ignore buffering */
|
|
|
|
/* From now on, we have stuff to dealloc, so jump to error label
|
|
* instead of returning */
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock) {
|
|
PyErr_SetString(PyExc_MemoryError, "unable to allocate lock");
|
|
goto error;
|
|
}
|
|
#endif
|
|
|
|
if (mode_char == 'r')
|
|
self->fp = BZ2_bzReadOpen(&bzerror, self->rawfp,
|
|
0, 0, NULL, 0);
|
|
else
|
|
self->fp = BZ2_bzWriteOpen(&bzerror, self->rawfp,
|
|
compresslevel, 0, 0);
|
|
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->mode = (mode_char == 'r') ? MODE_READ : MODE_WRITE;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
fclose(self->rawfp);
|
|
self->rawfp = NULL;
|
|
#ifdef WITH_THREAD
|
|
if (self->lock) {
|
|
PyThread_free_lock(self->lock);
|
|
self->lock = NULL;
|
|
}
|
|
#endif
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2File_dealloc(BZ2FileObject *self)
|
|
{
|
|
int bzerror;
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
break;
|
|
case MODE_WRITE:
|
|
BZ2_bzWriteClose(&bzerror, self->fp,
|
|
0, NULL, NULL);
|
|
break;
|
|
}
|
|
Util_DropReadAhead(self);
|
|
if (self->rawfp != NULL)
|
|
fclose(self->rawfp);
|
|
Py_Type(self)->tp_free((PyObject *)self);
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_getiter(). */
|
|
static PyObject *
|
|
BZ2File_getiter(BZ2FileObject *self)
|
|
{
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
return NULL;
|
|
}
|
|
Py_INCREF((PyObject*)self);
|
|
return (PyObject *)self;
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_iternext(). */
|
|
#define READAHEAD_BUFSIZE 8192
|
|
static PyObject *
|
|
BZ2File_iternext(BZ2FileObject *self)
|
|
{
|
|
PyBytesObject* ret;
|
|
ACQUIRE_LOCK(self);
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
return NULL;
|
|
}
|
|
ret = Util_ReadAheadGetLineSkip(self, 0, READAHEAD_BUFSIZE);
|
|
RELEASE_LOCK(self);
|
|
if (ret == NULL || PyBytes_GET_SIZE(ret) == 0) {
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
return (PyObject *)ret;
|
|
}
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2File_Type definition. */
|
|
|
|
PyDoc_VAR(BZ2File__doc__) =
|
|
PyDoc_STR(
|
|
"BZ2File(name [, mode='r', buffering=0, compresslevel=9]) -> file object\n\
|
|
\n\
|
|
Open a bz2 file. The mode can be 'r' or 'w', for reading (default) or\n\
|
|
writing. When opened for writing, the file will be created if it doesn't\n\
|
|
exist, and truncated otherwise. If the buffering argument is given, 0 means\n\
|
|
unbuffered, and larger numbers specify the buffer size. If compresslevel\n\
|
|
is given, must be a number between 1 and 9.\n\
|
|
Data read is always returned in bytes; data written ought to be bytes.\n\
|
|
");
|
|
|
|
static PyTypeObject BZ2File_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"bz2.BZ2File", /*tp_name*/
|
|
sizeof(BZ2FileObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2File_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
PyObject_GenericGetAttr,/*tp_getattro*/
|
|
PyObject_GenericSetAttr,/*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2File__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
(getiterfunc)BZ2File_getiter, /*tp_iter*/
|
|
(iternextfunc)BZ2File_iternext, /*tp_iternext*/
|
|
BZ2File_methods, /*tp_methods*/
|
|
0, /*tp_members*/
|
|
BZ2File_getset, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2File_init, /*tp_init*/
|
|
PyType_GenericAlloc, /*tp_alloc*/
|
|
PyType_GenericNew, /*tp_new*/
|
|
PyObject_Free, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2Comp. */
|
|
|
|
PyDoc_STRVAR(BZ2Comp_compress__doc__,
|
|
"compress(data) -> string\n\
|
|
\n\
|
|
Provide more data to the compressor object. It will return chunks of\n\
|
|
compressed data whenever possible. When you've finished providing data\n\
|
|
to compress, call the flush() method to finish the compression process,\n\
|
|
and return what is left in the internal buffers.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Comp_compress(BZ2CompObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
PY_LONG_LONG totalout;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#:compress", &data, &datasize))
|
|
return NULL;
|
|
|
|
if (datasize == 0)
|
|
return PyBytes_FromStringAndSize("", 0);
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"this object was already flushed");
|
|
goto error;
|
|
}
|
|
|
|
ret = PyBytes_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_RUN);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror != BZ_RUN_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_in == 0)
|
|
break; /* no more input data */
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (PyBytes_Resize(ret, bufsize) < 0) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
goto error;
|
|
}
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (PyBytes_Resize(ret,
|
|
(Py_ssize_t)(BZS_TOTAL_OUT(bzs) - totalout)) < 0)
|
|
goto error;
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2Comp_flush__doc__,
|
|
"flush() -> string\n\
|
|
\n\
|
|
Finish the compression process and return what is left in internal buffers.\n\
|
|
You must not use the compressor object after calling this method.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Comp_flush(BZ2CompObject *self)
|
|
{
|
|
int bufsize = SMALLCHUNK;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
PY_LONG_LONG totalout;
|
|
int bzerror;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_ValueError, "object was already "
|
|
"flushed");
|
|
goto error;
|
|
}
|
|
self->running = 0;
|
|
|
|
ret = PyBytes_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_FINISH);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_FINISH_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (PyBytes_Resize(ret, bufsize) < 0)
|
|
goto error;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0) {
|
|
if (PyBytes_Resize(ret,
|
|
(Py_ssize_t)(BZS_TOTAL_OUT(bzs) - totalout)) < 0)
|
|
goto error;
|
|
}
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
static PyMethodDef BZ2Comp_methods[] = {
|
|
{"compress", (PyCFunction)BZ2Comp_compress, METH_VARARGS,
|
|
BZ2Comp_compress__doc__},
|
|
{"flush", (PyCFunction)BZ2Comp_flush, METH_NOARGS,
|
|
BZ2Comp_flush__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2Comp_Type. */
|
|
|
|
static int
|
|
BZ2Comp_init(BZ2CompObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int compresslevel = 9;
|
|
int bzerror;
|
|
static char *kwlist[] = {"compresslevel", 0};
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|i:BZ2Compressor",
|
|
kwlist, &compresslevel))
|
|
return -1;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
goto error;
|
|
}
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock) {
|
|
PyErr_SetString(PyExc_MemoryError, "unable to allocate lock");
|
|
goto error;
|
|
}
|
|
#endif
|
|
|
|
memset(&self->bzs, 0, sizeof(bz_stream));
|
|
bzerror = BZ2_bzCompressInit(&self->bzs, compresslevel, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->running = 1;
|
|
|
|
return 0;
|
|
error:
|
|
#ifdef WITH_THREAD
|
|
if (self->lock) {
|
|
PyThread_free_lock(self->lock);
|
|
self->lock = NULL;
|
|
}
|
|
#endif
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2Comp_dealloc(BZ2CompObject *self)
|
|
{
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
BZ2_bzCompressEnd(&self->bzs);
|
|
Py_Type(self)->tp_free((PyObject *)self);
|
|
}
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2Comp_Type definition. */
|
|
|
|
PyDoc_STRVAR(BZ2Comp__doc__,
|
|
"BZ2Compressor([compresslevel=9]) -> compressor object\n\
|
|
\n\
|
|
Create a new compressor object. This object may be used to compress\n\
|
|
data sequentially. If you want to compress data in one shot, use the\n\
|
|
compress() function instead. The compresslevel parameter, if given,\n\
|
|
must be a number between 1 and 9.\n\
|
|
");
|
|
|
|
static PyTypeObject BZ2Comp_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"bz2.BZ2Compressor", /*tp_name*/
|
|
sizeof(BZ2CompObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2Comp_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
PyObject_GenericGetAttr,/*tp_getattro*/
|
|
PyObject_GenericSetAttr,/*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2Comp__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
BZ2Comp_methods, /*tp_methods*/
|
|
0, /*tp_members*/
|
|
0, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2Comp_init, /*tp_init*/
|
|
PyType_GenericAlloc, /*tp_alloc*/
|
|
PyType_GenericNew, /*tp_new*/
|
|
PyObject_Free, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Members of BZ2Decomp. */
|
|
|
|
#undef OFF
|
|
#define OFF(x) offsetof(BZ2DecompObject, x)
|
|
|
|
static PyMemberDef BZ2Decomp_members[] = {
|
|
{"unused_data", T_OBJECT, OFF(unused_data), READONLY},
|
|
{NULL} /* Sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2Decomp. */
|
|
|
|
PyDoc_STRVAR(BZ2Decomp_decompress__doc__,
|
|
"decompress(data) -> string\n\
|
|
\n\
|
|
Provide more data to the decompressor object. It will return chunks\n\
|
|
of decompressed data whenever possible. If you try to decompress data\n\
|
|
after the end of stream is found, EOFError will be raised. If any data\n\
|
|
was found after the end of stream, it'll be ignored and saved in\n\
|
|
unused_data attribute.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Decomp_decompress(BZ2DecompObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
PY_LONG_LONG totalout;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#:decompress", &data, &datasize))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_EOFError, "end of stream was "
|
|
"already found");
|
|
goto error;
|
|
}
|
|
|
|
ret = PyBytes_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzDecompress(bzs);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
if (bzs->avail_in != 0) {
|
|
Py_DECREF(self->unused_data);
|
|
self->unused_data =
|
|
PyBytes_FromStringAndSize(bzs->next_in,
|
|
bzs->avail_in);
|
|
}
|
|
self->running = 0;
|
|
break;
|
|
}
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_in == 0)
|
|
break; /* no more input data */
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (PyBytes_Resize(ret, bufsize) < 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
goto error;
|
|
}
|
|
bzs->next_out = BUF(ret);
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0) {
|
|
if (PyBytes_Resize(ret,
|
|
(Py_ssize_t)(BZS_TOTAL_OUT(bzs) - totalout)) < 0)
|
|
goto error;
|
|
}
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
static PyMethodDef BZ2Decomp_methods[] = {
|
|
{"decompress", (PyCFunction)BZ2Decomp_decompress, METH_VARARGS, BZ2Decomp_decompress__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2Decomp_Type. */
|
|
|
|
static int
|
|
BZ2Decomp_init(BZ2DecompObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, ":BZ2Decompressor"))
|
|
return -1;
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock) {
|
|
PyErr_SetString(PyExc_MemoryError, "unable to allocate lock");
|
|
goto error;
|
|
}
|
|
#endif
|
|
|
|
self->unused_data = PyBytes_FromStringAndSize("", 0);
|
|
if (!self->unused_data)
|
|
goto error;
|
|
|
|
memset(&self->bzs, 0, sizeof(bz_stream));
|
|
bzerror = BZ2_bzDecompressInit(&self->bzs, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->running = 1;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
#ifdef WITH_THREAD
|
|
if (self->lock) {
|
|
PyThread_free_lock(self->lock);
|
|
self->lock = NULL;
|
|
}
|
|
#endif
|
|
Py_CLEAR(self->unused_data);
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2Decomp_dealloc(BZ2DecompObject *self)
|
|
{
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
Py_XDECREF(self->unused_data);
|
|
BZ2_bzDecompressEnd(&self->bzs);
|
|
Py_Type(self)->tp_free((PyObject *)self);
|
|
}
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2Decomp_Type definition. */
|
|
|
|
PyDoc_STRVAR(BZ2Decomp__doc__,
|
|
"BZ2Decompressor() -> decompressor object\n\
|
|
\n\
|
|
Create a new decompressor object. This object may be used to decompress\n\
|
|
data sequentially. If you want to decompress data in one shot, use the\n\
|
|
decompress() function instead.\n\
|
|
");
|
|
|
|
static PyTypeObject BZ2Decomp_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"bz2.BZ2Decompressor", /*tp_name*/
|
|
sizeof(BZ2DecompObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2Decomp_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
PyObject_GenericGetAttr,/*tp_getattro*/
|
|
PyObject_GenericSetAttr,/*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2Decomp__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
BZ2Decomp_methods, /*tp_methods*/
|
|
BZ2Decomp_members, /*tp_members*/
|
|
0, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2Decomp_init, /*tp_init*/
|
|
PyType_GenericAlloc, /*tp_alloc*/
|
|
PyType_GenericNew, /*tp_new*/
|
|
PyObject_Free, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Module functions. */
|
|
|
|
PyDoc_STRVAR(bz2_compress__doc__,
|
|
"compress(data [, compresslevel=9]) -> string\n\
|
|
\n\
|
|
Compress data in one shot. If you want to compress data sequentially,\n\
|
|
use an instance of BZ2Compressor instead. The compresslevel parameter, if\n\
|
|
given, must be a number between 1 and 9.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
bz2_compress(PyObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int compresslevel=9;
|
|
char *data;
|
|
int datasize;
|
|
int bufsize;
|
|
PyObject *ret = NULL;
|
|
bz_stream _bzs;
|
|
bz_stream *bzs = &_bzs;
|
|
int bzerror;
|
|
static char *kwlist[] = {"data", "compresslevel", 0};
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "s#|i",
|
|
kwlist, &data, &datasize,
|
|
&compresslevel))
|
|
return NULL;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
return NULL;
|
|
}
|
|
|
|
/* Conforming to bz2 manual, this is large enough to fit compressed
|
|
* data in one shot. We will check it later anyway. */
|
|
bufsize = datasize + (datasize/100+1) + 600;
|
|
|
|
ret = PyBytes_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
return NULL;
|
|
|
|
memset(bzs, 0, sizeof(bz_stream));
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
bzerror = BZ2_bzCompressInit(bzs, compresslevel, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_FINISH);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_FINISH_OK) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (PyBytes_Resize(ret, bufsize) < 0) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
bzs->next_out = BUF(ret) + BZS_TOTAL_OUT(bzs);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0) {
|
|
if (PyBytes_Resize(ret, (Py_ssize_t)BZS_TOTAL_OUT(bzs)) < 0) {
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
}
|
|
}
|
|
BZ2_bzCompressEnd(bzs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(bz2_decompress__doc__,
|
|
"decompress(data) -> decompressed data\n\
|
|
\n\
|
|
Decompress data in one shot. If you want to decompress data sequentially,\n\
|
|
use an instance of BZ2Decompressor instead.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
bz2_decompress(PyObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
PyObject *ret;
|
|
bz_stream _bzs;
|
|
bz_stream *bzs = &_bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#:decompress", &data, &datasize))
|
|
return NULL;
|
|
|
|
if (datasize == 0)
|
|
return PyBytes_FromStringAndSize("", 0);
|
|
|
|
ret = PyBytes_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
return NULL;
|
|
|
|
memset(bzs, 0, sizeof(bz_stream));
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
bzerror = BZ2_bzDecompressInit(bzs, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzDecompress(bzs);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
if (bzs->avail_in == 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"couldn't find end of stream");
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (PyBytes_Resize(ret, bufsize) < 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
bzs->next_out = BUF(ret) + BZS_TOTAL_OUT(bzs);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0) {
|
|
if (PyBytes_Resize(ret, (Py_ssize_t)BZS_TOTAL_OUT(bzs)) < 0) {
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
}
|
|
}
|
|
BZ2_bzDecompressEnd(bzs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static PyMethodDef bz2_methods[] = {
|
|
{"compress", (PyCFunction) bz2_compress, METH_VARARGS|METH_KEYWORDS,
|
|
bz2_compress__doc__},
|
|
{"decompress", (PyCFunction) bz2_decompress, METH_VARARGS,
|
|
bz2_decompress__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
/* ===================================================================== */
|
|
/* Initialization function. */
|
|
|
|
PyDoc_STRVAR(bz2__doc__,
|
|
"The python bz2 module provides a comprehensive interface for\n\
|
|
the bz2 compression library. It implements a complete file\n\
|
|
interface, one shot (de)compression functions, and types for\n\
|
|
sequential (de)compression.\n\
|
|
");
|
|
|
|
PyMODINIT_FUNC
|
|
initbz2(void)
|
|
{
|
|
PyObject *m;
|
|
|
|
Py_Type(&BZ2File_Type) = &PyType_Type;
|
|
Py_Type(&BZ2Comp_Type) = &PyType_Type;
|
|
Py_Type(&BZ2Decomp_Type) = &PyType_Type;
|
|
|
|
m = Py_InitModule3("bz2", bz2_methods, bz2__doc__);
|
|
if (m == NULL)
|
|
return;
|
|
|
|
PyModule_AddObject(m, "__author__", PyUnicode_FromString(__author__));
|
|
|
|
Py_INCREF(&BZ2File_Type);
|
|
PyModule_AddObject(m, "BZ2File", (PyObject *)&BZ2File_Type);
|
|
|
|
Py_INCREF(&BZ2Comp_Type);
|
|
PyModule_AddObject(m, "BZ2Compressor", (PyObject *)&BZ2Comp_Type);
|
|
|
|
Py_INCREF(&BZ2Decomp_Type);
|
|
PyModule_AddObject(m, "BZ2Decompressor", (PyObject *)&BZ2Decomp_Type);
|
|
}
|