mirror of
https://github.com/python/cpython.git
synced 2025-08-30 13:38:43 +00:00
Issue #25626: Change zlib to accept Py_ssize_t and cap to UINT_MAX
The underlying zlib library stores sizes in “unsigned int”. The corresponding Python parameters are all sizes of buffers filled in by zlib, so it is okay to reduce higher values to the UINT_MAX internal cap. OverflowError is still raised for sizes that do not fit in Py_ssize_t. Sizes are now limited to Py_ssize_t rather than unsigned long, because Python byte strings cannot be larger than Py_ssize_t. Previously this could result in a SystemError on 32-bit platforms. This resolves a regression in the gzip module when reading more than UINT_MAX or LONG_MAX bytes in one call, introduced by revision 62723172412c.
This commit is contained in:
parent
d13cade381
commit
e99e97762c
7 changed files with 120 additions and 47 deletions
|
@ -3,6 +3,7 @@
|
|||
|
||||
import unittest
|
||||
from test import support
|
||||
from test.support import bigmemtest, _4G
|
||||
import os
|
||||
import io
|
||||
import struct
|
||||
|
@ -116,6 +117,14 @@ class TestGzip(BaseTest):
|
|||
self.assertEqual(f.tell(), nread)
|
||||
self.assertEqual(b''.join(blocks), data1 * 50)
|
||||
|
||||
@bigmemtest(size=_4G, memuse=1)
|
||||
def test_read_large(self, size):
|
||||
# Read chunk size over UINT_MAX should be supported, despite zlib's
|
||||
# limitation per low-level call
|
||||
compressed = gzip.compress(data1, compresslevel=1)
|
||||
f = gzip.GzipFile(fileobj=io.BytesIO(compressed), mode='rb')
|
||||
self.assertEqual(f.read(size), data1)
|
||||
|
||||
def test_io_on_closed_object(self):
|
||||
# Test that I/O operations on closed GzipFile objects raise a
|
||||
# ValueError, just like the corresponding functions on file objects.
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue